Bitte benutzen Sie diese Kennung, um auf die Ressource zu verweisen:
http://dx.doi.org/10.18419/opus-14707
Langanzeige der Metadaten
DC Element | Wert | Sprache |
---|---|---|
dc.contributor.author | Wang, Yao | - |
dc.contributor.author | Jiang, Yue | - |
dc.contributor.author | Hu, Zhiming | - |
dc.contributor.author | Ruhdorfer, Constantin | - |
dc.contributor.author | Bâce, Mihai | - |
dc.contributor.author | Bulling, Andreas | - |
dc.date.accessioned | 2024-07-24T14:51:04Z | - |
dc.date.available | 2024-07-24T14:51:04Z | - |
dc.date.issued | 2024 | de |
dc.identifier.issn | 2573-0142 | - |
dc.identifier.uri | http://nbn-resolving.de/urn:nbn:de:bsz:93-opus-ds-147265 | de |
dc.identifier.uri | http://elib.uni-stuttgart.de/handle/11682/14726 | - |
dc.identifier.uri | http://dx.doi.org/10.18419/opus-14707 | - |
dc.description.abstract | Question answering has recently been proposed as a promising means to assess the recallability of information visualisations. However, prior works are yet to study the link between visually encoding a visualisation in memory and recall performance. To fill this gap, we propose VisRecall++ - a novel 40-participant recallability dataset that contains gaze data on 200 visualisations and five question types, such as identifying the title, and finding extreme values.We measured recallability by asking participants questions after they observed the visualisation for 10 seconds.Our analyses reveal several insights, such as saccade amplitude, number of fixations, and fixation duration significantly differ between high and low recallability groups.Finally, we propose GazeRecallNet - a novel computational method to predict recallability from gaze behaviour that outperforms several baselines on this task.Taken together, our results shed light on assessing recallability from gaze behaviour and inform future work on recallability-based visualisation optimisation. | en |
dc.language.iso | en | de |
dc.relation.uri | doi:10.1145/3655613 | de |
dc.rights | info:eu-repo/semantics/openAccess | de |
dc.subject.ddc | 004 | de |
dc.title | VisRecall++: analysing and predicting visualisation recallability from gaze behaviour | en |
dc.type | conferenceObject | de |
ubs.fakultaet | Informatik, Elektrotechnik und Informationstechnik | de |
ubs.fakultaet | Fakultätsübergreifend / Sonstige Einrichtung | de |
ubs.institut | Institut für Visualisierung und Interaktive Systeme | de |
ubs.institut | Fakultätsübergreifend / Sonstige Einrichtung | de |
ubs.konferenzname | ACM Symposium on Eye Tracking Research & Applications (2024, Glasgow) | de |
ubs.publikation.noppn | yes | de |
ubs.publikation.seiten | 18 | de |
ubs.publikation.source | Proceedings of the ACM on Human-Computer Interaction 8 (2024), issue ETRA | de |
ubs.publikation.typ | Konferenzbeitrag | de |
Enthalten in den Sammlungen: | 05 Fakultät Informatik, Elektrotechnik und Informationstechnik |
Dateien zu dieser Ressource:
Datei | Beschreibung | Größe | Format | |
---|---|---|---|---|
wang24_etra.pdf | 2,37 MB | Adobe PDF | Öffnen/Anzeigen |
Alle Ressourcen in diesem Repositorium sind urheberrechtlich geschützt.