Work-in-Progress CHI 2015, Crossings, Seoul, Korea CogniMouse: On Detecting Users’ Task Completion Difficulty through Computer Mouse Interaction Marios Belk Eleni Christodoulou Abstract Dept. of Computer Science CITARD Services Ltd. University of Cyprus CY-2064 Nicosia, Cyprus CY-1678 Nicosia, Cyprus cseleni@citard-serv.com This paper presents a method for analyzing users’ computer mouse interaction data with the aim to implicitly identify task completion difficulty while interacting with a system. Computer mouse motion streams and users’ skin conductance signals, acquired via an in-house developed computer mouse, and users’ feedback were investigated as reactions to task difficulty raising events. A classification algorithm was developed, producing real-time user models of user hesitation states. Preliminary results of a study in progress with seven older adults at work (age 56+) provide initial indications about links between mouse triggering states of user hesitation and task completion difficulty. belk@cs.ucy.ac.cy George Samaras David Portugal Dept. of Computer Science CITARD Services Ltd. University of Cyprus CY-2064 Nicosia, Cyprus CY-1678 Nicosia, Cyprus davidsp@citard-serv.com cssamara@cs.ucy.ac.cy Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author. Copyright is held by the owner/author(s). CHI'15 Extended Abstracts, Apr 18-23, 2015, Seoul, Republic of Korea ACM 978-1-4503-3146-3/15/04. http://dx.doi.org/10.1145/2702613.2732881 Author Keywords Computer Mouse; Sensors; Older Adults; User Study. ACM Classification Keywords H.5.2. Information interfaces and presentation (e.g., HCI): User Interfaces. Introduction With the advent of highly dynamic and fast emerging technologies and software, users are required to adapt 1019 Work-in-Progress CHI 2015, Crossings, Seoul, Korea their mental models, skills and habitual ways of working. This requirement is further intensified, when combined with eventual age-related cognitive degradations of older adults [1], causing task completion difficulty and eventually a negative user experience. Assisting older adults while interacting with systems is of critical importance in today’s information society. Researchers and practitioners alike have shown an increased interest lately in understanding behavior patterns and possible difficulties of older adults imposed by current visual and interaction designs of interactive systems [2, 3]. A number of research works exist that proposed various intelligent user interfaces and systems for supporting older adults at work and motivating them to stay for longer active and productive in computerized working environments [4, 5]. An important challenge of such intelligent and assistive interactive systems is how to implicitly identify that users experience difficulties with a given task [6], and accordingly assist the users, e.g. by providing personalized and contextualized support [7]. In this context, this paper presents an implicit user data collection method for identifying users’ task completion difficulty by leveraging computer mouse motion streams and a skin conductance sensor that is embedded in an in-house developed computer mouse. This work is part of CogniWin project, which aims to provide an innovative personalized system, motivating older adults to stay for longer active, and improving their productivity in the workplace. The paper is structured as follows: next we present related works, and subsequently we describe the developed computer mouse, named CogniMouse. Afterwards, we present the methodology and preliminary results of a user study that evaluated the accuracy of the computer mouse’s task difficulty identification method. We conclude with practical implications and future prospects of this work. Related Work The literature proposes several systems that leverage users’ computer mouse interaction data in order to implicitly infer important information about the users’ behavior patterns. Principally, the raw mouse data being extracted and processed includes time-stamped and chronologically ordered sequences of a stream of data including the x-y cursor position of the mouse on the screen, mouse hover, mouse scrolling and mouse click events (left and right clicks) [8-11]. Research works have proposed various approaches for distinguishing users’ behavior patterns, such as user hesitation, reading by tracing text, clicking and scrolling [8, 9, 10]. Reeder and Maxion [8] proposed an automated method for identifying user hesitation with the aim to detect instances of user difficulty while interacting with a system. Based on a hesitation detector algorithm that takes as input a time-stamped, chronologically ordered data stream of mouse and keyboard events, user hesitation was defined as anomalously long pauses between events in a given data stream. Arroyo et al. [9] proposed a Web logging system that tracks mouse movements in Websites illustrating mouse trajectories of users’ interactions that indicate users’ reading behavior and hesitation in the menu area. A preliminary study of Mueller and Lockerd [10] revealed that users often move the mouse cursor to an empty white space of the Webpage in case they are hesitating in order to avoid accidental clicks on hyperlinks. 1020 Work-in-Progress CHI 2015, Crossings, Seoul, Korea Unlike the aforementioned works, we propose an innovative instrumented computer mouse for advanced user interface sensing; as well as a new technique for detecting user hesitation with sensor fusion, which is not solely based on input pauses, but instead on realtime computer mouse data analysis, which benefits from detecting when the user is actually touching the device. Moreover, we validate preliminarily our system by conducting a user study, which indicates potential links between user hesitation state and task completion difficulty. programming language, which has been chosen for several reasons: i) it provides a good tradeoff between efficiency, security and robustness; and ii) it provides a number of libraries for communication with hardware components, e.g. HID support. User Hesitation Classification Delay is a common consequence of users having difficulty with a particular task [8]. Accordingly, the developed method detects the level of difficulty in completion of tasks by the user, through the detection of hesitating behavior. The CogniMouse Architecture Figure 1. The CogniMouse architecture. To date, the main focus has been placed in the development of a classifier of user’s hesitation state according to the GSR sensor input, and mouse motion streams. In this context, a parser module receives and processes the raw data coming from the mouse sensors, which is then analyzed by the classifier algorithm. We have implemented a Bayesian Classifier to detect a general measurement of user’s hesitation. Different classification techniques could be employed [12], however the classification algorithm highly depends on the properties of the specific problem to tackle, and the application of classification methods in human state prediction from multiple sources of data has different number of features and challenges. Assuming the availability of user historical data and common interval of parameters which are deemed as usual for each individual, human expertise judgment beforehand can be leveraged to model relevant data, with an important impact on the inference process. This is the case of Bayesian inference, which makes sequential use of the Bayes’ formula, when more data becomes available, to calculate a posterior distribution. This provides means to make inference about an environment of interest described by a state given an observation, whose relationship is encoded by a joint probability distribution. Following the philosophy described above, CogniMouse is currently being developed using C# object-oriented Specifically, our model represents the probability or the level of certainty that the user is having difficulty in CogniMouse is a plug-and play Human Interface Device (HID) that connects to any host computer via a Universal Serial Bus (USB) as seen in Figure 1. Having driver support on all major operating systems, a custom HID protocol has been developed, which allows sending 64 bytes of different sensor data per packet. On the mouse buttons area, a transparent Galvanic Skin Response (GSR) sensor has been embedded to react to changes on the user’s skin response. From the outside, the current prototype design makes use of a commercial Microsoft Comfort Mouse 4500 (cf. Figure 2), which causes no barrier to the users’ acceptance. 1021 Work-in-Progress CHI 2015, Crossings, Seoul, Korea completing the task, given by P(Hes|ZC,V,GSR). To that end, we used three different inputs in our Bayesian Classifier: i) the number of zero-crossings of the motion vector over the last 4 seconds of mouse usage (ZC); ii) the average intensity of the velocity vector over the last 4 seconds of mouse usage (V); and iii) the galvanic skin response value (GSR). The zero-crossings’ input represents the number of times that the user changed motion direction when using the mouse. Since we are developing a general model and no personalized data of the user is available at this stage, we assume that there is no prior information about the user. Thus the prior distribution, P(Hes), is defined as uniform, where all decisions are equiprobable. Additionally, we need to define the likelihood functions, P(input|Hes), so as to model each input according to the expected hesitation. The hesitation state of the user is obtained by fusing all the inputs by applying Bayes Formula in each iteration of the following algorithm: Figure 2. The current prototype design of the computer mouse makes use of a commercial Microsoft Comfort Mouse 4500. | , , ∙ | | ∙ ∙ ∙ | ∙ Method of Study A user study was conducted with the aim to initially evaluate the accuracy of the user hesitation classification algorithm. Procedure Controlled laboratory sessions were conducted in which participants were required to perform a series of tasks (i.e., change settings of a Web browser) on a standard desktop computer (IBM Thinkcenter M73, 21’’ monitor), using a standard keyboard and the developed computer mouse which recorded the mouse motion stream data and users’ skin conductance signals. Screen capturing software and audio data based on the think-aloud protocol were recorded which were later analyzed by a user experience researcher aiming to identify users’ task completion difficulty. This information was critical since it serves as an evaluation basis for the user hesitation triggering events of CogniMouse. A pre-study questionnaire was also provided to the participants to rate their experience with computers, the frequency of using the Web browser in their daily work, and the frequency they use Web browser tools, i.e. General Settings, Privacy, Security and Advanced Settings. Users’ Task Completion Difficulty Identification Based on the screen capturing software and the thinkaloud audio data, users’ task completion difficulty while performing a particular task was empirically detected and evaluated based on the following criteria: i) user statements (e.g. “I don’t know …”); ii) user silence and inactivity of computer mouse and keyboard for a considerable period of time; and iii) user asks for help from the user experience researcher. Periods of task completion difficulty are noted as soon as one of the above criteria are met until the user clearly performs an action (e.g. mouse click) indicating that the user is not in a difficulty state anymore, or when the user states that he/she is going to perform an action or that the difficulty has been overcome (e.g. “OK, I will …”). Tasks All participants were required to complete three tasks within a Web browser (Google Chrome v.39). The following tasks were provided in a random sequence to all participants, which primarily require changing the 1022 Work-in-Progress CHI 2015, Crossings, Seoul, Korea Web browser’s settings: i) “Set your preferred Website as your Web browser’s start page”; ii) “Set the search engine to be used when searching from the Web browser’s search box (omnibox)”; iii) “Set your Web browser’s cookie policy to block Websites from writing any data on your computer”. These tasks were chosen since they are performed by the users using primarily the computer mouse, and require limited keyboard typing input. Figure 3. Means of hesitation values per user and task. Figure 4. Means of hesitation values per task and user group. Data Analysis Two types of data were analyzed: i) user hesitation output, triggered by the CogniMouse user hesitation classification algorithm; and ii) task completion difficulty as noted by the user experience researcher based on the screen capturing software and the thinkaloud audio data analysis. Participants A total of seven individuals (4 male and 3 female) participated in the study, and their age varied between 56 and 67 (mean age 62). All participants are using computers and accessing Websites on a daily basis. Based on the pre-study questionnaire, four participants consider themselves as average to experienced, while three consider themselves as novice to average in regards with knowledge and experience with computers. All participants use the Web browser for the daily work activities. In regards with frequency of handling the settings of the Web browser, three participants rated themselves as experienced, whereas the remaining four as novice. Analysis of Results For our analysis we have separated users into two groups: the High-difficulty Group that includes four participants that had significant task completion difficulties while performing all the three tasks, and the Low-difficulty Group that includes three participants that did not have any major task completion difficulties while performing the tasks. The categorization into these groups was based solely on the empirical analysis made based on the screen capture data and the thinkaloud audio analysis. Figure 3 illustrates the means of user hesitation values per user and task; High-difficulty Group includes User #1-4, and the Low-difficulty Group includes User #5-7. Figure 4 illustrates the means of user hesitation values per task and user group. An independent-samples t-test was run to determine if there were differences in user hesitation values between the High-difficulty and Low-difficulty user groups. There was homogeneity of variances, as assessed by Levene's test for equality of variances (p=0.475). Results revealed that users of the Highdifficulty Group triggered higher user hesitation scores (0.65±0.045) than those of the Low-difficulty Group (0.57±0.026). These were statistically significant different at 0.082 (t(5)=2.757, p=0.04, d=0.21). Furthermore, descriptive statistics reveal that users of the Low-difficulty Group scored lower user hesitation values in all three tasks (Figure 4). Conclusions and Future Work The purpose of this paper is to present results of a work in progress that aims to implicitly identify users’ task completion difficulty by leveraging computer mouse motion data and users’ skin conductance signals. A user study was conducted with seven older adults to investigate the accuracy of a user hesitation classification algorithm. Empirical data, based on screen capture information and think-aloud audio data, 1023 Work-in-Progress CHI 2015, Crossings, Seoul, Korea revealed that users having difficulty with particular tasks scored higher user hesitation values compared to users that did not face any major task completion difficulties. Although the analysis yielded statistical significant results, these results shall be confirmed in further studies with a larger sample and users with varying profiles and ages. Finally, we believe that even though this study is focused on older adults, in general, identifying task completion difficulty through the computer mouse could be suitable for other age groups. In the future, we intend to contextualize the mouse data so that the system provides adaptive support taking the user’s task into account, when difficulty is identified. Moreover, further sensors shall be embedded in CogniMouse to detect additional user behaviors, as well as information from other devices, such as an eye tracker, which will be integrated in order to trigger personalized assistance. Acknowledgements This work was partially carried out in the frame of the CogniWin project, funded by the EU Ambient Assisted Living Joint Program (AAL 2013-6-114) as well as the EU FP7 project Miraculous-Life (#611421) and the Cyprus Research Promotion Foundation project PersonaWeb (ΤΠΕ/ΠΛΗΡΟ/0311(ΒΙΕ)/10). References [1] Kausler, D.H., Salthouse, T.A., Saults, J.S. Temporal memory over the adult lifespan. The American Journal of Psychology 101, 2, 207-215. [2] Findlater, L., Froehlich, J., Fattal, K., Wobbrock, J., Dastyar, T. Age-related differences in performance with touchscreens compared to traditional mouse input. In Proc. CHI 2013, ACM Press (2013), 343-346. [3] Waycott, J., Vetere, F., Pedell, S. Designing communication technologies for socially isolated older people: Fieldwork reflections. In Proc. CHI 2013, Workshop on Designing for -and with- Vulnerable People, ACM Press. [4] Leung, R., Findlater, L., McGrenere, J., Graf, P., Yang, J. Multi-layered interfaces to improve older adults’ initial learnability of mobile applications. ACM Trans. Access. Comput. 3, 1, Article 1 (2010), 30 p. [5] Lindsay, S., Jackson, D., Schofield, G., Olivier, P. Engaging older people using participatory design. In Proc. CHI 2012, ACM Press (2012), 1199-1208. [6] Atterer, R., Wnuk, M., Schmidt, A. Knowing the user’s every move: user activity tracking for website usability evaluation and implicit interaction. In Proc. WWW 2006, ACM Press (2006), 203-212. [7] Brusilovsky, P., Kobsa, A., Nejdl, W. The adaptive web: methods and strategies of web personalization. Springer Verlag, Heidelberg, 2007. [8] Reeder, R. Maxion, R. User interface defect detection by hesitation analysis. In Proc. DSN 2006, IEEE Computer Society (2006), 61-72. [9] Arroyo, E., Selker, T., Wei, W. Usability tool for analysis of web designs using mouse tracks. Ext. Abstracts CHI 2006, ACM Press (2006), 484–489. [10] Mueller, F., Lockerd, A. Cheese: Tracking mouse movement activity on websites, a tool for user modeling. Ext. Abstracts CHI 01, ACM (2001), 279-280. [11] Huang, J., White, R., Dumais, S. No clicks, no problem: using cursor movements to understand and improve search. In Proc. CHI 2011, ACM Press (2011), 1225-1234. [12] Luz, J.M., Couceiro, M., Portugal, D., Rocha, R.P., Araújo, H., Dias, G. Comparison of classification methods for golf putting performance analysis. Comp. Intel. and Dec. Mak. 61, 35-45. 1024
© Copyright 2024