Experienced mental workload, perception of usability, their interaction and impact on task performance | PLOS ONE
Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Experienced mental workload, perception of usability, their interaction and impact on task performance

  • Luca Longo

    Roles Investigation, Methodology, Writing – original draft, Writing – review & editing

    luca.longo@dit.ie

    Affiliations School of Computing, College of health and sciences, Dublin Institute of Technology, Dublin, Ireland, ADAPT: The global centre of excellence for digital content and media innovation, Dublin, Ireland

Abstract

Past research in HCI has generated a number of procedures for assessing the usability of interacting systems. In these procedures there is a tendency to omit characteristics of the users, aspects of the context and peculiarities of the tasks. Building a cohesive model that incorporates these features is not obvious. A construct greatly invoked in Human Factors is human Mental Workload. Its assessment is fundamental for predicting human performance. Despite the several uses of Usability and Mental Workload, not much has been done to explore their relationship. This empirical research focused on I) the investigation of such a relationship and II) the investigation of the impact of the two constructs on human performance. A user study was carried out with participants executing a set of information-seeking tasks over three popular web-sites. A deep correlation analysis of usability and mental workload, by task, by user and by classes of objective task performance was done (I). A number of Supervised Machine Learning techniques based upon different learning strategy were employed for building models aimed at predicting classes of task performance (II). Findings strongly suggests that usability and mental workload are two non overlapping constructs and they can be jointly employed to greatly improve the prediction of human performance.

Introduction

In recent years, with the advent of the Internet and the explosion of web-based system development, the construct of usability has been invoked in many different ways. Research in the past decades has generated a number of procedures for assessing the usability of interactive systems. It is believed it is a multi-dimensional construct, encompassing several features. Frequently, for example, during usability inspection, there is a tendency to omit characteristics of the users, aspects of the context and peculiarities of the tasks. This tendency is reasonable and it justified by the complexity of usability as a construct and a lack of a model that unifies all of these factors. Taking into account features of users is fundamental for the User Modeling community in order to build systems that fit the specific background, knowledge and objectives of users [13]. Similarly, considering the context of use has a significant influence in the inference of meaningful assessments of usability [47]. Additionally, during the usability inspection process, accounting for the demands of the underlying task is essential for predicting user experience and informing the design of interactive systems [8, 9]. Building a cohesive model that incorporate user, context and task-specific factors is not obvious. Usability inspection should be accompanied by the assessment of one of all of these factors when possible. Beside Usability, another construct has a long research history in the field of Human Factors: the construct of human mental workload (MWL) [10, 11]. This is often referred to as cognitive load and I believe this can significantly contribute to the goal of informing interaction and web-design. MWL, with roots in Psychology, has been mainly adopted within the fields of Ergonomics with several application in the transportation and nuclear industries [12]. Its assessment is fundamental for predicting performance, which in turn is key for describing user experience and engagement. The link usability and mental workload is nowadays under explored. A few studies have attempted to apply the construct of MWL to explain usability [1318]. Despite this weak interest, not much has yet been done to explore their relationship empirically. The aim of this research is to empirically investigate the relationship between subjective perception of usability and mental workload with a particular focus on their impact on objective user performance, this being assessed through observation of tangible facts. Fig 1 depicts the main constructs employed in this research study and their relationship.

The remainder of this paper is divided into the following sections. Firstly, noteworthy definitions of usability and mental workload are provided, followed by an overview of the assessment techniques employed in Human-Computer Interaction (HCI). Related work at the intersection of usability and mental workload is presented, describing how the two constructs have been employed individually and conjointly. An empirical experiment is subsequently defined in the context of human-web interaction. This is aimed at exploring the relationship between the perception of usability of three popular web-sites (youtube, wikipedia and google) and the mental workload experienced by human participants after engaging with them. Results are presented and critically examined, demonstrating how these constructs are related and how they influence objective user performance. Findings are strengthen by measures of the reliability of the underlying instruments adopted. A summary concludes this paper suggesting future research, highlighting the contribution to knowledge and implications to the field of Human-Computer Interaction.

1 Materials and methods

Widely employed in the broader field of HCI, usability and mental workload are two constructs from the discipline of Human Factor, with no limpid and broad definitions. Since their inception, there has been an intense debate about their assessment and measurement [1921]. Although multidimensional and complex, their usefulness for describing the user experience and informing interaction, interface and system design is beyond doubt.

1.1 Usability

The amount of literature covering definitions of usability, its frameworks and methodologies is significant [2226]. An early definition by Eason [27] was ‘the degree to which users are able to use the system with the skills, knowledge, stereotypes and experience they can bring to bear’. This definition is mainly connected to the ease-of-use, however, there are more factors influencing usability. For example, a definition provided by the ISO (International Organisation for Standardisation), is ‘The extent to which a product can be used by specified users to achieve specified goals with effectiveness, efficiency, and satisfaction in a specified context of use [28]’ (ISO 9241 series, Section 8.1). It is not a single, one-dimensional property of a user interface, rather it is often associated with the functionalities of a product rather than being merely a feature of the user interface [29]. Usability, according to Nielsen [30], is a method for improving ease-of-use in the design of interactive systems and technologies. It embraces other concepts such as efficiency, learnability and satisfaction. Nielsen’s principles are frequently employed to evaluate the usability of interfaces [30]. The evaluation is an iterative process in which usability issues are systematically discovered through the application of his principles [29]. The main limitation associated to these principles is that they are focused on the user interface under examination, overlooking contextual factors, the cognitive and emotional state of the users at the time of usability assessment as well as the complexity or time-pressure of the underlying tasks.

Often when selecting an appropriate usability assessment instrument, it is desirable to consider the effort and expense that will be incurred in collecting and analysing data, as in the context of interaction and web-design. Designers are inclined to adopt subjective usability assessment techniques for rapidly collecting feedback from users [25]. On one hand, these self-reporting techniques can only be administered post-task, thus their reliability is under discussion when done on long tasks. Meta-cognitive limitations can also mitigate the accuracy of self-reporting and thus it is challenging to perform comparisons of different raters adopting an absolute scale. On the other hand, these techniques have demonstrated their appeals because of their sensitivity and their diagnostical capacity [25]. One of this technique is the System Usability Scale [31], a questionnaire that consists of ten questions (Table A1 in S1 Appendix). It is a highly cited usability assessment technique [32]. It is a very easy scale to administer, massively applied in different domaind, showing reliability to distinguishing usable and unusable systems and interfaces both with small and large sample sizes [3335]. Other self-reporting usability assessment techniques include the Questionnaire for User Interface Satisfaction (QUIS) [36], developed at the HCI lab at the University of Maryland, the Computer System Usability Questionnaire (CSUQ) and the Perceived Usefulness and Ease of Use scale (PUEU), both developed at IBM [37, 38]. The former was developed to assess the satisfaction of users with aspects of a computer interface [36]. It includes a demographic questionnaire, a measure of system satisfaction along six scales, as well as a hierarchy of measures of nine specific interface factors. Each of these factors relates to a user’s satisfaction with that particular aspect of an interface and to the factors that make up that facet, on a 9-point scale. The latters are a survey that consists respectively of of 19 questions on a seven-point Likert scale of ‘strongly disagree’ to ‘strongly agree’ [37] and 12 questions, from ‘extremely likely’ to ‘extremely unlikely’ [38]. Although it is more complex than other instruments, QUIS has shown high reliability across several interfaces [39]. Additional usability scales include the Purdue Usability Testing Questionnaire [40] containing 100 questions, the USE questionnaire [41], formed upon 30 questions [41]. Many other usability measures have been proposed and the reader is referred to [25, 35]. Eventually, a recent study suggests that despite the intensive use of the construct of usability in Human-Computer Interaction research, its usefulness to HCI theories as well as our understanding has been meager [42].

1.2 Mental workload

Human Mental Workload (MWL) is a design concept fundamental for exploring the interaction of people with technological devices, interfaces and systems [10, 43, 44]. This construct has a long history in Psychology with several applications in Human Factors, in domains such as transportation [9, 45, 46], safety-critical environments [47, 48], automation and manufacturing [49, 50], medicine and health-care [51, 52]. The principal reason for assessing mental workload is to quantify the cost associated to performing a cognitive task for predicting operator and/or system performance [53, 54]. It has been extensively documented that mental underload and overload can negatively influence performance [55]. On one hand, during information processing, when MWL is at a low level, humans may feel frustrated or often annoyed. On the other hand, when MWL is at a high level, high level of confusions can be reached by individuals with a consequent decrement in their performance while processing information and thus higher chances of making mistakes. Hence, designers who are involved in assessing human or system performance require clues about operator workload at all stages of system design and operation. These clues allow them to explore and evaluate additional design options [9]. On one hand, the difficulty of typical tasks executed on early-stage interactive systems might be initially high, due, for instance, to interface complexity. This is likely to impose high level of mental workload upon operators and thus making them experience low levels of performance [54]. This is translated in higher operator’s response time, more errors and fewer tasks are completed per unit of time [56]. On the other hand, early-stage interactive systems might be designed with simplicity in mind, initially shaping typical tasks that are likely to impose low levels of mental workload upon humans. This situation should be avoided too as it leads to difficulties in maintaining attention and promote increment in reaction time [54] with consequences on user engagement and experience. In summary, at an early design phase, a system/interface can be optimised taking mental workload into consideration, guiding designers in making appropriate structural changes [55].

MWL is not a simple and linear concept. Intuitively, it can be described as as the total cognitive work necessary for a human to accomplish a task over time [57]. It is believed that is is not ‘an elementary property, rather it emerges from the interaction between the requirements of a task, the circumstances under which it is performed and the skills, behaviours and perceptions of the operator’ [9]. This definition is merely practical, and many other factors play a role in mental workload variation. Formalising mental workload as a clear, linear construct is far from being trivial [11, 5862] The area of MWL measurement is as extensive as its several definitions and formalisations. Several assessment techniques have been proposed in the last fifty years. Researchers in applied domains have shaped a tendency towards the use of ad hoc, domain-dependent measure or pool of measures. This trend is justified by the multi-dimensional nature of mental workload. Several reviews attempted to organise the significant amount of knowledge behind measurement procedures [10, 55, 63]. However, three main clusters are believed, by the community of MWL, to represent the main measures [54, 6467]:

  • self-assessment or self-reporting measures;
  • task measures or objective performance measures;
  • physiological measures.

The class of self-assessment measures is often referred to as self-report measures. This category relies upon the subject perceived experience of the interaction with an underlying interactive system through a direct estimation of individual differences such as attitudes, emotional state and level of stress of the operator, the effort devoted to the task and its demands [9, 46, 6870]. It is strongly believed that only the human involved with a task can provide accurate and precise judgements about the MWL experienced. For this reason, self-assessment measures have been always appealing to many practitioners. The class of task performance measures is based on the assumption that the mental workload of an individual, interacting with an underlying system or interface, becomes relevant only if it impacts system performance. Example of measures include reaction time to a secondary task, task completion time, error rate, tapping regularity. In turn, this category appears as the most valuable options for designers [7173]. The category of physiological measures considers responses of the body gathered from the individual interacting with an underlying task and system. These responses are thought to be highly correlated to MWL. Their utility lies in the interpretation and analysis of psychological processes and their effect on the state of the body. Example of these measures include brain function measures, cardiac measures such as hear-rate, eye measures, such as pupil dilation/movement and muscle measures. The advantage behind measures belonging to this category is that they can be collected continuously over time, without demanding an explicit response by the operator [74, 75]. So far they have required specific equipment and trained operators to employ this equipment minimising their employability in real-world tasks [76]. However, this tendency is assisting to a shift thanks to the advances in sensor-based technologies to monitor physiological signals.

1.2.1 A focus on self-assessment measures.

Self-assessment measures of mental workload have in general low implementation requirements, they are often not intrusive and possess high degree of acceptability by end-users [69, 77]. These measures are usually multi-dimensional. Examples include the NASA’s Task Load Index (NASATLX) [9], the Subjective Workload Assessment Technique [70] and the Workload Profile [77]. Uni-dimensional measures also exist such as the the Rating Scale Mental Effort [78], the Copper-Harper scale [79], the Bedford scale [80] and the Subjective Workload Dominance Technique [81]. Among this, the NASATLX is probably the most popular self-reporting MWL technique [9]. This has been used in many empirical studies as for instance, to evaluate user interfaces in health-care or in e-commerce application and for the improvement of user satisfaction [82]. [83] investigated how the design of query interfaces is related to performance and stress during information-seeking tasks. Mental workload was assessed using the NASATLX and log data was used as objective indicator of performance to characterise searching behaviour. The Workload Profile [77], the NASATLX and the Subjective Workload Assessment Technique [70] have been compared in a user study to evaluate different web-based interfaces [84]. In general, these techniques have demonstrated a good internal reliability (Cronbach’s Alpha varying around.80) and external validity [85]. In this research study, the Nasa Task Load Index and the Workload profile techniques have been adopted. These self-reporting techniques are described in details respectively in section 1.4.2 and section 1.4.3.

1.3 Research at the intersection of usability and mental workload

Not a lot of research exist at the intersection of mental workload and usability. O’Brien and collaborators identified mental workload and usability as dimensions of the construct of user engagement, showing these are weakly correlated to each other [15]. A similar view can be found in the proposal presented in [86]. Lehmann et al. highlighted the usefulness of adopting different metrics for assessing user engagement, such as usability and cognitive engagement [14]. In a recent review, it was acknowledge that usability and performance are two key elements for describing user experience [16]. The above work clearly emphasises the usefulness of adopting mental workload with traditional usability assessment methods for explaining user experience and user engagement. The constructs of mental workload and usability have been jointly mentioned in an article to better design e-learning artefacts in medical education [87]. Nonetheless, the above contributions are mainly theoretical with little empirical value. Tracy and Albers attempted to use the construct of mental workload to test the usability of web-sites [17] and allowing the identification of those high-workload sub-areas of the interface that required attention. A similar study was aimed at investigating tapping as a measure of mental workload and website usability [13]. Gahangir et al. attempted to understand the convergence of usability and cognitive load in evaluating the performance of fully integrated assistive technology solutions [88] when adopted by blind people. Their study showed a high correlation of a secondary task performance measure, and three types of load (intrinsic, germane and extraneous) treated as cognitive load indexed, to usability, measured with a modification of some of the dimensions proposed in [89]. Similarly, another study employed an index of mental workload, namely the NASA-TLX, in conjunction with a measure of usability, namely the USE questionnaire for evaluating an interface for social robotic telepresence [90]. Unfortunately, the sample size of the data employed in the above empirical studies is not enough to draw any credible conclusion about the interaction mental workload-usability.

Despite research at the intersection of mental workload and usability is sparse and limited, a number of papers have been published discussing the relationship between usability and task performance. In [91], preference, a parameter of usability, and human performance were positively associated. However, it pointed out the existence of cases where users would not prefer an underlying interactive system whose design would seem more aligned to objective performance measures. This line of thought also appeared in the meta-analysis conducted in [92] over 73 usability studies. Here authors suggest that measures of users? perceptions of phenomena, in general, are not correlated with objective measures of the phenomena.

1.4 Design of experiments

An empirical study with human participants has been designed and executed. Users had to interact with 3 popular web-sites (youtube, google, wikipedia) and execute typical tasks over them (Table A5 in S1 Appendix). The aim was to explore the relationship between the perception of usability, the subjective mental workload experience by users and the achieved objective performance. One self-reporting procedure for measuring usability and two self-reporting methods for mental workload assessment have been selected:

  • the System Usability Scale (SUS) [31]
  • the Nasa Task Load Index (NASATLX), developed at NASA [9]
  • the Workload Profile (WP) [77], based on Multiple Resource Theory [93, 94].

No physiological procedures for mental workload measurement were included in the study. This was not considered because the goal of this research is to investigate the relationship between perception of usability, through a questionnaire, and assessment of mental workload with self-reporting measures. Five classes of user objective performance on tasks have been set (Table 1). These classes of objective performance are sometimes conditionally dependent (Fig 2) and the associated number (1–5) is not an indication of their strength or rank but merely a label. The detailed research hypotheses are defined in Table 2 and illustrated in Fig 3. The three measurement techniques are detailed below, followed by the definition of the research hypotheses.

thumbnail
Fig 2. Partial dependencies of objective performance classes.

https://doi.org/10.1371/journal.pone.0199661.g002

1.4.1 The system usability scale.

The System Usability Scale is a subjective usability assessment instrument that include a Likert scale, bounded in the range 1 to 5 [31]. Questions can be found in Table A1 in S1 Appendix. Individual scores are not meaningful on their own. For odd questions (SUSi with i = {1|3|5|7|9}), the score contribution is the scale position (SUSi) minus 1. For even questions (SUSi with i = {2|4|6|8|10}), the contribution is 5 minus the scale position. For comparison purposes, the SUS value is converted in the range [1..100] ∈ ℜ with i1 = {1, 3, 5, 7, 9}, i2 = {2, 4, 6, 8, 10}:

1.4.2 The NASA task load index.

The Nasa Task Load Index [9] belongs to the category of self-assessment measures. It has been validated in the aviation industry and other contexts in Ergonomics [9, 71] with several applications in many socio-technical domains. It is a combination of six factors believed to influence MWL (questions of Table A2 in S1 Appendix). Each factors is quantified with a subjective judgement coupled with a weight computed via a paired comparison procedure. Subjects are required to decide, for each possible pair (binomial coefficient, ) of the 6 factors, ‘which of the two contributed the most to mental workload during the task’, such as ‘Mental or Temporal Demand?’, and so forth. The weights w are the number of times each dimension was selected. In this case, the range is from 0 (not relevant) to 5 (more important than any other attribute). The final MWL score is computed as a weighed average, considering the subjective rating of each attribute di and the correspondent weights wi:

1.4.3 The workload profile.

The Workload Profile assessment procedure [77] is built upon the Multiple Resource Theory proposed in [93, 94]. In this theory, individuals are seen as having different capacities or ‘resources’ related to:

  • stage of information processing—perceptual/central processing and response selection/execution;
  • code of information processing—spatial/verbal;
  • input—visual and auditory processing;
  • output—manual and speech output.

Each dimension is quantified through subjective rates (questions of Table A3 in S1 Appendix) and subjects, after task completion, are required to rate the proportion of attentional resources used for performing a given task with a value in the range 0..1 ∈ ℜ. A rating of 0 means that the task placed no demand while 1 indicates that it required maximum attention. The aggregation strategy is a simple sum of the 8 rates d (averaged here, and scaled in [1..100] ∈ ℜ for comparison purposes):

1.4.4 Research hypotheses.

Three research hypotheses have been defined in this empirical research (Table 2). The first hypothesis is a non-directional statement, anticipating that usability, as measured by SUS, captures a different variance than MWL, as measured by NASATLX or WP. It is anticipated that there is a relatively random relationship between usability and mental workload, indicating two uncorrelated constructs.

The second and third hypotheses are directional predictive statements. In detail, the second hypothesis anticipates that the usability and mental workload measures can be successfully combined together to predict objective performance better than the individual measures.

The third hypothesis assumes that the dimensions used to form a usability or a mental workload index cannot be combined together to predict the objective performance better than the individual model-specific dimensions. In other words, a hybrid model incorporating both attributes of usability and mental workload will form a new, unknown construct that does not contribute to enhance the prediction of objective performance when predicted with the attributed of the individual models. For clarification purposes, the above hypotheses are stated in formal terms in Table 3.

thumbnail
Table 3. Formal description of research hypotheses (corr a correlation coefficient and acc the accuracy of a prediction).

https://doi.org/10.1371/journal.pone.0199661.t003

1.4.5 Participants and procedure.

Due to the fact that this research involved human participants, the study has been approved by the ethics committee of the University of Dublin, Trinity College where the experiment has been carried out. The study has been conducted according to the principles expressed in the Declaration of Helsinki. Participants have been properly instructed and have indicated that they consent to participate by signing the appropriate informed consent paperwork. All efforts have been made by the author to protect the privacy and anonymity of participants. Participants were recruited initially through the use of internal mailing list in the School of Computer Science and Statistics at Trinity College Dublin. However, since the response was not as expected, with a drop out rate of 80%, probably due to the length of the experiment itself, the recruitment process was extended to personal contacts of the first participants through word-of-mouth.

A sample of 46 human volunteers, fluent in English, eventually participated in the research after signing the consent form. Participants were divided into 2 groups of 23 each: those in group A were different to those in group B. Volunteers could not engage with instructors during the execution of the tasks and their trained was not required. Ages ranges from 20 to 35 years; 24 females and 22 males evenly distributed across the 2 groups (Total—Avg.: 28.6, Std. 3.98; g.A—Avg. 28.35, Std.: 4.22; g.B—Avg: 28.85, Std.: 3.70) all with a daily Internet usage of at least 2 hours. Volunteers were asked to execute a set of 9 information-seeking web-based tasks (Table A5 in S1 Appendix) in the most natural way, over 2 or 3 sessions of approximately 45/70 minutes each, on different non-consecutive days on three popular web-sites. The number of sessions were established according to their availability to participate in the research. Designed information-seeking tasks differed in terms of intrinsic complexity, time-pressure, time-limits, human interference and interruptions as well as demands on different modalities (visual, auditory, information-processing). Also, three popular web-sites were selected with the assumption that participants had previously interacted with them at least once. The rationale behind this is the expectation to observe situations of underload for more assiduous Internet users, given their familiarity with the underlying interface. Two groups were formed because designed tasks were going to be executed on original and run-time altered web-interfaces (through a CSS/HTML manipulation, as in Table A4 in S1 Appendix). The rationale behind this manipulation was to allow the formation of scenarios in which even assiduous Internet users were expected to perceive an higher mental workload. If non-popular web-sites would have been selected, the chances to spot scenarios of underload were minimal. This is because of the unfamiliarity of the users with the new web-sites and the higher effort that would have been required for the execution of the experimental tasks. Additionally, participants had to interact with those web-sites multiple times, executing different tasks one after each other. This means that, even with the same interface, complexity of tasks could have been perceived differently, given the increasing level of fatigue of participants, boredom or annoyance task after task. The run-time manipulation of web-sites was also planned as part of a larger research study [9597], to enable A/B testing of web-interfaces (not included here). Interface manipulation was not extreme, like making things very hard to read. Rather the goal was to manipulate the original interface to alter usability and task difficulty independently. The order of the tasks administered was the same for all the volunteers. Computerised versions of the SUS (Table A1 in S1 Appendix), the NASATLX (Table A2 in S1 Appendix) and the WP (Table A3 in S1 Appendix) instruments were administered shortly after task completion. Note that the question of the NASATLX related to the ‘physical load’ dimension was set to 0 as well as its weight as no physical effort was required. As a consequence, the resulting pairwise comparison procedure became shorter. Some participant did not execute all the tasks and the final dataset contains 390 cases.

2 Results

Table 4 and Fig 4 show the means and the standard deviations of the usability and the mental workload scores for each information-seeking task (Table A5 in S1 Appendix).

thumbnail
Table 4. Mental workload & usability—Groups A, B (G.A/G.B).

https://doi.org/10.1371/journal.pone.0199661.t004

2.1 Testing hypothesis 1

To test hypothesis 1, a correlation analysis of the usability versus the mental workload scores has been performed. From an initial analysis of the data depicted in Fig 5, it seems intuitive to assess a random relationship between the usability scores (SUS) and the mental workload scores (NASATLX, WP). This is statistically confirmed in Table 5 by the Pearson and Spearman correlation coefficients computed over the full dataset (Groups A, B). The Pearson coefficient was chosen for exploring a linear correlation between the two constructs while the Spearman correlation for investigating the existence of a monotonic relationship, not necessarily linear.

thumbnail
Table 5. Pearson and Spearman correlation coefficients of the usability and the mental workload scores.

https://doi.org/10.1371/journal.pone.0199661.t005

Although the perception of usability does not seem to correlate at all with the subjective mental workload experienced by participants, a further investigation of their relationship was performed on a task-basis. Fig 6 depicts the density plots of the correlations achieved between the usability and mental workload scores, while Table 6 formally list their magnitude. Note that each density function contains 9 values, one for each task. Additionally, although the description of the tasks was identical across groups, participants executed them over two difference interfaces, therefore tasks in group A where considered different than tasks in group B. Commonly, in behavioural and social sciences, there may be a greater contribution from complicating factors, as in the case of subjective, self-reported ratings. Therefore, correlations above 0.5 are regarded as very high, within [0.3 − 0.5] as medium/moderate and within [0.1 − 0.3] small (symmetrically to negative values) [98](page 82). In this analysis, only medium and high correlation coefficients are taken into account (highlighted in Table 6), discarding those demonstrating random relationship.

thumbnail
Fig 6. Density plots of the correlations by task—Group A, B.

https://doi.org/10.1371/journal.pone.0199661.g006

Yet, a clearer picture does not emerge and just for a few tasks some form of relationship exists between perception of usability and the mental workload experienced by participants. Fig 7 is aimed at visually help with the explanation of these cases, extract further information and possible interpretations on why the usability scores were moderately or highly correlated with the workload scores.

  • tasks 1/A and 4/B: WP scores seem to be moderately negatively correlated with the SUS scores. This might suggest that when the proportion of attentional resources being taxed by a task is moderated and decreases, the perception of good usability increases. In other words, when web-interfaces and the tasks executed over them require a moderate use of different stages, codes of information processing and input, output modalities (section 1.4.3), the usability of those interfaces is increasingly perceived as positive.
  • tasks 9/A and 9/B: the NASATLX scores are highly and positively correlated with the SUS scores. This might suggests that, even when time pressure is imposed upon tasks (description of task 9 in Table A5 in S1 Appendix) causing an increment in the workload experienced, and the perception of performance decreases because the answer of the task is not found, than the perception of usability is not affected if the task is pleasant and amusing (like task 9). In other words, even if the experienced mental workload increases but not excessively, and even if an interface is slightly altered (task 9 group B), the perception of positive usability is strengthened if tasks are enjoyable.
  • tasks 1/B, 4/B, 5/B, 7/B: the NASATLX scores are highly negatively correlated with the SUS scores. This might suggests that when the mental workload experienced by users increases, perhaps because tasks are not straightforward, the perception of usability can be negatively influenced even with a slight alteration of the interface.

The above interpretations do not aim to be exhaustive. Rather they are possible interpretations and are only confined to this study. Further empirical investigations are required prior to generalising these findings. To further strengthening the data analysis, an investigation of the correlation between the usability and the mental workload scores has been performed on a participant-basis (Table 7 and Fig 8).

As in the previous analysis (by task), just medium and high correlation coefficients (>0.3) are taken into account for a deeper exploration. Additionally, because the correlations listed in Table 6 were not able to systematically demonstrate common trends, the analysis on the individual-basis was more strict. In details, only those scores of participants for which a medium or high linear relationship (Pearson) and a monotonic relationship (Spearman) was found between both the two MWL indexes (NASA, WP) and the usability scores (SUS) was taken into consideration. The goal was to look for the presence of any peculiar pattern of user’s behaviour or a more complex deterministic structure. These participants are highlighted in Table 7 (1, 5, 11, 12, 16, 27, 31, 42, 44). The densities of their correlations are depicted in Fig 9.

thumbnail
Fig 9. Density plots of the correlations of selected users by task (top, groups A, B) and by users (bottom).

https://doi.org/10.1371/journal.pone.0199661.g009

From Fig 8 (bottom), a multimodal distribution of the correlations of the usability and mental workload scores emerges, with a big cluster of users close to −0.5 and a smaller one close to 0.5. Fig 10 show the linear scatterplots associated to these participants with a linear straight regression line and a local smoothing regression line (Lowess algorithm [99]). The former type of linear regression is parametric and assumes normal distribution of data, while the latter is non-parametric, it does not necessarily assumes normality of data and it aids the identification of patterns, increasing the ability to see a line of best fit over data. Outliers from scatterplots have not been removed due to the limited available points—9 points which are the maximum tasks executed by each participant.

thumbnail
Fig 10. Relationship between usability and mental workload scores for participants with moderate or high Pearson and Spearman correlation coefficients.

https://doi.org/10.1371/journal.pone.0199661.g010

Yet, no relational pattern between perception of usability and subjective mental workload clearly emerges. Nonetheless, by reviewing the mental workload scores (NASATLX and WP), it appears that the nine selected participants have all experienced, except a few outliers, optimal mental workload (on average between 20 and 72). These users did not perceive situations of underload or overload while executing the underlying tasks. Similarly, taking a closer look at the usability scores, these users did not perceive an extremely negative usability, with scores higher than 40. Except users 12 and 16, all the others seem to share a property: the lower the experienced mental workload is when between optimal ranges, the lower their perception of usability.

A final attempt to investigate the relationship between experienced workload and perceived usability has been performed by grouping scores by the objective performance class. Table 8 lists the correlations of the workload and usability scores by performance class.

thumbnail
Table 8. Correlations of the mental workload scores with the usability scores by performance class.

https://doi.org/10.1371/journal.pone.0199661.t008

Unfortunately, no trend emerges from Table 8 and Fig 11 with the correlations between experienced mental workload and perceived usability all close to zero, suggesting a random relationship between the two constructs.

thumbnail
Fig 11. Correlations of the mental workload scores and the usability scores by performance class.

https://doi.org/10.1371/journal.pone.0199661.g011

2.2 Testing hypothesis 2

In order to test hypothesis two, an investigation of the impact of the perception of usability and the experienced mental workload on the objective performance achieved by user has been conducted. In this context, objective performance refers to objective indicators of the performance reached by each volunteers who participated in the user study, categorised in 5 classes (section 1.4). During the experimental study, the measurement of the objective performance of some user was faulty. These cases were discarded and a new dataset with 390 valid cases was formed. The investigation of the impact of the perception of usability and the mental workload experienced by users on the 5 classes of objective performance was treated as a classification problem, employing supervised machine learning. The distribution of these five classes are depicted in Fig 12 and Table 9.

thumbnail
Fig 12. Original distribution of the objective performance classes.

https://doi.org/10.1371/journal.pone.0199661.g012

Clearly, the original class frequencies are unbalanced and this is likely to have a negative influence on the classification of the performance classes. To mitigate this risk, a new dataset has been built through oversampling, a technique to adjust class distributions and to correct for a bias in the original dataset. This techniques aims to reduce the negative impact of class unbalance on model fitting. The minority classes were randomly sampled (with replacement) in a way to achieve the same size of the majority class (Table 9). The two mental workload indexes (NASA and WP) and the usability index (SUS) were treated as independent variables (features) and they were both employed individually and in combination to induce models aimed at predicting the five classes of objective performance (Fig 13). Four different families of learning classification techniques were adopted to predict the objective performance:

  • information-based learning: decision trees (Recursive Partitioning) [100, 101]
    1. - with the Gini impurity index
    2. - with the information gain entropy measure
  • similarity-based learning: k-Nearest Neighbors (euclidean distance)
  • probability-based learning: Naive Bayes
  • error-based learning: Support Vector Machine [102, 103].
    1. - with a radial basis function kernel
    2. - with a polynomial function kernel

thumbnail
Fig 13. Independent features and classification techniques.

https://doi.org/10.1371/journal.pone.0199661.g013

The independent features were normalised with unity-based normalisation (min/max algorithm) in the range [0..1] ∈ ℜ to facilitate the training of models. 10-fold stratified cross validation has been used in the training phase therefore the oversampled dataset was divided in 10 folds and in each fold, the original ratio of the distribution of the objective performance classes (Fig 12, Table 9) was preserved. 9 folds were used for training a model and the remaining fold for testing it against accuracy. This was repeated 10 times shifting the testing fold. Through this approach, 10 models were induced each with an associated classification accuracy. Thus 10 accuracy values were generated for each machine learning technique and for each combination of independent features (Fig 14). Table 10 lists these values, for the individual models (containing only the mental workload or usability feature) against the combined models (containing both the mental workload and the usability features), grouped by classification technique and ordered by mean. Importantly, training sets (a combination of 9 folds) and test sets (the remaining holdout set) were kept the same across the classification techniques and the different combination of independent features (paired 10-fold CV). This was essential to perform a fair comparison of the different trained models using the same data of training and test sets.

thumbnail
Fig 14. Distribution of the accuracies of individual and combined induced models ordered by mean.

https://doi.org/10.1371/journal.pone.0199661.g014

thumbnail
Table 10. Ordered distributions of accuracies of trained models grouped by learning technique (combined highlighted).

https://doi.org/10.1371/journal.pone.0199661.t010

From Table 10, most of the combined models (highlighted in blue), achieved almost always a higher accuracy than the individual models. However, to formally test hypothesis 2, the 10-fold cross-validated paired Wilcoxon statistical test has been chosen for comparing two matched accuracy distributions and to assess whether their population mean ranks differ [104]. This is a paired difference test which is a non-parametric alternative to the paired Student’s t-test. This has been selected because the population of accuracies (obtained testing each holdout set) was assumed to be not normal. Table 11 lists the accuracies achieved by each induced model, ordered by mean and grouped by classification learning technique.

thumbnail
Table 11. Wilcoxon test of distributions of accuracies ordered by independent features with 95% confidence intervals (statistically significant different models highlighted).

https://doi.org/10.1371/journal.pone.0199661.t011

In most of the cases, the combined models always yielded statistically higher classification accuracies than the individual models. In particular, perception of usability alone, as measured by SUS was nearly always the worst in predicting objective performance. The addition of a mental workload index to it (either NASA or WP) significantly enhanced the prediction of objective performance. The experienced mental workload, according to the NASATLX measure was half of the times sufficient to predict objective performance alone. In the other cases, the perceived usability, measured by the SUS index, was able to add predictive capacity to the individual models. A similar behaviour occurred with the other measure of mental workload, namely the WP index, which was able to predict objective performance individually half of the times. This empirical evidence suggests that indexes of experienced mental workload and perceived usability can be jointly employed to explain objective performance better than when employed individually. In particular, experience mental workload seems to explain larger variance than perception of usability, when both taken into account, as independent variables, to predict classes of objective performance.

2.3 Testing hypothesis 3

In order to test hypothesis 3, an investigation of the impact of the attributes used to assess usability (by the SUS measure) and the attributes used to assess mental workload (by the NASA and the WP measures) on the objective performance achieved by user has been conducted. This is a similar experiment as the one conducted for testing hypothesis 2 (section 2.2), but instead of using overall usability and mental workload indexes, as independent features, their dimensions were used. In other words, the questions of Tables A1, A2, A3 in S1 Appendix, were used as independent features. The same classification techniques used in the experiment set for testing hypothesis 2 were employed.

Fig 15 depicts the distributions of the accuracies produced by the models induced by using the selected supervised machine learning classification techniques ordered by means, which are on average higher than the distributions of accuracies obtained in Fig 14. Analytics presented in Table 12 follow the same trend as the findings presented in Table 10 but, on average with higher accuracies. As in the experiment conducted for testing hypothesis 2, the 10-fold cross-validated paired Wilcoxon statistical test has been chosen for comparing two matched accuracy distributions and to assess whether their population mean ranks differ [104]. Table 13 lists these tests comparing models incorporating all the features of the original usability and mental workload instruments, individually and combined. Findings suggest that the incorporation of the factors used within the SUS usability assessment technique to the factors used within the NASA scale, in a joint model, in most of the cases did not significantly enhance the prediction of objective performance. However, when the factors belonging to SUS were added to the factors belonging to WP in a joint model, half of the times these did contribute to enhance the prediction of objective performance. Eventually, when mental workload factors, either belonging to the NASA or WP scales, were added in a joint model to the SUS factors, this almost always enhanced the prediction of objective performance.

thumbnail
Fig 15. Ordered distributions of accuracies of trained models by mean using full feature sets of original mental workload and usability assessment instruments.

https://doi.org/10.1371/journal.pone.0199661.g015

thumbnail
Table 12. Ordered distributions of accuracies of trained models using full feature sets of original mental workload and usability instruments (combined models highlighted).

https://doi.org/10.1371/journal.pone.0199661.t012

thumbnail
Table 13. Wilcoxon test of distributions of accuracies ordered by independent features with 95% confidence intervals using mental workload and usability attributes (statistically significant different models highlighted).

https://doi.org/10.1371/journal.pone.0199661.t013

2.4 Internal reliability of measurement scales

In order to enhance the reliability of the findings obtained in this empirical research, the Cronbach’s alpha measure was computed to test the internal consistency of the psychometric instruments used, namely the System Usability Scale (SUS) and the mental workload assessment instruments, namely the Nasa Task Load Index (NASATLX) and the Workload Profile scale (WP). The Cronbach’s alpha is aimed at assessing how well these scales consistently measures what they are supposed to measure. Fig 16 depicts the inter-item correlations of the above scales, while Table 14 lists the Cronbach’s Alpha coefficient for each scale.

thumbnail
Fig 16. Intra-correlations of mental workload and usability questionnaire items.

https://doi.org/10.1371/journal.pone.0199661.g016

thumbnail
Table 14. Cronbach’s Alpha of the mental workload and usability questionnaire items.

https://doi.org/10.1371/journal.pone.0199661.t014

Clearly, the System Usability Scale is the most reliable, with a Cronbach’s Alpha coefficient of 0.93, in line with findings present in the literature [33, 105]. The Nasa Task Load index follows, with a lower reliability of 0.65 (using only the close ended questions) and 0.54 (using also the binary pair-wise comparisons) that indicates a questionable internal consistency of the scale. However, this is also in line with studies in the literature [106]. An important point is that the NASATLX scale as predominantly used in the field of transportation and safety critical systems [9] with little application in Human-Computer Interaction and within the arena of modern daily digital systems. The reliability of the Workload Profile (0.64) was aligned to the reliability of the NASATLX and these two scales had a fair concurrent validity (Correlation of 0.55 from Table 14).

3 Discussion

The results obtained in the previous sections are summarised and they are aligned to the research hypotheses previously set in Tables 2 and 3. Findings are critically discussed, including information about their statistical significance. Eventually, their implication to the broader field of Human-Computer Interaction is described.

H1: Usability and Mental workload are two uncorrelated constructs capturing difference variance (measured by self-reporting techniques—SUS, NASA-TLX, WP).

This has been tested by an in depth correlation analysis, both parametric and nonparametric, which confirmed that the two constructs are not correlated. The obtained correlation coefficients of Table 7 suggest that there is no linear correlation between the perception of usability and the experienced mental workload by users, with Pearson coefficients close to zero. Similarly, data suggests there is no tendency for usability to either monotonically increase or decrease when mental workload increases, with Spearman coefficients close to zero. The significance levels obtained for these correlations confirm that SUS and the NASA are significantly uncorrelated, fully accepting hypothesis H1(a). However, the absent correlation obtained between SUS and WP cannot be confirmed by statistical significance, thus the acceptance of sub-hypothesis H1(b) is with reservation. The correlation analysis has been strengthened by computing sub-correlations of data respectively by experimental task, users and objective performance class (Tables 68). Once more, no consistent pattern on the relationship between perception of usability and mental workload emerged, enforcing the belief that the two constructs are uncorrelated and capture difference variance.

H2: A unified model incorporating a usability and a MWL measure can significantly enhance the accuracy of the prediction of objective performance than the individual usability and MWL models.

This has been tested by inducing unified and individual models, using four supervised machine learning classification techniques, to predict the objective performance achieved by users (five classes of performance of Table 1). These models were trained with 10-fold cross validation, a well known model validation approach aimed at assessing how the results of a statistical analysis will generalise to an independent data set. For each classification technique, 10 accuracies were computed forming a distribution. Distributions of accuracies were subsequently compared using the 10-fold cross-validated paired Wilcoxon statistical test to assess whether their population mean ranks differ, with a significance level of 0.05. The unified models, including the measure of usability and one measure of mental workload, as independent features, were most of the times able to predict objective user performance, as dependent feature, statistically significantly better than the individual models (containing only one independent features, either usability or MWL). In details, on one hand, when the NASA measure was added to the SUS measure, it was always effective in significantly enhance the accuracy of the predictions, therefore sub-hypotheses H2(a, b) can be fully accepted. On the other hand, when SUS was added to a measure of mental workload, it enhanced the prediction of objective performance only half of the times. As a consequence, sub-hypotheses H2(c, d) cannot be fully accepted. This suggests that the NASA-TLX, as a measure of mental workload, has a higher capacity to explain objective performance than SUS, a measure of usability.

H3: A hybrid model incorporating features of a usability measure and features of a mental workload measure can significantly enhance the prediction of objective performance than models incorporating only usability or mental workload features.

This has been tested by inducing models containing the features of the System Usability Scale and the features of a measure of mental workload (either NASA or WP), firstly individually and then combined (hybrid model). The same four supervised machine learning classification techniques have been employed to train models to predict the objective performance achieved by users (five classes of performance of Table 1). These models were trained again with 10-fold cross validation in order to assess how the results will generalise to independent data. For each classification technique, 10 accuracies were computed forming a distribution. Distributions of accuracies were again compared using the 10-fold cross-validated paired Wilcoxon statistical test to assess whether their population mean ranks differ, with a significance level of 0.05. The hybrid models achieved, half of the times, a significantly higher classification accuracy than the individual models. In details, features of mental workload, when added to features of usability, nearly always statistically significantly enhanced the prediction of objective performance, therefore in favour of the acceptance of sub-hypotheses H3(a, b). However, on one hand, when features of mental workload belonging to the WP measure were added to features belonging to the SUS measure, only half of the times these significantly increased the classification of objective performance. On the other hand, when features of mental workload belonging to the NASA measure were added to features belonging to the SUS measure, nearly always did not enhance the prediction of objective performance. These results suggest that sub-hypothesis H3(d) cannot be fully accepted and sub-hypothesis H3(c) has to be rejected. It turns out that the features used within the NASA Task Load Index measure are powerful in predicting objective performance alone strengthening the fact that usability and mental workload measure two different aspects of user experience.

Table 15 summarises the acceptance status of the sub-hypotheses, whether they can be accepted, rejected or are subject to uncertainty.

thumbnail
Table 15. Formal description of research hypotheses and their acceptance status (corr a correlation coefficient and acc the accuracy of the model’s prediction).

https://doi.org/10.1371/journal.pone.0199661.t015

In summary, empirical evidence from this study suggests that there is no relationship between the perception of usability and the mental workload experienced by users on a set of web-based tasks executed on selected interfaces. Findings suggests that the two constructs seem to describe two not overlapping phenomena, sharing very little variance. The implication of this is that perception of usability and levels of experienced mental workload could be jointly employed to enhance the description of user experience. This is particularly relevant, for instance, in those scenarios in which humans, interacting with technologies, are people with cognitive disabilities or elderly, or when a graphical interface is presented on a desktop screen or a mobile device. Here, perception of usability can be high, but underlying tasks might impose non optimal levels of mental workload on users. Similarly, users can experience optimal mental load while executing underlying tasks, but might not perceive underlying interactive systems usable. In turn, the consideration of usability and mental workload as two distinct constructs can aid designers to build interactive technologies better aligned to the human mental limited capacities and that can maximise human performance. For example, during design phases, a designer can perform A/B testing of an interactive system by assessing mental workload and usability. In turn, this will generate a richer spectrum of feedback that can be taken into account to improve system design and optimise the performance of its users. The contributions of this research are to offer a new perspective on the application of mental workload to traditional usability inspection methods, and a richer approach to explain the human-system interaction and support its design.

4 Conclusion

This study attempted to investigate the correlation between the perception of usability and the mental workload imposed by typical tasks executed over three popular web-sites: Youtube, Wikipedia and Google. A literature review on prominent definitions of usability and mental workload was presented, with a particular focus on the latter construct. A well known subjective instrument for assessing usability —the System Usability Scale —and two subjective mental workload assessment procedures —the NASA Task Load Index, and the Workload Profile —have been employed in a primary research study involving 46 subjects. The perception of the usability of the interfaces these subjected interacted upon and the mental workload they have experienced while executing a selection of tasks, over selected interfaces, does not seem to correlate. The obtained empirical evidence strongly supports that usability and mental workload are two non overlapping constructs. Findings suggest that these two constructs can be jointly employed to improve the prediction of human performance, thus enhancing the description of user experience. The implications to the broader field of Human-Computer Interaction include the provision of mental workload as an important concept relevant for the design of interactive technologies better aligned to the human mental limited capacities and that can maximise human performance.

Future work will be devoted to the replication of this primary research on different interfaces and interactive systems. Experiment will be conducted by considering a wider selection of cognitive tasks in terms of temporal length, context (controlled and real-world tasks), complexity and mental resources demand. Similarly, a wider range of subjects is planned, including people affected by motor or cognitive impairments as well elderlies. Eventually, a new hybrid construct for explaining human performance over interactive technologies is envisaged. This new construct will incorporate factors concerned usability as well as human mental workload into a novel unified metric expected to have a higher validity, sensitivity and precision than current ad-hoc measures of user experience.

Acknowledgments

This research was supported by the ADAPT Centre for Digital Content Technology which is funded under the SFI Research Centres Programme (Grant 13/RC/2106) and is co-funded under the European Regional Development Fund.

References

  1. 1. Desmarais MC, Baker RS. A review of recent advances in learner and skill modeling in intelligent learning environments. User Modeling and User-Adapted Interaction. 2012;22(1-2):9–38.
  2. 2. Addie J, Niels T. Processing resources and attention. In: Handbook of human factors in Web design. Lawrence Erlbaum Associates; 2005. p. 3424–3439.
  3. 3. Fischer G. User Modeling in Human-Computer Interaction. User Modeling and User-Adapted Interaction. 2001;11(1-2):65–86.
  4. 4. Alonso-Ríos D, Vázquez-García A, Mosqueira-Rey E, Moret-Bonillo V. A Context-of-Use Taxonomy for Usability Studies. International Journal of Human-Computer Interaction. 2010;26(10):941–970.
  5. 5. Baldauf M, Dustdar S, Rosenberg F. A survey on context-aware systems. International Journal of Ad Hoc and Ubiquitous Computing. 2007;2(4):263–277.
  6. 6. Abowd G, Dey A, Brown P, Davies N, Smith M, Steggles P. Towards a better understanding of context and context-awareness. In: Handheld and ubiquitous computing. Springer; 1999. p. 304–307.
  7. 7. Macleod M. Usability in Context: Improving Quality of Use. In: Human Factors in Organizational Design and Management, Proceedings of the International Ergonomics Association 4th International Symposium. Elsevier; 1994.
  8. 8. Hassenzahl M, Tractinsky N. User experience-a research agenda. Behaviour & information technology. 2006;25(2):91–97.
  9. 9. Hart SG. Nasa-task load index (Nasa-TLX); 20 years later. In: Human Factors and Ergonomics Society Annual Meeting. vol. 50. Sage Journals; 2006. p. 904–908.
  10. 10. Young MS, Brookhuis KA, Wickens CD, Hancock PA. State of science: mental workload in ergonomics. Ergonomics. 2015;58(1):1–17. pmid:25442818
  11. 11. Longo L. A defeasible reasoning framework for human mental workload representation and assessment. Behaviour & Information Technology. 2015;34(8):758–786.
  12. 12. Longo L, Leva MC. Human Mental Workload: Models and Applications: First International Symposium, H-WORKLOAD 2017, Dublin, Ireland, June 28-30, 2017, Revised Selected Papers; 2017.
  13. 13. Albers M. Tapping as a Measure of Cognitive Load and Website Usability. Proceedings of the 29th ACM international conference on Design of communication. 2011; p. 25–32.
  14. 14. Lehmann J, Lalmas M, Yom-Tov E, Dupret G. Models of User Engagement. In: Proceedings of the 20th International Conference on User Modeling, Adaptation, and Personalization. UMAP’12. Berlin, Heidelberg: Springer-Verlag; 2012. p. 164–175. Available from: http://dx.doi.org/10.1007/978-3-642-31454-4_14.
  15. 15. O’Brien HL, Toms EG. What is user engagement? A conceptual framework for defining user engagement with technology. Journal of the American Society for Information Science and Technology. 2008;59(6):938–955.
  16. 16. Saket B, Endert A, Stasko J. Beyond Usability and Performance: A Review of User Experience-focused Evaluations in Visualization. In: Proceedings of the Sixth Workshop on Beyond Time and Errors on Novel Evaluation Methods for Visualization. BELIV’16. New York, NY, USA: ACM; 2016. p. 133–142. Available from: http://doi.acm.org/10.1145/2993901.2993903.
  17. 17. Tracy JP, Albers MJ. Measuring Cognitive Load to Test the Usability of Web Sites. Usability and Information Design. 2006; p. 256–260.
  18. 18. Longo L. Subjective Usability, Mental Workload Assessments and Their Impact on Objective Human Performance. In: IFIP Conference on Human-Computer Interaction. Springer; 2017. p. 202–223.
  19. 19. Annett J. Subjective rating scales in ergonomics: a reply. Ergonomics. 2002;45(14):1042–1046.
  20. 20. Annett J. Subjective rating scales: science or art? Ergonomics. 2002;45(14):966–987. pmid:12569049
  21. 21. Baber C. Subjective evaluation of usability. Ergonomics. 2002;45(14):1021–1025. pmid:12569060
  22. 22. Bastien JC. Usability testing: a review of some methodological and technical aspects of the method. International journal of medical informatics. 2010;79(4):e18–e23. pmid:19345139
  23. 23. Shackel B. Usability—Context, Framework, Definition, Design and Evaluation. Interact with Computers. 2009;21(5-6):339–346.
  24. 24. Hollingsed T, Novick DG. Usability inspection methods after 15 years of research and practice. In: Proceedings of the 25th annual ACM international conference on Design of communication. ACM; 2007. p. 249–255.
  25. 25. Hornbaek K. Current practice in measuring usability: Challenges to usability studies and research. International Journal of Human-Computer Studies. 2006;64(2):79–102.
  26. 26. Gray WD, Salzman MC. Damaged merchandise? A review of experiments that compare usability evaluation methods. Human–computer interaction. 1998;13(3):203–261.
  27. 27. Eason KD. Information technology and organisational change. CRC Press; 2005.
  28. 28. Brooke J, Bevan N, Brigham F, Harker S, Youmans D. Usability statements and standardisation: Work in progress in ISO. In: Proceedings of the IFIP TC13 Third Interational Conference on Human-Computer Interaction. North-Holland Publishing Co.; 1990. p. 357–361.
  29. 29. Nielsen J. Usability Inspection Methods. In: Conference Companion on Human Factors in Computing Systems. CHI’95. New York, NY, USA: ACM; 1995. p. 377–378.
  30. 30. Nielsen J. Heuristic evaluation. In: Nielsen J, Mack RLE, editors. Usability Inspection Methods. Wiley & Sons, New York; 1994. p. 413–414.
  31. 31. Brooke J. SUS-A quick and dirty usability scale. Usability evaluation in industry. 1996;189(194):4–7.
  32. 32. Bangor A, Kortum PT, Miller JT. An Empirical Evaluation of the System Usability Scale. International Journal of Human-Computer Interaction. 2008;24(6):574–594.
  33. 33. Lewis JR, Sauro J. The factor structure of the system usability scale. In: International conference on human centered design. Springer; 2009. p. 94–103.
  34. 34. Borsci S, Federici S, Lauriola M. On the dimensionality of the System Usability Scale: a test of alternative measurement models. Cognitive processing. 2009;10(3):193–197. pmid:19565283
  35. 35. Tullis TS, Stetson JN. A Comparison of Questionnaires for Assessing Website Usability. In: Annual Meeting of the Usability Professionals Association; 2004. p. 1–12.
  36. 36. Slaughter LA, Harper BD, Norman KL. Assessing the Equivalence of Paper and On-line versions of the QUIS 5.5. In: 2nd Annual Mid-Atlantic Human Factors Conference; 1994. p. 87–91.
  37. 37. Lewis JR. IBM Computer Usability Satisfaction Questionnaires: Psychometric Evaluation and Instructions for Use. International Journal of Human-Computer Interaction. 1995;7:57–78.
  38. 38. Davis FD. Perceived usefulness, perceived ease of use, and user acceptance of information technology. MIS quarterly. 1989; p. 319–340.
  39. 39. Harper BD, Norman KL. Improving User Satisfaction: The Questionnaire for User Interaction Satisfaction Version 5.5. In: 1st Annual Mid-Atlantic Human Factors Conference; 1993. p. 224–228.
  40. 40. Lin HX, Choong YY, Salvendy G. A proposed index of usability: a method for comparing the relative usability of different software systems. Behaviour & information technology. 1997;16(4-5):267–277.
  41. 41. Lund AM. Measuring Usability with the USE Questionnaire12. Usability interface. 2001;8(2):3–6.
  42. 42. Tractinsky N. The Usability Construct: A Dead End? Human-Computer Interaction. 2018;33(2):131–177.
  43. 43. Hancock P. Whither Workload? Mapping a Path for Its Future Development. In: International Symposium on Human Mental Workload: Models and Applications. Springer; 2017. p. 3–17.
  44. 44. Longo L. Designing medical interactive systems via assessment of human mental workload. In: Computer-Based Medical Systems (CBMS), 2015 IEEE 28th International Symposium on. IEEE; 2015. p. 364–365.
  45. 45. Loft S, Sanderson P, Neal A, Mooij M. Modeling and predicting mental workload in en route air traffic control: Critical review and broader implications. Human Factors. 2007;49(3):376–399. pmid:17552304
  46. 46. De Waard D. The measurement of drivers’ mental workload. The Traffic Research Centre VSC, University of Groningen; 1996.
  47. 47. Jou YT, Yenn TC, Lin CJ, Yang CW, Chiang CC. Evaluation of operators? mental workload of human–system interface automation in the advanced nuclear power plants. Nuclear Engineering and Design. 2009;239(11):2537–2542.
  48. 48. Gao Q, Wang Y, Song F, Li Z, Dong X. Mental workload measurement for emergency operating procedures in digital nuclear power plants. Ergonomics. 2013;56(7):1070–1085. pmid:23654299
  49. 49. de Greef T, Lafeber H, van Oostendorp H, Lindenberg J. Eye movement as indicators of mental workload to trigger adaptive automation. Foundations of augmented cognition Neuroergonomics and operational neuroscience. 2009; p. 219–228.
  50. 50. Dey A, Mann DD. Sensitivity and diagnosticity of NASA-TLX and simplified SWAT to assess the mental workload associated with operating an agricultural sprayer. Ergonomics. 2010;53(7):848–857. pmid:20582766
  51. 51. Byrne A. Measurement of mental workload in clinical medicine: a review study. Anesthesiology and pain medicine. 2011;1(2):90. pmid:25729663
  52. 52. Carswell CM, Clarke D, Seales WB. Assessing mental workload during laparoscopic surgery. Surgical innovation. 2005;12(1):80–90. pmid:15846451
  53. 53. Wickens CD. Mental Workload: Assessment, Prediction and Consequences. In: International Symposium on Human Mental Workload: Models and Applications. Springer; 2017. p. 18–29.
  54. 54. Cain B. A review of the mental workload literature. Defence Research & Dev. Canada, Human System Integration; 2007.
  55. 55. Xie B, Salvendy G. Review and reappraisal of modelling and predicting mental workload in single and multi-task environments. Work and Stress. 2000;14(1):74–99.
  56. 56. Huey BM, Wickens CD. Workload transition: implication for individual and team performance. Washington, DC.: National Academy Press; 1993.
  57. 57. Longo L. Mental workload in medicine: foundations, applications, open problems, challenges and future perspectives. In: Computer-Based Medical Systems (CBMS), 2016 IEEE 29th International Symposium on. IEEE; 2016. p. 106–111.
  58. 58. Moustafa K, Luz S, Longo L. Assessment of mental workload: a comparison of machine learning methods and subjective assessment techniques. In: International Symposium on Human Mental Workload: Models and Applications. Springer; 2017. p. 30–50.
  59. 59. Rizzo L, Longo L. Representing and Inferring Mental Workload via Defeasible Reasoning: A Comparison with the NASA Task Load Index and the Workload Profile. In: Proceedings of the 1st Workshop on Advances In Argumentation In Artificial Intelligence co-located with XVI International Conference of the Italian Association for Artificial Intelligence (AI*IA 2017), Bari, Italy, November 16-17, 2017.; 2017. p. 126–140.
  60. 60. Rizzo L, Dondio P, Delany SJ, Longo L. Modeling mental workload via rule-based expert system: a comparison with NASA-TLX and workload profile. In: IFIP International Conference on Artificial Intelligence Applications and Innovations. Springer; 2016. p. 215–229.
  61. 61. Longo L. Formalising Human Mental Workload as a Defeasible Computational Concept. The University of Dublin, Trinity College; 2014.
  62. 62. Longo L, Barrett S. Cognitive effort for multi-agent systems. In: International Conference on Brain Informatics. Springer; 2010. p. 55–66.
  63. 63. Gopher D, Donchin E. Workload—An examination of the concept. In: Boff KR, Kaufman L, Thomas JP, editors. Handbook of perception and human performance. vol. 2. John Wiley & Sons; 1986. p. 41/1–41/49.
  64. 64. Tsang PS. Mental Workload. In: Karwowski W, editor. International Encyclopaedia of Ergonomics and Human Factors (2nd ed.). vol. 1. Taylor & Francis; 2006.
  65. 65. Wilson GF, Eggemeier TF. Mental workload measurement. In: Karwowski W, editor. International Encyclopedia of Ergonomics and Human Factors (2nd ed.). vol. 1. Taylor & Francis; 2006.
  66. 66. Young MS, Stanton NA. Mental Workload. In: Stanton NA, Hedge A, Brookhuis K, Salas E, Hendrick HW, editors. Handbook of Human Factors and Ergonomics Methods. CRC Press; 2004. p. 1–9.
  67. 67. Young MS, Stanton NA. Mental workload: theory, measurement, and application. In: Karwowski W, editor. International encyclopedia of ergonomics and human factors. vol. 1. 2nd ed. Taylor & Francis; 2006. p. 818–821.
  68. 68. Zhang Y, Luximon A. Subjective mental workload measures. Ergonomia. 2005;27(3):199–206.
  69. 69. Thomas EN. Psychometric properties of subjective workload measurement techniques: implications for their use in the assessment of perceived mental workload. Human factors. 1991;33:17–33.
  70. 70. Reid GB, Nygren TE. The Subjective Workload Assessment Technique: A scaling procedure for measuring mental workload. In: Hancock PA, Meshkati N, editors. Human Mental Workload. vol. 52 of Advances in Psychology. North-Holland; 1988. p. 185–218.
  71. 71. Rubio S, Diaz E, Martin J, Puente JM. Evaluation of Subjective Mental Workload: A Comparison of SWAT, NASA-TLX, and Workload Profile Methods. Applied Psychology. 2004;53(1):61–86.
  72. 72. Tsang PS, Vidulich MA. Mental Workload and Situation Awareness. In: Salvendy G, editor. Handbook of Human Factors and Ergonomics. John Wiley & Sons, Inc.; 2006. p. 243–268.
  73. 73. Eggemeier TF, O’Donnell RD. A Conceptual Framework for Development of a Workload Assessment Methodology. Defense Technical Information Center OAI-PMH Repository (United States); 1998.
  74. 74. O’ Donnel RD, Eggemeier TF. Workload assessment methodology. In: Boff KR, Kaufman L, Thomas JP, editors. Handbook of perception and human performance. vol. 2. New York, Wiley-Interscience; 1986. p. 42/1–42/49.
  75. 75. Wilson GF, Eggemeier FT. Psychophysiological assessment of workload in multi-task environments. Multiple-task performance. 1991;329360.
  76. 76. Kramer AF. Physiological metrics of mental workload: A review of recent progress. Multiple-task performance. 1991; p. 279–328.
  77. 77. Tsang PS, Velazquez VL. Diagnosticity and multidimensional subjective workload ratings. Ergonomics. 1996;39(3):358–381. pmid:8849491
  78. 78. Zijlstra FRH. Efficiency in work behaviour. Delft University, The Netherlands; 1993.
  79. 79. Cooper GE, Harper RP. The use of pilot ratings in the evaluation of aircraft handling qualities. Advisory Group for Aerospace Research & Development; 1969. AD689722, 567.
  80. 80. Roscoe AH, Ellis GA. A Subjective Rating Scale for Assessing Pilot Workload in Flight: a decade of Practical Use. Farnborough (UK): Royal Aerospace Establishment; 1990. 90019.
  81. 81. Vidulich MA, Ward Frederic G SJ. Using the Subjective Workload Dominance (SWORD) technique for projective workload assessment. Human Factors Society. 1991;33(6):677–691.
  82. 82. Schmutz P, Heinz S, Métrailler Y, Opwis K. Cognitive Load in Ecommerce Applications: Measurement and Effects on User Satisfaction. Advances in Human-Computer Interaction. 2009;2009:3/1–3/9.
  83. 83. Edwards A, Kelly D, Azzopardi L. The Impact of Query Interface Design on Stress, Workload and Performance. Springer International Publishing; 2015. p. 691–702.
  84. 84. Longo L, Rusconi F, Noce L, Barrett S. The importance of human mental workload in web-design. In: 8th International Conference on Web Information Systems and Technologies. SciTePress; 2012. p. 403–409.
  85. 85. Xiao Y, Wang Z, Wang M, Lan Y. The appraisal of reliability and validity of subjective workload assessment technique and NASA-task load index. Zhonghua lao dong wei sheng zhi ye bing za zhi = Zhonghua laodong weisheng zhiyebing zazhi = Chinese journal of industrial hygiene and occupational diseases. 2005;23(3):178–181. pmid:16124892
  86. 86. Longo L. Human-computer interaction and human mental workload: Assessing cognitive engagement in the world wide web. In: IFIP Conference on Human-Computer Interaction. Springer; 2011. p. 402–405.
  87. 87. Davids M, Halperin M, Chikte U. Optimising cognitive load and usability to improve the impact of e-learning in medical education. African Journal of Health Professions Education. 2015;7(2):147–152.
  88. 88. Hossain G, Shaik AS, Yeasin M. Cognitive load and usability analysis of R-MAP for the people who are blind or visual impaired. In: Proceedings of the 29th ACM international conference on Design of communication. ACM; 2011. p. 137–144.
  89. 89. Nielsen J. The usability engineering life cycle. Computer. 1992;25(3):12–22.
  90. 90. Kiselev A, Loutfi A. Using a mental workload index as a measure of usability of a user interface for social robotic telepresence. In: 2nd Workshop of Social Robotic Telepresence in Conjunction with IEEE International Symposium on Robot and Human Interactive Communication 2012; 2012.
  91. 91. Nielsen J, Levy J. Measuring Usability: Preference vs. Performance. Commun ACM. 1994;37(4):66–75.
  92. 92. Hornbæk K, Law ELC. Meta-analysis of correlations among usability measures. In: Proceedings of the SIGCHI conference on Human factors in computing systems. ACM; 2007. p. 617–626.
  93. 93. Wickens CD, Hollands JG. Engineering Psychology and Human Performance. 3rd ed. Prentice Hall; 1999.
  94. 94. Wickens CD. Multiple resources and mental workload. Human Factors. 2008;50(2):449–454. pmid:18689052
  95. 95. Longo L. Human-Computer Interaction and Human Mental Workload: Assessing Cognitive Engagement in the World Wide Web. In: INTERACT (4); 2011. p. 402–405.
  96. 96. Longo L. Formalising human mental workload as non-monotonic concept for adaptive and personalised web-design. In: International Conference on User Modeling, Adaptation, and Personalization. Springer; 2012. p. 369–373.
  97. 97. Longo L, Dondio P. On the Relationship between Perception of Usability and Subjective Mental Workload of Web Interfaces. In: IEEE/WIC/ACM International Conference on Web Intelligence and Intelligent Agent Technology, WI-IAT 2015, Singapore, December 6-9, Volume I; 2015. p. 345–352.
  98. 98. Cohen J. Statistical power analysis for the behavioral sciences. Lawrence Erlbaum Associates.; 1988.
  99. 99. Cleveland WS. Robust locally weighted regression and smoothing scatterplots. American Statistical Association. 1979;74:829–836.
  100. 100. Therneau TM, Atkinson EJ, et al. An introduction to recursive partitioning using the RPART routines. Technical report Mayo Foundation; 1997.
  101. 101. Breiman L, Friedman J, Stone CJ, Olshen RA. Classification and regression trees. CRC press; 1984.
  102. 102. Bennett KP, Campbell C. Support Vector Machines: Hype or Hallelujah? SIGKDD Explor Newsl. 2000;2(2):1–13.
  103. 103. Karatzoglou A, Meyer D. Support Vector Machines in R. Journal of Statistical Software. 2006;15(9):1–32.
  104. 104. Wilcoxon F. Individual Comparisons by Ranking Methods. Biometrics Bulletin. 1945;1(6):80–83.
  105. 105. Bangor A, Kortum PT, Miller JT. An empirical evaluation of the system usability scale. Intl Journal of Human–Computer Interaction. 2008;24(6):574–594.
  106. 106. Hart SG, Staveland LE. Development of NASA-TLX (Task Load Index): results of empirical and theoretical research. In: Hancock PA, Meshkati N, editors. Human Mental Workload. vol. 52 of Advances in Psychology. North-Holland; 1988. p. 139–183.