ArticlePDF Available

Measuring Usability with the USE Questionnaire

Authors:
Measuring Usability with the USE Questionnaire
12
By Arnold M. Lund
There are a variety of issues that tend to recur in the life of a user interface designer. I recall the first
time I was asked to write performance requirements for a user interface. How should I go about
deciding on an acceptable level of errors or an acceptable speed of accomplishing a standard task? How
do I know whether I have improved an interface design enough? Of the many problems that need fixing,
which ones should take priority? How do I even know whether improving the user interface of a product
is going to have an impact on sales? At one company, we sold usability so successfully one of the
business units declared they wanted to label each product with a “usability seal of approval.” How
would one go about determining when to award such a seal?
Over the years I have worked with colleagues at Ameritech (where the work began), U.S. WEST
Advanced Technologies, and most recently Sapient to create a tool that has helped in dealing with some
of these questions. The tool that we developed is called the USE Questionnaire. USE stands for
Usefulness, Satisfaction, and Ease of use. These are the three dimensions that emerged most strongly in
the early development of the USE Questionnaire. For many applications, Usability appears to consist of
Usefulness and Ease of Use, and Usefulness and Ease of Use are correlated. Each factor in turn drives
user satisfaction and frequency of use. Users appear to have a good sense of what is usable and what is
not, and can apply their internal metrics across domains.
General Background
Subjective reactions to the usability of a product or application tend to be neglected in favor of
performance measures, and yet it is often the case that these metrics measure the aspects of the user
experience that are most closely tied to user behavior and purchase decisions. While some tools exist
for assessing software usability, they typically are proprietary (and may only be available for a fee).
More importantly, they do not do a good job of assessing usability across domains. When re-engineering
began at Ameritech, it became important to be able to set benchmarks for product usability and to be
able to measure progress against those benchmarks. It also was critical to ensure resources were being
used as efficiently as possible, and so tools to help select the most cost-effective methodology and the
ability to prioritize design problems to be fixed by developers became important. Finally, it became clear
that we could eliminate all the design problems and still end up with a product that would fail in the
marketplace.
It was with this environment as a background that a series of studies began at Ameritech. The first one
was headed by Amy Schwartz, and was a collaboration of human factors, market research in our largest
marketing organization, and a researcher from the University of Michigan business school. Building on
1
Published in Lund, A. M. (2001). Measuring usability with the USE questionnaire. Usability Interface, 8(2), 3-6
(www.stcsig.org/usability/newsletter/index.html).
2
For more detail, contact Arnie Lund at alund@acm.org or amlundjr@gmail.com.
that research, I decided to develop a short questionnaire that could be used to measure the most
important dimensions of usability for users, and to measure those dimensions across domains. Ideally it
should work for software, hardware, services, and user support materials. It should allow meaningful
comparisons of products in different domains, even though testing of the products happened at
different times and perhaps under different circumstances. In the best of all worlds, the items would
have a certain amount of face validity for both users and practitioners, and it would be possible to
imagine the aspects of the design that might influence ratings of the items. It would not be intended to
be a diagnostic tool, but rather would treat the dimensions of usability as dependent variables.
Subsequent research would assess how various aspects of a given category of design would impact
usability ratings.
The early studies at Ameritech suggested that a viable questionnaire could be created. Interestingly, the
results of those early studies were consistent with studies conducted in the MIS and technology
diffusion areas, which also had identified the importance of and the relationship between Usefulness,
Satisfaction, and Ease of Use. Furthermore, the rich research tradition in these other areas provides
theory that may be extended to explain the relationships. This is an area that provides a link between
academic research and practice, and it is informed by several disciplines. Some work has already been
published suggesting that at least one publicly available tool drawn from earlier research can be applied
effectively to software interfaces.
How It Developed
The first step in identifying potential items for the questionnaire was to collect a large pool of items to
test. The items were collected from previous internal studies, from the literature, and from
brainstorming. The list was then massaged to eliminate or reword items that could not be applied across
the hardware, software, documentation, and service domains. One goal was to make the items as simply
worded as possible, and as general as possible. As rounds of testing progressed, standard psychometric
techniques were used to weed out additional items that appeared to be too idiosyncratic or to improve
items through ongoing tweaking of the wording. In general, the items contributing to each scale were of
approximately equal weight, the Chronbach's Alphas were very high, and for the most part the items
appeared to tap slightly different aspects of the dimensions being measured.
The questionnaires were constructed as seven-point Likert rating scales. Users were asked to rate
agreement with the statements, ranging from strongly disagree to strongly agree. Various forms of the
questionnaires were used to evaluate user attitudes towards a variety of consumer products. Factor
analyses following each study suggested that users were evaluating the products primarily using three
dimensions, Usefulness, Satisfaction, and Ease of Use. Evidence of other dimensions was found, but
these three served to most effectively discriminate between interfaces. Partial correlations calculated
using scales derived for these dimensions suggested that Ease of Use and Usefulness influence one
another, such that improvements in Ease of Use improve ratings of Usefulness and vice versa. While
both drive Satisfaction, Usefulness is relatively less important when the systems are internal systems
that users are required to use. Users are more variable in their Usefulness ratings when they have had
only limited exposure to a product. As expected from the literature, Satisfaction was strongly related to
the usage (actual or predicted). For internal systems, the items contributing to Ease of Use for other
products actually could be separated into two factors, Ease of Learning and Ease of Use (which were
obviously highly correlated). The items that appeared across tests for the three factors plus Ease of
Learning are listed below. The items in italics loaded relatively less strongly on the factors.
Usefulness
It helps me be more effective.
It helps me be more productive.
It is useful.
It gives me more control over the activities in my life.
It makes the things I want to accomplish easier to get done.
It saves me time when I use it.
It meets my needs.
It does everything I would expect it to do.
Ease of Use
It is easy to use.
It is simple to use.
It is user friendly.
It requires the fewest steps possible to accomplish what I want to do with it.
It is flexible.
Using it is effortless.
I can use it without written instructions.
I don't notice any inconsistencies as I use it.
Both occasional and regular users would like it.
I can recover from mistakes quickly and easily.
I can use it successfully every time.
Ease of Learning
I learned to use it quickly.
I easily remember how to use it.
It is easy to learn to use it.
I quickly became skillful with it.
Satisfaction
I am satisfied with it.
I would recommend it to a friend.
It is fun to use.
It works the way I want it to work.
It is wonderful.
I feel I need to have it.
It is pleasant to use.
Work to refine the items and the scales continues. There is some evidence that for websites and certain
consumer products there is an additional dimension of fun or aesthetics associated with making a
product compelling. For the dependent variables of primary interest, however, these items appear to be
reasonably robust. A short form of the questionnaire is easily constructed by using the three or four
most heavily weighted items for each factor.
Conclusion
While the questionnaire has been used successfully by many companies around the world, and as part
of several dissertation projects, the development of the questionnaire is still not over. For the reasons
cited, this is an excellent starting place. The norms I have developed over the years have been useful in
determining when I have achieved sufficient usability to enable success in the market. To truly develop a
standardized instrument, however, the items should be taken through a complete psychometric
instrument development process. A study I have been hoping to run is one that simultaneously uses the
USE Questionnaire and other questionnaires like SUMI or QUIS to evaluate applications. Once a publicly
available (i.e., free) standardized questionnaire is available that applies across domains, a variety of
interesting lines of research are possible. The USE Questionnaire should continue to be useful as it
stands, but I hope the best is yet to come.

Supplementary resource (1)

... Sessions were conducted in-person or through Zoom and facilitated by study staff trained in usability and field testing of eHealth platforms. For usability testing, participants completed the 30-item USE questionnaire [65] which contains 4 subscales assessing usefulness (eg, "It helps me be more effective"), ease of use (eg, "It is easy to use"), ease of learning (eg, "I learned to use it quickly"), and satisfaction (eg, "I am satisfied with it") on an 8-point Likert scale (1=strongly disagree to 8=strongly agree). For field testing, we used the "think aloud" method [66], which encourages participants to vocalize any thoughts or comments while performing or immediately after performing a task, such as reading or watching a video. ...
... The Usefulness, Satisfaction, and Ease of Use (USE) questionnaire [65] and sociodemographic and medical characteristics were analyzed with descriptive statistics. We applied rapid qualitative analysis (RQA) to analyze the focus groups and usability and field-testing sessions following Watkins' guidelines [67]. ...
Article
Background Allogeneic hematopoietic stem cell transplantation (allo-SCT) is an effective treatment for various hematologic cancers, though it often results in severe side effects and psychological distress, which can negatively impact health outcomes. Integrative therapies like mindfulness-based stress reduction (MBSR), mindfulness meditation (MM), and music therapy (MT) yield promising results in enhancing both psychosocial outcomes (eg, reducing anxiety and depression) and physiological adaptation (eg, decreasing inflammation) in cancer patients. Objective We developed and refined, using focus groups and environmental and field testing, an eHealth-delivered mindfulness-based music therapy (eMBMT) intervention aimed at improving health-related quality of life, symptom burden (ie, pain, fatigue, and sleep), disease activity (ie, chronic graft-versus-host disease, cytomegalovirus activation, and infections) and psychosocial (ie, depression, anxiety, and cancer-specific distress) and physiological adaptation (ie, inflammation and immune reconstitution) tailored to adults receiving allo-SCT. Methods eMBMT intervention content is grounded in MT, MM, and MBSR, developed by a multidisciplinary team, and adapted for adults undergoing allo-SCT. eMBMT content was refined through focus groups and usability and field testing. Focus groups used a semistructured interview guide, while field testing used the “think aloud” method. Usability was evaluated using the 30-item Usefulness, Satisfaction, and Ease of Use (USE) questionnaire. Descriptive statistics analyzed the USE questionnaire and participant characteristics, while rapid qualitative analysis was applied to focus groups and field-testing sessions. Survivors eligible to participate in the focus groups and usability and field testing were adults ( > 18 years old) who received an allo-SCT ( < 36 months) for myelodysplastic syndrome, acute myeloid leukemia, or chronic myeloid leukemia, and were in remission for greater than 3 months. Results During the focus groups, participants (n=11; mean age 43.6, SD 17.8 years) provided qualitative feedback highlighting the shock of diagnosis, challenges during hospitalization, and coping strategies posttreatment. The eMBMT platform received positive evaluations for usefulness (mean 6.47, SD 0.29), ease of use (mean 6.92, SD 0.60), and satisfaction (mean 6.16, SD 0.82). Key themes from field testing highlighted the significance of social support, hope, and maintaining an active lifestyle. Suggestions for improvement included incorporating more representative content, reducing text, enhancing guidance, offering diverse music options, and streamlining blood sample collection. Conclusions The eMBMT intervention is a comprehensive, user-friendly eHealth tool tailored to the unique needs of allo-SCT patients. The positive feedback and identified areas for improvement underscore its potential to enhance well-being, symptom management, and overall quality of life for cancer survivors. A future pilot randomized controlled trial will further evaluate the feasibility, acceptability, and preliminary efficacy of the eMBMT intervention in improving health-related quality of life, symptom burden, disease activity, and psychosocial and physiological adaptation.
... Parents can use the Digital Malay Ethnoparenting application by selecting the desired menu. Then, parents were asked to fill in the usability questionnaire developed by (Lund, 2001) as a summative evaluation. The results of this summative evaluation are significant because they will determine the success and usability of the prototype following the goals to be achieved by this research. ...
... In the next step, beta testing involved five lecturers from the University of Riau and University of Education Indonesia and five parents of early-age children in Pekanbaru City, Riau Province. Meanwhile, 48 parents were involved in the validation stage, and the data obtained from the Usefulness, Satisfaction, and Ease of Use or USE questionnaire developed by Lund (2001) were examined using the statistical program SPSS. The methodology used in the questionnaire uses the 5-point Likert scale to produce comprehensive and precise results. ...
Article
Full-text available
This research aims to design and develop a Digital Malay Ethnoparenting application based on Riau Malay culture. Using the research and development method with a quantitative and qualitative approach, the research stages follow the ADDIE model. (Analysis, Design, Development, Implementation, Evaluation). The results of the analysis indicate the need for culturally-based parenting content that is relevant to the digital lifestyle of millennial parents. The application content is designed to include the values of Tunjuk Ajar Melayu, which contains moral and spiritual guidance. The application prototype was tested in alpha and beta phases, involving 48 parents in Pekanbaru. The trial showed that this application is innovative, easy to use, and meets the needs of modern parenting. Expert evaluation ensures the system's cultural validity and functionality. The research results confirm the potential of the application to preserve the cultural values of Riau Malay through technology, bridging the need for modernization without leaving behind the local cultural heritage.
... As metodologias usadas nas avaliações dos JEDs são: Game Experience Questionnaire -GEQ (E12) [Ijsselsteijn et al. 2007]; Multiple Intelligence Questionnaire -MIPQ (E12) [Tirri e Nokelainen 2011]; Instructional Materials Motivation Survey -IMMS (E14); Usefulness, Satisfaction, and Ease of Use -USE (E14) [Lund 2001]; System Usability Scale -SUS (E18) [Brooke 1996]; Handheld Augmented Reality Usability Scale -HARUS (E18) [Santos et al. 2014]; Emotional Self-Assessment Manikin -emoti-SAM (E22) [Hayashi et al. 2016]; Model for the Evaluation of Educational Games and Activities -MEEGA+ (E22) [Petri et al. 2020]; e Instrumento para Avaliação da Qualidade de Jogos Educacionais -IAQJEd (E24) [Coutinho 2017]. Para realizar as avaliações, os autores utilizaram experimentos controlados (E5, E9, E12), aplicação de pré-testes e póstestes (E5, E7, E8, E12), e aplicação de questionários após o contato dos estudantes com os JEDs (E3, E13, E14, E16, E18, E19, E21, E22, E24, E25). ...
Conference Paper
Full-text available
O ensino de Circuitos Lógicos, embora essencial nos cursos de Computação, enfrenta desafios pela complexidade dos conceitos, sendo os jogos educacionais uma alternativa eficaz para promover engajamento e facilitar a aprendizagem. Este trabalho apresenta um mapeamento sistemático da literatura sobre jogos educacionais digitais aplicados ao ensino de Circuitos Lógicos. São discutidas questões técnicas, como o gênero, os elementos de jogo e as tecnologias de desenvolvimento, além de aspectos pedagógicos, como os conteúdos abordados, o perfil do público-alvo, as metodologias de avaliação e os impactos alcançados. Entre os principais resultados, destacam-se a concentração dos jogos em poucos conteúdos, a ausência de rigor metodológico na avaliação de sua eficácia, a indisponibilidade de acesso público para grande parte dos jogos e a carência de abordagens voltadas à inclusão e à acessibilidade.
... Several researchers have developed questionnaires [25,29,34,42] on the basis of the TAM [56]. Other questionnaires [13,30,32,47] utilized for usability assessment include the mHealth App Usability Questionnaire (MAUQ) [3] and the Usefulness, Satisfaction and Ease of Use (USE) questionnaire [55]. ...
Article
Full-text available
Cyber threats targeting educational institutions have become increasingly sophisticated, requiring proactive detection and rapid response measures. This study presents an integrated framework that combines an AI-driven threat detection system with the Malware Information Sharing Platform (MISP) to enhance cyber defenses in school networks. Leveraging a Design Science Research methodology, the project unfolds across four phases: system design and architecture, prototype implementation, pilot testing, and evaluation. In the system design phase, requirements were gathered to develop an end-to-end architecture on Amazon Web Services (AWS), encompassing real-time data ingestion through Amazon Kinesis and Amazon Simple Email Service, data storage on Amazon S3, and AI-based anomaly and phishing detection using AWS Lambda. The prototype stage featured the integration of Random Cut Forest (RCF) for unsupervised anomaly detection and DistilBERT for phishing classification, enabling near real-time analysis of network and email data streams. MISP was hosted on Amazon EC2 and integrated with external threat feeds via STIX/TAXII, creating a closed-loop system that continuously refines shared Indicators of Compromise (IoCs). Pilot testing involved three schools in Camarines Norte, where simulated attack scenarios validated the system’s practical effectiveness. Results revealed high detection accuracy, with reduced false positives once IoCs were regularly enriched through MISP. The automated alerting workflow significantly shortened time-to-detection and time-to-response when compared to traditional security approaches. Quantitative metrics confirmed improvements in detection speed, precision, and recall, while qualitative feedback highlighted the system’s ease of use, scalability, and cost efficiency. These findings underscore the potential of AI-enhanced solutions underpinned by threat intelligence sharing, illustrating a robust, sustainable approach to improving cybersecurity in the education sector.
Article
Modern sport technology is a rapidly evolving scientific field utilizing leading techniques, such as the internet of things (IoT), augmented reality (AR), virtual reality (VR) and the use of many kinds of sensors. The feedback and the support provided by technology seem to have the potential to bring sufficient changes in sports training following a new trend, which is trying to integrate modern technology into sports. The aim of this study is to integrate responsive feedback and internet of things (IoT) technology into sports mainly in order to support the training of novice bowling athletes, raise their engagement with their sport and boost their performance. Following this direction, this research utilizes a system of sensors to apply step tracking of a bowling athlete during a throw ball attempt. The proposed solution is a European-funded project (ERASMUS+) supporting the “Europe 2020 strategy” to motivate people to transform their learning experiences into a beneficial way of constructing better performance. The results of the proposed system evaluation showed that bowlers and their coaches highlighted its usability and its usefulness. Thus, the proposed system may prove a valuable tool supporting athletes in general, and more particularly, novices’ training in bowling.
Article
Introduction: The UDL framework provides a promising method for establishing inclusive and accessible learning environments by satisfactorily addressing the diverse requirements of all students. The framework has garnered substantial recognition in educational policies and practices, even though additional research is required to obtain a comprehensive understanding of its implementation and efficacy. Objectives: This study seeks to evaluate the applicability of the UDL paradigm in literacy instruction for remedial teachers. Methods: The usability assessment was adapted to focus on the domains of usefulness, ease of use, ease of learning, and satisfaction. Data were gathered via an online questionnaire administered to 25 specialist teachers in remedial education, operating in both urban and rural settings. The data were analyzed utilizing Aiken’s V, and a 7-point Likert scale was adopted for streamlined answer collecting and analysis. Results: The results indicate that the UDL literacy framework exhibits significant usefulness, as evidenced by robust consensus across all assessed categories. The usefulness domain exhibits Aiken’s V values ranging from 0.8 to 0.88, underscoring its efficacy in fulfilling user requirements. The domains of ease of use and ease of learning exhibit high agreement scores of 0.8 to 0.85 and 0.8 to 0.84, respectively, signifying that the framework is user-friendly and comprehensible. The satisfaction domain, with ratings between 0.8 and 0.84, indicates that consumers regard the framework as worthwhile and advantageous. Conclusions: These results indicate that the UDL literacy framework is an effective and practical approach for improving literacy education in remedial programs.
ResearchGate has not been able to resolve any references for this publication.