Background
Version 1.0
In 1987, Ben Shneiderman presented a questionnaire that directed user attention to focus on their subjective rating of the human-computer interface. While this questionnaire was a strong step towards focus on users' evaluations of an interface, no empirical work had been done to assess its reliability or validity.Version 2.0
This original questionnaire consisted of 90 questions in total. Of these questions, 5 were concerned with rating a user's overall reaction of the system. The remaining 85 were organized into 20 groups which, in turn, consisted of a main component question followed by related subcomponent questions.Version 3.0
QUIS Version 2.0 was modified and expanded to three major sections. In the first section, there were three questions concerned with the type of system under evaluation and the amount of time spent on that system. In the second section, four questions focused on the user's past computer experiences. The last section, section III, included the modified version of QUIS Version 2.0, now containing 103 questions. These modifications included changing the 1-10 rating scale to be from 1-9, which 0 used as "not applicable". This also simplified future data entry for the questionnaire since a maximum rating would no longer require two keystrokes (as it would have in "10"). This in turn would reduce response bias from subjects.Version 4.0
Chin, Norman and Shneiderman (1987) administered the QUIS Version 3.0 and a subsequent revised Version 4.0 to an introductoryVersion 5.0
Another study using QUIS Version 5.0 was carried out with a local PC User's Group. In order to look at ratings across products, the participants were divided into 4 groups. Each group rated a different product. The products were: # a product the rater liked # a product the rater disliked # a command line system (CLS) # a Menu Driven Application (MDA) This investigation examined the reliability and discriminability of the questionnaire. In terms of discriminability, the researchers compared the ratings for software that was liked vs. the ratings for the software that was disliked. Lastly, a comparison between a mandatory CLS with that of a voluntarily chosen MDA was made. The researchers found that the overall reliability of QUIS Version 5.0 using Cronbach's alpha was .939.Version 5.5
Even though the QUIS Version 5.0 was a powerful tool for interface evaluation, interface issues limited the utility of the on-line version. Previous versions of the QUIS have been laid out in a very linear fashion, in which one question would be shown on each screen. However, this format is unable to capture the hierarchical nature of the question sets in the QUIS. This in turn limits the continuity between questions. QUIS 5.5 presented related sets of questions on the same screen. This helped improve question continuity within a set and reduced the amount of time subjects would spend navigating between questions.Version 5.5 - Development of the Web Based QUIS
Standard HTML forms were used to let users interact with the QUIS Version 5.5. The online version's style is very similar to the paper version of the questionnaire. The online version displayed multiple questions per page and comment areas at the end of each section. In order to ensure that users considered each question, a response was required for each question (users were able to answer "Not Applicable"). Client-side JavaScript was used to both validate and format the user's responses. The data for each section of the QUIS was time stamped and recorded on the client computer. At the end of the questionnaire the data from all sections of the QUIS were gathered together and send as a single piece back to the server where the QUIS was deployed. This method of data collection ensures that only completed questionnaires were entered, and prevents concurrency issues between users.Slaughter, L. A., Harper, B. D. & Norman, K. L. (1994). Assessing the Equivalence of Paper and On-line versions of the QUIS 5.5. Proceedings of the 2nd Annual Mid-Atlantic Human Factors Conference, (pp. 87-91), Washington, D.C.Version 5.5 Paper vs. Online Study
This study compared responses from paper and on-line formats of the QUIS Veresion 5.5. The majority of studies were interested in assessing equivalence between computerized and paper forms of tests. Overall, the results have not indicated significant differences. Twenty subjects evaluated WordPerfect© using both the paper and online formats of the QUIS Version 5.5. Each administration of the QUIS was preceded by a practice session to refamiliarize the subject with the interface. As the researchers expected, the format of the questionnaire did not affect users' ratings. However, it was of note that subjects using the online format wrote more in the comment sections than those who used the paper format. Also, the comments made by subjects using the online format provided better feedback in terms of problems, strengths, and examples. These results indicated that the online QUIS format provides more higher-quality information to developers, researchers and human factors experts than the paper-pencil format.Version 6.0
QUIS Version 5.5 was expanded into Version 6.0 and used for the study of the AVR "Guardian" system.Version 7.0
The QUIS Version 7.0 is an updated and expanded version of the previously validated QUIS 5.5. It is arranged in a hierarchical format and contains: (1) a demographic questionnaire, (2) six scales that measure overall reaction ratings of the system, (3) four measures of specific interface factors: screen factors, terminology and system feedback, learning factors, system capabilities, and (4) optional sections to evaluate specific components of the system. These specific components include: # technical manuals and online help # on-line tutorials # multimedia # Internet access # software installation Additional space allowing the rater to make comments regarding the interface is also included within the questionnaire. The comment space is headed by a statement that prompts the rater to comment on each of the specific interface factors.Current
In addition to English, the QUIS 7.0 is currently available in the following languages: German, Italian, Portuguese (Brazilian), and Spanish.Competitors
* Software Usability Measurement Inventory (SUMI): Contains 50 items for 5 aspects *References
{{Reflist Usability User interfaces