Information & Management 44 (2007) 104–113 www.elsevier.com/locate/im A comparison of Magal’s service quality instrument with SERVPERF Hollis Landrum a,1,2, Victor R. Prybutok b,3, Xiaoni Zhang c,* a U.S. Army Engineer Research and Development Center, CEWES-IM-R, 3909 Halls Ferry Road, Vicksburg, MS 39180, United States Information Technology and Decision Sciences Department, College of Business Administration, Denton, TX 76203-5249, United States c Department of Business Informatics, College of Informatics, Northern Kentucky University, Highland Heights, KY 41099, United States b Received 26 May 2005; received in revised form 8 April 2006; accepted 3 November 2006 Available online 11 December 2006 Abstract The role of service quality has become critical to the success of organizations. Therefore, it is important to use a reliable instrument to measure information service quality. SERVQUAL is a popular instrument for doing this, but, though widely used, it has been criticized for its reliability and validity. Use of the performance measures from SERVQUAL to form SERVPERF has addressed some of these issues. However, Magal’s instrument on information center success was found effective within a service context because of its service orientation. In our study, we therefore compared Magal’s instrument with the SERVPERF instrument in predicting satisfaction and usefulness; we found that Magal’s instrument had predictive advantages in determining future usefulness and satisfaction. An important result was to show that our results supported the use of Magal’s instrument as an alternative to SERVQUAL for researchers and managers interested in service quality assessment. # 2006 Elsevier B.V. All rights reserved. Keywords: Service quality; Usefulness; User self-efficacy; Information quality; SERVQUAL; SERVPERF 1. Introduction Service quality matters in every industry: all companies recognize its importance because it affects customer loyalty and satisfaction [24]. Thus, it is necessary to use a reliable instrument to measure information service quality. SERVQUAL has been applied to various settings and different users but it has * Corresponding author. Tel.: +1 859 572 6408; fax: +1 859 572 6627. E-mail addresses: hlandrum@att.net (H. Landrum), prybutok@unt.edu (V.R. Prybutok), zhangx@nku.edu (X. Zhang). 1 Tel.: +1 601 634 3561. 2 Retired. 3 Tel.: +1 940 565 4767; fax: +1 940 565 4935. 0378-7206/$ – see front matter # 2006 Elsevier B.V. All rights reserved. doi:10.1016/j.im.2006.11.002 been criticized for its reliability and validity [23,34,44]. Modified versions of SERVQUAL have only used its performance measures or SERVPERF. Magal’s [27] instrument also addressed service quality. It could be applied to the information service industry and its validity has been proven. Also, while Magal’s instrument uses 16 items, SERVPERF needs 21 items. We therefore decided to compare these instruments in an information service context to provide insight into the effectiveness of the two instruments. Many researchers have suggested that service quality should be included as a measure of IS success [36,46] but despite numerous studies, few have performed empirical tests on the relationship between service quality and IS success factors. One notable exception was that of Landrum and Prybutok [25]. H. Landrum et al. / Information & Management 44 (2007) 104–113 In this work we relied on portions of Landrum and Prybutok’s instrument and built upon this prior work to further our understanding of service quality measures, we: (1) validated Magal’s and SERVPERF’s dimensions, (2) compared the predictive validity of Magal and the SERVPERF instrument using satisfaction and usefulness as the dependent variables. 2. Literature review 2.1. Service quality The most influential instrument in measuring service quality has been SERVQUAL, developed in 1988 by Parasuraman et al. [33]; it contains 22 items and has been widely used for measuring service quality in marketing. It features five service dimensions— tangibles, reliability, responsiveness, assurance, and empathy. SERVQUAL was grounded in the Gaps Model [34] which stated that expectations are subjective and consist of user wishes or beliefs that a service provider should exhibit certain characteristics. Customers form their judgment of service performance through interaction with the providers. The gap or difference between customer expectations and perceptions about the provided service results in the customers’ perceptions of service quality. Many studies have validated the use of SERVQUAL in several applications; for example, Van Dyke et al. [45] suggest the use of an IS-context-modified version of the instrument to assess the quality of services supplied by an ISP. However, there are different beliefs on the appropriateness of using SERVQUAL in an IS context and there is some debate on the appropriateness of the instrument itself: prior studies have questioned its dimensionality; for example, Nitecki [29] used a three dimensional model rather than the five proposed by Zeithaml et al. [48]. Recently, Jiang et al. [22] surveyed IT professionals and concluded that four dimensions were more appropriate than five. Furthermore, Wright and Martin [47] investigation of service quality in the higher education sector using importance–performance technique determined that reliability was not significant but that contact, tangibles, and response were significant components of online library services. In addition, others doubted the predictive power of using gap theory and preferred SERVPERF, the performance-only approach based simply on customers’ perceptions of the service provider performance. They contended that performance alone provides better 105 predictive power than SERVQUAL, which uses the gapbased scale: indeed many studies have provided evidence that performance scores exhibit better reliability and validity than do difference scores [3,5]. We therefore used only performance scores in examining SERVPERF for its dimensionality, convergent, and reliability measures. 2.2. Magal’s instrument Two instruments, one by Essex and Magal [13] for measuring information center success and another by Seddon and Kiew [40] for measuring IS success, were adapted in our study. Both were based on the well known instrument of Ives et al. [21] based on work by Bailey and Pearson [4] for measuring IS satisfaction. Magal made a number of modifications to the instrument to reflect aspects believed important to information center success and also changed the scale from a semantic differential to a seven-point Likert scale ranging from low to high for both importance and performance of each item. Magal’s model was based on three dimensions: quality of information center staff service, quality of user-developed applications, and user self-sufficiency. Substituting information quality for quality of userdeveloped applications, the dimensions for success of libraries became: quality of library staff service, quality of information received, and user self-sufficiency. Because of the nature of information, it is difficult to separate the tangible from the intangible process of delivery [17]. However, since information was the product or outcome, it could be considered separate from, and not a subset of service. In a study of a library reference service, Murfin and Gugelchuk [28] found that library customers appeared to distinguish between satisfaction with service provided and satisfaction with the information received. In our work, service therefore referred to the delivery process, and the variables associated with service were related to how the service was delivered by the staff. Variables associated with information, on the other hand, referred to the outcome or nature of the information product, such as its comprehensiveness or accuracy. 2.3. Service quality and satisfaction Customer satisfaction is the result of a product or service exceeding what the customers expected [31,32]. Hernon and Whitman defined satisfaction as a sense of contentment that arose from an actual experience [18]. Several researchers, such as Pitt et al. [37], proposed 106 H. Landrum et al. / Information & Management 44 (2007) 104–113 that service quality was a missing factor in the DeLone and McLean’s 1992 model and, in 2003, they modified it to include service quality as a success measure [11]. In fact, if service quality is high users are satisfied. 2.4. Usefulness and satisfaction Usefulness is experienced when the user believes that the system has enhanced his or her job performance [20]. Satisfaction occurs when a system enhances a user’s performance [15]. The theoretical relationship between these originated in DeLone and McLean’s [10] model that posited satisfaction and use as outcome measures of IS success. Seddon and Kiew [40] replaced use with usefulness and showed usefulness and satisfaction as two outcomes of IS. They also hypothesized that perceived usefulness leads to user satisfaction. The theoretical relationship between usefulness and satisfaction was then validated; for example, Rai et al. [38] confirmed this relationship and in an ERP context, usefulness was shown to be positively related to satisfaction [49]. Hsu and Chiu examined post-adoption cognitive beliefs and factors influencing e-service continuance and their results suggested that satisfaction was determined by perceived usefulness [19]. We therefore investigated the relationship between Magal’s instrument and SERVPERF for both usefulness and satisfaction because both instruments are potentially justified as success outcomes that relate to service quality. Fig. 1 shows the structural model using SERVQUAL dimensions to predict Fig. 2. Model 2 Magal’s instrument. usefulness and satisfaction and Fig. 2 shows the structural model using Magal’s dimensions to predict usefulness and satisfaction. 3. Methodology 3.1. SERVQUAL instrument and its constructs We used the SERVPERF items from the SERVQUAL instrument to measure service quality with minimum wording changes; the instrument had five service quality dimensions—tangibles, reliability, responsiveness, assurance, and empathy. The 1988 version of SERVQUAL consisted of 22 questions on expectations of service and 22 identical questions on performance of service. Five were used to measure tangibles, five to measure reliability, five to measure responsiveness, four to measure assurance, and three to measure empathy. Three changes were made to the 1994 items [35]. First, the item about the maintenance of error free records was deleted. Second, the item about keeping customers informed when services would be performed was reassigned from the responsiveness to the reliability dimension. Third, the item about convenience in the hours of operation was moved from the empathy to the tangibles dimension. In our study, we therefore used the performance items (SERVPERF) from the 1994 version of PZB’s SERVQUAL with only 21 items and minimal adaptation to accommodate the IS studied. 3.2. Magal’s constructs Fig. 1. Model 1 SERVQUAL satisfaction. We adapted Magal’s 16 item set to our study with minor modifications to the wording of the items so that they referred to a library rather than an information center. The instrument consisted of seven items to H. Landrum et al. / Information & Management 44 (2007) 104–113 measure staff service quality, four to measure information quality, and five to measure user self-sufficiency. 3.3. Usefulness The measures for usefulness were adopted from Davis’s study [9]: six were used. 3.4. Satisfaction Measurements of satisfaction vary; some studies have used only one item [42], while others have multiple items [12,26]. In general studies have examined performances and expectations [7]. In our study we used an outcome-based definition of satisfaction and thus our measures included four seven-point overall service satisfaction scales dealing with value, effectiveness, efficiency, and overall satisfaction. 4. Data analysis 4.1. Data collection The survey (see Appendix A) was given to randomly selected customers of two Army Corps of Engineers research centers. These centers have similar missions and similar types of knowledge-worker customers, and all of them pay for the services and information they receive. Technically speaking, these facilities are research libraries; they function as information service centers and research support systems for engineers and scientists. There is physical access to the facilities but a great deal of electronic access occurs through an online public access catalog that enables users to locate information and make requests to the help desk for technical and topical support (and determine service quality). The main difference between the two sites is size. Customers were chosen randomly from user lists at each site. Questionnaires were distributed and returned Table 1 Descriptive statistics Information quality User self-sufficient Staff service quality Tangible Reliability Responsive Assurance Empathy Satisfaction Usefulness Mean Standard deviation Reliability 6.08 5.27 6.02 5.03 5.93 6.18 5.96 5.95 5.67 5.38 0.89 1.14 0.92 1.04 0.95 0.94 0.93 1.00 1.12 1.19 0.93 0.89 0.91 0.84 0.91 0.89 0.87 0.92 0.93 0.96 107 anonymously in sealed envelopes. Respondents rated the perceived performance of each SERVPERF item on a Likert seven-point scale ranging from low to high. Because we are interested in determining the users’ opinion of the quality of the information service, it was necessary to survey the users. During the pilot study, we randomly selected half of the users at the small library site. For the full scale survey we mailed out surveys to the users at the small site that had not participated in the pilot and to the entire population of users at the large site. In addition we took care to ensure that the respondent sample was consistent with the populations at each facility. Table 2 Convergent validity of the constructs in model 1 Constructs and indicators Standardized loadings Tangibility (TAN) Tan1 Tan2 Tan3 Tan4 Tan5 0.74 0.79 0.82 0.82 0.55 Reliability (REL) Rel1 Rel2 Rel3 Rel4 Rel5 0.89 0.89 0.94 0.95 0.74 Responsiveness (RES) Res1 Res2 Res3 0.81 0.95 0.94 Assurance (ASU) As1 As2 As3 As4 0.85 0.83 0.89 0.85 Empathy (EMP) Emp1 Emp2 Emp3 Emp4 0.85 0.91 0.89 0.92 Usefulness U1 U2 U3 U4 U5 U6 0.91 0.92 0.95 0.95 0.92 0.87 Satisfaction Sat1 Sat2 Sat3 Sat4 0.92 0.97 0.97 0.89 108 H. Landrum et al. / Information & Management 44 (2007) 104–113 Table 3 Discriminant validity of the constructs in model 1 Construct pair Tangible Tangible Tangible Tangible Reliability Reliability Reliability Responsive Responsive Assurance Useful Useful Useful Useful Useful Useful Satisfaction Satisfaction Satisfaction Satisfaction Satisfaction * Reliability Responsive Assurance Empathy Responsive Assurance Empathy Assurance Empathy Empathy Tangible Reliability Responsive Assurance Empathy Satisfaction Tangible Reliability Responsive Assurance Empathy DChi-square DDegrees of freedom Discriminant validity 280 360 276 304 68 196 402 84 108 19 497 1337 452 632 834 864 338 872 388 506 789 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 Yes* Yes* Yes* Yes* Yes* Yes* Yes* Yes* Yes* Yes* Yes* Yes* Yes* Yes* Yes* Yes* Yes* Yes* Yes* Yes* Yes* Significant at 1%. A total of 385 usable responses were received representing a response rate of 37%. Over 70% of the respondents were men and 68% engineers or scientists. These percentages closely reflect the demographics of the study sites. More than 75% had used the research center more than six times in the past year, and at least 77% indicated they relied on center staff frequently when seeking information. This suggested that the respondents’ experience qualified them to evaluate the variables meaningfully. We performed a z test on the difference in proportions of gender and occupations between respondents and nonrespondents and the mean difference between the respondents and the non-respondents. The p values for the z test were all above 0.05, which supported the contention that there was no significant difference between the respondents and non-respondents for these demographic variables. In addition, non-response bias was evaluated by comparing responses from early and late respondents [2]. t-Test results on age, gender, and occupations between late and early responses showed no significant differences at 1% level. Table 4 Structural equations for model 1 Convergent validity was examined by using the t-tests for the factor loadings. Anderson and Gerbing [1] stated that convergent validity was established if all factor loadings for the indicators measuring the same construct were statistically significant. Our results showed that all t-tests were significant for all the indicators in Table 2. Discriminant validity was evaluated by fixing the correlation between various constructs at 1.0 and reassessing the modified model [16,41]. Significant differences in the Chi-square statistics for the con- R2 Equations ** ** ** Useful = 0.24TAN + 0.27REL + 0.49RES + 0.14AS + 0.75EMP*** Sat = 0.31TAN*** + 0.11REL* + 0.058RES + 0.19AS* + 0.26EMP** + 0.43Useful*** * ** *** Significant at 0.05 level. Significant at 0.01 level. Significant at 0.001 level. 0.50 0.72 4.2. Analytical methods The focus of our research was to compare Magal’s instrument and SERVPERF in predicting usefulness. To perform the analytical computations we used LISREL because all the constructs were well established and its confirmatory factor analysis technique was suitable [6,8]. Table 1 shows the descriptive statistics for each of the research constructs. The reliability for all the constructs ranged from 0.84 to 0.96, exceeding the suggested value of 0.7 [30]. 4.3. Validation of model 1 H. Landrum et al. / Information & Management 44 (2007) 104–113 Table 5 Convergent validity of the constructs in model 2 Constructs and indicators Standardized loadings Staff service quality SSQ1 SSQ2 SSQ3 SSQ4 SSQ5 SSQ6 SSQ7 0.88 0.91 0.88 0.89 0.84 0.69 0.75 Information quality IQ1 IQ2 IQ3 IQ4 0.93 0.97 0.97 0.84 User self-sufficiency USS1 USS2 USS3 USS4 0.84 0.88 0.85 0.88 Satisfaction Sat1 Sat2 Sat3 Sat4 0.93 0.98 0.92 0.91 Usefulness U1 U2 U3 U4 U5 U6 0.89 0.92 0.94 0.93 0.93 0.90 109 Table 4 shows the structural equations for model 1. The fit indices for model 1 were NFI = 0.93, NNFI = 0.95, GFI = 0.90, and AGFI = 0.85. Gefen et al. stated that GFI and AGFI above 0.8 were acceptable fit indices [14]. 4.4. Validity of model 2 In addition to the loadings shown in Table 5 which supported the convergent validity of Magal’s dimension, there were methodologies used to assess discriminant validity [39,43]. Table 6 supports the discriminant validity among the three dimensions within Magal’s instrument. Table 7 shows the structural equations for model 2. The fit indices for model 2 using Magal’s dimensions were NFI = 0.91, NNFI = 0.92, GFI = 0.85, AGFI = 0.82. 5. Discussion strained and unconstrained models indicated high discriminant validity among constructs. Table 3 shows the results of construct pair tests among the seven constructs, providing support for their discriminant validity. The results of the analysis for model 1 supported the convergent and discriminant validity of the five dimensions of SERVPERF. They also showed that the SERVPERF dimensions explained 50% of the variance in usefulness. When predicting usefulness, four dimensions: tangible, reliability, responsive, and empathy of SERVPERF were significant and that the assurance dimension was not significant. The coefficients were 0.24, 0.27, 0.49, and 0.75, respectively. The magnitude of these coefficients indicated the relative importance in predicting usefulness. Empathy played the most important role and the next most important dimension was responsiveness. Our LISREL results for model 1 also showed that the SERVPERF dimensions explained 72% of variance in satisfaction. When predicting it, four dimensions: tangible, reliability, assurance, and empathy of SERVPERF were significant and the responsive Table 6 Discriminant validity of the constructs in model 2 Construct pair Staff service quality Staff service quality Information quality Usefulness Usefulness Usefulness Usefulness Satisfaction Satisfaction Satisfaction * Significant at 1%. Information quality User self-sufficiency User self-sufficiency Staff service quality Information quality User self-sufficiency Satisfaction Staff service quality Information quality User self-sufficiency DChi-square DDegrees of freedom Discriminant validity 692 589 692 767 711 602 549 488 754 493 1 1 1 1 1 1 1 1 1 1 Yes* Yes* Yes* Yes* Yes* Yes* Yes* Yes* Yes* Yes* 110 H. Landrum et al. / Information & Management 44 (2007) 104–113 Table 7 Structural equations for model 2 Equations R2 Useful = 0.45SSQ*** + 0.088IQ + 0.33USS*** Sat = 0.24SSQ*** + 0.087IQ + 0.16USS** + 0.5Useful*** Useful = 0.73SSQ*** Sat = 0.4SSQ*** + 0.54Useful *** 0.61 0.79 0.54 0.76 * Significant at 0.05 level. Significant at 0.01 level. *** Significant at 0.001 level. ** dimension was not significant. The coefficients for the predictors: tangible, assurance, empathy, and usefulness were 0.31, 0.11, 0.19, 0.26, and 0.43, respectively. Model 2 showed the relationships of the three components of Magal’s instrument with usefulness and satisfaction. The LISREL analysis confirmed the convergent and discriminant validity of Magal’s instrument, which explained 61% of the variance in usefulness. Staff service quality and user self-sufficient components were significant predictors while information quality was not significant. The coefficients for staff service quality and user self-sufficiency were 0.45 and 0.33, respectively. Magal’s instrument also explained 79% of the variance in satisfaction. These results showed that staff service quality, user selfsufficiency, and usefulness were significant predicators. The coefficients for staff service quality, user selfsufficiency, and usefulness were 0.24, 0.16, and 0.5, respectively. In this model information quality was not a significant predictor for either usefulness or satisfaction. When staff service quality was used alone to predict usefulness and satisfaction, all the fit indices were comparable to model 2 but the R2 squares were 54 and 76% for usefulness and satisfaction. Comparisons of the models in Table 7 showed that including user selfsufficiency in the model only improved R2 7% for usefulness and 3% for satisfaction. Comparing models 1 and 2 showed that the five dimensions of SERVPERF explained 50% of variance in usefulness whereas staff service quality explained 54% of variance in usefulness. In addition, the five dimensions of SERVPERF explained 72% of the variance in satisfaction whereas staff service quality explained 76% of the variance in satisfaction. These results also showed that in both instances staff service quality was superior to SERVPERF in predicting usefulness and satisfaction. In addition, these findings showed that staff service quality explained the majority of the variance in usefulness. In Magal’s instrument, there were only seven items to measure staff service quality. If we were interested in service quality comparison where staff was a relevant measure, the seven item staff service quality provided a more parsimonious instrument than the complete set of 16 items Magal’s instrument and 21 items SERVPERF. Our results showed that Magal’s instrument explained more variance than SERVPERF when predicting usefulness. The higher R2 derived from Magal’s instrument was potentially attributed to its three distinct components: staff service quality, information quality, and user self-sufficiency. The three constructs in Magal’s instrument represented perceptions about the service provider, system itself, and user characteristics. In essence, Magal’s instrument captured a complete service provider, system, and users’ interaction, and, as a result, provided a higher predictive power than SERVPERF. On the other hand, although SERVPERF has five dimensions, these measured the same underlying construct—service quality. Therefore, when SERVPERF was used to predict usefulness, there was only one construct that predicted usefulness whereas Magal’s instrument contained three independent constructs related to usefulness. One explanation for our finding was that, because Magal’s instrument provided a greater number of non-redundant statistically significant independent variables for analysis, it provided higher predictive validity. 6. Conclusion Using the right instrument to determine service quality helps management evaluate service performance and provides the ability to use the analysis to design better service operations and delivery. In our study, we compared SERVPERF and Magal’s instrument. First, our analysis showed that both Magal’s instrument and SERVPERF are valid instruments. However, Magal’s provided higher predictive validity for usefulness and satisfaction as a measure of success. We also found that though only one component of Magal’s instrument, staff service quality, was used to predict usefulness and satisfaction, staff service quality still produced higher predictive validity for usefulness and satisfaction. In addition to instrument validation, we compared the predictive validity of the two instruments. Because Magal’s instrument contained only seven items to measure staff service quality whereas SERVPERF contained 21, the parsimonious nature of Magal’s instrument with its higher predictive validity makes it a better choice in evaluating service quality. H. Landrum et al. / Information & Management 44 (2007) 104–113 Our results also provided strong support for using the seven-item measure of staff service quality from Magal’s instrument, because it provided a parsimonious instrument with a relatively high predictive capability for both usefulness and satisfaction. 111 However, it is only appropriate to assess overall service quality when staffing is relevant. Because of fewer items in evaluating service quality, management may find this small instrument appealing for data collection. Appendix A. Survey instrument We want to survey your impressions about how the library facility performs. For each of the questions below, please rate your perception of the facility’s performance by circling your assessment on the 1–7 scale, with 1 as low and 7 as high. If you do not know how to rate the performance you should circle 0 for ‘‘unknown’’. Please do not omit any questions. Magal’s instrument by dimension Staff service quality 1. The staff at this facility maintain good relations with the users 2. The staff at this facility maintain good communication with the users 3. The staff are technically competent at this facility 4. The staff at this facility have a cooperative attitude 5. The staff at this facility provide fast response/turnaround time 6. The facility allows convenient access 7. There is relevance of the facility’s support to needs Low-to-high Unknown 1234567 1234567 1234567 1234567 1234567 1234567 1234567 0 0 0 0 0 0 0 Information quality 8. The information received is accurate 9. The information received is complete 10. The information received is relevant 11. The information received is current 1234567 1234567 1234567 1234567 0 0 0 0 User self-sufficiency 12. 13. 14. 15. 16. 1234567 1234567 1234567 1234567 1234567 0 0 0 0 0 1234567 1234567 1234567 1234567 1234567 0 0 0 0 0 SERVPERF by dimension Tangibles Items 1. 2. 3. 4. 5. I feel like a participant with the facility I have a clear understanding of how to use the facility Appropriate training on the use of the facility’s resources is provided I am able to use the facility independently I feel in control when using the facility This facility has modern equipment This facility looks appealing The staff at this facility look neat and professional The documentation, such as signs, handouts, and brochures are appealing This facility provides convenient hours of operation Reliable items 6. This facility provides service as promised 7. I can depend on the staff at this facility to handle user service problems 8. The staff at this facility perform the right service the first time 9. The staff at this facility provide service at the promised time 10. The staff at this facility keep users informed about when services will be performed 1234567 1234567 1234567 1234567 1234567 0 0 0 0 0 Responsiveness items 11. The staff at this facility provide prompt service to users 12. The staff at this facility are willing to help users 13. The staff at this facility are ready to respond to users’ requests 1234567 1234567 1234567 0 0 0 Assurance items 14. 15. 16. 17. The The The The staff staff staff staff at at at at this this this this facility facility facility facility are courteous instill confidence in users make users feel secure in their transactions have the knowledge to answer users’ questions 1234567 1234567 1234567 1234567 0 0 0 0 Empathy items 18. 19. 20. 21. The The The The staff staff staff staff at at at at this this this this facility facility facility facility provide individual attention to users have the users’ best interests at heart deal with users in a caring fashion understand the needs of users 1234567 1234567 1234567 1234567 0 0 0 0 1234567 1234567 0 0 Satisfaction 1. The service at this facility provides value 2. The service at this facility is effective 112 H. Landrum et al. / Information & Management 44 (2007) 104–113 Appendix A (Continued ) Usefulness Low-to-high Unknown 3. The service at this facility is efficient 4. I am satisfied with service at this facility 1234567 1234567 0 0 1. 2. 3. 4. 5. 6. 1234567 1234567 1234567 1234567 1234567 1234567 0 0 0 0 0 0 This This This This This This facility facility facility facility facility facility allows me to accomplish my tasks faster improves my ability to do research enhances my effectiveness enables me to be more productive makes it easier to do research is useful References [1] J.C. Anderson, D.W. Gerbing, Structural equation modeling in practice: a review and recommended two step approach, Psychological Bulletin 103 (3), 1988, pp. 411–423. [2] J.S. Armstrong, T.S. Overton, Estimating nonresponse bias in mail surveys, Journal of Marketing Research 14 (3), 1977, pp. 396–402. [3] E. Babakus, G.W. Boller, An empirical assessment of the SERVQUAL scale, Journal of Business Research 24 (3), 1992, pp. 253–268. [4] J.E. Bailey, S.W. Pearson, Development of a tool for measuring and analyzing computer user satisfaction, Management Science 29 (5), 1983, pp. 530–545. [5] M.K. Brady Jr., J. Cronin Jr., R.R. Brand, Performance-only measurement of service quality: a replication and extension, Journal of Business Research 55 (1), 2002, pp. 27–31. [6] D.T. Campbell, D.W. Fiske, Convergent and discriminant validation by the multitrait-multimethod matrix, Psychological Bulletin 56 (1), 1959, pp. 81–105. [7] A. Caruana, Service loyalty: the effects of service quality and the mediating role of customer satisfaction, European Journal of Marketing 36 (7/8), 2002, pp. 811–828. [8] W. Chin, Issues and opinion on structural equation modeling, MIS Quarterly 22 (1), 1998, pp. 7–16. [9] F. Davis, Perceived usefulness, perceived ease of use, and user acceptance of information technology, MIS Quarterly 13 (3), 1989, pp. 318–340. [10] W.H. DeLone, E.R. McLean, Information systems success: the quest for the dependent variable, Information Systems Research 3 (1), 1992, pp. 60–95. [11] W.H. DeLone, E.R. McLean, The DeLone and McLean model of information systems success: a ten-year update, Journal of Management Information Systems 19 (4), 2003, pp. 9– 30. [12] W.J. Doll, W. Xia, G. Torkzadeh, A confirmatory factor analysis of the end-user computing satisfaction index, MIS Quarterly 18 (4), 1994, pp. 453–461. [13] P.A. Essex, S.R. Magal, Determinants of information center success, Journal of Management Information Systems 15 (2), 1998, pp. 95–117. [14] D. Gefen, E. Karhanna, D.W. Straub, Trust and TAM in online shopping: an integrated model, MIS Quarterly 27 (1), 2003, pp. 51–90. [15] M. Gelderman, The relation between user satisfaction, usage of information systems and performance, Information & Management 34 (1), 1998, pp. 1–11. [16] E.E. Ghiselli, J.P. Campbell, S. Zedick, Measurement Theory for the Behavior Sciences, Freeman, San Francisco, CA, 1981. [17] J. Goldhar, STI dissemination: issues and opportunities, in: W.R. King, G. Zaltman (Eds.), Marketing Scientific and Technical Information, Westview Press, Colorado, 1979 , pp. 23–27. [18] P. Hernon, J. Whitman, Delivering Satisfaction and Service Quality: A Customer-based Approach for Libraries, American Library Association, Chicago, 2001. [19] M.-H. Hsu, C.-M. Chiu, Predicting electronic service continuance with a decomposed theory of planned behavior, Behavior & Information Technology 23 (5), 2004, pp. 359–373. [20] M. Igbaria, N. Zinateli, P. Cregg, A. Cavaye, Personal computing: acceptance factors in small firms: a structural equation model, MIS Quarterly 21 (3), 1997, pp. 279–305. [21] B. Ives, M.H. Olson, J. Baroudi, The measurement of user information satisfaction, Communications of the ACM 26 (10), 1983, pp. 785–793. [22] J.J. Jiang, G. Klein, C.L. Carr, Measuring information system service quality: SERVQUAL from the other side, MIS Quarterly 26 (2), 2002, pp. 145–166. [23] W.J. Kettinger, C.C. Lee, Pragmatic perspectives on the measurement of information systems, MIS Quarterly 21 (2), 1997, pp. 223–240. [24] J. Kim, J. Lee, K. Han, M. Lee, Businesses as buildings: metrics for the architectural quality of Internet businesses, Information Systems Research 13 (3), 2002, pp. 239–254. [25] H. Landrum, V.R. Prybutok, A service quality and success model for the information service industry, European Journal of Operational Research 156 (3), 2004, pp. 628–642. [26] H.-H. Lin, Y.-S. Wang, An examination of the determinants of customer loyalty in mobile commerce contexts, Information & Management 43 (3), 2006, pp. 271–282. [27] S.R. Magal, A model for evaluating information center success, MIS Quarterly 8 (1), 1991, pp. 91–106. [28] E.M. Murfin, G.M. Gugelchuk, Development and testing of a reference transaction assessment instrument, College & Research Libraries 48 (4), 1987, pp. 314–336. [29] D. Nitecki, Changing the concept and measure of service quality in academic libraries, Journal of Academic Librarianship 22 (3), 1996, pp. 181–190. [30] J.C. Nunnally, Psychometric Theory, first ed., McGraw-Hill, New York, 1967. [31] R.L. Oliver, A cognitive model of the antecedents and consequences of satisfaction decisions, Journal of Marketing Research 17 (4), 1980, pp. 460–469. [32] R. Oliver, Satisfaction: A Behavioral Perspective on the Consumer, New York, McGraw-Hill, 1997. [33] A. Parasuraman, L.L. Berry, V.A. Zeithaml, SERVQUAL: a multiple-item scale for measuring consumer perceptions of service, Journal of Retailing 64 (1), 1988, pp. 12–40. H. Landrum et al. / Information & Management 44 (2007) 104–113 [34] A. Parasuraman, L.L. Berry, V.A. Zeithaml, Refinement and reassessment of the SERVQUAL scale, Journal of Retailing 67 (4), 1990, pp. 420–450. [35] A. Parasuraman, L.L. Berry, V.A. Zeithaml, Alternative scales for measuring service quality: a comparative assessment based on psycholometric and diagnostic criteria, Journal of Retailing 70 (3), 1994, pp. 201–230. [36] L.F. Pitt, R.T. Watson, C.B. Kavan, Measuring information systems service quality: concerns for a completer canvas, MIS Quarterly 21 (2), 1997, pp. 209–221. [37] L. Pitt, R.T. Richard, C.B. Bruce, Service quality: a measure of information systems effectiveness, MIS Quarterly 19 (2), 1995, pp. 173–187. [38] A. Rai, S.S. Lang, R.B. Welker, Assessing the validity of IS a success model: an empirical test and theoretical analysis, Information Systems Research 13 (1), 2002, pp. 50–69. [39] A.L. Sanjay, S. Devaraj, An empirical comparison of statistical construct validation approaches, IEEE Transactions on Engineering Management 48 (3), 2001, pp. 319–329. [40] P. Seddon, M. Kiew, A partial test and development of the DeLone and McLean model of IS success, in: J.I. DeGross, S.L. Huff, M.C. Munro (Eds.), in: Proceedings of the 15th International Conference on Information Systems, Vancouver, British Columbia, Canada, 1994, pp. 99–110. [41] A.H. Segars, V. Grover, Re-examining perceived ease of use and usefulness: a confirmatory factor analysis, MIS Quarterly 17 (4), 1993, pp. 517–525. [42] F. Selness, H. Hansen, The potential hazard of self-service in developing customer loyalty, Journal of Service Research 4 (2), 2001, pp. 79–90. [43] D.W. Straub, Validating instruments in MIS research, MIS Quarterly 13 (2), 1989, pp. 147–165. [44] T.P. Van Dyke, L.A. Kappelman, V.R. Prybutok, Measuring information systems service quality: concerns on the use of the SERVQUAL questionnaire, MIS Quarterly 21 (2), 1997, pp. 195–208. [45] T.P. Van Dyke, V.R. Prybutok, L. Kappelman, Cautions on the use of the SERVQUAL measure to assess the quality of information systems services, Decision Sciences 30 (3), 1999, pp. 877–892. [46] R.T. Watson, L.F. Pitt, C.B. Kavan, Measuring information systems service quality: lessons from two longitudinal case studies, MIS Quarterly 22 (1), 1998, pp. 61–79. [47] C. Wright, O.N. Martin, Service quality evaluation in the higher education sector: an empirical investigation of students’ perceptions, Higher Education Research & Development 21 (1), 2002, pp. 23–39. [48] V.A. Zeithaml, A. Parasuraman, L.L. Berry, Delivering Quality Service: Balancing Customer Perceptions and Expectations, Macmillan, New York, NY, 1990. 113 [49] M. Zrivan, N. Pliskin, R. Levin, Measuring user satisfaction and perceived usefulness in the ERP context, Journal of Computer Information Systems 45 (3), 2005, pp. 43–52. Hollis Landrum retired as an information systems management specialist with the US Army Corps of Engineers at the Engineer Research and Development Center. He received a PhD in information science from the University of North Texas in 1999. He has published numerous articles in information and library science. In addition, he has published articles in national music journals and has had several plays produced. He has won numerous awards for his work with the Army Corps of Engineers, including the Commander’s Award for Civilian Service, one of the highest civilian awards in the Army. Victor R. Prybutok is a regents professor of decision sciences in the Information Technology and Decision Sciences Department and Director of the Center for quality and productivity in the College of Business Administration at the University of North Texas. He received a PhD in from Drexel University in 1984. He is a senior member of the American Society for Quality (ASQ) and is an ASQ certified quality engineer, certified quality auditor, and certified quality manager. Journals where his published articles have appeared include The American Statistician, Communications of the ACM, Communications in Statistics, Data Base, Decision Sciences, European Journal of Operational Research, IEEE Transactions on Engineering Management, MIS Quarterly, OMEGA: The International Journal of Management Science, and Operations Research. In addition, he serves on the editorial board of the Quality Management Journal Xiaoni Zhang is an assistant professor of business informatics at the Northern Kentucky University. She received her PhD in business computer information systems from the University of North Texas in 2001. Her research interests include software volatility, Web usability, e-commerce systems, enterprise systems and mobile technology. She is a member of Decision Science Institute and Association for Information System. Her publications appear in IEEE Transactions on Engineering Management, Communication of the ACM, International Conference of Information Systems, and other journals.