Saturday, March 30, 2019
SERVQUAL MODEL as a Service Quality Measure
SERVQUAL MODEL as a Service Quality Measure1.0 IntroductionA great occupy of inspection and repair- timbre look for in re centime decades has been devoted to the development of measures of utility smell. In situation, the SERVQUAL cats-paw (Parasuraman et al., 1988) has been widely applied and valued by academics and practicing managers (Buttle, 1996). However, several studies create place potential difficulties with the hire of SERVQUAL (Carman, 1990 Cronin and Taylor, 1992 Asubonteng et al., 1996 Buttle, 1996 Van Dyke et al., 1997 Llosa et al., 1998). These difficulties comport related to the use of alleged(prenominal) variation go a ports, the ambiguity of the definition of consumer expectations, the st competency of the SERVQUAL graduated table over time, and the dimensionality of the instrument. As a core of these upbraidings, questions subscribe been raised regarding the use of SERVQUAL as a measure of religious go whole tone.1.1 The SERVQUAL exceedWhen the SERVQUAL plate was develop by Parasuraman et al. (1985, 1988), theiraim was to provide a generic instrument for amount proceeds select across a broad range of portion categories. Relying on information from 12 contract groups of consumers, Parasuraman et al. (1985) pass overed that consumers valuated service woodland by comparing expectations (of service to be received) with perceptions (of service actually received) on ten dimensions tangibles, dependableness, responsiveness, communication, credibility, security, competence, rationality/knowing customers, adroitness, and access. In a later (Parasuraman et al. (1988) exercise, the authors cut back the passe-partout ten dimensions to volt(1) tangibles (the appearance of physical facilities, equipment, and personnel)(2) reliability (the ability to commit the promised service dep quitably and accurately)(3) responsiveness (the willingness to help customers and provide nimble service)(4) empathy (the provision of i ndividual cargon and attention to customers) and(5) assurance (the knowledge and discretion of employees and their ability to inspire trust and confidence).Each dimension is measured by foursome to cinque accompaniments (making a total of 22 items across the five dimensions). Each of these 22 items is measured in both ways(1) the expectations of customers disturbing a service and(2) the perceived levels of service actually provided.In making these measurements, respondents atomic quash 18 asked to indicate their degree of agreement with certain statements on a s until now-point Likert-type scale (1 powerfully disagree to 7 strongly agree). For each item, a so-called kerfuffle explanation (G) is and soly calculated as the difference amidst the raw perception-of- murder score (P) and the raw expectations score (E). The greater the gap score (calculated as G P minus E), the high uper the score for perceived service prize.Chapter 2 Literature Review2.0 IntroductionDespit e the far-flung use of the SERVQUAL model to measure service quality, several suppositional and experimental criticisms of the scale have been raised. Buttle (1996) summa organised the major criticisms of SERVQUAL in devil broad categories theoretical and ope symmetrynal.Theoretical issues compriseParadigmatic objections SERVQUAL is based on a disconfirmation double rather than an attitudinal epitome and SERVQUAL fails to draw on established economic, statistical and mental theory.Gaps model there is little evidence that customers assess service quality in terms of P E gaps.Process orientation SERVQUAL focuses on the process of service delivery, non the solvents of the service encounter.Dimensionality SERVQUALs five dimensions argon not universals the cast of dimensions comprising SQ is contextualized items do not always lode on to the operators which unity would a priori expect and there is a high degree of intercorrelation betwixt the five RATER dimensions.Operationa l criticisms includeExpectations the term expectation is polysemic consumers use banners separate than expectations to evaluate SQ and SERVQUAL fails to measure absolute SQ expectations.Item composition four or five items can not capture the variability within each SQ dimension.Moments of truth (MOT) customers assessments of SQ whitethorn vary from MOT to MOT.Polarity the reversed augury of items in the scale causes respondent error.Scale points the seven-point Likert scale is flawed.deuce administrations two administrations of the instrument cause boredom and confusion.Variance extracted the over SERVQUAL score accounts for a disappointing proportion of item pas seuls.The above criticism will be discussed below.2.1 Paradigmatic objections (Theoretical Criticisms)Two major criticisms have been raised. First, SERVQUAL has been inappropriately based on an expectations disconfirmation model rather than an attitudinal model of SQ. Second, it does not build on existent knowledge in economics, statistics and psychology. SERVQUAL is based on the disconfirmation model widely take in the customer rejoicing publications. In this literature, customer satisfaction (CSat) is operationalised in terms of the relationship in the midst of expectations (E) and outcomes (O). If O matches E, customer satisfaction is predicted. If O exceeds E, then customer delight may be produced. If E exceeds O, then customer dissatisfaction is indicated. harmonize to Cronin and Taylor (1992 1994) SERVQUAL is paradigmatically flawed because of its ill-judged adoption of this disconfirmation model. Perceived quality, they claim, is beat out conceptualised as an view. They criticise Parasuraman et al. for their hesitancy to define perceived SQ in attitudinal terms, even though Parasuraman et al. (1988) had early claimed that SQ was similar in m whatsoever ways to an attitude.Cronin and Taylor signify Researchers have attempted to differentiate service quality from consumer sati sfaction, even opus exploitation the disconfirmation format to measure perceptions of service quality this approach is not consistent with the differentiation expressed betwixt these constructs in the satisfaction and attitude literatures.Iacobucci et al.s (1994) review of the debate surrounding the conceptual and operational differences between SQ and CSat concludes that the constructs have not been consistently defined and differentiate from each otherwise in the literature. She suggests that the two constructs may be connected in a number of ways. First, they may be both different operationalisations of the uniform construct, evaluation. Second, they may be orthogonally related, i.e. they may be entirely different constructs. Third, they may be conceptual cousins. Their family connections may be dependent on a number of other considerations, including for example, the duration of the evaluation. Parasuraman et al. (1985) have described satisfaction as more situation- or enc ounter- item, and quality as more ho add upic, developed over a protracted period of time, although they offer no empirical evidence to support this contention. SQ and CSat may likewise be related by time order. The pre governing ruling is that SQ is the logical predecessor to CSat, but this remains unproven. Cronin and Taylors critique draws support from Olivers (1980) research which suggests that SQ and CSat ar distinct constructs but ar related in that satisfaction mediates the effect of prior-period perceptions of SQ and causes revised SQ perceptions to be formed. SQ and CSat may also be differentiated by virtue of their content. Whereas SQ may be thought of as high in cognitive content, CSat may be more heavily slopped with affect (Oliver, 1993). Cronin and Taylor suggest that the adequacy-importance model of attitude measurement should be adopted for SQ research. Iacobucci et al. (1994) add the observation that in somewhat general psychological sense, it is not clear wh at short-term evaluations of quality and satisfaction are if not attitudes. In turn, Parasuraman et al. (1994) have vigorously defended their position, claiming that critics seem to discount prior conceptual act as in the SQ literature, and suggest that Cronin and Taylors rifle does not justify their claim that the disconfirmation paradigm is flawed.In other work, Cronin and Taylor (1994) explanation that Recent conceptual advances suggest that the disconfirmation-based SERVQUAL scale is measuring neither service quality nor consumer satisfaction. Rather, the SERVQUAL scale appears at best an operationalisation of entirely one of the many forms of expectancy disconfirmation.A different concern has been raised by Andersson (1992). He objects to SERVQUALs failure to draw on former social science research, particularly economic theory, statistics, and psychological theory. Parasuraman et al.s work is exceedingly inductive in that it moves from historically situated observation to general theory.Andersson (1992) claims that Parasuraman et al. forsake the principle of scientific continuity and deduction. Among specific criticisms are the future(a)First, Parasuraman et al.s steering technology takes no account of the costs of improving service quality. It is nave in assuming that the peripheral revenue of SQ improvement always exceeds the marginal cost. (Aubrey and Zimbler, 1983., Crosby., 1979, Juran., 1951 and Masser., 1957) have addressed the issue of the costs/benefits of quality improvement in service settings.)Second, Parasuraman et al. collect SQ data using ordinal scale methods (Likert scales) except perform analyses with methods suited to interval-level data (factor analysis).Third, Parasuraman et al. are at the absolute end of the street regarding possibilities to use statistical methods. Ordinal scales do not go away for investigations of common product-moment correlations. Interdependencies among the dimensions of quality are difficult to des cribe. SERVQUAL studies cannot answer questions such as Are there elasticities among the quality dimensions? Is the customer value of improvements a elongate or non-linear function?Fourth, Parasuraman et al. fail to draw on the large literature on the psychology of perception.2.2 Gaps ModelA related set of criticisms refer to the value and humbleing of gaps identified in the disconfirmation model. Babakus and Boller (1992) tack the use of a gap approach to SQ measurement intuitively appealing but hazard that the difference scads do not provide any spare information beyond that already contained in the perceptions agent of the SERVQUAL scale. They found that the dominant contributor to the gap score was the perceptions score because of a generalised reception tendency to rate expectations high.Churchill and Surprenant (1982), in their work on CSat, also excogitate whether gap measurements contribute anything new or of value given that the gap is a direct function of E and P. It has also been noted that firearm conceptually, difference scores might be sensible, they are problematic in that they are notoriously unreliable, even when the measures from which the difference scores are derived are themselves highly reliable (Iacobucci et al., 1994).Also, in the context of CSat, Oliver (1980) has pondered whether it might be preferent to consider the P E scores as raw differences or as ratios. No work has been get acrossed using a ratio approach to measure SQ. Iacobucci et al. (1994) take a different tack on the incorporation of E-measures. They suggest that expectations might not exist or be formed clearly enough to serve as a standard for evaluation of a service experience. Expectations may be formed concurrently with service enjoyment. Kahneman and Miller (1986) have also proposed that consumers may form experience-based norms subsequently service experiences, rather than expectations before.A further issue raised by Babakus and Inhofe (1991) is th at expectations may attract a social desirability response bias. Respondents may experience motivated to adhere to an I-have-high-expectations social norm. Indeed, Parasuraman et al. report that in their interrogation of the 1988 version the majority of expectations scores were above half dozen on the seven-point scale. The general mean expectation was 6.22 (Parasuraman et al., 1991b).Teas (1993a 1993b 1994) has pondered the meaning of identified gaps. For example, there are cardinal ways of producing P E gaps of -1 (P = 1, E = 2 P = 2, E = 3 P = 3, E = 4 P = 4, E = 5 P = 5, E = 6 P = 6, E = 7). Do these fastened gaps mean equal perceived SQ? He also notes that SERVQUAL research thus far has not established that all service providers within a consideration or picking set, e.g. all car-hire firms do, in fact, share the corresponding expectations ratings across all items and dimensions.A further criticism is that SERVQUAL fails to capture the dynamics of changing expectations . Consumers learn from experiences. The inference in much of Parasuraman et al.s work is that expectations rise over time. An E-score of seven in 1986 may not necessarily mean the same as an E-score in 1996. Expectations may also fall over time (e.g. in the health service setting). Grnroos (1993) recognises this weakness in our understanding of SQ, and has called for a new phase of service quality research to focus on the dynamics of service quality evaluation. Wotruba and Tyagi (1991) agree that more work is needed on how expectations are formed and changed over time.Implicit in SERVQUAL is the assumption that positive and negative disconfirmations are symmetrically valent. However, from the customers perspective, failure to butt on expectations often seems a more world-shattering outcome than success in meeting or exceeding expectations (Hardie et al., 1992). Customers will often criticise inadequate service performance and not praise elisional performance.Recently, Cronin and Taylor (1992) have tried a performance-based measure of SQ, dubbed SERVPERF, in four industries (banking, pest control, dry modify and fast food). They found that this measure explained more of the variance in an general measure of SQ than did SERVQUAL. SERVPERF is composed of the 22 perception items in the SERVQUAL scale, and therefore excludes any consideration of expectations. In a later defence of their argument for a perceptions-only measure of SQ, Cronin and Taylor (1994) acknowledge that it is possible for researchers to infer consumers disconfirmation through arithmetic way of life (the P E gap) but that consumer perceptions, not calculations, govern behavior. Finally, a group of researchers, including Zeithaml herself (Boulding et al., 1993), has recently rejected the value of an expectations-based or gap-based model in purpose that service quality was only influenced by perceptions.2.3 Process orientationSERVQUAL has been criticized for nidus on the process of se rvice delivery rather than outcomes of the service encounter. Grnroos (1982) identified three servings of SQ technical, operating(a) and reputational quality. Technical quality is concerned with the outcome of the service encounter, e.g. have the dry cleaners got rid of the stain? Functional quality is concerned with the process of service delivery, e.g. were the dry cleaners counter provide courteous? Reputational quality is a reflection of the corporate image of the service organization. While technical quality focuses on what, functional quality focuses on how and involves consideration of issues such as the behaviour of customer contact round, and the despatch of service. Critics have argued that outcome quality is missing from Parasuraman et al.s formulation of SQ (Cronin and Taylor, 1992 Mangold and Babakus, 1991 Richard and Allaway, 1993). Richard and Allaway (1993) time-tested an augmented SERVQUAL model which they claim incorporates both process and outcome components , and comment that the challenge is to determine which process and outcome quality attributes of SQ have the great impact on choice1. Their research into Domino Pizzas process and outcome quality employed the 22 Parasuraman etal. (1988) items, circumscribed to suit context, and the following six outcome items(1) Dominos has delicious home-delivery pizza.(2) Dominos has nutritious home-delivery pizza.(3) Dominos home-delivery pizza has flavourful sauce.(4) Dominos provides a generous amount of toppings for its home-delivery pizza.(5) Dominos home-delivery pizza is make with superior ingredients.(6) Dominos prepared its home-delivery pizza crust exactly the way I like it.These researchers found that the process-only items borrowed and adapted from SERVQUAL accounted for only 45 per cent of the variance in customer choice the full inventory, inclusive of the six outcome items, accounted for 71.5 per cent of variance in choice. The difference between the two is significant at the 0.0 01 level. They conclude that process-and-outcome is a better predictor of consumer choice than process, or outcome, alone. In defense of SERVQUAL, Higgins et al., (1991) have argued that outcome quality is already contained within these dimensions reliability, competence and security.2.4 DimensionalityCritics have raised a number of significant and related questions active the dimensionality of the SERVQUAL scale. The most(prenominal) serious are concerned with the number of dimensions and their stability from context to context. There seems to be general agreement that SQ is a second-order construct, that is, it is factorially complex, being composed of several first-order variables 2. SERVQUAL is composed of the five RATER 3 factors. There are however, several alternative conceptualizations of SQ. As already noted, Grnroos (1984) identified three components technical, functional and reputational quality Lehtinen and Lehtinen (1982) also identify three components interactive, p hysical and corporate quality Hedvall and Paltschik (1989) identify two dimensions willingness and ability to serve, and physical and psychological access Leblanc and Nguyen (1988) list five components corporate image, internal organisation, physical support of the service producing system, staff/customer interaction, and the level of customer satisfaction.Parasuraman et al. (1988) have claimed that SERVQUAL provides a basic material body through its expectations/perceptions format encompassing statements for each of the five service quality dimensions. The skeleton, when necessary, can be adapted or supplemented to fit the characteristics or specific research needs of a particular organization.In their 1988 paper, Parasuraman et al. also claimed that the last 22-item scale and its five dimensions have sound and stable psychometric properties. In the 1991b revision, Parasuraman et al. found evidence of consistent factor structure across five independent samples. In other words, they make claims that the five dimensions are generic across service contexts. Indeed, in 1991, Parasuraman et al. claimed that SERVQUALs dimensions and items represent core evaluation criteria that decease specific companies and industries (1991b) 4.2.5 Number of dimensionsWhen the SERVQUAL instrument has been employed in modified form, up to nine distinct dimensions of SQ have been revealed, the number varying according to the service sector under investigation. One study has even produced a single-factor solution. 9 factors accounted for 71 per cent of SQ variance in Carmans (1990) hospital research access service, tangible accommodations, tangible food, tangible privacy, nursing care, explanation of treatment, access and courtesy afforded visitors, discharge planning, and patient accounting (billing)5.Five factors were distinguished in Saleh and Ryans (1992) work in the hotel industry conviviality, tangibles, reassurance, avoid sarcasm, and empathy. The first of these, co nviviality, accounted for 62.8 per cent of the overall variance the second factor, tangibles, accounted for a further 6.9 per cent the five factors together accounted for 78.6 per cent. This is strongly suggestive of a two-factor solution in the hospitality industry. The researchers had initially presume that the factor analysis would confirm the SERVQUAL dimensions but this failed to be the case.Four factors were extracted in Gagliano and Hathcotes (1994) investigation of SQ in the retail clothing sector personal attention, reliability, tangibles and convenience. Two of these have no correspondence in SERVQUAL. They conclude the skipper SERVQUAL scale does not perform as well as expected in apparel intensity retailing. Three factors were identified in Bouman and van der Wieles (1992) research into car avail customer kindness, tangibles and faith 6. The authors were not able to find the same dimensions for settle service quality as did Berry et al.One factor was accept in Ba bakus et al.s (1993b) survey of 635 utility company customers. Analysis essentially produced a single-factor model of SQ which accounted for 66.3 per cent of the variance. The authors advance several possible explanations for this unidimensional result including the nature of the service, (which they describe as a low-involvement service with an ongoing consumption experience), non-response bias and the use of a single expectations/perceptions gap scale. These researchers concluded With the exception of findings reported by Parasuraman and his colleagues, empirical evidence does not support a five-dimensional concept of service quality.In summary, Babakus and Boller (1992) commented that the domain of service quality may be factorially complex in some industries and very simple and unidimensional in others. In effect, they claim that the number of SQ dimensions is dependent on the particular service being offered. In their revised version, Parasuraman et al. (1991b) suggest two rea sons for these anomalies. First, they may be the product of differences in data collection and analysis procedures. A more plausible explanation is that differences among empirically derived factors across takingss may be primarily due to across-dimension similarities and/or within dimension differences in customers evaluations of a specific company involved in each setting.Spreng and Singh (1993) have commented on the lack of discrimination between several of the dimensions. In their research, the correlation between self-assertion and Responsiveness constructs was 0.97, indicating that they were not separable constructs. They also found a high correlation between the combined Assurance-Responsiveness construct and the Empathy construct (0.87). Parasuraman et al. (1991b) had earlier found that Assurance and Responsiveness items loaded on a single factor and in their 1988 work had found mediocre intercorrelations among the five dimensions of 0.23 to 0.35.In testing their revised version (Parasuraman et al., 1991b), Parasuraman and colleagues found that the four items under Tangibles skint into two distinct dimensions, one pertaining to equipment and physical facilities, the other to employees and communication materials. They also found that Responsiveness and Assurance dimensions showed considerable overlap, and loaded on the same factor. They suggested that this was a product of imposing a five-factor constraint on the analyses. Indeed, the additional degrees of freedom allowed by a subsequent six-factor solution generated distinct Assurance and Responsiveness factors.Parasuraman et al., (1991a) have now accepted that the five SERVQUAL dimensions are coordinated as evidenced by the need for oblique rotations of factor solutionsto declare the most interpretable factor patterns. One fruitful area for rising research, they conclude, is to explore the nature and causes of these interrelationships. It therefore does appear that both contextual fortune and analytical processes have some bearing on the number of dimensions of SQ.2.6 contextual stabilityCarman (1990) tested the generic qualities of the SERVQUAL instrument in three service settings a tyre retailer, a business school placement contract and a alveolar consonant school patient clinic. Following Parasuraman et al.,s suggestion, he modified and augmented the items in the original ten-factor SERVQUAL scale to suit the three contexts. His factor analysis identified between five and seven underlying dimensions. According to Carman, customers are at least partly context-specific in the dimensions they employ to evaluate SQ. In all three cases, Tangibles, Reliability and gage were present 7. Responsiveness, a major component in the RATER scale, was relatively weak in the dental clinic context.Carman also commented Parasuraman, Zeithaml and Berry combined their original Understanding and admission price dimensions into Empathy our results did not find this to be an appropriate combination. In particular he found that if a dimension is very important to customers they are likely to be decomposed into a number of sub-dimensions. This happened for the placement centre where Responsiveness, Personal attention, opening and thingamajig were all identified as separate factors. According to Carman, this indicates that researchers should work with the original ten dimensions, rather than adopt the revised five-factor Parasuraman et al., (1988) model.2.7 Item loadingsIn some studies (e.g. Carman, 1990), items have not loaded on the factors to which they were expected to belong. Two items from the Empathy stamp battery of the Parasuraman et al., (1988) instrument loaded heavily on the Tangibles factor in a study of dental clinic SQ. In the tyre retail study, a Tangibles item loaded on to Security in the placement centre a Reliability item loaded on to Tangibles. An item concerning the ease of making appointments loaded on to Reliability in the dental clinic cont ext, but Security in the tyre store context. He also found that only two-thirds of the items loaded in the same way on the expectations battery as they did in the perceptions battery. Carman supplies other examples of the same phenomena, and suggests that the unexpected results indicate both face rigorousness and a construct hardship problem. In other words, he warns against importing SERVQUAL into service setting contexts without modification and validity checks.Among his specific recommendations is the following We recommend that items on Courtesy and Access be retained and that items on some dimensions such as Responsiveness and Access be expand where it is believed that these dimensions are of particular importance. He also reports specific Courtesy and Access items which performed well in terms of nomological and construct validity.Carman (1990) further suggested that the factors, Personal attention, Access or Convenience should be retained and further contextualised researc h work be done to identify their significance and meaning.2.8 Item correlationsConvergent validity and discriminant validity are important considerations in the measurement of second-order constructs such as SERVQUAL. One would associate a high level of convergent validity with a high level of intercorrelations between the items selected to measure a single RATER factor. Discriminant validity is indicated if the factors and their component items are independent of each other (i.e. the items load heavily on one factor only). Following their modified replication of Parasuraman et al.,s work, Babakus and Boller (1992) conclude that rules for convergence and discrimination do not indicate the earth of the five RATER dimensions.The best scales have a high level of intercorrelation between items comprising a dimension (convergent validity). In their development work in four sectors (banking, credit-card company, repair and maintenance company, and long-distance telecommunications company ) Parasuraman et al., (1988) found inter-item reliability coefficients (alphas) varying from 0.52 to 0.84. Babakus and Boller (1992) report alphas which are broadly consistent with those of Parasuraman, varying from 0.67 to 0.83 (see Table III). In their 1991b version, Parasuraman et al. report alphas from 0.60 to 0.93, and observe that every alpha value obtained for each dimension in the final study is higher than the corresponding values in theoriginal study. They attribute this improvement to their rewording of the 22 scale items.Spreng and Singh (1993), and Brown et al., (1993) are highly critical of the questionable application of alphas to difference scores. They evaluate the reliability of SERVQUAL using a measure specifically designed for difference scores (Lord, 1963). Spreng and Singh conclude that there is not a great deal of difference between the reliabilities correctly calculated and the more common alpha calculation, an observation with which Parasuraman et al., (1993 ) concurred when they wrote The collective conceptual and empirical evidence neither demonstrates clear superiority for the non-difference score format nor warrants abandoning the difference score format.2.9 Expectations (Operational Criticisms)Notwithstanding the more fundamental criticism that expectations play no significant role in the conceptualization of service quality, some critics have raised a number of other concerns about the operationalization of E in SERVQUAL.In their 1988 work, Parasuraman et al. defined expectations as desires or wants of consumers, i.e. what they feel a service provider should offer rather than would offer (emphasis added). The expectations component was designed to measure customers normative expectations (Parasuraman et al., 1990), and is similar to the ideal standard in the customer satisfaction/dissatisfaction literature (Zeithaml et al., 1991).Teas (1993a) found these explanations somewhat vague and has questioned respondents recitation of th e expectations battery in the SERVQUAL instrument. He believes that respondents may be using any one of six interpretations (Teas, 1993b)(1) Service attribute importance. Customers may respond by rating the expectations statements according to the importance of each.(2) Forecasted performance. Customers may respond by using the scale to predict the performance they would expect.(3) Ideal performance. The optimal performance what performance can be.(4) Deserved performance. The performance level customers, in the light of their investments, feel performance should be.(5) Equitable performance. The level of performance customers feel they ought to receive given a perceived set of costs.(6) Minimum tolerable performance. What performance moldiness be? Each of these interpretations is somewhat different, and Teas contends that a considerable percentage of the variance of the SERVQUAL expectations measure can be explained by the difference in respondents interpretations.Accordingly, the expectations component of the model lacks discriminant validity. Parasuraman et al. (1991b 1994) have responded to these criticisms by redefining expectations as the service customers would expect from gauzy service organizations, rather than normative expectations of service providers, and by vigorously defend their inclusion in SQ research. Iacobucci et al. (1994) want to drop the term expectations from the SQ vocabulary. They prefer the generic label standard, and believe that several standards may operate simultaneously among them ideals, my most desired combination of attributes, the industry standard of a nominal average competitor, deserved SQ, and brand standards based on past experiences with the brand.Some critics have questioned SERVQUALs failure to access customer evaluations based on absolute standards of SQ. The instrument asks respondents to report their expectations of excellent service providers within a class (i.e. the measures are relative rather than absolute). It has be
Subscribe to:
Post Comments (Atom)
No comments:
Post a Comment
Note: Only a member of this blog may post a comment.