INTRODUCTION

Consumer-based brand equity (CBBE) is defined as the added value that a brand name endows upon a product as a result of a firm's marketing efforts, measured from the psychology approach. It has been considered as a sustainable competitive advantage that positively affects companies’ future profits and long-term cash flow.1 Numerous measures of CBBE2, 3 have been used since it was first conceptualized by Aaker4 and Keller.5 Aaker4 identified five components of brand equity as awareness, associations, perceived quality, loyalty and other proprietary assets such as patents and trademarks. Keller5 focused on brand knowledge, including awareness and unique favorable beliefs. More CBBE components such as willing to pay a price premium and uniqueness were later identified by CBBE researchers.3 Most of them are developed to differentiate consumers from each other.

Wang and Finn6 used Multivariate Generalizability Theory (MGT) to develop the first brand level CBBE scale. It was designed to maximally differentiate brands from each other and reliably capture their differences in a brand portfolio context. MGT approach quantifies all the design factors involved and increases the validity and reproducibility of the CBBE estimates. But it has no interest in correcting the observed measures (that is, brand ratings) for differences among the various elements (such as individual respondents, dimensions, items) encountered by each brand during a data collection design. In order to achieve a contextually unbiased, objective estimate of brand equity, individual differences, together with other extraneous sources, must be accounted for. Removing the biases can also make the outcome of the measurement to be generalized beyond the specific elements of the measurement, such as the particular consumers, dimensions, items and so on. To address this issue, we introduce Many-facet Item Response Theory (MFIRT) to complement the information provided by MGT analysis.

In general, the goal of MGT is to examine measurement scale in terms of some index of reliability, deal with the precision of observed raw data and optimize a future data collection design by minimizing the influence of identified sources of error variances, whereas the purpose of MFIRT is to correct the observed measure for extraneous differences within each facet of a data collection design.7 These two approaches are complementary. MGT approach provides general, group-level information particularly in making overall decisions about questionnaire design and the unit of analysis is the total score, whereas MFIRT offers more specific information at the individual respondent, item, dimension and interaction level and the unit of analysis is individual brand rating. These two measurement approaches, used together, allow for balanced and thorough interpretations of brand performance assessment data and offer better ways of improving performance measurement.

The current research focuses on how to apply MFIRT to get objective and unbiased estimates of brands in a brand portfolio context (including both master brands such as Coca-Cola and sub-brands such as Diet Coke and Coke Zero) and how sub-brands differentiate themselves from each other and from the master brands in terms of the brand equity value. As in Wang and Finn's paper (in press), brand portfolio in the current work is defined as a branded house rather than a house of brands.8 A branded house consists of a single master brand and its sub-brands. For example, Disney is the master brand and Disney World, Disney Land, Disney Picture and Disney DVD can be seen as sub-brands. In contrast, a house of brands contains independent, unconnected brands. Consumers may not realize that the independent and unconnected house of brands belong to the same company's portfolio. For example, when consumers buy Folgers Coffee, most are probably unaware that Folgers is a division of Procter & Gamble.

METHODOLOGY

Item Response Theory (IRT) is an advanced testing theory that has been accepted as a better alternative to Classical Test Theory (CTT) in the areas of education psychology and psychometrics.9, 10 It is the simplest two-facet (that is, consumers and items) version of the more general MFIRT. The marketing literature has relied largely on the application of CTT-based approaches for long time (see Balasubramaniam and Kamakura11 and Singh et al12 as exceptions). More recently, more applications of IRT in the marketing literature have started to appear.13, 14, 15 But these are still traditional IRT modeling in terms of facets. For example, de Jong et al13 examined measurement invariance in a cross-country study using an IRT model. In their study, consumers were nested within countries and country was treated as an extraneous factor rather than a research focus. There was no separate country facet in their model.

Traditional IRT models cannot fully capture the complexity of CBBE (that is, multi-dimensional with multiple sources of systematic variance) especially in a brand portfolio context. We introduce MFIRT, originally developed by Linacre16 in educational psychology, which extends IRT models from two facets (that is, consumers and items), to model additional facets (such as brands and dimensions) in the responses. Moreover, we combine MFIRT method with the Rating scale model17 to accommodate both the multi-facet and the ordinal character of CBBE data.

In order to estimate the brand equity of master brands and sub-brands simultaneously after accounting for extraneous sources of variance (for example, individual differences, dimensions, items and response categories), we specify the model as follows:

where P rbmdik and Prbmdi(k−1) are the probability of choosing category k and k−1, respectively, on item i in dimension d for person r on sub-brand b which belongs to master brand m. P r is the criticality of person r; B b is the brand equity of sub-brand b; M m is the brand equity of master brand m; I i is the threshold of item i; D d is the dimensional threshold of dimension d; C k is the threshold of category k relative to category k−1.

B b and M m are the major research interest. After accounting for extraneous factors (such as P r , I i and D d ) that are confounded with B b and M m in the brand equity evaluations, B b and M m give us contextually unbiased estimates. Individual differences are examined by P r . A high threshold for P r means that this particular consumer tends to give lower responses to all the brands evaluated, other things (including brand equity) being equal. It could be interpreted that this consumer is harder to please. By introducing C k , the proposed approach makes a weaker ordinal assumption about the categories.18 In particular, ordered step calibration is designed to capture the phenomenon that as a consumer moves up the leniency continuum or as a brand moves up the brand equity continuum, each response category in turn becomes the most probable response. Evidence of disordered average measurement and/or step calibration questions the functioning of the Likert scale. We use I i and D d to examine the different characteristics of items nested within dimensions. A high threshold for I i or D d means it is harder for consumers to agree with a specific item/dimension. A low threshold means an item/dimension is agreed to more easily.

Moreover, in the model, (a) the contribution of each facet (that is, consumers, brands, items nested within dimensions and response categories) to the evaluations is captured in a single parameter with a value independent of all other parameters within the frame of reference. It is the brand parameter value (that is, a brand's equity) that a CBBE measurement study is majorly intended to determine. This single parameterization is necessary if, for example, brands are to be arranged in one order of merit. (b) The parameters combine additively to produce the evaluations. Additivity implies that all the parameters share one linear scale. Not only does linearity assist brand managers in understanding the meaning underlying the data, it also provides a useful basis for further analysis, because many commonly used statistical procedures assume linearity in their data. (c) The estimate of any parameter is dependent on the accumulation of all responses in which it participates, but is independent of the particular value of any of those responses. This axiom of ‘local independence’ allows the statistical estimates of the measures to be as free as possible of which particular consumers evaluate which particular brands on which particular dimensions and items.19 Therefore, the proposed approach not only has meaning generalized beyond the local details of the evaluation situation,20 but also has the capability to accommodate more flexible data collection designs and handle missing data.

EMPIRICAL APPLICATION

The fully crossed (brands by respondents) data set used by Wang and Finn6 is examined here to demonstrate the complementary roles played by MFIRT to MGT analysis. It was collected in 2008 from 254 North American undergraduate students, who received course credit for their participation. The students evaluated six soft drink brands (that is, Classic Coke, Diet Coke, Coke Zero, Pepsi, Diet Pepsi and Diet Pepsi Max). The brand evaluation instrument consisted of 25 items (as in Table 3) capturing the five dimensions (that is, brand awareness, brand associations, brand loyalty, perceived quality and uniqueness). All items were positively worded and used a 7-point Likert scale from ‘strongly disagree’ (1) to ‘strongly agree’ (7).

We choose a Bayesian approach and use WinBUGs to estimate the brand portfolio model. The total brand equity of a sub-brand is modeled as the sum of its individual brand equity and the equity it gets from its master brand.

Bayesian analysis requires the specification of the models based on prior distributions for model parameters and hence random samples are simulated from the posterior distributions through simulated Markov chain procedures.21 To assess the convergence and Monte Carlo variances, we use two chains with differing starting values and calculate the Gelman-Rubin statistic, as modified by Brooks and Gelman,22 which compares within- to between-chain variability. The ratio R=B/W<1.05 for all parameter estimates signifies that convergence is attained. A total of 12 000 iterations were obtained with the first 4000 as burn-in.

Model identification is attained by imposing necessary constraints. The vectors of brand equity dimension thresholds (D d ) sum to zero and the main effect for master brand Coca-Cola is fixed at 1. The sub-brand main effects (B b ) are also set to sum to zero within each family brand group. Similarly, item thresholds (I i ) are set to sum to zero within each dimension. Consumer criticality (P r ) is drawn from a normal distribution with the mean fixed to zero.

In Table 1, we list estimates of CBBE for all six sub-brands with the two master brands after accounting for the influence of persons, items and dimensions.

Table 1 Brand equity measures from MFIRT (master brands and sub-brands)

When comparing the two master brands, the Coke family has higher CBBE than the Pepsi family (1 versus 0.807). After accounting for the master brands, Classic Coke as a sub-brand itself has the highest CBBE (0.266) and Diet Coke has the lowest one (−0.187). After combining with the brand equity sub-brands get from their master brands, Classic Coke still has the highest brand equity in total (1.266), followed by Original Pepsi (1.036). The sub-brand with the lowest combined brand equity is Diet Pepsi Max (0.634). From the results, we can see that the sub-brands within a brand portfolio do not have the same brand equity. In our empirical examination of CBBE, the sub-brands are different from their master brands and from each other.

Estimates of respondents’ criticality (P r ), item threshold (I i ) and dimension threshold (D d ) are also shown in Tables 2, 3 and 4.

Table 2 MFIRT analysis of consumers – Brand portfolio
Table 3 MFIRT analysis of items – Brand portfolio
Table 4 MFIRT analysis of dimensions – Brand portfolio

The mean criticality of consumers is −0.005, with a standard deviation of 0.283. The range is from −0.926 to 1.083. Higher value means more critical to give an affirmative response to the desirable brand equity statement. The threshold estimates of items and dimensions are shown in Tables 3 and 4. An item/dimension receiving a high raw score suggests that this item/dimension has higher threshold for consumers to give affirmative response. The range of dimension estimates is from −0.485 for brand awareness to 0.119 for uniqueness. That is, brand awareness is the easiest dimension to agree with whereas uniqueness is the most difficult one. Item estimates also vary within and across dimensions. The step calibrations are in ascending order (that is, −0.51, −0.33, −0.15, −0.06, 0.12, 0.33 and 0.61) implying that as one brand moves up the brand equity continuum (or one consumer moves up to be less critical), each category in turn becomes the most probable response.7 Thus, the Likert scale functions well in the current study.

DISCUSSION AND IMPLICATION

Followers of the traditional measurement paradigm23 implicitly assume CBBE scales have only one purpose – scaling people. While common in psychology, it is not always appropriate in marketing especially for CBBE measurement. Wang and Finn6 introduced MGT and developed a brand level scale to fill this gap. It is designed to maximally differentiate brands from each other and reliably capture their differences. This is the first scale that has been developed for CBBE measurement area at the brand level. MGT approach introduced in their study quantifies all the design factors involved and increases the validity and reproducibility of the CBBE estimates. However, MGT has little interest in the brand equity reporting per se. Aggregate evaluations (such as means) are still used for CBBE comparison among brands. An average is a joint property of all dimensions and items in a particular scale and the particular individuals sampled. When dimensions (and items) are added to or dropped from a scale, or respondents change, the average will change. Therefore, the average CBBE at the brand level could be at best contextually biased and at worst an unreliable and invalid estimate. The average also inappropriately assumes the responses are interval-scaled instead of ordinal-scaled as they really are. MFIRT analysis addresses the limitations of this aggregate CBBE evaluation approach. It models brand equity (at the brand level) as a separate term, independent of particular dimensions, items and individuals involved in the evaluation. The ordinal nature of the rating scale is also examined. This model characteristic of independence not only provides better estimates of brand equity but also offers flexibility in terms of data collection.

The present research has several implications. First, it provides a complementary framework for researchers to further understand and measure CBBE beyond the CTT and MGT approaches. It is a new and comprehensive perspective and offers a lot of opportunities to be further developed (for example, examining antecedents and consequences at either consumer or brand level) and adapted to other research questions (such as scaling the interactions between brands and consumers). Second, the proposed scale increases the objectivity and managerial relevance of CBBE. Managers can use this newly developed scale to get reliable brand estimates of CBBE that can be compared with the results from other brand equity measurement approaches (that is, finance and economics approaches) to identify the existence of inconsistency and investigate the reason. It helps managers precisely monitor the response from the consumer side on their brand marketing moves, the crucial step leading to the success of any marketing program, and facilitates timely adjustments. For managers who are interested in the relationships between CBBE and other marketing strategic variables, this work provides unbiased estimates that have high reliability and validity when scaling brands. Using these high-quality brand CBBE estimates as inputs, the correlations they get and the inference they make about the relationships are unbiased. Third, by explicitly separating brand parameters from those for consumers, dimensions and items, this research makes effort to bridge the gap between objective measures and subjective measures in the brand equity measurement area. Researchers are interested in the ‘true’ scores of variables of interest (CBBE). However, they only have observed scores, which are scores that confound the true scores with multiple extraneous sources of variation such as the different response patterns of individual consumers. Some of them may have either lower or higher probability to use extreme response categories (such as 1s and 7s in a 1-7 Likert scale). The current research not only conceptually separates the true value from other sources of variance but also introduces the methodologies into marketing (that is, MFIRT) that enables the separation. Finally, the current research provides a tool for brand portfolio managers to measure the CBBE of their master brands and sub-brands and also offers useful information about the comparative strength of the master brands and their sub-brands.

This research has some limitations that we plan to address in future research. The nature of the dimensions (that is, formative versus reflective) and the possible interactions (such as interactions between brands and consumers) are not recognized in the current MFIRT analysis. The nature of the dimensions can be examined by testing a MFIRT measurement model that incorporates a hierarchical structure. Substantial interactions between brands and consumers are commonly observed in marketing literature and practice. An interaction term can be introduced into a CBBE MFIRT measurement model to quantify this effect.