1 Introduction

In various applied problems the response variable is presented in three category ordinal levels. For instance, in marketing research a dependent variable of overall satisfaction is commonly measured in ordinal Likert 10-point scale from the worst to the best values. With several upper levels corresponding to satisfaction and bottom levels to dissatisfaction, it is a trinomial outcome of satisfaction–middle–dissatisfaction, or positive–neutral–negative segments modeled by various predictors. Modeling for such a trinomial variable is usually performed by a pair of ordinal logistic regressions or by two multinomial share models [5, 18, 19]. Many examples considered in the main statistical packages include: large-medium-small size of soda that people buy in relation to other meals and demographics; gold–silver–bronze medaling in Olympic sport, with relevant predictors of training hours, diet, age, and popularity of this kind of activity in athletes’ home country; three categories of decision by college juniors asked if they are unlikely, somewhat likely, or very likely to apply to graduate school, with possible predictors of parental educational status, public or private undergraduate college, and GPA. Other examples of ordered response are presented in bond ratings, surveys with responses of agree–no opinion–disagree, high–mid–low levels of state spending on government programs, the levels none–partial–full of insurance coverage chosen, and employment status ‘non-employed’–‘employed part-time’–‘fully employed’ [4, pp. 824–827]. Tri-state neural networks [20] and fatigue profiles in experimental physiology [1] have been studied by logistic models with an intermediate and eventually reached levels of saturation. It is interesting to note that division of data to three groups has been also studied for continuous and discrete modeling ([3, 6], and references within). Modeling for triple outcome can be performed in different regression and non-regression approaches [2, 7, 11, 12]. Multinomial, ordinal logistic model, and other mentioned techniques are complicated numerical procedures which require special software.

The current work shows that the problem of three level modeling can be reduced to a much simpler and convenient one binomial logit model, so the categorical outcome can be transformed into one probability scale. It can be done in the approach based on a special data transformation used in the Best–Worst scaling or MaxDiff technique known in marketing research, which is a contemporary method for the prioritization of items proposed in [1416], developed and applied in numerous works (for instance, [810, 21, 22]. In MaxDiff, the positive-neutral data subset is stacked with the negative-neutral subset in which the signs of predictors are changed to opposite. The binary dependent variable equals one for positive-negative outcomes and zero for neutral outcomes, respectively. Logistic regression is constructed where the positive category predictions are close to 1, negative are close to 0, and neutral are in the middle of its continuous 0–1 scale.

2 MaxDiff and Trinomial Modeling

A general multinomial logit (MNL) defines a choice among several outcomes and can be described by the probability model:

$$\begin{aligned} p_k =\frac{\exp \left( {a_1^{(k)} x_1 + \cdots +a_n^{(k)} x_n } \right) }{1+\exp \left( {a_1^{(2)} x_1 + \cdots +a_n^{(2)} x_n } \right) + \cdots +\exp \left( {a_1^{(m)} x_1 + \cdots +a_n^{(m)} x_n } \right) }, \end{aligned}$$
(1)

where \(x_{j}\) are predictor variables (\(j=1, 2, {\ldots }, n\)), and \(a_j^{(k)}\) are parameters of the probabilities of each k-th choice among all m of them (\(m=3\) in the trinomial outcome). For the sake of identification, the first share parameters are taken as reference, so put to zero.

For the binomial dependent variable the common tool for modeling is logistic regression

$$\begin{aligned} p=\frac{\exp \left( {a_1 x_1 + \cdots +a_n x_n } \right) }{1+\exp \left( {a_1 x_1 + \cdots +a_n x_n } \right) }. \end{aligned}$$
(2)

After finding parameters, the choice probabilities are estimated by (2) as the continuous values in the 0–1 interval. In discrete choice modeling (DCM), particularly with MaxDiff data, the binary outcome of the Best item versus non-best items (1 vs. 0) can be modeled by the logistic regression (2). Choice of the Worst item in MaxDiff can be considered in a similar DCM model (2). For a simultaneous estimation by all Best and Worst choices in one combined dataset the following property is applied: if to change signs of all predictor variables then probability estimated by the logit model (2) equals \(1\hbox {-}p\) which corresponds to absence of a binary event. Indeed, consider the transformation of sign change:

$$\begin{aligned} \frac{\exp \left( {-(a_1 x_1 + \cdots +a_n x_n )} \right) }{1+\exp \left( {-(a_1 x_1 + \cdots +a_n x_n )} \right) }=1-\frac{\exp \left( {a_1 x_1 + \cdots +a_n x_n } \right) }{1+\exp \left( {a_1 x_1 + \cdots +a_n x_n } \right) }=1-p.\nonumber \\ \end{aligned}$$
(3)

The design matrix for modeling of the Worst item can contain all predictors with opposite sign, and the binary output is defined as the Worst versus non-worst items (also 1 vs. 0).

In practical MaxDiff modeling, two DCM design matrices with the rows defined by Best-Neutral segment and by Worst-Neutral segment (with opposite signs of predictors) are combined into one total matrix of choices [17].

This approach can be applied to data with trinomial outcome of positive-neutral-negative segments considering the positive segment as the Best one, and negative as Worst one. In the combined matrix, the positive segment has the original values of predictors and the value 1 in the binary outcome, the negative segment has the opposite sign of the original values of predictors and the value 1 in the binary outcome, and the neutral segment has doubled rows of original neutral segment and the same with opposite signs of predictors and 0 value of the binary outcome. For a more explicit presentation, let us express the positive-neutral segment and the negative-neutral segment of data in matrix form:

$$\begin{aligned} \left( {\begin{array}{c} 1 \\ 0 \\ \end{array}} \right) \sim \left( {\begin{array}{c} X_{positive} \\ X_{neutral} \\ \end{array}} \right) , \qquad \left( {\begin{array}{c} -1 \\ 0 \\ \end{array}} \right) \sim \left( {\begin{array}{c} X_{negative} \\ X_{neutral} \\ \end{array}} \right) . \end{aligned}$$
(4)

Then the total stacked segments for the binary logit model can be written as follows:

$$\begin{aligned} \left( {\begin{array}{c} 1 \\ 0 \\ 1 \\ 0 \\ \end{array}} \right) \sim \left( {\begin{array}{c} X_{positive} \\ X_{neutral} \\ -X_{negative} \\ -X_{neutral} \\ \end{array}} \right) . \end{aligned}$$
(5)

This stacked data is modeled in one logistic regression (2) which defines choice of positive or negative outcomes versus non-chosen cases. With design (5), the positive and negative values of the binary predictors push the outcomes with the values 1 to the sides of maximum and minimum probability (2), respectively, while zero values yield the medium part of the logit curve.

Thus, a trinomial outcome can be considered via the binary logistic model, in one probability scale. Another important feature of modeling in one logit regression consists in possibility to obtain the analytical closed-form solution as described in [8, 13].

3 Numerical Example

Consider data from a real marketing research project with nineteen hundred observations, two dozen predictors, and trinomial outcome dependent variable of 320 negative, 926 neutral, and 681 positive cases. We can try to apply ordinary least squares, also to use three binary logit models of each level versus the others, with additional normalizing the predicted values for each response to one. The third technique corresponds to direct modeling of the MNL regression (1), and the fourth approach is the suggested trinomial modeling via the binary logit. By the scale of probability predicted in logistic regression for all three levels we can classify observations using Receiver Operating Characteristic, or ROC curve. An example of it is shown in Fig. 1. With sliding thresholds we identify the cut-off for neutral segment, and the other two segments, respectively.

Fig. 1
figure 1

ROC for Negative class: Sensitivity versus Specificity, or true positive rate (TPR) versus false positive rate (FPR)

Table 1 presents the results of described four approaches in cross-tables of the observed and predicted three classes. The bottom row in Table 1 shows the hit rate, or proportion of the total correct prediction on diagonal of cross-tables to the total number of observations. We see that the hit rate is very similar by all models, with a slightly better value for MNL which uses twice more parameters for data fitting. Interesting to note that trinomial binary model yields the best prediction of the negative values for the smallest total count in the data.

Table 1 Prediction of trinomial outcome by several models

4 Summary

The work describes a convenient approach to modeling trinomial ordinal categorical outcome via one binary logit regression. Theoretical features and practical application of this model are discussed, including a possibility of constructing logit regression in analytical closed-form solution. The considered approach suggests a new technique which is useful for problems requiring presentation of the trinomial events in one probability scale and helps in practical data fit and analysis.