PLoS ONE
Home Using survival prediction techniques to learn consumer-specific reservation price distributions
Using survival prediction techniques to learn consumer-specific reservation price distributions
Using survival prediction techniques to learn consumer-specific reservation price distributions

Competing Interests: The authors have declared that no competing interests exist.

Article Type: Research Article Article History
Abstract

A consumer’s “reservation price” (RP) is the highest price that s/he is willing to pay for one unit of a specified product or service. It is an essential concept in many applications, including personalized pricing, auction and negotiation. While consumers will not volunteer their RPs, we may be able to predict these values, based on each consumer’s specific information, using a model learned from earlier consumer transactions. Here, we view each such (non)transaction as a censored observation, which motivates us to use techniques from survival analysis/prediction, to produce models that can generate a consumer-specific RP distribution, based on features of each new consumer. To validate this framework of RP, we run experiments on realistic data, with four survival prediction methods. These models performed very well (under three different criteria) on the task of estimating consumer-specific RP distributions, which shows that our RP framework can be effective.

Jin,Haider,Greiner,Wei,Häubl,and Hutson: Using survival prediction techniques to learn consumer-specific reservation price distributions

1 Introduction

1.1 Motivation

Reservation price (RP) is the highest price a consumer is willing to pay for one unit of a certain product or service [1], which is an important and widely used concept in both the economics and marketing literature. It is critical for designing various pricing strategies, such as personalized pricing [2, 3], one-to-one promotion [4], and optimal pricing [5]. Many other fields, including auction [6, 7], ad exchange [8], negotiation, and the design and pricing of bundles [9, 10], also heavily rely on accurate estimations of consumers’ RPs.

For example, suppose that we are interested in setting the price of a certain product ω to achieve maximum profit when selling it to a certain population. If we know the reservation price ri of each subject i, then we can easily compute the overall purchasing probability function PPF(v) over price v for this specific population as

where n is the total number of consumers and I{ϕ} = 1 if the proposition ϕ is true, and 0 otherwise; see Fig 1.

The overall purchasing probability function–(left) for 4 individuals; (right) over a population.
Fig 1

The overall purchasing probability function–(left) for 4 individuals; (right) over a population.

If we knew PPF(v), we could achieve the maximum expected profit by setting the price of ω to be

where c is the production cost of ω. (Throughout, we use the term “production cost” to mean the incremental cost of producing a single unit).

Moreover, if we are allowed to sell ω at different prices to different subjects–i.e., first degree price discrimination [11]–then the seller’s best strategy for maximum profit is to sell ω to the subjects at their individual reservation prices–i.e., using Fig 1, sell to Person 1 at $13.20, sell to Person 2 at $15.60, etc. (here, we assume the production cost is under $10).

In the scenario of e-commerce, which has enjoyed booming development recently, the online retailers also have great interest in designing pricing strategies, understanding consumers’ purchasing decisions, doing one-to-one promotion and so on, which rely on accurate estimation/elicitation of consumers’ RPs. Additionally, online retailers usually have more information available than the traditional offline ones about their consumers, such as consumer-specific information (demographics, this consumer’s historical transactions and so on) and historical transactions of other consumers, which may be related to consumers’ RPs. This motivates us to find ways to better estimate consumers’ RPs with this available information.

1.2 Contributions

In this paper, we propose a novel framework of formulating the RP estimation problem, which involves explicitly defining a consumer’s RP as a random variable conditional on the consumer’s features [12]. This probabilistic framework not only captures the inherent uncertainty of RP, but also allows us to use stochastic models to express the relationship between RP and consumer-specific features.

We take a machine learning approach to this “RP estimation” task: inferring a probabilistic RP model from information about previous consumers—notably their transaction data (the observation that a consumer decided to “buy” a product at some specified price) and non-transaction data (the observation that a consumer intentionally decided to “not buy” a product at some specified price). Here, each non-transaction instance refers to an explicit decision to not purchase a product at the offered price–e.g., when a consumer puts an item in the shopping cart or the wish list, but does not buy it.

We can then apply this learned model to a specific consumer, to produce that consumer’s specific posterior distribution over RP, and use it to predict whether a consumer is likely to purchase the product. Note this approach does not require individual consumers to directly report their RPs, and so avoids some of the problems associated with that alternative approach; these relate to the first two drawbacks described in Subsection 3.1.

Beyond these foundations, implementing this required three novel contributions: First, we note that the purchasing (resp., non-purchasing) observations correspond to right censored (resp., left censored) observations in the survival analysis setting, which motivates us to utilize various survival techniques to learn a model that maps the features of a consumer to his/her RP, from historical (non-)transaction data—note the term “(non-)transaction data” refers to both transaction data and non-transaction data. (We discuss below how this relates to other marketing/auction results that view earlier data as being censored). Second, we provide empirical evidence (using several appropriate datasets) that this framework is effective in producing consumer-specific pricing, which can lead to greater profit than fixed pricing. Third, we introduce a fairly new learning tool for survival analysis, multi-task logistic regression (MTLR), to the marketing community, and demonstrate that this MTLR system is competitive–either outperforming or matching many standard tools, across three different measures.

This work is relevant to the marketing community as it means that a seller can first learn an RP-model for a specific product, based only on data that is often readily available–the (non-)transaction logs, along with some consumer description data–then apply the resulting model to accurately estimate the individual RP distribution of that product for a novel consumer, even if that consumer has not bought the product of interest before, or even is completely new.

Note this paper focuses on the task of estimating a consumer’s reservation price, but not about how a seller would use that information. We briefly touch on this topic in Section 6.4.

1.3 Other marketing applications of survival analysis ideas

As a final preliminary comment, note that prior work has explored ways to use ideas from survival analysis to tackle a variety of other marketing phenomena. For background, Hosmer et al. [13] provides a nice general introduction to survival analysis in general, and Wang et al. [14] summarizes many machine learning techniques and evaluation metrics for survival analysis. One obvious example is predicting when a customer will cease his/her relationship with a company–this is called “customer attrition” or “customer churn.” This corresponds exactly to survival analysis, as it is predicting the time to an event, where (right) censoring means a consumer is still with the company [12, 15]. The present work relates in terms of censoring of items, but differs as it deals with price, rather than time; it also considers both left- and right-censoring, etc.; Table 2 shows the connections. We note that some project, including [16, 17], connect this time to cost: Given that the treatment cost for a patient accumulates over time, if the study ends before the treatment is complete for a patient (or when that patient is lost to follow-up), we will not know his/her total cost–i.e., that person’s lifetime-medical-cost is left-censored. The present work differs as (1) our reasons for under-bounding the cost is not due to temporal truncation, but rather a model learned from earlier observations about (non)transactions of other consumers, being applied to a current consumer, which is (2) both left- and right-censored, and is (3) personalized, based on consumer features.

Ganchev et al. [18] also uses survival analysis techniques (here the Kaplan-Meier estimator) as a way to deal with the problem of order dispersion in “dark pools,” a relatively new kind of equities exchange in which traders seek to “invisibly” trade large volumes at market prices. Note that these applications are different from our goal, of predicting a consumer’s individual reservation price.

The present research is also similar to prior work on auctions: Blum et al. [19] observed that a bidder, at a sealed-bid auction, can use the auction outcomes to provide censored information about the other participants, which can be used to approximate their underlying bid distribution. Cesa-Bianchi et al. [20] and Amin et al. [21] use a similar observation in their analyses of second-price auctions–here, “only if we win the click do we observe the actual competing price; otherwise, we only know our bid was too low.” Our results differ by (1) considering individual consumer purchases, rather than winning multi-consumer auctions (and hence reservation price, rather than reservation bid), and (2) producing a model that involves learned non-linear combinations of consumer features (and so can estimate RPs of completely new consumers).

1.4 Outline

Section 3 describes our framework of RP estimation. Section 3.1 first summarizes the related literature, to place our work. Subsection 3.2 then introduces the formal definition of RP, and Subsection 3.3, the decision model that formulates the relationship between consumers’ purchasing decisions and their RPs. Subsection 3.4 illustrates our way to collect (non-)transaction data, which can be used to learn the RP distributions.

Section 4 first describes the relationship between the RP estimation problem and survival analysis problem. Then Subsection 4.2 introduces four survival models that can be used to estimate RPs: viz., Kaplan-Meier Estimator, Cox proportional hazard model, accelerated failure time model, and the MTLR model.

Section 5 describes how we collected the needed data and some basic information about the four datasets. We also discuss several potential problems of data quality and ways to address them. (We used a survey to collect the relevant information; its questions appear in Appendix in S1 File).

Section 6 presents empirical results of using various survival models to estimate a consumer’s RP, under three different evaluation criteria: the mean absolute error of the RP predictions, the classification accuracy of predicting specific purchases, and estimating the profit obtained with a simple pricing strategy. All results are based on ten-times repeated ten-fold cross validation. The strong performance of these models in estimating consumer-specific RPs supports the effectiveness of our novel framework. This section also provides the features found to be most relevant to the prediction. Finally, Section 7 discusses three potential directions for future work, and Section 8 summarizes our contributions.

2 Methods

This study used the data about reservation price that we collected using a survey on Amazon Mechanical Turk. This study received written approval from the Research Ethics Office at the University of Alberta (Number: Pro00048923_REN1).

3 Framework of reservation price estimation

The common understanding of RP–i.e., the highest price a consumer is willing to pay for a certain unit of product or service–indicates that consumers’ purchasing decisions on a certain product are closely related to their RPs of the product. Some methods (e.g., BDM; see below) require actual purchasing to obtain accurate estimates of RPs. However, we may be able to avoid directly asking consumers to report their RPs, by instead inferring their RPs from their purchasing decisions (and the decisions of other consumers); this information is much easier to collect in practice. Therefore, in this section, we propose a consumer decision model that formulates the way consumers reach purchasing decisions and how it is related to their RPs. This decision model and a corresponding way of collecting data make up our framework of RP estimation. Within this framework, we can design new methods or utilize existing methods to learn the RP distributions from the observations of consumers’ purchasing decisions, i.e., (non-)transaction data.

3.1 Previous analyses of reservation prices

As revealing the true RPs will put consumers at a disadvantage in making deals with sellers, they would not voluntarily reveal this information. This had led to a huge amount of research efforts in designing incentive compatible methods for eliciting fixed-point RPs [6, 2225]. In general, methods like the Becker-DeGroot-Marschak (BDM) method strive to make the consumers realize that revealing the true RP is their best strategy (see Fig 2), which is the key to accurate elicitation of consumers’ RPs.

The Becker-DeGroot-Marschak method.
Fig 2

The Becker-DeGroot-Marschak method.

However, it would be unrealistic to assume that a consumer’s RP for a product always stays the same. Wang et al. [25] also note that there is even uncertainty within an individual’s RP, due to the consumer’s uncertainty about his/her own preference [26] and the product performance [27].

Therefore, several different interpretations of the RPs have been proposed [2830], which are associated with different probabilities of purchasing (see Fig 3):

    Floor RP: the maximum price at or below which the consumer will buy with 100% probability [28].

    Indifferent RP: the price at which a consumer is indifferent between the money and the product—i.e., s/he will buy it with 50% probability [29].

    Ceiling RP: the minimum price at or above which the consumer will never buy it—i.e., s/he has 0% probability of buying [30].

Three interpretations of reservation price: Floor RP, indifferent RP, and ceiling RP.
Fig 3

Three interpretations of reservation price: Floor RP, indifferent RP, and ceiling RP.

Furthermore, ICERANGE [25] embraced the inherent uncertainty of RP, by viewing a consumer’s RP as a price range instead of a single price point, which here means simultaneously eliciting several price points associated with different purchasing probabilities.

However, none of these methods deal with the challenge raised in the e-commerce scenario, i.e., how to utilize the extra information in e-commerce to help the task of inferring consumers’ RPs. What is worse, they also suffer from several drawbacks, which make them ineffective in the e-commerce scenario:

    Consumers have little patience and no motivation to participate in the elicitation activity.

    It is hard to validate if consumers realize that their best strategy is to tell their true RPs, which may lead to inaccurate elicitation of RPs.

    These methods have no capability of generalize beyond individual-level RP predictions–i.e., each new consumer must go through the whole elicitation procedure to estimate his/her RP for the product of interest.

Therefore, we need a model that can overcome these drawbacks and can effectively utilize the new information available in e-commerce setting to help the task of estimating consumers’ RPs.

3.2 Stochastic setting of reservation price

These arguments motivate us to use a probabilistic interpretation of RPs [12]. Here, we let X denote the random vector representing the features of consumers and x denote a certain vector of feature values, corresponding to a single consumer. We also formally define two crucial random variables:

Definition 1 (Consumer-specific reservation price) For a certain product ω, the consumer-specific RP Rω|x0 is a random variable conditioned on the features of the consumer X=x.

Definition 2 (Consumer-specific purchasing decision) If a product ω is offered at price v, the consumer-specific purchasing decision Aω,v|x{1,0} is a binary random variable, conditioned on the features of the consumer X=x. (Note that Aω,v1|x and Aω,v2|x are two different random variables, for v1v2. Also, by convention, we will identify the value 1 with “buy” and 0 with “not_buy”).

3.3 Consumer decision model

In this section, we propose a decision-making model that describes how the consumer’s purchasing decision Aω,v|x is related to the RP Rω|x. When a consumer with features X=x is faced with a given offer–i.e., a specific product ω is being offered at price v–s/he reaches her/his purchasing decision aAω,v|x in a two-step procedure

    Step 1. Draw an “instant RP”: an instant RP r is drawn from the distribution of Rω|x.

    Step 2. Make a decision:

    i.e.,

That is, we assume that after drawing an instant RP rRω|x, the customer’s decision is determined by the relationship between r and v (see Fig 4). Then it is explicit that the relationship between the purchasing decision random variable Aω,v|x and the reservation price random variable Rω|x is

How consumers reach a purchasing decision when offered product ω at price v.
Fig 4

How consumers reach a purchasing decision when offered product ω at price v.

We can also derive the purchasing probability function PPFω(x,v), i.e., the probability that consumer x will buy product ω at price v, to be

where FRω|x(·) is the cumulative distribution function (CDF) of consumer xs RP for product ω.

Note that this decision-making process nicely matches the common understanding of RP–i.e., the highest price a consumer is willing to pay for a unit of a certain product or service [12].

3.4 Data collection and format

As shown in the previous section, RPs and purchasing decisions are closely related to each other, which suggests that we can indirectly infer consumers’ RPs for a certain product from the purchasing decisions of these (and other) consumers. Therefore, for a certain product ω, instead of directly asking consumers for their instant RPs ri, we instead collect (non-)transaction data–i.e., the decisions of many (earlier) consumers ai,ω on whether they will purchase ω at various different prices v. Each observation in the dataset Dω of product ω is a vector in the format (xi,vi,ω,ai,ω). An example dataset appears in Table 1.

Table 1
An example dataset of some product ω.
features of consumers alternativesprice vi,ωdecision ai,ω
agegendermonthly income
18male$200$3.500
26female$3000$5.001
28female$2000$4.50

Traditional RP models require consumers to be highly involved in a sophisticated elicitation procedure in order to make them understand that telling the true RP is their optimal choice. However, our data collection process is significantly simpler and does not make assumptions about consumers’ understanding, as we do not ask consumers to report their RPs directly.

4 Reservation price prediction models

This section first summarizes the main ideas from survival analysis (including censoring), to show how we can use survival prediction techniques to learn the consumer-specific RP distribution from the (non-)transaction data. We then introduce three popular survival models and one recent effort from the machine learning community for predicting subject-specific survival distributions.

4.1 Relation to survival analysis

Typically survival analysis focuses on time-to-event data, where the variable of interest is the death/event time T. In general, survival models try to learn the survival function,

from event and censored (left, right, or interval) data. This task differs from ordinary regression as it must deal with censored observations at training time, which are incomplete observations of the event time T.

Fig 5 suggests a training sample, perhaps for a breast cancer study, where the variable of interest is the patient’s time of relapse. Here, we know the actual time when some patients relapse (P#1 and P#2). Some other patients may still be non-relapsed when the study ends (e.g., P#3), and others may drop out of the study (P#4, P#5); here, we will never know her actual time of relapse T. All we know is that her relapse time is after her censored time–call it cr(P#i)–which is only partial information about T. This is called right censoring, as the unknown event time T is on the “right” side of the right-censored time, cr(P#i)–i.e., T > cr(P#i). For yet other patients, the unknown true event time is before a certain time cl(P#j). For example, imagine finding that a patient gets her first examination six months after the study began, and is then diagnosed as having already experienced a relapse. In this case, all we know is that the relapse happened in the first six months. This is called left censoring; see P#6 and P#7 in Fig 5. Here, we only know that the subject had already relapsed at the measured time, but not when she had relapsed–i.e., T < cr(P#j). Finally, we may sometimes know that the unknown true event time is in a certain time range [t1,t2]–e.g., perhaps if patients take yearly examinations, then if a patient is diagnosed as having a relapse, then we only know that the relapse time is in the previous one year. See P#8.

Uncensored observations.
Fig 5

Uncensored observations.

(where the patient’s actual time of death is observed), as well as right, left and interval censored events, where the time of death can only be bounded (providing a lower bound for right censored, and upper bounds for left censored, and an interval for interval censored). The “walking” figure shows the patient is alive at that time (the blue circle means this is the last time the patient was known to be alive), the angel emerging means that patient died at that time, and the tombstone symbol means the patient is known to be dead at that time (although s/he might have died earlier).

Now observe that a consumer’s RP is similar to a patient’s survival time, in that the observations (purchasing transactions) provide only censored versions of the information needed. Consider a purchasing transaction–e.g., when MrA bought a pen for $2. Here, we only know that MrA’s true RP for this pen is greater than or equal to $2. Similarly, if MsB chooses not to buy this pen for $5, this “non–purchasing transaction” means that MsB’s true RP is less than the $5. That is, for rRω|x:

    Purchasing transaction: aω = 1 ⇔ rωv

    Non-purchasing transaction: aω = 0 ⇔ rω < v

If we take the RP R as the variable of interest, instead of time T, then the purchasing (resp., non-purchasing) observations in the RP setting are equivalent to right censored (resp., left censored) observations in the survival analysis setting, as each purchase means the price was a lower bound of the consumer’s true reservation price (just like the right-censored time is a lower bound of a patient’s time of death); and each non-purchase means the offered price is an upper bound of consumer’s true reservation price (just like the left-censored time exceeds the upper bound of a patient’s time of death). To be more clear, Table 2 shows the matching relationship between the terminologies in these two settings.

Table 2
Matching terminology and symbols.
Survival AnalysisReservation Price
Event time

T
Reservation price

R
Survival distribution

f(t)
RP distribution

f(v)
Survival function

ST(t)
Purchasing probability function

PPF(v)
Left censored observation

c = l
Non-purchasing transaction

a = 0
Right censored observation

c = r
Purchasing transaction

a = 1
Uncensored observation

c = u
Observation of true RP (not used in this model)

The final line illustrates one important difference between these models, in that most survival studies include complete observations about some subjects–i.e., we know when some patients actually died. In our RP estimation task, however, we have no complete observations of consumers’ RPs at all, as every instance is either left- or right-censored. (Section 1.3 summarized other works that made similar connections).

Nevertheless, this connection allows us to utilize survival models to learn the distribution of R using the (non-)purchasing transactions. Fig 6 illustrates the whole learning system and how it works on new consumers.

Illustration of the data-generation, learning process and how it works on new instances.
Fig 6

Illustration of the data-generation, learning process and how it works on new instances.

4.2 Survival models

This subsection introduces several (three classic, and one recent) survival models: Kaplan-Meier estimator [31], Cox proportional hazard model [32], accelerated failure time model [33], and the multi-task logistic regression (MTLR) model [34]. In the context of survival analysis, the Kaplan-Meier model estimates an entire population’s survival distribution whereas these other models estimate individual survival distributions (ISDs) [35], and hence we refer to these other models as ISD models.(Of course, these are only a few of the many different survival analysis models; we chose only these four models as they are standard, and/or representative of major classes we want to examine). We will test their performance in Section 6, to evaluate the effectiveness of our framework of RP.

4.2.1 Kaplan-Meier estimator

The Kaplan-Meier (KM) estimator [31] is an empirical non-parametric model that estimates the survival function S(t), Eq 5. This tool is widely used in clinical studies for comparing the survival curves of two subpopulations in order to identify the risk factors–i.e., the features important to survival.

For a dataset consisting of only uncensored and right censored data, the empirical estimate of S(t) is

where τ1,τ2,…,τK are the set of all K distinct death times in the dataset, dj is the number of deaths at time τj, and rj is the number of subjects at risk right before τj (i.e., the number of subjects who died or were censored at or after τj).

Since an RP dataset only consists of left- and right-censored data, but no event data, we have to resort to the Expectation-Maximization approach of Turnbull [36] to estimate the survival curve, where both left and right censored data are treated as interval censored data.

Notice that the KM estimator does not consider the features of the subjects, which means it predicts the same survival curve for all subjects and thus is not personalized. We still include it here for completeness, as it is one of the most widely accepted and used models in survival analysis. We implemented this using the ic_np function in the R package icenReg [37] for the Turnball estimator.

4.2.2 Cox proportional hazards model

The Cox proportional hazards (Cox) model is a semi-parametric model designed for comparing the survival time of two populations or to identify the risk factors critical to survival [32]. Unlike the KM model, the Cox model uses the subject’s features and works with the hazard function(The hazard function is also called the “failure rate” since it reflects the subject’s instantaneous rate of failure).

instead of the survival function, where f(t) = Pr(T = t) is the probability that the patient dies at time t.

The Cox model models the hazard function as

where λ0(t) is the baseline hazard function, and θ is learned from a data sample. Here, as the relative influence of each feature xi,k depends linearly (“proportionally”) on the corresponding coefficient θk (albeit in the exponent), this is called a proportional hazard model.

One of the advantages of this model is that we can estimate θ by maximum partial likelihood estimation [32], which requires no knowledge of the baseline hazard λ0(t). This simplifies the task of identifying the risk factors. However when it comes to the prediction task, the proportional hazard assumption restricts the shapes of predicted survival curves of all patients to be essentially the same, as shown in Fig 7(a). This means its predictions on subjects’ survival rates might not be calibrated [34].

Survival curves for several patients from North Alberta Cancer Dataset [34], generated by (a) Cox model; (b) AFT model; (c) MTLR.
Fig 7

Survival curves for several patients from North Alberta Cancer Dataset [34], generated by (a) Cox model; (b) AFT model; (c) MTLR.

Similar to our analysis of Kaplan-Meier model (Subsection 4.2.1), we again treat both left and right censored data as interval censored data and utilize the Cox model designed for interval censored data to estimate our model [38, 39]. Specifically, we use the icenReg [37] package again but use the ic_sp function.

4.2.3 Accelerated failure time model (Tobit model)

The accelerated failure time model (AFT) is a parametric model that directly models the distribution of T with some parametric distribution [33], as shown below:

where δ is the scale parameter and ϵ is the error term. Different distributions of ϵ yield different forms of the AFT model. The commonly used distributions for ϵ include the Weibull distribution, log-logistic distribution, and the log-normal distribution, and the normal distribution. With N(0,1) (Gaussian distribution with zero mean and unit variance), the AFT model is actually equivalent to the well-known Tobit model in the economics literature [40].

In the AFT model, the effect of covariates is to accelerate/decelerate the scale of life time, while in the Cox model, the effect of covariates is to multiply the hazard by a constant. Fig 7(b) shows an example of the predicted survival curves of four patients in a survival dataset, by the AFT model with ϵ following the log-normal distribution. In our experiments, we fit the AFT model with ϵ following the log-normal distribution using the function survreg in the R package survival [41].

4.2.4 Multi-task logistic regression

Multi-task logistic regression (MTLR) is a recent effort from the machine learning community to produce a patient-specific survival function, which works well, according to several criteria [34]. Unlike the earlier models, MTLR does not make any explicit assumptions nor restrictions about the hazard function nor the shape of survival curves, meaning the MTLR survival curves of different individuals can be very different and can intersect with each other. This offers greater prediction capacity and flexibility. Fig 7(c) shows the predicted survival functions from MTLR for four patients [34].

MTLR first discretizes the continuous time axis into K+1 time points {τ0,τ1,τ2,…,τK}, with τ0 = 0 and τK = ∞, and then transforms the survival function prediction task into a sequence of binary probabilistic classification tasks, by constructing initially a logistic regression model for each time point τj,j = 1,…,K − 1:

where θj and bj are the parameters associated with the jth time point and yj = I{T < τj} indicates if the subject x has incurred an event before τj. (We earlier explored various rules for setting K, and found that KN, where N = Ne+Nc is the total number of individuals in the study, works effectively).

Then if we (for now) treat the classifiers as independent, we have the probability mass function (PMF) of y as

However, as we must prevent the case that yj = 1 and yj+1 = 0 from holding (that is, after someone dies, that person cannot come back alive–the “No Zombie” rule), the normalization term is the summation of the unnormalized “probability” of these K legal ys, which are (1, 1,…,1, 1), (0, 1,…,1, 1),…, (0, 0,…,0, 1), and (0, 0,…,0, 0). The final form of the PMF (probability mass function) of T is

where Θ=(θ1,θ2,...,θK-1), B = (b1,b2bK−1) and
is the normalization term.

Then one can derive the log-likelihood function of a dataset D={[xi,ti]}, where the first Ne instances are uncensored and the remaining Nc are right and/or left censored:

where yk(ti) = I{ti < τk}, and also cj(ti) = I{ti < τj} for right censored observations and cj(ti) = I{tiτj−1} for left censored observations.

MTLR is accessible as a web-tool (http://pssp.srv.ualberta.ca/) but our experiments use the R implementation provided in the MTLR [42] package.

5 Dataset

5.1 Data collection

While there are many datasets of financial transactions, essentially all report only the actual purchases, but not the “non-purchases”—i.e., they do not report situations where a consumer has declined an offer. For our stochastic RP setting, we need a dataset that contains both purchases and non-purchases. (While the donation dataset used in KDD Cup 1998 [43] does provide “non-donate” transactions, these non-donations only happen when a donor’s “reservation donation” is zero, which means that this dataset provides no meaningful left-censored observations).(While relatively few datasets record such non-transactions, they are clearly available, as many websites routinely collect this information about the users for future marketing advertisements [44, 45]. We wonder if many existing datasets do not collect is just because none of the standard analyses have demonstrated a benefit to this information. Perhaps the results of this research will motivate future researchers to collect this important information).

We therefore designed and executed our own online survey on Qualtrics, asking subjects from Amazon Mechanical Turk to provide information about themselves, and about their interest in purchasing each of four different specific-types of chocolate bar.

Here, we acquire one datasetand use a one-hot encoded feature to identify the brand of chocolate. For each consumer, we collected 41 features, e.g., the consumer’s demographics information, and preference towards the chocolate brand and flavor, the time when s/he ate her/his last meal and so on. Note that the subjects did not purchase any product in the survey; they just provided information, for which they were paid. (Note (1) we obtained the appropriate ethics permission for this study with human participants; (2) the dataset is publicly available at https://github.com/haiderstats/reservation-prices; and (3) more details about the survey appear in Appendix A in S1 File, or visit https://qtrial2014.az1.qualtrics.com/SE/?SID=SV_0kycgJjTgOj5Z8p).

The purpose of this survey is to collect relevant information, to help us in evaluating our various models. Therefore, we directly ask each participant i to provide his/her instant RP ri,ω for the product ω; we then used this to determine their responses ai,ω to certain offers, i.e., (non-)transaction data, following the decision-making process proposed in Subsection 3.2. N.b., our learning algorithms do NOT use those ri,ω’s–instead, they just use the (non-)transaction data; see Fig 6. We only used the collected ri,ω values as a way to evaluate our learners.

To explore the utility of the features (the Survey Questions; see Appendix A in S1 File) for our RP estimation task, we ran a simple Cox “univariate feature selection” process on this data, to identify which individual features are “relevant”, at p < 0.05; see Table 3. Note however that none of the learning algorithms used that data.

Table 3
Features that Cox-feature-selection considered relevant, at p < 0.05.
IdQuestion
-Brand of Chocolate
A.3.9How likely will you recommend this chocolate to your friends?
A.3.10How tasty do you believe the chocolate is?
A.4.3.eWhat kind of features of chocolate do you like—fruit flavor?
A.4.4On average, how much do you pay on chocolate in each grocery shopping?
A.4.5For all the possible chocolates available (weight: 100g), what is the highest price you are willing to pay?
A.4.6.eWhen making chocolate purchase decisions, how important is the shape & looking of chocolate on affecting your decision?
A.5.4What is your employment status?

5.2 Data quality

To ensure that our data quality is good and the reported RPs are accurate, our online survey included five attention-check questions, one RP understanding question and a two-step RP elicitation procedure [46]. We eliminated any subject who failed any attention-check or RP understanding question or who showed any inconsistency in his/her answers about RP. We also eliminated blatantly ridiculous responses—e.g., a subject willing to pay $10000 for a 100g chocolate bar.

This left 722 responses (out of 1080 submissions) qualified for each of the four chocolates, leading to an overall dataset size of 722 × 4 = 2888 instances, with 41 features describing consumer preferences/demographics and a one-hot encoded feature identifying the chocolate brand. Table 4 reports the median, mean (and standard deviation) of the consumer’s reported RPs.

Table 4
Median, mean (+ std) RP for the chocolate in the four datasets, over the 722 consumers; also retail price for each.
LindtGodivaValrhonaHersheys
Median of RP (sω)4.503.992.991.25
Mean of RP3.884.842.941.48
Std of RP1.892.922.081.05
Retail Price6.0010.007.502.00

While we tried to produce a dataset with good quality, the hypothetical response bias [25] cannot be completely avoided, as there were no real purchases. Fortunately, as our goal is to evaluate the performance of survival models within our novel framework of RP, this systematic bias will not be a serious issue. When online retailers later collect (non-)transaction data in practice, the consumers will be making purchases, which will mitigate this hypothetical response bias.

5.3 Generation of (non-)transaction data

After acquiring the true RPs and features of the consumers, we simulated a (non-)transaction data collection session by first sampling one query price vi,ω for each consumer from a stretched Chi-Square distribution–i.e., for each dataset, we first we set the parameter k in χk to be the mean of the RPs, and then used a linear mapping to match the variance of the distribution χk2 with the variance of the RPs. We then determined the consumer’s response ai,ω following the decision-making process defined in Subsection 3.2–i.e., the consumer’s purchasing decision is simply

This led to a dataset where each row is described by 41 features, the brand indicator, and also an offer price, vi,ω and response bit, ai,ω, for each ith consumer for the ωth brand. That is, the format of this dataset strictly conforms with the example dataset shown in Table 1 (see also Fig 6).

Note in particular that the dataset does not include the consumer’s RP ri,ω nor did we use the true RP data in training nor in the hyper-parameter selection via cross validation. The true RP data is only available in the final testing phase for evaluating the RP prediction performance. The whole procedure is outlined in Fig 8.

Procedure of generating (non-)transaction data.
Fig 8

Procedure of generating (non-)transaction data.

6 Experimental results

While Concordance (aka C-index) is a fairly standard measure for evaluating survival models [13], this section presents three other measures that are more relevant for this marketing task–including expected profit. All results are based on ten-times repeated ten-fold cross-validation where, for each of our four survival models (KM, Cox, AFT, and MTLR), we train a model on 9/10 of the subjects; we then use that learned model to produce a “RP distribution” for each of the remaining 1/10 of the subjects—we then repeat this process ten times. With the exception of KM, these models train on one dataset containing all four types of chocolates. Since KM is a population-based model we build four separate KM models–one for each type of chocolate. For MTLR, within each fold we execute an internal three-fold cross validation to select the best hyper-parameter, e.g., regularization constant. All significance tests presented use the two-sided t-test statistic given by Bouckaert and Frank [47], which corrects for the high Type II error and low replicability of significance tests involving cross-validation. For transparency all experimental code and data are publicly available (https://github.com/haiderstats/reservation-prices).

6.1 Mean absolute error

Given the learned CDF F^Rω|xi(v) of consumer xi’s RP for product ω, we use the median RP as the prediction for consumer xi’s RP value:

(We use the median price point as the RP prediction, as it is more robust than mean). As we have collected the consumers’ true instant RP ri,ω, we can compute the mean absolute error (MAE) of our predicted RPs,

where N is the number of consumers. Note that we cannot use this criterion in internal cross-validation to select hyper-parameters, because the learners do not have access to the true RP.

Fig 9 (and Table 6 in Appendix B of S1 File) shows the ten-times repeated ten-fold cross validation MAE for KM and the 3 ISD models, as well as a “cheating” baseline model. This “cheating” baseline utilizes the consumers’ true RPs, which are not available to the learners, and computes the median value of RP in the training set to use as its prediction on the test set. We also include the “Base” model, which is the best “single price” (not personalized) model possible, given the reservation price information: Here, we first compute the true RPs of the consumers in the training set then use its median value as the offer price to each consumer in the test set. (This is oracle-based as the learners do not know the consumers’ true RPs).

Mean absolute error (MAE) over ten-times repeated ten-fold CV.
Fig 9

Mean absolute error (MAE) over ten-times repeated ten-fold CV.

Despite this, we still found that the ISD models (AFT, Cox, MTLR) achieve better (that is, lower) MAE than the cheating baseline and the KM model, across all brands of chocolate. Specifically, for the overall average error (far right subfigure of Fig 9), two-sided t-tests show the performance of all ISD models were significantly better than KM and the baseline, p < 0.001, but no ISD model was significantly better than another. (While the “cheating model” did know a lot about the current consumer, it did not use information about the other consumers; our results show that models learned from just the “legal” information about those other consumers, can do better that ones based on “illegal” information about the current consumer).

6.2 Binary classification accuracy

This evaluation criterion tests if the learned models can accurately predict the consumer’s response to our offer of ω at price v. This too is very important in real applications. Here, each predictor predicts the response using

We then compute the classification accuracy as

Fig 10 shows that the classification accuracies of all four survival models (even the non-personalized KM) are significantly better (p < 0.001) than the “random guess” baseline, i.e.,

Ten-times repeated ten-fold cross validation classification accuracy, for each chocolate individually, and overall.
Fig 10

Ten-times repeated ten-fold cross validation classification accuracy, for each chocolate individually, and overall.

Additionally, in overall accuracy (far right subfigure of Fig 10), all ISDs outperformed KM, all p < 0.001, but did not significantly outperform one another. Table 7 in Appendix B of S1 File, provides the detailed information.

6.3 Profit using a simple pricing strategy

This section explores whether survival models within our RP framework can lead to real profit in practice. We use a pricing strategy that is simple and intuitive, which aims to maximize the expected profit and relies on good estimates of the PPF.

As we have a predicted purchasing probability function PPFω^(xi,v)=1-F^Rω|xi(v) for each consumer xi, we know the predicted expected profit by offering ω at price v to xi would be (v-c)·PPF^ω(xi,v) where c is the seller’s cost to produce ω. Here, the seller should therefore offer the product ω to xi at the price v^i(c) with maximum expected profit:

The true mean profit PFTω(c) for product ω, with production cost c, is

Unfortunately, due the price variability among retailers, there is no single retail price of the chocolates; moreover, we realized that the consumers in our population probably had limited interaction with the products–perhaps only through this survey. This motivated us to set the retail price sω of each chocolate bar ω to be its median reservation price, over the consumers–see Table 4.

Additionally, we also do not know the production costs. We therefore considered a range of possible production costs, at different proportions of the retail price cω∈ {0.10sω,0.15sω,…,0.85sω,0.90sω}. Note that Gilbert [48] claims that the most likely cost, for chocolate, is ≈0.75sω, based on the retail costs and margin accounting for 28% of the retail cost of chocolate.

Below we consider 4 different ways to determine how much the seller should charge for this product, v^i(c); see Table 5. So far, we have considered survival-based methods, including some that are personalized (ISD models: AFT, Cox, MTLR), and one that is not (KM). We also consider using the retail price (median reservation price sω) as a fixed price used for all customers.

Table 5
Different approaches to set prices.
PersonalizedNot Personalized
Survival-basedAFT, Cox, MTLRKM
Not Survival-basedLDA, LR, NBRetail Price

For comparison, we also evaluate the performance of three typical machine learning (ML) based probabilistic classifiers–viz., naïve Bayes (NB), logistic regression (LR) and linear discriminant analysis (LDA)–on the profit criterion, as these models can also be used to estimate PPF, though in a different manner: Here, for each product ω, we define the purchasing decision variable as A ∈ {0,1} and use (X,V) as the input variables, where V is the product price variable. That is, for each ρ{NB,LR,LDA}, the learned PPFρ model corresponds to

(To simplify notation, we do not include the ω).

Note the probability that a consumer will purchase a product should be monotonic in the offer price–i.e., if there is a 10% chance that a consumer will purchase an item if it is offered at $1, it should not be 50% at $2. We intentionally did not consider any nonlinear models as we found that they did not always exhibit this required property.

For each type of chocolate, for each of these 3 + 1 + 3 + 1 = 8 methods (Table 5), we computed the (10-times 10-fold CV) average profit for each of the 17 proposed costs cω. Fig 11 summarizes these average profit across all brands of chocolate, both by survival models (on right) versus ML models (on left), where each plot also includes the retail price model (the profit associated with selling the product at the standard price–see Table 4–in pink) as a reference. For completeness, this profit, for each chocolate brand individually, is shown in Figs 12–15 in Appendix B of S1 File.

Overall average profit 10-times 10-fold cross validation results.
Fig 11

Overall average profit 10-times 10-fold cross validation results.

On left, the average profit generated by the ML models– i.e., LR, LDA, and NB. On right, the average profit for the survival models–i.e., KM, AFT, Cox, and MTLR. The retail price model (in pink) is included in both sections for reference. The vertical dashed line is placed at the most likely real proportion of cost (0.75). See details in Table 8 in S1 File.

For the ML models, LR is strictly better than both the LDA and NB models, but still worse than the retail price model for cω ∈ [0.10,0.65]sω. Similarly, both AFT and KM are worse than using the retail price for cω < 0.70sω. The Cox and MTLR models either match or are superior to the retail price model–Cox and MTLR make a significantly higher profit (p < 0.05) than the retail model at majority of costs–all except cω ∈ [0.30,0.55]sω. Cox and MTLR’s profits do not significantly differ from each other at any cost but are significantly higher than AFT and LR at all costs (p < 0.01). This result strongly supports our idea of estimating RP on the individual level and reinforces the approach of using survival models for such estimation.

6.4 Discussion

First, we saw that ISD models (AFT, Cox, MTLR), with no knowledge of the true RP, can beat the cheating baseline on the MAE evaluation criterion. This shows that even without direct measurement of consumers’ true RPs, but only censored observations (of other consumers), one can still produce pretty good estimates of a consumer’s RP. It suggests that our way of collecting data may work in practice for the RP estimation task.

Second, on all three evaluation criteria, the performance of the ISD models are generally much better than the non-personalized one, i.e., KM. This is strong empirical support for modeling consumer-specific RPs, rather than a uniform RP. Moreover, the fact that the personalized MTLR and Cox models obtained significantly more profit than the retail price at a large majority of cost points, further bolsters this personalized approach.

Third, the strong performance of the MTLR and Cox models (both much better than the baselines) with respect to the ACC and PFT evaluation criteria, suggests that our way of estimating RP can be helpful in the real-world challenge of predicting if consumers would accept an offer or not, which will lead to higher profits. This is extremely useful for online retailers who want to conduct private promotions or general first degree price discrimination [49].

Note that we were initially surprised that our results were so good–indeed, apparently better than companies, who certainly must have seriously explored this pricing issue. We attribute our success to two factors: First, we were able to offer different prices to different consumers, which is probably not an option for products like chocolate bars. Second, we also had access to a number of characteristics for each consumer, which certainly helped our predictor; this might not always be available. We view our analysis as a proof-of-concept, to further demonstrate that it is possible to improve profit by offering personalized prices, assuming certain consumer features are known.

We also found that MTLR and Cox did extremely well—better than the other survival models considered—on maximizing profits. This suggests that MTLR and Cox are likely good choices for predicting RP distribution predictions, in general. Besides right censored and event data, MTLR can also handle left censored and interval censored data without modifications, while most packages of KM, AFT, and Cox only deal with event and right censored data. For this same reason, many other recent survival models (such as Random Survival Forests [50] and many recent deep learning models [16, 5154]) cannot be applied without major modification.

The capability to more accurately infer how much specific consumers might be willing to pay for a particular product–i.e., their individual reservation prices—is undoubtedly beneficial to sellers. However, converting this informational advantage into economic value for a firm is not a straightforward endeavor. In many settings, (near) perfect price discrimination–where the seller attempts to offer the product to each consumer at his/her exact reservation price (or just below it), as long as it exceeds the incremental cost of production–is not feasible due to the firm’s limited ability to make an individual consumer a strictly “private” price offer and prevent consumers from sharing price information among each other. Moreover, consumers may have a choice among competing offers by multiple sellers, and thus concurrently available offers must be incorporated along with individuals’ reservation prices. In addition, a person’s reservation price for a specific product may vary over time (e.g., as a function of other purchase or consumption decisions), presenting yet another challenge for sellers who intend to use inferred consumer-specific reservation prices to guide the implementation of personalized pricing.

Finally, the prospect of using individual reservation prices—accurately inferred based on prior behavior—to generate personalized price offers has important ethical implications. Is it socially acceptable for a seller to charge different consumers a different price for the same product? On one hand, one could argue that such a practice might be unfair to consumers. On the other hand, though, the are many instances where some form of price discrimination is occurring in the marketplace (e.g., based on when, where, or what quantity consumers buy, their prior purchases/loyalty, or their demographic characteristics), evidently without significant consumer backlash. It is difficult to predict how marketplace norms, and in particular consumers’ views towards sellers engaging in massively personalized pricing, might evolve. As a final observation on this issue, note that personalized pricing might actually enhance social welfare by providing a mechanism for serving some (e.g., economically underprivileged) consumers who would not be able to afford a product if the seller (was required to) set a single, uniform price, by offering it to these individuals at prices that they can afford.

7 Future work

7.1 Transaction-specific RP estimation

We anticipate it would be straightforward to integrate the product features Y in the dataset, allowing us to estimate (consumer, product)-specific reservation prices–i.e., estimating R|x,y instead of R|x. As R|x,y1 may be related to R|x,y2, this might allow us to transfer the knowledge of RP between similar products, possibility allowing us to estimate a consumer’s RP for a new product, as long as we know the product features. However, as we currently only have have data about four similar products, it is not realistic for us to experiment on this task.

A more ambitious goal is to include other information, such as transaction time and transaction location. Our framework can easily model this case, as well.

7.2 Relevant behavioral indicators

We can consider including other relevant behavioral indicators in the model estimation. A consumer’s decision process involves spending time inspecting various offers provided by multiple sellers, whether s/he considers a particular option as a viable candidate (e.g., adding an option to the shopping cart or wish list), or whether s/he revisits certain preferred options. The combination of certain behavioral features can help predict reservation prices that are particularly profitable for a specific segment of consumers. Online retailers, such as Amazon and Alibaba, follow similar practices by using targeted ads towards consumers who have inspected similar items. Nevertheless, such pricing strategies require companies to have significant resources for exploiting such behavioral indicators and assumes consumers have limited information about competing products offered by other companies.

7.3 Unbalanced data

In our study, we set the retail price to be the median reservation price, implying that half of consumers would purchase the product at the retail price. In real settings, most consumers will turn down most products—that is, most consumers will not accept most offers from online retailers. This means that most datasets will be (seriously) unbalanced [55], where the degree of unbalance will depend on several factors, such as the promotion strategy, distribution of offer price and the product itself. We plan to further study this direction, to see if survival models can handle such very unbalanced datasets.

7.4 Online predictor

Suppose we have two consumers xi and xj where xi and xj are very similar, then we find that xi declines our offer for ω at vi = $5. Should we then offer ω to xj at a price higher than $5? Probably not, as xi and xj are similar. This example argues that we should generate the offers sequentially, utilizing the previous responses, as this may be better than generating the offers {vi} in a batch mode. This leads to many interesting contextual bandit issues, and associated analyses [56]. We plan to extend our system to this on-line context.

8 Contributions

Motivated by the new demands of e-commerce, we propose a novel framework of estimating consumer-specific reservation price, which consists of a consumer decision-making model, and a corresponding data collection method.

This framework has three major advantages over the traditional elicitation methods in the marketing literature, which help it meet the new demands of the e-commerce scenario:

    It captures the inherent uncertainty of reservation price, consistent with the discussion in Talluri and Van Ryzin [12], etc.

    It connects the RP estimation task to survival prediction, which allows us to use survival models (standard and novel) to perform an individual-level RP prediction based on consumer-specific information.

    It is much easier and more practical for on-line retailers to implement our framework than the traditional elicitation method, as our data collection method does not ask consumers to report their true RPs, but indirectly infer a consumer’s RP based on historical (non-)transaction data of other consumers.

The experimental results show that survival prediction models, especially Cox and MTLR, perform well on this task under three different criteria. This empirically suggests that our framework of learning an RP prediction model is meaningful and could be very useful in practice. Given this success based on a relatively small dataset, we anticipate that others may try this approach on larger datasets, with even greater success.

References

SKalish, PNelson. A comparison of ranking, rating and reservation price measurement in conjoint analysis. Marketing Letters. 1991;2(4):327335. 10.1007/BF00664219

VChoudhary, AGhose, TMukhopadhyay, URajan. Personalized pricing and quality differentiation. Management Science. 2005;51(7):11201130. 10.1287/mnsc.1050.0383

NArora, XDreze, AGhose, JDHess, RIyengar, BJing, et al. Putting one-to-one marketing to work: Personalization, customization, and choice. Marketing Letters. 2008;19(3-4):305321. 10.1007/s11002-008-9056-z

GShaffer, ZJZhang. Competitive one-to-one promotions. Management Science. 2002;48(9):11431160. 10.1287/mnsc.48.9.1143.172

RKohli, VMahajan. A reservation-price model for optimal pricing of multiattribute products in conjoint analysis. Journal of Marketing Research. 1991; p. 347354. 10.1177/002224379102800309

WVickrey. Counterspeculation, auctions, and competitive sealed tenders. The Journal of finance. 1961;16(1):837. 10.1111/j.1540-6261.1961.tb02789.x

Mohri M, Medina AM. Learning Theory and Algorithms for Revenue Optimization in Second–Price Auctions with Reserve. arXiv preprint arXiv:13105665. 2013;.

Balseiro S, Feldman J, Mirrokni V, Muthukrishnan S. Yield optimization of display advertising with ad exchange. In: Proceedings of the 12th ACM conference on Electronic commerce. ACM; 2011. p. 27–28.

RVenkatesh, VMahajan. 11 The design and pricing of bundles: a review of normative guidelines and practical approaches. Handbook of pricing research in marketing. 2009; p. 232. 10.4337/9781848447448.00020

10 

BLSimonin, JARuth. Bundling as a strategy for new product introduction: Effects on consumers’ reservation prices for the bundle, the new product, and its tie-in. Journal of Business Research. 1995;33(3):219230. 10.1016/0148-2963(94)00071-L

11 

DUlph, NVulkan. Electronic commerce and competitive first-degree price discrimination. University of Bristol, Department of Economics; 2000.

12 

KTTalluri, GJVan Ryzin. The theory and practice of revenue management. vol. 68. Springer Science & Business Media; 2006.

13 

DWHosmer, SLemeshow, SMay. Applied survival analysis. Wiley Blackwell; 2011.

14 

Wang P, Li Y, Reddy CK. Machine learning for survival analysis: A survey. arXiv preprint arXiv:170804649. 2017;.

15 

DVan den Poel, BLariviere. Customer attrition analysis for financial services using proportional hazard models. European journal of operational research. 2004;157(1):196217. 10.1016/S0377-2217(03)00069-9

16 

Ye X, Henk HJ. An Introduction to Recently Developed Methods for Analyzing Censored Cost Data. ISPOR Connections. 2017;.

17 

YHuang. Cost analysis with censored data. Medical care. 2009;47(7 Suppl 1):S115. 10.1097/MLR.0b013e31819bc08a

18 

KGanchev, YNevmyvaka, MKearns, JWVaughan. Censored exploration and the dark pool problem. Communications of the ACM. 2010;53(5):99107. 10.1145/1735223.1735247

19 

ABlum, YMansour, JMorgenstern. Learning Valuation Distributions from Partial Observation. In: AAAI; 2015. p. 798804.

20 

NCesa-Bianchi, CGentile, YMansour. Regret minimization for reserve prices in second-price auctions. IEEE Transactions on Information Theory. 2015;61(1):549564. 10.1109/TIT.2014.2365772

21 

KAmin, MKearns, PKey, ASchwaighofer. Budget optimization for sponsored search: Censored learning in MDPs. In: Uncertainty in Artificial Intelligence; 2012.

22 

GMBecker, MHDeGroot, JMarschak. Measuring utility by a single-response sequential method. Behavioral science. 1964;9(3):226232. 10.1002/bs.3830090304

23 

KMMiller, RHofstetter, HKrohmer, ZJZhang. How should consumers’ willingness to pay be measured? An empirical comparison of state-of-the-art approaches. Journal of Marketing Research. 2011;48(1):172184. 10.1509/jmkr.48.1.172

24 

KJedidi, ZJZhang. Augmenting conjoint analysis to estimate consumer reservation price. Management Science. 2002;48(10):13501368. 10.1287/mnsc.48.10.1350.272

25 

TWang, RVenkatesh, RChatterjee. Reservation price as a range: An incentive-compatible measurement approach. Journal of Marketing Research. 2007;44(2):200213. 10.1509/jmkr.44.2.200

26 

GWFischer, MFLuce, JJia. Attribute conflict and preference uncertainty: Effects on judgment time and error. Management Science. 2000;46(1):88103. 10.1287/mnsc.46.1.88.15131

27 

RTRust, JJInman, JJia, AZahorik. What you don’t know about customer-perceived quality: the role of customer expectation distributions. Marketing Science. 1999;18(1):7792. 10.1287/mksc.18.1.77

28 

JRHauser, GLUrban. The value priority hypotheses for consumer budget plans. Journal of Consumer Research. 1986;12(4):446. 10.1086/208529

29 

SMoorthy, BTRatchford, DTalukdar. Consumer information search revisited: Theory and empirical analysis. Journal of Consumer Research. 1997; p. 263277. 10.1086/209482

30 

Varian HR. Microeconomic Analysis. Norton International edition. W.W. Norton; 1992. Available from: http://books.google.ca/books?id=m20iQAAACAAJ.

31 

ELKaplan, PMeier. Nonparametric estimation from incomplete observations. Journal of the American statistical association. 1958;53(282):457481. 10.1080/01621459.1958.10501452

32 

DRCox, et al. Regression models and life tables. JR stat soc B. 1972;34(2):187220.

33 

LWei. The accelerated failure time model: a useful alternative to the Cox regression model in survival analysis. Statistics in medicine. 1992;11(14-15):18711879. 10.1002/sim.4780111409

34 

CNYu, RGreiner, HCLin, VBaracos. Learning Patient-Specific Cancer Survival Distributions as a Sequence of Dependent Regressors. In: NIPS; 2011. p. 18451853.

35 

Haider H, Hoehn B, Davis S, Greiner R. Effective Ways to Build and Evaluat Individual Survival Distributions. arXiv preprint arXiv:181111347. 2018;.

36 

BWTurnbull. Nonparametric estimation of a survivorship function with doubly censored data. Journal of the American Statistical Association. 1974;69(345):169173. 10.1080/01621459.1974.10480146

37 

CAnderson-Bergman. icenReg: Regression Models for Interval Censored Data in R. Journal of Statistical Software. 2017;81(12):123. 10.18637/jss.v081.i12

38 

WPan. Extending the iterative convex minorant algorithm to the Cox model for interval-censored data. Journal of Computational and Graphical Statistics. 1999;8(1):109120. 10.1080/10618600.1999.10474804

39 

Henschel V, Heifi C, Mansmann U. intcox: Compendium to apply the iterative convex minorant algorithm to interval censored event data; 2009.

40 

WHGreene. Econometric analysis. Pearson Education India; 2003.

41 

Therneau TM. A Package for Survival Analysis in S; 2014. Available from: http://CRAN.R-project.org/package=survival.

42 

Haider H. MTLR: Survival Prediction with Multi-Task Logistic Regression; 2018. Available from: https://CRAN.R-project.org/package=MTLR.

43 

SHettich, SDBay. The UCI KDD Archive [http://kdd.ics.uci.edu]. Irvine, CA: University of California, Department of Information and Computer Science; 1999. Available from: http://www.kdnuggets.com/meetings/kdd98/kdd-cup-98.html#data.

44 

CASummers, RWSmith, RWReczek. An audience of one: Behaviorally targeted ads as implied social labels. Journal of Consumer Research. 2016;43(1):156178. 10.1093/jcr/ucw012

45 

BJing. Behavior-based pricing, production efficiency, and quality differentiation. Management Science. 2017;63(7):23652376. 10.1287/mnsc.2016.2463

46 

Smith S. 4 Ways to Ensure Valid Responses for your Online Survey; 2013. http://www.qualtrics.com/blog/online-survey-valid-responses/.

47 

Bouckaert RR, Frank E. Evaluating the replicability of significance tests for comparing learning algorithms. In: Pacific-Asia Conference on Knowledge Discovery and Data Mining. Springer; 2004. p. 3-12.

48 

Gilbert119 CL. Value chain analysis and market power in commodity processing with application to the cocoa and coffee sectors. Governance, coordination and distribution along commodity value chains. 2007; p. 267.

49 

Shiller BR. First Degree Price Discrimination Using Big Data; 2013.

50 

HIshwaran, UBKogalur, EHBlackstone, MSLauer, et al. Random survival forests. The annals of applied statistics. 2008;2(3):841860. 10.1214/08-AOAS169

51 

JLKatzman, UShaham, ACloninger, JBates, TJiang, YKluger. DeepSurv: personalized treatment recommender system using a Cox proportional hazards deep neural network. BMC medical research methodology. 2018;18(1):24. 10.1186/s12874-018-0482-1

52 

Luck M, Sylvain T, Cardinal H, Lodi A, Bengio Y. Deep learning for patient-specific kidney graft survival analysis. arXiv preprint arXiv:170510245. 2017;.

53 

Lee C, Zame WR, Yoon J, van der Schaar M. Deephit: A deep learning approach to survival analysis with competing risks. In: Thirty-Second AAAI Conference on Artificial Intelligence; 2018.

54 

Giunchiglia E, Nemchenko A, van der Schaar M. RNN-SURV: A Deep Recurrent Model for Survival Analysis. In: International Conference on Artificial Neural Networks. Springer; 2018. p. 23–32.

55 

SKotsiantis, DKanellopoulos, PPintelas, et al. Handling imbalanced datasets: A review. GESTS International Transactions on Computer Science and Engineering. 2006;30(1):2536.

56 

GBartók, CSzepesvari. Partial monitoring with side information. In: Algorithmic Learning Theory. Springer; 2012. p. 305319.