Abstract
This paper examines the role of a subdivided or aggregated online review system to help online takeaway vendors select the most appropriate information strategy. First, we develop two models to depict the interaction between online vendors’ information strategies and consumers’ responses. Second, we take the multidimensional product attributes with their corresponding weights into consideration and illustrate that the sensitivity to product misfits, instead of the relative importance of product attributes, dominates profit maximization. Third, we make a comparison to find the most appropriate scenario to adopt a full or partial information strategy. When a large number of consumers satisfy the delivery time performance, an aggregated review system will be a better choice. Otherwise, vendors are advised to host a subdivided review system. Finally, we universally identify a variance boundary in the rating-star review system, which not only prevents consumers from expressing their real feelings but also makes observing consumer feedback and strategic adjustments inconvenient for online vendors.
Similar content being viewed by others
References
Anderson, M., & Anderson, M. (2014). 88% of consumers trust online reviews as much as personal recommendations. Retrieved November 18, 2015.
Anderson, S. P., De Palma, A., & Thisse, J. F. (1992). Discrete choice theory of product differentiation. Cambridge: MIT Press.
Cao, Q., Duan, W., & Gan, Q. (2011). Exploring determinants of voting for the “helpfulness” of online user reviews: A text mining approach. Decision Support Systems, 90(2), 511–515.
Celik, L. (2014). Information unraveling revisited: Disclosure of horizontal attributes. The Journal of Industrial Economics, 62(1), 113–136.
Chen, P. Y., Hong, Y., & Liu, Y. (2017). The value of multidimensional rating systems: Evidence from a natural experiment and randomized experiments. Management Science. https://doi.org/10.1287/mnsc.2017.2852.
Chen, Y., & Xie, J. (2008). Online consumer review: Word-of-mouth as a new element of marketing communication mix. Management Science, 54(3), 477–491.
Chen, Y., Yang, S., & Wang, Z. (2016). Service cooperation and marketing strategies of infomediary and online retailer with eWOM effect. Information Technology and Management, 17(2), 109–118.
Dimoka, A., Hong, Y., & Pavlou, P. A. (2012). On product uncertainty in online markets: Theory and evidence. MIS Quarterly, 36(2), 395–426.
Dolan, R., Conduit, J., Fahy, J., & Goodman, S. (2016). Social media engagement behavior: A uses and gratifications perspective. Journal of Strategic Marketing, 24(3–4), 261–277.
Floh, A., Koller, M., & Zauner, A. (2013). Taking a deeper look at online reviews: The asymmetric effect of valence intensity on shopping behaviour. Journal of Marketing Management, 29(5–6), 646–670.
Floyd, K., Freling, R., Alhoqail, S., Cho, H. Y., & Freling, T. (2014). How online product reviews affect retail sales: A meta-analysis. Journal of Retailing, 90(2), 217–232.
Ge, Y., & Li, J. (2015). Measure and mitigate the dimensional bias in online reviews and ratings. In Proceedings of the 36th international conference on information systems, Fort Worth, TX.
Goldhaber, M. H. (1997). The attention economy and the net. First Monday, 2(4). http://firstmonday.org/article/view/519/440.
Gu, Z., & Xie, Y. (2013). Facilitating fit revelation in the competitive market. Management Science, 59(5), 1196–1212.
Hao, L., Li, X., Tan, Y., & Xu, J. (2011). The economic value of ratings in app market. Available at SSRN 1892584.
Herrmann, P., Kundisch, D., Zimmermann, S., & Nault, B. (2015). How do different sources of the variance of consumer ratings matter? Economics of information systems. Fort Worth, TX: ICIS 2015.
Ho, Y. C., Wu, J., & Tan, Y. (2017). Disconfirmation effect on online rating behavior: A structural model. Information Systems Research, 28(3), 626–642.
Hu, N., Pavlou, P. A., & Zhang, J. (2007). Why do online product reviews have a J-shaped distribution? Overcoming biases in online word-of-mouth communication. Communications of the ACM, 52(10), 144–147.
Hu, N., Pavlou, P. A., & Zhang, J. (2017). On self-selection biases in online product reviews. MIS Quarterly, 41(2), 449–471.
Hu, N., Zhang, J., & Pavlou, P. A. (2009). Overcoming the J-shaped distribution of product reviews. Communications of the ACM, 52(10), 144–147.
Jiang, Y., & Guo, H. (2015). Design of consumer review systems and product pricing. Information Systems Research, 26(4), 714–730.
Kahneman, D., & Tversky, A. (1984). Choices, values, and frames. American Psychologist, 39(4), 341.
Kuan, K. K., Hui, K. L., Prasarnphanich, P., & Lai, H. Y. (2015). What makes a review voted? An empirical investigation of review voting in online review systems. Journal of the Association for Information Systems, 16(1), 48.
Kuksov, D., & Xie, Y. (2010). Pricing, frills, and customer ratings. Marketing Science, 29(5), 925–943.
Kwark, Y., Chen, J., & Raghunathan, S. (2014). Online product reviews: Implications for retailers and competing manufacturers. Information systems research, 25(1), 93–110.
Kwon, B. C., Kim, S. H., Duket, T., Catalán, A., & Yi, J. S. (2015). Do people really experience information overload while reading online reviews? International Journal of Human-Computer Interaction, 31(12), 959–973.
Lee, J. (2013). What makes people read an online review? The relative effects of posting time and helpfulness on review readership. Cyberpsychology, Behavior, and Social Networking, 16(7), 529–535.
Lee, J., Park, D. H., & Han, I. (2008). The effect of negative online consumer reviews on product attitude: An information processing view. Electronic Commerce Research and Applications, 7(3), 341–352.
Lee, E. J., & Shin, S. Y. (2014). When do consumers buy online product reviews? Effects of review quality, product type, and reviewer’s photo. Computers in Human Behavior, 31(1), 356–366.
Lelis, S., & Howes, A. (2011). Informing decisions: How people use online rating information to make choices. In Proceedings of the SIGCHI conference on human factors in computing systems (pp. 2285–2294). ACM.
Li, X., & Hitt, L. M. (2008). Self-selection and information role of online product reviews. Information Systems Research, 19(4), 456–474.
Li, X., & Hitt, L. M. (2010). Price effects in online product reviews: An analytical model and empirical analysis. MIS Quarterly, 34(4), 809–831.
Liu, Q., Huang, S., & Zhang, L. (2016). The influence of information cascades on online purchase behaviors of search and experience products. Electronic Commerce Research, 16(4), 553–580.
Liu, Q. B., & Karahanna, E. (2017). The dark side of reviews: The swaying effects of online product reviews on attribute preference construction. MIS Quarterly, 41(2), 427–448.
Mudambi, S. M., & Schuff, D. (2010). What makes a helpful review? A study of customer reviews on Amazon.com. MIS Quarterly, 34(1), 185–200.
Park, D. H., Lee, J., & Han, I. (2006). Information overload and its consequences in the context of online consumer reviews. In PACIS 2006 proceedings (p. 28).
Park, D. H., & Lee, J. (2009). eWOM overload and its effect on consumer behavioral intention depending on consumer involvement. Electronic Commerce Research and Applications, 7(4), 386–398.
Punj, G. (2012). Consumer decision making on the web: A theoretical analysis and research guidelines. Psychology & Marketing, 29(10), 791–803.
Purnawirawan, N., Eisend, M., De Pelsmacker, P., & Dens, N. (2015). A meta-analytic investigation of the role of valence in online reviews. Journal of Interactive Marketing, 31, 17–27.
Research and Markets. (2016). Global delivery and takeaway food market 2016–2020 with Delivery Hero, Just Eat, Foodpanda, Takeaway & Grubhub Dominating. http://www.businesswire.com/news/home/20160921006145/en/Global-Delivery-Takeaway-Food-Market-2016-2020-Delivery. Accessed 21 Sept 2016.
Sicilia, M., & Ruiz, S. (2010). The effects of the amount of information on cognitive responses in online purchasing tasks. Electronic Commerce Research and Applications, 9(2), 183–191.
Sun, M. (2011). Disclosing multiple product attributes. Journal of Economics & Management Strategy, 20(1), 195–224.
Sun, M. (2012). How does the variance of product ratings matter? Management Science, 58(4), 696–707.
Sun, H., & Xu, L. (2016). Online reviews and collaborative service provision: A signal-jamming model. Production and Operations Management. https://doi.org/10.1111/poms.12592.
Sutton, J. (1986). Vertical product differentiation: Some basic themes. The American Economic Review, 76(2), 393–398.
Technavio Research. (2016). Top 5 vendors in the delivery and takeaway food market from 2016 to 2020. http://www.businesswire.com/news/home/20160905005097/enTop-5-Vendors-Delivery-Takeaway-Food-arket. Accessed 5 Sept 2016.
The New York Times. (2012). The best book reviews money can buy. http://www.nytimes.com/2012/08/26/business/book-reviewers-for-hire-meet-a-demand-for-online-raves.html. Accessed 25 Aug 2012.
Yin, D., Bond, S. D., & Zhang, H. (2014). Anxious or angry? Effects of discrete emotions on the perceived helpfulness of online reviews. MIS Quarterly, 38(2), 539–560.
Yin, D., Mitra, S., & Zhang, H. (2016). When do consumers value positive vs. negative reviews? An empirical investigation of confirmation bias in online word of mouth. Information Systems Research, 27(1), 131–144.
Yu, J. (2015) Delivery service assortment and product pricing in online retailing: The impact of pricing flexibility and customer rating (Doctoral dissertation, University of British Columbia).
Zhang, K. Z., Cheung, C. M., & Lee, M. K. (2014). Examining the moderating effect of inconsistent reviews and its gender differences on consumers’ online shopping decision. International Journal of Information Management, 34(2), 89–98.
Ziegele, M., & Weber, M. (2015). Example, please! Comparing the effects of single customer reviews and aggregate review scores on online shoppers’ product evaluations. Journal of Consumer Behaviour, 14(2), 103–114.
Zou, P., Yu, B., & Hao, Y. (2011). Does the valence of online consumer reviews matter for consumer decision making? The moderating role of consumer expertise. Journal of Computers, 6(3), 484–488.
Funding
Funding was provided by National Natural Science Foundation of China (Grant No. 71771184), Humanities and Social Science Talent Plan (Grant No. ER42015060002), Natural Science Basic Research Plan in Shaanxi Province of China (Grant No. 2017JQ7012), Chinese Fundamental Research Funds for the Central Universities (Grant No. JB180602) and Innovation Fund of Xidian University.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
On behalf of all authors, the corresponding author states that there is no conflict of interest.
Appendices
Appendix A
Summary of notations
Notation | Definition |
---|---|
\(v\) | Product valuation |
\(\hat{v}_{i}\) | Customers’ expected product valuation in period i before consumption |
αv | Product quality |
(1 − α)v | Product delivery time |
α | Relative importance of product attributes |
τ | Customers’ unit misfit cost, where 0 ≤ τ ≤ v |
\(\hat{\tau }\) | Customers’ expected unit misfit cost |
β | Sensitivity to mismatched attributes |
βτ | Quality misfit cost, where 0 ≤ βτ ≤ αv |
(1 − β)τ | Time misfit cost, where 0 ≤ (1 − β)τ ≤ (1 − α)v |
x t | Customers’ taste used to measure quality misfit, where 0 ≤ xt ≤ 1 |
x u | Customers’ time urgency used to measure time misfit, where 0 ≤ xu ≤ 1 |
f | The likelihood of customers dissatisfied with the delivery time |
(1 − f) | The probability of customers satisfied with the delivery time |
\(u_{i}^{subdivided}\) | Customers’ utility before consumption in period i for the subdivided review system |
\(u_{i}^{aggregated}\) | Customers’ utility before consumption in period i for the aggregated review system |
\(D_{i}^{subdivided}\) | Customers’ demand in period i for the subdivided review system |
\(D_{i}^{aggregated}\) | Customers’ demand in period i for the aggregated review system |
M q | Average rating of product quality in the subdivided review system |
M t | Average rating of delivery time in the subdivided review system |
M a | Average rating in the aggregated review system |
V q | Rating variance in product quality in the subdivided review system |
V t | Rating variance in delivery time in the subdivided review system |
V a | Rating variance in the aggregated review system |
V adt | Rating variance caused by accepted delivery time in the aggregated review system |
\(P_{i}^{subdivided}\) | Product price in period i for the subdivided review system |
\(P_{i}^{aggregated}\) | Product price in period i for the aggregated review system |
\(x_{t,i}^{u = 0}\) | Customers’ taste with zero time misfit |
\(x_{u,i}^{t = 0}\) | Customers’ time urgency with zero quality misfit |
\(\pi_{i}^{subdivided}\) | Online vendors’ profit in period i for the subdivided review system |
\(\pi_{i}^{aggregated}\) | Online vendors’ profit in period i for the aggregated review system |
Appendix B
2.1 B.1 Proof of Proposition 1
To check the effects of the average rating and variance on the second-period equilibrium outcomes in the subdivided review system, we first calculate the first-order conditions of price. From Eq. (9), we have the following:
It shows that the best second-period price increases with the average ratings and variances of both product attributes.
Second, we solve the first-order conditions of demand. As shown below,
we know that the second-period demand increases with the average ratings caused by the quality and delivery time, but it decreases with the variances of both product attributes.
The first-order conditions of profit prove that the second-period profit increases with the average ratings of the quality and delivery time. Only when \(\sqrt {3V_{q} } + \sqrt {3V_{t} }\) is larger than \(2M_{q} + 2M_{t}\), the profit increases with the variances of both product attributes. However, \(\sqrt {3V_{q} } + \sqrt {3V_{t} }\) cannot exceed \(2M_{q} + 2M_{t}\). Assuming that a is the highest rating score and b is the lowest rating score in any rating scale, as we know, when customers publish extreme ratings, it results in the largest variance. Thus, we have \(V_{q} = V_{t} = \left( {\frac{a - b}{2}} \right)^{2} = M_{q} = M_{t} \,\) and \(3\sqrt {V_{q} } < M_{q} ,{ 3}\sqrt {V_{t} } < M_{t}\). This means that a variance boundary exists in current rating stars review systems.
2.2 B.2 Proof of Proposition 2
To check the impacts of the average ratings and variances on the second-period equilibrium outcomes in the aggregated review system, we first calculate the first-order conditions of price. From Eq. (13), we have the following:
It describes that the best second-period price increases with the aggregated average rating and the variance caused by accepted delivery time and decreases with the aggregated rating variance otherwise.
Second, we solve the first-order conditions of demand. As seen below,
we know second-period demand increases with the aggregated average rating. However, it decreases with the aggregated rating variance and the variance caused by accepting delivery time.
The first-order conditions of profit illustrate that the second-period profit increases with the aggregated average rating and decreases with the aggregated rating variance. Only when the variance caused by the accepting delivery time is higher than a threshold, i.e., \(V_{acc} > 2\left( {V_{a} + M_{a}^{2} } \right),\) the profit increase. However, \(V_{acc}\) cannot exceed \(2\left( {V_{a} + M_{a}^{2} } \right),\) which is restricted by the variance boundary.
2.3 B.3 Proof of Proposition 3
Regardless of which types of online review systems online vendors apply, the true value of the product is fixed. From Eqs. (6) and (12), we obtain the relationship between rating scores and variances as \(\left( {M_{q} + M_{t} + \sqrt {3V_{q} } + \sqrt {3V_{t} } } \right)^{3} = \frac{{\left[ {V_{a} + M_{a}^{2} + \sqrt {2V_{adt} \left( {M_{a}^{2} + V_{a} } \right)} } \right]^{3} }}{{M_{a}^{3} }}.\) Given the rating distribution of the aggregated review system, the aggregated average rating Mq is \(\left( {1 - f} \right)\left( {v - \frac{{\beta x_{t} \tau + \left( {1 - \beta } \right)x_{u} \tau }}{3}} \right)\), and the aggregated rating variance Va is \(\left( {1 - f} \right)f\left( {v - \frac{{\beta x_{t} \tau + \left( {1 - \beta } \right)x_{u} \tau }}{3}} \right)^{2} .\) Based on the results of Mq and Va, \(\frac{f}{1 - f}\) is equal to \(\frac{{V_{a} }}{{V_{a} + M_{a}^{2} }}.\) Because \(\beta x_{t} \tau + \left( {1 - \beta } \right)x_{u} \tau = 2\left( {\sqrt {3V_{q} } + \sqrt {3V_{t} } } \right),\) the variance caused by accepting the delivery time can be rewritten as \(V_{acc} = \frac{{2M_{a}^{2} \left( {\sqrt {3V_{q} } + \sqrt {3V_{t} } } \right)^{2} }}{{9\left( {V_{a} + M_{a}^{2} } \right)}}.\) Thus, we are able to simplify Eq. (17) as follows,
It indicates that the profit difference depends on \(V_{a} \left( {M_{a}^{2} + V_{a} } \right)^{2} - M_{a}^{6}\), that is, \(f\left( {1 - f} \right)^{3} .\) When the likelihood of a dissatisfied delivery time f* > 0.317, the subdivided review system arouses much more profits than the aggregated review system does.
Let \(A = \left[ {V_{a} + M_{a}^{2} + \sqrt {2V_{acc} \left( {M_{a}^{2} + V_{a} } \right)} } \right]^{3} \left[ {V_{a} \left( {M_{a}^{2} + V_{a} } \right)^{2} - M_{a}^{6} } \right]\) and \(B = 54V_{a} \left( {M_{a}^{2} + V_{a} } \right)^{2} M_{a}^{3} \left( {\sqrt {3V_{q} } + \sqrt {3V_{t} } } \right)^{2}\), substituting these terms back to \(\Delta \pi_{2} = \pi_{2}^{subdivided*} - \pi_{2}^{aggregated*}\) yields \(\Delta \pi_{2} = \frac{{A\left[ {\beta x_{t} + \left( {1 - \beta } \right)x_{u} } \right]^{2} }}{{\beta \left( {1 - \beta } \right)B}}\). Thus, we derive \(\frac{{\partial \Delta \pi_{2} }}{\partial \beta } = \frac{{\beta x_{t} + \beta x_{u} - x_{u} }}{{\left[ {\beta \left( {1 - \beta } \right)B} \right]^{2} }}.\) Since \(\frac{{\partial^{2} \Delta \pi_{2} }}{{\partial \beta^{2} }} = \frac{{x_{u}^{2} x_{t} + x_{t}^{2} x_{u} }}{{\left[ {\beta \left( {1 - \beta } \right)B} \right]^{4} }} > 0\), we know that if customers’ sensitivity to mismatched attributes \(\beta^{*} = \frac{{x_{u} }}{{x_{t} + x_{u} }}\), online vendors have the minimum profit in the second period. Similarly, let \(C = 2\hat{v}^{3} \left[ {f - \left( {1 - f} \right)^{3} } \right]\) and \(D = 27\tau^{2} f\), substituting these terms back to \(\Delta \pi_{ 1} = \pi_{ 1}^{subdivided*} - \pi_{ 1}^{aggregated*}\) yields \(\Delta \pi_{ 1} = \frac{C}{{\beta \left( {1 - \beta } \right)D}}\). Thus, we derive \(\frac{{\partial \Delta \pi_{1} }}{\partial \beta } = \frac{{ - \left( {1 - 2\beta } \right)D}}{{\left[ {\beta \left( {1 - \beta } \right)D} \right]^{ 2} }}\). Since \(\frac{{\partial^{ 2} \Delta \pi_{1} }}{{\partial \beta^{ 2} }} = \frac{{ 2D^{2} \left[ {\beta \left( {1 - \beta } \right)} \right]^{ 3} + 4\left( {1 - 2\beta } \right)^{2} }}{{D\left[ {\beta \left( {1 - \beta } \right)} \right]^{ 3} }} > 0\), when \(\beta^{*} = 0.5\), first-period profit reaches the lowest point. In terms of these results, we can deduce that online vendors’ profits for both online review systems are directly related to customers’ sensitivity to mismatched attributes in any periods.
Rights and permissions
About this article
Cite this article
Wang, H., Du, R., Li, J. et al. Subdivided or aggregated online review systems: Which is better for online takeaway vendors?. Electron Commer Res 20, 915–944 (2020). https://doi.org/10.1007/s10660-018-9314-y
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10660-018-9314-y