Guest Speaker: Dr. Jinzhi Bu (Massachusetts Institute of Technology)
Time/Date: 9:30 am Wed. 13th, Jan. 2021 (Beijing Time)
Length: 1.5 hours
Meeting ID: 628 844 49676
Password: 451895
Click Link: https://zoom.com.cn/j/62884449676
ABSTRACT
Classical statistical learning distinguishes between offline learning and online learning. Motivated by the idea of bridging the gap between these two types of learning tasks, this work investigates the impact of pre-existing offline data on the online learning in the context of a dynamic pricing problem. We consider a seller offering a single product with an infinite amount of inventory over a selling horizon. The demand in each period is determined by the price of the product according to a linear demand model with unknown parameters. We assume that the seller has some offline data before the start of the selling horizon, and wants to utilize both the pre-existing offline data and the sequentially-revealed online data to minimize the regret of the online learning process. We characterize the joint effect of the size, location and dispersion of the offline data on the optimal regret of the online learning. Our results reveal surprising transformations of the optimal regret rate with respect to the size of the offline data, which we refer to as phase transitions. In addition, our results also demonstrate that the location and dispersion of the offline data have an intrinsic effect on the optimal regret, which is quantified via the inverse-square law.