All Dates/Times are Australian Eastern Standard Time (AEST)

# Technical Program

## Paper Detail

 Paper ID D6-S1-T3.2 Paper Title A Lower Bound for Regret in Logistic Regression Authors Gil I. Shamir, Google, United States; Wojciech Szpankowski, Purdue, United States Session D6-S1-T3: Classification II Chaired Session: Monday, 19 July, 22:00 - 22:20 Engagement Session: Monday, 19 July, 22:20 - 22:40 Abstract We study logistic regression with binary features in which the number (or degree) of occurring features determines the label probability. This model fits one of social networks, where the number of friends determines the likelihood of outcomes instead of the identity of the friends, or more generally, a graph model, where the degree of a node can determine its behavior. It includes the case in which weights can be viewed as i.i.d. (e.g., in Bayesian modeling). For such a model, we introduce the maximal minimax regret that we analyze using a unique combination of analytic combinatorics and information theory. More importantly, the resulting regret is a general lower bound for the pointwise regret of a general logistic regression over all algorithms (learning distributions). We show that the introduced worst case (maximum over feature sequences) maximal minimax regret grows asymptotically as $(d/2) \log (T/d) +(d/2) \log(\pi/2) +O(d/\sqrt{T})$ for dimensionality $d=o(\sqrt{T})$, which is a lower bound for a regret of a general logistic regression. We extend our results to loss functions other than logistic loss and non-binary labels.