Research
You can find my Google Scholar page here.
Research in Distribution Estimation
Distribution estimation concerns the problem of estimating an unknown probability distribution from a finite set of samples. The empirical estimator—equivalently, the maximum likelihood estimator (MLE)—is known to be asymptotically optimal. However, alternative procedures such as add-constant estimators, the modified Good–Turing estimator, and the profile maximum likelihood (PML) estimator can demonstrate improved performance under various choices of loss functions. These estimators are natural, in the sense that they assign equal probability to symbols that appear the same number of times in the observed sample.
In our work, we characterize the unavoidable estimation error incurred by the class of natural estimators. We further develop strategies that leverage some side information to beat the performance of natural estimators.
-
Estimating Error in Natural Distribution Estimation
H. Balasundaram, A. Thangaraj.
Annual Allerton Conference on Communication, Control, and Computing, 2025. Conference Paper. Slides. -
Distribution Estimation with Side Information
H. Balasundaram, A. Thangaraj.
Submitted to International Symposium on Information Theory, 2026. arXiv.
Research in Online Convex Optimization
The problem of Online Convex Optimization (OCO), where a policy selects an action before observing a convex cost function, is well understood, with a regret bound of O(√T). We study Constrained Online Convex Optimization (COCO), a natural extension of OCO in which, after each action, the policy observes not only a cost function but also a constraint function. Since it may be impossible to satisfy the constraint at every step online, the goal is to simultaneously minimize regret while controlling the cumulative constraint violation.
Prior work by Vaze and Sinha established that both regret and total constraint violation can both be bounded by O(√T), a result that was widely believed to be optimal. In this work, we refute that belief and show that, in two dimensions, it is possible to achieve:
- Regret: O(√T)
- Constraint violation: O(T^(1/3))
Our analysis is geometric in nature. At each step, we demonstrate that either the area or the perimeter of the constraint sets shown decreases. Since the initial region has finite area and perimeter, this process cannot continue indefinitely. This yields a simultaneous upper bound on both regret and constraint violation.
- Breaking the $O(\sqrt{T})$ Cumulative Constraint Violation Barrier while Achieving $O(\sqrt{T})$ Static Regret in Constrained Online Convex Optimization
H. Balasundaram, Karthick Krishna Mahendran, Rahul Vaze.
arXiv
Research in Information Theory
- Learning to Transmit Over Unknown Erasure Channels with Empirical Erasure Rate Feedback
H. Balasundaram, K. Jagannathan.
arXiv.
Research on Stable Matchings
-
Generalized Capacity Planning for the Hospital-Residents Problem
H. Balasundaram, G. Limaye, M. Nasre, and A. Raja.
Theoretical Computer Science, 2026. Journal Paper. -
Stability Notions for Hospital Residents with Sizes
H. Balasundaram, Krishnashree J. B., G. Limaye, M. Nasre.
Foundations of Software Technology and Theoretical Computer Science 2025. Conference Paper.