Hi, I'm Alex Havrilla, a PhD student at Georgia Tech studying generative modeling in both theory and practice. My research aims to leverage insights from both disciplines to advance each other, ultimately unifying them under a single conceptual framework. I have interned at FAIR, Microsoft, and am a cofounder of the open-source RLHF research group CarperAI.

I graduated from Carnegie Mellon University with a joint MS/BS in mathematics and an additional major in computer science.

Picture

Research

My research covers a broad range of topics intersecting with generative modeling including LLMs, Reinforcement Learning, Diffusion Models, and statistical/approxmation theory for generative models. Recently, I’ve been deeply engaged in exploring ways to enhance the reasoning capabilities of Language Models (LLMs) for knowledge discovery, utilizing techniques from reinforcement learning. On the theoretical front, I focus on approximation and statistical theories for generative models, with a strong emphasis on validating these theories through empirical observations.

Papers

A. Havrilla, S. C. Raparthy, C. Nalmpantis, J. Dwivedi-Yu, M. Zhuravinskyi, E. Hambro, R. Raileanu, GLoRe: When, Where, and How to Improve LLM Reasoning via Global and Local Refinements. To appear in ICML 2024. [pdf]

A. Havrilla, Y. Du, S. C. Raparthy, C. Nalmpantis, J. Dwivedi-Yu, M. Zhuravinskyi, E. Hambro, S. Sukhbaatar, R. Raileanu, Teaching Large Language Models to Reason with Reinforcement Learning. To appear on ArXiv. [pdf]

T. Sawada, D. Paleka, A. Havrilla, P. Vidas, A. Kranias, P. Tadepilli, A. Komatsuzaki, ARB: An Advanced Reasoning Benchmark for Large Language Modeling. To appear in Neurips 2023 MathAI workshop. [pdf]

H. Liu, A. Havrilla, R. Lai, W. Liao. Deep Nonparmaetric Estimation of Intrinsic Data Structures by Chart Autoencoders: Generalization Error and Robustness. To appear in Journal of Applied and Computation Harmonic Analysis. [pdf]

A. Havrilla, M. Zhuravynski, A. Tiwari, J. Tow, E. Kim, Q. Anthony, S. Biderman, L. Castricato. trlX: A Framework for Large Scale Reinforcement Learning from Human Feedback. To appear in EMNLP 2023. [pdf]

A. Havrilla, K. Rojas, W. Liao, M. Tao, Dual-FNO UNet: Scale-Robust Diffusion Model for Zero-Shot Super-Resolution Image Generation. To appear in Neurips 2023 workshop on diffusion models. [pdf]

A. Havrilla, M. Iyer, Training Large Language Models with Noisy Algorithmic Chain of Thought. To appear in ICML 2023 worksohp on Symbolic and Data driven methods for reasoning in NLP. [pdf]

S. Biderman, A. Tikiwari, L. Castricato, E. Hallahan, A. Havrilla, Q. Anthony, E. Raff, What Makes a Good Multimodal Embedding Space? Theoretical and Empirical Insights from Topology. To appear on ArXiv.

A. Havrilla, L. Castricato, M. Shabuland, I. Yang, S. Frazier, M. Riedl, Robust Preference Learning for Storytelling via Contrastive Reinforcement Learning. To appear on ArXiv. [pdf]

B. Dahal, A. Havrilla, M. Chen, T. Zhao, W. Liao, On Deep Generative models for Approximation and Estimation of Distributions on Manifolds. To appear in Neurips 2022. [pdf]

A. Havrilla, P. Nayar, T. Tkocz, Khinchin-type inequalities via Hadamard’s factorisation. To appear in International Mathematics Research Notices. [pdf]

A. Havrilla, T. Tkocz, Sharp Khinchin-type inequalities for symmetric discrete uniform random variables. To appear in Israel J. Math. [pdf]

My Thesis on Sharp Khintchine type Inequalities and some New Results [pdf]