Research
Papers, preprints, and ongoing projects. The through-line is sparse recovery applied wherever it fits — optimizers, attention, KV cache, interpretability, privacy.
TMLR 2024 · arXiv 2402.19016
Federated learning via OMP under DP constraints. Near-SOTA at 8–10% of dense parameters with formal convergence guarantees.
arXiv 2212.13071
Optimal client sampling strategy for differentially private federated networks.
ICML Workshop on Theory of Differential Privacy 2022
ICML Workshop on AI for Science 2022 · IMPS 2022 (Oral)
ICML Workshop on Socially Responsible ML 2021
Halley Deg-5 T=3 optimizer outperforms GramMuon T=5 on WikiText-103 and enwiki8. Running experiments on H100 and RTX PRO 6000 Blackwell. Targets ICML-tier venue.
OMP and FoBa applied uniformly across FFN activations, attention mechanisms, and optimizer orthogonalization. FoBa-Gate as SwiGLU replacement (p=0.012, Cohen’s d=2.16). OMP-attention with ridge-regularized backward (3× better validation perplexity than softmax).
Statistical power analysis via metric learning.