Jared Fernandez Jared Fernandez
Energy Considerations of Large Language Model Inference and Efficiency Optimizations
Efficient Hardware Scaling and Diminishing Returns in Large-Scale Training of Language Models
Gradient Localization Improves Lifelong Pretraining of Language Models
The Framework Tax: Disparities Between Inference Efficiency in Research and Deployment
Model Successor Functions
Tools Fail: Detecting Silent Errors in Faulty Tools
Language Models Need Inductive Biases to Count Inductively
DiffusionPID: Interpreting Diffusion via Partial Information Decomposition
VISREAS: Complex Visual Reasoning with Unanswerable Questions
Skews in the Phenomenon Space Hinder Generalization in Text-to-Image Generation
Training Vision-Language Transformers from Captions
WebQA: Multihop and Multimodal QA
Rosa Vitiello Technically Rosanna
2025 | Sheryl Matthew | Undergrad | → PhD @ USC | ![]() |
2024 | Minyoung Hwang | Intern | → PhD @ MIT | ![]() |
2022 | Xiang Fan | Undergraduate | → PhD @ UWashington | ![]() |
2022 | Andy Liu | Intern | → PhD @ CMU | ![]() |
2022 | Deogratias Amani | Intern | → | ![]() |
2022 | Josh Zhanson | MLT | → Data Scientist @ MSFT | ![]() |
2022 | Xiaoyu (Erin) Zhang | 5th year MS | → PhD @ UIUC | ![]() |
2021 | Feiyu (Gavin) Zhu | Undergrad | → PhD @ CMU | ![]() |
2021 | Blessing Bassey | Intern | → | ![]() |
2021 | Melanie Sclar | Intern | → PhD @ UWashington | ![]() |