PhD student at OSI LAB of KAIST AI working on large language models.

💌 [email protected] 🌐 itsnamgyu @ { Google Scholar Twitter LinkedIn GitHub }

I’m interested in novel approaches to improve the efficiency and helpfulness of LLMs.

Recently, I investigated the distillation of chain-of-thought (CoT) reasoning from 100B+ LLMs to small language models [C4], and applied this to hate speech detection [C5].

I am actively working on how to improve LLMs, focusing on these two questions: (1) how to handle the variance in computation/modeling capacity requirements of each token, and (2) how to incentivize the model to express/consider its own uncertainty.






Honors & Activities