Taehyeon Kim

 Prev: Google Research (NYC), Qualcomm AI, DynamoFL (YCW22)
 PhD Student @ KAIST AI [OSI Lab]
 Located in Seoul, South Korea
 PhD in Graduate School of AI (KAIST)
 MS in Data Science (KAIST)
 BS in Mathematical Science (KAIST)
[Minor-Intellectual Property]
I am a Ph.D. Student in Optimization & Statistical Inference (OSI) Lab @ KAIST, advised by Prof. Se-Young Yun. I worked as a PhD Intern @ Google ResearchQualcomm AI, and DynamoFL (YCW22 Startup). My expected graduation date is Feb. 2025.  Contact: potter32 [at], kimtaehyeon610 [at] (permanent)


My goal is to tackle trustworthy and real-world AI/ML challenges
LLM: Instruction Tuning, Instructive Decoding, Speculative Decoding, Parallel Decoding
CV/NLP: Knowledge Distillation & Learning with Noisy Labels
Data Heterogeneity: Federated Learning & AutoML & Semi-Supervised Learning
Efficiency: Efficient Deep Learning
Game-Changing Research

️ News

Jun. 2024
Jan. 2024
 1 Accepted @ ICML2024W: Blockwise Parallel Decoding (Speculative Decoding)
 1 Accepted @ ICLR2024 (Spotlight): Instruction Following on Large Language Model
Dec. 2023
 1 Accepted @ NeurIPS2023W: Instruction Tuning & Instruction Following
 1 Accepted @ NeurIPS2023: Semi-Supervised Federated Object Detection
 Attending NeurIPS 2023 @ NOLA, US  
 1 Accepted @ AAAI2024: Few-shot & Domain Generalization
Oct. 2023
 Working with Google Research NYC

Working at/with

2023.01 - 2023.05
Semi-Supervised Object Detection
Federated Learning
Eric Lin
2021.06 - 2021.12
Neural Architecture Search
Knowledge Distillation
Heesoo Myeong
2017.03 - 2018.02
Trajectory Prediction
Jaegil Lee

 Publications & Technical Reports

Conference and Journal

Preprints (Under Review)

 Exploring and Improving Blockwise Parallel Drafts
Taehyeon Kim, Ananda Theertha Suresh, Kishore Papineni, Michael Riley, Sanjiv Kumar, Adrian Benton
Marc Bartholet*, Taehyeon Kim*, Ami Beuret, Se-Young Yun, Joachim M. Buhmann
 FLR: Label-Mixture Regularization for Federated Learning with Noisy Labels
Taehyeon Kim, Donggyu Kim, Se-Young Yun
 Towards Fast Multilingual LLM Inference: Speculative Decoding and Specialized Drafters
Euiin Yi*, Taehyeon Kim*, Hongseok Jeung, Du-Seong Chang, Se-Young Yun
Namgyu Ho*, Sangmin Bae*, Taehyeon Kim, Hyunjik Jo, Yireun Kim, Tal Schuster, Adam Fisch, James Thorne, Se-Young Yun

 Leadership Awards  Activities

 Research Projects

 Invited Talks

 Services & Others