About

“Experience feeds us, opportunity grows us, relationships inspire us, and the world shapes us.”

I am a research scientist in Artificial Intelligence, studying the representation dynamics of large Transformer models: how internal representations form, align, and can be mapped into other spaces. My work spans pre-training, tokenization, model architecture, and model compression, with a focus on the mechanisms that shape model behavior across scales.

Grounded in rigorous empirical methods, I believe industrial research matters most when it narrows the gap between what is theoretically possible and what is practically deployable, extracting maximum value from every FLOPS, byte, and human interaction. In a life lived only once, I am committed to building a career I can look back on without regret, defined by curiosity, integrity, and lasting contributions.


Work Experience

NAVER Cloud : Dec.2021 - Present / Foundation Research, Leader (Research Scientist)

  • Pathfindings of HyperCLOVA X modeling (Technical Report: 2024, 2025)
  • Omni-foundation models and multi-linguality of large-scale Transformers.
  • Pre-training efficiency & Transformers architecture.
  • Model compression & quantization.
  • Also serves as Mission Lead for Omni-Pathfinding and Technical Writing.
  • 2025: Appointed Team Lead, Foundation Research Team, Hyperscale AI Division
  • 2024: Promoted to a technical leadership role, Foundation Research Team
  • 2023: Promoted to a technical leadership role, Hyperscale AI Efficiency Team
  • Alongside my full-time role at NAVER, I am pursuing a Ph.D. in AI at Korea Advanced Institute of Science and Technology (KAIST) under Prof. Jinwoo Shin, supported by a NAVER Cloud grant.

LG Energy Solution : Nov.2020 - Nov.2021 / Business Strategy, Data Scientist

Nepes : Jan.2018 - Oct.2020 / Future Intelligence, Research Scientist

  • 2019: Promoted to an assistant manager.
  • Military service (South Korea).

Recent Publications

*equal contribution

Lead author

  1. Exploiting Vocabulary Frequency Imbalance in Language Model Pre-training.
    Woojin Chung*, Jeonghoon Kim *.
    NeurIPS 2025.
  2. Peri-LN: Revisiting Normalization Layer in the Transformer Architecture.
    Jeonghoon Kim, Byeongchan Lee, Cheonbok Park, Yeontaek Oh, Beomjun Kim, Taehwan Yoo, Seongjin Shin, Dongyoon Han, Jinwoo Shin, Kang Min Yoo.
    ICML 2025.
  3. LRQ: Optimizing Post-Training Quantization for Large Language Models by Learning Low-Rank Weight-Scaling Matrices.
    Jung Hyun Lee*, Jeonghoon Kim *, June Yong Yang, Se Jung Kwon, Eunho Yang, Dongsoo Lee.
    NAACL 2025.
  4. Cross-lingual Collapse: How Language-Centric Foundation Models Shape Reasoning in Large Language Models.
    Cheonbok Park*, Jeonghoon Kim *, Joosung Lee, Sanghwan Bae, Jaegul Choo, Kang Min Yoo.
    Arxiv (under review, 2025).
  5. Rethinking Channel Dimensions to Isolate Outliers for Low-bit Weight Quantization of Large Language Models.
    Jung Hwan Heo*, Jeonghoon Kim *, Beomseok Kwon, Byeongwook Kim, Se Jung Kwon, Dongsoo Lee.
    ICLR2024.
  6. Memory-Efficient Fine-Tuning of Compressed Large Language Models via sub-4-bit Integer Quantization.
    Jeonghoon Kim * , Jung Hyun Lee*, Sungdong Kim, Joonsuk Park, Kang Min Yoo, Se Jung Kwon, Dongsoo Lee.
    NeurIPS 2023.
  7. FlexRound: Learnable Rounding based on Element-wise Division for Post-Training Quantization.
    Jung Hyun Lee*, Jeonghoon Kim *, Se Jung Kwon, Dongsoo Lee.
    ICML 2023.

Co-author

  1. HyperCLOVA X THINK Technical Report
    HyperCLOVA X Team.
    Technical report (Arxiv, 2025).
  2. ReGUIDE: Data Efficient GUI Grounding via Spatial Reasoning and Search.
    Hyunseok Lee, Jeonghoon Kim, Beomjun Kim, Jihoon Tack, Chansong Jo, Jaehong Lee, Cheonbok Park, Sookyo In, Jinwoo Shin, Kang Min Yoo.
    Arxiv (under review, 2025).
  3. LUT-GEMM: Quantized Matrix Multiplication based on LUTs for Efficient Inference in Large-Scale Generative Language Models.
    Gunho Park, Baeseong Park, Minsub Kim, Sungjae Lee, Jeonghoon Kim, Beomseok Kwon, Se Jung Kwon, Byeongwook Kim, Youngjoo Lee, Dongsoo Lee.
    ICLR2024.
  4. HyperCLOVA X Technical Report
    HyperCLOVA X Team.
    Technical report (Arxiv, 2024).
  5. Winning Both the Accuracy of Floating Point Activation and the Simplicity of Integer Arithmetic.
    Yulhwa Kim, Jaeyong Jang, Jehun Lee, Jihoon Park, Jeonghoon Kim, Byeongwook Kim, Baeseong park, Se Jung Kwon, Dongsoo Lee, Jae-joon Kim.
    ICLR 2023.
  6. AlphaTuning: Quantization-Aware Parameter-Efficient Adaptation of Large-Scale Pre-Trained Language Models.
    Se Jung Kwon, Jeonghoon Kim, Jeongin Bae, Kang Min Yoo, Jin-Hwa Kim, Baeseong Park, Byeongwook Kim, Jung-Woo Ha, Nako Sung, Dongsoo Lee.
    Findings of EMNLP 2022.

Reviewing

  • International Conference on Machine Learning
  • Neural Information Processing Systems
  • International Conference on Learning Representations
  • Association for Computational Linguistics.