I am a PhD Student at Institute of Science Tokyo (formerly Tokyo Tech). My research lies at the intersection of HPC and Machine Learning, specifically focusing on distributed training and low-precision training (FP8/NVFP4) for Large Language Models.
I am a core contributor to the Swallow Project, a Japanese LLM development initiative, where I maintain the pre-training library and lead large-scale training experiments.
- [Jan 2026] My paper "Rewriting Pre-Training Data Boosts LLM Performance in Math and Code" has been accepted to ICLR 2026! ๐
I am actively looking for Research Internship opportunities in the US If you are interested in my work on LLM systems and low-precision training, please reach out!
- ๐ Website: https://okoge-kaz.github.io/
- ๐ Google Scholar: Citations Profile
- ๐ผ LinkedIn: kazuki-fujii
- ๐ฆ X (Twitter): @kazukifujii




