Skip to content
View Hiroki11x's full-sized avatar
🦙
🦙

Organizations

@jphacks @rioyokotalab @crest-deep @TITAMAS @RotaPlusPlus @Agents-NY @ArtHackDay-Plus1 @MLHPC

Block or report Hiroki11x

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 100 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
Hiroki11x/README.md

About

Hi! I'm Hiroki Naganuma, a Ph.D. candidate in Computer Science at Université de Montréal and Mila - Quebec AI Institute, advised by Prof. Ioannis Mitliagkas. My research focuses on large-scale parallelization in machine learning and the analysis of training dynamics in deep neural networks. I recently completed student researcher role at Google DeepMind (Mountain View), where I worked under the guidance of Dr. George E. Dahl on learning rate scheduling. Previously, I completed a summer internship at Microsoft Research (Redmond), where I worked on efficient pretraining algorithms for large language models with Dr. Philipp Witte and Dr. Russell J. Hewett. In the summer of 2025, I will be a research intern at Meta - AI and Systems Co-Design (Menlo Park) (host: Dr. Hao-Jun Michael Shi and Dr. Parameswaran Raman).

I am a recipient of the Masason Foundation Fellowship. I earned my B.Sc. (2017) and M.Sc. (2019) from the Tokyo Institute of Technology as Valedictorian, where I had the privilege of working closely with Prof. Rio Yokota and an outstanding group of collaborators.

Collaboration & Mentorship: If you're interested in my work, feel free to reach out!

📄 CV (Last updated: April 2025)

Stats

My github stats Top Langs

Pinned Loading

  1. horovod/horovod horovod/horovod Public

    Distributed training framework for TensorFlow, Keras, PyTorch, and Apache MXNet.

    Python 14.5k 2.3k

  2. LossLandscapeGeometry LossLandscapeGeometry Public

    No Wrong Turns: The Simple Geometry Of Neural Networks Optimization Paths (ICML2024)

    Shell 8

  3. Optimizer_Comparison_OOD Optimizer_Comparison_OOD Public

    Empirical Study on Optimizer Selection for Out-of-Distribution Generalization (TMLR2023)

    Python 5

  4. ConjugateGradient_GAN ConjugateGradient_GAN Public

    Conjugate Gradient Method for Generative Adversarial Networks (AISTATS2023)

    Python 3 1

  5. MLHPC/wandb_tutorial MLHPC/wandb_tutorial Public

    38 1

  6. Timm_OOD_Calibration Timm_OOD_Calibration Public

    An Empirical Study of Pre-trained Model Selection for Out-of-Distribution Generalization and Calibration (TMLR2025)

    Python 2