P-1 AI
Machine Learning Engineer - Training & Infrastructure
P-1 AI, San Francisco, California, United States, 94199
About P-1 AI
We are building an engineering AGI. We founded P-1 AI with the conviction that the greatest impact of artificial intelligence will be on the built world—helping mankind conquer nature and bend it to our will. Our first product is Archie, an AI engineer capable of quantitative and spatial reasoning over physical product domains that performs at the level of an entry‑level design engineer. We aim to put an Archie on every engineering team at every industrial company on earth. About The Role
We’re looking for an experienced engineer to take ownership of
LLM training operations
across our applied research team. Your focus will be on
making large‑scale GPU training run reliably, efficiently, and fast
on a dedicated mid‑size GPU cluster and possibly on cloud platforms as well. You’ll work closely with researchers and ML engineers developing new models and agentic systems, ensuring their experiments scale smoothly across multi‑node GPU clusters. From debugging NCCL deadlocks to optimizing FSDP configs, you’ll be the go‑to person for training infrastructure and performance. What You’ll Do
Own the training pipeline for large‑scale LLM fine‑tuning and post‑training workflows Configure, launch, monitor, and debug multi‑node distributed training jobs using FSDP, DeepSpeed, or custom wrappers Contribute to upstream and internal forks of training frameworks like TorchTune, TRL, and Hugging Face Transformers Tune training parameters, memory footprints, and sharding strategies for optimal throughput Work closely with infra and systems teams to maintain the health and utilization of our GPU clusters (e.g., Infiniband, NCCL, Slurm, Kubernetes) Implement features or fixes to unblock novel use cases in our LLM training stack About You
3+ years working with large‑scale ML systems or training pipelines Deep familiarity with PyTorch, especially distributed training via FSDP, DeepSpeed, or DDP Comfortable navigating training libraries like TorchTune, Accelerate, or Trainer APIs Practical experience with multi‑node GPU training, including profiling, debugging, and optimizing jobs Understanding of low‑level components like NCCL, Infiniband, CUDA memory, and model partitioning strategies You enjoy bridging research and engineering—making messy ideas actually run on hardware Nice to Have
Experience maintaining Slurm, Ray, or Kubernetes clusters Past contributions to open‑source ML training frameworks Exposure to model scaling laws, checkpointing formats (e.g., HF sharded safetensors vs. distcp), or mixed precision training Familiarity with on‑policy reinforcement learning setups with inference (policy rollouts) as part of the training loop, such as GRPO, PPO, or A2C Experience working at a startup Interview Process
Initial screening – Head of Talent (30 mins) Hiring manager interview – Head of AI (45 mins) Technical interview – AI Chief Scientist and/or Head of AI (45 mins) Culture fit / Q&A (maybe in person) – with co‑founder & CEO (45 mins) Seniority level
Mid‑Senior level Employment type
Full‑time Job function
Engineering and Information Technology Industries
Software Development
#J-18808-Ljbffr
We are building an engineering AGI. We founded P-1 AI with the conviction that the greatest impact of artificial intelligence will be on the built world—helping mankind conquer nature and bend it to our will. Our first product is Archie, an AI engineer capable of quantitative and spatial reasoning over physical product domains that performs at the level of an entry‑level design engineer. We aim to put an Archie on every engineering team at every industrial company on earth. About The Role
We’re looking for an experienced engineer to take ownership of
LLM training operations
across our applied research team. Your focus will be on
making large‑scale GPU training run reliably, efficiently, and fast
on a dedicated mid‑size GPU cluster and possibly on cloud platforms as well. You’ll work closely with researchers and ML engineers developing new models and agentic systems, ensuring their experiments scale smoothly across multi‑node GPU clusters. From debugging NCCL deadlocks to optimizing FSDP configs, you’ll be the go‑to person for training infrastructure and performance. What You’ll Do
Own the training pipeline for large‑scale LLM fine‑tuning and post‑training workflows Configure, launch, monitor, and debug multi‑node distributed training jobs using FSDP, DeepSpeed, or custom wrappers Contribute to upstream and internal forks of training frameworks like TorchTune, TRL, and Hugging Face Transformers Tune training parameters, memory footprints, and sharding strategies for optimal throughput Work closely with infra and systems teams to maintain the health and utilization of our GPU clusters (e.g., Infiniband, NCCL, Slurm, Kubernetes) Implement features or fixes to unblock novel use cases in our LLM training stack About You
3+ years working with large‑scale ML systems or training pipelines Deep familiarity with PyTorch, especially distributed training via FSDP, DeepSpeed, or DDP Comfortable navigating training libraries like TorchTune, Accelerate, or Trainer APIs Practical experience with multi‑node GPU training, including profiling, debugging, and optimizing jobs Understanding of low‑level components like NCCL, Infiniband, CUDA memory, and model partitioning strategies You enjoy bridging research and engineering—making messy ideas actually run on hardware Nice to Have
Experience maintaining Slurm, Ray, or Kubernetes clusters Past contributions to open‑source ML training frameworks Exposure to model scaling laws, checkpointing formats (e.g., HF sharded safetensors vs. distcp), or mixed precision training Familiarity with on‑policy reinforcement learning setups with inference (policy rollouts) as part of the training loop, such as GRPO, PPO, or A2C Experience working at a startup Interview Process
Initial screening – Head of Talent (30 mins) Hiring manager interview – Head of AI (45 mins) Technical interview – AI Chief Scientist and/or Head of AI (45 mins) Culture fit / Q&A (maybe in person) – with co‑founder & CEO (45 mins) Seniority level
Mid‑Senior level Employment type
Full‑time Job function
Engineering and Information Technology Industries
Software Development
#J-18808-Ljbffr