Browse 33 exciting jobs hiring in Distributed Training now. Check out companies hiring such as Character.ai, Adobe, Awesome Motive in Cleveland, Santa Rosa, Seattle.
Help advance state-of-the-art AI safety by researching, implementing, and integrating alignment and evaluation techniques for large language models at a high-growth consumer AI company.
Lead research and engineering efforts in generative AI and foundation models to create next-generation visual and multimodal creative experiences at Adobe.
Specter is hiring an ML Infrastructure Engineer to design and scale training pipelines, optimized model serving, and continuous production workflows for real-time edge perception systems.
Lead high-impact, product-aligned experiments on foundation models using PyTorch and distributed training to improve real-world customer outcomes at Liquid AI.
Lead development and productization of large-scale deep learning models for digital biology at NVIDIA, building microservices and agentic AI solutions for life-sciences applications.
Lead large-scale LLM training and synthetic data pipelines at Periodic Labs to build scientifically knowledgeable models and scale training across supercomputing infrastructure.
Periodic Labs seeks an experienced Distributed Training Engineer to optimize and operate frontier-scale LLM training systems that power AI-driven scientific research.
James Talarico’s U.S. Senate campaign is seeking a Distributed Organizing Manager to build and run an inclusive, high-impact distributed volunteer program across phone, text, and virtual channels.
NVIDIA seeks a Senior AI Software Engineer to extend Megatron Core and NeMo frameworks through distributed training innovations, performance tuning, and scalable tooling for large-scale LLM and multimodal model workflows.
Help build and scale the ML infrastructure that powers GTV's real-time generative video features, working closely with research and product teams in San Francisco.
NVIDIA is seeking a Senior Software Engineer to drive JAX core design and performance optimizations for high‑performance deep learning on NVIDIA hardware.
Experienced SRE with distributed systems and LLM experience needed to design and operate scalable, reliable managed AI services for a mission-driven, sustainability-focused AI infrastructure company.
Lead engineering efforts on LinkedIn's AI Platform to scale model training, feature engineering, and high-performance model serving for large language and recommendation models.
Experienced ML scientist needed to research, develop, and productionize advanced deep learning models (with emphasis on language models) for Turnitin's global education products in a fully remote US role.
Join Quilter's ML team to design and scale large, high-performance models and training pipelines that automate PCB layout and optimization for real-world hardware design.
Quilter is hiring a Senior ML Ops Engineer to own and scale production ML training and inference infrastructure for automated PCB layout generation.
Join Voltai as a Machine Learning Engineer to build and deploy high-performance LLMs, retrieval pipelines, and agentic frameworks tailored to semiconductor and electronics design.
Lead architecture and operation of scalable ML platform infrastructure at NVIDIA to empower researchers and engineers to train and deploy large-scale models on powerful GPU systems.
Boson AI seeks an experienced research engineer to optimize training and inference pipelines on GPU clusters using CUDA/Triton, PyTorch, and distributed optimization techniques.
Lead the strategy and roadmap for Lambda’s networking backbone, building secure, low-latency connectivity and high-performance networking services for large-scale AI and HPC workloads.
Lead development and deployment of large-scale diffusion models to generate realistic 3D worlds and assets at an early-stage San Francisco AI startup.
NVIDIA seeks an experienced Machine Learning Engineer to design and build high-quality LLM training datasets and tooling to improve foundation model performance across multi-modal and safety-critical use cases.
An early-stage AI company developing video foundation models seeks an Applied AI Researcher to train and distill large-scale diffusion transformer models on-site in San Francisco.
Field AI seeks a Robotics Research AI Engineer to develop robot-learning methods and foundation-model solutions that get deployed on real robots in challenging field environments.
Help power AGI research by building ML infrastructure and tools that make researchers and GPUs dramatically more productive in a hybrid San Francisco role.
Agtonomy is hiring a Senior Software Engineer to scale ML infrastructure and productionize distributed training, data pipelines, and model deployment for autonomy on heavy equipment.
Cognitiv seeks a Machine Learning Research Scientist Intern to help optimize DNNs and LLM integrations for high-performance AdTech applications while gaining hands-on research and engineering experience.
Apply advanced reinforcement and imitation learning to build production RL/IL systems for multi-arm robotic food manipulation at Chef Robotics.
Lead the optimization and scaling of distributed training infrastructure for foundation models, improving wall-clock convergence by tuning data pipelines, kernels, and multi-node systems.
Prime Intellect is hiring a Growth Operations specialist to build content-led demand generation, SEO authority, and revenue operations that drive PQLs and marketplace growth.
Help scale state-of-the-art video generation models by designing and shipping CUDA/Triton kernels, PyTorch integrations, and end-to-end performance improvements at Mirage's NYC HQ.
Join a research team building agentic capabilities for ChatGPT, contributing to research, large-scale training, evaluations, and production deployment in a hybrid San Francisco role.
At EchoTwin AI, contribute to multimodal model development and deployment by building and optimizing vision-language systems that enable real-time urban infrastructure insights.
Below 50k*
0
|
50k-100k*
1
|
Over 100k*
28
|