NVIDIA is seeking a Forward Deployed Engineer to join our AI Accelerator team, working directly with strategic customers to implement and optimize pioneering AI workloads! You will provide hands-on technical support for advanced AI implementations, complex distributed systems, and ensure customers achieve optimal performance from NVIDIA's AI platform across diverse environments. We work directly with the world's most innovative AI companies to solve their toughest technical challenges.
What you will be doing:
In this role, you will implement innovative solutions that push the boundaries of what's possible with AI infrastructure while directly impacting customer success with breakthrough AI initiatives!
Technical Implementation: Design and deploy custom AI solutions including distributed training, inference optimization, and MLOps pipelines across customer environments
Customer Support: Provide remote technical support to strategic customers, optimize AI workloads, diagnose and resolve performance issues, and guide technical implementations through virtual collaboration
Infrastructure Management: Deploy and manage AI workloads across DGX Cloud, customer data centers, and CSP environments using Kubernetes, Docker, and scheduling systems for GPU
Performance Optimization: Profile and optimize large-scale model training and inference workloads, implement monitoring solutions, and resolve scaling challenges
Integration Development: Build custom integrations with customer systems, develop APIs and data pipelines, and implement enterprise software connections
End-user Documentation: Create implementation guides, documentation for resolution approaches and standard methodologies for complex AI deployments
What we need to see:
8+ years of experience in customer-facing technical roles (Solutions Engineering, DevOps, ML Infrastructure Engineering)
BS, MS, or Ph.D. in CS, CE, EE (related technical field) or equivalent experience.
Strong proficiency with Linux systems, distributed computing, Kubernetes, and GPU scheduling
AI/ML experience supporting inference workloads and training at large-scale
Programming skills in Python, with experience in PyTorch, TensorFlow, or similar AI frameworks
Customer engagement ability to work effectively with technical teams under high-pressure situations
Ways to stand out from the crowd:
NVIDIA ecosystem experience with DGX systems, CUDA, NeMo, Triton, or NIM
Cloud platforms hands-on experience with AWS, Azure, or GCP AI services
MLOps expertise with containerization, CI/CD pipelines, and observability tooling
Infrastructure as code experience with Terraform, Ansible, or similar automation tools
Enterprise software integration experience with Salesforce, ServiceNow, or similar platforms
With competitive salaries and a generous benefits package, NVIDIA is widely considered to be one of the technology world’s most desirable employers; we have some of the most forward-thinking and hardworking people in the world working for us and, due to unparalleled growth, our best-in-class teams are rapidly growing.
#LI-Hybrid
Your base salary will be determined based on your location, experience, and the pay of employees in similar positions. The base salary range is 168,000 USD - 264,500 USD for Level 4, and 200,000 USD - 322,000 USD for Level 5.You will also be eligible for equity and benefits.
If an employer mentions a salary or salary range on their job, we display it as an "Employer Estimate". If a job has no salary data, Rise displays an estimate if available.
NVIDIA invites motivated undergraduate and graduate students to participate in a 12-week Deep Learning Computer Architecture internship, contributing to GPU, deep learning, and high-performance computing projects.
NVIDIA seeks ambitious Bachelor's, Master's, and PhD students for 12-week Mixed Signal and Digital Circuit Design internships to contribute to impactful hardware projects using modern EDA tools and lab test equipment.
Lead engineering and modernization efforts for HR and payroll systems at NBCUniversal/Versant, ensuring reliable operations through a complex spin-off and preparing systems for future cloud migrations.
The Oregon DOJ's Child Support Technology Team is hiring a senior Release Manager to design and operate release pipelines, automate batch processes, and ensure secure, compliant deployments of the Origin system.
Monarch is hiring a hands-on Infrastructure & MLOps Engineer to build and operate scalable cloud and AI infrastructure that powers their personal finance platform.
Mainstay is hiring a senior full-stack engineer to build and ship production-ready LLM features, retrieval systems, and scalable backend services for AI-driven real-estate workflows.
Experienced SFCC developer needed to build and maintain PopSockets' ecommerce platform, driving SFRA development, integrations, and mentoring a small team in a remote-first environment.
Senior Backend Engineer to build scalable healthcare infrastructure at WHOOP, developing APIs and services that connect members to data science-driven health insights.
Lead the design and automation of our AWS-based platform and developer tooling to enable secure, scalable, and highly observable delivery of payments and practice-management services.
Wyetech is hiring an experienced Software Integration Engineer 3 to automate and sustain HPC telemetry, metrics, and software integrations for mission-critical federal systems.
Siepe seeks a hands-on Lead Engineer to architect and lead the next-generation document processing and LLM inference pipeline for its Financial Intelligence Platform.
Lead the backend and cloud engineering for GPTOps, building scalable, secure C# services on Azure and Kubernetes to power LLM-driven features across the Delinea platform.
Experienced staff-level engineer wanted to design and deliver cloud-native, data-rich systems (Python/TypeScript/AWS/SQL) supporting rights and contracts across NBCUniversal.
Senior-level geospatial software engineer needed to architect and build scalable, cloud-native geospatial systems using Python and AWS for a mission-driven small business.
Work on Exa's Agentic Search Team to build high-compute search infrastructure and full-stack features that combine LLMs, web-scale crawling, and interactive frontends.
NVIDIA is a publicly traded, multinational technology company headquartered in Santa Clara, California. NVIDIA's invention of the GPU in 1999 sparked the growth of the PC gaming market, redefined computer graphics, and ignited the era of modern AI.
156 jobs