Software Engineer, Inference - TL (San Francisco) Job at OpenAI, San Francisco, CA

Y0pmeTB4UEVXMEhKa1pyRnFiMmRycXF0dEE9PQ==
  • OpenAI
  • San Francisco, CA

Job Description

About the Team

Our team brings OpenAI's most capable research and technology to the world through our products. We empower consumers, enterprises, and developers alike to access state-of-the-art AI models - unlocking new capabilities across productivity, creativity, and more. We focus on high-performance model inference and accelerating research through efficient and reliable infrastructure.

About the Role

We're looking for a hands-on Tech Lead to drive the design, optimization, and scaling of our inference systems. In this role, you'll lead engineering efforts to ensure our largest models run with exceptional efficiency in high-throughput, low-latency environments. You'll be responsible for shaping our CUDA strategy, driving performance at the kernel level, and collaborating across teams to deliver end-to-end production readiness.

In this role, you will:

  • Lead the design and implementation of core inference infrastructure for serving frontier AI models in production.

  • Own and optimize CUDA-based systems and kernels to maximize performance across our fleet.

  • Partner with researchers to integrate novel model architectures into performant, scalable inference pipelines.

  • Build tooling and observability to detect bottlenecks, guide system tuning, and ensure stable deployment at scale.

  • Collaborate cross-functionally to align technical direction across research, infra, and product teams.

  • Mentor engineers on GPU performance, CUDA development, and distributed inference best practices.

You may thrive in this role if you:

  • Have deep expertise in CUDA, including writing and optimizing high-performance kernels for inference or training workloads.

  • Have experience leading complex engineering efforts, particularly at the systems and performance layer of large-scale ML infrastructure.
    Understand the full inference stack - from model loading and memory management to communication libraries and deployment orchestration.

  • Are comfortable working in large, distributed GPU environments and debugging performance issues across hardware and software layers.

  • Have strong familiarity with PyTorch and NVIDIA's GPU software stack (NCCL, NVLink, MIG, etc.).

  • Take a systems-level view, but aren't afraid to dive into low-level code when performance is on the line.

Bonus:

  • Experience with inference frameworks like TensorRT, vLLM, SGLang, or custom model parallelism infrastructure.

  • Familiarity with TPU, AMD GPUs, ROCm, HIP, TensorRT-LLM, Ray Serve, Megatron, MPI, or Horovod.

  • Familiarity with profiling tools (Nsight, nvprof, or custom observability stacks).
    Background in HPC or large-scale distributed systems engineering.

About OpenAI

OpenAI is an AI research and deployment company dedicated to ensuring that general-purpose artificial intelligence benefits all of humanity. We push the boundaries of the capabilities of AI systems and seek to safely deploy them to the world through our products. AI is an extremely powerful tool that must be created with safety and human needs at its core, and to achieve our mission, we must encompass and value the many different perspectives, voices, and experiences that form the full spectrum of humanity.

We are an equal opportunity employer and do not discriminate on the basis of race, religion, national origin, gender, sexual orientation, age, veteran status, disability or any other legally protected status.

OpenAI Affirmative Action and Equal Employment Opportunity Policy Statement

For US Based Candidates: Pursuant to the San Francisco Fair Chance Ordinance, we will consider qualified applicants with arrest and conviction records.

We are committed to providing reasonable accommodations to applicants with disabilities, and requests can be made via this link.

OpenAI Global Applicant Privacy Policy

At OpenAI, we believe artificial intelligence has the potential to help people solve immense global challenges, and we want the upside of AI to be widely shared. Join us in shaping the future of technology.

Compensation Range: $460K - $685K

#J-18808-Ljbffr

Job Tags

Full time,

Similar Jobs

Leading Edge Connections

Chat Representative (Remote) Work from home Job at Leading Edge Connections

 ...! All our teams are 100% Native Virtual (work at home) and always will be! No more stuffy offices...  ...Pop and Country ~1-3 years call center chat experience required ~ Technical...  ...customer issues ~ Great understanding of text and chat acronyms ~ Have excellent grammar... 

Navy Federal Credit Union

Financial Advisor - Lexington Park Job at Navy Federal Credit Union

 ...of investment products Administer and document account activity and execution in accordance with NFFG process and procedures and FINRA, NASAA, SEC regulations Conduct review of accounts to communicate account status, explain activities and inform clients of new offerings... 

Western New York Public Broadcasting Association

Internship - Radio News Job at Western New York Public Broadcasting Association

 ...Job Description Job Description Buffalo Toronto Public Media is your local source for television and radio - WNED PBS, WNED KIDS, WNED Create, WBFO NPR, WNED Classical, and WBFO The Bridge! We reach Western New York and Southern Ontario and can be streamed from... 

Acentra Health

Clinical Review Manager (RHIA/RHIT Certification - Fully Remote) Job at Acentra Health

 ...its Privacy and Security Rules.QualificationsRequired QualificationsAHIMA hospital-based coding credentials required.RHIA or RHIT Certification5+ years of hospital clinical coding experience required.3+ years of hospital utilization review experience required.... 

Reef Systems

Certified Strength and Conditioning Coach (CSCS) - HPO - Fort Bliss, TX Job at Reef Systems

Certified Strength and Conditioning Coach (CSCS) HPO Fort Bliss, TX Reef Systems is seeking a fully qualified certified strength and conditioning coach to support the 7 ASOS at Fort Bliss, TX. Introduction : The Human Performance Optimization (HPO) mission...