Back to Home

Key Responsibilities and Required Skills for Deep Learning Performance Architect

💰 $160,000 - $240,000

AIMachine LearningEngineeringPerformanceMLOps

🎯 Role Definition

The Deep Learning Performance Architect is a senior technical leader focused on maximizing the runtime efficiency, scalability, and robustness of deep learning models across diverse hardware and deployment environments. This role blends machine learning model understanding, systems-level profiling, compiler optimizations, and close collaboration with infra and product teams to design, implement, and validate performance improvements that reduce latency, increase throughput, and lower operational costs for production AI services.

Key responsibilities include performance profiling and instrumentation, model optimization (quantization, pruning, kernel fusion), integration with accelerators and inference runtimes (TensorRT, ONNX Runtime, XLA), building performance CI and benchmarking pipelines, and mentoring engineers to adopt best practices for performant model design and deployment.


📈 Career Progression

Typical Career Path

Entry Point From:

  • Senior Machine Learning Engineer with a focus on inference and productionization
  • Senior Systems Engineer / GPU Software Engineer specializing in CUDA or accelerator runtimes
  • Research Engineer or ML Researcher working on model efficiency, compilers, or systems

Advancement To:

  • Principal Deep Learning Performance Architect / Distinguished Engineer
  • Director of ML Infrastructure or Head of AI Performance Engineering
  • VP of Engineering, ML Platforms or Chief Architect for AI Systems

Lateral Moves:

  • ML Infrastructure / MLOps Lead
  • Inference Platform Engineering Manager
  • Compiler or Accelerator Software Architect

Core Responsibilities

Primary Functions

  • Lead end-to-end performance engineering initiatives for deep learning models in production, owning profiling, root-cause analysis, and delivery of measurable latency and throughput improvements across CPU, GPU, TPU, and edge devices.
  • Design, implement, and maintain robust benchmarking and performance CI pipelines that automatically measure model latency, throughput, memory usage, and cost across multiple hardware targets and software stacks.
  • Profile and analyze model execution using low-level tools (nvprof/nsight, perf, VTune), framework profilers (PyTorch Profiler, TensorFlow Profiler), and custom instrumentation to identify hotspots, kernel inefficiencies, and memory bottlenecks.
  • Translate model-level observations into systems-level optimizations: operator fusion, kernel tuning, data-layout transformations, and memory planning to reduce peak memory and improve compute utilization.
  • Implement and validate model quantization strategies (post-training and quantization-aware training), mixed precision (FP16/BF16), and reduced-precision kernels to lower latency and inference cost while maintaining accuracy targets.
  • Architect and integrate compiler toolchains (XLA, TVM, Glow, MLIR) and inference runtimes (TensorRT, ONNX Runtime, Triton) to produce highly optimized binaries for targeted accelerators and cloud GPUs.
  • Optimize distributed training and inference pipelines by tuning communication libraries (NCCL, Gloo), overlap of compute and communication, and collective algorithms to scale models efficiently across many nodes.
  • Lead kernel-level performance work, author or tune CUDA/C++ kernels, and work with hardware abstractions to extract maximum throughput from accelerators while ensuring numerical stability and correctness.
  • Drive model architecture trade-offs with ML researchers and model teams to make models more hardware-friendly (e.g., replacing expensive attention patterns, reducing memory-bound operators, or restructuring compute graphs).
  • Build and validate end-to-end performance SLAs and SLOs for AI services, translating business latency and cost objectives into engineering tasks and monitoring targets.
  • Partner with platform and SRE teams to instrument production inference paths with observability, tracing, and alerting for performance regressions and capacity planning.
  • Create and maintain actionable documentation, best-practice guides, and templated pipelines for model teams to deploy high-performance models reproducibly across environments.
  • Lead performance-driven code reviews and mentor engineers on profiler usage, optimization patterns, and understanding of hardware/software trade-offs.
  • Evaluate and benchmark new hardware accelerators, chips, and cloud instance types, producing cost-performance analyses and recommendation reports for procurement and migration decisions.
  • Drive adoption of model optimization techniques such as pruning, operator fusion, graph rewriting, and model distillation to reduce compute and memory footprint while preserving accuracy and robustness.
  • Create reproducible experiments and A/B tests to validate that performance optimizations do not negatively impact model quality or user-facing metrics, and collaborate with data-science stakeholders to assess trade-offs.
  • Implement automated fallback strategies and multi-precision execution flows that choose the optimal runtime path based on latency, throughput, and accuracy constraints at inference time.
  • Collaborate with security, privacy, and compliance teams to ensure performance changes maintain or enhance data protection, deterministic behavior, and auditability in regulated environments.
  • Lead cross-functional efforts to containerize and package optimized models and runtimes for easy deployment across cloud, on-prem, and edge environments, including CI/CD integration.
  • Stay current with state-of-the-art techniques in model acceleration, compilers, pruning, quantization, and hardware advances, and evangelize new approaches to engineering and research teams.
  • Estimate, prioritize, and manage performance-related roadmaps and deliverables, balancing short-term wins with longer-term compiler or architecture investments to align with business objectives.
  • Troubleshoot and remediate production performance incidents, conduct postmortems, and implement systemic changes to prevent regressions and improve mean time to resolution.
  • Contribute to open-source performance tooling, libraries, or model zoos when appropriate to accelerate community-driven optimizations and leverage external advances.

Secondary Functions

  • Support ad-hoc data requests and exploratory data analysis.
  • Contribute to the organization's data strategy and roadmap.
  • Collaborate with business units to translate data needs into engineering requirements.
  • Participate in sprint planning and agile ceremonies within the data engineering team.
  • Provide subject-matter-expert input into procurement and vendor evaluations for inference runtimes, hardware accelerators, and managed AI compute services.
  • Assist recruiting teams in interviewing and evaluating candidates for ML performance and systems roles.

Required Skills & Competencies

Hard Skills (Technical)

  • Deep proficiency with deep learning frameworks (PyTorch, TensorFlow) and their profiling/graph IRs; ability to read and modify execution graphs and operator implementations.
  • Strong experience optimizing inference runtimes and integrating with accelerators using TensorRT, ONNX Runtime, XLA, TVM, Glow, or Triton.
  • Expert-level knowledge of GPU programming (CUDA, cuDNN, NCCL), memory management, and kernel optimization; experience writing and tuning CUDA/C++ kernels.
  • Practical experience with quantization (PTQ, QAT), mixed precision (FP16/BF16), pruning, operator fusion, and model compression techniques.
  • Hands-on knowledge of compiler toolchains (XLA, TVM, MLIR, LLVM) and experience driving optimizations through graph transforms and code generation.
  • Familiarity with distributed training and inference patterns, communication tuning, and scaling strategies across multi-node clusters.
  • Proficiency with performance profiling and observability tools: nvprof/nvtx/nsight systems, perf, VTune, PyTorch/TensorFlow profilers, and distributed tracing systems.
  • Solid experience with containerization and deployment tools (Docker, Kubernetes) and inference platforms (Triton, KFServing, Seldon) for scalable model serving.
  • Strong systems programming skills in C/C++ and Python; experience building production-grade libraries, bindings, and performance-sensitive microservices.
  • Knowledge of hardware architectures beyond GPUs (TPU, FPGA, NPU, ASICs) and the constraints/trade-offs when mapping models to those accelerators.
  • Experience with model format standards and conversion tools: ONNX, ONNX-ML, savedmodel, TorchScript, and experience debugging format-related performance issues.
  • Familiarity with cloud GPU/accelerator offerings (AWS EC2 GPU/Inferentia, GCP TPUs, Azure ML) and cost-performance analysis across instance types.
  • Experience implementing performance CI, reproducible benchmarking suites, and automated regression detection for model performance.
  • Understanding of numerical stability, precision trade-offs, and validation/testing strategies to ensure model correctness post-optimization.
  • Familiarity with data pipeline and preprocessing optimizations (data loading, batching, prefetching, zero-copy) that impact overall inference throughput.

Soft Skills

  • Strong collaboration and communication skills to align cross-functional stakeholders (researchers, infra, SRE, product) on performance goals and trade-offs.
  • Proven ability to lead technical projects and mentor engineers, driving adoption of performance best practices across teams.
  • Analytical mindset with strong troubleshooting techniques and a bias for data-driven decisions when evaluating optimizations and trade-offs.
  • Ability to distill complex hardware and compiler constraints into clear, actionable recommendations for non-specialist stakeholders.
  • Prioritization skills and product-minded thinking to balance latency/cost/accuracy trade-offs aligned to business metrics.
  • Patience and persistence to iterate on low-level optimizations, reproduce edge-case regressions, and validate fixes in production contexts.
  • Strong documentation and knowledge sharing skills to produce reproducible playbooks, runbooks, and internal training materials.
  • Comfort working in ambiguous, fast-changing technical domains and ability to rapidly prototype and evaluate new performance approaches.

Education & Experience

Educational Background

Minimum Education:

  • Bachelor's degree in Computer Science, Electrical Engineering, Applied Mathematics, or equivalent practical experience with significant systems/ML work.

Preferred Education:

  • Master's or PhD in Computer Science, Machine Learning, Computer Architecture, High-Performance Computing, or related fields with emphasis on systems, compilers, or ML efficiency research.

Relevant Fields of Study:

  • Computer Science
  • Machine Learning or Artificial Intelligence
  • Computer Engineering / Electrical Engineering
  • High Performance Computing / Compilers

Experience Requirements

Typical Experience Range: 6+ years in software engineering, with at least 3–5 years specifically focused on deep learning performance, inference engineering, or accelerator/runtime optimization.

Preferred:

  • 8+ years experience including leadership of cross-functional performance projects, published work or contributions to open-source ML performance tools, and proven delivery of production performance gains (e.g., >2x latency reduction or cost savings at scale).
  • Demonstrated track record optimizing models and runtimes for multiple hardware platforms (GPU/TPU/edge) and implementing production benchmarking and CI systems.