The role focuses on optimizing AI models for efficiency, involving GPU/CPU code profiling, high-performance programming, and developing performance tools.
About Luma AI
About the Role
Responsibilities
Experience
Luma's mission is to build multimodal AI to expand human imagination and capabilities. We believe that multimodality is critical for intelligence. To go beyond language models and build more aware, capable and useful systems, the next step function change will come from vision. So we are working on training and scaling up multimodal foundation models for systems that can see and understand, show and explain, and eventually interact with our world to effect change.
The Performance Optimization team at Luma is dedicated to maximizing the efficiency and performance of our AI models. Working closely with both research and engineering teams, this group ensures that our cutting-edge multimodal models can be trained efficiently and deployed at scale while maintaining the highest quality standards.
- Profile and optimize GPU/CPU/Accelerator code for maximum utilization and minimal latency
- Write high-performance PyTorch, Triton, CUDA, deferring to custom PyTorch operations if necessary
- Develop fused kernels and leverage tensor cores and modern hardware features for optimal hardware utilization on different hardware platforms
- Optimize model architectures and implementations for distributed multi-node production deployment
- Build performance monitoring and analysis tools and automation
- Research and implement cutting-edge optimization techniques for transformer model
- Expert-level proficiency in Triton/CUDA programming and GPU optimization
- Strong PyTorch skills
- Experience with PyTorch kernel development and custom operations
- Proficiency with profiling tools (NVIDIA Nsight, torch profiler, custom tooling)
- Deep understanding of transformer architectures and attention mechanisms
- (Preferred) Experience with compilers/exporters such as torch.compile, TensorRT, ONNX, XLA
- (Preferred) Experience optimizing inference workloads for latency and throughput
- (Preferred) Experience with Triton compiler and kernel fusion techniques
- (Preferred) Knowledge of warp-level intrinsics and advanced CUDA optimization
Your applications are reviewed by real people.
CompensationThe base pay range for this role is $187,500 – $395,000 per year.
About LumaLuma’s mission is to build unified general intelligence that can generate, understand, and operate in the physical world.
We believe that multimodality is critical for intelligence. To go beyond language models and build more aware, capable and useful systems, the next step function change will come from vision. So, we are working on training and scaling up multimodal foundation models for systems that can see and understand, show and explain, and eventually interact with our world to effect change.
Top Skills
Cuda
Nvidia Nsight
Onnx
PyTorch
Tensorrt
Torch Profiler
Triton
Xla
Similar Jobs
Artificial Intelligence • Consumer Web • Digital Media • Information Technology • Social Impact • Software
As a Senior Quality Platform Engineer, you will develop and maintain quality infrastructure, improve developer experience, and implement quality engineering practices to ensure scalable, efficient testing workflows.
Top Skills:
AWSAzureCircleCICypressDockerGCPGithub ActionsGitlabJavaJavaScriptJestJunitKubernetesPlaywrightPythonRubyTypescript
Artificial Intelligence • Fintech • Payments • Business Intelligence • Financial Services • Generative AI
The Account Executive will drive outbound sales, establish relationships with C-level executives, negotiate contracts, and support SME growth through innovative financial solutions.
Top Skills:
Google SuiteLinkedin Sales NavigatorOutreachSalesforceZoominfo
Artificial Intelligence • Fintech • Payments • Business Intelligence • Financial Services • Generative AI
Responsible for managing end-to-end partner relationships, driving partner revenue, and executing go-to-market strategies while building a strong partner ecosystem.
Top Skills:
HubspotSalesforceZoominfo
What you need to know about the Toronto Tech Scene
Although home to some of the biggest names in tech, including Google, Microsoft and Amazon, Toronto has established itself as one of the largest startup ecosystems in the world. And with over 2,000 startups — more than 30 percent of the country's total startups — Toronto continues to attract new businesses. Be it helping entrepreneurs manage their finances, simplifying business operations by automating payroll or assisting pharmaceutical companies in launching new drugs, the city's tech scene is just getting started.


