10 open positions available
The job involves managing hardware, optimizing system performance, and building diagnostic tools for high-performance computing environments. | Requires deep understanding of hardware, low-level OS interactions, and system optimization, which are not reflected in your experience. | The Opportunity At Luma AI, "full-stack" has a distinct meaning. It means understanding everything from the generative model down to the silicon it runs on. We are pushing the physical limits of current hardware to train Omni models that understand the world. This requires a level of engineering rigor that standard cloud environments simply do not demand. We are looking for engineers who are tired of high-level abstractions and want to work on the metal that powers the AI revolution. Where You Come In You will operate at the jagged edge where software meets hardware. Standard cloud providers abstract away the complexity; we embrace it. You will be responsible for maximizing efficiency from our heterogeneous fleet of NVIDIA and AMD accelerators. This role is about precision, performance, and the relentless pursuit of system optimization in a multi-vendor supercomputing environment. What You Will Build • The Bare Metal Stack: Manage and optimize the lifecycle of bare-metal servers, ensuring that our OS, drivers, and firmware are tuned for peak AI performance. • High-Throughput Interconnects: Engineer the software configurations for our InfiniBand and RoCE fabrics, solving the intricate data movement challenges that define modern distributed training. • Performance Diagnostics: Build the tooling to visualize what is happening inside the cluster, turning opaque hardware counters into actionable signals for debugging latency and throughput. The Profile We Are Looking For • Low-Level Fluency: You are not afraid of the kernel. You understand interrupts, memory management, and how the OS interacts with peripheral devices. • Hardware Curiosity: You understand that software doesn't run in a vacuum. You are interested in the physical constraints of GPUs, networking cards, and storage subsystems. • First-Principles Reasoning: When a system behaves unexpectedly, you don't just restart it; you investigate the physics of the failure to ensure it is solved permanently.
Design and implement backend systems for autonomous AI agents, focusing on memory, planning, and tool integration. | Experience in building complex, event-driven, non-deterministic systems and leading technical product definitions. | The Opportunity Luma AI is defining the future of autonomous creativity. We are moving beyond chat interfaces to build intelligent agents that act as creative partners, executing complex, multi-step workflows on behalf of the user. We are looking for a foundational engineer to architect the brain of our product platform, the reasoning engine, memory systems, and tool-use capabilities that power our agents. Where You Come In You will define the cognitive architecture of our platform. This role involves building the backend systems that allow AI agents to plan, execute, and evaluate complex creative tasks autonomously. You will operate with high agency, inventing new patterns for state management and orchestration that do not yet exist in standard frameworks. What You Will Build Cognitive Infrastructure: Architect the backend systems for agent memory, planning, and context management, enabling long-horizon interactions for creative projects. Autonomous Workflows: Design the orchestration layer that allows models to chain multiple tools, video generation, editing, and reasoning, into seamless, magical user experiences. Novel API Patterns: Invent new protocols for asynchronous, stateful communication between the client and our agentic backend. The Profile We Are Looking For Technical Leadership: You have demonstrated the ability to lead the technical definition of undefined product areas, taking abstract concepts like autonomous creativity and turning them into concrete engineering roadmaps. Orchestration Expertise: You have deep experience building complex, event-driven architectures and are familiar with the challenges of non-deterministic systems. Builder DNA: You prefer writing your own orchestration logic over relying on off-the-shelf abstractions, and you thrive in environments where you define the primitives.
Supporting research and development projects related to polymers and materials engineering, including literature review, testing, and analysis. | Bachelor's degree in Materials Science or Engineering with experience in polymers, material testing, and data analysis; no mention of software or system architecture skills required for this role. | The Opportunity Luma AI operates as a full-stack AI lab, training foundational models and building the products that utilize them. We have secured the capital and compute resources of a major tech company, raising over $1.3 billion, while maintaining the agile, high-impact structure of a focused startup. This unique position allows our core engineering team to operate with immense leverage, deploying resources that few other companies can match to solve the hardest problems in multimodal intelligence. Where You Come In You will serve as a technical authority within the engineering organization, architecting the foundational systems that power our generative media platform. This is a hands-on leadership role where you will define the technical strategy for our backend infrastructure, ensuring it can support massive-scale inference and complex agentic workflows. You will work across the entire stack, from the database layer to the orchestration of our proprietary models. What You Will Build Distributed Systems Architecture: Design and implement high-throughput systems capable of coordinating massive GPU clusters and serving millions of concurrent users with low latency. The Intelligence Layer: Architect the sophisticated backend logic that translates frontier research into reliable product capabilities, solving novel challenges in asynchronous processing and state management. Technical Strategy: Lead the definition of our backend roadmap, making high-stakes decisions on technology selection, schema design, and infrastructure scalability. The Profile We Are Looking For Systems Generalist: You possess deep, versatile expertise across the backend spectrum, with the ability to operate from the kernel level up to the API contract. Architectural Authority: You have a track record of designing complex systems from first principles, demonstrating the judgment to navigate trade-offs between speed, reliability, and scale in a high-growth environment. Technical Depth: Expert-level command of Python or Go, with profound knowledge of Kubernetes, distributed databases, and cloud infrastructure.
Develop and implement full-stack features for AI-powered applications, integrating research breakthroughs into user-facing products. | Proficiency in Python or Go, React, TypeScript, and experience shipping complex products from scratch. | The Opportunity Luma AI is building the next era of creative tools, powered by our own Omni models that see, hear, and understand the world. We combine the resources of a capitalization-heavy AI lab with the speed of a small, elite product team. This environment guarantees that you will sit at the intersection of Product, Design, and Research, translating raw model capabilities into magical user experiences that define a new category of software. Where You Come In You will be a foundational member of the product engineering team, responsible for the end-to-end delivery of our most ambitious features. This role requires a builder who can traverse the entire stack. You will partner directly with researchers to productionize novel capabilities, ensuring that the complexity of the model is invisible to the user. What You Will Build End-to-End Features: You will own the full lifecycle of product development, from designing the data model and API endpoints to implementing the client-side logic and UI components. Novel Interactions: You will invent new paradigms for interacting with generative AI, bridging the gap between asynchronous backend processes and responsive, high-fidelity frontend interfaces. Research Integration: You will rapidly prototype and ship features that leverage our latest research breakthroughs, turning experimental capabilities into stable, polished product workflows The Profile We Are Looking For Full-Stack Fluency: You are an expert in Python or Go on the backend and possess strong proficiency with modern frontend frameworks like React and TypeScript. Craft Obsession: You care deeply about the feel of the product. You understand that in an AI-native application, latency, reliability, and interaction design are inextricably linked. Proven Builder: You have a history of shipping complex, user-facing products from scratch, thriving in environments where you define the spec as you build.
Support research-to-product pipelines and develop material testing protocols. | Requires expertise in backend systems, APIs, cloud infrastructure, and technical leadership, which are not present in your profile. | The Opportunity Luma AI is a full-stack AI lab building Multimodal AGI. To solve the hardest problems in intelligence, models must learn from audio, video, and images. We train our own foundational models and build the products that sit on top of them. We offer the rare combination of massive compute resources and funding typically found at major tech companies, coupled with the agility and individual impact of a small, early-stage startup. Where You Come In You will join a small, high-autonomy team to architect the systems that transform our frontier research into a creative partner for millions of users. This role sits at the intersection of Engineering, Product, and Research, requiring you to translate raw model capabilities into fast, reliable, and magical product experiences. What You Will Build The Logic of Creativity: Architect the backend systems for Dream Machine, moving beyond simple API calls to build complex, asynchronous workflows that power generative media. Research-to-Product Pipelines: Collaborate directly with researchers to productionize novel capabilities, such as camera control or visual reasoning. End-to-End Architecture: Own features from the database schema to the client-facing API, ensuring low latency for high-fidelity media generation. The Profile We Are Looking For Technical Leadership: You have a proven ability to navigate ambiguity and lead technical definition, taking projects from a vague concept to a shipped product without needing established roadmaps. High Agency Builder: You have a history of creating, launching, and building new systems from scratch. You thrive in environments where you own the stack from the database to the interface. Product Instincts: You care deeply about the end-user experience and understand how backend decisions impact the feel of the product. Technical Fluency: You have deep expertise in Python or Go, REST/gRPC APIs, and modern cloud infrastructure.
Design and maintain scalable AI training and inference platforms, ensuring high availability and operational excellence. | Experience with large-scale ML workloads, distributed training, GPU resource management, and automation in complex environments. | The Opportunity Luma AI is training the multimodal models that will define the next era of intelligence. Unlike other software companies, our product roadmap is driven by research breakthroughs. This requires a symbiotic relationship between our infrastructure engineers and our research scientists. We provide the massive compute resources necessary to compete at the top tier of AI, with a team structure that ensures you are in the room where the models are designed. Where You Come In You will build the platform that enables scientific discovery. Your work will directly accelerate the velocity of our research team, ensuring they have a stable, performant, and scalable environment to train and test the next generation of Omni models. You will translate the complex requirements of large-scale ML workloads into robust infrastructure reality. What You Will Build • Research Platforms: Design and maintain the scheduling and orchestration systems that allow researchers to launch and manage massive training jobs with ease. • Observability for Intelligence: Implement deep observability stacks that provide transparency into cluster health, allowing us to predict and prevent interruptions to critical training runs. • Scalable Inference: Architect the production systems that serve our models to the world, balancing the high availability required for consumer products with the massive compute intensity of generative AI. The Profile We Are Looking For • Service Orientation: You understand that reliable infrastructure is the enabler of innovation, and you care deeply about the developer experience of the researchers you support. • Operational Excellence: You have a track record of maintaining high availability in complex, distributed environments, using automation to reduce toil. • ML Infrastructure Fluency: You are familiar with the unique demands of AI workloads, including the management of GPU resources and the intricacies of distributed training.
Maintain and scale GPU infrastructure clusters by developing monitoring, automation, and fault-tolerant systems to ensure high availability and performance. | Master's degree plus 2+ years experience with AWS, Kubernetes, CI/CD pipelines (GitLab, Jenkins), Terraform, Vault PKI, Python and shell scripting for automation. | Position: Staff Software Engineer - Reliability Job Description Luma AI runs on thousands of H100 GPUs across multiple providers and clusters for Training, Data Processing and Inference. Working with the Infrastructure and Research teams, the Staff Software Engineer – Reliability maintains the health of our GPU clusters, developing the monitoring and management tools necessary to maximize their performance. Specific Job Duties Include The Following • Collaborate with researchers and engineers to specify the availability, performance, correctness, and efficiency requirements of the current and future versions of our GPU infrastructure. (15%) • Work with multiple GPU cloud providers to scale up, scale down, maintain and monitor our GPUs in many clusters. (20%) • Design and implement solutions to ensure the scalability of our infrastructure to meet rapidly increasing demands. (15%) • Implement and manage monitoring systems to proactively identify issues and anomalies in our production environment. (10%) • Implement fault-tolerant and resilient design patterns to minimize service disruptions. (10%) • Build and maintain automation tools to streamline repetitive tasks and improve system reliability. (15%) • Participate in an on-call rotation to respond to critical incidents and ensure 24/7 system availability alongside other infrastructure developers. (5%) • Develop and maintain service level objectives (SLOs) and service level indicators (SLIs) to measure and ensure system reliability. (10%) Job Requirements Requires a Master’s degree (or foreign equivalent) in Computer Science, Information Technology, Electronic Engineering, or related field of study, plus 2 years of experience in the job offered, Software Engineer, or a related occupation. Position requires at least 2 years of experience in the following skills: • CI/CD pipelines and automation using AWS and Kubernetes. • CI/CD using gitlab, docker, Kubernetes. • AWS infrastructure using Terraform, packer. • Certs LifeCycle Management using vault with PKI. • Python to automate every aspect of the pipeline. • Python and Shell Scripts for build automation. • Jenkins servers for continuous integration.
Lead product positioning, launch, and adoption by crafting messaging, running go-to-market campaigns, enabling sales, and translating customer insights into product strategy. | 5-8 years in product marketing or related roles ideally in B2B SaaS or AI tech, strong storytelling and analytical skills, experience with product launches, and ability to collaborate cross-functionally. | At Luma, we’re building multimodal AI to expand human imagination and capabilities. We believe the next leap in AI won’t come from language alone, but from models that can see, understand, create, and interact with the world visually. Dream Machine is just the beginning. We’re creating an intelligent creative partner that helps people express what’s in their mind’s eye without needing the perfect words. Think: generative AI that’s not just helpful, but beautifully intuitive and deeply visual. The Role We need our first Senior Product Marketing Manager (PMM) to lead how our product is positioned, launched, and adopted in the market. This role sits at the intersection of product, marketing, and sales — turning technical capabilities into clear, compelling value for customers and establishing the foundation of the PMM function at Luma. What a Typical Week Might Look Like: Message & Motivate: Craft narratives, value propositions, and messaging frameworks that resonate with creators, technical users, and executives. Launch & Learn: Own go-to-market strategy for new features and models, running campaigns and launches end-to-end in partnership with product and growth. Enable & Empower: Build sales tools, pitch decks, one-pagers, and case studies that help teams tell the Luma story with clarity and impact. Listen & Translate: Conduct customer interviews, competitive analysis, and market research, then funnel insights back to influence roadmap and positioning. Create & Educate: Write blogs, produce tutorials, and ship marketing content that simplifies complex AI into stories that inspire and inform. Measure & Iterate: Track launch and campaign performance, analyze adoption metrics, and optimize messaging and tactics based on data. Your Background Looks Something Like: The Essentials 5–8 years in product marketing, product management, or a related role, ideally in B2B SaaS or creative/AI tech Proven success leading end-to-end product launches and go-to-market campaigns Strong storytelling skills — able to turn complex technical details into customer-centric narratives Analytical and data-driven mindset, tying insights to strategy and measurable outcomes Skilled at cross-functional collaboration and influencing without authority Comfortable thriving in a fast-moving, creative, and high-ownership environment What Would Make You Exceptional: Experience in AI/ML or generative AI products Background in design-forward SaaS or creative tools High-growth or scale-up experience where you built frameworks while delivering results Technical literacy (machine learning concepts, data workflows) MBA or technical degree (e.g., Computer Science, Engineering) You Might Really Thrive Here If… Zero-to-one is your happy place; you love building functions and processes where none exist You geek out on AI and can translate complex concepts into stories that resonate with both engineers and execs Your idea of fun is turning raw product features into campaigns that make customers say “wow” “Iterate, ship, learn” feels more natural than “draft, revise, approve” You believe great storytelling can move markets, and you want a front-row seat in shaping a new category
Build and own secure, scalable backend enterprise systems including authentication, team management, billing, and analytics dashboards. | Senior or staff-level engineer with deep Python backend experience, security focus, containerization knowledge, and history of shipping enterprise-grade products. | About Luma AI Luma's mission is to build multimodal AI to expand human imagination and capabilities. We believe that multimodality is critical for intelligence. To go beyond language models and build more aware, capable, and useful systems, the next step function change will come from vision. So, we are working on training and scaling up multimodal foundation models for systems that can see and understand, show and explain, and eventually interact with our world to effect change. Where You Come In As Luma's generative AI is adopted by the world's leading creative companies, we are building a new, foundational team to create the secure, scalable, and reliable systems these partners require. This is a zero-to-one opportunity for a versatile and product-minded backend engineer to build our enterprise platform from the ground up. You will be a founding engineer on this new team, taking high-level requirements and turning them into robust, production-grade features. This is a high-agency role for an engineer who thrives on the challenge of serving sophisticated customers and has a proven history of shipping enterprise-grade systems in a fast-paced environment. What You'll Do Architect and Own Luma's core enterprise authentication layer, including features like SSO and role-based access control. Build and Launch backend systems for enterprise team management, including invites, shared credit pools, and billing integrations. Design and Ship high-traffic, secure APIs that power our enterprise offerings and partner integrations. Develop analytics dashboards and reporting features to provide enterprise customers with critical insights into their teams' usage and performance. Collaborate closely with our security, product, and design teams to harden our infrastructure and ensure our systems meet the highest standards for privacy and data protection. Who You Are You have experience as a Senior or Staff-level software engineer, with a proven record of shipping impactful, user-focused products at scale. You possess deep, hands-on proficiency in Python, as our entire backend stack is Python-based. You have experience designing, building, and maintaining scalable backend systems and high-traffic REST APIs. You have a "security mindset" and have experience working with security departments to build secure, enterprise-ready products. You have hands-on experience with containerization technologies like Docker and a medium-level understanding of Kubernetes. You thrive in fast-paced startup environments and have high agency, with a proven ability to operate effectively in the face of ambiguity. What Sets You Apart (Bonus Points) You have been an early or founding engineer at a startup focused on building backend systems for enterprise customers You have worked at a company with a strong API-first business model. Experience building enterprise applications or authentication systems from scratch.
Ship new model architectures by integrating them into the inference engine and empower the product team to create groundbreaking features through user-friendly APIs. Build sophisticated scheduling systems to optimally leverage GPU resources while maintaining CI/CD pipelines for model processing and internal tooling. | Strong generalist Python skills and extensive experience with Kubernetes and Docker are required. Experience with high performance large scale ML systems and multimedia processing is a plus. | Luma’s mission is to build multimodal AI to expand human imagination and capabilities. We believe that multimodality is critical for intelligence. To go beyond language models and build more aware, capable and useful systems, the next step function change will come from vision. So, we are working on training and scaling up multimodal foundation models for systems that can see and understand, show and explain, and eventually interact with our world to effect change. Role & Responsibilities Ship new model architectures by integrating them into our inference engine Empower our product team to create groundbreaking features by developing user-friendly APIs and interaction patterns Build sophisticated scheduling systems to optimally leverage our expensive GPU resources while meeting internal SLOs Build and maintain CI/CD pipelines for processing/optimizing model checkpoints, platform components, and SDKs for internal teams to integrate into our products/internal tooling. Background Strong generalist Python skills Experience with queues, scheduling, traffic-control, fleet management at scale. Extensive experience with Kubernetes and Docker. Bonus points if you have experience with high performance large scale ML systems (>100 GPUs) and/or Pytorch experience. Bonus points if you have experience with ffmpeg and multimedia processing. Tech stack Must have Python Kubernetes Redis S3-compatible Storage Nice to have Pytorch CUDA Ffmpeg
Create tailored applications specifically for Luma AI with our AI-powered resume builder
Get Started for Free