NE

Nebius

4 open positions available

1 location
1 employment type
Actively hiring
full-time

Latest Positions

Showing 4 most recent jobs
NE

Senior Technical Product Manager Token Factory - Inference

NebiusAnywherefull-time
View Job
Compensation$120K - 160K a year

Lead product roadmap and delivery of scalable ML inference capabilities, collaborate with engineering and research teams, support customer PoCs, and translate ML workflows into product requirements. | 3-5 years product management in cloud infrastructure or ML platforms, strong technical foundation in ML systems, familiarity with ML inference tools, and strong cross-team communication skills. | Why work at Nebius Nebius is leading a new era in cloud computing to serve the global AI economy. We create the tools and resources our customers need to solve real-world challenges and transform industries, without massive infrastructure costs or the need to build large in-house AI/ML teams. Our employees work at the cutting edge of AI cloud infrastructure alongside some of the most experienced and innovative leaders and engineers in the field. Where we work Headquartered in Amsterdam and listed on Nasdaq, Nebius has a global footprint with R&D hubs across Europe, North America, and Israel. The team of over 800 employees includes more than 400 highly skilled engineers with deep expertise across hardware and software engineering, as well as an in-house AI R&D team. The role In this role, you will lead the definition, development, and delivery of Nebius Token Factory’s inference capabilities, focusing on highly scalable, production-grade machine learning systems. You will be responsible for shaping the direction of our inference platform, driving product decisions that balance performance, reliability, and real-world customer needs. This includes working closely with engineering and research teams to design and optimize real-time and batch inference workflows, supporting customer PoCs, and translating technical challenges into clear product requirements. You will work directly with customers and internal stakeholders to understand ML workflows at scale, identify bottlenecks, and define features that improve latency, throughput, orchestration, and deployment efficiency. You will also guide product adoption by delivering intuitive tools and robust infrastructure that solve complex inference problems across diverse use cases. This role requires a strong technical foundation in ML systems and a product mindset oriented toward execution, clarity, and long-term scalability. You are welcome to work remotely from the US. Your responsibilities will include: Own the product roadmap for Nebius Token Factory inference capabilities, focusing on high-load, production-grade ML scenarios. Be involved in customer PoCs involving distributed ML model deployment, inference orchestration, and optimization. Work closely with engineering and research teams to shape scalable infrastructure for real-time and batch inference. Act as the technical voice in customer conversations, translating ML workflows into product requirements. Drive product adoption by delivering tools and features that solve real-world inference problems at scale. We expect you to have: 3–5 years of product management experience, ideally in cloud infrastructure, ML platforms, or developer tools. Strong technical foundation (e.g. Computer Science or Engineering degree) with ability to dive deep into model architectures and serving systems. Familiarity with modern ML inference tools and frameworks (e.g., Triton Inference Server, vLLM, SGLang, TensorRT-LLM, Dynamo, KServe, Ray Serve). Proven track record of delivering technically complex products that support distributed and high-throughput ML pipelines. Strong communicator with experience working across engineering, research, and customer-facing teams. It will be an added bonus if you have: Deep understanding of modern ML architectures, including transformer-based models and their inference characteristics. Experience delivering or supporting ML solutions in production as part of a customer-facing or solutions role. Knowledge of MLOps or AIOps cycles, including observability, performance optimization, and continuous delivery of ML systems. What we offer Competitive salary and comprehensive benefits package. Opportunities for professional growth within Nebius. Flexible working arrangements. A dynamic and collaborative work environment that values initiative and innovation. We’re growing and expanding our products every day. If you’re up to the challenge and are excited about AI and ML as much as we are, join us!

Product Management
Machine Learning Systems
Cloud Infrastructure
ML Inference Tools (Triton, KServe)
Distributed ML Pipelines
Technical Communication
Customer-Facing Solutions
Direct Apply
Posted 1 day ago
Nebius

Senior Site Reliability Engineer — AI Studio (Inference Platform)

NebiusAnywherefull-time
View Job
Compensation$120K - 180K a year

Own the reliability, performance, and observability of the AI inference platform stack, including telemetry pipelines, autoscaling, infrastructure-as-code, and incident remediation. | Deep fluency with Kubernetes, Prometheus, Grafana, Terraform, scripting in Python or Bash, experience with distributed backend failures, and preferably GPU-heavy workload or MLOps background. | Why work at Nebius Nebius is leading a new era in cloud computing to serve the global AI economy. We create the tools and resources our customers need to solve real-world challenges and transform industries, without massive infrastructure costs or the need to build large in-house AI/ML teams. Our employees work at the cutting edge of AI cloud infrastructure alongside some of the most experienced and innovative leaders and engineers in the field. Where we work Headquartered in Amsterdam and listed on Nasdaq, Nebius has a global footprint with R&D hubs across Europe, North America, and Israel. The team of over 800 employees includes more than 400 highly skilled engineers with deep expertise across hardware and software engineering, as well as an in-house AI R&D team. AI Studio is a part of Nebius Cloud, one of the world’s largest GPU clouds, running tens of thousands of GPUs. We are building an inference platform that makes every kind of foundation model — text, vision, audio, and emerging multimodal architectures — fast, reliable, and effortless to deploy at massive scale. To deliver on that promise, we need an engineer who can make the platform behave flawlessly under extreme load and recover gracefully when the unexpected happens. In this role you will own the reliability, performance, and observability of the entire inference stack. Your day starts with designing and refining telemetry pipelines — metrics, logs, and traces that turn hundreds of terabytes of signal into clear, actionable insight. From there you might tune Kubernetes autoscalers to squeeze more efficiency out of GPUs, craft Terraform modules that bake resilience into every new cluster, or harden our request-routing and retry logic so even transient failures go unnoticed by users. When incidents do arise, you’ll rely on the automation and runbooks you helped create to detect, isolate, and remediate problems in minutes, then drive the post-mortem culture that prevents recurrence. All of this effort points toward a single goal: scaling the platform smoothly while hitting aggressive cost and reliability targets. Success in the role calls for deep fluency with Kubernetes, Prometheus, Grafana, Terraform, and the craft of infrastructure-as-code. You script comfortably in Python or Bash, understand the nuances of alert design and SLOs for high-throughput APIs, and have spent enough time in production to know how distributed back-ends fail in the real world. Experience shepherding GPU-heavy workloads — whether with vLLM, Triton, Ray, or another accelerator stack — will serve you well, as will a background in MLOps or model-hosting platforms. Above all, you care about building self-healing systems, thrive on debugging performance from kernel to application layer, and enjoy collaborating with software engineers to turn reliability into a feature users never have to think about. If the idea of safeguarding the infrastructure that powers tomorrow’s multimodal AI energizes you, we’d love to hear your story. What we offer • Competitive salary and comprehensive benefits package. • Opportunities for professional growth within Nebius. • Flexible working arrangements. • A dynamic and collaborative work environment that values initiative and innovation. We’re growing and expanding our products every day. If you’re up to the challenge and are excited about AI and ML as much as we are, join us!

Kubernetes
Prometheus
Grafana
Terraform
Python
Bash
Distributed Systems
Cloud Infrastructure
Microservices
AWS
Azure
Verified Source
Posted about 1 month ago
NE

Key Customers Solutions Architect

NebiusAnywherefull-time
View Job
Compensation$215K - 275K a year

Serve as a technical advisor for AI GPU cloud customers, optimize GPU workloads, collaborate with sales and product teams, and manage customer relationships. | 5-10+ years in cloud or AI/ML roles, proficiency with IaC tools, Kubernetes, Python, GPU computing, and strong communication and customer skills. | Why work at Nebius Nebius is leading a new era in cloud computing to serve the global AI economy. We create the tools and resources our customers need to solve real-world challenges and transform industries, without massive infrastructure costs or the need to build large in-house AI/ML teams. Our employees work at the cutting edge of AI cloud infrastructure alongside some of the most experienced and innovative leaders and engineers in the field. Where we work Headquartered in Amsterdam and listed on Nasdaq, Nebius has a global footprint with R&D hubs across Europe, North America, and Israel. The team of over 800 employees includes more than 400 highly skilled engineers with deep expertise across hardware and software engineering, as well as an in-house AI R&D team. The role Nebius seeks a Key Customers Solutions Architect to support key and strategic Nebius GPU Cloud services customers. In this role, you will be a trusted technical advisor, helping clients design, deploy, and scale AI solutions while managing large-scale GPU workloads involving hundreds to thousands of GPUs. You will also collaborate with sales and product teams to drive growth and enhance customer satisfaction. You’re welcome to work remotely from the United States or Canada. Your responsibilities will include: Serve as the primary technical point of contact, troubleshooting and resolving complex AI/ML. Guide customers in optimizing GPU performance for ML training and inference workloads, ensuring seamless integration and scalability. Partner with the sales team to identify new opportunities, promote the latest products, and deliver technical presentations. Act as a bridge to product teams, providing customer feedback, relaying feature requests, and ensuring alignment with customer requirements. Engage with internal and external stakeholders, negotiate solutions, and effectively drive alignment to address customer challenges. We expect you to have: Experience: 5 - 10 + years in roles like Cloud Solutions Architect, Technical Account Manager, or Customer Engineer, with hands-on experience in cloud services and AI/ML workloads. Proficiency in Infrastructure as Code (IaC) tools like Terraform and Ansible. Experience with Kubernetes and Python programming. Solid understanding of GPU computing, including ML training, inference workloads, and GPU stacks (e.g., CUDA, OpenCL). Customer-centric approach with a proven ability to build trust and foster long-term relationships. Strong ability to explain technical concepts to technical and non-technical audiences. It will be an added bonus if you have: Hands-on experience with HPC/ML orchestration frameworks (e.g., Slurm, Kubeflow). Experience with deep learning frameworks (e.g., PyTorch, TensorFlow). Familiarity with ML tools from NVIDIA, AWS, Azure, and Google Cloud providers. Strong project management skills, with the ability to prioritize tasks and deliver on deadlines. Proven experience mentoring technical teams and driving team growth. Expertise in stakeholder negotiation to support problem resolution and ensure seamless collaboration. Key Employee Benefits: Health Insurance: 100% company-paid medical, dental, and vision coverage for employees and families. 401(k) Plan: Up to 4% company match with immediate vesting. Parental Leave: 20 weeks paid for primary caregivers, 12 weeks for secondary caregivers. Remote Work Reimbursement: Up to $85/month for mobile and internet. Disability & Life Insurance: Company-paid short-term, long-term, and life insurance coverage. Compensation We offer competitive salaries, ranging from 215k - 275k OTE (On-Target Earnings) and equity based on your experience, skills, and location. Join Nebius Today! What we offer Competitive salary and comprehensive benefits package. Opportunities for professional growth within Nebius. Hybrid working arrangements. A dynamic and collaborative work environment that values initiative and innovation. We’re growing and expanding our products every day. If you’re up to the challenge and are excited about AI and ML as much as we are, join us!

Cloud services
AI/ML workloads
GPU computing
Infrastructure as Code (Terraform, Ansible)
Kubernetes
Python programming
Customer relationship management
Direct Apply
Posted 2 months ago
Nebius

Senior Partner Manager, Venture Capital

NebiusAnywherefull-time
View Job
Compensation$215K - 275K a year

Manage and grow venture capital partnerships, source and onboard AI startups, design partner initiatives, track KPIs, and collaborate cross-functionally to drive startup program adoption. | 7+ years in strategic partnerships or business development with 5+ years in startup/venture ecosystem, strong VC network, AI/cloud knowledge, and excellent communication skills. | Why work at Nebius Nebius is leading a new era in cloud computing to serve the global AI economy. We create the tools and resources our customers need to solve real-world challenges and transform industries, without massive infrastructure costs or the need to build large in-house AI/ML teams. Our employees work at the cutting edge of AI cloud infrastructure alongside some of the most experienced and innovative leaders and engineers in the field. Where we work Headquartered in Amsterdam and listed on Nasdaq, Nebius has a global footprint with R&D hubs across Europe, North America, and Israel. The team of over 800 employees includes more than 400 highly skilled engineers with deep expertise across hardware and software engineering, as well as an in-house AI R&D team. The role Nebius is looking for an entrepreneurial, strategic, and partnership-focused professional to join our global startup team as a Senior Partner Manager, Venture Capital. In this role, you’ll take ownership of growing and managing our U.S. startup pipeline by developing strong relationships with leading venture capital firms, accelerators, and startup communities. You’ll focus on driving adoption of our established startup program by sourcing high-potential AI startups, co-creating value-added initiatives with partners, and delivering impactful ecosystem activations. You’ll be a strategic thinker and doer—comfortable managing complex partnerships, growing the deal pipeline, and delivering exceptional experiences that deepen Nebius’ role as the cloud partner of choice for AI-native startups in the U.S. Your responsibilities will include: Develop and Manage Venture Capital Partnerships • Build and maintain strong relationships with leading U.S. venture capital firms, accelerators, and key startup ecosystem partners. • Act as a trusted Nebius representative, growing visibility and influence in the VC and startup ecosystem. Drive and Grow Startup Pipeline • Source, qualify, and onboard high-potential AI startups into Nebius’ established startup program. • Collaborate with sales and marketing teams to ensure clear tracking, strong engagement, and effective conversion into long-term Nebius users. Create and Deliver Value-Added Initiatives • Design and execute tailored partner initiatives, joint campaigns, and enablement resources. • Develop compelling GTM materials such as case studies and playbooks to support adoption and success. Measure and Optimize Partnership Success • Own and track KPIs including lead generation, startup acquisition, activation rates, and revenue contribution. • Use data and feedback to evaluate impact, identify opportunities, and optimize strategies. Collaborate Across Teams • Partner with internal stakeholders—including marketing, sales, product, and solutions architecture—to align goals and ensure seamless execution. • Champion the voice of the U.S. startup and VC community within Nebius. We expect you to have: • Minimum 7+ years of professional experience in strategic partnerships, business development, or sales & GTM, with at least 5 years working in the startup or venture ecosystem. • Deep understanding of VC and startup dynamics, investment models, and growth strategies. • Proven ability to drive qualified pipeline growth and deliver measurable business impact through partnerships. • Excellent organizational, communication, and presentation skills. • Ability to work independently and collaboratively in a fast-paced, cross-functional environment. • Familiarity with the AI and cloud computing landscape, with the ability to learn new technologies quickly. • Genuine passion for the startup ecosystem, with consistent participation in the VC or technology events and conferences. • Experience using CRM tools for pipeline management, internal communication, and workflow automation. It will be an added bonus if you have: • 5+ years of experience working in a VC firm, accelerator, as a startup founder, or as an early team member of a venture-backed AI startup. • Established, high-trust network within the U.S. venture capital ecosystem. • Experience in technology partnerships or selling AI/cloud solutions. • Strong understanding of AI infrastructure, cloud solutions, and the needs of AI/ML startups. Key employee benefits in the US: • Health insurance: 100% company-paid medical, dental, and vision coverage for employees and families. • 401(k) plan: Up to 4% company match with immediate vesting. • Parental leave: 20 weeks paid for primary caregivers, 12 weeks for secondary caregivers. • Remote work reimbursement: Up to $85/month for mobile and internet. • Disability & life insurance: Company-paid short-term, long-term and life insurance coverage. Compensation We offer competitive salaries, ranging from $215k - $275k OTE + equity based on your experience. What we offer • Competitive salary and comprehensive benefits package. • Opportunities for professional growth within Nebius. • Hybrid working arrangements. • A dynamic and collaborative work environment that values initiative and innovation. We’re growing and expanding our products every day. If you’re up to the challenge and are excited about AI and ML as much as we are, join us!

Strategic partnerships
Business development
Venture capital ecosystem knowledge
Pipeline growth
CRM tools
AI and cloud computing familiarity
Communication and presentation
Verified Source
Posted 2 months ago

Ready to join Nebius?

Create tailored applications specifically for Nebius with our AI-powered resume builder

Get Started for Free

Ready to have AI work for you in your job search?

Sign-up for free and start using JobLogr today!

Get Started »
JobLogr badgeTinyLaunch BadgeJobLogr - AI Job Search Tools to Land Your Next Job Faster than Ever | Product Hunt