$Not specified
You will be responsible for implementing and maintaining fine-tuning pipelines and automating data preprocessing workflows. Collaborating with cross-functional teams, you will ensure reproducible and efficient training environments.
The ideal candidate should have strong programming skills in Python and practical experience with the Hugging Face ecosystem. Familiarity with fine-tuning techniques and experience in building production-ready training scripts are also essential.
This position is posted by Jobgether on behalf of a partner company. We are currently looking for a Machine Learning Engineer in the United States. This role provides the opportunity to build and optimize training pipelines for cutting-edge AI models, focusing on practical, production-ready implementations rather than academic research. You will work on automating fine-tuning processes and enabling vendors to deploy adaptable AI models efficiently across diverse hardware. Collaborating closely with backend engineers, DevOps, and cross-functional teams, you will ensure reproducible, high-performance training workflows. This position is ideal for engineers passionate about ML engineering, system optimization, and building tools that have direct real-world impact. You will operate in a dynamic, fast-paced environment, contributing to the next generation of AI infrastructure while maintaining a strong emphasis on quality, scalability, and reproducibility. Accountabilities: In this role, you will be responsible for: Implementing and maintaining LoRA/QLoRA fine-tuning pipelines using PyTorch and Hugging Face Transformers. Developing logic for incremental training and adapter stacking, producing clean, versioned “delta packs.” Automating data preprocessing workflows, including tokenization, formatting, and filtering for user datasets. Building training scripts and workflows that integrate with orchestration backends via REST/gRPC or job queues. Implementing monitoring hooks (loss curves, checkpoints, evaluation metrics) feeding dashboards for real-time tracking. Collaborating with DevOps and backend engineers to ensure reproducible, portable, and efficient training environments. Writing tests to guarantee reproducibility, correctness, and reliability of adapter outputs. Participating occasionally in on-site meetings for discussions and collaborative problem-solving. The ideal candidate will have: Strong programming skills in Python with hands-on experience in PyTorch. Practical experience with Hugging Face ecosystem (Transformers, Datasets, PEFT). Familiarity with LoRA/QLoRA or parameter-efficient fine-tuning techniques. Understanding of mixed-precision training (FP16/BF16) and memory optimization techniques. Experience building production-ready training scripts with reproducibility, logging, and error handling. Comfortable working in Linux GPU environments (CUDA, ROCm). Ability to collaborate with engineers across disciplines, including non-ML specialists. Preferred Qualifications: Experience with bitsandbytes, xformers, or flash-attention. Familiarity with distributed training frameworks (multi-GPU, NCCL, DeepSpeed, or Accelerate). Prior work in MLOps or packaging ML pipelines for deployment. Contributions to open-source ML libraries. This role offers: Competitive compensation and equity potential. Flexible remote work with occasional on-site collaboration. The chance to work on product engineering for AI, focusing on building robust, real-world ML solutions. Exposure to cutting-edge AI infrastructure and collaborative, high-performing technical teams. Jobgether is a Talent Matching Platform that partners with companies worldwide to efficiently connect top talent with the right opportunities through AI-driven job matching. When you apply, your profile goes through our AI-powered screening process designed to identify top talent efficiently and fairly. 🔍 Our AI evaluates your CV and LinkedIn profile thoroughly, analyzing your skills, experience, and achievements. 📊 It compares your profile to the job’s core requirements and past success factors to determine your match score. 🎯 Based on this analysis, we automatically shortlist the 3 candidates with the highest match to the role. 🧠 When necessary, our human team may perform an additional manual review to ensure no strong profile is missed. The process is transparent, skills-based, and free of bias — focusing solely on your fit for the role. Once the shortlist is completed, we share it directly with the company that owns the job opening. The final decision and next steps (such as interviews or additional assessments) are then made by their internal hiring team. Thank you for your interest! #LI-CL1
This job posting was last updated on 10/4/2025