Senior LLM Engineer

Maitai
Full-time$100k-225k/year (USD)Redwood City, United States
Apply Now

High-level

Join Maitai to reshape how enterprise companies build with open-source LLMs. You’ll be at the forefront, driving cutting-edge innovations in model fine-tuning, distillation, and automation to continuously enhance LLM performance. You’ll collaborate directly with founders, engineers, and enterprise customers, building the core management layer that defines enterprise AI infrastructure. We're scaling rapidly and looking for engineers who deeply understand open-source LLM ecosystems and can confidently automate and optimize model improvements at scale.

Low-level

You will lead the fine-tuning, distillation, and deployment of open-source LLMs tailored for enterprise customers. Your role involves:

  • Preparing, optimizing, and managing large-scale datasets for model training and continuous improvement.

  • Developing and automating sophisticated fine-tuning pipelines to enhance model accuracy, reliability, and inference speed.

  • Distilling models to smaller, faster, and more efficient variants without compromising performance.

  • Implementing new platform features in Python and Go to facilitate seamless dataset curation, correction, and augmentation.

Who You Are

  • You've spent quite a bit of time in Unsloth notebooks and know your way around fine-tuning opensource models like llama, gemma, mistral, etc.

  • You recognize dataset balancing as an art as much as a science.

  • You subscribe to r/LocalLLaMA.

  • You've got a local model humming on your MacBook Pro to power Cursor.

Why Join Us?

  • Massive technical challenges – Pioneering automated, continuous improvements for enterprise-grade opensource LLMs.

  • Ownership and Impact – Drive architecture decisions, shape core product offerings, and influence company strategy from day one.

  • Elite, collaborative team – Join a fast-moving environment working alongside top-tier engineers.

  • Equity Upside – Early-stage, meaningful equity ownership.

  • Zero Red Tape – Ship fast, iterate faster, and enjoy working without heavy processes or Jira epics.

More You Need To Know

  • In-person role in downtown Redwood City, CA. Caltrain or parking pass, lunches, and Starbucks/Philz coffee provided.

  • Engineers own their product decisions. Engage directly with customers, set your own specs, and deliver meaningful features.

About Us

Maitai ensures LLMs never fail by optimizing reliability, speed, and resilience. Acting as an intelligent proxy, we apply real-time autocorrections, route requests intelligently, and fine-tune models for maximum performance. We're experiencing explosive growth, are well-capitalized, and seizing a massive opportunity to redefine how enterprises build with AI. Our platform delivers AI models that significantly outperform closed-source alternatives in speed and accuracy, supported by robust online guardrails. Leading YC startups and public enterprises trust Maitai to manage their LLM infrastructure.

About the interview

  1. Quick Chat (15-minute Video Call)

  2. Let’s discuss your experience, interests, and ambitions.

  3. Tech Discussion

  4. Get on a call and talk tech. What's going on in the industry, what have you worked with recently, latest model you've fine-tuned, last meetup you were at, etc.

  5. Hands-On Technical

  6. Join us at our office to work through a problem with our team.

  7. In-person Meetup

About Maitai

Maitai manages the LLM stack for enterprise companies, enabling the fastest and most reliable inference. The future of enterprise AI revolves around mosaics of small, domain-specific models powering powerful, responsive agents, and Maitai is well positioned to capture the market. If you're looking at getting in early with a company redefining how large companies build with AI, then let's talk.

Application Requirements

experience fine-tuning open-source LLMs (e.g., LLaMA, Gemma, Mistral), comfort with Unsloth or similar tools, dataset preparation and balancing skills, ability to develop automated fine-tuning pipelines, knowledge of model distillation techniques, strong Python and Go programming skills, experience deploying models at scale, familiarity with LLM infrastructure, interest in open-source model ecosystems, hands-on experience with local LLM setups, ability to work in-person in Redwood City, CA, willingness to engage directly with customers and set specs, strong interest in AI infrastructure for enterprise use.