Research Engineer II
Microsoft
Copilot Tuning is a new product that enables organizations to customize large language models (LLMs) using tenant data, unlocking task-specific agents tailored to real enterprise workflows. We are a team advancing how models are adapted, evaluated, and deployed within Microsoft 365 Copilot, bridging cutting-edge research with production systems to transform the LLM experience in the enterprise.
We are looking for candidates for a Research Engineer II role who are passionate about translating research ideas into scalable, reliable ML systems. This role is ideal for candidates with strong engineering fundamentals and experience working with machine learning systems, along with an interest in prototyping, experimentation, and system-building. The candidate will focus on turning research prototypes into production-ready capabilities, building data pipelines, developing and iterating on model tuning workflows, and operationalizing LLM-based systems that meet real-world performance, safety, and quality requirements.
The candidate will work closely with applied scientists and engineers to bring research innovations into production, rapidly iterating on ideas, validating them with data, and integrating them into robust services. This includes enabling task-specific agents that reflect organizational knowledge, improving consistency and efficiency, and ensuring solutions generalize across diverse enterprise scenarios while maintaining strong security and compliance guarantees.We provide a nurturing environment for engineers excited about working at the intersection of research and production. We are looking for candidates who excel in problem-solving, experimentation, and system design, and who can navigate ambiguity while driving ideas from concept to shipped experience in a fast-paced, collaborative setting.
Microsoft's mission is to empower every person and every organization on the planet to achieve more. As employees, we come together with a growth mindset, innovate to empower others, and collaborate to realize our shared goals. Each day we build on our values of respect, integrity, and accountability to create a culture of inclusion where everyone can thrive at work and beyond.
Responsibilities
- Design and build inferencing‑time orchestration services that dynamically adapt Copilot behavior based on tenant context, user intent, and enterprise policies.
- Develop runtime systems that integrate LLM inferencing with Microsoft 365 Substrate signals, APIs, and permission models for grounded Copilot responses.
- Implement inferencing pipelines that support tool selection, prompt composition, and policy enforcement during Copilot execution.
- Enable tenant‑safe Copilot execution by incorporating identity, compliance boundaries, and contextual grounding into inferencing workflows.
- Build telemetry and feedback loops to evaluate inferencing performance across latency, relevance, grounding quality, and safety metrics.
- Partner with platform teams to support real‑time routing, experimentation, and configuration of Copilot inferencing paths
Qualifications
Required Qualifications:
- Bachelor's Degree in Computer Science or related technical field AND 2+ years technical engineering experience with coding in languages including, but not limited to, C, C++, C#, Java, JavaScript, or Python
- OR equivalent experience.
Preferred Qualifications:
- Master's Degree in Computer Science or related technical field AND 3+ years technical engineering experience with coding in languages including, but not limited to, C, C++, C#, Java, JavaScript, or Python
- OR Bachelor's Degree in Computer Science or related technical field AND 5+ years technical engineering experience with coding in languages including, but not limited to, C, C++, C#, Java, JavaScript, or Python
- OR equivalent experience.
- Experience developing and scaling machine learning systems from research prototypes to production, including experimentation, evaluation, UI, and iteration on models and pipelines.
- Experience collaborating with research scientists to implement, validate, and productize systems, ensuring robustness, reproducibility, and measurable impact.
- Experience building and supporting ML-powered services, including model deployment, monitoring, evaluation pipelines, and maintaining reliability and performance in production environments.
- Experience with cloud platforms (e.g., Azure, AWS, or GCP) and modern engineering practices (e.g., CI/CD, testing, code reviews) applied to ML systems and experimentation workflows.
- Familiarity with Generative AI concepts such as large language models (LLMs), prompt engineering, fine-tuning, or Retrieval-Augmented Generation (RAG), and experience adapting these techniques into real-world applications.
Software Engineering IC3 - The typical base pay range for this role across the U.S. is USD $100,600 - $199,000 per year. There is a different range applicable to specific work locations, within the San Francisco Bay area and New York City metropolitan area, and the base pay range for this role in those locations is USD $131,400 - $215,400 per year.
Certain roles may be eligible for benefits and other compensation. Find additional benefits and pay information here:
https://careers.microsoft.com/us/en/us-corporate-pay
This position will be open for a minimum of 5 days, with applications accepted on an ongoing basis until the position is filled.
Microsoft is an equal opportunity employer. All qualified applicants will receive consideration for employment without regard to age, ancestry, citizenship, color, family or medical care leave, gender identity or expression, genetic information, immigration status, marital status, medical condition, national origin, physical or mental disability, political affiliation, protected veteran or military status, race, ethnicity, religion, sex (including pregnancy), sexual orientation, or any other characteristic protected by applicable local laws, regulations and ordinances. If you need assistance with religious accommodations and/or a reasonable accommodation due to a disability during the application process, read more about requesting accommodations.