Job Overview :
Master Works is looking for a highly skilled LLM Engineer to join our AI Core Delivery team within the AI & Analytics CoE. In this role, you will develop, deploy, and optimize advanced language model capabilities, including Retrieval-Augmented Generation (RAG) pipelines and agentic AI systems. You will play a key role in scaling AI Core as the central enterprise AI platform by ensuring production-ready performance, robust observability, and seamless integration of LLMs with internal and external systems.
Key Responsibilities
- Model Development & Optimization
- Build and optimize RAG pipelines for efficient document processing, indexing, and retrieval.
- Implement agentic AI logic for tool integration, task automation, and external data access.
- Enhance query understanding, response ranking, and AI-generated output quality.
- Develop evaluation benchmarks to measure LLM performance, accuracy, and relevance.
- LLMOps & Deployment
- Collaborate with API developers and backend engineers to deliver secure, observable LLM endpoints.
- Deploy and manage LLMs in cloud-native or hybrid environments with support for scalability and multi-cloud readiness.
- Implement CI / CD workflows for model updates, rollback mechanisms, and performance tuning.
- Backend & Data Integration
- Integrate LLM capabilities with enterprise data sources and knowledge bases.
- Support multi-source retrieval and embedding strategies to improve AI responses.
- Design pipelines that ensure low latency, high throughput, and fault tolerance.
- Security & Compliance
- Apply enterprise-grade authentication, access management, and encryption standards.
- Ensure compliance with internal governance policies, data privacy, and audit requirements.
- Continuous Improvement
- Monitor system performance, debug issues, and fine-tune models based on real-world feedback.
- Stay updated on advancements in LLM architectures, open-source tools, and evaluation frameworks.
Requirements
Bachelor’s or Master’s degree in Computer Science, AI, Data Science, or related field.Hands-on experience with LLM development and deployment (OpenAI, Anthropic, Hugging Face, LLaMA, etc.).Strong understanding of Retrieval-Augmented Generation (RAG) , embeddings, and vector databases.Experience with cloud environments (AWS, Azure, GCP) and infrastructure-as-code (IaC).Proficiency with Python or similar languages for AI / ML pipeline development.Familiarity with container orchestration (Docker, Kubernetes) and CI / CD practices.Knowledge of observability tools for monitoring AI performance in production environments.Preferred Skills
Experience integrating agentic AI capabilities with external APIs and automation tools.Strong grasp of prompt engineering, contextual memory systems, and LLM evaluation .Exposure to enterprise AI security and compliance standards .