Insignia's looking for a Machine Learning Engineer who has gone beyond prompt-tuning and fine-tuning, someone who's architected, deployed, and optimized RAG-based systems and LLM pipelines in real-world applications.
Our Office at Meruya, West Jakarta
What You'll Do:
- Design and deployment of production-grade RAG and LLM systems
- Optimize retrieval accuracy, context quality, and model performance at scale
- Build robust data pipelines for ingestion, chunking, embedding, and indexing
- Collaborate with data engineers, software teams, and product leads to integrate AI into core features
- Set best practices for prompt engineering, evaluation, and monitoring
Who You Are:
3+ years in ML engineering, with experience on LLMs and RAG systems
Hands-on experience with LangChain, LlamaIndex, vector databases (Pinecone, Weaviate, FAISS)
Strong in Python, PyTorch/TensorFlow, and MLOps tools (MLflow, Kubernetes, Docker)
Experience deploying models on AWS, GCP, or Azure or hybrid environments
Familiar with evaluation frameworks, A/B testing, and latency optimization
Bonus: Experience with fine-tuning, LoRA, or distillation for domain-specific performance
Practical mindset you care about reliability, cost, and impact, not just novelty
Why Join Us?
Opportunity to work on impactful AI/ML projects.
Collaborative and innovative team environment.
Career growth, health insurance, meals provided, etc.