About the job LLM Applications Engineer
LLM Applications Engineer
Location: New York | San Francisco | Munich | London (In-Person)
Employment Type: Full-Time
Base Salary: $130,000 – $175,000
Overview
We are hiring an LLM Applications Engineer to build and deploy production-grade LLM-powered systems.
This is a hybrid AI infrastructure and product engineering role. You will design and implement RAG pipelines, vector retrieval systems, agentic workflows, and full-stack LLM-powered product experiences. The role requires hands-on ownership across backend Python systems and React-based frontend applications.
This position is for engineers who move beyond prototypes and ship robust, scalable LLM systems into production.
What Youll Do
-
Design and deploy production-grade RAG (Retrieval-Augmented Generation) pipelines
-
Build and optimize vector retrieval systems
-
Implement agentic LLM workflows and orchestration layers
-
Develop full-stack product experiences powered by LLMs
-
Design clean, scalable APIs and asynchronous processing systems
-
Connect LLM systems to structured data sources, including SQL databases and data engines
-
Collaborate closely with product and engineering teams to ship LLM-first features
What Were Looking For
Core Requirements
-
2+ years of full-stack web development experience
-
Proficiency in Python and JavaScript or TypeScript
-
Experience with LLM orchestration frameworks such as LangChain, LlamaIndex, or Haystack
-
Hands-on experience building production RAG pipelines and retrieval systems
-
Strong API design and asynchronous processing fundamentals
-
Ability to operate across backend infrastructure and frontend UI
-
Computer Science degree from a top-tier program
Strong Plus
-
Built and deployed RAG pipelines in live production environments
-
Experience working with scientific, technical, or research datasets
-
Strong product mindset with LLM-first feature development
-
Experience integrating LLM systems with SQL databases and broader data infrastructure
Who This Is Not For
This role is not a fit if you:
-
Have only academic or prototype-level LLM exposure
-
Have exclusively backend-only or frontend-only experience
-
Lack experience with vector databases or retrieval systems
-
Have not deployed LLM systems into production environments
What Success Looks Like
-
Production-grade RAG pipelines running reliably at scale
-
Clean, well-architected retrieval and orchestration systems
-
Seamless integration between LLM backends and frontend product experiences
-
Measurable product impact from LLM-powered features
-
Ownership of end-to-end LLM application architecture
Why This Role Is Unique
This is not a research-only AI role and not a traditional full-stack position.
You will sit at the intersection of AI infrastructure and product, building systems that combine retrieval, orchestration, and real-world user interfaces. If you want to ship meaningful LLM-powered products — not just experiment with models — this is an opportunity to own that architecture end to end.