Applied AI Engineer - Agent Intelligence & Retrieval
Added 6 hours agoRole Overview
We're seeking an Applied AI Engineer to join our fast-paced team to design, develop, and implement new features for Littlebird, our AI teammate for Mac and Android (Windows soon). We're building a personal AI that connects your entire digital life, protecting your focus from clutter and surfacing what you need, the moment you need it. Check out the recent TechCrunch news on our $11M seed funding and what users are saying about the product.
We're a small, async-first team that values craft and ownership. Our engineers live at the intersection of genuine research curiosity and production discipline: you'll push the frontier in areas like conversational inference, ranking, and sentiment analysis, then wrestle those ideas into systems that are fast, lean, and built to scale.
The Role
This role is about making our agent smarter, faster, and more reliable. You will live in the core of our AI, obsessing over the quality of our retrieval, the precision of our ranking, and the logic of our agent.
Some of the hard problems you'll solve:
Master the Art of Retrieval & Re-ranking: Our current hybrid search pipeline is functional but will need to scale with customer growth. You will own its evolution.
- Solve the "Broad Query" Problem: How do you make retrieval work just as well for "what did I do last week?" as it does for a specific, targeted question? This involves query analysis, decomposition, and potentially multiple retrieval strategies.
- Optimize the Ranking Stack: You'll experiment with and productionize new re-ranking models to crush our latency bottlenecks. You'll fine-tune our ranking strategy to better blend sparse and dense retrieval signals.
- Develop Intelligent Pruning: How do you shrink the context passed to the LLM by 80% without losing the critical 1% of information that leads to the right answer? You'll design and test sophisticated context pruning and summarization techniques.
Engineer Better Agentic Reasoning: Our agent uses a multi-step, tool-calling approach to solve problems. Debugging and improving it is a core challenge.
- Context Engineering: You'll become an expert in "prompt-level" performance, figuring out the optimal way to structure and present context to the LLM to minimize hallucinations and improve reasoning.
- Debugging Complex Agentic Flows: You'll be a detective, tracing the root cause of agent failures through layers of tool calls, context retrieval, and LLM responses to understand where things went wrong and how to fix them.
What we're looking for:
- Strong OOP programming skills and prior experience writing production software in python, Typescript, or C++
- A deep, intuitive understanding of information retrieval and modern RAG pipelines. You've likely built a few from scratch.
- Hands-on experience with vector databases, hybrid search, and re-ranking models.
- An experimental, data-driven mindset. You're comfortable running A/B tests, analyzing metrics, and iterating quickly to improve performance.
- A pragmatic approach. You're focused on shipping tangible improvements to the AI's quality, not just chasing SOTA benchmarks.
Benefits
- Remote-friendly work environment
- Collaborative team culture
- Opportunity to shape infrastructure decisions
- Competitive compensation packages including stock and health benefits, paid time off, and parental leave. 401k options for all US-based employees.
- Flexible working hours across multiple time zones