Jobs.ca
Jobs.ca
Language
Lattice logo

Senior AI Software Engineer

Lattice9 days ago
Remote
Canada
CA$90,849 - CA$121,132/yearly
Senior Level

About the role

Who you are

  • 5+ years of professional software engineering experience with significant time spent on production AI/ML systems
  • Deep hands-on experience with LLM-based systems: prompt engineering, RAG pipelines, agent orchestration, evaluation metrics, and model fine-tuning
  • Proven ability to work with data and understand statistics, especially in experiments
  • Proven ability to build and operate agentic AI systems in production: multi-step workflows, multi-agent topologies, and the failure modes that come with them
  • Strong command of AI evaluation: you’ve built eval frameworks before, you know the difference between a good eval and a vanity metric, and you have opinions about it
  • Production-grade Python engineering: clean, maintainable, testable code
  • LangGraph or comparable agent orchestration frameworks. You’ve built real agent workflows with it, not just tutorials
  • LangSmith or comparable LLM observability tooling for tracing, evaluation, and debugging
  • Reads AI papers & blogs regularly and is a trusted source of AI trends
  • Vector databases (Pinecone or similar) and retrieval system design
  • AWS ecosystem or other cloud infrastructure (ex GCP). Comfortable with lambdas, queues, and cloud-native architecture
  • Familiarity with TypeScript is a plus. Our full-stack engineers use it and cross-pollination is valuable
  • Clear eyes: you see problems as they are, not as you’d like them to be. You surface hard truths early and address them directly
  • Ship, shipmate, self: you prioritize the product and your teammates. Low ego, high ownership
  • You’re as comfortable in ambiguity as you are in well-defined problems: early foundations mean you’ll encounter both
  • Strong technical communication: you can debate evaluation methodology with an AI lead and explain it clearly to an EM in the same afternoon
  • Experience with RLHF, LoRA, or other model adaptation techniques
  • Background in traditional ML (supervised/unsupervised, neural networks) and knowing when an LLM is overkill
  • Experience with MLOps tooling: MLflow, DataDog, CI/CD pipelines for model deployment
  • Published work, conference talks, or open-source contributions in AI/ML
  • Experience in HR tech, people analytics, or other domains where data quality and trust are critical

What the job involves

  • Our AI Engineering team is building the systems that power how AI works across Lattice. We’ve laid the foundations: traces are flowing and evals are running - and we’re now focused on defining how our AI products are measured, improved, and trusted at scale. This is a high-ownership role where you’ll help shape evaluation methodology, agent architecture, and the core systems that determine how AI performs in production
  • Design and ship a robust, end-to-end AI evaluation framework, covering offline evals, production tracing, and human-in-the-loop feedback loops, connected across all of Lattice’s AI use cases
  • Define and instrument the metrics that actually matter: agent task completion, hallucination rates, response quality, user engagement, and downstream business outcomes
  • Build and maintain evaluation datasets, test harnesses, and automated scoring pipelines to catch regressions before they ship
  • Identify and surface the drivers of agent quality improvement, giving the team clear signals on where to invest
  • Architect and implement reusable agent infrastructure: multi-turn conversation workflows, recommendation services, LLM DAGs, and standardized agent topology patterns using LangGraph
  • Build and scale RAG pipelines and retrieval infrastructure, including vector store management and retrieval quality optimization
  • Make principled build vs. buy decisions across LLM providers, agent frameworks, and evaluation tooling, balancing capability, cost, latency, and vendor risk
  • Contribute to production AI systems with a strong focus on reliability, observability, and performance, not just prototypes
  • Own projects end-to-end: scope them, drive them to completion, and bring in the right people at the right time
  • Partner with engineering leads and managers to inform technical direction on agent quality and evaluation strategy you’ll be expected to hold intelligent, substantive conversations about methodology, not just implementation
  • Raise the AI engineering bar across the broader team through code review, documentation, and thoughtful technical debate

About Lattice

Software Development
501-1000

The People Platform to manage people and their performance. Because when people thrive, business thrives. #becausepeople

We’re on a mission to make work meaningful. If you are passionate about the intersection where great cultures meet high performance, please join us. We’re hiring for a number of positions! Learn about joining our team on our Careers tab or by emailing us at hello@lattice.com.

Similar jobs you might like