Summary
Join our team as an AI Systems Engineer and help us push the boundaries of AI in engineering. You'll be part of a small, agile team where your contributions will have an immediate impact. This role involves building and improving AI pipelines, experimenting with LLM tuning, enhancing AI-generated outputs, and owning features from prototype to production. You'll work with various technologies, iterate quickly, and research new tools. We're looking for someone AI-savvy with production Python coding experience and a strong understanding of LLMs and NLP.
Requirements
- You’re AI-savvy, you’ve worked with LLMs and understand how they function under the hood
- You have a builder mindset. You’ve shipped real Python code to production in a team environment and are comfortable with backend frameworks (FastAPI, Flask, etc.)
- You know how to engineer LLM prompts, validate outputs, and iterate quickly to get high quality results
- You have experience using Natural Language Processing (NLP) techniques to extract, transform, and parse textual data into meaningful representations suitable for downstream LLM-based operations
- You’re used to experimenting and prototyping in Notebooks
- You have strong DevOps fundamentals, and experience with CI/CD & cloud services (AWS, GCP, Azure)
- You have experience with monitoring tooling and troubleshooting production issues
- You’re self-directed, adaptable, and love wearing multiple hats—R&D one day, demoing & debugging your latest pipelines with customers the next
- You can communicate complex technical concepts to non-technical folks (we may ask you to explain how an LLM works under-the-hood)
- You care about how your work impacts users and drives business value
- You’re always testing new tools or reading up on the latest AI trends
Responsibilities
- Build and improve dynamic AI pipelines by designing, writing & deploying production Python code
- Experiment with various LLM tuning strategies to answer complex qualitative questions (e.g. can AI help identify which tasks are blockers or risks)
- Boost the quality of AI-generated outputs—whether it’s improving summaries, surfacing insights, or generating new categories from scratch
- Own end-to-end features: from scrappy prototype to stable, production-ready deployment
- Configure, maintain and deploy distributed application services to cloud environments
- Get your hands dirty across backend, infrastructure, and AI/ML workflows
- Iterate fast: tweak prompts, tune models, test outputs, and constantly improve
- Research new tools, techniques and frameworks to keep us ahead of the curve
Preferred Qualifications
- Experience with TensorFlow, PyTorch, or deploying open-source LLMs (Llama, Mistral, etc.) on your own infra
- Knowledge of graph databases or vector databases
- Hands-on with serverless (AWS Lambda) or cloud-native tooling (Kubernetes, Docker)
- An academic or practical background in ML and/or Natural Language Processing (NLP) or computer science
- Ideally based in Vancouver (but we’re open to remote across the Americas)
Benefits
- Flexible remote work + unlimited vacation (we actually take it)
- Annual learning & development budget (conferences, books, courses)
- Health & wellness perks
- Top-tier gear—whatever you need to do your best work
- A no-ego, collaborative team that’s serious about building something great
Disclaimer: Please check that the job is real before you apply. Applying might take you to another website that we don't own. Please be aware that any actions taken during the application process are solely your responsibility, and we bear no responsibility for any outcomes.