AI Engineer – AI Observability and Quality
Quick Summary
Overview About the Role & TeamIDC is building the next generation of AI-powered intelligence platforms that transform how technology decisions get made. Our platform re-imagines the way decision-makers discover and interact with trusted research and data, and is foundational to IDC's future.
6+ years of software engineering experience, with significant work in testing infrastructure, ML evaluation, or quality systems Experience building evaluation or testing frameworks for LLM-based or ML-based systems Understanding of how to measure…
About the Role
~1 min read- Design and build the evaluation infrastructure that ensures the platform's AI systems produce accurate, well-sourced, high-quality responses
- Build automated test suites that validate answer quality across agent pipeline changes
- Develop regression detection systems that catch quality degradation before it reaches users
- Create evaluation frameworks that measure response accuracy, citation correctness, and source quality
- Work closely with the product team to translate quality criteria into measurable, automatable test scenarios
- Build cost and latency monitoring that tracks the operational efficiency of AI pipeline execution
- Define evaluation standards and practices that scale as the platform and team grow
- 6+ years of software engineering experience, with significant work in testing infrastructure, ML evaluation, or quality systems
- Experience building evaluation or testing frameworks for LLM-based or ML-based systems
- Understanding of how to measure response quality for generative AI: accuracy, groundedness, citation correctness, relevance
- Proficiency in Python
- Ability to operate independently and define your own roadmap. You will be the first hire in this function
- Experience working at the intersection of engineering and product, translating qualitative quality criteria into quantitative measurements
- Experience with LLM evaluation frameworks (e.g., RAGAS, DeepEval, or custom)
- Familiarity with LLM observability tools (e.g., Langfuse, LangSmith, Weights & Biases)
- Background in statistical methods for quality measurement (significance testing, distribution analysis)
- Experience building A/B testing or experimentation infrastructure
- Background in search relevance evaluation or information retrieval metrics
What We Offer
~1 min readWhat We Offer
~1 min readAt IDC, we are committed to fair and equitable pay practices. Employees are compensated equitably for their work, aligned with their skills and experience. Salary and incentive structures are determined through a rigorous process that considers experience, education, certifications, role-specific requirements, internal equity, and verified Canada market data from an independent third-party partner.The expected total annual compensation, depending on location and experience, is between 100,000 CAD - 143,000 CAD, inclusive of base salary and variable compensation.
IDC is committed to providing equal employment opportunities for all qualified persons. Employment eligibility verification required. We participate in E-Verify.
#LI-JF1#LI-Remote
Location & Eligibility
Listing Details
- Posted
- March 30, 2026
- First seen
- May 6, 2026
- Last seen
- May 13, 2026
Posting Health
- Days active
- 0
- Repost count
- 0
- Trust Level
- 11%
- Scored at
- May 6, 2026
Signal breakdown
Please let idccareers-canada-idg know you found this job on Jobera.
3 other jobs at idccareers-canada-idg
View all →Explore open roles at idccareers-canada-idg.
Stay ahead of the market
Get the latest job openings, salary trends, and hiring insights delivered to your inbox every week.
No spam. Unsubscribe at any time.