Job Description
Perplexity serves tens of millions of users daily with reliable, high-quality answers grounded in an LLM-first search engine and our specialized data sources. We aim to use the latest models as they are released, but the intelligence frontier is a jagged one, and popular benchmarks do not effectively cover our use cases. In this role, you will build specialized evals to improve answer quality across Perplexity, covering search-based LLM answers and other scenarios popular with our users.
Responsibilities
Architect and maintain automated evaluation pipelines to assess answer quality across Perplexity's products, ensuring high standards for accuracy and helpfulness
Design evaluation sets and methods specifically to measure the impact of tool calls (particularly web search retrieval) on the final answer's quality
Develop VLM-based solutions to programmatically evaluate how final answers render visually across different platforms and devices
Continuously review public benchmarks and academic evaluations for their applicability to the Perplexity product, adapting and incorporating them into our regular performance measurements
Operate within a small, high-impact team where your evaluation metrics directly shape product changes, collaborating closely with technical leadership to measure and improve Answer Quality
Qualifications
PhD or MS in a technical field or equivalent experience
4+ years of experience in data science or machine learning
Strong proficiency in Python and SQL (expected to write production-grade code)
Experience building within a modern cloud data stack, specifically AWS and Databricks
Comfortable with agentic coding workflows and using AI-assisted development tools to iterate faster
Preferred Qualifications
1+ years of experience working with LLMs at scale, specifically with LLM-as-a-judge setups
Prior experience working on customer-facing web products or consumer apps, with real user traffic at scale
A strong research background, with experience applying research methods to real-world ML problems
Experience defining evaluation metrics (e.g., factual consistency, hallucination rate, retrieval precision) and building ground truth datasets
Perplexity
6 jobs posted
About the job
Feb 13, 2026
Mar 15, 2026
Similar Jobs
Visa
24 days agoData Scientist
Warsaw, PolandView detailsEpic Games
30 days agoSenior Data Scientist (FamXP)
Cary, North CarolinaView detailsRobinhood
30 days agoData Scientist
Ljubljana, SIView detailsHP
30 days agoData Scientist
Chennai, Tamil Nadu, IndiaView detailsPaypal
26 days agoData Scientist
Bangalore, Karnataka, IndiaView detailsStripe
25 days agoData Scientist
N/AView detailsNasdaq
26 days agoData Scientist
United StatesView detailsGrab
24 days agoSenior Data Scientist (Analytics)
Singapore, SingaporeView detailsAirbnb
24 days agoSenior Data Scientist, Pricing
Remote$177K - $208K/yrView detailsAnthropic
24 days agoData Scientist
New York City, NYSan Francisco, CASeattle, WA$275K - $370K/yrView details
Looking for something different?
Browse all AI jobs