InterviewStack.io LogoInterviewStack.io

A and B Test Design Questions

Designing and running A and B tests and split tests to evaluate product and feature changes. Candidates should be able to form clear null and alternative hypotheses, select appropriate primary metrics and guardrail metrics that reflect both product goals and user safety, choose randomization and assignment strategies, and calculate sample size and test duration using power analysis and minimum detectable effect reasoning. They should understand applied statistical analysis concepts including p values confidence intervals one tailed and two tailed tests sequential monitoring and stopping rules and corrections for multiple comparisons. Practical abilities include diagnosing inconclusive or noisy experiments detecting and mitigating common biases such as peeking selection bias novelty effects seasonality instrumentation errors and network interference and deciding when experiments are appropriate versus alternative evaluation methods. Senior candidates should reason about trade offs between speed and statistical rigor plan safe rollouts and ramping define rollback plans and communicate uncertainty and business implications to technical and non technical stakeholders. For developer facing products candidates should also consider constraints such as small populations cross team effects ethical concerns and special instrumentation needs.

MediumTechnical
0 practiced
Given: observed absolute lift = 0.8 percentage points (i.e., conversion increases from 5.0% to 5.8%) with 95% CI [0.2, 1.4] percentage points, and average revenue per conversion = $50. Estimate incremental revenue per 100,000 users and explain the uncertainty to stakeholders. Show calculations and caveats.
HardTechnical
0 practiced
Design an experiment results pipeline that supports real-time monitoring, historical aggregations, and reproducible post-hoc analysis. Describe the data sources, ingestion and transformation steps, storage models (e.g., event logs and aggregated tables), key tables/views for analysts, and how to enable rerunning analyses for auditing and reproducibility.
EasyTechnical
0 practiced
Before analyzing experiment outcomes, describe how you would verify that randomization worked. Include at least two statistical checks, practical thresholds for concern, and how to interpret failures of those checks.
HardTechnical
0 practiced
An experiment personalizes offers using sensitive attributes (e.g., protected class proxies). Describe engineering and analysis safeguards to protect privacy and prevent discriminatory treatment. Include instrumentation approaches, logging and access controls, subgroup fairness checks, and when to withhold or limit rollout based on fairness signals.
MediumTechnical
0 practiced
You plan to check experiment results daily and may stop early if the treatment looks winning. Explain the statistical risks of 'peeking' and describe two formal approaches that allow interim analyses without inflating Type I error (name and briefly describe each). When would you prefer an alpha-spending approach over a Bayesian monitoring approach in a product environment?

Unlock Full Question Bank

Get access to hundreds of A and B Test Design interview questions and detailed answers.

Sign in to Continue

Join thousands of developers preparing for their dream job.