InterviewStack.io LogoInterviewStack.io

A and B Test Design Questions

Designing and running A and B tests and split tests to evaluate product and feature changes. Candidates should be able to form clear null and alternative hypotheses, select appropriate primary metrics and guardrail metrics that reflect both product goals and user safety, choose randomization and assignment strategies, and calculate sample size and test duration using power analysis and minimum detectable effect reasoning. They should understand applied statistical analysis concepts including p values confidence intervals one tailed and two tailed tests sequential monitoring and stopping rules and corrections for multiple comparisons. Practical abilities include diagnosing inconclusive or noisy experiments detecting and mitigating common biases such as peeking selection bias novelty effects seasonality instrumentation errors and network interference and deciding when experiments are appropriate versus alternative evaluation methods. Senior candidates should reason about trade offs between speed and statistical rigor plan safe rollouts and ramping define rollback plans and communicate uncertainty and business implications to technical and non technical stakeholders. For developer facing products candidates should also consider constraints such as small populations cross team effects ethical concerns and special instrumentation needs.

HardTechnical
50 practiced
Baseline conversion has been drifting upward by about 0.5% per week due to seasonality during your planned experiment. How do you adjust sample size and duration calculations to maintain power and avoid bias in estimated treatment effect? Describe both design and analysis strategies.
HardTechnical
48 practiced
You need to detect whether an observed uplift is driven by a novelty effect that fades over time. Design an experiment and analysis strategy to distinguish novelty from persistent treatment effects. Include time windows, decay modeling, and decision rules.
EasyTechnical
58 practiced
You are evaluating a new recommendation algorithm expected to increase click-through-rate (CTR). Formulate clear null and alternative hypotheses for an A/B test comparing the new algorithm (treatment) to the current algorithm (control). Specify the metric, unit of analysis (user/session), directionality, and what statistical evidence would lead you to reject the null in favor of the alternative.
HardTechnical
44 practiced
Developer-facing product: a change affects SDK behavior used by internal teams (small user base, cross-team impacts). Propose instrumentation, experiment design, and communication plan to run safe, informative tests while minimizing disruption to dependent teams.
HardTechnical
54 practiced
You have 5 treatment arms and multiple business-critical metrics. Describe a sequential rollout and decision policy that balances exploration speed, statistical rigor, and business risk. Include how you'd handle early winners/losers and what thresholds trigger promotion of an arm to production.

Unlock Full Question Bank

Get access to hundreds of A and B Test Design interview questions and detailed answers.

Sign in to Continue

Join thousands of developers preparing for their dream job.