InterviewStack.io LogoInterviewStack.io

Probability and Statistical Inference Questions

Covers fundamental probability theory and statistical inference from first principles to practical applications. Core probability concepts include sample spaces and events, independence, conditional probability, Bayes theorem, expected value, variance, and standard deviation. Reviews common probability distributions such as normal, binomial, Poisson, uniform, and exponential, their parameters, typical use cases, computation of probabilities, and approximation methods. Explains sampling distributions and the Central Limit Theorem and their implications for estimation and confidence intervals. Presents descriptive statistics and data summary measures including mean, median, variance, and standard deviation. Details the hypothesis testing workflow including null and alternative hypotheses, p values, statistical significance, type one and type two errors, power, effect size, and interpretation of results. Reviews commonly used tests and methods and guidance for selection and assumptions checking, including z tests, t tests, chi square tests, analysis of variance, and basic nonparametric alternatives. Emphasizes practical issues such as correlation versus causation, impact of sample size and data quality, assumptions validation, reasoning about rare events and tail risks, and communicating uncertainty. At more advanced levels expect experimental design and interpretation at scale including A B tests, sample size and power calculations, multiple testing and false discovery rate adjustment, and design choices for robust inference in real world systems.

HardTechnical
0 practiced
You train a deep generative model but some training features are missing at random. Compare approaches for handling missingness: marginalizing missing values via Monte Carlo EM, multiple imputation followed by training, and variational inference that treats missing entries as latent variables. Discuss computational trade-offs and convergence diagnostics for each.
EasyTechnical
0 practiced
Explain the Central Limit Theorem (CLT) in practical terms for an AI engineer: when you estimate the mean validation loss from mini-batches, why does CLT matter and under what conditions can you apply normal-based confidence intervals to batch-level statistics? Mention sample size and finite-variance assumptions.
MediumTechnical
0 practiced
As an AI engineer, contrast Bayesian and frequentist inference. Give an example in anomaly detection where a Bayesian approach with an informative prior helps stabilize rare-event rate estimates, and explain how to choose/assess priors in practice.
EasyTechnical
0 practiced
Explain why correlation does not imply causation. Provide an AI-related example where a feature (e.g., user location) correlates with churn but is not causal. Describe at least two strategies (experimental or observational) you could use to test for causality before using this feature in automated interventions.
EasyTechnical
0 practiced
Define the terms sample space, event, and independence in probability theory. Using a binary classification example, let the sample space be all examples, event A = model predicts positive, event B = true label is positive. Explain what independence of A and B would imply for a classifier, whether a well-performing classifier's A and B are typically independent, and how you would empirically test independence on a labeled dataset.

Unlock Full Question Bank

Get access to hundreds of Probability and Statistical Inference interview questions and detailed answers.

Sign in to Continue

Join thousands of developers preparing for their dream job.