Time Series Interview Questions

Dan Lee's profile image
Dan LeeData & AI Lead
Last updateMarch 13, 2026

Time series questions appear in nearly every data scientist interview at companies like Netflix, Uber, and Meta because real business problems involve temporal data. Whether you're forecasting demand, detecting payment anomalies, or building recommendation systems, you need to handle sequential patterns, seasonality, and the cardinal rule of never learning from the future.

What makes time series interviews brutal is that subtle mistakes create plausible wrong answers. You might confidently explain how to build a 7-day rolling average feature, but if you don't account for late-arriving data or timezone normalization, your model will leak future information and fail in production. Interviewers specifically design questions around these edge cases because they separate candidates who have shipped real systems from those who only know textbook examples.

Here are the top 25 time series questions organized by the skills that matter most: data preparation fundamentals, stationarity concepts, forecast evaluation, anomaly detection, and financial markets basics.

Advanced25 questions

Time Series Interview Questions

Top Time Series interview questions covering the key areas tested at leading tech companies. Practice with real questions and detailed solutions.

Data ScientistQuantitative ResearcherMetaGoogleNetflixUberTwo SigmaCitadelGoldman SachsAmazon

Time Series Basics: Indexing, Resampling, Lags, and Splits

Interviewers start with indexing, resampling, and lag features because most candidates fail on data preparation basics. You'll get a messy dataset with mixed timezones, missing values, and late-arriving updates, then be asked to create features for a forecasting model.

The key insight is that temporal data leakage is silent and deadly. Your training metrics look great, but production performance crashes because your rolling averages accidentally included future information or your timestamp alignment was off by one period. Practice building features with strict temporal ordering under realistic data quality constraints.

Time Series Basics: Indexing, Resampling, Lags, and Splits

Start by showing you can handle timestamps correctly: sorting, time zones, missing intervals, lag features, and time based train validation splits. Candidates struggle because small leakage or alignment mistakes silently break model evaluation in interviews.

You have event logs with columns user_id, event_ts, and value, where event_ts is a string with mixed time zones (some end in Z, others have -0700). You need hourly aggregates per user and a 24 hour lag feature for modeling, what exact timestamp normalization and ordering steps do you apply to avoid silent leakage or misalignment?

UberUberMediumTime Series Basics: Indexing, Resampling, Lags, and Splits

Sample Answer

Most candidates default to parsing timestamps into naive datetimes and calling resample, but that fails here because the same wall clock hour means different instants across time zones. You should parse as timezone aware, convert everything to a single canonical zone like UTC, then sort by (user_id, event_ts) and de duplicate if your source can emit repeats. Aggregate on the normalized timestamp, then build the 24 hour lag using a time aware shift, for example groupby user and shift by 24 rows only if you have a complete hourly index, otherwise use a time based merge or asof join on $t-24\text{h}$. Finally, verify alignment by checking that the lagged feature timestamp is strictly earlier than the target timestamp for every row.

Practice more Time Series Basics: Indexing, Resampling, Lags, and Splits questions

Stationarity, Decomposition, and Transformations

Stationarity questions test whether you can diagnose time series properties and choose appropriate transformations. Netflix and Uber love asking about revenue or ride demand series that have trends, seasonality, and heteroskedasticity all mixed together.

Most candidates memorize that you difference to remove trends and log-transform to stabilize variance, but they can't decide the right order of operations or validate their choices. The secret is using ACF patterns and residual diagnostics to confirm your transformations actually worked, not just applying cookbook recipes.

Stationarity, Decomposition, and Transformations

In interviews, you are tested on whether you can diagnose trend, seasonality, and changing variance and decide what to difference, detrend, or transform. Many candidates can recite definitions but cannot justify choices using ACF, PACF, and residual checks.

You are forecasting daily rides for a city. The series has an upward trend and a strong weekly pattern, the ACF decays slowly and has big spikes at lags 7, 14, 21. What transformations would you apply before fitting an ARIMA-type model, and how would you validate that you did enough?

UberUberMediumStationarity, Decomposition, and Transformations

Sample Answer

Use seasonal differencing at lag 7 and then difference once more if trend remains, and only add a variance-stabilizing transform if residual variance grows with level. The slow ACF decay suggests a unit root or trend, and the repeated spikes at multiples of 7 suggest weekly seasonality, so $\nabla_7 x_t = x_t - x_{t-7}$ is a first move, followed by $\nabla x_t$ if needed. You validate by checking that the residual ACF has no significant structure and that the mean and variance look stable over time. If PACF and ACF of the differenced series cut off or decay quickly, and residuals look like white noise, you are done.

Practice more Stationarity, Decomposition, and Transformations questions

Classical Forecasting: ARIMA, SARIMA, and Exponential Smoothing

Forecast evaluation separates candidates who have deployed models from those who haven't. You'll design backtesting frameworks, choose appropriate metrics, and handle the practical constraints of production forecasting systems.

The trap here is evaluation methodologies that look rigorous but contain subtle biases. Many candidates suggest standard train/test splits or use MAPE without thinking about zero values and outliers. Real systems need rolling origin backtests that respect information availability and metrics that align with business impact.

Classical Forecasting: ARIMA, SARIMA, and Exponential Smoothing

You will often be asked to pick and tune a classical model under time pressure, then defend your assumptions and backtesting setup. Candidates stumble when translating business context into orders, seasonal terms, and diagnostics like Ljung Box and prediction intervals.

Forecast Evaluation, Backtesting, and Production Constraints

Anomaly detection questions focus on threshold setting, false positive control, and distinguishing real signals from noise. Companies like Citadel and Two Sigma ask these because trading and payment systems need reliable alerting with minimal false alarms.

Candidates typically propose simple statistical thresholds without considering seasonality, autocorrelation, or multiple testing problems. The winning approach involves seasonal decomposition, dynamic thresholds that adapt to volatility, and statistical corrections when monitoring hundreds of series simultaneously.

Forecast Evaluation, Backtesting, and Production Constraints

Expect questions that probe how you evaluate forecasts without leaking future information, including rolling origin backtests and metrics like MAPE, sMAPE, and pinball loss. It is easy to give a metric name, harder to explain when it fails and how you would monitor drift in production.

You are forecasting daily rides for 200 cities and you have 2 years of history. How would you design a rolling origin backtest so every forecast only uses information available at that time, and how would you aggregate results across cities?

UberUberMediumForecast Evaluation, Backtesting, and Production Constraints

Sample Answer

Reason through it: First, you pick an initial training window, for example first 12 months, and a forecast horizon $h$, for example 7 days. Then you define cutoffs $t_1, t_2, \dots$ and for each cutoff you refit using data up to $t_i$, forecast $\hat{y}_{t_i+1:t_i+h}$, and score against realized $y_{t_i+1:t_i+h}$, never touching data beyond $t_i$ during fitting or feature construction. You roll forward by a fixed step, often 1 day or 1 week, and keep the training policy fixed, expanding window or fixed length, based on how you will train in production. For aggregation, you report both macro averages (each city equal weight) and volume weighted averages (weight by actual demand) because the former reflects robustness and the latter reflects business impact.

Practice more Forecast Evaluation, Backtesting, and Production Constraints questions

Anomaly Detection and Change Point Analysis

Financial time series questions appear at quant funds and fintech companies, testing your understanding of returns, volatility, and market microstructure effects. These roles demand precision because small modeling errors become large trading losses.

The classic mistake is building strategies that work in backtests but fail live due to survivorship bias, transaction costs, or lookahead bias in feature construction. Interviewers probe whether you understand bid-ask spreads, market impact, and the difference between research environments and production trading systems.

Anomaly Detection and Change Point Analysis

When you design anomaly detection, you must balance false positives, delayed detection, and seasonality, then explain thresholds and alerting logic. Interviews reveal gaps when candidates ignore multiple testing, autocorrelation, or how to label and evaluate anomalies realistically.

You own alerting for hourly payment failures, with strong day of week seasonality and autocorrelation. How do you set an anomaly threshold that controls false positives while still catching a 5 to 10 percent sustained increase within 2 hours?

UberUberHardAnomaly Detection and Change Point Analysis

Sample Answer

This question is checking whether you can separate signal from predictable structure, then tune detection under latency constraints. You should first model expected failures with seasonality and correlation, for example a regression or state space model with day of week effects, then monitor residuals $r_t = y_t - \hat{y}_t$ rather than raw $y_t$. Set thresholds on standardized residuals using an estimate of one step ahead uncertainty, and add a persistence rule like 2 consecutive breaches to reduce noise driven alerts. Calibrate the cutoff to a target false alarm rate using backtests on clean periods, and report the tradeoff curve between detection delay and false positives.

Practice more Anomaly Detection and Change Point Analysis questions

Financial Time Series and Market Microstructure Basics

Financial Time Series and Market Microstructure Basics

For quant roles, interviews check whether you understand returns, volatility clustering, and non IID behavior, plus how backtests can lie due to lookahead and survivorship bias. Candidates struggle most with articulating a robust research workflow, from feature design to risk controls and validation.

You are modeling daily returns for a single stock to forecast next-day volatility, do you use simple returns or log returns, and how does that choice affect aggregation and your interpretation of extreme moves?

Goldman SachsGoldman SachsEasyFinancial Time Series and Market Microstructure Basics

Sample Answer

The standard move is to use log returns $r_t = \log(P_t/P_{t-1})$ because they add nicely over time, so multi-day return is approximately $\sum r_t$. But here, the exception is large moves and low prices, because log returns and simple returns diverge when $|R_t|$ is big, and risk metrics like drawdowns are naturally in simple return space. If you care about compounding exactly, simple returns multiply, log returns add, choose based on what you report and how you aggregate. For microstructure-heavy data, your return definition also interacts with bid-ask bounce, so consistent price conventions matter as much as the formula.

Practice more Financial Time Series and Market Microstructure Basics questions

How to Prepare for Time Series Interviews

Build Features With Real Messiness

Practice creating lag and rolling features on datasets with missing timestamps, late-arriving data, and timezone issues. Use pandas with explicit UTC conversion and forward-fill strategies that respect information availability at prediction time.

Master ACF Pattern Recognition

Train yourself to diagnose stationarity issues from autocorrelation plots. Plot the ACF of raw data, first differences, and log differences until you can instantly recognize trend persistence, seasonal patterns, and white noise signatures.

Design Backtests That Mirror Production

Always propose rolling origin validation with realistic data cutoffs that simulate how your model would perform with only historical information. Account for training time, feature computation delays, and the business constraints of when predictions are actually needed.

Question Every Evaluation Metric

For each forecasting scenario, think through why MAPE, RMSE, or accuracy might give misleading results. Practice explaining metric failure modes to non-technical stakeholders and proposing alternatives that align with business costs.

Simulate Multiple Testing Problems

When designing anomaly detection for multiple time series, always address how you'll control family-wise error rates. Practice calculating Bonferroni corrections and explaining why naive per-series thresholds will flood you with false positives.

How Ready Are You for Time Series Interviews?

1 / 6
Time Series Basics: Indexing, Resampling, Lags, and Splits

You inherit a dataset of user signups with timestamps in local time across multiple time zones. The model will use daily features and predict next-day signups. What is the best preprocessing approach to avoid leakage and time alignment bugs?

Frequently Asked Questions

How much time series depth do I need for interviews in Data Scientist or Quantitative Researcher roles?

You should be comfortable with stationarity, autocorrelation, differencing, seasonality, and basic forecasting baselines like ARIMA and exponential smoothing. You also need to explain validation for time dependent data, including walk-forward backtesting and why random splits leak information. For quant roles, expect deeper questions on stochastic processes, state space models, and how assumptions break in real markets.

Which companies tend to ask the most time series interview questions?

You will see the most time series depth at hedge funds, proprietary trading firms, market makers, and quant research teams at banks. Time series also shows up heavily at ad tech, forecasting, and operations heavy companies, like logistics, marketplaces, and subscription businesses. If the team owns forecasting, anomaly detection, or experimentation with sequential data, you should expect time series questions.

Do I need to code for time series interviews, and what kinds of coding tasks appear?

Yes, you often need to code, especially to build features like lags and rolling statistics, run walk-forward evaluation, or implement a simple forecasting baseline. You may also be asked to diagnose leakage or compute metrics over horizons and groups. Practice these patterns with realistic prompts at datainterview.com/coding.

How do time series interviews differ between Data Scientist and Quantitative Researcher roles?

As a Data Scientist, you will usually focus on practical forecasting workflows, feature engineering, monitoring, and model selection under business constraints. As a Quantitative Researcher, you will be pushed harder on statistical assumptions, dependence structure, microstructure effects, and rigorous backtesting that avoids look-ahead bias. You should be able to discuss both predictive performance and whether your inference is valid under autocorrelation.

How can I prepare for time series interviews if I have no real-world time series experience?

You can build a small portfolio project that uses public data, like energy demand, retail sales, or macro indicators, and demonstrate walk-forward backtests with clear baselines. Make sure you show how you handle seasonality, missing timestamps, and exogenous variables, and include a simple error analysis by horizon. Use datainterview.com/questions to practice explaining your approach and tradeoffs clearly.

What are the most common mistakes candidates make in time series interviews?

You often lose points by using random train test splits, leaking future information through scaling or feature windows, or tuning on the test set instead of a rolling validation. Another common mistake is ignoring seasonality and calendar effects, or evaluating with a single aggregate metric that hides horizon specific errors. You should also avoid claiming causality from correlated time series without addressing confounding and autocorrelation.

Dan Lee's profile image

Written by

Dan Lee

Data & AI Lead

Dan is a seasoned data scientist and ML coach with 10+ years of experience at Google, PayPal, and startups. He has helped candidates land top-paying roles and offers personalized guidance to accelerate your data career.

Connect on LinkedIn