InterviewStack.io LogoInterviewStack.io

Stream Processing and Event Streaming Questions

Designing and operating systems that ingest, process, and serve continuous event streams with low latency and high throughput. Core areas include architecture patterns for stream native and event driven systems, trade offs between batch and streaming models, and event sourcing concepts. Candidates should demonstrate knowledge of messaging and ingestion layers, message brokers and commit log systems, partitioning and consumer group patterns, partition key selection, ordering guarantees, retention and compaction strategies, and deduplication techniques. Processing concerns include stream processing engines, state stores, stateful processing, checkpointing and fault recovery, processing guarantees such as at least once and exactly once semantics, idempotence, and time semantics including event time versus processing time, watermarks, windowing strategies, late and out of order event handling, and stream to stream and stream to table joins and aggregations over windows. Performance and operational topics cover partitioning and scaling strategies, backpressure and flow control, latency versus throughput trade offs, resource isolation, monitoring and alerting, testing strategies for streaming pipelines, schema evolution and compatibility, idempotent sinks, persistent storage choices for state and checkpoints, and operational metrics such as stream lag. Familiarity with concrete technologies and frameworks is expected when discussing designs and trade offs, for example Apache Kafka, Kafka Streams, Apache Flink, Spark Structured Streaming, Amazon Kinesis, and common serialization formats such as Avro, Protocol Buffers, and JSON.

HardSystem Design
42 practiced
Design state backend and checkpointing strategy for a streaming job that must hold session state for 200M users at ~1KB per user (~200GB). Compare using RocksDB local state with incremental checkpoints to S3 vs using an external DB (Cassandra/Scylla) for state. Discuss recovery time, checkpoint size, IOPS, compaction, and operational implications.
HardTechnical
37 practiced
Design a schema governance process for a company with hundreds of producers writing Avro/Protobuf to Kafka. Include steps for schema registration, automated compatibility checks in CI, approval workflows, emergency rollbacks, and preventing producers from bypassing the schema registry.
EasyTechnical
37 practiced
Compare Avro, Protocol Buffers (Protobuf), and JSON as serialization formats for event streams. Explain trade-offs in size, parsing speed, schema evolution capabilities, and runtime compatibility. When and why would you use a schema registry?
MediumTechnical
43 practiced
Your analytics team requests reprocessing six months of raw events after a bug fix in enrichment logic. Design an operational plan and pipeline changes to support efficient and safe reprocessing: how you would replay events, avoid duplicate results in analytics store, handle schema changes, estimate cost, and monitor progress.
MediumTechnical
39 practiced
For windowed aggregations, discuss how window size, checkpoint interval, batching, and parallelism influence latency and throughput. Provide concrete knobs you would adjust to optimize for low latency vs for high throughput and explain trade-offs for state size and recoverability.

Unlock Full Question Bank

Get access to hundreds of Stream Processing and Event Streaming interview questions and detailed answers.

Sign in to Continue

Join thousands of developers preparing for their dream job.