Episode 23 — Compare time series and survival analysis goals without mixing assumptions

In this episode, we’re going to clear up a confusion that shows up a lot when people first learn predictive modeling with time involved, which is the temptation to treat all time-related problems as the same kind of problem. Time series forecasting and survival analysis both involve time, and both can use similar words like hazard, risk, and prediction, but they are trying to answer different questions and they rely on different assumptions. If you blend them together carelessly, you can end up building a model that sounds sophisticated but is answering the wrong question. Beginners often notice that one approach predicts a value over time while the other predicts something about an event, yet they may not realize how deep that difference goes. The goal here is to help you separate the goals, understand what each method considers an outcome, and recognize the kinds of data structures that fit each approach. Once you can do that, you will know when you are forecasting a changing signal and when you are modeling the timing of a specific event.

Before we continue, a quick note: this audio course is a companion to our course companion books. The first book is about the exam and provides detailed information on how to pass it best. The second book is a Kindle-only eBook that contains 1,000 flashcards that can be used on your mobile device or Kindle. Check them both out at Cyber Author dot me, in the Bare Metal Study Guides Series.

A time series problem is usually about predicting the future values of a measurement that continues over time, like daily energy usage, hourly traffic, or weekly sales. The outcome is a value at a time point, and that value exists whether or not something dramatic happens, because it is part of an ongoing process. You can often imagine the data as a single line that never stops, where each time step produces a new observation, and the goal is to extend that line forward. The structure usually emphasizes ordering and dependency, meaning that past values help predict future values, and the series can have trends, seasonality, and lags. In many time series settings, you can evaluate a forecast by comparing predicted values to actual values across a horizon, such as the next day, next week, or next month. The key idea is that the process keeps generating values, and you are predicting the next part of the sequence.

Survival analysis is different because the main outcome is not a continuous stream of values, but the time until an event happens. The event might be a machine failure, a customer canceling a subscription, a patient experiencing a health outcome, or a loan going into default. What makes this distinct is that the event is typically treated as a one-time transition, meaning once it happens, you are no longer observing the same kind of future outcome for that subject. In survival analysis, each subject has a timeline that starts at some origin point, and you care about how long it takes until the event, if it happens at all during your observation window. That makes the data feel more like many separate timelines rather than one long line, because you have multiple subjects, each with its own duration. The goal is often to estimate how risk changes over time and to predict the distribution of event times, not to forecast a numeric value at each time point.

A major concept that makes survival analysis special is censoring, and understanding censoring is one of the fastest ways to see why survival analysis is not just time series with a different label. Censoring happens when you do not observe the event for some subjects during the time you were watching them, so you only know that their event time is greater than a certain duration. For example, if you track customers for six months and some customers do not cancel during that period, you cannot say their time-to-cancel is six months, because it might be much longer. You only know they survived at least six months without the event. Time series forecasting typically does not treat missing future values that way, because you are not modeling time-to-event for each subject, you are modeling the next values of a continuing process. When you ignore censoring and treat those cases as if the event time equals the observation time, you introduce a bias that can seriously distort the model’s understanding of risk.

Another concept that distinguishes survival analysis is the survival function, which describes the probability that the event has not happened by a given time. Closely related is the hazard function, which describes the instantaneous rate of event occurrence at a moment in time, given that the event has not occurred yet. You can think of hazard as a way of describing how risk is changing as time passes, and it does not require you to forecast a continuous measurement at every time step. In time series, you might care about the expected value at the next step, and you might model variability around that expectation, but you are not typically trying to describe the probability of a one-time event having occurred by each time. That difference matters because it changes what the model outputs mean and what kinds of evaluation are appropriate. A forecast error metric for time series does not naturally capture the quality of predicted event timing, especially when censoring is involved.

It also helps to notice that time series forecasting often assumes that the process can be described in terms of patterns in the sequence itself, while survival analysis often assumes that the event timing is influenced by features of the subject and possibly how those features change over time. In time series, your primary input might be the past values of the same series, plus maybe external variables like weather or promotions. In survival analysis, your primary input is often covariates, meaning subject-level features such as age, contract type, device model, or usage behavior, and the outcome is the duration until the event. Some survival settings include time-varying covariates, which means the features can change as time passes, but even then the target is still time-to-event, not the next value in a continuous series. This is why saying both are time-based is true but incomplete, because the unit of analysis and the target structure are different. If you confuse the unit of analysis, you can easily build a model that answers a question you did not mean to ask.

A common mixing mistake happens when someone takes a time series of counts, like daily incident tickets, and tries to interpret spikes as events in the survival sense without defining a clear subject and a one-time event. A daily ticket count is an ongoing measurement that can go up and down, and there is no single moment where it permanently transitions into an event state. Survival analysis needs a definition like “time until the first critical incident after deployment” for each system, or “time until a customer churns” for each customer, because those are one-time events per subject. Another mixing mistake is when someone has churn data and tries to forecast a churn indicator as a time series without accounting for censoring and without recognizing that once churn happens, the subject stops producing observations in the same way. You can still model churn using other approaches, but if you pretend it is just a regular sequence of values, you risk training on impossible information and evaluating with misleading metrics. These mistakes are not about being bad at math; they are about not being precise about what time means in the problem.

You can also contrast how each approach handles the idea of prediction horizons and what it means to be wrong. In time series, you might predict the next 30 days of values, and error is the difference between predicted and actual values at each day. Being off by 10 units today is a concrete numeric error, and being off tomorrow is another numeric error, so you can average those errors and summarize accuracy. In survival analysis, being wrong might mean predicting that an event would occur earlier than it did, or later than it did, or predicting high risk for someone who never experiences the event during the observed window. Because censoring means you do not always know the true event time, you cannot score predictions the same way you score time series forecasts, and that is why survival analysis uses specialized evaluation ideas like concordance, which focuses on relative ordering of risks rather than exact event times for everyone. The evaluation choices reflect the fact that the target is fundamentally different.

Another assumption difference shows up in how you think about independence and repeated measurements. Time series points are strongly dependent because each time step follows from the previous one, and you expect that dependency, which is why lag features matter. Survival analysis often treats each subject as a separate observation with its own timeline, and while subjects can be correlated in real life, the modeling setup often assumes that subject timelines are independent given the covariates. That assumption is not always perfect, but it is part of the standard framework. If you mistakenly treat each day of a subject’s life as independent rows without respecting that they come from the same person, you can distort survival modeling. Conversely, if you treat different customers as if they are time steps in a single sequence, you can distort time series reasoning. Temporal thinking is not just about having timestamps; it is about choosing the right structure for what the data represents.

It’s also important to compare what each approach implies about what information is allowed to influence the prediction. In time series forecasting, it is common to use past values and known future inputs, like a schedule of planned promotions, but you must not use future values of the target itself. In survival analysis, you can use baseline covariates known at the start, and sometimes you can use time-varying covariates up to the current time, but you cannot use information that would only be known after the event occurs. If you accidentally include a feature that is recorded only after a failure, like a repair code, you create leakage that is especially dangerous in time-to-event settings because it can appear to predict the event with near-perfect accuracy. The more you separate the target definition from the available information timeline, the safer your model becomes. This is one reason survival analysis education emphasizes careful definition of time origin, event definition, and censoring rules.

A helpful way to decide between these approaches is to ask what the business or operational question is really asking, because the question often implies the correct framework. If the question is “What will the value be next week?” you are in time series territory because you want a future value of an ongoing measurement. If the question is “How long until it happens?” you are in survival territory because you want a time-to-event distribution. If the question is “Will it happen within the next 30 days?” you might use survival concepts to compute that probability, or you might frame it as a classification problem with a fixed window, but you should be aware of what you lose when you fix the window and ignore censoring beyond it. The more you can state the question precisely, the easier it is to choose the correct assumptions. The danger comes when you try to force one framework to answer the other framework’s question without adjusting the data structure and evaluation accordingly.

There is also an important mindset difference: time series often focuses on modeling the process itself, while survival analysis often focuses on modeling the risk of transition from one state to another. Time series is comfortable with the idea that the process continues forever, producing values that fluctuate, and forecasting extends the process forward. Survival analysis is comfortable with the idea that each subject has a life course that may end in an event, and the key uncertainty is when that end point happens. That means survival analysis naturally supports questions about expected remaining time, changing risk over tenure, and the impact of covariates on risk. Time series naturally supports questions about expected future levels, seasonal planning, and detecting anomalies as deviations from expected trajectories. Both can be used in decision-making, but they answer different types of decisions, and that difference shows up in how you define targets and features.

By the end of this topic, the goal is that you can look at a dataset and immediately ask whether you have one long evolving signal or many subjects with possible event times. If you have repeated measurements over time and you care about forecasting future values, you are likely in time series territory, and you should respect ordering, seasonality, and lag. If you have an event definition and you care about how long until that event, you are likely in survival territory, and you should respect censoring, time origin, and risk interpretation. Mixing assumptions usually happens when someone uses the language of one framework while building the data structure of the other, and the result is confusion about what the model output means. Keeping the goals separate does not limit you; it actually gives you more freedom because you can pick the right tool for the question and evaluate it honestly. When you can articulate these differences clearly, you are thinking like someone who understands time in data rather than someone who is just hoping the model will figure it out.

Episode 23 — Compare time series and survival analysis goals without mixing assumptions
Broadcast by