Code & Cure

#7 - Predicting No-Shows: The Surprising Science Behind Missed Appointments

Vasanth Sarathy & Laura Hagopian

Why do so many doctor’s appointments end in empty waiting rooms? Nearly one in four scheduled visits turn into no-shows, disrupting care, wasting resources, and straining already overburdened systems. But a new study shows we might be able to see these gaps coming—and stop them.

By analyzing over a million healthcare visits, researchers used machine learning to uncover surprising predictors of missed appointments. The top signal? How far in advance the appointment was booked. Appointments scheduled more than 60 days out had the highest odds of being missed—more telling than age, income, or insurance status. Other key factors included continuity with the same provider, a patient’s past attendance, distance to the clinic, and even the weather.

This episode unpacks how models like random forests and gradient boosting sift through massive datasets to identify no-show risks—not just for populations, but for individual patients. These insights open the door to smarter, more personalized interventions: tighter scheduling windows, transportation support, or ensuring patients see familiar faces.

Tune in to explore how AI could help healthcare systems run smoother, deliver more timely care, and keep more patients from vanishing in the first place.

References:

Predicting Missed Appointments in Primary Care: A Personalized Machine Learning Approach
Wen-Jan Tuan, Yifang Yan, Bilal Abou Al Ardat, Todd Felix and Qiushi Chen
Annals of Family Medicine, July/August 2025 

Credits: 

Theme music: Nowhere Land, Kevin MacLeod (incompetech.com)
Licensed under Creative Commons: By Attribution 4.0
https://creativecommons.org/licenses/by/4.0/

Speaker 1:

Can we actually predict who might miss their doctor's appointment? Turns out yes, and these might not be the reasons you're expecting.

Speaker 2:

Hello and welcome back. My name is Vasant Sarathy and I am here today with Laura Hagopian. Hey Laura, how's it going today?

Speaker 1:

I'm doing well. How are you?

Speaker 2:

I'm good, I'm good. What are we talking about? What's on the docket for today?

Speaker 1:

Well, we are talking about missed appointments and late cancellations and whether or not those could be predicted, Because if we know, you know who is going to no-show or why people might not show up to their appointments, we can potentially do something about it.

Speaker 2:

Right, right, right, and that happens frequently and it's probably a really important issue to tackle too, because it probably affects their health.

Speaker 1:

Yeah, I mean in studies. The range of appointments missed in primary care practices is anywhere in the eight to 24% range. Wow that's like that could be almost a quarter of patients not not coming to their appointment or canceling within 24 hours.

Speaker 1:

Yeah, yeah, which is like very interesting to hear because I feel like it's so hard to get a doctor's appointment. So I'm I'm curious to dig into this and understand better. You know what is happening that people aren't showing up or that they're canceling with very little notice for their appointments. Yeah, and we found at.

Speaker 1:

What I think was interesting in this study was not only do they look at the stuff that would be normally available in your electronic health record things like demographic data you know their sex, their age, their race, their language but they also looked at things like their healthcare utilization how long it had been since they had a scheduled their appointment, who their clinician was, how long it had been since they had a scheduled their appointment, who their clinician was, how long their clinician had been practicing. But they also looked at geosocial and environmental context. So they looked at things like oh, how far is this person from their doctor's office? And they looked at their area deprivation index and they looked at things like geez, was it raining that day? What was the temperature? Was there any snowfall? And so when you think about something like this holistically, this makes a lot of sense, right? It's not just the information about the patient, but about all the things that are surrounding them that could affect whether or not they show up for an appointment.

Speaker 2:

Yeah, and this sort of, to me at least, reminds me of the Florence Nightingale discussions we had before Folks can refer to, I think podcast. It was number three, but we talked about how other factors are really relevant and potentially could be more important than what you think of. Is the reason why people cancel these appointments.

Speaker 1:

Like might people cancel because it's a really hot or a really cold day? The answer is the answer is yes actually. So it's, it's interesting. And they used, you know, machine learning to try to figure out what the relationships might be or what are the things that could actually predict who's going to not show up for their appointment. So I'd love to hear more from you about the machine learning models they used. And I read this fancy term called gradient boost and I want to know what it means.

Speaker 2:

Yeah, I think this paper was great because folks who are new to machine learning or folks who are picking up machine learning to try to learn how to do it, this is actually kind of a nice paper because it takes you all the way from data to models to types of things to evaluate, all the way through, and kind of shows you how this sort of analysis should be done, and so I really enjoyed reading this paper. And it's not like it's using the most fanciest, latest tool. It's not like a chat GPT-LLM type of machine learning here. This is the kind of stuff that you know folks in universities are learning when they take their intro to machine learning class. Now, of course, the versions that they're learning in the intro to machine learning class are a little bit more simple than what you know is more sophisticated here in this paper. But it's those basic ideas and I think it's worthwhile taking a step back. We've talked in this in our podcast here about AI right generally, but before language models and before all of the big neural net revolutions, there was more basic machine learning systems, and machine learning at its very core went from AI that used to just be computer programs that humans wrote to take some data in and produce some useful insight out and those programs were authored by hand to machine learning models in the 90s that took a bunch of data and learned patterns in those data and were able to do something useful with them.

Speaker 2:

And in this particular paper, we're focusing more on predictive analytics and the idea here is that you have all of these different pieces of data right. You just mentioned all these different factors, everything from the person's information to healthcare utilization, to geo, you know kind of other kinds of information, geo-environmental information. All of that is relevant information potentially. But you also know what you're trying to do is do some prediction with that and your prediction could be. In this particular case it was one of three things it could be either the person is a no-show. The person is, you know, going to cancel the appointment late under 24 hours I think is what they picked or is going to complete the appointment late under 24 hours I think is what they picked or is going to complete the appointment and show up right.

Speaker 2:

So this sort of task in the machine learning world at some level is called a classification task. You're taking a bunch of information and you're deciding whether it fits into one of these three categories no show, completed late. And a similar task would be like identifying whether an email is spam or not spam. It's two classes spam or not spam and the email has all this text and other information that you're providing it. So that's roughly the broad subtask that this paper is working on, and if you think about the data that's coming in, I think it's worthwhile. Whenever you think about machine learning models, it's worthwhile to always think about data first, then the model it ends up being. Actually, you know, 80 to 90% of the work in building a good machine learning model actually just isn't working with data and that's really messy. It's you're getting your hands dirty working with really.

Speaker 1:

So dirty, yeah, dirty in the that's funny. Dirty really. So. So dirty, yeah, dirty in the that's funny dirty in the compute computational sense. I mean, you're talking to someone who you know was hands deep in bodily fluids all the time.

Speaker 2:

So when you talk about getting dirty with data yeah, well, um, but so dirty in that sense, in the sense that it's not nice clean programs, it's. You're working and trying to clean up all the data to make it make it sort of palatable for the, for the machine learning models. So let's talk a second about data. So you have all the stuff about person's gender, maybe where they live, and all of that stuff, and there's different types of it. So there's types of data that's categorical, that has specific answers to a question, that has specific categories for each entry. There could be data that is just numerical you could have the temperature of the day and that's just a number and there's other kinds of, you know, particular kinds of data itself that goes into each of these fields.

Speaker 2:

So if you think of the data as a giant spreadsheet, each column is a different factor or feature, and we talked about different categories of factors before, but each one of those is a separate column and there's a separate column right at the end of this spreadsheet that is assigned for the prediction, which could be no show, late or completed, right. So that's three categories, that's one column, that's the final answer. But we have all these other columns of data and every row you can think of as a different patient. And so you have all this information now in these different rows. You have thousands upon hundreds of thousands of rows.

Speaker 1:

I don't actually yeah, there were over a million visits and they used most of them to train the model, but they think they tested the model on about 20%.

Speaker 2:

Yeah, and that is typically what people do is just to hold out about 20% of the data and then they train the model on the remaining so that the holdout data isn't so the model can be really tested properly and you're not just like testing on stuff that the model is already kind of memorized and so that 80% of the data is still a large amount of data that they have. But this data is very heterogeneous. Like I said before, there's different types of data, but they're all just big giant table right, and so there's that. Then there's a separate issue of quality of data. There might be missing elements here and there. Maybe some information is dropped out. That is like hard engineering work to try to get all of that right, because the models are only going to do as well as the data is. So if the data is messy and has all these problems with them, the model isn't going to perform as well either.

Speaker 1:

Yeah, that makes sense.

Speaker 2:

And so if you want to think about the data we talked about, all these different factors there looks like in this paper at least, they use a couple dozen, maybe more factors, which is also called features, which I'll use interchangeably here. But but the machine learning models are also dependent. The performance of these machine learning models are also dependent on how many of those there are. So there's all these factors, so let's get into the models themselves. So broadly, the very simplest way to think about machine learning models is something called linear regression, and linear regression is….

Speaker 1:

I remember that from math class. There you go.

Speaker 2:

But even if you haven't taken any math class, it's kind of easy to visualize.

Speaker 2:

Because if you had a chart, imagine a chart in front of you with a whole bunch of dots on it If you had a chart that had a bunch of dots that roughly look like a straight line, then you can draw a line right through those dots and say, okay, this line kind of represents my dots right.

Speaker 2:

And the tighter they are and closer they are to the line, the better the line fits those dots, the easier it is for me to draw the line Right. But if you had it scattered everywhere, then you're not going to be able to draw a line as easily and it's more complicated. But maybe it's not a line, maybe those dots are kind of a weird S shape, and then with your pencil you could draw an S shape covering most of the dots and then you know that you're pretty close to the right answer. So that's machine. Charting is essentially a big piece of it is doing that kind of work. It's for the machine to automatically figure out what the slope, what the angle of that line is and where that line should live in that chart. And instead of you having to figure it out. It does it by just looking at the data and using some math.

Speaker 1:

But there's all these factors here. It's not just like one factor that you're charting right, yeah, or they could interact with each other.

Speaker 2:

So now this is where the notion of dimensions come in. And when I asked you to imagine a chart, that's just two dimensions, right, you have the vertical part of the chart and the horizontal part of the chart, and your dots, that kind of, are represented in that middle of that piece of paper, but you have multiple dozen factors, so it's a multiple dozen dimensional chart. You can't imagine that in your head, nobody can.

Speaker 1:

I don't know what you're talking about.

Speaker 2:

Yeah, it's a wrinkle in time, right, but the thing is, you don't have to imagine that in your head. That's why you have the math right, the math works out. But that's essentially what they're trying to do is fit a line or whatever higher dimensional version of that is on those dots, now what they. So linear regression just gives you a line. But here what we want is we want it to say hey, this is a no-show, it is a cancellation, or it's a late, late cancellation, or, if it's a completion, right, we don't. We don't want a line, we want three, one of three answers we want the output we want to figure out what's going to happen before it actually does.

Speaker 2:

We want to predict the future, but we don't want to predict. It's not like predicting the cost of a house, where we have another number that we can put on that line. It's not just a number, it's one of these three outcomes, right? So what we want is something that tells us one of these three answers, and that's slightly different. It requires a little different math, and that's one of the models that they use is called logistic regression, which is the same idea as linear regression, except that you also then have to decide, based on the line, whether or not it's one of these three categories. The easiest way to look at it would be the line might tell you where on the chart it is, but you might have a threshold. You might say you might cut it off and say above this number it's a yes and below this number it's a no, or above this number it's a no-show, between these numbers it's a cancellation. So you might have to decide what the threshold is that makes it work.

Speaker 2:

Now the machine learning work for logistic regression is around that this paper also had other machine learning models. I also want to just take a step back and throw more jargon at you.

Speaker 1:

Oh, thanks.

Speaker 2:

Yeah, that's what I was looking for, but a umbrella term for all of this stuff here that we're talking about is called supervised learning, and supervised just exactly means that you give it the right answer, you tell it. Here's the thing that I want you to take in, and here's the type of output I want you to give me and I want you to take in, and here's the type of output I want you to give me, and I give you a whole bunch of right answers. You know problems and right answers, data, and so that's called supervised learning.

Speaker 1:

Like I'm the supervisor or the data is the supervisor, the data is the supervisor in a sense. Yeah.

Speaker 2:

And the machine updates how much to bend that line based on as it processes each piece of data, it updates that information and it tries to reduce the error in how it predicts the line and the way. So that's logistic regression. There's other models here and they do it in different ways. There's this whole set of models called decision trees, which is like asking 20 questions. So you, instead of going through and drawing a line, what you try to do is look at each of the factors and you decide on a threshold and you say if it's above, if the person is above 36, maybe they will be a no-show.

Speaker 1:

And then you imagine building out a whole 20 question, except for each of this, for each of these things, like what was their prior, like their prior no-show rate, uh is high, so they're likely to go down this tree and now I know if they had a late cancellation before then I'm going to keep going down the they're likely to no shower or cancel or whatever it is.

Speaker 2:

Right, and if they're above a certain age and they live near Paris and they do this, you can imagine going down that tree in that fashion and that's called a decision tree. It's very powerful model actually. It works really well for many cases and it allows interpretation because after the fact, if you get an answer, you can sort of trace the path that it took to come at that answer in a sense, and understand like what it took into account as it came up with the answer.

Speaker 1:

So it's not necessarily causal, but it's like explainable.

Speaker 2:

Like, we understand why this happened.

Speaker 1:

It's not some sort of black box out there. We can figure out why we think this person is going to be a no-show, for example.

Speaker 2:

And just to be clear, even logistic regression is very explainable in the sense that for every factor, you assign a weight on how much you should be placing on that. At the end of the training process, the machine has learned I need to weigh this factor so much, I need to weigh this other factor so much, I need to weigh this other factor so much so the model, but just by definition, will tell you how much it weighs every single factor. So in a sense it's explainable. And so, going back to decision trees, so that's like asking 20 questions. Now imagine you had a jury of people, each of them asking 20 different questions, variations of these same 20 questions for different pieces of the data, and then at the end, taking a vote. That's what random forests are. The word forest comes from the fact that you have many trees. Get it.

Speaker 1:

Yes, yes, I get it. Haha, that was a dad joke.

Speaker 2:

Yeah, but that's it.

Speaker 1:

So you could ask the questions in different orders and then get slightly different results, and then you pull it all together.

Speaker 2:

And then again you learn what questions to ask, when to ask them, and so on, and so that's random forests and they're really powerful. Gradient boost also works. It works kind of differently. Think of it as a series of coaches, where the first person kind of tells you roughly what to do in a very loose way, and you kind of do it, you make all these mistakes, and the second person looks at the mistakes you made and teaches you something different. So in each iteration it's getting the errors that you made to make it slightly better, to make the model slightly better. It's a little bit different in the way it works, but it's essentially that's. The core idea is that you utilize the. You kind of start kind of rough and then you kind of get finer and finer and more and more correct as you go on. But you start with something that's kind of rough, if that makes sense, and so that's so those. But gradient boost to some degree is doesn't care about the underlying model, although we do talk about it in terms of decision trees. It could be applied to other models too, but that's kind of roughly the broad world of, and then you have neural networks. Sorry, I didn't mention that because I think they did work with. I think there's a. I've thrown a lot at you. Right now I don't know what you're talking about, but I think one thing to keep in mind as you process all of this is why do we need so many models Like? Why do we need all of these different things?

Speaker 2:

Well, some of the things to keep in mind when machine learning engineers think about these things is, like I said before, the nature of the data. How many lines of data do you have? How many patients do you have? How many lines of data do you have? How many patients do you have? How many features do you have?

Speaker 2:

Imagine I gave you an image and told you tell me if it's a cat or not. If you think about the actual image, it's a bunch of pixels, and if it's a high resolution image, it's millions of pixels, where each pixel is red, green or blue. Right, that's kind of how images are made up. That's a million factors. That's way too many for any decision tree or any of these other models to work.

Speaker 2:

Neural networks really shine in those scenarios, but because of that you lose interpretability, because they're all giant black boxes, so you don't have the benefit, whereas decision trees work really well in these settings where you have tabular data. It doesn't matter how many data points you have, but you have a reasonable number of factors. But you have something that is very heterogeneous. If you think about the image again, it's all the same, right. Every pixel is kind of the same. It's not heterogeneous. Every single pixel is either red, green or blue, whereas decision trees are better equipped to handle more heterogeneous data. The reason gradient boost is doing really well in this model and even random forest is that there might be inherent interaction between the factors.

Speaker 1:

Yeah, of course they're not independent. Yeah.

Speaker 2:

In a linear model they are all kind of separate. So, like in a logistic regression model, you can say that remember I talked about putting it to a line. But what if you can't put it to a line? What if it's some complex shape that it actually has to capture? Then you have interactions between these factors and logistic regression isn't as good anymore.

Speaker 1:

I mean that makes sense to me, right. If it's like a hot day and I live far from the clinic and I have to walk there because my area deprivation index is, you know, a certain number, then I'm just going to be less likely to go. So it does make sense that if we can capture the interactions between the factors, we can kind of understand it all better.

Speaker 2:

Yeah, and I think that's a way to think about why there exists so many different types of models is they solve slightly different problems with the data. They're better suited for certain kinds of data, and often the trade-off is interpretability, and so you know that depends on the application.

Speaker 1:

Right, and the interpretability here is key, because the whole point is to understand why people don't show up or why people cancel late, so that you can intervene on that right, so that you can do something about it.

Speaker 2:

Yeah, no, exactly. And I do want to take a second to talk also about how one judges these models. How do we know a model is doing well versus not well? And there's all these different metrics that people use. People often use the term accuracy, Like how accurate is it? Did it get it right? But there's more to that term accuracy than meets the eye. Right, you could have the model predict a no-show and in fact the person was a no-show. Or you could have the model predict something else and the person was a no-show. Or you could have the model predict something else and the person was a no-show. Or you could have the model predict a no-show and the person was something else.

Speaker 2:

right, there's like different ways of going wrong, in other words. So machine learning engineers use the terms true positives and true negatives to describe scenarios where the machine got it right, where the prediction and the underlying truth was the same. True positive means machine predicted true and the underlying truth is also true. I think an easy way to envision this is actually not with our example here, but with the spam email the machine, spam detector. The machine predicted spam and it actually was spam. The machine, spam detector. The machine predicted spam and it actually was spam. Versus the machine. And that's a true positive and a true negative. Is machine predicted not spam and it really was not spam. But there's also false positives, which is the machine predicted spam but it actually wasn't, and the other way around.

Speaker 2:

So what that means all of that is to say that there is other metrics like that that the machine learning engineers use to build on these core ideas. They use things like precision and recall and all of these ideas and again, all of these ideas. They might seem really complex, but at the very core of it, all they're trying to do is decide if whatever the machine is predicting, first of all, is that right, all the spams that it's predicted? Are they really all spams and all the spams that exist out there? Is it really catching all of them? And they're after these sorts of metrics and they use fancy terms like position and recall to capture these ideas. But that's the very core of it. But they don't stop just there. Remember I mentioned before that we're not just drawing a line, we're trying to like decide where to put the threshold to decide if something qualifies as spam versus not spam, and you know the sort of when, where to draw the line.

Speaker 2:

That line is arbitrarily drawn at some level.

Speaker 2:

And what you know in this paper they've done a good job of is to say, okay, what if we just pick various lines? We just drew all kinds of lines and we, we calculated for each of those lines, we just we evaluated how well the model did, and then we draw a graph out of that and say, okay, I've tried all of these different thresholds, all of these different ways to cut up my data. Does the model do well regardless? And so that's one of the big, I think one of the nice pieces here in this paper as well, which is a standard practice, but I think it's nice to note, because what it does for you, it tells you that can the model actually discriminate between those three cases, in our case the no-show, the cancellation and the completion?

Speaker 2:

How well can it actually discriminate between those three cases, in our case the no-show, the cancellation and the completion? How well can it actually discriminate between them? And so you need to be able to think about these other kinds of mathematical measures, and that's again. These papers are filled with mathematics for that reason, because they want to get at these very core ideas. How accurate is it? Does it catch everything I want to catch and can it really tell apart these different groups Right?

Speaker 1:

And I think they did. Actually, when you look at the factors, they really were able to figure out what the most influential factors were. The number one most influential factor drumroll please was actually the scheduled lead time. So if you schedule an appointment and it was more than 60 days out, you were much less likely to show up for it, which I mean, I don't know, I've definitely missed things that were pretty far out of my calendar, and you can imagine perhaps that if someone had such a non acute issue, maybe it went away by then or they just, you know, it fell off their radar as being something important or whatever it was. So schedule lead time was the number one factor, but there were a lot of other factors they found at sort of the population level, like how long a provider had been practicing, what was the age of the patient? Was the visit with the patient's own PCP? I think that one's interesting because I really like seeing my own PCP. I don't like seeing other people in the practice, right?

Speaker 1:

Other things that were important factors here were whether or not someone had gone to their other appointments, right?

Speaker 1:

Were they a no-show in the past or had they canceled in the past because they were more likely to then again.

Speaker 1:

And then the geosocial context turned out to be important, right when they lived, what their area deprivation index was, how far it was from their home to the clinic, and, like I mentioned before, environmental factors like the temperature played a role as well, and so they were able to see, okay, at the population level. There are all these factors that are at play, but I think one of the things that was very interesting to me is not only could they see that across the whole population of the more than 100,000 patients they looked at, but they were actually able to drill down and say oh at. But they were actually able to drill down and say oh, vasant, like you didn't go to your appointment or you were a late cancellation, and we know why for you and your reason wasn't the lead time, because your lead time was short, your reason was the temperature and the distance from the clinic or whatever it was. They were able to, like, drill down into individuals and figure out which of those features, which of those factors, were the most important.

Speaker 2:

Yeah, and that is something that you know is called the Shapley value, which is a mathematical term, and you know it's one thing that I love about this is that it's getting at the, it's breaking apart the features. We talked about all the factors, but identifying which ones played the role to make the model predict what it predicted is what Shapley calculations are all about and that's what they use to determine what you just said and what I like about the Shapley value is that the way it does it is again going back to our spreadsheet and having all of the factors to be different columns. What they essentially do is try to see how accurate the model is. If I just took out a certain column and they play around with that, I'm sort of oversimplifying, but the basic idea is that you take columns out and then you see what happens. And then you take other columns out and you see what happens. You take groups of columns out and see what happens and you sort of average out all of that and then you realize, okay, based on all of these, these are the factors that really contributed to the globally to the model. That means I'm not talking about specific patients here, but broadly looking at all of the data that the model has looked at. It's making its predictions highly influenced by some of these. These are the following factors that it's making its decisions on, you know, highly influenced by these. So that's at the global level.

Speaker 2:

But, like you said, the same Shapley value can also be used for an individual patient. So if I come in with a new piece of data, I have this trained model ready to go. It's sitting on my table and I come in and I put a new line, new row in that spreadsheet for a new patient. I can do a Shapley type calculation there too, where I literally change a column. This is how they do it.

Speaker 2:

Actually. They take one of the columns that they're interested in. They do this for every single column, but they pick one column and then they just randomly replace it by some other value from some other piece of data from their existing data set and then they make variations, perturbations of that data point and then see what happens to the accuracy and the prediction. Does the prediction change, how much does the prediction change by? And so that's kind of how they're able to do that. But that's right. But I do want to note that it is not the case that the model is saying hey, you showed up late because you don't go out in hot days or whatever. Right, whatever the factor was, that was deemed.

Speaker 1:

Temperature yeah, yeah.

Speaker 2:

Temperature right. I'm not saying that that is the factor, but just for hypothetical sake. Yeah, but that's not what the model is saying. The model is not saying it. That's not a causal model. It's not a causal prediction no-transcript. The cause no, and it's a. It's a factor in relation to the trained model that was trained on all the other data that was there before.

Speaker 1:

Which, like, makes sense, right, the the fact, some of the factors, like having a no-show in the past. It doesn't cause you to have a no-show now, right, but it does maybe explain, if you've had no-shows before, that you might be more likely to have one again.

Speaker 2:

Yeah, and the purpose of the Shapley values is to explain the model predictions so you understand the model better. It is not meant to serve as a explainer of the human behavior involved.

Speaker 1:

Yeah, I mean that makes sense. But at the same time my mind is going to okay, like, what do we do with this information? What do we do with this population level information? What do we do with this individual level information? And so there are certain things like age was found to be important in the model. Okay, we can't change their age, right.

Speaker 1:

But there are other things like the lead time to appointment, where, okay, maybe it's not causal, but like, what if there's? Maybe, what if it is a little bit right? And what can we do about it? Because that's always my question is how can we reduce the no-shows, how can we reduce the late cancellation? So what if, in this case, a clinic tried out and said, hey, we're going to make sure we book all our appointments for this subpopulation within 60 days and see if that changes our no-show or late cancellation rate? Or let's see if we can get people in with their own PCPs better rather than having cross-coverage.

Speaker 1:

Or maybe, when you think about clinic operations, perhaps we can't predict the hot day or whatever. But there are situations where you could say, hmm, I'm expecting a lot of no-shows or late cancellations, maybe we can squeeze an extra patient in here. So, even though they're not causal, it's like maybe they're actionable or you could experiment with pieces of this and try to figure out is there something we can optimize here? Can we get you know? One of the things we didn't talk too much about was like could we get you know an interpreter of someone who doesn't speak English is less likely to show up, or could we get transportation for someone who who's you know, has a hard time getting into clinic? All of these things are associated with it. So, even though it doesn't prove it's causal, in some cases, like in my mind, I'm like well, yeah, but it could be, yeah yeah, yeah, no, absolutely.

Speaker 2:

I think that it serves as at least ideas, right, ideas for potential interventions, and it's funny because it's meant to be a predictive model, but maybe it's a descriptive model of what's happening and at some weird level and that's kind of what you're suggesting at these correlations could mean something. We don't know yet, but it's worthwhile investigating them.

Speaker 1:

Right, yeah, so I think this is a great paper because it really does show, you can predict for the most part who's going to miss or not show up, who's going to be a late cancellation. And these machine learning models really help, especially when we give them lots of cleaned up, diverse data sets that involve more than just what's in the electronic health record and they don't quite explain why. But you know, we could try to take some of that information and apply it at the population level, at the patient level, at the clinic level, to intervene and try to reduce the number of no-shows and late cancellations.

Speaker 2:

Yeah, I'd be very curious to see what the authors here move forward with. Do they build something from it and work with that and actually do those next steps that you just talked about?

Speaker 1:

Yeah Well, thanks so much for joining us. Thank you. See you next time.

People on this episode