This video, we're going to be talking about reinforcement schedules. So unlike in classical conditioning where learning is essentially just based on repetition, something happening over and over again, research has shown that operant conditioning can change based on the patterns of reinforcement. We're going to go over all those different types of patterns in this video. Now, I do want to give you a heads-up this is going to be a very terminology-heavy video. There are a lot of different terms to learn.
However, the terms are all quite intuitive, and once you know what they mean, all of these are going to be pretty easy to understand, but it can be a bit tricky at first. So definitely take your time with this one, kind of slow it down if you need to, and I promise it will totally make sense in the end. Alright. So our first type of reinforcement is called interval reinforcement, and interval reinforcement is essentially dependent on the length of time that elapses between reinforcement. So this is completely dependent on time.
It is not dependent on the animal's behavior or responses in any way. And we have two types of interval reinforcement. Our fixed interval schedule or fixed interval reinforcement essentially means that behavior is going to be reinforced at a fixed or a set time. So, for example, a rat might get a treat every five minutes. It does not matter how many times in that five-minute span the rat pushes that little lever.
The behavior does not matter. It's only going to come after five minutes. Humans also experience this if you've ever gotten a paycheck. You probably get a paycheck once a week or twice a month or whatever it is, and it's not super dependent on your behavior, right?
Maybe one week you kind of slack off a little bit. Maybe one week you take a day off. Maybe one week you work super hard, but that paycheck is going to come at the same time no matter what your behavior looked like. So that is fixed interval. Now we can also have variable interval, and this would be when behavior is going to be reinforced at a variable amount of time since the last reinforcement.
So again, these terms are pretty intuitive once you can kind of get the hang of them. So essentially, this could be, you know, a rat getting a treat about every five minutes, maybe it comes every four minutes, the next time it comes at six minutes. The idea is that the animal cannot quite anticipate exactly when it's going to come.
That time is now variable. This is also kind of like fishing. Obviously, there is some skill and behavior involved in fishing. I don't want to imply that there isn't, but there are plenty of days where you, you know, put your line in the water and some days you get a response after one minute, some days you get a response after five minutes, and it does not depend on your behavior at all. So that would be a variable interval schedule.
Now, moving away from interval reinforcement, we can also have something called ratio reinforcement, and ratio reinforcement is more dependent on the ratio of responses to reinforcement. So now, responses and behavior are what matters. So with interval reinforcement, time is important. With ratio reinforcement, the response or the behavior is more important, and again we have two types fixed and variable. So with a fixed ratio schedule, behavior is going to be reinforced after a fixed number of repetitions.
So, for example, the rat might get a treat every five responses. Or this is also how customer loyalty programs work. So if you've ever been part of one of these, maybe you have to buy, you know, 10 coffees and then you get a free coffee, right? But this is entirely dependent on your behaviors and how many behaviors you do. It doesn't matter if it takes you a week to buy 10 coffees.
It could take you a month to buy 10 coffees, right? The time doesn't matter, just your behavior. So that's fixed ratio. Now with variable ratio schedules, you probably know what's coming here, but behavior is going to be reinforced based on a variable number of responses or repetitions. So now that rat is going to get a treat about every five responses, could come after three and then it could come after seven.
The idea again is that the animal cannot quite predict it. That is going to be a variable, responding rate. This is also how slot machines work. So you can imagine, people just keep on pulling those levers, right, and you might get a jackpot after a hundred lever pulls. It can take a really long time, but you know that that reinforcement is coming if you just do enough of the behavior.
That's kind of the idea here. So that is variable ratio schedules or variable ratio reinforcement. So to summarize this, when we're thinking about our reinforcement schedules, quite often you will see them depicted on a graph like this where we're looking at time on the x-axis and the number of responses on the y-axis. And this is not going to be real data, but the data usually comes out looking kind of like this where what we see is that ratio schedules, both variable ratio and fixed ratio, usually lead to the fastest learning and the most responses. Okay?
So and this makes sense, right? Because with these ratio schedules, behavior is what matters. Responding is the most important thing here, and so you're going to have both humans and animals learning very quickly to do more of the behaviors in order to get more of the rewards. So again, ratio schedules usually lead to faster learning and higher response rates. With interval schedules, we usually see a little bit slower learning and less responding overall.
And this totally makes sense, right, because, again, with these interval schedules, the only thing that matters is time. These are not dependent on the animal's behavior at all. So why would you have a high response rate if your responses don't actually matter? Right? There's no point in pushing the lever a thousand times if that reward is just coming after five minutes if you push it once or push it a hundred times.
It doesn't matter. Right? Now one more kind of specific thing to note is that, like we said, the ratio schedules lead to kind of the fastest learning and the most responding, and out of those variable ratio schedules tend to be the absolute most effective. So they lead to the fastest learning, the highest response rates, and they are the most resistant to extinction. And again, think of those slot machines just to kind of keep that in your mind to help you remember that these are kind of the most resistant to extinction.
We've all seen that stereotype of a person who's gambling or using slot machines and they will respond hundreds if not thousands of times just anticipating that reinforcement, right? So these schedules can lead to really high rates of response, and they're very hard to extinguish once they are learned. Alright. Great job with that one, guys, and I will see you in our next one. Bye bye.