{"id":187,"date":"2013-11-21T18:54:31","date_gmt":"2013-11-21T18:54:31","guid":{"rendered":"http:\/\/opentextbc.ca\/introductiontopsychology\/?post_type=chapter&#038;p=187"},"modified":"2016-11-17T21:11:46","modified_gmt":"2016-11-17T21:11:46","slug":"7-2-changing-behavior-through-reinforcement-and-punishment-operant-conditioning","status":"publish","type":"chapter","link":"https:\/\/opentextbc.ca\/introductiontopsychology\/chapter\/7-2-changing-behavior-through-reinforcement-and-punishment-operant-conditioning\/","title":{"raw":"8.2 Changing Behaviour through Reinforcement and Punishment: Operant Conditioning","rendered":"8.2 Changing Behaviour through Reinforcement and Punishment: Operant Conditioning"},"content":{"raw":"<div class=\"bcc-box bcc-highlight\">\r\n<h3>Learning Objectives<\/h3>\r\n<ol>\r\n \t<li>Outline the principles of operant conditioning.<\/li>\r\n \t<li>Explain how learning can be shaped through the use of reinforcement schedules and secondary reinforcers.<\/li>\r\n<\/ol>\r\n<\/div>\r\nIn classical conditioning the organism learns to associate new stimuli with natural biological responses such as salivation or fear. The organism does not learn something new but rather begins to perform an existing behaviour in the presence of a new signal. <strong>Operant conditioning<\/strong>, on the other hand, is <em>learning that occurs based on the consequences of behaviour<\/em> and can involve the learning of new actions. Operant conditioning occurs when a dog rolls over on command because it has been praised for doing so in the past, when a schoolroom bully threatens his classmates because doing so allows him to get his way, and when a child gets good grades because her parents threaten to punish her if she doesn\u2019t. In operant conditioning the organism learns from the consequences of its own actions.\r\n<h2>How Reinforcement and Punishment Influence Behaviour: The Research of Thorndike and Skinner<\/h2>\r\nPsychologist Edward L. Thorndike (1874-1949) was the first scientist to systematically study operant conditioning. In his research Thorndike (1898)\u00a0observed cats who had been placed in a \u201cpuzzle box\u201d from which they tried to escape (\"Video Clip: Thorndike\u2019s Puzzle Box\"). At first the cats scratched, bit, and swatted haphazardly, without any idea of how to get out. But eventually, and accidentally, they pressed the lever that opened the door and exited to their prize, a scrap of fish. The next time the cat was constrained within the box, it attempted fewer of the ineffective responses before carrying out the successful escape, and after several trials the cat learned to almost immediately make the correct response.\r\n\r\nObserving these changes in the cats\u2019 behaviour led Thorndike to develop his<strong> law of effect<\/strong>, <em>the principle that responses that create a typically pleasant outcome in a particular situation are more likely to occur again in a similar situation, whereas responses that produce a typically unpleasant outcome are less likely to occur again in the situation<\/em> (Thorndike, 1911).\u00a0The essence of the law of effect is that successful responses, because they are pleasurable, are \u201cstamped in\u201d by experience and thus occur more frequently. Unsuccessful responses, which produce unpleasant experiences, are \u201cstamped out\u201d and subsequently occur less frequently.\r\n\r\n<em>When Thorndike placed his cats in a puzzle box, he found that they learned to engage in the important escape behaviour faster after each trial. Thorndike described the learning that follows reinforcement in terms of the law of effect.<\/em>\r\n\r\n<a href=\"http:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2014\/09\/Thorndike\u2019s-Puzzle-Box-video.png\"><img src=\"https:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2014\/09\/Thorndike\u2019s-Puzzle-Box-video-150x150.png\" alt=\"&quot;&quot;\" class=\"alignright wp-image-2664 size-thumbnail\" height=\"150\" width=\"150\" \/><\/a><strong><a href=\"http:\/\/www.youtube.com\/v\/BDujDOLre-8\">Watch: \"Thorndike\u2019s Puzzle Box\" [YouTube]<\/a>: <\/strong>http:\/\/www.youtube.com\/watch?v=BDujDOLre-8\r\n\r\nThe influential behavioural psychologist B. F. Skinner (1904-1990) expanded on Thorndike\u2019s ideas to develop a more complete set of principles to explain operant conditioning. Skinner created specially designed environments known as <em>operant chambers<\/em> (usually called <em>Skinner boxes<\/em>) to systematically study learning.<strong> A Skinner box<\/strong> (operant chamber) is <em>a structure that is big enough to fit a rodent or bird and that contains a bar or key that the organism can press or peck to release food or water. It also contains a device to record the animal\u2019s responses<\/em> (Figure 8.5).\r\n\r\nThe most basic of Skinner\u2019s experiments was quite similar to Thorndike\u2019s research with cats. A rat placed in the chamber reacted as one might expect, scurrying about the box and sniffing and clawing at the floor and walls. Eventually the rat chanced upon a lever, which it pressed to release pellets of food. The next time around, the rat took a little less time to press the lever, and on successive trials, the time it took to press the lever became shorter and shorter. Soon the rat was pressing the lever as fast as it could eat the food that appeared. As predicted by the law of effect, the rat had learned to repeat the action that brought about the food and cease the actions that did not.\r\n\r\nSkinner studied, in detail, how animals changed their behaviour through reinforcement and punishment, and he developed terms that explained the processes of operant learning (Table 8.1, \"How Positive and Negative Reinforcement and Punishment Influence Behaviour\"). Skinner used the term <strong>reinforcer\u00a0<\/strong>to refer to <em>any event that strengthens or increases the likelihood of a behaviour,<\/em> and the term <strong>punisher<\/strong>\u00a0to refer to <em>any event that weakens or decreases the likelihood of a behaviour<\/em>. And he used the terms <em>positive<\/em> and <em>negative<\/em> to refer to whether a reinforcement was presented or removed, respectively. Thus, <strong>positive reinforcement <\/strong><em>strengthens a response by presenting something pleasant after the response,<\/em> and <strong>negative reinforcement<\/strong> <em>strengthens a response by reducing or removing something unpleasant<\/em>. For example, giving a child praise for completing his homework represents positive reinforcement, whereas taking Aspirin to reduce the pain of a headache represents negative reinforcement. In both cases, the reinforcement makes it more likely that behaviour will occur again in the future.\r\n\r\n[caption id=\"attachment_1067\" align=\"aligncenter\" width=\"400\"]<a href=\"http:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2013\/11\/skinnerbox.jpg\"><img class=\"wp-image-1067\" alt=\"&quot;&quot;\" src=\"http:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2013\/11\/skinnerbox-1024x512.jpg\" height=\"200\" width=\"400\" \/><\/a> Figure 8.5 Skinner Box. B. F. Skinner used a Skinner box to study operant learning. The box contains a bar or key that the organism can press to receive food and water, and a device that records the organism\u2019s responses.[\/caption]\r\n<table><caption>Table 8.1 How Positive and Negative Reinforcement and Punishment Influence Behaviour.<\/caption>\r\n<thead>\r\n<tr>\r\n<td style=\"text-align: center\" colspan=\"4\"><a href=\"#skiptable8.1\">[Skip Table]<\/a><\/td>\r\n<\/tr>\r\n<tr>\r\n<th>Operant conditioning term<\/th>\r\n<th>Description<\/th>\r\n<th>Outcome<\/th>\r\n<th>Example<\/th>\r\n<\/tr>\r\n<\/thead>\r\n<tbody>\r\n<tr>\r\n<td>Positive reinforcement<\/td>\r\n<td>Add or increase a pleasant stimulus<\/td>\r\n<td>Behaviour is strengthened<\/td>\r\n<td>Giving a student a prize after he or she gets an A on a test<\/td>\r\n<\/tr>\r\n<tr>\r\n<td>Negative reinforcement<\/td>\r\n<td>Reduce or remove an unpleasant stimulus<\/td>\r\n<td>Behaviour is strengthened<\/td>\r\n<td>Taking painkillers that eliminate pain increases the likelihood that you will take painkillers again<\/td>\r\n<\/tr>\r\n<tr>\r\n<td>Positive punishment<\/td>\r\n<td>Present or add an unpleasant stimulus<\/td>\r\n<td>Behaviour is weakened<\/td>\r\n<td>Giving a student extra homework after he or she misbehaves in class<\/td>\r\n<\/tr>\r\n<tr>\r\n<td>Negative punishment<\/td>\r\n<td>Reduce or remove a pleasant stimulus<\/td>\r\n<td>Behaviour is weakened<\/td>\r\n<td>Taking away a teen\u2019s computer after he or she misses curfew<\/td>\r\n<\/tr>\r\n<\/tbody>\r\n<\/table>\r\n<p id=\"skiptable8.1\">Reinforcement, either positive or negative, works by increasing the likelihood of a behaviour. <strong>Punishment<\/strong>, on the other hand, refers to <em>any event that weakens or reduces the likelihood of a behaviour<\/em>. <strong>Positive punishment<\/strong>\u00a0<em>weakens a response by presenting something unpleasant after the response<\/em>, whereas<strong> negative punishment<\/strong>\u00a0<em>weakens a response by reducing or removing something pleasant<\/em>. A child who is grounded after fighting with a sibling (positive punishment) or who loses out on the opportunity to go to recess after getting a poor grade (negative punishment) is less likely to repeat these behaviours.<\/p>\r\nAlthough the distinction between reinforcement (which increases behaviour) and punishment (which decreases it) is usually clear, in some cases it is difficult to determine whether a reinforcer is positive or negative. On a hot day a cool breeze could be seen as a positive reinforcer (because it brings in cool air) or a negative reinforcer (because it removes hot air). In other cases, reinforcement can be both positive and negative. One may smoke a cigarette both because it brings pleasure (positive reinforcement) and because it eliminates the craving for nicotine (negative reinforcement).\r\n\r\nIt is also important to note that reinforcement and punishment are not simply opposites. The use of positive reinforcement in changing behaviour is almost always more effective than using punishment. This is because positive reinforcement makes the person or animal feel better, helping create a positive relationship with the person providing the reinforcement. Types of positive reinforcement that are effective in everyday life include verbal praise or approval, the awarding of status or prestige, and direct financial payment. Punishment, on the other hand, is more likely to create only temporary changes in behaviour because it is based on coercion and typically creates a negative and adversarial relationship with the person providing the reinforcement. When the person who provides the punishment leaves the situation, the unwanted behaviour is likely to return.\r\n<h2>Creating Complex Behaviours through Operant Conditioning<\/h2>\r\nPerhaps you remember watching a movie or being at a show in which an animal \u2014 maybe a dog, a horse, or a dolphin \u2014 did some pretty amazing things. The trainer gave a command and the dolphin swam to the bottom of the pool, picked up a ring on its nose, jumped out of the water through a hoop in the air, dived again to the bottom of the pool, picked up another ring, and then took both of the rings to the trainer at the edge of the pool. The animal was trained to do the trick, and the principles of operant conditioning were used to train it. But these complex behaviours are a far cry from the simple stimulus-response relationships that we have considered thus far. How can reinforcement be used to create complex behaviours such as these?\r\n\r\nOne way to expand the use of operant learning is to modify the schedule on which the reinforcement is applied. To this point we have only discussed a <strong>continuous reinforcement schedule<\/strong>, in which <em>the desired response is reinforced every time it occurs<\/em>; whenever the dog rolls over, for instance, it gets a biscuit. Continuous reinforcement results in relatively fast learning but also rapid extinction of the desired behaviour once the reinforcer disappears. The problem is that because the organism is used to receiving the reinforcement after every behaviour, the responder may give up quickly when it doesn\u2019t appear.\r\n\r\nMost real-world reinforcers are not continuous; they occur on a <strong>partial (or intermittent) reinforcement schedule <\/strong>\u2014 <em>a schedule in which the responses are sometimes reinforced and sometimes not<\/em>. In comparison to continuous reinforcement, partial reinforcement schedules lead to slower initial learning, but they also lead to greater resistance to extinction. Because the reinforcement does not appear after every behaviour, it takes longer for the learner to determine that the reward is no longer coming, and thus extinction is slower. The four types of partial reinforcement schedules are summarized in Table 8.2, \"Reinforcement Schedules.\"\r\n<table><caption>Table 8.2 Reinforcement Schedules.<\/caption>\r\n<thead>\r\n<tr>\r\n<td style=\"text-align: center\" colspan=\"3\"><a href=\"#skiptable8.2\">[Skip Table]<\/a><\/td>\r\n<\/tr>\r\n<tr>\r\n<th>Reinforcement schedule<\/th>\r\n<th>Explanation<\/th>\r\n<th>Real-world example<\/th>\r\n<\/tr>\r\n<\/thead>\r\n<tbody>\r\n<tr>\r\n<td>Fixed-ratio<\/td>\r\n<td>Behaviour is reinforced after a specific number of responses.<\/td>\r\n<td>Factory workers who are paid according to the number of products they produce<\/td>\r\n<\/tr>\r\n<tr>\r\n<td>Variable-ratio<\/td>\r\n<td>Behaviour is reinforced after an average, but unpredictable, number of responses.<\/td>\r\n<td>Payoffs from slot machines and other games of chance<\/td>\r\n<\/tr>\r\n<tr>\r\n<td>Fixed-interval<\/td>\r\n<td>Behaviour is reinforced for the first response after a specific amount of time has passed.<\/td>\r\n<td>People who earn a monthly salary<\/td>\r\n<\/tr>\r\n<tr>\r\n<td>Variable-interval<\/td>\r\n<td>Behaviour is reinforced for the first response after an average, but unpredictable, amount of time has passed.<\/td>\r\n<td>Person who checks email for messages<\/td>\r\n<\/tr>\r\n<\/tbody>\r\n<\/table>\r\n<p id=\"skiptable8.2\">Partial reinforcement schedules are determined by whether the reinforcement is presented on the basis of the time that elapses between reinforcement (interval) or on the basis of the number of responses that the organism engages in (ratio), and by whether the reinforcement occurs on a regular (fixed) or unpredictable (variable) schedule. In a <strong>fixed-interval schedule<\/strong>, <em>reinforcement occurs for the first response made after a specific amount of time has passed<\/em>. For instance, on a one-minute fixed-interval schedule the animal receives a reinforcement every minute, assuming it engages in the behaviour at least once during the minute. As you can see in Figure 8.6, \"Examples of Response Patterns by Animals Trained under Different Partial Reinforcement Schedules,\" animals under fixed-interval schedules tend to slow down their responding immediately after the reinforcement but then increase the behaviour again as the time of the next reinforcement gets closer. (Most students study for exams the same way.) In a <strong>variable-interval schedule<\/strong>, <em>the reinforcers appear on an interval schedule, but the timing is varied around the average interval, making the actual appearance of the reinforcer unpredictable<\/em>. An example might be checking your email: you are reinforced by receiving messages that come, on average, say, every 30 minutes, but the reinforcement occurs only at random times. Interval reinforcement schedules tend to produce slow and steady rates of responding.<\/p>\r\n\r\n\r\n[caption id=\"attachment_185\" align=\"aligncenter\" width=\"400\"]<a href=\"http:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2013\/11\/555f0fdcff3a6dbbe56a86acd40f3463.jpg\"><img src=\"http:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2013\/11\/555f0fdcff3a6dbbe56a86acd40f3463-1024x411.jpg\" alt=\"&quot;&quot;\" class=\"wp-image-185\" height=\"161\" width=\"400\" \/><\/a> Figure 8.6 Examples of Response Patterns by Animals Trained under Different Partial Reinforcement Schedules. Schedules based on the number of responses (ratio types) induce greater response rate than do schedules based on elapsed time (interval types). Also, unpredictable schedules (variable types) produce stronger responses than do predictable schedules (fixed types).[\/caption]\r\n\r\nIn a <strong>fixed-ratio schedule<\/strong>, <em>a behaviour is reinforced after a specific number of responses<\/em>. For instance, a rat\u2019s behaviour may be reinforced after it has pressed a key 20 times, or a salesperson may receive a bonus after he or she has sold 10 products. As you can see in Figure 8.6, \"Examples of Response Patterns by Animals Trained under Different Partial Reinforcement Schedules,\" once the organism has learned to act in accordance with the fixed-ratio schedule, it will pause only briefly when reinforcement occurs before returning to a high level of responsiveness. A <strong>variable-ratio schedule<\/strong>\u00a0<em>provides reinforcers after a specific but average number of responses<\/em>. Winning money from slot machines or on a lottery ticket is an example of reinforcement that occurs on a variable-ratio schedule. For instance, a slot machine (see Figure 8.7, \"Slot Machine\") may be programmed to provide a win every 20 times the user pulls the handle, on average. Ratio schedules tend to produce high rates of responding because reinforcement increases as the number of responses increases.\r\n\r\n[caption id=\"attachment_4371\" align=\"aligncenter\" width=\"400\"]<a href=\"http:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2014\/10\/Figure-8-7.jpg\"><img src=\"http:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2014\/10\/Figure-8-7-1024x682.jpg\" alt=\"&quot;&quot;\" class=\"wp-image-4371\" height=\"267\" width=\"400\" \/><\/a> Figure 8.7 Slot Machine. Slot machines are examples of a variable-ratio reinforcement schedule.[\/caption]\r\n\r\nComplex behaviours are also created through <strong>shaping<\/strong>, <em>the process of guiding an organism\u2019s behaviour to the desired outcome through the use of successive approximation to a final desired behaviour<\/em>. Skinner made extensive use of this procedure in his boxes. For instance, he could train a rat to press a bar two times to receive food, by first providing food when the animal moved near the bar. When that behaviour had been learned, Skinner\u00a0would begin to provide food only when the rat touched the bar. Further shaping limited the reinforcement to only when the rat pressed the bar, to when it pressed the bar and touched it a second time, and finally to only when it pressed the bar twice. Although it can take a long time, in this way operant conditioning can create chains of behaviours that are reinforced only when they are completed.\r\n\r\nReinforcing animals if they correctly discriminate between similar stimuli allows scientists to test the animals\u2019 ability to learn, and the discriminations that they can make are sometimes remarkable. Pigeons have been trained to distinguish between images of Charlie Brown and the other Peanuts characters (Cerella, 1980),\u00a0and between different styles of music and art (Porter &amp; Neuringer, 1984; Watanabe, Sakamoto &amp; Wakita, 1995).\r\n\r\nBehaviours can also be trained through the use of <em>secondary reinforcers<\/em>. Whereas a <strong>primary reinforcer<\/strong>\u00a0includes <em>stimuli that are naturally preferred or enjoyed by the organism, such as food, water, and relief from pain<\/em>, a <strong>secondary reinforcer<\/strong>\u00a0(sometimes called <em>conditioned reinforcer<\/em>) is a <em>neutral event that has become associated with a primary reinforcer through classical conditioning<\/em>. An example of a secondary reinforcer would be the whistle given by an animal trainer, which has been associated over time with the primary reinforcer, food. An example of an everyday secondary reinforcer is money. We enjoy having money, not so much for the stimulus itself, but rather for the primary reinforcers (the things that money can buy) with which it is associated.\r\n<div class=\"bcc-box bcc-success\">\r\n<h3>Key Takeaways<\/h3>\r\n<ul>\r\n \t<li>Edward Thorndike developed the law of effect: the principle that responses that create a typically pleasant outcome in a particular situation are more likely to occur again in a similar situation, whereas responses that produce a typically unpleasant outcome are less likely to occur again in the situation.<\/li>\r\n \t<li>B. F. Skinner expanded on Thorndike\u2019s ideas to develop a set of principles to explain operant conditioning.<\/li>\r\n \t<li>Positive reinforcement strengthens a response by presenting something that is typically pleasant after the response, whereas negative reinforcement strengthens a response by reducing or removing something that is typically unpleasant.<\/li>\r\n \t<li>Positive punishment weakens a response by presenting something typically unpleasant after the response, whereas negative punishment weakens a response by reducing or removing something that is typically pleasant.<\/li>\r\n \t<li>Reinforcement may be either partial or continuous. Partial reinforcement schedules are determined by whether the reinforcement is presented on the basis of the time that elapses between reinforcements (interval) or on the basis of the number of responses that the organism engages in (ratio), and by whether the reinforcement occurs on a regular (fixed) or unpredictable (variable) schedule.<\/li>\r\n \t<li>Complex behaviours may be created through shaping, the process of guiding an organism\u2019s behaviour to the desired outcome through the use of successive approximation to a final desired behaviour.<\/li>\r\n<\/ul>\r\n<\/div>\r\n<div class=\"bcc-box bcc-info\">\r\n<h3>Exercises and Critical Thinking<\/h3>\r\n<ol>\r\n \t<li>Give an example from daily life of each of the following: positive reinforcement, negative reinforcement, positive punishment, negative punishment.<\/li>\r\n \t<li>Consider the reinforcement techniques that you might use to train a dog to catch and retrieve a Frisbee that you throw to it.<\/li>\r\n \t<li>Watch the following two videos from current television shows. Can you determine which learning procedures are being demonstrated?\r\n<ol>\r\n \t<li><a href=\"http:\/\/www.break.com\/usercontent\/2009\/11\/the-office-altoid-experiment-1499823\"><em>The Office<\/em><\/a>: http:\/\/www.break.com\/usercontent\/2009\/11\/the-office-altoid- experiment-1499823<\/li>\r\n \t<li><a href=\"http:\/\/www.youtube.com\/watch?v=JA96Fba-WHk\"><em>The Big Bang Theory <\/em>[YouTube]<\/a>: http:\/\/www.youtube.com\/watch?v=JA96Fba-WHk<\/li>\r\n<\/ol>\r\n<\/li>\r\n<\/ol>\r\n<\/div>\r\n<h2>References<\/h2>\r\nCerella, J. (1980). The pigeon\u2019s analysis of pictures.\u00a0<i>Pattern Recognition, 12<\/i>, 1\u20136.\r\n\r\nKassin, S. (2003). <a href=\"http:\/\/wps.prenhall.com\/hss_kassin_essentials_1\/15\/3933\/1006917.cw\/index.html\">Essentials of psychology<\/a>. Upper Saddle River, NJ: Prentice Hall. Retrieved from Essentials of Psychology Prentice Hall Companion Website: http:\/\/wps.prenhall.com\/hss_kassin_essentials_1\/15\/3933\/1006917.cw\/index.html\r\n\r\nPorter, D., &amp; Neuringer, A. (1984). Music discriminations by pigeons.\u00a0<i>Journal of Experimental Psychology: Animal Behavior Processes, 10<\/i>(2), 138\u2013148.\r\n\r\nThorndike, E. L. (1898).\u00a0<i>Animal intelligence: An experimental study of the associative processes in animals.<\/i>\u00a0Washington, DC: American Psychological Association.\r\n\r\nThorndike, E. L. (1911).\u00a0<i><a href=\"http:\/\/www.archive.org\/details\/animalintelligen00thor\">Animal intelligence: Experimental studies<\/a>.<\/i>\u00a0New York, NY: Macmillan. Retrieved from\u00a0http:\/\/www.archive.org\/details\/animalintelligen00thor\r\n\r\nWatanabe, S., Sakamoto, J., &amp; Wakita, M. (1995). Pigeons\u2019 discrimination of painting by Monet and Picasso.\u00a0<i>Journal of the Experimental Analysis of Behaviour, 63<\/i>(2), 165\u2013174.\r\n<h2>Image Attributions<\/h2>\r\n<strong>Figure 8.5:<\/strong> \"<a href=\"http:\/\/en.wikipedia.org\/wiki\/File:Skinner_box_photo_02.jpg\">Skinner box<\/a>\" (http:\/\/en.wikipedia.org\/wiki\/File:Skinner_box_photo_02.jpg) is licensed under the CC BY SA 3.0 license (http:\/\/creativecommons.org\/licenses\/by-sa\/3.0\/deed.en). \"<a href=\"http:\/\/en.wikipedia.org\/wiki\/File:Skinner_box_scheme_01.png\">Skinner box scheme<\/a>\" by Andreas1 (http:\/\/en.wikipedia.org\/wiki\/File:Skinner_box_scheme_01.png) is licensed under the CC BY SA 3.0 license (http:\/\/creativecommons.org\/licenses\/by-sa\/3.0\/deed.en)\r\n\r\n<strong>Figure 8.6:<\/strong> Adapted from Kassin (2003).\r\n\r\n<strong>Figure 8.7:<\/strong>\u00a0 \"<a href=\"http:\/\/commons.wikimedia.org\/wiki\/File:HardRockCasinoSlotMachines.jpg\">Slot Machines in the Hard Rock Casino<\/a>\" by Ted Murpy (http:\/\/commons.wikimedia.org\/wiki\/File:HardRockCasinoSlotMachines.jpg) is licensed under <a href=\"http:\/\/creativecommons.org\/licenses\/by\/2.0\/deed.en\">CC BY 2.0.<\/a> (http:\/\/creativecommons.org\/licenses\/by\/2.0\/deed.en).","rendered":"<div class=\"bcc-box bcc-highlight\">\n<h3>Learning Objectives<\/h3>\n<ol>\n<li>Outline the principles of operant conditioning.<\/li>\n<li>Explain how learning can be shaped through the use of reinforcement schedules and secondary reinforcers.<\/li>\n<\/ol>\n<\/div>\n<p>In classical conditioning the organism learns to associate new stimuli with natural biological responses such as salivation or fear. The organism does not learn something new but rather begins to perform an existing behaviour in the presence of a new signal. <strong>Operant conditioning<\/strong>, on the other hand, is <em>learning that occurs based on the consequences of behaviour<\/em> and can involve the learning of new actions. Operant conditioning occurs when a dog rolls over on command because it has been praised for doing so in the past, when a schoolroom bully threatens his classmates because doing so allows him to get his way, and when a child gets good grades because her parents threaten to punish her if she doesn\u2019t. In operant conditioning the organism learns from the consequences of its own actions.<\/p>\n<h2>How Reinforcement and Punishment Influence Behaviour: The Research of Thorndike and Skinner<\/h2>\n<p>Psychologist Edward L. Thorndike (1874-1949) was the first scientist to systematically study operant conditioning. In his research Thorndike (1898)\u00a0observed cats who had been placed in a \u201cpuzzle box\u201d from which they tried to escape (&#8220;Video Clip: Thorndike\u2019s Puzzle Box&#8221;). At first the cats scratched, bit, and swatted haphazardly, without any idea of how to get out. But eventually, and accidentally, they pressed the lever that opened the door and exited to their prize, a scrap of fish. The next time the cat was constrained within the box, it attempted fewer of the ineffective responses before carrying out the successful escape, and after several trials the cat learned to almost immediately make the correct response.<\/p>\n<p>Observing these changes in the cats\u2019 behaviour led Thorndike to develop his<strong> law of effect<\/strong>, <em>the principle that responses that create a typically pleasant outcome in a particular situation are more likely to occur again in a similar situation, whereas responses that produce a typically unpleasant outcome are less likely to occur again in the situation<\/em> (Thorndike, 1911).\u00a0The essence of the law of effect is that successful responses, because they are pleasurable, are \u201cstamped in\u201d by experience and thus occur more frequently. Unsuccessful responses, which produce unpleasant experiences, are \u201cstamped out\u201d and subsequently occur less frequently.<\/p>\n<p><em>When Thorndike placed his cats in a puzzle box, he found that they learned to engage in the important escape behaviour faster after each trial. Thorndike described the learning that follows reinforcement in terms of the law of effect.<\/em><\/p>\n<p><a href=\"http:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2014\/09\/Thorndike\u2019s-Puzzle-Box-video.png\"><img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2014\/09\/Thorndike\u2019s-Puzzle-Box-video-150x150.png\" alt=\"&quot;&quot;\" class=\"alignright wp-image-2664 size-thumbnail\" height=\"150\" width=\"150\" srcset=\"https:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2014\/09\/Thorndike\u2019s-Puzzle-Box-video-150x150.png 150w, https:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2014\/09\/Thorndike\u2019s-Puzzle-Box-video-65x65.png 65w, https:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2014\/09\/Thorndike\u2019s-Puzzle-Box-video.png 200w\" sizes=\"auto, (max-width: 150px) 100vw, 150px\" \/><\/a><strong><a href=\"http:\/\/www.youtube.com\/v\/BDujDOLre-8\">Watch: &#8220;Thorndike\u2019s Puzzle Box&#8221; [YouTube]<\/a>: <\/strong>http:\/\/www.youtube.com\/watch?v=BDujDOLre-8<\/p>\n<p>The influential behavioural psychologist B. F. Skinner (1904-1990) expanded on Thorndike\u2019s ideas to develop a more complete set of principles to explain operant conditioning. Skinner created specially designed environments known as <em>operant chambers<\/em> (usually called <em>Skinner boxes<\/em>) to systematically study learning.<strong> A Skinner box<\/strong> (operant chamber) is <em>a structure that is big enough to fit a rodent or bird and that contains a bar or key that the organism can press or peck to release food or water. It also contains a device to record the animal\u2019s responses<\/em> (Figure 8.5).<\/p>\n<p>The most basic of Skinner\u2019s experiments was quite similar to Thorndike\u2019s research with cats. A rat placed in the chamber reacted as one might expect, scurrying about the box and sniffing and clawing at the floor and walls. Eventually the rat chanced upon a lever, which it pressed to release pellets of food. The next time around, the rat took a little less time to press the lever, and on successive trials, the time it took to press the lever became shorter and shorter. Soon the rat was pressing the lever as fast as it could eat the food that appeared. As predicted by the law of effect, the rat had learned to repeat the action that brought about the food and cease the actions that did not.<\/p>\n<p>Skinner studied, in detail, how animals changed their behaviour through reinforcement and punishment, and he developed terms that explained the processes of operant learning (Table 8.1, &#8220;How Positive and Negative Reinforcement and Punishment Influence Behaviour&#8221;). Skinner used the term <strong>reinforcer\u00a0<\/strong>to refer to <em>any event that strengthens or increases the likelihood of a behaviour,<\/em> and the term <strong>punisher<\/strong>\u00a0to refer to <em>any event that weakens or decreases the likelihood of a behaviour<\/em>. And he used the terms <em>positive<\/em> and <em>negative<\/em> to refer to whether a reinforcement was presented or removed, respectively. Thus, <strong>positive reinforcement <\/strong><em>strengthens a response by presenting something pleasant after the response,<\/em> and <strong>negative reinforcement<\/strong> <em>strengthens a response by reducing or removing something unpleasant<\/em>. For example, giving a child praise for completing his homework represents positive reinforcement, whereas taking Aspirin to reduce the pain of a headache represents negative reinforcement. In both cases, the reinforcement makes it more likely that behaviour will occur again in the future.<\/p>\n<figure id=\"attachment_1067\" aria-describedby=\"caption-attachment-1067\" style=\"width: 400px\" class=\"wp-caption aligncenter\"><a href=\"http:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2013\/11\/skinnerbox.jpg\"><img loading=\"lazy\" decoding=\"async\" class=\"wp-image-1067\" alt=\"&quot;&quot;\" src=\"http:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2013\/11\/skinnerbox-1024x512.jpg\" height=\"200\" width=\"400\" srcset=\"https:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2013\/11\/skinnerbox-1024x512.jpg 1024w, https:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2013\/11\/skinnerbox-300x150.jpg 300w, https:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2013\/11\/skinnerbox-65x32.jpg 65w, https:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2013\/11\/skinnerbox-225x112.jpg 225w, https:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2013\/11\/skinnerbox-350x175.jpg 350w, https:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2013\/11\/skinnerbox.jpg 1200w\" sizes=\"auto, (max-width: 400px) 100vw, 400px\" \/><\/a><figcaption id=\"caption-attachment-1067\" class=\"wp-caption-text\">Figure 8.5 Skinner Box. B. F. Skinner used a Skinner box to study operant learning. The box contains a bar or key that the organism can press to receive food and water, and a device that records the organism\u2019s responses.<\/figcaption><\/figure>\n<table>\n<caption>Table 8.1 How Positive and Negative Reinforcement and Punishment Influence Behaviour.<\/caption>\n<thead>\n<tr>\n<td style=\"text-align: center\" colspan=\"4\"><a href=\"#skiptable8.1\">[Skip Table]<\/a><\/td>\n<\/tr>\n<tr>\n<th>Operant conditioning term<\/th>\n<th>Description<\/th>\n<th>Outcome<\/th>\n<th>Example<\/th>\n<\/tr>\n<\/thead>\n<tbody>\n<tr>\n<td>Positive reinforcement<\/td>\n<td>Add or increase a pleasant stimulus<\/td>\n<td>Behaviour is strengthened<\/td>\n<td>Giving a student a prize after he or she gets an A on a test<\/td>\n<\/tr>\n<tr>\n<td>Negative reinforcement<\/td>\n<td>Reduce or remove an unpleasant stimulus<\/td>\n<td>Behaviour is strengthened<\/td>\n<td>Taking painkillers that eliminate pain increases the likelihood that you will take painkillers again<\/td>\n<\/tr>\n<tr>\n<td>Positive punishment<\/td>\n<td>Present or add an unpleasant stimulus<\/td>\n<td>Behaviour is weakened<\/td>\n<td>Giving a student extra homework after he or she misbehaves in class<\/td>\n<\/tr>\n<tr>\n<td>Negative punishment<\/td>\n<td>Reduce or remove a pleasant stimulus<\/td>\n<td>Behaviour is weakened<\/td>\n<td>Taking away a teen\u2019s computer after he or she misses curfew<\/td>\n<\/tr>\n<\/tbody>\n<\/table>\n<p id=\"skiptable8.1\">Reinforcement, either positive or negative, works by increasing the likelihood of a behaviour. <strong>Punishment<\/strong>, on the other hand, refers to <em>any event that weakens or reduces the likelihood of a behaviour<\/em>. <strong>Positive punishment<\/strong>\u00a0<em>weakens a response by presenting something unpleasant after the response<\/em>, whereas<strong> negative punishment<\/strong>\u00a0<em>weakens a response by reducing or removing something pleasant<\/em>. A child who is grounded after fighting with a sibling (positive punishment) or who loses out on the opportunity to go to recess after getting a poor grade (negative punishment) is less likely to repeat these behaviours.<\/p>\n<p>Although the distinction between reinforcement (which increases behaviour) and punishment (which decreases it) is usually clear, in some cases it is difficult to determine whether a reinforcer is positive or negative. On a hot day a cool breeze could be seen as a positive reinforcer (because it brings in cool air) or a negative reinforcer (because it removes hot air). In other cases, reinforcement can be both positive and negative. One may smoke a cigarette both because it brings pleasure (positive reinforcement) and because it eliminates the craving for nicotine (negative reinforcement).<\/p>\n<p>It is also important to note that reinforcement and punishment are not simply opposites. The use of positive reinforcement in changing behaviour is almost always more effective than using punishment. This is because positive reinforcement makes the person or animal feel better, helping create a positive relationship with the person providing the reinforcement. Types of positive reinforcement that are effective in everyday life include verbal praise or approval, the awarding of status or prestige, and direct financial payment. Punishment, on the other hand, is more likely to create only temporary changes in behaviour because it is based on coercion and typically creates a negative and adversarial relationship with the person providing the reinforcement. When the person who provides the punishment leaves the situation, the unwanted behaviour is likely to return.<\/p>\n<h2>Creating Complex Behaviours through Operant Conditioning<\/h2>\n<p>Perhaps you remember watching a movie or being at a show in which an animal \u2014 maybe a dog, a horse, or a dolphin \u2014 did some pretty amazing things. The trainer gave a command and the dolphin swam to the bottom of the pool, picked up a ring on its nose, jumped out of the water through a hoop in the air, dived again to the bottom of the pool, picked up another ring, and then took both of the rings to the trainer at the edge of the pool. The animal was trained to do the trick, and the principles of operant conditioning were used to train it. But these complex behaviours are a far cry from the simple stimulus-response relationships that we have considered thus far. How can reinforcement be used to create complex behaviours such as these?<\/p>\n<p>One way to expand the use of operant learning is to modify the schedule on which the reinforcement is applied. To this point we have only discussed a <strong>continuous reinforcement schedule<\/strong>, in which <em>the desired response is reinforced every time it occurs<\/em>; whenever the dog rolls over, for instance, it gets a biscuit. Continuous reinforcement results in relatively fast learning but also rapid extinction of the desired behaviour once the reinforcer disappears. The problem is that because the organism is used to receiving the reinforcement after every behaviour, the responder may give up quickly when it doesn\u2019t appear.<\/p>\n<p>Most real-world reinforcers are not continuous; they occur on a <strong>partial (or intermittent) reinforcement schedule <\/strong>\u2014 <em>a schedule in which the responses are sometimes reinforced and sometimes not<\/em>. In comparison to continuous reinforcement, partial reinforcement schedules lead to slower initial learning, but they also lead to greater resistance to extinction. Because the reinforcement does not appear after every behaviour, it takes longer for the learner to determine that the reward is no longer coming, and thus extinction is slower. The four types of partial reinforcement schedules are summarized in Table 8.2, &#8220;Reinforcement Schedules.&#8221;<\/p>\n<table>\n<caption>Table 8.2 Reinforcement Schedules.<\/caption>\n<thead>\n<tr>\n<td style=\"text-align: center\" colspan=\"3\"><a href=\"#skiptable8.2\">[Skip Table]<\/a><\/td>\n<\/tr>\n<tr>\n<th>Reinforcement schedule<\/th>\n<th>Explanation<\/th>\n<th>Real-world example<\/th>\n<\/tr>\n<\/thead>\n<tbody>\n<tr>\n<td>Fixed-ratio<\/td>\n<td>Behaviour is reinforced after a specific number of responses.<\/td>\n<td>Factory workers who are paid according to the number of products they produce<\/td>\n<\/tr>\n<tr>\n<td>Variable-ratio<\/td>\n<td>Behaviour is reinforced after an average, but unpredictable, number of responses.<\/td>\n<td>Payoffs from slot machines and other games of chance<\/td>\n<\/tr>\n<tr>\n<td>Fixed-interval<\/td>\n<td>Behaviour is reinforced for the first response after a specific amount of time has passed.<\/td>\n<td>People who earn a monthly salary<\/td>\n<\/tr>\n<tr>\n<td>Variable-interval<\/td>\n<td>Behaviour is reinforced for the first response after an average, but unpredictable, amount of time has passed.<\/td>\n<td>Person who checks email for messages<\/td>\n<\/tr>\n<\/tbody>\n<\/table>\n<p id=\"skiptable8.2\">Partial reinforcement schedules are determined by whether the reinforcement is presented on the basis of the time that elapses between reinforcement (interval) or on the basis of the number of responses that the organism engages in (ratio), and by whether the reinforcement occurs on a regular (fixed) or unpredictable (variable) schedule. In a <strong>fixed-interval schedule<\/strong>, <em>reinforcement occurs for the first response made after a specific amount of time has passed<\/em>. For instance, on a one-minute fixed-interval schedule the animal receives a reinforcement every minute, assuming it engages in the behaviour at least once during the minute. As you can see in Figure 8.6, &#8220;Examples of Response Patterns by Animals Trained under Different Partial Reinforcement Schedules,&#8221; animals under fixed-interval schedules tend to slow down their responding immediately after the reinforcement but then increase the behaviour again as the time of the next reinforcement gets closer. (Most students study for exams the same way.) In a <strong>variable-interval schedule<\/strong>, <em>the reinforcers appear on an interval schedule, but the timing is varied around the average interval, making the actual appearance of the reinforcer unpredictable<\/em>. An example might be checking your email: you are reinforced by receiving messages that come, on average, say, every 30 minutes, but the reinforcement occurs only at random times. Interval reinforcement schedules tend to produce slow and steady rates of responding.<\/p>\n<figure id=\"attachment_185\" aria-describedby=\"caption-attachment-185\" style=\"width: 400px\" class=\"wp-caption aligncenter\"><a href=\"http:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2013\/11\/555f0fdcff3a6dbbe56a86acd40f3463.jpg\"><img loading=\"lazy\" decoding=\"async\" src=\"http:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2013\/11\/555f0fdcff3a6dbbe56a86acd40f3463-1024x411.jpg\" alt=\"&quot;&quot;\" class=\"wp-image-185\" height=\"161\" width=\"400\" srcset=\"https:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2013\/11\/555f0fdcff3a6dbbe56a86acd40f3463-1024x411.jpg 1024w, https:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2013\/11\/555f0fdcff3a6dbbe56a86acd40f3463-300x120.jpg 300w, https:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2013\/11\/555f0fdcff3a6dbbe56a86acd40f3463-65x26.jpg 65w, https:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2013\/11\/555f0fdcff3a6dbbe56a86acd40f3463-225x90.jpg 225w, https:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2013\/11\/555f0fdcff3a6dbbe56a86acd40f3463-350x140.jpg 350w, https:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2013\/11\/555f0fdcff3a6dbbe56a86acd40f3463.jpg 1680w\" sizes=\"auto, (max-width: 400px) 100vw, 400px\" \/><\/a><figcaption id=\"caption-attachment-185\" class=\"wp-caption-text\">Figure 8.6 Examples of Response Patterns by Animals Trained under Different Partial Reinforcement Schedules. Schedules based on the number of responses (ratio types) induce greater response rate than do schedules based on elapsed time (interval types). Also, unpredictable schedules (variable types) produce stronger responses than do predictable schedules (fixed types).<\/figcaption><\/figure>\n<p>In a <strong>fixed-ratio schedule<\/strong>, <em>a behaviour is reinforced after a specific number of responses<\/em>. For instance, a rat\u2019s behaviour may be reinforced after it has pressed a key 20 times, or a salesperson may receive a bonus after he or she has sold 10 products. As you can see in Figure 8.6, &#8220;Examples of Response Patterns by Animals Trained under Different Partial Reinforcement Schedules,&#8221; once the organism has learned to act in accordance with the fixed-ratio schedule, it will pause only briefly when reinforcement occurs before returning to a high level of responsiveness. A <strong>variable-ratio schedule<\/strong>\u00a0<em>provides reinforcers after a specific but average number of responses<\/em>. Winning money from slot machines or on a lottery ticket is an example of reinforcement that occurs on a variable-ratio schedule. For instance, a slot machine (see Figure 8.7, &#8220;Slot Machine&#8221;) may be programmed to provide a win every 20 times the user pulls the handle, on average. Ratio schedules tend to produce high rates of responding because reinforcement increases as the number of responses increases.<\/p>\n<figure id=\"attachment_4371\" aria-describedby=\"caption-attachment-4371\" style=\"width: 400px\" class=\"wp-caption aligncenter\"><a href=\"http:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2014\/10\/Figure-8-7.jpg\"><img loading=\"lazy\" decoding=\"async\" src=\"http:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2014\/10\/Figure-8-7-1024x682.jpg\" alt=\"&quot;&quot;\" class=\"wp-image-4371\" height=\"267\" width=\"400\" srcset=\"https:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2014\/10\/Figure-8-7-1024x682.jpg 1024w, https:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2014\/10\/Figure-8-7-300x200.jpg 300w, https:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2014\/10\/Figure-8-7-65x43.jpg 65w, https:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2014\/10\/Figure-8-7-225x150.jpg 225w, https:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2014\/10\/Figure-8-7-350x233.jpg 350w, https:\/\/opentextbc.ca\/introductiontopsychology\/wp-content\/uploads\/sites\/9\/2014\/10\/Figure-8-7.jpg 1599w\" sizes=\"auto, (max-width: 400px) 100vw, 400px\" \/><\/a><figcaption id=\"caption-attachment-4371\" class=\"wp-caption-text\">Figure 8.7 Slot Machine. Slot machines are examples of a variable-ratio reinforcement schedule.<\/figcaption><\/figure>\n<p>Complex behaviours are also created through <strong>shaping<\/strong>, <em>the process of guiding an organism\u2019s behaviour to the desired outcome through the use of successive approximation to a final desired behaviour<\/em>. Skinner made extensive use of this procedure in his boxes. For instance, he could train a rat to press a bar two times to receive food, by first providing food when the animal moved near the bar. When that behaviour had been learned, Skinner\u00a0would begin to provide food only when the rat touched the bar. Further shaping limited the reinforcement to only when the rat pressed the bar, to when it pressed the bar and touched it a second time, and finally to only when it pressed the bar twice. Although it can take a long time, in this way operant conditioning can create chains of behaviours that are reinforced only when they are completed.<\/p>\n<p>Reinforcing animals if they correctly discriminate between similar stimuli allows scientists to test the animals\u2019 ability to learn, and the discriminations that they can make are sometimes remarkable. Pigeons have been trained to distinguish between images of Charlie Brown and the other Peanuts characters (Cerella, 1980),\u00a0and between different styles of music and art (Porter &amp; Neuringer, 1984; Watanabe, Sakamoto &amp; Wakita, 1995).<\/p>\n<p>Behaviours can also be trained through the use of <em>secondary reinforcers<\/em>. Whereas a <strong>primary reinforcer<\/strong>\u00a0includes <em>stimuli that are naturally preferred or enjoyed by the organism, such as food, water, and relief from pain<\/em>, a <strong>secondary reinforcer<\/strong>\u00a0(sometimes called <em>conditioned reinforcer<\/em>) is a <em>neutral event that has become associated with a primary reinforcer through classical conditioning<\/em>. An example of a secondary reinforcer would be the whistle given by an animal trainer, which has been associated over time with the primary reinforcer, food. An example of an everyday secondary reinforcer is money. We enjoy having money, not so much for the stimulus itself, but rather for the primary reinforcers (the things that money can buy) with which it is associated.<\/p>\n<div class=\"bcc-box bcc-success\">\n<h3>Key Takeaways<\/h3>\n<ul>\n<li>Edward Thorndike developed the law of effect: the principle that responses that create a typically pleasant outcome in a particular situation are more likely to occur again in a similar situation, whereas responses that produce a typically unpleasant outcome are less likely to occur again in the situation.<\/li>\n<li>B. F. Skinner expanded on Thorndike\u2019s ideas to develop a set of principles to explain operant conditioning.<\/li>\n<li>Positive reinforcement strengthens a response by presenting something that is typically pleasant after the response, whereas negative reinforcement strengthens a response by reducing or removing something that is typically unpleasant.<\/li>\n<li>Positive punishment weakens a response by presenting something typically unpleasant after the response, whereas negative punishment weakens a response by reducing or removing something that is typically pleasant.<\/li>\n<li>Reinforcement may be either partial or continuous. Partial reinforcement schedules are determined by whether the reinforcement is presented on the basis of the time that elapses between reinforcements (interval) or on the basis of the number of responses that the organism engages in (ratio), and by whether the reinforcement occurs on a regular (fixed) or unpredictable (variable) schedule.<\/li>\n<li>Complex behaviours may be created through shaping, the process of guiding an organism\u2019s behaviour to the desired outcome through the use of successive approximation to a final desired behaviour.<\/li>\n<\/ul>\n<\/div>\n<div class=\"bcc-box bcc-info\">\n<h3>Exercises and Critical Thinking<\/h3>\n<ol>\n<li>Give an example from daily life of each of the following: positive reinforcement, negative reinforcement, positive punishment, negative punishment.<\/li>\n<li>Consider the reinforcement techniques that you might use to train a dog to catch and retrieve a Frisbee that you throw to it.<\/li>\n<li>Watch the following two videos from current television shows. Can you determine which learning procedures are being demonstrated?\n<ol>\n<li><a href=\"http:\/\/www.break.com\/usercontent\/2009\/11\/the-office-altoid-experiment-1499823\"><em>The Office<\/em><\/a>: http:\/\/www.break.com\/usercontent\/2009\/11\/the-office-altoid- experiment-1499823<\/li>\n<li><a href=\"http:\/\/www.youtube.com\/watch?v=JA96Fba-WHk\"><em>The Big Bang Theory <\/em>[YouTube]<\/a>: http:\/\/www.youtube.com\/watch?v=JA96Fba-WHk<\/li>\n<\/ol>\n<\/li>\n<\/ol>\n<\/div>\n<h2>References<\/h2>\n<p>Cerella, J. (1980). The pigeon\u2019s analysis of pictures.\u00a0<i>Pattern Recognition, 12<\/i>, 1\u20136.<\/p>\n<p>Kassin, S. (2003). <a href=\"http:\/\/wps.prenhall.com\/hss_kassin_essentials_1\/15\/3933\/1006917.cw\/index.html\">Essentials of psychology<\/a>. Upper Saddle River, NJ: Prentice Hall. Retrieved from Essentials of Psychology Prentice Hall Companion Website: http:\/\/wps.prenhall.com\/hss_kassin_essentials_1\/15\/3933\/1006917.cw\/index.html<\/p>\n<p>Porter, D., &amp; Neuringer, A. (1984). Music discriminations by pigeons.\u00a0<i>Journal of Experimental Psychology: Animal Behavior Processes, 10<\/i>(2), 138\u2013148.<\/p>\n<p>Thorndike, E. L. (1898).\u00a0<i>Animal intelligence: An experimental study of the associative processes in animals.<\/i>\u00a0Washington, DC: American Psychological Association.<\/p>\n<p>Thorndike, E. L. (1911).\u00a0<i><a href=\"http:\/\/www.archive.org\/details\/animalintelligen00thor\">Animal intelligence: Experimental studies<\/a>.<\/i>\u00a0New York, NY: Macmillan. Retrieved from\u00a0http:\/\/www.archive.org\/details\/animalintelligen00thor<\/p>\n<p>Watanabe, S., Sakamoto, J., &amp; Wakita, M. (1995). Pigeons\u2019 discrimination of painting by Monet and Picasso.\u00a0<i>Journal of the Experimental Analysis of Behaviour, 63<\/i>(2), 165\u2013174.<\/p>\n<h2>Image Attributions<\/h2>\n<p><strong>Figure 8.5:<\/strong> &#8220;<a href=\"http:\/\/en.wikipedia.org\/wiki\/File:Skinner_box_photo_02.jpg\">Skinner box<\/a>&#8221; (http:\/\/en.wikipedia.org\/wiki\/File:Skinner_box_photo_02.jpg) is licensed under the CC BY SA 3.0 license (http:\/\/creativecommons.org\/licenses\/by-sa\/3.0\/deed.en). &#8220;<a href=\"http:\/\/en.wikipedia.org\/wiki\/File:Skinner_box_scheme_01.png\">Skinner box scheme<\/a>&#8221; by Andreas1 (http:\/\/en.wikipedia.org\/wiki\/File:Skinner_box_scheme_01.png) is licensed under the CC BY SA 3.0 license (http:\/\/creativecommons.org\/licenses\/by-sa\/3.0\/deed.en)<\/p>\n<p><strong>Figure 8.6:<\/strong> Adapted from Kassin (2003).<\/p>\n<p><strong>Figure 8.7:<\/strong>\u00a0 &#8220;<a href=\"http:\/\/commons.wikimedia.org\/wiki\/File:HardRockCasinoSlotMachines.jpg\">Slot Machines in the Hard Rock Casino<\/a>&#8221; by Ted Murpy (http:\/\/commons.wikimedia.org\/wiki\/File:HardRockCasinoSlotMachines.jpg) is licensed under <a href=\"http:\/\/creativecommons.org\/licenses\/by\/2.0\/deed.en\">CC BY 2.0.<\/a> (http:\/\/creativecommons.org\/licenses\/by\/2.0\/deed.en).<\/p>\n","protected":false},"author":1,"menu_order":1,"template":"","meta":{"pb_show_title":"on","pb_short_title":"","pb_subtitle":"","pb_authors":[],"pb_section_license":""},"chapter-type":[],"contributor":[],"license":[],"class_list":["post-187","chapter","type-chapter","status-publish","hentry"],"part":447,"_links":{"self":[{"href":"https:\/\/opentextbc.ca\/introductiontopsychology\/wp-json\/pressbooks\/v2\/chapters\/187","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/opentextbc.ca\/introductiontopsychology\/wp-json\/pressbooks\/v2\/chapters"}],"about":[{"href":"https:\/\/opentextbc.ca\/introductiontopsychology\/wp-json\/wp\/v2\/types\/chapter"}],"author":[{"embeddable":true,"href":"https:\/\/opentextbc.ca\/introductiontopsychology\/wp-json\/wp\/v2\/users\/1"}],"version-history":[{"count":27,"href":"https:\/\/opentextbc.ca\/introductiontopsychology\/wp-json\/pressbooks\/v2\/chapters\/187\/revisions"}],"predecessor-version":[{"id":4633,"href":"https:\/\/opentextbc.ca\/introductiontopsychology\/wp-json\/pressbooks\/v2\/chapters\/187\/revisions\/4633"}],"part":[{"href":"https:\/\/opentextbc.ca\/introductiontopsychology\/wp-json\/pressbooks\/v2\/parts\/447"}],"metadata":[{"href":"https:\/\/opentextbc.ca\/introductiontopsychology\/wp-json\/pressbooks\/v2\/chapters\/187\/metadata\/"}],"wp:attachment":[{"href":"https:\/\/opentextbc.ca\/introductiontopsychology\/wp-json\/wp\/v2\/media?parent=187"}],"wp:term":[{"taxonomy":"chapter-type","embeddable":true,"href":"https:\/\/opentextbc.ca\/introductiontopsychology\/wp-json\/pressbooks\/v2\/chapter-type?post=187"},{"taxonomy":"contributor","embeddable":true,"href":"https:\/\/opentextbc.ca\/introductiontopsychology\/wp-json\/wp\/v2\/contributor?post=187"},{"taxonomy":"license","embeddable":true,"href":"https:\/\/opentextbc.ca\/introductiontopsychology\/wp-json\/wp\/v2\/license?post=187"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}