Discriminative stimulus training and selective stimulus control in rats

Authors Avatar
Discriminative stimulus training and selective stimulus control in rats.

Name: Mark Miller

ID number: 3330061

Course: PSYC 111

Lab Group: 103

Tutor: Jacqueline Harris

Date Due: 9 May 2002

Abstract

The aim of the experiment was to show that rats demonstrated stimulus discrimination and selective stimulus control during operant conditioning. The first hypothesis was the subject would learn to discriminate between the VR16 conditions that signal reinforcement and the EXT conditions. It was also hypothesised that the stimulus used to discriminate between VR16 and EXT would either be the light or the tone, not a combination.

The participant in this experiment was a 16-month-old, female, Spague - Dawley albino rat that was randomly selected from a group of 20. The apparatus used was an operant chamber, which delivered two stimuli (a light and a tone) to the subject, and a reinforcer of diluted condensed milk. During the first week of experimentation the subject underwent discrimination training, this was followed by a series of probe trials in the second week.

The results from the first week showed the subject learned that no reinforcement was given during EXT, because the rate of responding decreased. The second week's results showed that high tone was the stimulus used to discriminate between the stimuli. These results supported both the hypotheses, and it was concluded that rats do demonstrate stimulus discrimination and selective stimulus control.

The major theorists for the development of operant conditioning were Edward Thorndike (1910), John Watson (1914), and Burrhus Skinner (1938) (Huitt and Hummel, 1997). They proposed that learning is the result of the application of consequences following overt behaviour; that is, subjects begin to connect certain responses with certain stimuli. This led Thorndike to conclude that the probability of a specific response reoccurring is changed according to the consequences following the response, and he labelled this learning conditioning (Carlson and Buskist 1997, Huitt and Hummel, 1997).

In 1910, Thorndike used the notion of 'consequences' to teach cats and dogs to manipulate a latch in a "puzzle-box", to activate a door and escape (Huitt and Hummel, 1997). The consequence was either punishment or reward (Carlson and Buskist, 1997). Thorndike measured the time it took the animal to escape over various trials, and over time he noted that the animals latency to escape decreased consistently until it would activate the lever immediately after being placed in the box (Huitt and Hummel, 1997). The reward of being freed from the box somehow strengthened the association between the stimulus of being in the box and the appropriate action (Huitt and Hummel, 1997). Thorndike concluded that the reward strengthened the stimulus-response associations (Carlson and Buskist, 1997). He then went on to formulate his "law of effect", which can be summarised by saying that an animal is more likely to repeat a response if the result is favourable, and less likely to repeat the action if the consequences were not favourable (Carlson and Buskist, 1997). There were two possible consequences of a behaviour, reinforcement or punishment. These could be divided into two sub-categories, positive (sometimes called pleasant) and negative (sometimes called aversive). These could be added to or taken away from the environment in order to change the probability of a given response occurring again (Carlson and Buskist, 1997. Werzburg University). Punishment decreases the repetition of behaviour and reinforcement usually increases the likelihood of response being repeated.

A stimulus that acts as an indicator to the subject, suggesting that a reinforcer is available is said to be a discriminative stimulus (Gleitman, 1995). A discriminative stimulus affects the subjects' behaviour considerably (Gleitman, 1995), as it influences the likelihood of a response occurring (Carlson and Buskist, 1997). Reynolds (1961) conducted experiments where two pigeons learned to tap a red key with a white triangle. To determine which was the discriminative stimulus, he tested the two birds with either a plain red key or a plain key with just a white triangle. Reynolds (1961) found that the first bird used the red key as the discriminative stimulus and the second bird used the white triangle to discriminate between stimuli. This experiment is also an example of selective stimulus control, where each pigeon selected which stimulus it believed was responsible for producing the reinforcer.
Join now!


To effectively study how a subject behaves in a given environment and to certain stimuli, it was necessary to establish a schedule of reinforcement, which is a set of guidelines saying how often the subject is reinforced (Gleitman, 1995). Stimuli could be presented to the environment according to a schedule of which there were two categories: continuous and intermittent (Gleitman, 1995), or not at all using extinction. Continuous reinforcement simply means that the behaviour is followed by a consequence each time it occurs. Intermittent schedules were based either on the passage of time (interval schedules) or the number ...

This is a preview of the whole essay