"

Unit 2 Key Terms

Psychology of Learning

Unit 2: Operant Conditioning & Choice

Key Terms

Accumulative Record: A graph that shows the cumulative number of responses over time; the slope indicates response rate, with steeper slopes showing higher rates.

Algorithm: A methodical rule that guarantees that a problem will be solved; provides systematic procedures that, if followed correctly, always produce correct solutions.

Anchoring: A bias in decision-making whereby our judgments are influenced by a reference point which is given, even when that reference point is arbitrary & irrelevant.

Area Under the Curve (AUC): An empirical measure of discounting that plots subjective value against delay; larger AUC indicates less steep discounting (more self-control).

Arousal: A general state of physiological & psychological activation ranging from deep sleep to high excitement; affects performance according to the Yerkes-Dodson law.

Automatic Reinforcement: Reinforcement that occurs when the behavior itself is the reinforcing stimulus.

Availability Heuristic: A mental shortcut which makes us believe that things that come easily to mind occur more often than things that do not come easily to mind.

Avoidance: A type of negative reinforcement in which the organism’s response prevents an aversive stimulus from occurring.

Avoidance Learning: Learning to make a response that prevents an aversive stimulus from occurring; maintained by negative reinforcement even though the aversive stimulus is never experienced.

Backward Chaining: A method of teaching a behavioral chain by starting with the last step & working backward; the learner first masters the final response closest to reinforcement.

Baseline: The normal rate of behavior before any experimental manipulation; used as a comparison point for measuring behavior change.

Behavior Modification: The application of learning principles, particularly operant conditioning, to change behavior in applied settings.

Behavioral Contrast: A change in response rate in one component of a multiple schedule that occurs when reinforcement conditions change in the other component.

Behavioral Economics: The study of how organisms allocate their limited resources, including time & money; combines principles from behavioral psychology with microeconomics to predict decision-making.

Bias: In the Generalized Matching Law, a parameter (b) representing a consistent preference for one alternative independent of reinforcement rates.

Bounded Rationality: The concept that human rationality is limited by cognitive constraints including finite information-processing capacity, limited working memory, & time pressures.

Break-and-Run Pattern: A response pattern on fixed-ratio schedules characterized by a pause after reinforcement followed by a high, steady rate of responding until the next reinforcement.

Certainty Effect: The tendency to overweight certain outcomes relative to merely probable outcomes, even when expected values favor the uncertain option.

Chaining: A procedure for establishing a sequence of behaviors in which each response produces a stimulus that serves as both a conditioned reinforcer for the preceding response & a discriminative stimulus for the next response.

Choice: A form of decision-making which involves selecting among alternatives; differs from judgment in requiring selection rather than estimation.

Choice Architecture: The design of environments & contexts in which people make decisions; modifying choice architecture is more effective than relying on willpower.

Closed-Loop Movement: Movement that relies on continuous sensory feedback for guidance & correction; typically slower movements that can be adjusted during execution.

Compensatory Strategies: Decision-making strategies that involve trading low values on one dimension for high values on another; weaknesses on some dimensions can be offset by strengths on others.

Concurrent Schedule: A reinforcement schedule in which two or more schedules are available simultaneously, each associated with a different response.

Continuous Motor Skills: Motor skills with no distinct beginning or end point, such as swimming, running, or steering; require ongoing adjustments throughout performance.

Continuous Reinforcement (CRF): A schedule in which every correct response is reinforced; also called FR-1.

Cumulative Recorder: A device that produces a continuous record of responding over time by moving paper at a constant rate while a pen moves upward with each response.

Decision Tree: A visual tool that maps out all possible choices, outcomes, & their consequences, helping decision-makers systematically evaluate options.

Decision-Making Under Certainty: Decisions made in which the factors that determine the outcomes of the different choice alternatives are known with complete accuracy.

Decision-Making Under Risk: Decisions made in which the factors that determine the outcomes of the different choice alternatives occur with known probabilities.

Decision-Making Under Uncertainty: Decisions made in which the probabilities of the different factors which affect the outcomes of the different choice alternatives are not known with precision or cannot be determined.

Delay Discounting: The decrease in the subjective value of a reward as the delay to receiving that reward increases; also called temporal discounting.

Descriptive Models of Decision-Making: Models that address how humans actually make decisions, describing observed behavior rather than prescribing optimal behavior.

Differential Reinforcement: Reinforcing responses that meet a specific criterion while withholding reinforcement for responses that do not meet the criterion.

Discrete Motor Skills: Motor skills with a recognizable beginning & end point, such as throwing a ball or pressing a button; can be performed as single units.

Discriminative Stimulus (SD): A stimulus that signals the availability of reinforcement for a particular response; sets the occasion for responding.

Distributed Practice: Practice sessions that are spaced apart with rest intervals between them; generally produces better long-term retention than massed practice.

Drive Reduction Theory: Hull’s theory that reinforcement works by reducing biological drives; problematic because some reinforcers increase arousal rather than reduce drives.

Ego Depletion: The state in which willpower has been temporarily depleted through prior self-control efforts, making subsequent self-control attempts more likely to fail.

Elastic Demand: In behavioral economics, consumption that decreases substantially when price increases; indicates the reinforcer has available substitutes.

Electrical Stimulation of the Brain (ESB): Direct stimulation of brain regions that can serve as a powerful reinforcer; rats will cross electrified floors or forgo food to obtain ESB.

Escape: A type of negative reinforcement in which the organism’s response terminates an ongoing aversive stimulus.

Expected Utility: Calculated by multiplying the probability of each outcome times the utility (value) of that outcome, then summing across all outcomes.

External Imagery: Mental imagery from a third-person perspective, as if watching yourself perform on video; emphasizes visual form & technique.

Extinction: The procedure of withholding reinforcement for a previously reinforced response, resulting in a decrease in the response.

Extinction Burst: A temporary increase in the frequency, duration, or intensity of a response when extinction is first implemented.

Fast Strategy: A life history strategy adaptive in high mortality, unstable environments characterized by early reproduction, steep delay discounting, & high risk-taking.

Fast-and-Frugal Trees: Shortcuts for decision-making involving a series of yes/no questions to come to a decision very quickly; use minimal information for sequential binary decisions.

Fixed Action Pattern: An innate, stereotyped behavior sequence triggered by a specific stimulus; can interfere with operant conditioning when tokens trigger instinctive behaviors.

Fixed Interval (FI): A schedule in which reinforcement is delivered for the first response after a fixed period of time has elapsed since the last reinforcement.

Fixed Ratio (FR): A schedule in which reinforcement is delivered after a fixed number of responses.

Forward Chaining: A method of teaching a behavioral chain by starting with the first step & proceeding to later steps; the learner first masters the initial response in the sequence.

Framing: A bias in decision-making whereby our decisions are influenced by the way a question is asked or a choice is presented.

Future-Uncertainty: Uncertainty about the long-term consequences of behavior; one of two types of uncertainty (along with present-uncertainty) that reduce prudent decision-making.

Gambler’s Fallacy: A bias in judgment whereby our estimations of probability are heavily influenced by the most recent outcomes, leading us to believe past outcomes affect future independent events.

Generalized Matching Law: Baum’s extension of the Matching Law that incorporates bias & sensitivity parameters to account for systematic deviations from perfect matching.

Heuristic: A mental shortcut that helps decision-makers make judgments & choices quickly in the face of uncertainty; works from partial information.

Hyperbolic Equation: The mathematical function V = A / (1 + kD) that best describes delay discounting, where V = discounted value, A = actual value, k = discount rate, & D = delay.

Impulsivity: Preference for smaller-immediate rewards over larger-delayed rewards; demonstrated in paradigms where organisms choose immediate but smaller reinforcement.

Indifference Points: Combinations of options where someone is equally likely to choose either alternative; reveal how steeply a person discounts future rewards.

Inelastic Demand: In behavioral economics, consumption that remains relatively constant despite price increases; indicates the reinforcer has few substitutes or is essential.

Information Cascade: A situation where sequential decision-makers ignore their own information & copy predecessors, potentially propagating errors through populations.

Information Overload: A state where too much information is available, exceeding cognitive capacity to process it effectively, leading to poorer decisions.

Instinctive Drift: The tendency for learned behaviors to drift toward instinctive behaviors over time; demonstrated by Breland & Breland with raccoons & pigs.

Intermittent Reinforcement: A schedule in which only some responses are reinforced; also called partial reinforcement.

Internal Imagery: Mental imagery from a first-person perspective, experiencing the kinesthetic sensations of performing the movement; emphasizes feeling the body move.

Intuition: A type of thinking that involves understanding that is quick & effortless & often involves insight; generally, people cannot verbalize the thought processes.

Inverted-U Theory: The theory that performance is optimal at moderate levels of arousal & deteriorates at both low & high arousal levels; also called the Yerkes-Dodson law.

Ironic Errors Theory: Wegner’s theory that trying not to make a specific error can increase the likelihood of making that error, especially under pressure or cognitive load.

Judgment: A type of decision-making in which one estimates the likelihood of an event; examples include estimating probability of rain or assessing chances of success.

Knowledge of Performance (KP): Feedback about the movement pattern or technique used to produce a response; information about how the movement was executed.

Knowledge of Results (KR): Feedback about the outcome of a movement in terms of the environmental goal; information about whether the goal was achieved.

Latent Learning: Learning that occurs without obvious reinforcement & is not demonstrated until reinforcement is introduced; demonstrated by Tolman & Honzik with rats in mazes.

Law of Diminishing Marginal Value: The principle that each additional unit of a commodity provides less additional utility than the previous unit; explains why variety in choices can maximize total utility.

Life History Theory (LHT): An analytical framework designed to study the diversity of life history strategies; predicts how organisms should allocate limited resources among competing demands.

Local Reinforcement Rate: The reinforcement rate currently being received from a particular alternative, as opposed to the overall or average rate.

Loss Aversion: The phenomenon where people are more motivated to avoid losses than to acquire equivalent gains; a loss of $1000 is more painful than a gain of $1000 is joyful.

Massed Practice: Practice sessions that are continuous with little or no rest between trials; can lead to fatigue & poorer long-term retention.

Matching Law: The principle that in a two-choice situation, the percentage of responses made on each alternative will match the percentage of reinforcers received from each alternative.

Maximum Utility Theory: The economic principle that rational consumers make choices to maximize total utility or satisfaction from their resources.

Melioration: A theory proposing that organisms shift behavior toward whichever alternative provides better local reinforcement rates, even when this doesn’t maximize global outcomes.

Mental Imagery: Cognitive rehearsal of a movement without physical execution; involves activation of motor areas in the brain & can improve performance.

Momentary Maximization Theory: A theory of choice that argues organisms make choices to maximize their satisfaction at the present moment rather than considering long-term consequences.

Motor Programs: Pre-structured sets of motor commands that can be executed without continuous feedback; allow rapid, automatic movement sequences.

Motor Trace: In Adams’s theory, the memory representation responsible for initiating movement; selects & initiates the motor commands.

Movement Sequences: Series of movements performed in a fixed order with specific timing; can be cyclical (walking) or non-cyclical (typing a word).

Mowrer’s Two-Factor Theory: The theory that avoidance learning involves two processes: classical conditioning creates fear of warning stimuli, & operant conditioning maintains avoidance responses that reduce fear.

Multiple Schedule: A schedule in which two or more schedules alternate, each associated with a different discriminative stimulus.

Need Reduction Theory: The theory that reinforcement works by satisfying biological needs; problematic because saccharin reinforces without satisfying nutritional needs.

Negative Punishment: The removal of a stimulus following a response that decreases the future probability of that response; also called omission training.

Negative Reinforcement: The removal of an aversive stimulus following a response that increases the future probability of that response.

Negative Transfer: When prior learning interferes with new learning; occurs when old & new tasks require different responses to similar stimuli.

Normative Models of Decision-Making: Models that address how humans should make decisions to be rational & maximize expected outcomes; establish standards for optimal decision-making.

Open-Loop Movement: Movement that is executed without ongoing sensory feedback; typically rapid, ballistic movements that cannot be corrected once initiated.

Operant Behavior: Behavior that operates on the environment to produce consequences; behavior that is controlled by its consequences.

Operant Conditioning: A type of learning in which behavior is modified by its consequences; responses that produce favorable consequences are strengthened while those that produce unfavorable consequences are weakened.

Optimal Foraging: Behavior patterns in which animals make choices that maximize their net resource intake per unit time; involves balancing current site quality against search costs.

Optimization Theory: A theory of choice behavior which assumes that consumers spend their resources in ways that maximize their utility.

Overconfidence: A bias in decision-making whereby we are more confident than correct; an overestimation of the accuracy of one’s beliefs & judgments.

Overmatching: A deviation from matching in which animals show a stronger preference for the better alternative than matching predicts; occurs when sensitivity parameter a > 1.

Partial Reinforcement Effect (PRE): The finding that behaviors reinforced on intermittent schedules are more resistant to extinction than behaviors reinforced continuously.

Perceptual Trace: In Adams’s theory, the memory representation of what correct performance feels like; serves as a reference standard for error detection.

Positive Punishment: The presentation of an aversive stimulus following a response that decreases the future probability of that response.

Positive Reinforcement: The presentation of a stimulus following a response that increases the future probability of that response.

Positive Transfer: When prior learning facilitates new learning; occurs when old & new tasks share similar stimulus-response relationships.

Post-Reinforcement Pause (PRP): A pause in responding that occurs after reinforcement delivery, particularly on fixed schedules.

Precommitment: A situation where the individual makes a decision well before the actual time where consequences would be given & removes the ability to change that decision later.

Preference Reversals: Changes in what a person prefers as time passes, even though objective consequences remain constant; explains why people plan well but execute poorly.

Premack Principle: The principle that high-probability behaviors can reinforce low-probability behaviors; a more preferred activity can serve as a reinforcer for a less preferred activity.

Present-Uncertainty: Uncertainty about the immediate consequences of behavior; one of two types of uncertainty (along with future-uncertainty) that reduce prudent decision-making.

Prestige Bias: The tendency to copy choices of successful, high-status individuals.

Probability Discounting: The loss of value of a reinforcer that is to be received only probabilistically; a certain reward is worth more than an uncertain reward of equal objective value.

Probability of Reciprocation: The subjective likelihood that one’s cooperative behavior will be matched by another (in social cooperation) or that present sacrifice will yield future benefit (in self-control); the crucial variable distinguishing self-control from social cooperation dilemmas.

Probability Theory: A branch of mathematics that deals with random events & allows for the best average prediction to be made; provides formal methods for reasoning about uncertain outcomes.

Progressive Ratio Schedule: A schedule in which the ratio requirement increases systematically after each reinforcement; used to assess motivation or reinforcer efficacy.

Prospect Theory: A descriptive theory of choice under risk showing that people evaluate options relative to reference points & weight losses more heavily than equivalent gains.

Prudent Decision-Making: Decision-making driven by long-term consequences rather than immediate outcomes; essentially synonymous with self-control in the context of repeated choices.

Pseudocertainty Effect: Treating conditional certainty (certainty within an uncertain scenario) as if it were absolute certainty.

Punishment: A consequence that decreases the future probability of the behavior it follows.

Rapid Adjusting Procedure (Titration): A procedure that systematically adjusts the value of one option based on previous choices until an indifference point is found.

Recognition Heuristic: A mental shortcut that biases us towards choosing those things that are familiar to us, assuming familiarity indicates quality or importance.

Reflection Effect: The reversal of risk preferences between gain & loss domains—risk aversion for gains but risk seeking for losses.

Regret Theory: A descriptive model of decision-making that assumes decision-makers experience feelings of regret & try to avoid this feeling when making decisions.

Reinforcement: A consequence that increases the future probability of the behavior it follows.

Reinforcement Relativity: The principle that whether an activity serves as a reinforcer depends on its probability relative to other available activities, not on absolute properties.

Reinforcement Schedule: A rule specifying which responses will be reinforced; determines the pattern & timing of reinforcement delivery.

Representativeness Heuristic: A mental shortcut used to assess probability whereby the more an object appears to represent a class of objects, the more likely we judge it to belong to that class.

Response Chain: A sequence of behaviors in which each response produces stimuli that serve as discriminative stimuli for the next response in the chain.

Response Deprivation Theory: The theory that restricting access to any activity below its baseline level creates conditions for that activity to serve as a reinforcer.

Reward Deficiency Syndrome: A theory suggesting that some individuals have underactive reward systems, leading them to seek excessive stimulation through addictive behaviors.

Satisficing: A theory of decision-making arguing that decision-makers work to satisfy basic needs & achieve acceptable outcomes rather than searching exhaustively for the absolute best option.

Scallop Pattern: A response pattern on fixed-interval schedules characterized by a pause after reinforcement followed by gradually accelerating responding as the interval elapses.

Schema Theory: Schmidt’s theory that motor learning involves developing generalized rules (schemas) relating movement parameters to outcomes, enabling transfer to novel movements.

Self-Control: Control of behavior using internal controls rather than external controls; involves choosing larger-later rewards over smaller-sooner rewards.

Sensitivity: In the Generalized Matching Law, a parameter (a) representing how responsive the organism is to differences in reinforcement rates.

Shaping: The reinforcement of successive approximations to a target behavior; used to establish behaviors not currently in the organism’s repertoire.

Slow Strategy: A life history strategy adaptive in low mortality, stable environments characterized by delayed reproduction, shallow delay discounting, & risk aversion.

Social Cooperation Problems: Decision situations where one choice is easy (often selfish) & another is harder but offers superior long-term consequences for the community.

Social Discounting: The phenomenon where people care less about socially distant others than about themselves & close others; correlated with temporal discounting.

Social Learning: Acquiring information about options, outcomes, & strategies by observing others rather than through personal direct experience.

Social Proof: The tendency to infer that popular choices are good choices based on others’ behavior.

Spontaneous Recovery: The reappearance of an extinguished response after a period of time following extinction.

Stochastic Models of Choice: Models that treat preferences as though they have a random component, allowing understanding of how a person can prefer different options on different occasions.

Subjective Utility Theory: An extension of expected utility theory used when probabilities cannot be determined in advance; decision-makers estimate subjective probabilities based on beliefs & experience.

Sunk Cost Fallacy: The tendency to continue investing in something because of past investments that cannot be recovered, rather than basing decisions solely on future costs & benefits.

Sunk Costs: Past investments (time, money, effort) that have already been incurred & cannot be recovered regardless of future decisions.

Superstitious Behavior: Behavior that is accidentally reinforced by coincidental pairing with reinforcement, though there is no causal relationship between the behavior & the reinforcer.

Take-the-Best Heuristic: A mental shortcut based on a single good reason, ignoring all other information; identify the most important cue & make your decision based on that cue alone.

Three-Term Contingency: The basic unit of operant analysis consisting of the discriminative stimulus (SD), the response (R), & the consequence (SR); also called the ABCs of behavior.

Time Horizons: Points in time in the future beyond which we do not understand or plan; the ultimate form of delay discounting where consequences fail to be represented.

Tragedy of the Commons: A conflict where each individual, seeking their own selfish best interest, utilizes common resources in a way that is ultimately destructive to the community as a whole.

Transfer of Training: The influence of prior learning on new learning; can be positive (facilitating) or negative (interfering).

Type A Personality: A personality pattern characterized by competitiveness, time urgency, hostility, & achievement striving; associated with maximizing decision strategies.

Type B Personality: A personality pattern characterized by a relaxed, patient, & easygoing approach; associated with satisficing decision strategies.

Undermatching: A deviation from matching in which animals express relative indifference to the alternatives in a two-choice procedure; occurs when sensitivity parameter a < 1.

Universal Human Morality: A human condition described by C.S. Lewis shared by all individuals; an internal control which helps keep selfish behavior in check.

Utilitarianism: A theory of morality that argues the best choice is always the one which maximizes the utility of all people that will be affected by the choice.

Utility: A measure used in economics to describe the subjective satisfaction, value, or benefit gained from consuming a good or service; in psychology, represents reinforcing value.

Variable Interval (VI): A schedule in which reinforcement is delivered for the first response after a variable period of time has elapsed, averaging a specified duration.

Variable Ratio (VR): A schedule in which reinforcement is delivered after a variable number of responses, averaging a specified number.

Willpower: The cognitive resource required for self-regulation; like a muscle that gets fatigued with use but can be strengthened through regular exercise.

Yerkes-Dodson Law: The principle that optimal performance occurs at moderate arousal levels, with the optimal level varying based on task complexity.

License

Psychology of Learning TxWes Copyright © by Jay Brown. All Rights Reserved.