Hostname: page-component-cd9895bd7-p9bg8 Total loading time: 0 Render date: 2024-12-23T05:12:29.745Z Has data issue: false hasContentIssue false

External validity of individual differences in multiple cue probability learning: The case of pilot training

Published online by Cambridge University Press:  01 January 2023

Nadine Matton*
Affiliation:
ENAC/University of Toulouse, France. ENAC, 7 Avenue Edouard Belin, 31055 Toulouse Cedex 4, France
Éric Raufaste
Affiliation:
University of Toulouse, France
Stéphane Vautier
Affiliation:
University of Toulouse, France
*
Rights & Permissions [Opens in a new window]

Abstract

Individuals differ in their ability to deal with unpredictable environments. Could impaired performances on learning an unpredictable cue-criteria relationship in a laboratory task be associated with impaired learning of complex skills in a natural setting? We focused on a multiple-cue probability learning (MCPL) laboratory task and on the natural setting of pilot training. We used data from three selection sessions and from the three corresponding selected pilot student classes of a national airline pilot selection and training system. First, applicants took an MCPL task at the selection stage (N = 556; N = 701; N = 412). Then, pilot trainees selected from the applicant pools (N = 44; N = 60; N = 28) followed the training for 2.5 to 3 yrs. Differences in final MCPL performance were associated with pilot training difficulties. Indeed, poor MCPL performers experienced almost twice as many pilot training difficulties as better MCPL performers (44.0% and 25.0%, respectively).

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BY
The authors license this article under the terms of the Creative Commons Attribution 3.0 License.
Copyright
Copyright © The Authors [2013] This is an Open Access article, distributed under the terms of the Creative Commons Attribution license (http://creativecommons.org/licenses/by/3.0/), which permits unrestricted re-use, distribution, and reproduction in any medium, provided the original work is properly cited.

1 Introduction

Individuals permanently have to learn to adapt to non-deterministic environments. The weather, stock exchange shares, presidential elections or the efficacy of medical care depend on so many factors that they cannot be considered as totally predictable. However, some people are less efficient than others in dealing with noisy environments. Indeed most education systems put only little emphasis on learning to cope with unpredictability or with noisy information. Therefore it is not surprising to hear from examples of mathematicians who persist in taking suboptimal stock market decisions (see the example of John Allen Paulos cited by Stanovich, 2009). Could one detect such difficulties in dealing with uncertainty in real life with a laboratory cognitive task simulating learning in an unpredictable environment?

The field of aviation is specially illustrative of situations involving the ability to deal with unpredictable events. For pilots especially, the necessary skills cannot be learned solely by explicit instruction and by acquisition of declarative knowledge. In particular, acquiring flying skills involves learning to infer relationships between cues (nature of clouds, wind force, physiological sensations, visual cues of surrounding environment..) and criteria (aircraft speed, altitude,...) through repeated experiences. Pilot students have to learn to infer which cues are positively or negatively related to aircraft attitude and which cues are irrelevant in a given situation. Some pilot students need more flying hours than others, and some of them never complete pilot training. In the US Air Force, for example, despite selection of the best applicants for pilot training, some pilot students fail or have difficulties during training (e.g., Carretta, 2011).

Learning in nondeterministic environments has widely been studied using the Multiple Cue Probability Learning (MCPL) task, which grew out of probabilistic functionalism (Brunswik, 1955, 1956). Learners have to predict criterion states from states of cues through exposure to successive multiple cue-criterion configurations (for reviews, see Reference Hammond and StewartHammond & Stewart, 2001; Reference Karelaia and HogarthKarelaia & Hogarth, 2008). Uncertainty in the tasks comes from the non-deterministic relationship between cues and criteria. Large individual differences are usually found in the final performance in such tasks (for a recent study including individual analyses, see Reference Speekenbrink and ShanksSpeekenbrink & Shanks, 2010).

In the present paper we aimed at exploring individual differences in MCPL within a pilot selection context and at relating them to pilot training outcome. More precisely, we assessed the proportion of pilot students who experienced difficulties during training for various subgroups of students classified by their MCPL performance. The next sections present elements of the MCPL paradigm, airline pilot selection and training, and the general principle of the empirical studies that will be presented.

1.1 The laboratory task: MCPL

1.1.1 The MCPL paradigm

Learning in uncertain environments has usually been approached through MCPL tasks. In a typical MCPL experiment, participants must discover a cue-criterion relationship through a series of trials. The basic task for the participant is to learn to make predictions of a criterion from cues, after successive trials representing values of the cues and the criterion. For example, a trader must predict share values given some financial cues to help him decide. Various types of feedback can be provided to participants. In the present paper, we focus on situations where only outcome feedback is available, that is, situations where the observed value of the criterion is provided after the participant gives his prediction on each trial. In the preceding example, the trader can be told which value was actually reached by the share. In the case of nondeterministic relationships, the actual value of the criterion differs from the target rule prediction by some random value that changes from trial to trial. Since outcome feedback does not reflect the target rule perfectly, participants must infer the target rule despite the noise.

Other kinds of feedback have been proposed in the literature, such as providing characteristics of the cue-criterion relationships (task information), providing the person’s cue-utilization (cognitive information), or providing the relations between the person’s perception of environment and the environment (e.g., Balzer, Doherty, & O’Connor, 1989; Reference Newell, Weston, Tunney and ShanksNewell, Weston, Tunney, & Shanks, 2009). In complex MCPL tasks (e.g., high number of cues or non linear relations between cues and criterion) outcome feedback is not necessarily helpful for the learning process (Reference HarveyHarvey, 2011). Nevertheless, we chose to focus on outcome feedback because we believe it is more representative of real-world situations, and because this type of feedback may be helpful in the kind of tasks used in the present studies, with few and uncorrelated cues (Reference Hogarth and KarelaiaHogarth & Karelaia, 2007).

1.1.2 Individual differences in MCPL

A standard view holds that MCPL involves a hypothesis testing activity in which the individual constructs hypotheses from memory about the relationship between the cues and the criterion and tests them with the available data (Reference BrehmerBrehmer, 1980; Reference Lindberg and BrehmerLindberg & Brehmer, 1977). In non-deterministic MCPL tasks where outcome feedback is provided, given its noisy nature, the individual has also to resist frustration, and defense mechanisms may be involved in order to reduce the generated anxiety (Reference SmedslundSmedslund, 1955). Individual differences in MCPL performances have notably been found between pathological and non-pathological groups. For instance, schizophrenics’ performances were increasingly impaired as the number of cues augmented (Reference GillisGillis, 1969). Depressed individuals also demonstrated difficulties in applying consistently a particular cognitive strategy and in utilizing new and more relevant information (Reference PostPost, 1978). Furthermore, paranoid individuals manifested greater difficulty in ignoring irrelevant aspects of the environment compared to non-paranoid individuals (Reference Gillis and DavisGillis & Davis, 1973). On non-pathological individuals, no consistent differences in performances have been found between cognitively simple and complex participants (Reference WintersWinters, 1970) or between individuals varying in cognitive styles, using the Myers-Briggs Type Indicator (Reference Ruble and CosierRuble & Cosier, 1990). More generally, MCPL studies have shown large variability in the strategies used by participants (e.g., Reference Gluck, Shohamy and MyersGluck, Shohamy, & Myers, 2002; Reference Meeter, Myers, Shohamy, Hopkins and GluckMeeter, Myers, Shohamy, Hopkins, & Gluck, 2006).

The impact of the nature of the task on the performances has been widely studied. Linear relationships are easier to learn than non linear ones (e.g., Reference Hammond and SummersHammond & Summers, 1965). Performances are also better when the proportion of noise is smaller (e.g., Reference Peterson and UlehlaPeterson & Ulehla, 1964). Tasks with mixed cues, i.e., with some cues being positively and other cues being negatively related to the criterion, have been shown to be sensitive to age and working memory capacity differences. Young adults were compared to old people, to young children and adolescents. Consistently, young adults were the group with the highest MCPL performances (Reference Chasseigne, Ligneau, Grau, Le Gall, Roque and MulletChasseigne et al., 2004; Reference Lafon, Chasseigne and MulletLafon, Chasseigne, & Mullet, 2004). The authors hypothesized that tasks with mixed cues involve the inhibition of the prepotent direct relation response and the coordination of the different cue values, which is supposed to load heavily on executive functioning. Reference Rolison, Evans, Walsh and DennisRolison, Evans, Walsh, and Dennis (2011) found that individuals with high working memory capacity (WMC) performed better on tasks that contained positive and negative cues than individuals with low WMC, but high-WMC individuals performed no better in tasks containing only positive cues.

1.2 Learning in a natural setting: Pilot Training

For a student with no flying experience, airline pilot training lasts 2.5 to 3 years and consists of theoretical and practical training. After taking theory examinations for the Airline Transport Licence on aeronautical knowledge, pilot students train for pilot licences (Commercial Pilot Licence with the qualification for Instrument Rules flights, Multiple Engine aircraft, or Multi-Crew Cooperation). Practical training is composed of flying hours with a flight instructor and simulator flights grouped in several phases (manoeuvrability, radio-navigation, instrument flight rules,...). At the end of each phase, a check flight assesses the flying skills of the pilot student. In case of difficulties, additional flying hours or the exclusion from the training are decided by the training organization. Given the high cost of flying hours (training one single student costs about 250 K€, i.e., ≈ 320 K$), pilot training organizations are interested in limiting the number of additional flying hours and the number of training failures.

1.2.1 Pilot Selection

The pilot selection process is usually composed of successive steps. The most common selection tools are cognitive ability tests, psychomotor tests, group exercises and individual interviews (Reference Carretta, Retzlaff, Paul and KingCarretta, Retzlaff, Callister, & King, 1998; Reference Goeters, Maschke and EissfeldGoeters, Maschke, & Eissfeld, 2004; Reference MartinussenMartinussen, 1996). Various psychological dimensions are traditionally evaluated: Spatial ability, numerical ability, verbal ability, attentional ability, multitasking, decision making, cooperation, communication, leadership, and other personality measurements. However, to our knowledge, in these selection processes the ability to learn to deal with uncertainty has never been assessed directly.

Much research has been focused on assessing the relationship between student performance at the selection tests and pilot training outcome, i.e., the predictive validity of the selection tests (Reference Burke, Hobson and LinskyBurke, Hobson, & Linsky, 1997; Reference CarrettaCarretta, 2011; Reference Carretta and ReeCarretta & Ree, 1994, 2003; Reference DamosDamos, 1993; Reference MartinussenMartinussen, 1996; Reference Martinussen and TorjussenMartinussen & Torjussen, 1998; Reference Olea and ReeOlea & Ree, 1994; Reference Park and LeePark & Lee, 1992; Reference Ree and CarrettaRee & Carretta, 1996; Schmidt & J. E. Hunter, 1998; Reference Stauffer and ReeStauffer & Ree, 1996).

Most predictive validity studies use correlations between selection test scores and training outcome. The training outcome is evaluated through flying grades, instructor assessments or pass/fail criteria. Correlations between selection tests and training outcome typically ranged between r = .15 and r = .40. The best predictors were composite scores based on cognitive and psychomotor tests (e.g., r = .37, Martinussen, 1996). Nevertheless, predictive validity of the pilot selection test scores has declined since the 1960s; for instance, the mean correlation between mechanical ability scores and pilot training outcome decreased from r = .32 to r = .14 between 1940–1960 and 1961–1990 (D. R. Reference Hunter and BurkeHunter & Burke, 1994). Thus, it is important to better comprehend the causes of failure. An investigation of these causes in a pilot training organization showed that the pilot students who had difficulties during practical training were not necessarily the worst performers on the cognitive ability tests used at the selection stage (Reference MattonMatton, 2008). Indeed, practical flying training involves different processes than those required to perform well on traditional cognitive ability tests. Student pilots have in particular to learn to deal with uncertain elements (e.g., weather, nearby traffic, engine failures, etc.) and to make decisions based on incomplete data. In some cases, flight instructors noted that pilot students had difficulty facing the unexpected and/or had difficulty identifying the most relevant information and got lost in details. Thus, it seemed beneficial to assess the candidates’ ability to adapt to uncertainty through the MCPL paradigm and to evaluate the relationship with pilot training outcome.

1.3 Logic of the studies

The idea was to relate MCPL performance to pilot training outcome. The studies were carried out in an actual pilot selection and training context. The French Air Transport Pilot Training School, “ENAC”, offers each year the opportunity to 20 to 80 young students to receive free theoretical and practical airline pilot training. In this organization, pilot selection comprises three steps: written-academic tests (mathematics, physics, English), cognitive-ability tests and final tests (group exercises completed by individual interviews, and an oral English exam). Among those students, two thirds are eliminated at the first step (written-academic tests). After the final step, around 10% of the sample who took the cognitive-ability tests are selected for training. As the samples of pilot trainees were small, we collected data of three selection sessions. Two studies were conducted:

  • First, an individual differences study was carried out on applicant data from three sessions. A MCPL task was added in an actual pilot selection context to assess individual differences in the ability to learn to adapt to unpredictable environments. Based on the MCPL literature, the majority of applicants were expected to perform such a task successfully. Therefore, pilot students who would perform poorly on the MCPL task might have some particular difficulty in dealing with uncertainty or with noisy information.

  • Second, an external validity study was performed on training data for the three corresponding selected pilot students classes. Final pilot training outcome was collected 2.5 to 3 yrs after selection and coded as Success (those who succeeded the training without any problem) or Difficulty (those who received additional training hours or who failed). Finally, pilot training outcome was related to MCPL performances.

The main methodology was approximately analogous for the three sessions and their commonalities are now described.

1.3.1 Participants

The three samples of the individual difference study consisted of the applicants who were taking the yearly examination for admission to the ENAC pilot training, all young adults and mostly males. The external validity study was conducted on the cumulated three small samples of pilot trainees recruited after the whole selection process.

1.3.2 Cues, criteria and their relationships

Given the interesting results on MCPL individual differences with mixed-cues tasks (Reference Chasseigne, Ligneau, Grau, Le Gall, Roque and MulletChasseigne et al., 2004; Reference Lafon, Chasseigne and MulletLafon et al., 2004; Reference Rolison, Evans, Dennis and WalshRolison, Evans, Dennis, & Walsh, 2012; Reference Rolison, Evans, Walsh and DennisRolison, Evans, Walsh, et al., 2011) we chose to use a combination of positive and negative cue-criterion relationships. Moreover, as linear relationships are easier to learn than non linear ones (Reference Brehmer and QvanstromBrehmer & Qvanstrom, 1976), given our objective of detecting poor MCPL performances, we chose to use linear cue-criterion relationships. Indeed, a difficulty detected on an easy task is potentially more meaningful than on a difficult task. For the same reason we chose to use uncorrelated cues, as cue redundancy usually impaired MCPL performance (Reference Karelaia and HogarthKarelaia & Hogarth, 2008), even though in real piloting settings, many cues are inter-correlated and redundant.

1.3.3 Task, apparatus and procedure

The MCPL task was inserted at the end of the cognitive-ability testing step, but MCPL results were not taken into account for the selection decision (the applicants were not informed of this). As the selection process itself was being renewed at that time, the cognitive ability tests differed widely across sessions. They will be detailed in the methodological section of each study and in appendices A, B and C.

The MCPL tasks consisted of 60 trials within a specific time frame. A progress bar representing remaining time was shown at the bottom of the screen, which certainly induced time pressure. On each trial, the cues were presented as vertical bars of continuously varying height (up to 350 pixels) on a 15" CRT computer screen with a 1024× 768 resolution. Participants provided their prediction by setting the height of a response-bar using the mouse. After clicking on a validation button, they received the outcome feedback through a fourth bar (see Figure 1). MCPL stimuli were constructed from a linear regression in the form y=cue1−cue2+e, with e an error term from a standard normal distribution. Cues and outcome feedback values were then transformed to vary from 0 to 350 pixels. The first five trials were used for familiarization. Importantly, all participants were instructed that perfect performance was almost impossible to attain, due to some random factors.

Figure 1: MCPL task with two cues. From left to right, the first two bars represented the cues. The third bar was the individual’s response and the fourth bar was the feedback given.

1.3.4 Analyses

Following Brunswik, achievement (noted r a after Reference Hursch, Hammond and HurschHursch, Hammond, & Hursch, 1964) denotes the correlation between a participant’s responses and the corresponding criteria. In order to assess the ability to learn the probabilistic relationship we focused on the MCPL final performance, but also on the initial performance in order to have a reference point. Individual initial and final performances were summarized through two correlations: ra 1, the achievement of the first 20 trialsFootnote 1 and ra Last, the achievement of the 20 last trials treated by the applicant. As the five first trials were familiarization trials, ra 1 ranged from trials #6 to #25. Given that all applicants did not complete the 60 trials within the time-limit, the 20 last trials could differ from one individual to another. Nevertheless, for each applicant ra Last corresponded to the achievement level reached after benefitting from the maximum amount of learning time.

To assess the discriminant validity of MCPL performance, we computed correlations between MCPL performance and cognitive ability tests scores. As the batteries of cognitive ability tests differed across selection sessions, we computed the standardized sums of z-scores of all cognitive ability tests per session, which we denoted Z cog.

2 Individual differences Study on Selection Data

MCPL individual differences were studied using three selection sessions.

2.1 Session 1: Low Uncertainty, Two Cues

Session 1 was conducted during the 2006 pilot selection.

2.1.1 Participants

At the selection stage, 556 applicants took the MCPL task, all aged between 18 and 31 years old (M = 21.0, sd = 2.48) and 91.2% male. Forty four pilot students (90.9% male, M age = 20.6) were admitted after the selection process.

2.1.2 Task, apparatus and procedure

The cue-criterion multiple correlation was high (R e = .96). Cues were linearly related and individual ecological validities were positive, .63 and negative, -.72. The cue inter-correlation was <.01. The whole task was limited to 10 minutes. The average number of trials completed was 58.3 (sd = 5.1).

Six cognitive ability tests were administered before the MCPL task: A spatial ability test, a mechanical comprehension test, a perceptual speed test, a numerical ability test, a reasoning test and a divided attention test (see Appendix A for more details).

2.1.3 Results

Large individual differences on initial and final performances were observed among applicants (see Figure 2). On the whole, applicants did learn the cue-criterion relationship (see Table 1) as they started at a mean initial correlation of ra 1 = .42 and ended at a significantly higher mean final correlation of ra Last = .73 (p <.001). Final performance was significantly related to the number of trials completed within the time limit (r(554)=.26, p <.001). Total time spent after the five familiarization trials varied from 3.1 to 9.5 min (M = 6.9, sd = 1.4). Interestingly, the majority of applicants had high final performances, thus poor final performances were rare in this population.

Figure 2: Histograms of initial (ra 1) and final (ra Last) MCPL performances for the three studies, each correlation being computed over 20 trials.

Table 1: Descriptive statistics for initial (ra 1) and final (ra Last MCPL performances for the three studies

2.2 Session 2: Low Uncertainty, Three Cues

Session 1 showed that substantive differences in the learning of a non-deterministic relationship between cues and criteria could be observed among pilot candidates, even with a small amount of noise in the relationship that had to be learnt (R e = .96). Large individual differences in the learning profiles were identified, differing by initial and final performance. Study 2 aimed at replicating the differences in MCPL performances during the 2007 examination. Indeed, applicants of this selection process are known to be well informed of the tests used at the preceding selection session (through Internet forums for example). Therefore it was necessary to change the relationship to be learnt in the MCPL task. We chose to increase the difficulty by adding an irrelevant cue as a new cue.

2.2.1 Participants

At the selection stage, 701 applicants took the MCPL task, all aged between 18 and 31 years old (M = 20.7, sd = 2.24) and 88.4% male. Sixty pilot students (86.7% male, M age = 20.4) were admitted after the selection process (see Appendix B for more details).

2.2.2 Task, design and analyses

Three cues were used. The multiple cue-criterion correlation was similar to that of Study 1 (i.e., R e = .96). Individual cue-criterion correlations were positive (R P = .74), negative (R N = −.70), and almost null (R I = .09, I for “Irrelevant”). Cue inter-correlations were nonsignificant. Participants had 15 minutes to complete the task. The average number of trials completed was close to that of Session 1 (M = 58.6, sd = 4.6).

Ten cognitive ability tests were administered before the MCPL task: Two spatial ability tests, two mechanical comprehension tests, two perceptual speed tests, one numerical ability test, one reasoning test, one verbal ability test and a divided attention test (see Appendix B for more details).

2.2.3 Results

As in Study 1, large individual differences on initial and final performances were observed among applicants (see Figure 2). On the whole, applicants did learn the cue-criterion relationship (see Table 1) as they started at a mean initial correlation of ra 1 = .31 and ended at a significantly higher mean final correlation of ra Last = .52 (p <.001). Final performance was again significantly related to the number of trials completed within the time limit (r(699)=.21, p <.001). Total time spent after the five familiarization trials was significantly longer than in Session 1 (M = 10.0, sd = 2.5, miN = 3.0 and max=14.1, t(1255)=26.3, p <.001)), which was predictable given the increase of available time. Initial and final performances were lower than in Session 1, so the addition of an irrelevant cue increased the task difficulty, although the available time was increased (from 10 to 15 min). As in Session 1, the majority of applicants had high final performances, thus the poor final performances were rare in this population too.

2.3 Session 3: Replication of Session 2

Session 3 was conducted during the 2010 examination. We chose to use exactly the same task as in Session 2, i.e., with low uncertainty and three mixed cues. As a result, we could cumulate data of these three sessions for the external validity study.

2.3.1 Participants

At the selection stage, 412 applicants took the MCPL task, all aged between 18 and 30 years old (M = 21.5, sd = 2.71) and 91.7% male. Twenty eight pilot students (92.9% male, M age = 20.3) were admitted after the selection process.

2.3.2 Task, design and analyses

The task was strictly identical to that of Study 2: Three cues, high multiple cue-criterion correlation (R e = .95) and individual cue-criterion correlations were positive (R P = .74), negative (R N = −.69), and almost null (R I = .08). Cue inter-correlations were nonsignificant. Participants had 15 minutes to complete the task. The average number of trials completed was close to that of Session 2 (M = 59.4, sd = 3.9).

Fourteen cognitive ability tests were administered before the MCPL task: Two spatial ability tests, one mechanical comprehension test, two perceptual speed tests, two numerical ability tests, three reasoning tests, three verbal ability tests and one multitasking test (see Appendix C for more details).

2.3.3 Results

The results replicated those of Session 2, as mean initial and final performances reached similar levels (see Table 1). Again, large individual differences among initial and final performances were observed and poor final performances were rare (see Figure 2). Final performance was again significantly related to the number of trials completed within the time limit (r(410)=.13, p <.01). However, total time spent after the five familiarization trials was significantly lower than in Session 2 (M = 7.6, sd = 2.8, miN = 2.8 and max=14.1, t(1111)=15.1, p <.001)).

3 External Validity Study on Training Data

3.1 Participants

We combined pilot training data for the three pilot student groups (N = 44, N = 60 and N = 28), so the sample consisted of N = 132 pilot students (87.8% male, M age = 20.5 and sdage = 1.2). All of them came from scientific preparatory classes for competitive admission to elite universities. Among them, 30.0% had experienced difficulties during practical pilot training leading to complementary flying hours and/or exclusion from the training.

3.2 Analyses

Firstly, to get a picture of the nature of the relationship between MCPL performance and pilot training outcome, we applied a method described by Reference Hosmer and LemeshowHosmer and Lemeshow (2000). It consisted in creating intervals for the MCPL performance and computing the frequency of occurrence of pilot training difficulties within each group. We chose to use quartiles, so each group size was sufficient to compute representative frequencies (N = 33). Thus, we created four MCPL performance categories labeled “poor”, “medium”, “high” and “very high”. As the tasks used in the three selection sessions were equivalent in level of global uncertainty (R e = .96), we combined MCPL performance and pilot training data of the three classes.

Secondly, we investigated the potential role of individual differences in cognitive ability test scores in the association between MCPL performance and pilot training difficulty. We computed the standardized sums of z-scores of cognitive ability tests, Z cog, for each session and corresponding Cronbach’s alpha. The association between the performance on the cognitive ability tests and the MCPL performance was assessed through the correlation between the Z cog aggregated across the three sessions and the Fisher transformed ra 1 and ra Last of all applicants. Then, we asked whether Z cog could have moderating effects on the association between MCPL and pilot training difficulty by deriving partial contingency tables at various levels of Z cog. Given the small sample size, we categorized Z cog in two subgroups by a median split and derived the partial contingency tables (e.g., see Agresti, 2002, p. 47–54 for the methodology of partial association). Z cog score was also added as a predictor in logistic regressions of pilot training outcome on MCPL performance to assess the potential confounding effect of individual differences in Z cog scores on the relationship between the two variables. Four models of logistic regression were fitted to the data. Model M1 used raw ra Last, i.e., the fine grained variability of MCPL final performances. In Model M2 we used simplified predictor data, corresponding to the four categories of MCPL final performance defined by the mean of each quartile. Model M3 tested the significance of MCPL initial performance categorized in quartiles in the same way. M4 tested the significance of final MCPL performance and Z cog both categorized in quartiles.

3.3 Results

3.3.1 Relationship between MCPL and pilot training

Descriptive statistics of MCPL performance for the sample of pilot students showed large individual differences in initial and final MCPL performance (see Table 2). Final MCPL performance and the frequency of pilot training difficulty were associated (see Figure 3 and Table 3). Indeed, the highest rate of difficulty during training was observed for the group of poorest MCPL final performances. Moreover, the training difficulty rate was non significantly different for medium, high and very high MCPL final performances (27%, N = 33 vs. 24%, N = 33, p = .78). The pattern suggested a cutoff around ra Last = .50. If applied (see Table 4), this cutoff would lead to a significant difference between training difficulty rates for the two subgroups of poor vs. good MCPL performers (44% below cutoff, N = 33 vs. 25% above cutoff, N = 99, p = .03). On the other hand, there was no evidence of an association between MCPL initial performance and training outcome. Indeed, after applying the same cutoff (ra 1 = .50), the difficulty rate was not significantly different for the two subgroups (31% below cutoff, N = 72 vs. 31% above cutoff, N = 60, ns).

Figure 3: Proportion of pilot training difficulties for MCPL final performance, ra Last, grouped by quartiles. Dots = group means. Individual performances are represented by "+" (difficulty during pilot training) and "–" (pilot training success).

Table 2: Descriptive statistics for initial (ra 1) and final (ra Last) MCPL performances for pilot students

Table 3: Frequency Table of Pilot Training Difficulty Rate by MCPL Performance Group

Table 4: Frequency Contingency Tables of Pilot Training Difficulty Rate by MCPL dichotomized Performance Group

Note. MCPL final and initial performance has been dichotomized in poor vs. good following the cutoff of ra Last = .50 and ra 1 = .50 respectively.

3.3.2 Interaction with other cognitive ability tests

Cronbach’s alpha of Z cog were acceptable for the three applicant groups (.78, .75 and .83, respectively), revealing acceptable internal consistency of this measurement. The correlation between the Fisher transformed initial and final MCPL performances and the standardized sums of the cognitive ability tests scores on the whole applicant data was low, although significant (at p <.001), r(1667)=.15 for ra 1 and r(1667)=.16 for ra Last. Therefore the differences in cognitive ability tests could account for only 2.2% of the differences of MCPL performances, highlighting discriminant validity of MCPL performance against the actual batteries of tests used. Thus, MCPL performance was not redundant over Z cog.

One might ask whether Z cog has a moderating effect on the association between MCPL and pilot training difficulty. Thus, we produced the two partial contingency tables for the subgroups of “high” vs. “low” cognitive ability and for two categories of MCPL performance using the cutoff of ra Last = .50 (see Table 5). The global odds ratio was θ = 2.40, and conditional odds ratio of both categories of Z cog were similar (θHighCog = 2.38 and θLowCog = 2.46). So, the odds of pilot training difficulty for those who performed poorly on the MCPL task were 2.4 times the odds for those performing good, regardless of their performance at cognitive ability tests.

Table 5: Partial Contingency Tables of Pilot Training Difficulty Rate by MCPL Performance Group at two levels of general cognitive ability (measured by Z cog)

Note. MCPL performance has been dichotomized in poor vs. good following the cutoff of ra Last = .50.

The results of logistic regressions confirmed the significance of MCPL final performance on pilot training outcome (see Table 6). Indeed, while controlling for differences in Z cog, MCPL final performance categorized by quartiles was a statistically significant predictor of pilot training outcome (with p = .03). On the other hand, MCPL initial performance, categorized by quartiles, was not a significant predictor of pilot training outcome. Furthermore, MCPL raw final performance was not a significant predictor of pilot training (probably due to the lack of power resulting from the use of the fine grained continuous ra Last variable). Moreover, differences in Z cog were not predictive of pilot training difficulty neither with the full scale nor categorized by quartiles. This is not surprising as pilot students were selected on the basis of Z cog, so we did not expect differences in Z cog to be highly predictive of pilot training outcome.

Table 6: Logistic Regression of MCPL Performance on Pilot Training Difficulty, including Scores on Cognitive Ability Tests as a Predictor

Note. Pilot training outcome was coded 1 for Difficult and 0 for Success. Model M1 used raw MCPL final performance, M2 used MCPL final performance categorized in quartiles (replaced by the mean value of each group), M3 used MCPL initial performance categorized in quartiles and M4 used MCPL final performance and Z cog categorized in quartiles (replaced by the mean value of each group).

* significant at 5%. Sample size N = 132

The poor predictive power of MCPL initial performance compared to final performance suggested that the predictive value of the final MCPL performance could be attributed to what had been learnt at the end of task time-limit. More precisely, individual differences in initial performance seemed to result from some random factors (good or bad luck at the first trials). Indeed, among those who started poorly (N = 72 with ra 1<.50), more than half (58%) ended at ra Last≥.50, 80% of which succeeded the pilot training.

4 General Discussion

The present studies were conducted in the context of airline pilot selection and training. The data were obtained from three selection sessions and the three corresponding selected pilot students classes. Initially, Multiple Cue Probability Learning (MCPL) performance was assessed at the time of selection. Then, pilot training outcome (success/difficulty) was collected after a three-year follow-up. All MCPL tasks used a mix of positive and negative cues, and for two of them, an irrelevant cue. The results showed large individual differences in initial and final MCPL performances, with a majority of applicants achieving high levels of performance. The level of MCPL final performance could be related to the outcome of airline pilot training. The frequency of training difficulties (i.e., additional flying hours and/or exclusion from the training) was highest for the group of poor MCPL performers compared to medium/high/very high MCPL performers. Poor MCPL performance was associated with a final achievement inferior to ra=.50. Moreover, poor initial MCPL performance (i.e., at the beginning of the task) could not be significantly related to difficulties during pilot training.

4.1 External validity of MCPL performances

To our knowledge, the relationship between MCPL performance and performance in a real setting of high skilled training had not been assessed before. Despite the significant difference of pilot training success rate for poor vs. good MCPL performers, the relationship between MCPL performance and pilot training outcome was not deterministic. A poor final MCPL performance does not necessarily imply difficulties in pilot training. Indeed, the causes of pilot training difficulties are surely complex. Moreover, as pointed out by Klayman (1984), probability learning tasks could fail to capture some important features of learning in natural environments, such as the discovery of new valid predictive cues, and the incorporation of these new cues into the learner’s predictive model. Thus, we did not expect the training difficulties to be systematically related to difficulties in the MCPL task. Nevertheless poor final MCPL performance was significantly related to a higher proportion of pilot training difficulty. Moreover, initial MCPL performance could not predict pilot training outcome, even though the initial performance was not intended to do so. Indeed, initial MCPL performance corresponded to the trials where participants tested their first hypotheses. For instance, if participants thought first of positive cue-criterion relationships (as is often the case, Brehmer, 1977), their poor initial performances would not be symptomatic of cognitive impairment.

What could be the underlying cognitive processes that led the pilot students to have difficulty both in the MCPL task and during the pilot training? A lack of ability to generate different hypotheses (or a too “sparse hypothesis space”, Reference Navarro and PerforsNavarro & Perfors, 2011) or the perseveration in a wrong hypothesis (e.g., Reference Dunbar and SussmanDunbar & Sussman, 1995). Unfavorable personality characteristics could also be invoked. Perhaps MCPL tasks assess the degree of ambiguity tolerance in a behavioral way. Unfortunately, no ambiguity tolerance test was present in the battery of tests at the selection stage. Another interpretation could be that the experimental conditions of the high stake selection setting overloaded the executive functions of the poor MCPL learners, thus preventing them from functioning efficiently in a controlled mode (e.g., Reference Keinan, Friedland, Kahneman and RothKeinan, Friedland, Kahneman, & Roth, 1999). Stressors are known to promote the use of simple strategies, even in individuals accustomed to using complex solutions (Reference Van Hiel and MervieldeVan Hiel & Mervielde, 2007). Thus, we could explain learning failures by a disruption of executive processing due to an emotional reaction linked to a high-stakes and stressful examination. In a dual-process perspective, Rolison, Evans, Dennis, et al. (2012) suggested that learning about positive cues would involve automatic processes whereas learning about negative cues would involve controlled processes. From that view point, an interpretation of our results could be that poor MCPL performers would have difficulty in getting involved in controlled processes in stressful situations. Similarly, such difficulty could occur in real-life flight situations.

4.2 Poor MCPL performances among young adults

Consistent with previous findings from the literature, we found large individual differences in MCPL performance. Nevertheless, we could have expected better final performances. For instance with a similar MCPL task, young adults (N = 98, aged from 18 to 25) were all able to learn a two-cue mixed probability learning task (Reference Chasseigne, Ligneau, Grau, Le Gall, Roque and MulletChasseigne et al., 2004). Lafon et al. (2004) showed that children of 5 to 10 years old had difficulty in learning the negative relationship and that only the young adults (aged from 17 to 27) learned efficiently how to use the negative cue correctly. It is noteworthy that, in these two studies, participants had 150 trials and no time limit (between 30 and 40 min to complete the task). In our studies, time was limited (10 or 15 min), which induced some time pressure and could explain the non optimal final performance. Therefore, our MCPL performances are more likely to reflect a rate of learning in uncertainty, than an ability to deal with uncertainty. Moreover, results of Lafon et al. (2004) and Chasseigne et al. (2004), showed some improvement after 60 trials (2 first blocks). For these authors, the presence of the negative cue and the coordination of the two cue values involved greater demands on the executive control, thus providing an interpretation for the poor performances of both very young or very old participants. This hypothesis is also consistent with the findings of Rolison, Evans, Walsh, et al. (2011), who found that working memory capacity was correlated with performance on MCPL tasks containing negative cues.

4.3 MCPL and other cognitive abilities

As noted by Reference Weaver and StewartWeaver and Stewart (2012) “despite over 300 studies of MCPL (Reference HolzworthHolzworth, 1999), MCPL has not been connected to the intelligence or learning literature” (p. 403). Reference Weaver and StewartWeaver and Stewart (2012) found a correlation (r(98)=.29, p <.01) between scores of an inductive reasoning test and performance on a three-mixed-cue MCPL task with low uncertainty (R e = .9). Overall correlation with a composite score of other cognitive ability tests in our studies was also significant but smaller (r(1667)=.16, p <.001). The larger sample size of our studies would tend to make us cautious regarding the medium correlation obtained by Reference Weaver and StewartWeaver and Stewart (2012). Nevertheless, the correlation we found is positive, indicating that some part of the variance observed in MCPL performance may be attributed to what is usually called general cognitive ability.

4.4 Practical Implications

The practical implication of the present finding in a selection setting is quite straightforward. MCPL tasks in a selection setting could be useful to detect applicants with difficulty learning in uncertainty. However, the MCPL tasks used in the present studies involved perceptual skills (as cues, response and outcome feedback were represented through colored bars), and the question remains open how as to generalize to learning uncertainty in tasks involving cognitive skills.

The practical implications can range between two extremes. At one extreme, poor MCPL performance could alert the selection practitioners and incite them to further investigate those applicants’ ability to deal with uncertainty (during the interviews for example). At the other extreme, the selection organization could eliminate poor MCPL performers. From a purely organizational point of view, the minimization of the training difficulty risk would justify this decision despite of unavoidable wrong eliminations.

It is unclear whether MCPL tasks could help diagnose complex learning deficiencies or if an individual MCPL profile could be useful for an instructor to adapt his training method to the student. These questions are now opened by this research.

Appendix A: Cognitive ability tests of Session 1

  1. 1. Spatial test (S1). This paper-and-pencil test was composed of three time-limited sub-tests (180 s, 300 s, 210 s) measuring the following abilities: (a) perceptual speed (an identical picture test of 25 items), (b) spatial relations (a picture rotation test of 20 items), and (c) visualization (a block counting test of 15 items). Total scores varied from 0 to 85 (number of correct answers).

  2. 2. Mechanical movement test (M1). This paper-and-pencil test presented 36 situations to evaluate, from a mechanical point of view, with a choice of 4 possible answers for each situation. The test was time limited (25 min) and scores ranged from 0 to 42 (number of correct answers).

  3. 3. Attentional ability test (A1). In this paper-and-pencil test applicants had to detect three target signs among eight in a page containing 1560 signs. Time was limited (10 min) and scores ranged from −600 to 600 (number of correct answers minus number of omissions).

  4. 4. Numerical test (N1). This paper-and-pencil test presented 30 arithmetic problems where applicants had to choose the correct answer among 5. They could write intermediate calculations on rough paper provided. This test was time limited (35 min) and scores ranged from 0 to 30 (number of correct answers).

  5. 5. Reasoning test (R1). This paper-and-pencil test presented 48 reasoning problems where applicants had to calculate a distance while taking into account one to three additional rules. The correct answer had to be chosen among 5. Test was time limited (10 min) and scores ranged from 0 to 48 (number of correct answers).

  6. 6. Divided attention test (TAD). This computer-based test was composed of four stages in which four tasks were successively added (from only one task at the first stage to four tasks at the fourth stage). A specific box was connected to a computer and comprised all the elements necessary to interact with the software. The first task was a pursuit task where applicants had to maintain a cross in a circle with a lever on the box while compensating for some pseudo-random movements. Performance of the pursuit task corresponded to the mean Euclidian distance between the cross and the middle of the circle. The second task was a monitoring task consisting in maintaining the level of a gauge at the middle of a rectangle with a second lever on the box. Performance of the monitoring task corresponded to the mean distance of the gauge level from the middle of the rectangle. The third task was a detection task in which applicants had to push on one of four boutons on the box when one of four corresponding squares becomes red (instead of blue or green). Performance of the detection task consisted of the number of correct and incorrect actions. The fourth task was a mental calculation task where the applicants had to enter the result of a simple calculation (e.g., 15+9−12) through the numerical keypad of the box. A composite score was calculated taking into account the performances for each task at the various stages.

Appendix B: Cognitive ability tests of Session 2

  1. 1. Spatial test (S1). This test was identical to S1 from Session 1.

  2. 2. Spatial test (S2). This paper-and-pencil test was a test of visualization in three dimensions. Applicants had to chose which of four three-dimensional forms could be made by folding a specified two-dimensional model. This test was time limited (15 min) and scores varied from 0 to 30 (number of correct answers).

  3. 3. Mechanical movement test (M1). This test was identical to M1 from Session 1.

  4. 4. Mechanical movement test (M2). This paper-and-pencil test presented 30 situations to evaluate from a mechanical point of view, with a choice of 3 possible answers for each situation. This test was time limited (15 min) and scores ranged from 0 to 30 (number of correct answers).

  5. 5. Attentional ability test (A1b). This test was a parallel form of A1 from Session 1.

  6. 6. Instrument reading test (A2). This test consisted in reading the value indicated by six instruments (e.g., speed, oil pressure) and choosing the correct answer among 5. Seventy items had to be completed in 10 min. Scores varied from 0 to 70 (number of correct answers).

  7. 7. Numerical test (N1). This test was identical to N1 from Session 1, except that it was computer-based.

  8. 8. Inductive reasoning test (R2). In this test, applicants had to induce analogies and differences among abstract figures and to decide whether a given figure belonged to one of two groups of figures or not. The test comprised 105 items and was time limited (30 min). Scores ranged from 0 to 105 (number of correct answers).

  9. 9. Verbal comprehension test (V1). This computer-based test consisted in reading texts and and answering comprehension questions by choosing the correct answer among 5. Ten texts and three questions per test had to be completed in 35 min. Scores varied from 0 to 30 (number of correct answers).

  10. 10. Divided attention test (TAD). This test was identical to TAD from Session 1.

Appendix C: Cognitive ability tests of Session 3

  1. 1. Spatial test (S2). This test was identical to S2 from Session 2, except that it was computer-based.

  2. 2. Spatial test (S3). This computer-based test consisted in rotating mentally a figure following instructions and choosing the correct answer among 5. Sixty items had to be completed in 15 min. Scores ranged from 0 to 60 (number of correct answers).

  3. 3. Mechanical movement test (M2). This test was identical to S2 from Session 2 , except that it was computer-based.

  4. 4. Numerical test (N1). This test was identical to N1 from Session 1, except that it was computer-based.

  5. 5. Numerical test (N2). This computer-based test consisted in computing mental calculations without rough paper and choosing the correct answer among 10. Forty items had to be completed in 20 min. Scores ranged from 0 to 40 (number of correct answers).

  6. 6. Inductive reasoning test (R2). This test was identical to S2 from Session 2, except that it was computer-based.

  7. 7. Verbal comprehension test (V1b). This test was a parallel form of V1 from Session 2.

  8. 8. Inductive reasoning test (R3). In this computer-based test applicants had to induce the rule(s) that governed a set of three abstract figures and to chose the correct figure that best completed the set among 6. Thirty six items had to be completed in 35 min. Scores ranged from 0 to 36 (number of correct answers).

  9. 9. Inductive reasoning test (R4). In this computer-based test applicants had to induce the rule(s) that governed a set of eight abstract figures and to chose the correct figure among 8. Thirty items had to be completed in 10 min. Scores ranged from 0 to 30 (number of correct answers).

  10. 10. Verbal ability test (V2). In this computer-based test applicants had to chose the correct synonym of a word among 6. Forty three items had to be completed in 10 min. Scores ranged from 0 to 43 (number of correct answers).

  11. 11. Verbal ability test (V3). In this computer-based test applicants had to find the odd one out of a series of six words. Fifty items had to be completed in 15 min. Scores ranged from 0 to 50 (number of correct answers).

  12. 12. Attention test (A3). This computer-based test was composed of two stages. In the first stage applicants had to count the number of target signs and chose the correct answer among 7. In the second stage, applicants had to count target signs following a rule given in the instructions and chose the correct answer among 7. In each stage, ten items had to be treated in 5 min. Scores ranged from 0 to 20 (number of correct answers).

  13. 13. Attention test (A4). In this computer-based test applicants had first to memorize four target numbers or letters and their locations, and second to detect their presence in four sets of 128 letters or numbers. One hundred and sixty items had to be completed in 24 min. Scores ranged from 0 to 160 (number of correct answers).

  14. 14. Divided attention test (TGP). The principles of this test were analogous to those of TAD from Sessions 1 and 2. Four tasks that were successively added at four stages. The pursuit task consisted in pursuing a moving circle with a cross through a first joystick. The monitoring task consisted in maintaining the level of four gauges inside an interval by using the second joystick. The detection task consisted in pushing on one of nine keyboard keys when a target letter appeared in the corresponding zone. The mental calculation consisted in simple arithmetic calculations (e.g., deducing a distance from speed and time). The composite score was again calculated taking into account the performances for each task at the various stages.

Appendix D: Cognitive ability test score correlations

Correlations between test scores and MCPL initial and final performances (after Fisher transformation).

Note. * significant at 5%.

** significant at 1%.

*** significant at 0.1%. Sample sizes are N = 556, N = 701 and N = 412.

Footnotes

Note. * significant at 5%.

** significant at 1%.

*** significant at 0.1%. Sample sizes are N = 556, N = 701 and N = 412.

1 Twenty trials seemed to be a good compromise between the minimum number of trials required to compute a correlation and the maximum number of last trials that represent the final performance.

References

Agresti, A. (2002). Categorical data analysis. Hoboken, NJ:John Wiley & Sons.CrossRefGoogle Scholar
Balzer, W. K., Doherty, M. E., & O’Connor, R. (1989). Effects of cognitive feedback on performance. Psychological Bulletin, 106, 410433.10.1037/0033-2909.106.3.410CrossRefGoogle Scholar
Brehmer, B. (1977). Learning complex rules in probabilistic inference tasks: iii. the effect of cue validity (tech. rep. No. 117).Google Scholar
Brehmer, B. (1980). In one word: not from experience. Acta Psychologica, 45, 233241.CrossRefGoogle Scholar
Brehmer, B., & Qvanstrom, G. (1976). Information integration and subjective weights in multiple-cue judgments. Organizational Behavior and Human Decision Processes, 17, 118126.CrossRefGoogle Scholar
Brunswik, E. (1955). Representative design and probabilistic theory in a functional psychology. Psychological Review, 62, 193217.CrossRefGoogle Scholar
Brunswik, E. (1956). Perception and the representative design of psychological experiments. University of California Press: Berkeley.10.1525/9780520350519CrossRefGoogle Scholar
Burke, E. F., Hobson, C., & Linsky, C. (1997). Large sample validations of three general predictors of pilot training success. International Journal of Aviation Psychology, 7, 225234.CrossRefGoogle ScholarPubMed
Carretta, T. R. (2011). Pilot candidate selection method: still an effective predictor of us air force pilot training performance. Aviation Psychology and Applied Human Factors, 1, 38.CrossRefGoogle Scholar
Carretta, T. R., & Ree, M. J. (1994). Pilot-candidate selection method: sources of validity. International Journal of Aviation Psychology, 4, 103117.CrossRefGoogle Scholar
Carretta, T. R., & Ree, M. J. (2003). Pilot selection methods. In P. S. Tsang & M. A. Vidulich (Eds.), Principles and practice of aviation psychology (pp. 357396). Mahwah, NJ: Lawrence Erlbaum Associates.Google Scholar
Carretta, T. R., Retzlaff, D., Paul, Callister, J. D., & King, R. E. (1998). A comparison of two U.S. Air force pilot aptitude tests. Aviation, Space, and Environment Medicine, 69, 931935.Google Scholar
Chasseigne, G., Ligneau, C., Grau, S., Le Gall, A., Roque, M., & Mullet, E. (2004). Aging and probabilistic learning in single- and multiple-cue tasks. Experimental Aging Research, 30, 2345.CrossRefGoogle ScholarPubMed
Damos, D. L. (1993). Using meta-analysis to compare the predictive validity of single- and multiple-task measures to flight performance. Human Factors, 35, 615628.CrossRefGoogle Scholar
Dunbar, K., & Sussman, D. (1995). Toward a cognitive account of frontal lobe function: simulating frontal lobe deficits in normal subjects. In Structure and functions of the human prefrontal cortex (pp. 289304). New York Academy of Sciences.Google Scholar
Gillis, J. S. (1969). Schizophrenic thinking in a probabilistic situation. Psychological Record, 19, 211224.CrossRefGoogle Scholar
Gillis, J. S., & Davis, K. E. (1973). The effects of psychoactive drugs on complex thinking in paranoid and non paranoid schizophrenics: an application of the multiple-cue model to the study of disordered thinking. In L. Rappoport & D. A. Summers (Eds.), Human judgment and social interaction (pp. 170184). New York, USA: Rinehart and Winston, Inc.Google Scholar
Gluck, M. A., Shohamy, D., & Myers, C. E. (2002). How do people solve the "weather prediction" task?: individual variability in strategies for probabilistic category learning. Learning and Memory, 9, 408418.10.1101/lm.45202CrossRefGoogle Scholar
Goeters, K.-M., Maschke, P., & Eissfeld, H. (2004). Ability requirements in core aviation professions: job analyses of airline pilots and air traffic controllers. In K.-M. Goeters (Ed.), Aviation psychology: practice and research. Hampshire: Ashgate.Google Scholar
Hammond, K. R., & Stewart, T. R. (Eds.). (2001). The essential brunswik: beginnings, explications, applications. Cary, NC: Oxford University Press.CrossRefGoogle Scholar
Hammond, K. R., & Summers, D. A. (1965). Cognitive dependence on linear and non-linear cues. Psychological Review, 72, 215224.10.1037/h0021798CrossRefGoogle Scholar
Harvey, N. (2011). Learning judgment and decision making from feedback. In M. K. Dhami, A. Schlottmann & M. Waldmann (Eds.), Judgment and decision making as a skill: learning, development, and evolution (pp. 406– 464). Cambridge: Cambridge University Press.Google Scholar
Hogarth, R. M., & Karelaia, N. (2007). Heuristic and linear models of judgment: matching rules and environments. Psychological Review, 114, 733758.CrossRefGoogle ScholarPubMed
Holzworth, R. J. (1999). Annotated bibliography of cue probability learning studies (http://www.albany.edu/cpr/brunswik/resources/mcplbib.doc). Department of Psychology, University of Connecticut.Google Scholar
Hosmer, D. W., & Lemeshow, S. (2000). Applied logistic regression (Second edition). Wiley.CrossRefGoogle Scholar
Hunter, D. R., & Burke, E. F. (1994). Predicting aircraft pilot- training success: a meta-analysis of published research. International Journal of Aviation Psychology, 4, 297313.CrossRefGoogle Scholar
Hursch, C. J., Hammond, K. R., & Hursch, J. L. (1964). Some methodological considerations in multiple-cue probability studies. Psychological Review, 71, 4260.CrossRefGoogle ScholarPubMed
Karelaia, N., & Hogarth, R. M. (2008). Determinants of linear judgment: a meta-analysis of lens model studies. Psychological Bulletin, 134, 404426.CrossRefGoogle ScholarPubMed
Keinan, G., Friedland, N., Kahneman, D., & Roth, D. (1999). The effect of stress on the suppression of erroneous competing responses. Anxiety, stress, and coping, 12, 455476.CrossRefGoogle ScholarPubMed
Klayman, J. (1984). Learning from feedback in probabilistic environments. Acta Psychologica, 56, 8192.10.1016/0001-6918(84)90009-XCrossRefGoogle Scholar
Lafon, P., Chasseigne, G., & Mullet, E. (2004). Functional learning among children, adolescents and young adults. Journal of Experimental Child Psychology, 88, 334347.CrossRefGoogle ScholarPubMed
Lindberg, L. A., & Brehmer, B. (1977). Effects of task information and active feedback control in inductive inference (tech. rep. No. 123).Google Scholar
Martinussen, M. (1996). Psychological measures as predictors of pilot performance: a meta-analysis. International Journal of Aviation Psychology, 6, 120.CrossRefGoogle ScholarPubMed
Martinussen, M., & Torjussen, T. (1998). Pilot selection in the norwegian air force: a validation and meta-analysis of the test battery. International Journal of Aviation Psychology, 8, 3345.CrossRefGoogle Scholar
Matton, N. (2008). Approches psychométrique et cognitive des différences individuelles d’aptitudes: application à la sélection des pilotes de ligne. (Doctoral dissertation, University of Toulouse).Google Scholar
Meeter, M., Myers, C. E., Shohamy, D., Hopkins, R. O., & Gluck, M. A. (2006). Strategies in probabilistic categorization: results from a new way of analyzing performance. Learning and Memory, 13, 230239.CrossRefGoogle ScholarPubMed
Navarro, D. J., & Perfors, A. F. (2011). Hypothesis generation, sparse categories, and the positive test strategy. Psychological Review, 118, 120134.CrossRefGoogle ScholarPubMed
Newell, B. R., Weston, N. J., Tunney, R. J., & Shanks, D. R. (2009). The effectiveness of feedback in multiple-cue probability learning. Quarterly Journal of Experimental Psychology, 62, 890908.CrossRefGoogle ScholarPubMed
Olea, M. M., & Ree, M. J. (1994). Predicting pilot and navigator criteria: not much more than g. Journal of Applied Psychology, 79, 845851.CrossRefGoogle Scholar
Park, K. S., & Lee, S. W. (1992). A computer-aided aptitude test for predicting flight performance of trainees. Human Factors, 34, 189204.CrossRefGoogle Scholar
Peterson, C., & Ulehla, Z. J. (1964). Uncertainty, inference difficulty, and probability learning. Journal of Experimental Psychology, 67, 523530.CrossRefGoogle ScholarPubMed
Post, P. D. (1978). The cognitive functioning of depressives in a multiple-cue, probabilistic task. (Doctoral dissertation, ProQuest Information and Learning).Google Scholar
Ree, M. J., & Carretta, T. R. (1996). Central role of g in military pilot selection. International Journal of Aviation Psychology, 6, 111123.CrossRefGoogle ScholarPubMed
Rolison, J. J., Evans, J. S. B. T., Dennis, I., & Walsh, C. R. (2012). Dual-processes in learning and judgment: evidence from the multiple-cue probability learning paradigm. Organizational Behavior and Human Decision Processes, 118, 189202.CrossRefGoogle Scholar
Rolison, J. J., Evans, J. S. B. T., Walsh, C. R., & Dennis, I. (2011). The role of working memory capacity in multiple-cue probability learning. The Quarterly Journal of Experimental Psychology, 64, 14941514.CrossRefGoogle ScholarPubMed
Ruble, T. L., & Cosier, R. A. (1990). Effects of cognitive styles and decision setting on performance. Organizational Behavior and Human Decision Processes, 46, 283295.CrossRefGoogle Scholar
Schmidt, F. L., & Hunter, J. E. (1998). The validity and utility of selection methods in personnel psychology: practical and theoretical implications of 85 years of research findings. Psychological Bulletin, 124, 262274.CrossRefGoogle Scholar
Smedslund, J. (1955). Multiple-probability learning: an inquiry into the origins of perception. Oslo, Norway: Akademisk forlag.Google Scholar
Speekenbrink, M., & Shanks, D. R. (2010). Learning in a changing environment. Journal of Experimental Psychology: General, 139, 266298.CrossRefGoogle Scholar
Stanovich, K. E. (2009). What intelligence tests miss: the psychology of rational thought. New Haven, CT US: Yale University Press.Google Scholar
Stauffer, J., & Ree, M. J. (1996). Predicting with logistic or linear regression: will it make a difference in who is selected for pilot training? International Journal of Aviation Psychology, 6, 233241.CrossRefGoogle Scholar
Van Hiel, A., & Mervielde, I., (2007). The search for complex problem-solving strategies in the presence of stressors. Human Factors, 49, 10721082.CrossRefGoogle ScholarPubMed
Weaver, E. A., & Stewart, T. R. (2012). Dimensions of judgment: factor analysis of individual differences. Journal of Behavioral Decision Making, 25, 402413.CrossRefGoogle Scholar
Winters, E. P. (1970). Person perception in multiple-cue probability learning task as a function of cognitive complexity and inferential set. Dissertation Abstracts International, 30(12-B), 5681.Google Scholar
Figure 0

Figure 1: MCPL task with two cues. From left to right, the first two bars represented the cues. The third bar was the individual’s response and the fourth bar was the feedback given.

Figure 1

Figure 2: Histograms of initial (ra1) and final (raLast) MCPL performances for the three studies, each correlation being computed over 20 trials.

Figure 2

Table 1: Descriptive statistics for initial (ra1) and final (raLast MCPL performances for the three studies

Figure 3

Figure 3: Proportion of pilot training difficulties for MCPL final performance, raLast, grouped by quartiles. Dots = group means. Individual performances are represented by "+" (difficulty during pilot training) and "–" (pilot training success).

Figure 4

Table 2: Descriptive statistics for initial (ra1) and final (raLast) MCPL performances for pilot students

Figure 5

Table 3: Frequency Table of Pilot Training Difficulty Rate by MCPL Performance Group

Figure 6

Table 4: Frequency Contingency Tables of Pilot Training Difficulty Rate by MCPL dichotomized Performance Group

Figure 7

Table 5: Partial Contingency Tables of Pilot Training Difficulty Rate by MCPL Performance Group at two levels of general cognitive ability (measured by Zcog)

Figure 8

Table 6: Logistic Regression of MCPL Performance on Pilot Training Difficulty, including Scores on Cognitive Ability Tests as a Predictor

Figure 9

Note. *

Supplementary material: File

S1930297500003685sup001.csv

Matton et al. supplementary material 1

Download S1930297500003685sup001.csv(File)
File 29.5 KB
Supplementary material: File

S1930297500003685sup002.csv

Matton et al. supplementary material 2

Download S1930297500003685sup002.csv(File)
File 989.6 KB
Supplementary material: File

S1930297500003685sup003.csv

Matton et al. supplementary material 3

Download S1930297500003685sup003.csv(File)
File 46.2 KB
Supplementary material: File

S1930297500003685sup004.csv

Matton et al. supplementary material 4

Download S1930297500003685sup004.csv(File)
File 1.3 MB
Supplementary material: File

S1930297500003685sup005.csv

Matton et al. supplementary material 5

Download S1930297500003685sup005.csv(File)
File 37.4 KB
Supplementary material: File

S1930297500003685sup006.csv

Matton et al. supplementary material 6

Download S1930297500003685sup006.csv(File)
File 425.5 KB