Next Article in Journal
Childhood Trauma and Malevolent Creativity in Chinese College Students: Moderated Mediation by Psychological Resilience and Aggression
Next Article in Special Issue
Task Switching: On the Relation of Cognitive Flexibility with Cognitive Capacity
Previous Article in Journal
Quantity and Quality in Scientific Productivity: The Tilted Funnel Goes Bayesian
Previous Article in Special Issue
Metacognition, Mind Wandering, and Cognitive Flexibility: Understanding Creativity
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Is Flexibility More than Fluency and Originality?

Institute of Psychology and Pedagogy, Ulm University, Albert-Einstein Allee 47, 89081 Ulm, Germany
*
Author to whom correspondence should be addressed.
Submission received: 5 May 2022 / Revised: 23 September 2022 / Accepted: 27 October 2022 / Published: 1 November 2022
(This article belongs to the Special Issue Cognitive Flexibility: Concepts, Issues and Assessment)

Abstract

:
Flexibility (i.e., the number of categorically different ideas), fluency (i.e., the answer quantity), and originality (i.e., the quality of ideas) are essential aspects of the ability to think divergently. Theoretically, fluency and ideational flexibility tasks are akin to one another. However, flexibility was also considered to be uniquely related to working memory capacity due to the task requirements involved in generating diverse answers (e.g., self-monitoring, suppression, and category generation). Given that the role of working memory is strengthened in flexibility tasks relative to fluency and originality tasks, flexibility should be more strongly related with working memory. Additionally, mental speed should show a similar pattern of results because mental speed has been previously related to task complexity. Based on a sample of N = 409 adults (Mage = 24.01 years), we found in latent variable models that fluency/originality strongly predicts flexibility and accounts for 61% of its variance. Creative flexibility was unrelated to working memory and mental speed after controlling for fluency/originality. Additionally, the residual of a latent flexibility factor was unrelated to self-reported creative activities. We concluded that flexibility, as measured here, can be deemed primarily a method factor that did not show value over and above fluency/originality as assessed in traditional fluency and originality tasks. We discussed perspectives for disentangling trait and method variance in flexibility tasks.

1. Introduction

The established understanding of divergent thinking—one way to assess creative thinking—is based upon the fluency (i.e., the number of ideas), the originality (i.e., the quality of ideas), and the flexibility (i.e., the diversity or variety of ideas) in idea production (e.g., Guilford 1956, 1967; Reiter-Palmon et al. 2019; Weiss et al. 2021b). Elaboration as a further aspect is rarely studied (Hornberg and Reiter-Palmon 2017). These aspects of divergent thinking can be studied in terms of latent constructs as they are not directly observable but inferred from variables that are observable (e.g., the quantity of answers regarding an alternate use item). In this manuscript, we therefore understood fluency, originality, and flexibility as latent variables that depict the prominent aspects of divergent thinking and have been linked with a number of ability and personality covariates as well as with creative outcomes (e.g., Batey and Furnham 2006; Benedek et al. 2012b; Feist 1998; Weiss et al. 2021a).

1.1. Understanding Flexibility

Flexibility is a cognitive ability that helps humans to achieve flexible behavior in different environments (Ionescu 2012). Flexibility has been defined based on different ideas (see Ionescu 2012): (a) as the cognitive ability to adapt to changing demands (e.g., task set-switching (Colzato et al. 2009; Monsell 2003)), (b) as a combination of flexibility and persistence (Nijstad et al. 2010), or (c) as one property of the cognitive system (e.g., modification of processes based on a change in task demands (Deák 2004)). Ionescu (2012) unified these views and ideas by describing flexibility as a general characteristic that is based on the interaction of knowledge, attention, monitoring, and executive functions with environmental cues such as task demands. When studied as an aspect of the ability to think divergently, flexibility is usually assessed in two instantiations: ideational flexibility and associative flexibility. Ideational flexibility describes the use of a variety of ideational categories (Guilford 1967; Runco 1986; Runco and Okuda 1991), whereas associative flexibility includes the ability to build a diversified association chain (Benedek et al. 2012b). However, in terms of understanding flexibility as an aspect of divergent thinking, it is predominantly conceptualized as ideational flexibility, and we will restrict our discussion to this aspect of flexibility.
Ideational flexibility is a prominent aspect in models of cognitive abilities (e.g., Carroll 1993) and features saliently in theories of creativity (e.g., Nijstad et al. 2010). Factor analytic studies of creative-thinking abilities (Wilson et al. 1954) list flexibility as a key aspect of idea production and creativity. It is described as the “adaptability to changing instructions”, the “freedom to inertia of thought”, and the “spontaneous shifting of sets” (Wilson et al. 1954, p. 298). In models of cognitive ability, flexibility is mostly described as an aspect of the general retrieval ability in the area of idea generation/production—such as in the Cattell-Horn-Carroll model (CHC, McGrew 2009). Within Carroll’s (1993) discussion of abilities in the domain of idea generation, several flexibility factors are listed in addition to fluency and originality factors. They date back to measures proposed in the Kit of Reference Tests for Cognitive Factors (Ekstrom et al. 1976; French et al. 1963) as summarized in Table 1. These factors all tap ideational flexibility and include the “shifting of sets” (Wilson et al. 1954) while differing between verbal and figural content.
The tests listed in Table 1 were all scored with regard to the number of spontaneous changes from one category to another (within the predefined categories (Carroll 1993)). The same scoring schemes of the predefined categories are still used in contemporary divergent thinking tasks that are instructed in a hybrid way for fluency as well as flexibility (e.g., Jäger et al. 1997; Reiter-Palmon et al. 2019) or specifically for flexibility (Runco and Okuda 1991). In fact, these tasks all seem inspired by the factor descriptions listed in Table 1 or the task materials published either by Wilson et al. (1954) or within the French Kits.
As described above, one idea for understanding flexibility is viewing it as a function of flexibility and persistence within a task. Both determinants are deemed essential for creativity and can mutually compensate each other. Originality can then emanate from the number of categories generated in a task but also from persisting in developing ideas within the categories in more depth (e.g., Nijstad et al. 2010, Zhang et al. 2020). Flexibility (and persistence) can therefore be understood as concepts that promote the generation of an original set of ideas (Nijstad et al. 2010). Flexibility is deemed a cognitive process that reflects the ease of subjects to switch to a different approach or different perspective. Persistence is defined as a process that reflects the degree of task-directed focused maximal cognitive effort and how this effort is sustained over time and distraction. Flexibility and persistence are seen as being substantially due to attention control and working memory—the former presumably being more strongly implied in the case of flexibility. Within models of divergent thinking, the flexibility of a person is usually assessed through the number of predefined categories used in a response. Persistence can be expressed as the mean number of solutions generated within a response category. Obviously, both persistence and flexibility are strongly related to fluency measures of the same task. In fact, multiplying the persistence with the flexibility score should deliver the sum of all solutions generated. In this sense, the dual pathway model is not a theory of creativity but a simple decomposition of fluency performance.
Previous factor analytical considerations have also strengthened the conclusion that meaningful individual differences can be equally found in fluency, originality, and flexibility (e.g., Carroll 1993; Wilson et al. 1954). However, the prominent usage of solely creative fluency indicators in the assessment of divergent thinking (Hargreaves and Bolton 1972) as well as the confounding effects of fluency in assessing originality and flexibility (e.g., Forthmann et al. 2020) appear to point to the direction that fluency is—in terms of reliability, validity, and scoring—a better indicator for capturing individual differences than originality or flexibility.
From a psychometric perspective, a key feature of flexibility is the approach to performance appraisals that stresses the diversity of responses. Evidently, unusual procedures—for instance, instructions that are incongruent with the performance appraisal—could be subject to explicit modelling. Evidently, fluency and originality are best understood as dispositional trait factors, whereas different methods of instructing subjects—for instance, performance appraisals that apply fluency scoring to an originality task—could be incorporated into overarching measurement models as so-called method factors (Eid et al. 2017; Weiss et al. 2021b). The strength of such method factors could provide us with an idea of how strongly such inconsistencies can affect the measurement of fluency and originality. Equally plausible, the instruction to provide persistent responses in a number of flexibility tests could also strengthen the relevance of some performance aspects relative to others.
In sum, ideational flexibility has been theoretically and empirically linked to divergent thinking (fluency and originality)—going back to several cognitive theories—but also to creative and original outcomes (Benedek et al. 2012b; Guilford 1967; Johnson-Laird 1988; Nijstad et al. 2010). Therefore, it is not surprising that flexibility and other aspects of divergent thinking—such as fluency and originality—are highly related. However, the correlations reported are so strong that the existence of flexibility beyond a less complex factor for fluency is questionable. In the literature, such high correlations with fluency have often led to understanding divergent thinking as unidimensional (e.g., Weiss et al. 2021a) or as evidence in favor of the so-called equal odds that describe the number of qualitative creative ideas as a linear function of fluency (e.g., Forthmann et al. 2020). Scores, such as those reported in Benedek et al. (2012a), that present a correlation on the manifest level of r = .86 between ideational fluency and ideational flexibility might be highly fluency confounded. Please note, that in Benedek and colleagues (and, unfortunately, many more studies) the tasks were scored so that fluency, originality, and flexibility were all assessed in any one task (Reiter-Palmon et al. 2019). Obviously, this causes stochastic dependencies that are at odds with standard psychometric models. More importantly, instructions to participants are inevitably opaque. Participants are often not aware if they should demonstrate fluency, originality, or flexibility. Often, they are simply instructed to be “creative” and are then coded for fluency, originality, and flexibility at the same time. Such unwanted stochastic dependencies can be avoided by instructing tasks only regarding one of these aspects of divergent thinking. Other studies replicated the strong relations between fluency and flexibility (e.g., r = .98, Shen et al. 2018) but also share this operational opaqueness. The key issue is that the instruction (e.g., be flexible) is mostly not congruent with the scoring (e.g., flexibility) (Reiter-Palmon et al. 2019). Therefore, published evidence can be deemed insufficient for estimating the amount of variance in flexibility measures that is accounted for by other factors such as originality and fluency. Other things being equal, more conclusive evidence should be based on tasks in which the scoring is congruent with the instruction given to participants.
Theoretically, fluency and ideational flexibility tasks are akin to one another. This means that, even though they capture different aspects of divergent thinking (quantity vs. diversity), they are theoretically related as diversity without quantity is not possible. It is also the case that a key feature—i.e., to produce a solution or an idea—is the same in both tasks, and instead of quickly producing a chain of conceptually similar responses, flexibility instructions stress diversity. Therefore, both concepts should be correlated and should show some uniqueness. In fact, once a categorization for responses is developed, fluency tasks can be instructed as flexibility tasks. Similarly, ideational flexibility tasks can be instructed as ordinary fluency tasks if the quantity of solutions is stressed and the diversity of responses is omitted from the instructions.
Besides, the ability to retrieve information from long-term storage is crucial for creative fluency and hence also for retrieving information from different categories in flexibility (Carroll 1993; Silvia et al. 2013; Unsworth et al. 2013). Verbal fluency and retrieval can be understood as hinging upon four ingredients (Rosen and Engle 1997). First, activation spreads in long-term memory starting with a cue delivered in the instruction. In flexibility tasks, more than one chain of activation might need to be pursued, thereby making such tasks harder, at least by virtue of putting a stronger load on cognitive control and working memory. Second, as a task progresses, more and more generated responses must be monitored to avoid repetitions and omissions. These executive functions are supposedly harder in tasks in which several chains of solutions are pursued, i.e., if solutions across a number of categories are generated instead of exhausting solutions from one category. Third, subjects will usually adhere to Grice’s maxims (Grice 1975) and avoid useless repetitions, although they are often not instructed as such. Therefore, generated solutions must be supervised, and this activity might put a higher strain on working memory if solutions are generated from different response categories. This should be the case because the response categories also need supervision in flexibility tasks apart from the solutions. The fourth ingredient is the self-generation of category cues to access new items (Rosen and Engle 1997). This includes that previously retrieved answers must be suppressed in order to access new answer categories. If this is not done, someone will resample already used categories and only generate answers within the same category.
Taken together, verbal fluency partly depends on working memory. In contrast, ideational flexibility should depend more strongly on working memory. Participants who are low in working memory should have more difficulties than those with high working memory in successfully monitoring and supervising solutions. They should also struggle to maintain response chains and the self-generation of cues. Given that the role of working memory is strengthened in flexibility tasks relative to fluency tasks, the difference between subjects with low and high working memory should be of higher magnitude in flexibility tasks. Similarly, mental speed should be more highly related with flexibility than with fluency. Flexibility tasks require more complex cognitive processes—maintaining response chains, etc.—that are not only demanding working memory capacity but also mental speed. Previous research has shown that general task complexity is related with mental speed and that this complexity is best understood as working memory requirements (e.g., Goecke et al. 2021; Sheppard and Vernon 2008). This implies that more complex tasks—as they require a greater amount of working memory—should show a higher correlation with mental speed than less complex tasks. An assumption based on that idea would be that flexibility, as the more complex task, is more strongly related with mental speed than fluency. However, on the contrary, Forthmann et al. (2019) have shown that general retrieval—which can be seen as the less complex task—is more strongly related with general mental speed than creative ideation (i.e., divergent thinking).

1.2. The Present Study

The two contrasting perspectives we wanted to present, therefore, allowed contrasting predictions. For example, suppose flexibility primarily constitutes a method factor. In that case, the relation between fluency and flexibility would be below unity, and the residual of a latent flexibility factor would show no meaningful relations with relevant covariates (working memory, mental speed, and creative activities). Similar ideas have been published in terms of the equal odds studying the confounding effects of fluency (Forthmann et al. 2020). Based on the ideas of equal odds, (Forthmann et al. 2020), a flexibility ratio score that is uncorrelated with fluency would fit the idea of such a method factor if both fluency and flexibility were scored on the same task. In the present paper, we used different tasks for assessing fluency and flexibility; therefore, the application of an equal odds scoring approach—despite its theoretical proximity to our ideas—could not be directly transferred to this analysis.
From the contrasting perspective, flexibility constitutes a lower-order factor of general retrieval ability just like fluency and originality, which all contain unique trait variance (Carroll 1993). As a result, the relationship between verbal fluency and flexibility should be below unity, and the residual of a latent flexibility factor should show meaningful relations with working memory. In order to overcome the obstacles of previous studies, we used instructions that were congruent with the scoring of all fluency, originality, and flexibility tasks. Based on structural equation modeling, we were able to test the above-described contrasting predictions. If flexibility is a unique construct, it is predicted by fluency/originality only to a limited degree. Based on the literature that reports correlations between flexibility and both working memory and mental speed, we would further assume that the flexibility factor shows significant relations with these constructs if it contains unique trait variance. Therefore, we first reported the correlations between fluency, originality, flexibility, working memory, and mental speed on a manifest level to replicate previous findings. Next, we went beyond previous research by testing if flexibility has unique trait variance beyond fluency/originality in a structural equation model.

2. Method

The following sections provide information on the study design, sample, and measures. The data and scripts that can be used to reproduce all analyses can be found online in OSF repository [https://osf.io/kh49m/].

2.1. Design and Participants

The current analysis was based on a larger study that included two studies conducted in three German cities (see Weiss et al. 2021a). The study was conducted in a computerized manner in computer laboratories. The lab session included—depending on the city—two to five hours of testing. All tasks used in this manuscript were administered in all cities. In Bamberg and Ulm, the test battery included five hours of testing and a variety of further covariates (e.g., the broad measurement of crystallized intelligence, insight tasks, personality, etc. (for a comprehensive list, see Weiss et al. 2021a)), while the test battery in Greifswald only included the creativity tasks and some intelligence indicators as this study was complemented by an EEG measurement (Kaur et al. 2020, 2021). The five-hour testing included various small and larger breaks to prevent fatigue. Additionally, the participants completed another two hours of online tests at home. The study at hand only included the data collected in the lab sessions. Included are various indicators for divergent thinking (fluency and originality (see Weiss et al. 2021a)), working memory capacity (Schmitz et al. 2018; Wilhelm et al. 2013), mental speed (Schmitz and Wilhelm 2016), flexibility (Schoppe 1975), and creative activities (Diedrich et al. 2018) that were used for the current analysis. Tests and questionnaires that were not subject to this study were not further described but can be found in Weiss et al. (2021a) and Goecke et al. (2020).
The participants were recruited through various channels (e.g., mailing lists and announcements in public places) and received monetary rewards for their participation in the study. The original sample included N = 457 participants. Due to missingness in the flexibility tasks, N = 33 participants were excluded from the analysis. The remaining sample N = 424 was cleaned regarding multivariate outliers. This procedure was performed using the Mahalanobis distance (see Meade and Craig 2012) which shows the standardized distance of a data point from the mean of the multivariate distribution. The Mahalanobis distance was calculated for multivariate outliers in flexibility, working memory, and mental speed. We excluded n = 15 outliers that showed a Mahalanobis distance > 15. The outliers were mostly low performers in the working memory and mental speed tasks who arguably had problems understanding the instructions.
The final sample used that was reported here included a total of N = 409 participants. A total of 31% of the participants were male, and the mean age of the sample was 24.01 years (ranging from 18 to 49 years (Hartshorne and Germine 2015)).

2.2. Measures

2.2.1. Fluency and Originality

Fluency/originality as indicators of divergent thinking were assessed based on six tasks. Four tasks were indicators for fluency (including one figural fluency indicator), and two tasks were indicators for originality. The tasks were all open-ended and scored by two or three human raters (Reiter-Palmon et al. 2019). Descriptive statistics and intraclass correlations (ICC; Shrout and Fleiss 1979) can be found in Weiss et al. (2021a). The human rating included counting appropriate answers whenever the task was instructed for fluency (Reiter-Palmon et al. 2019). In the originality tasks, the human raters were trained to score each answer on a five-point Likert scale with regard to the uniqueness/uncommonness, remoteness, and cleverness of an answer (Silvia 2008; Silvia et al. 2009). Absent or inappropriate answers were coded as zero. Missing values in single tasks were due to computer problems and were deemed to be missing completely at random. The interrater reliability coefficients were mostly high, so aggregated scores across the human raters were used. In the following sections, the tasks measuring fluency and originality are described in more detail.
Fluency. All fluency tasks were open-ended and had a time limit. Participants were instructed to provide as many appropriate answers as possible within the given time. The similar attributes tasks (e.g., “Name as many things that you can that are ‘uneatable for humans’”) are similar to the alternate uses tasks and were based on items from the verbal creativity test (Schoppe 1975). Inventing names for abbreviations (e.g., “Invent names for the abbreviation: ‘T-E-F’”) was also adapted from Schoppe (1975). Another task was adapted and translated from the Kit of Reference for Cognitive Factors (Ekstrom et al. 1976). This task measured retrieval fluency (e.g., “Name as many household items”). Figural fluency comprised four paper–pencil tasks (e.g., “Draw as many objects as you can be based on a circle and a rectangle”) that were taken from the Berliner Intelligenzstruktur-Test für Jugendliche: Begabungs- und Hochbegabungsdiagnostik (Berlin Structure-of-Intelligence test for Youth: Diagnosis of Talents and Giftedness; Jäger 2006).
Originality. Originality was measured based on two open-ended indicators. In both tasks, the participants were instructed to provide a single answer that was very unique and original. Combining objects (e.g., “Combine two objects in order to build a door stopper in your house”) was adapted from the Kit of Reference Tests for Cognitive Factors (Ekstrom et al. 1976) and translated from English to the German language. The other task provoked the production of original nicknames (e.g., “Invent a combining objects for a bathtub”) and was adapted from (Schoppe 1975).

2.2.2. Creative Flexibility

Flexibility was assessed based on six items from the four-word sentence task (Schoppe 1975). In this task, four letters were provided (e.g., T-G-F-U) that should be used to build a four-word sentence using words that begin with these letters (e.g., “Theo goes up furiously”/“Ulrike fears the goose”). The letters could be used in different orders, and no filler words were allowed. The participants were instructed to come up with as many different sentences as they could in a limited time period.
The task was coded for flexibility by three human raters. Flexibility can be scored by counting the number of predefined categories into which responses of a participant can be classified (Reiter-Palmon et al. 2019) or by counting the number of switches between such categories (Nusbaum and Silvia 2011). In the present study, we counted the number of categories used. Sentences were coded as flexible solutions if they fell semantically, syntactically, and/or in terms of content in different categories. Sentences from participants that were not classified into a novel category or sentences that did not adhere to the instructions (number of words, etc.) were coded as zero. Table 2 shows the items that were used along with the mean values of flexibility and the inter-rater reliability (Shrout and Fleiss 1979).

2.2.3. Working Memory Capacity

Working memory was measured using a recall-1-back task (Schmitz et al. 2018; Wilhelm et al. 2013). These tasks were deployed using verbal stimuli (WMv) and figural stimuli (WMf). In the letter WMv task, the participants were presented with a letter in a box and were asked to type in that letter as soon as a new letter appeared. This was conducted with one to three boxes, implying that, with more boxes, the location of the previous letter had to be memorized next to remembering the current symbol. The tasks included a training phase with feedback and a test phase. In WMf, figures were displayed within a 3 × 3 matrix. Participants were instructed to indicate the figure that appeared last in the matrix at a given position while remembering the current stimulus. The task included a training phase with 21 trials and a test phase including 66 classifications. In both tasks, participants were, therefore, asked to identify the position where the same symbol occurred last while memorizing the current stimulus (see also Wilhelm et al. 2013).

2.2.4. Mental Speed

The mental speed tasks included two verbal indicators and one figural indicator based on the comparison tasks (Schmitz and Wilhelm 2016). In these tasks, we presented two triplets of figures or letters simultaneously. The participants were instructed to decide if both triplets were identical. The task consisted of two blocks of 40 trials each. We used the reciprocal reaction time as an indicator within the measurement models. This score displays the correct answers per time.

2.2.5. Inventory of Creative Activities and Achievement

The ICAA (Diedrich et al. 2018; Jauk et al. 2014) measures everyday creative activities in eight domains of creativity. The measurement includes the frequency of everyday creative activities based on the biographical assessment of the frequency of such behavior. As the ICAA’s long version showed problems regarding model fit, reliability, measurement invariance, and convergent validity, we used a short scale (S-ICA) that was compiled using the meta-heuristic ant colony optimization (Steger et al. 2022). The S-ICA included 8 items, one from each domain, and showed good model fit (CFI = .93; RMSEA = .05).

2.3. Statistical Analyses

Measurement Models

We computed several measurement models that were included in a larger structural model. For this, we used similar procedures to those described in Weiss et al. (2021a): for evaluating the fit of all models, we used the comparative fit index (CFI), the root mean square error of approximation (RMSEA), and the standardized root mean square residual (SRMR) (Hu and Bentler 1999). Applying these fit indices, CFI ≥ .95, RMSEA ≤ .06, and SRMR ≤ .08, indicated a very good fit. However, fit indices above CFI > .90 and RMSEA < .07 were acceptable. The statistical analysis was based on r software (version 3.6.2), using mostly the packages lavaan (Rosseel 2012) for all latent variable models and psych (Revelle 2018) for the outlier analysis and further descriptive statistics. All measurement models were estimated with the maximum likelihood (ML) estimator, and the structural model was estimated based on the robust maximum likelihood (MLR) estimator. We used the full information ML estimator to handle missing values (Schafer and Graham 2002). As a reliability estimate, we used McDonald’s ω (McDonald 1999; Raykov and Marcoulides 2011). The factor saturation (ω) for a factor indicates how much variance is accounted for by a latent variable in all underlying indicators (Brunner et al. 2012).
Fluency and originality were modeled using a one-factor model as the latent correlation between fluency and originality was extremely high (r = .75), and a model including two correlated factors did not fit the data significantly better than a one-factor model (see Weiss et al. 2020; Weiss et al. 2021a). The previous studies using this data also showed that a separated originality factor failed to show significant variance. Therefore, to reduce the complexity of the structural model, we decided to stick to a one-factor model that captured fluency/originality in a single latent factor. This model fits the data well (χ2(9) = 15.50; p = .08; CFI = .99; RMSEA = .04; and SRMR = .03). The reliability of the latent factor displaying fluency/originality was acceptable as ω = .74. Flexibility was modeled based on six items. The model fits the data well (χ2(9) = 18.72; p = .03; CFI = .99; RMSEA = .05; and SRMR = .02) and showed good reliability as ω = .89. The factor loadings were all > .74. The model is schematically displayed in Figure 1.

3. Results

Based on the assumption that participants low in working memory should have more difficulties than subjects high in working memory in successfully monitoring and supervising solutions and that they should also struggle more with maintaining response chains and the self-generation of cues, we first analyzed the correlations between flexibility, working memory, and mental speed on a manifest level. In the first step, we correlated all variables on a manifest level. The results are presented in the scatterplot in Figure 2. Flexibility was significantly related with all fluency and originality indicators, although the correlations were higher with fluency (indicators: sa, inv, rf, and fig) than with originality (indicators: co and ni). Replicating previous results, flexibility also showed significant correlations with both working memory indicators and the indicator for figural mental speed. These relations were also salient in a confirmatory factor model that included flexibility, working memory, and mental speed (χ2(34) = 47.37; p = .06; CFI = .99; and RMSEA = .03). Flexibility and mental speed were unrelated (r = .12), but working memory and flexibility did show a meaningful bivariate relation (r = .28).
In the next step, we estimated a larger model that included all of the variables in question. Here, flexibility was predicted by fluency/originality. To test if the variance left in flexibility was related with any of the covariates, we then correlated the residual of the latent flexibility factor with working memory, mental speed, and creative activities. As described above, if flexibility primarily constitutes a method factor, the relation between fluency and flexibility will be below unity, and the residual of a latent flexibility factor will show no meaningful relations with relevant covariates (working memory, mental speed, and creative activities). On the other hand, if flexibility constitutes a lower-order factor, it will contain unique trait variance. As a result, the relationship between verbal fluency and flexibility should be below unity, and the residual of a latent flexibility factor should show meaningful relations with working memory.
This model (Figure 3) fits the data well: χ2(244) = 423.84; p = .00; CFI = .93; RMSEA = .04; and SRMR = .06. Factor saturation was acceptable (ωflu/org = .70, ωFlex = .89, ωS-ICA = .71, ωWM = .49, and ωMS = .86). Fluency/originality explained 61% of the variance in creative flexibility. As a further measure of creativity, the self-reported creative activities were not significantly related to working memory and mental speed but, as expected, were related with fluency/originality (r = .44). Fluency/originality, on the other side, was moderately correlated with mental speed (r = .20) and working memory (r = .29). The residual of the latent flexibility factor was not significantly related to any of the covariates (working memory, mental speed, or creative activities). Figure 3 displays the structural model, including these predictions and correlations.
A very similar pattern was found when the model presented in Figure 3 was estimated without the originality indicators: χ2(201) = 348.80; p = .00; CFI = .94; RMSEA = .04; and SRMR = .06. In this case, a latent fluency factor predicted the flexibility factor slightly stronger (β = .81), while the residual of the latent flexibility factor was again not significantly related to any of the covariates (working memory, mental speed, or creative activities).

4. Discussion

Flexibility of thought is an important topic in creativity research. It has been described as a key ingredient of creativity (Nijstad et al. 2010), and it is theoretically akin to the fluency of ideas. However, flexibility has also been related with working memory capacity as it requires monitoring, suppression, and supervising responses (Rosen and Engle 1997). This raises the question of whether flexibility primarily constitutes a method factor or if it contains unique trait variance. In the next section, we summarized and interpreted our findings regarding flexibility.

4.1. Flexibility—A Method Factor

In previous studies, ideational flexibility has often been studied detached from other aspects of divergent thinking. For example, Benedek and colleagues found that ideational flexibility is significantly predicted by intelligence and inhibition (Benedek et al. 2012a). Their paper also assessed fluency, but they decided to reduce the model complexity by not modeling fluency and flexibility in one model as they were very highly correlated. Their finding underpins the theoretical idea that fluency and flexibility are akin to one another. Indeed, in many instances, fluency tasks can be turned into flexibility measures (and vice versa) by modifying instructions. It is also important that very strongly related variables are—ceteris paribus—unlikely to show divergent relations with other variables.
In the first step, we replicated previously found correlations on a manifest level and found that flexibility is related with divergent thinking, working memory, and mental speed. This finding suggested that participants low in working memory have more difficulties than subjects high in working memory in successfully monitoring and supervising solutions, and they should also struggle more with maintaining and updating previously generated responses and with the self-generation of cues to access new possible answers (Rosen and Engle 1997). Despite the initial assumption that mental speed is important in flexibility, because of the task complexity (Goecke et al. 2021), we did not find a significant relationship between flexibility and mental speed on a latent level. This finding converged with previous studies that report no or only small relations between mental speed and fluency/originality (e.g., Weiss et al. 2021a). On the other side, other studies report shared variance between mental speed and divergent thinking (Forthmann et al. 2019; Preckel et al. 2011). Indeed, the unspeeded relative to the speeded administration of divergent thinking tasks improved the quality of ideas (Forthmann et al. 2018). In the present data, fluency and flexibility tasks were speeded, but the originality tasks were unspeeded.
Given the theoretical proximity of fluency/originality and flexibility, previous results showing a high correlation between them might be misleading as fluency/originality are often not controlled for (e.g., Benedek et al. 2012a). Therefore, we next tested if flexibility can be subsumed below fluency/originality. The results showed that flexibility is strongly but not perfectly predicted by fluency/originality. This finding matched previous results (Benedek et al. 2012a). In an extended model, we were able to show that the residual of a latent flexibility factor shows no meaningful relations with working memory, mental speed, and creative activities. Given that flexibility could be subsumed below a more general fluency/originality factor and that its specificity was unrelated to previously suggested covariates, we suggested that the most adequate interpretation of flexibility is that it represents one out of a number of task classes of divergent thinking that did not show a meaningful specificity in the present study. Our findings, therefore, suggested that ideational flexibility is best understood as a method factor with no incremental validity.

4.2. Future Directions of Research

The findings implied that the aspects of divergent thinking—as discussed and measured hitherto—might need some revision. Creative fluency is the most prominent factor in previous research and this manuscript, and previous research shows that originality and flexibility are of limited utility over and above fluency (Hargreaves and Bolton 1972). However, a number of articles have provided evidence for the theoretical and practical importance of originality and flexibility, and based on novel automated scoring approaches (e.g., Reiter-Palmon et al. 2019) or the adaption of older scoring approaches (e.g., Forthmann and Doebler 2022), an incremental utility of originality and flexibility is conceivable. This implies that either the tasks or associated performance appraisal standards must be improved in order to better capture unique variance of flexibility and originality. Alternatively, the concept of divergent thinking should be revised towards a construct that is best depicted based on indicators of fluency. Such a revision would imply losing theoretically important aspects of creativity, such as the creation of novel ideas, and would therefore narrow the construct to retrieving previously learned information. Prior to discarding flexibility from divergent thinking, we propose a number of interventions to strengthen the flexibility measurement.
Even though this multivariate study included a variety of indicators that assessed fluency and originality, the indicator that was used to assess flexibility was limited to a single task with six items (four-word sentences; Schoppe 1975). This limited the findings as the employment of a single paradigm exacerbates the distinction between trait and method variance. Methodological specificities of this ideational flexibility paradigm might be crucial for the results reported. Therefore, administering additional flexibility measures, preferably also including associational flexibility (Benedek et al. 2012b), the flexibility of use, and indicators for figural flexibility (Ekstrom et al. 1976; French et al. 1963) in a replication and extension of the present results, would deliver more conclusive results concerning the utility of flexibility. Studying a broader range of flexibility tasks would allow the testing of whether or not flexibility is best understood as a flexibility method factor below an overarching fluency/originality factor. If we pursue such research, we should be open to the result that the diversity of responses (usually collected in measures that could serve as fluency tasks if instructed as such) is not essential for capturing individual differences in divergent thinking beyond fluency/originality. In this case, it would be better to focus on originality as that facet seems to build the true core of creativity, but it often has limitations in terms of measurement. The conclusion that flexibility does not capture variance beyond fluency/originality does not necessarily contradict theories such as the dual pathway theory (Nijstad et al. 2010). This model could—for instance—be further addressed by assessing if flexibility mediates the relationship between fluency and originality. On the other side, we should be open to the perspective that the concept of flexibility is mostly reflecting increased demands on attention control and working memory rather than performance aspects that should be deemed intrinsic to creativity. If that were the case, dismissing flexibility from the realm of creativity would be the correct conclusion.
According to Ockham’s razor, flexibility tasks need to capture something relevant that goes beyond what is measured with fluency tasks; if not, it makes sense to stick to fluency tasks because they are the more basic concept, and they are cheaper to administer and score (Hargreaves and Bolton 1972). Costs of scoring are usually substantial, given that all responses in flexibility tasks require human attention. Studies that capitalize on technological developments to evaluate responses from flexibility tasks could greatly facilitate and thereby cut the cost of scoring flexibility tasks. One such technological approach builds upon a semantic analysis of answers (Cosgrove et al. 2021). In the semantic analysis, the participants’ responses can be used to generate semantic networks displaying the connectivity of various nodes (words) in the semantic space (Cosgrove et al. 2021; Li et al. 2021). Such lexical network structures can categorize answers and express flexibility based on different indices, such as clustering coefficients, average path lengths between clusters and words, network modularity, and network diameter (Boccaletti et al. 2006; Kenett et al. 2016). Similar approaches are imaginable for processing images by blending images or comparing various image vectors and properties with one another (Rafner et al. 2020).
The flexibility within lexical network structures was mostly restricted to single words in the last years. Additionally, most of these studies use fluency tasks (that are instructed for fluency), but they are later scored for the semantic flexibility of participants (Kenett et al. 2016). This implies that flexibility is assessed based on the responses given to a fluency task, which is incongruent with the instruction (Reiter-Palmon et al. 2019). Studying semantic networks in flexibility tasks with congruent instructions (i.e., participants are instructed to maximize the diversity of their responses) would be of great interest. In addition, most semantic network approaches calculate networks for different groups (e.g., low creativity and high creativity (Li et al. 2021) or low intelligence and high intelligence (Kenett et al. 2016)). Such a procedure requires an a priori set threshold that allows for splitting the sample into low and high ability groups. This comes along with a loss of information. Therefore, future lexical network studies should focus on dimensional approaches. For instance, flexibility networks could be established along the working memory continuum using locally weighted structural equation models (Hildebrandt et al. 2016).

4.3. Conclusions and Outlook

The central question of this manuscript was to assess if flexibility primarily constitutes a method factor or if it contains unique trait variance. Our study showed that the residual of a latent flexibility factor is not meaningfully predicted by working memory, mental speed, and creative activities. This was in line with the idea that flexibility is best understood as one out of a number of divergent thinking concepts and that its specificity is not predictively useful (i.e., it is a method factor). As measured in this study, flexibility had no incremental validity above and beyond the broad assessment of fluency and originality. Therefore, flexibility mostly appeared as a twist to the more prominent fluency retrieval tasks but as a twist with an increased price tag attached to its use. Future research could either replicate and extend the present findings or could use technological tools to facilitate economic use of flexibility measures. We see potential in the analysis of semantic flexibility that captures a variety of semantic indices. Such an analysis should allow for dimensional considerations of variables (such as intelligence). Within such a novel approach, performance indicators, such as clustering coefficients, average path lengths between clusters and words, and network modularity, might be uniquely predicted by working memory and mental speed. Such indices have the potential to capture the supervision of generated answers, the suppression of previously used categories, and the activation of different category chains that might not all be perfectly captured in the traditional human count data.

Author Contributions

Conceptualization, S.W. and O.W.; methodology, S.W. and O.W.; formal analysis, S.W.; investigation, S.W., resources, O.W.; data curation, S.W.; writing—original draft preparation, S.W.; writing—review and editing, O.W.; visualization, S.W.; supervision, O.W.; project administration, S.W.; funding acquisition, O.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study did not require further ethical approval in accordance with the declaration of Helsinki.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

We reported all data exclusions. The decisions surrounding final sample sizes were presented in the sample sections. The research objectives and hypothesis were not preregistered. The data and all required scripts are available at https://osf.io/kh49m/.

Acknowledgments

We thank Andrea Hildebrandt, Ulrich Schroeders, Yadwinder Kaur and Diana Steger for recruitment and data collection, as well as helping conceptualizing creativity measurement. Besides, we thank all research assistants who helped during data acquisition and human coding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Batey, Mark, and Adrian Furnham. 2006. Creativity, Intelligence, and Personality: A Critical Review of the Scattered Literature. Genetic, Social, and General Psychology Monographs 132: 355–429. [Google Scholar] [CrossRef] [PubMed]
  2. Benedek, Mathias, Fabiola Franz, Moritz Heene, and Aljoscha C. Neubauer. 2012a. Differential effects of cognitive inhibition and intelligence on creativity. Personality and Individual Differences 53: 480–85. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Benedek, Mathias, Tanja Könen, and Aljoscha C. Neubauer. 2012b. Associative abilities underlying creativity. Psychology of Aesthetics, Creativity, and the Arts 6: 273–81. [Google Scholar] [CrossRef]
  4. Boccaletti, Stefano, Vito Latora, Yamir Moreno, Martin Chavez, and D-U. Hwang. 2006. Complex networks: Structure and dynamics. Physics Reports 424: 175–308. [Google Scholar] [CrossRef]
  5. Brunner, Martin, Gabriel Nagy, and Oliver Wilhelm. 2012. A Tutorial on Hierarchically Structured Constructs: Hierarchically Structured Constructs. Journal of Personality 80: 796–846. [Google Scholar] [CrossRef]
  6. Carroll, John B. 1993. Human Cognitive Abilities: A Survey of Factor-Analytic Studies. Cambridge: Cambridge University Press. [Google Scholar]
  7. Colzato, Lorenza S., Mariëtte Huizinga, and Bernhard Hommel. 2009. Recreational cocaine polydrug use impairs cognitive flexibility but not working memory. Psychopharmacology 207: 225–34. [Google Scholar] [CrossRef] [Green Version]
  8. Cosgrove, Abigail L., Yoed N. Kenett, Roger E. Beaty, and Michele T. Diaz. 2021. Quantifying flexibility in thought: The resiliency of semantic networks differs across the lifespan. Cognition 211: 104631. [Google Scholar] [CrossRef]
  9. Deák, Gedeon O. 2004. The Development of Cognitive Flexibility and Language Abilities. In Advances in Child Development and Behavior. Amsterdam: Elsevier, Bd. 31, S. 271–327. [Google Scholar] [CrossRef]
  10. Diedrich, Jennifer, Emanuel Jauk, Paul J. Silvia, Jeffrey M. Gredlein, Aljoscha C. Neubauer, and Mathias Benedek. 2018. Assessment of real-life creativity: The Inventory of Creative Activities and Achievements (ICAA). Psychology of Aesthetics, Creativity, and the Arts 12: 304–16. [Google Scholar] [CrossRef]
  11. Eid, Michael, Christian Geiser, Tobias Koch, and Moritz Heene. 2017. Anomalous results in G-factor models: Explanations and alternatives. Psychological Methods 22: 541–62. [Google Scholar] [CrossRef]
  12. Ekstrom, Ruth B., J. W. French, Harry Horace Harman, and D. Derman. 1976. Manual for Kit of Factor-Referenced Cognitive Tests. Princeton: Educational Testing Service. [Google Scholar]
  13. Feist, Gregory J. 1998. A meta-analysis of personality in scientific and artistic creativity. Personality and Social Psychology Review 2: 290–309. [Google Scholar] [CrossRef]
  14. Forthmann, Boris, and Philipp Doebler. 2022. Fifty years later and still working: Rediscovering Paulus et al.’s (1970) automated scoring of divergent thinking tests. Psychology of Aesthetics, Creativity, and the Arts. [Google Scholar] [CrossRef]
  15. Forthmann, Boris, Carsten Szardenings, and Heinz Holling. 2020. Understanding the confounding effect of fluency in divergent thinking scores: Revisiting average scores to quantify artifactual correlation. Psychology of Aesthetics, Creativity, and the Arts 54: 123–33. [Google Scholar] [CrossRef]
  16. Forthmann, Boris, Corinna Lips, Carsten Szardenings, Jana Scharfen, and Heinz Holling. 2018. Are Speedy Brains Needed when Divergent Thinking is Speeded-or Unspeeded? The Journal of Creative Behavior 54: 123–33. [Google Scholar] [CrossRef]
  17. Forthmann, Boris, David Jendryczko, Jana Scharfen, Ruben Kleinkorres, Mathias Benedek, and Heinz Holling. 2019. Creative ideation, broad retrieval ability, and processing speed: A confirmatory study of nested cognitive abilities. Intelligence 75: 59–72. [Google Scholar] [CrossRef]
  18. French, John W., Ruth B. Ekstrom, and Leighton A. Price. 1963. Manual for Kit of Reference Tests for Cognitive Factors (Revised 1963). Princeton: Education Testing Service. [Google Scholar]
  19. Goecke, Benjamin, Florian Schmitz, and Oliver Wilhelm. 2021. Binding Costs in Processing Efficiency as Determinants of Cognitive Ability. Journal of Intelligence 9: 18. [Google Scholar] [CrossRef]
  20. Goecke, Benjamin, Selina Weiss, Diana Steger, Ulrich Schroeders, and Oliver Wilhelm. 2020. Testing competing claims about overclaiming. Intelligence 81: 101470. [Google Scholar] [CrossRef]
  21. Grice, H. Paul. 1975. Logic and Conservation. In Speech Acts. New York: Academic Press, pp. 41–58. [Google Scholar]
  22. Guilford, Joy Paul. 1956. The Structure of Intellect. Psychological Bulletin 53: 267–93. [Google Scholar] [CrossRef] [PubMed]
  23. Guilford, Joy Paul. 1967. The Nature of Human Intelligence. New York: McGraw-Hill. [Google Scholar]
  24. Hargreaves, D. J., and N. Bolton. 1972. Selecting Creativity Tests for Use in Research. British Journal of Psychology 63: 451–62. [Google Scholar] [CrossRef]
  25. Hartshorne, Joshua K., and Laura T. Germine. 2015. When Does Cognitive Functioning Peak? The Asynchronous Rise and Fall of Different Cognitive Abilities Across the Life Span. Psychological Science 26: 433–43. [Google Scholar] [CrossRef]
  26. Hildebrandt, Andrea, Oliver Lüdtke, Alexander Robitzsch, Christopher Sommer, and Oliver Wilhelm. 2016. Exploring Factor Model Parameters across Continuous Variables with Local Structural Equation Models. Multivariate Behavioral Research 51: 257–58. [Google Scholar] [CrossRef]
  27. Hornberg, Jason, and Roni Reiter-Palmon. 2017. Creativity and the big five personality traits: Is the relationship dependent on the creativity measure? In The Cambridge Handbook of Creativity and Personality Research. Edited by Gregory J. Feist, Roni Reiter-Palmon and James C. Kaufman. Cambridge: Cambridge University Press, S. 275–93. [Google Scholar]
  28. Hu, Li-tze, and Peter M. Bentler. 1999. Cutoff criteria for fit indexes in covariance structure analysis: Conventional criteria versus new alternatives. Structural Equation Modeling: A Multidisciplinary Journal 6: 1–55. [Google Scholar] [CrossRef]
  29. Ionescu, Thea. 2012. Exploring the nature of cognitive flexibility. New Ideas in Psychology 30: 190–200. [Google Scholar] [CrossRef]
  30. Jauk, Emanuel, Mathias Benedek, and Aljoscha C. Neubauer. 2014. The Road to Creative Achievement: A Latent Variable Model of Ability and Personality Predictors: The road to creative achievement. European Journal of Personality 28: 95–105. [Google Scholar] [CrossRef] [PubMed]
  31. Jäger, Adolf Otto. 2006. Berliner Intelligenzstruktur-Test für Jugendliche Begabungs-und Hochbegabungsdiagnostik: BIS-HB. Göttingen: Hogrefe. [Google Scholar]
  32. Jäger, Adolf Otto, Heinz-Martin Süß, and André Beauducel. 1997. Berliner Intelligenzstruktur-Test: BIS-Test. Göttingen: Hogrefe. [Google Scholar]
  33. Johnson-Laird, P. N. 1988. Freedom and constraint in creativity. In The Nature of Creativity. Edited by Robert J. Sternberg. Cambridge: Cambridge University Press, S. 202–19. [Google Scholar]
  34. Kaur, Yadwinder, Guang Ouyang, Werner Sommer, Selina Weiss, Changsong Zhou, and Andrea Hildebrandt. 2020. What Does Temporal Brain Signal Complexity Reveal About Verbal Creativity? Frontiers in Behavioral Neuroscience 14: 146. [Google Scholar] [CrossRef] [PubMed]
  35. Kaur, Yadwinder, Selina Weiss, Changsong Zhou, Rico Fischer, and Andrea Hildebrandt. 2021. Exploring Neural Signal Complexity as a Potential Link between Creative Thinking, Intelligence, and Cognitive Control. Journal of Intelligence 9: 59. [Google Scholar] [CrossRef]
  36. Kenett, Yoed N., Roger E. Beaty, Paul J. Silvia, David Anaki, and Miriam Faust. 2016. Structure and flexibility: Investigating the relation between the structure of the mental lexicon, fluid intelligence, and creative achievement. Psychology of Aesthetics, Creativity, and the Arts 10: 377–88. [Google Scholar] [CrossRef]
  37. Li, Yangping, Yoed N. Kenett, Weiping Hu, and Roger E. Beaty. 2021. Flexible Semantic Network Structure Supports the Production of Creative Metaphor. Creativity Research Journal 33: 209–23. [Google Scholar] [CrossRef]
  38. McDonald, Roderick P. 1999. Test Theory: A Unified Treatment. Mahwah: L. Erlbaum Associates. [Google Scholar]
  39. McGrew, Kevin S. 2009. CHC theory and the human cognitive abilities project: Standing on the shoulders of the giants of psychometric intelligence research. Intelligence 37: 1–10. [Google Scholar] [CrossRef]
  40. Meade, Adam W., and S. Bartholomew Craig. 2012. Identifying careless responses in survey data. Psychological Methods 17: 437–55. [Google Scholar] [CrossRef]
  41. Monsell, Stephen. 2003. Task switching. Trends in Cognitive Sciences 7: 134–40. [Google Scholar] [CrossRef]
  42. Nijstad, Bernard A., Carsten KW De Dreu, Eric F. Rietzschel, and Matthijs Baas. 2010. The dual pathway to creativity model: Creative ideation as a function of flexibility and persistence. European Review of Social Psychology 21: 34–77. [Google Scholar] [CrossRef]
  43. Nusbaum, Emily, and Paul J. Silvia. 2011. Are intelligence and creativity really so different? Fluid intelligence, executive processes, and strategy use in divergent thinking. Intelligence 39: 36–45. [Google Scholar] [CrossRef] [Green Version]
  44. Preckel, Franzis, Christina Wermer, and Frank M. Spinath. 2011. The interrelationship between speeded and unspeeded divergent thinking and reasoning, and the role of mental speed. Intelligence 39: 378–88. [Google Scholar] [CrossRef]
  45. Rafner, Janet, Arthur Hjorth, Sebastian Risi, Lotte Philipsen, Charles Dumas, Michael Mose Biskjær, Lior Noy, Kristian Tylén, Carsten Bergenholtz, Jesse Lynch, and et al. 2020. crea.blender: A Neural Network-Based Image Generation Game to Assess Creativity. Paper presented at the 2020 Annual Symposium on Computer-Human Interaction in Play, Virtual, November 2–4; Extended Abstracts. pp. 340–44. [Google Scholar] [CrossRef]
  46. Raykov, Tenko, and George A. Marcoulides. 2011. Introduction to Psychometric Theory. London: Routledge. [Google Scholar]
  47. Reiter-Palmon, Roni, Boris Forthmann, and Baptiste Barbot. 2019. Scoring divergent thinking tests: A review and systematic framework. Psychology of Aesthetics, Creativity, and the Arts 13: 144–52. [Google Scholar] [CrossRef]
  48. Revelle, William. 2018. psych: Procedures for Psychological, Psychometric, and Personality Research. R package Version 1. Evanston: Northwestern University. [Google Scholar]
  49. Rosen, Virginia M., and Randall W. Engle. 1997. The Role of Working Memory Capacity in Retrieval. Jorunal of Experimental Psychology: General 126: 211–27. [Google Scholar] [CrossRef]
  50. Rosseel, Yves. 2012. Lavaan: An R package for structural equation modeling and more. Version 0.5–12 (BETA). Journal of Statistical Software 48: 1–36. [Google Scholar] [CrossRef] [Green Version]
  51. Runco, Mark A. 1986. Flexibility and Originality in Children’s Divergent Thinking. The Journal of Psychology 120: 345–52. [Google Scholar] [CrossRef]
  52. Runco, Mark A., and Shawn M. Okuda. 1991. The instructional enhancement of the flexibility and originality scores of divergent thinking tests. Applied Cognitive Psychology 5: 435–41. [Google Scholar] [CrossRef]
  53. Schafer, Joseph L., and John W. Graham. 2002. Missing data: Our view of the state of the art. Psychological Methods 7: 147–77. [Google Scholar] [CrossRef]
  54. Schmitz, Florian, and Oliver Wilhelm. 2016. Modeling Mental Speed: Decomposing Response Time Distributions in Elementary Cognitive Tasks and Correlations with Working Memory Capacity and Fluid Intelligence. Journal of Intelligence 4: 13. [Google Scholar] [CrossRef] [Green Version]
  55. Schmitz, Florian, Dominik Rotter, and Oliver Wilhelm. 2018. Scoring Alternatives for Mental Speed Tests: Measurement Issues and Validity for Working Memory Capacity and the Attentional Blink Effect. Journal of Intelligence 6: 47. [Google Scholar] [CrossRef] [Green Version]
  56. Schoppe, Karl-Josef. 1975. Verbaler Kreativitäts-Test-VKT: Ein Verfahren zur Erfassung verbal-produktiver Kreativitätsmerkmale. Göttingen: Verlag für Psychologie CJ Hogrefe. [Google Scholar]
  57. Shen, Wangbing, Bernhard Hommel, Yuan Yuan, Liu Chang, and Wei Zhang. 2018. Risk-Taking and Creativity: Convergent, but Not Divergent Thinking Is Better in Low-Risk Takers. Creativity Research Journal 30: 224–31. [Google Scholar] [CrossRef] [Green Version]
  58. Sheppard, Leah D., and Philip A. Vernon. 2008. Intelligence and speed of information-processing: A review of 50 years of research. Personality and Individual Differences 44: 535–51. [Google Scholar] [CrossRef]
  59. Shrout, Patrick E., and Joseph L. Fleiss. 1979. Intraclass Correlations: Uses in Assessing Rater Reliability. Psychological Bulletin 86: 420–28. [Google Scholar] [CrossRef]
  60. Silvia, Paul J. 2008. Creativity and Intelligence Revisited: A Latent Variable Analysis of Wallach and Kogan. Creativity Research Journal 20: 34–39. [Google Scholar] [CrossRef]
  61. Silvia, Paul J., Christopher Martin, and Emily C. Nusbaum. 2009. A snapshot of creativity: Evaluating a quick and simple method for assessing divergent thinking. Thinking Skills and Creativity 4: 79–85. [Google Scholar] [CrossRef] [Green Version]
  62. Silvia, Paul J., Roger E. Beaty, and Emily C. Nusbaum. 2013. Verbal fluency and creativity: General and specific contributions of broad retrieval ability (Gr) factors to divergent thinking. Intelligence 41: 328–40. [Google Scholar] [CrossRef]
  63. Steger, D., S. Weiss, and O. Wilhelm. 2022. The short inventory of creative activities (S-ICA): Compiling a short scale using ant colony optimization. Creativity Research Journal. [Google Scholar] [CrossRef]
  64. Unsworth, Nash, Gene A. Brewer, and Gregory J. Spillers. 2013. Working memory capacity and retrieval from long-term memory: The role of controlled search. Memory and Cognition 41: 242–54. [Google Scholar] [CrossRef]
  65. Weiss, Selina, Diana Steger, Ulrich Schroeders, and Oliver Wilhelm. 2020. A reappraisal of the threshold hypothesis of creativity and intelligence. Journal of Intelligence 8: 38. [Google Scholar] [CrossRef]
  66. Weiss, Selina, Diana Steger, Yadwinder Kaur, Andrea Hildebrandt, Ulrich Schroeders, and Oliver Wilhelm. 2021a. On the Trail of Creativity: Dimensionality of Divergent Thinking and Its Relation With Cognitive Abilities, Personality, and Insight. European Journal of Personality 35: 291–314. [Google Scholar] [CrossRef]
  67. Weiss, Selina, Oliver Wilhelm, and Patrick Kyllonen. 2021b. An improved taxonomy of creativity measures based on salient task attributes. Psychology of Aesthetics, Creativity, and the Arts. [Google Scholar] [CrossRef]
  68. Wilhelm, Oliver, Andrea Hildebrandt, and Klaus Oberauer. 2013. What is working memory capacity, and how can we measure it? Frontiers in Psychology 4: 433. [Google Scholar] [CrossRef] [Green Version]
  69. Wilson, Robert C., J. P. Guilford, Paul R. Christensen, and Donald J. Lewis. 1954. A factor-analytic study of creative-thinking abilities. Psychometrika 19: 297–311. [Google Scholar] [CrossRef]
  70. Zhang, Weitao, Zsuzsika Sjoerds, and Bernhard Hommel. 2020. Metacontrol of human creativity: The neurocognitive mechanisms of convergent and divergent thinking. NeuroImage 210: 116572. [Google Scholar] [CrossRef]
Figure 1. Schematic measurement model of fluency/originality (Model A) and flexibility (Model B).
Figure 1. Schematic measurement model of fluency/originality (Model A) and flexibility (Model B).
Jintelligence 10 00096 g001aJintelligence 10 00096 g001b
Figure 2. Correlations on a manifest level: flexibility, fluency, originality, working memory, and mental speed. Note: Statistically significant correlations are marked with * (p < .05), ** (p < .01), and *** (p < .001). Fluency indicators: sa = similar attributes; inv = inventing names; rf = retrieval fluency; and fig = figural fluency. Originality indicators: co = combining objects and ni = nicknames. WMC = working memory and MS = mental speed.
Figure 2. Correlations on a manifest level: flexibility, fluency, originality, working memory, and mental speed. Note: Statistically significant correlations are marked with * (p < .05), ** (p < .01), and *** (p < .001). Fluency indicators: sa = similar attributes; inv = inventing names; rf = retrieval fluency; and fig = figural fluency. Originality indicators: co = combining objects and ni = nicknames. WMC = working memory and MS = mental speed.
Jintelligence 10 00096 g002
Figure 3. Structural model: fluency/originality, flexibility, working memory, mental speed, and creative activities. Note: S-ICA = short inventory of creative activities; Distur. Flex. = residual of latent flexibility; WM = working memory; and MS = mental speed. Non-significant correlations between S-ICA and WM/MS are not depicted.
Figure 3. Structural model: fluency/originality, flexibility, working memory, mental speed, and creative activities. Note: S-ICA = short inventory of creative activities; Distur. Flex. = residual of latent flexibility; WM = working memory; and MS = mental speed. Non-significant correlations between S-ICA and WM/MS are not depicted.
Jintelligence 10 00096 g003
Table 1. Flexibility factors from the Kit of Reference Tests for Cognitive Factors (Ekstrom et al. 1976; French et al. 1963).
Table 1. Flexibility factors from the Kit of Reference Tests for Cognitive Factors (Ekstrom et al. 1976; French et al. 1963).
FactorDescription
Figural Adaptive Flexibility“The ability to change set in order to meet new requirements imposed by figural problems” (p. 49)
Semantic Spontaneous Flexibility“The ability to produce a diversity of verbally expressed ideas in a situation that is relatively unrestricted” (p. 50)
Figural Flexibility“The ability to change set in order to generate new and different solutions to figural problems” (p. 73)
Flexibility of Use“The mental set necessary to think of different uses for objects” (p. 197)
Table 2. Descriptive statistics for flexibility indicators, including ICCs.
Table 2. Descriptive statistics for flexibility indicators, including ICCs.
ItemMean (SD)MinMaxICC
#1 (T-G-F-U)3.08 (1.69)08.99
#2 (B-H-K-N)3.72 (1.96)010.99
#3 (R-D-T-B)2.87 (1.73)09.67.99
#4 (S-L-G-H)3.47 (1.84)010.98
#5 (A-P-F-K)2.82 (1.63)08.99
#6 (M-W-U-N)2.97 (1.73)08.33.99
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Weiss, S.; Wilhelm, O. Is Flexibility More than Fluency and Originality? J. Intell. 2022, 10, 96. https://0-doi-org.brum.beds.ac.uk/10.3390/jintelligence10040096

AMA Style

Weiss S, Wilhelm O. Is Flexibility More than Fluency and Originality? Journal of Intelligence. 2022; 10(4):96. https://0-doi-org.brum.beds.ac.uk/10.3390/jintelligence10040096

Chicago/Turabian Style

Weiss, Selina, and Oliver Wilhelm. 2022. "Is Flexibility More than Fluency and Originality?" Journal of Intelligence 10, no. 4: 96. https://0-doi-org.brum.beds.ac.uk/10.3390/jintelligence10040096

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop