Summary. An analysis of performance on the Jouve–Cerebrals Test of Induction (JCTI) and four GAMA subtests (Matching, Analogies, Sequences, Construction) points to a single dominant source of individual differences rather than two separate abilities. With N = 118, factor-analytic evidence favors a general reasoning factor that subsumes both spatial–temporal and abstract problem-solving demands. Any apparent “two-factor” pattern is better explained by task-specific variance and sampling noise than by distinct latent abilities.
Background
The study set out to examine how nonverbal tasks cohere psychometrically. Although it is intuitive to split reasoning into spatial–temporal manipulation versus abstract relation finding, the question is empirical: do the data support multiple dimensions once the shared variance among tests is modeled?
Key Insights
“Two factors” do not hold up.
- One general factor dominates. Principal-axis factoring with parallel analysis retained a single factor that explained roughly 42% of observed variance. Loadings were substantial across tasks (e.g., Construction > JCTI > Analogies ≈ Sequences > Matching), indicating broad overlap in what these measures capture.
- “Two factors” do not hold up. Forcing a two-factor solution produced weaker fit, near-zero correlations between the putative factors, and notable cross-loadings (e.g., Analogies loading on both). This pattern signals an unstable partition rather than meaningful separable constructs.
- Task flavor ≠ distinct ability. Construction shows the highest saturation with the general factor, likely because spatial visualization is highly g-loaded. Matching shows the lowest loading but still aligns with the same latent dimension. Differences across subtests reflect measurement emphasis, not separate abilities.
Significance
For practitioners, the safest interpretation is that these tasks index a common nonverbal reasoning capacity. Total scores are therefore more defensible than fine-grained “profiles” carved from a small set of indicators. In educational and clinical contexts, this supports using a compact battery like JCTI + selected GAMA subtests to obtain a stable index of general reasoning without overinterpreting subtest scatter.
Future Directions
Use confirmatory models (bifactor, correlated factors) and report omega-hierarchical and explained common variance to quantify general-factor saturation.
- Replicate with larger, demographically diverse samples and add multiple indicators per hypothesized facet (≥3 per factor) to test whether reliable subdimensions emerge when the battery is expanded.
- Use confirmatory models (bifactor, correlated factors) and report omega-hierarchical and explained common variance to quantify general-factor saturation.
- Compare predictive utility of total versus putative subscale scores for outcomes such as STEM coursework, technical training performance, and complex problem-solving tasks.
Conclusion
The data do not warrant splitting performance into spatial–temporal and abstract reasoning factors. A single, robust general reasoning factor accounts for the common variance across JCTI and GAMA tasks; any residual differences look task-specific rather than factorially distinct.
Reference
Jouve, X. (2018). Exploring underlying factors in cognitive tests: Spatial-temporal reasoning and abstract reasoning abilities. Cogn-IQ Research Papers. https://pubscience.org/ps-1mFWV-3f180b-jGlP
People Also Ask
What are explore the validity and reliability of the jcti, and its strong correlations with sat math and rist scores.?
The Jouve–Cerebrals Test of Induction (JCTI) is a nonverbal measure of inductive reasoning. Using data from N = 2,306 examinees, this study assessed score reliability and concurrent validity against external benchmarks. Findings indicate stable internal consistency and strong convergence with quantitative and nonverbal indicators, supporting use in educational and vocational decision-making.
Read more →What are evaluating factor retention in exploratory factor analysis?
Determining the optimal number of factors to retain in exploratory factor analysis (EFA) has long been a subject of debate in social sciences research. Finch (2020) addresses this challenge by comparing the performance of fit index difference values and parallel analysis, a well-established method in this field. The study offers valuable insights into how these approaches perform under varying conditions, particularly with categorical and normally distributed indicators.
Read more →What are cognitive abilities, not math skills, predict wealth for preterm adults?
This study, authored by Jaekel et al. (2019), examines the relationship between being born very preterm (VP) or with very low birth weight (VLBW), cognitive abilities, and wealth accumulation in adulthood. By tracking participants from birth to 26 years of age, the research provides key insights into how early cognitive abilities influence long-term economic outcomes.
Read more →What is the underlying dimensions of cognitive abilities: a multidimensional scaling analysis of jcces and gama subtests?
This study aimed to investigate the relationships between tasks of the Jouve Cerebrals Crystallized Educational Scale (JCCES) and General Ability Measure for Adults (GAMA) using multidimensional scaling (MDS) analysis. The JCCES measures Verbal Analogies, Mathematical Problems, and General Knowledge, while the GAMA assesses nonverbal cognitive abilities through Matching, Analogies, Sequences, and Construction tasks. A total of 63 participants completed both assessments. MDS analysis revealed a 2-dimensional solution, illustrating a diagonal separation between nonverbal and verbal abilities, with Mathematical Problems slightly closer to the verbal side. Seven groups were identified, corresponding to distinct cognitive processes. The findings suggest that JCCES and GAMA tasks are not independent and share common underlying dimensions. This study contributes to a more nuanced understanding of cognitive abilities, with potential implications for educational, clinical, and research settings. Future research should address the study's limitations, including the small sample size and potential methodological constraints.
Read more →Why is background important?
The study set out to examine how nonverbal tasks cohere psychometrically. Although it is intuitive to split reasoning into spatial–temporal manipulation versus abstract relation finding, the question is empirical: do the data support multiple dimensions once the shared variance among tests is modeled?
How does key insights work in practice?
One general factor dominates. Principal-axis factoring with parallel analysis retained a single factor that explained roughly 42% of observed variance. Loadings were substantial across tasks (e.g., Construction > JCTI > Analogies ≈ Sequences > Matching), indicating broad overlap in what these measures capture. “Two factors” do not hold up. Forcing a

