92 results on '"Abad FJ"'
Search Results
2. When proxy-driven learning is no better than random: The consequences of representational incompleteness.
- Author
-
Gu, Q, Zobel, J, Vázquez-Abad, FJ, Lin, P, Gu, Q, Zobel, J, Vázquez-Abad, FJ, and Lin, P
- Abstract
Machine learning is widely used for personalisation, that is, to tune systems with the aim of adapting their behaviour to the responses of humans. This tuning relies on quantified features that capture the human actions, and also on objective functions-that is, proxies - that are intended to represent desirable outcomes. However, a learning system's representation of the world can be incomplete or insufficiently rich, for example if users' decisions are based on properties of which the system is unaware. Moreover, the incompleteness of proxies can be argued to be an intrinsic property of computational systems, as they are based on literal representations of human actions rather than on the human actions themselves; this problem is distinct from the usual aspects of bias that are examined in machine learning literature. We use mathematical analysis and simulations of a reinforcement-learning case study to demonstrate that incompleteness of representation can, first, lead to learning that is no better than random; and second, means that the learning system can be inherently unaware that it is failing. This result has implications for the limits and applications of machine learning systems in human domains.
- Published
- 2022
3. 4CPS-111 Clinical and economic impact of infliximab biosimilar Inflectra in rheumatoid arthritis, psoriatic arthropathy and ankylosing spondylitis naïve and switched patients: 5 years of follow-up
- Author
-
Borras, J, primary, Gracia-Pérez, A, additional, Valcuende Rosique, A, additional, Castera, MDe, additional, Rosique, JD, additional, and Abad, FJ, additional
- Published
- 2020
- Full Text
- View/download PDF
4. Measurement invariance of the WHOQOL-AGE questionnaire across three European countries
- Author
-
Santos D, Abad FJ, Maria Miret, Chatterji S, Olaya B, Zawisza K, Koskinen S, Leonardi M, Haro JM, Ayuso-Mateos JL, and Caballero FF
- Subjects
Measurement invariance ,Multi-group confirmatory factor analysis ,WHOQOL-AGE ,Bifactor model ,Quality of life - Abstract
PURPOSE: Developing valid and reliable instruments that can be used across countries is necessary. The present study aimed to test the comparability of quality of life scores across three European countries (Finland, Poland, and Spain). METHOD: Data from 9987 participants interviewed between 2011 and 2012 were employed, using nationally representative samples from the Collaborative Research on Ageing in Europe project. The WHOQOL-AGE questionnaire is a 13-item test and was employed to assess the quality of life in the three considered countries. First of all, two models (a bifactor model and a two-correlated factor model) were proposed and tested in each country by means of confirmatory factor models. Second, measurement invariance across the three countries was tested using multi-group confirmatory factor analysis for that model which showed the best fit. Finally, differences in latent mean scores across countries were analyzed. RESULTS: The results indicated that the bifactor model showed more satisfactory goodness-of-fit indices than the two-correlated factor model and that the WHOQOL-AGE questionnaire is a partially scalar invariant instrument (only two items do not meet scalar invariance). Quality of life scores were higher in Finland (considered as the reference category: mean = 0, SD = 1) than in Spain (mean = - 0.547, SD = 1.22) and Poland (mean = - 0.927, SD = 1.26). CONCLUSIONS: Respondents from Finland, Poland, and Spain attribute the same meaning to the latent construct studied, and differences across countries can be due to actual differences in quality of life. According to the results, the comparability across the different considered samples is supported and the WHOQOL-AGE showed an adequate validity in terms of cross-country validation. Caution should be exercised with the two items which did not meet scalar invariance, as potential indicator of differential item functioning.
- Published
- 2018
5. Effectiveness of infliximab, adalimumab and golimumab for non-infectious refractory uveitis in adults
- Author
-
Borrás-Blasco J, Casterá DE, Cortes X, Abad FJ, Rosique-Robles JD, and Mallench LG
- Subjects
adalimumab ,TNF ,uveitis ,golimumab ,infliximab - Abstract
Aim: To discuss the available data regarding the off-label uses of anti-TNF agents in non-infectious uveitis. Data source: A literature search was performed in Medline through PubMed from January 2001 to January 2014. Study selection and data extraction: English-language articles about uveitis treatment with anti-TNF drugs in adult patients were reviewed. Data synthesis: The use of anti-TNF-alpha drugs for treatment of several refractory manifestations of refractory uveitis in adult patients is increasing. However, due to the lack of evidence from randomized controlled trials, the use of anti-TNF in uveitis remains "off-label" in most countries. There is no trial-based evidence to support it except for the experience provided by cases and case series. This experience, which is continuously increasing, has yielded encouraging results. Anti-TNF-alpha drugs, such as infliximab, adalimumab, and golimumab, are reasonably effective for controlling ocular inflammation and sparing patients corticosteroid treatment in non-infectious refractory uveitis. Approximately 80% of patients on infliximab, adalimumab, or golimumab were able to achieve sustained control of inflammation by 6 months. Conclusion: Anti-TNF-alpha therapy is effective in inducing clinical remission for refractory uveitis, with a relatively low rate of treatment-ending adverse events. However, randomized and controlled trials are required to adequately assess the maintained clinical efficacy and safety profile in the long term of anti-TNF agents for non-infectious refractory uveitis.
- Published
- 2015
6. GM-004 Economic impact associated with a Biological Treatment Prioritisation protocol in rheumatoid arthritis patients in Sagunto Hospital: Abstract GM-004 Table 1
- Author
-
Borras, J, primary, Casterá, DE, additional, Abad, FJ, additional, and Rosique-Robles, JD, additional
- Published
- 2014
- Full Text
- View/download PDF
7. Setwise and Filtered Gibbs Samplers for Teletraffic Analysis
- Author
-
Andrew, LLH, Qian, G, Vazquez-Abad, FJ, Andrew, LLH, Qian, G, and Vazquez-Abad, FJ
- Abstract
A setwise Gibbs sampler (SGS) method is developed to simulate stationary distributions and performance measures of network occupancy of Baskett-Chandy-Muntz-Palacios (BCMP) telecommunication models. It overcomes the simulation difficulty encountered in applying the standard Gibbs sampler to closed BCMP networks with constant occupancy constraints. We show Markov chains induced by SGS converge to the target stationary distributions. This article also investigates the filtered Gibbs sampler (FGS) as an efficient method for estimating various network performance measures. It shows that FGS's efficiency is considerable, but may be improperly overestimated. A more conservative performance estimator is then presented.
- Published
- 2010
8. HIV patients' gastrointestinal tolerability and treatment satisfaction after switching from lopinavir/ritonavir (LPV/r) SGC to co-formulated LPV/r tablets
- Author
-
Borras-Blasco, J, primary, Rosique-Robles, JD, additional, Belda, A, additional, Abad, FJ, additional, and Castera, MDE, additional
- Published
- 2008
- Full Text
- View/download PDF
9. Measurement of epistemological beliefs: psychometric properties of the EQEBI test scores.
- Author
-
Ordonez XG, Ponsoda V, Abad FJ, and Romero SJ
- Abstract
This article proposes a new test (called the EQEBI) for the measurement of epistemological beliefs, integrating and extending the Epistemological Questionnaire (EQ) and the Epistemic Beliefs Inventory (EBI). In Study 1, the two tests were translated and applied to a Spanish-speaking sample. A detailed dimensionality exploration, by means of the monotone homogeneity model and confirmatory factor analysis, provides unexpected dimensionality results. Therefore, a new test is proposed, and the dimensionality and psychometric properties of the test's scores are determined. The new test retains four of the five original dimensions. The EQEBI has 27 items, and its score reliability is higher than that of the original EQ and EBI. In Study 2, the EQEBI was applied to another sample to verify the psychometric properties of the obtained scores. The expected four unidimensional scales are confirmed. The scales are calibrated with the graded response model. Recommendations are offered for using the scales for epistemological beliefs assessment. [ABSTRACT FROM AUTHOR]
- Published
- 2009
- Full Text
- View/download PDF
10. Parapharyngeal abscess in a patient receiving etanercept.
- Author
-
Borrás-Blasco J, Nuñez-Cornejo C, Gracia-Perez A, Rosique-Robles JD, Casterá MD, Viosca E, and Abad FJ
- Published
- 2007
- Full Text
- View/download PDF
11. Generational changes on the Draw-a-Man test: a comparison of Brazilian urban and rural children tested in 1930, 2002 and 2004.
- Author
-
Colom R, Flores-Mendoza CE, and Abad FJ
- Abstract
Although gains in generational intelligence test scores have been widely demonstrated around the world, researchers still do not know what has caused them. The cognitive stimulation and nutritional hypotheses summarize the several diverse potential causes that have been considered. This article analyses data for a sample of 499 children tested in 1930 and one equivalent sample of 710 children tested 72 years later, the largest gap ever considered. Both samples comprised children aged between 7 and 11 who were assessed by the Draw-a-Man test in the city of Belo Horizonte, Brazil. Further, one additional sample of 132 children was assessed in 2004 in a rural area very similar in several diverse factors to the 1930 urban sample. The results are consistent with both the cognitive stimulation and the nutritional hypotheses. [ABSTRACT FROM AUTHOR]
- Published
- 2007
- Full Text
- View/download PDF
12. Complex span tasks, simple span tasks, and cognitive abilities: a reanalysis of key studies.
- Author
-
Colom R, Rebollo I, Abad FJ, and Shih PC
- Abstract
There is great interest in the relationships between memory span tasks and cognitive abilities. However, the causes underlying their correlation remain unknown. In the present article, five key data sets were reanalyzed according to two criteria: They must consider complex span tasks (so-called working memory [WM] tasks) and simple span tasks (so-called short-term memory [STM] tasks), and they must comprise cognitive ability measures. The obtained results offer several points of interest. First, memory span tasks should be conceived from a hierarchical perspective: They comprise both general and specific components. Second, the general component explains about four times the variance explained by the specific components. Third, STM and WM measures are closely related. Fourth, STM and WM measures share the same common variance with cognitive abilities. Finally, the strong relationship usually found between memory span tasks and cognitive abilities could be tentatively interpreted by the component shared by STM and WM--namely, the capacity for temporarily preserving a reliable memory representation of any given information. [ABSTRACT FROM AUTHOR]
- Published
- 2006
- Full Text
- View/download PDF
13. Abilities that explain the intelligence decline: Evidence from the WAIS-III
- Author
-
Escorial, S., Rebollo, I., Garcia, Lf, Roberto Colom, Abad, Fj, and Juan-Espinosa, M.
14. Dimensionality assessment in the presence of wording effects: A network psychometric and factorial approach.
- Author
-
Garcia-Pardina A, Abad FJ, Christensen AP, Golino H, and Garrido LE
- Subjects
- Humans, Factor Analysis, Statistical, Self Report, Models, Statistical, Computer Simulation, Data Interpretation, Statistical, Psychometrics methods
- Abstract
This study proposes a procedure for substantive dimensionality estimation in the presence of wording effects, the inconsistent response to regular and reversed self-report items. The procedure developed consists of subtracting an approximate estimate of the wording effects variance from the sample correlation matrix and then estimating the substantive dimensionality on the residual correlation matrix. This is achieved by estimating a random intercept factor with unit loadings for all the regular and unrecoded reversed items. The accuracy of the procedure was evaluated through an extensive simulation study that manipulated nine relevant variables and employed the exploratory graph analysis (EGA) and parallel analysis (PA) retention methods. The results indicated that combining the proposed procedure with EGA or PA achieved high accuracy in estimating the substantive latent dimensionality, but that EGA was superior. Additionally, the present findings shed light on the complex ways that wording effects impact the dimensionality estimates when the response bias in the data is ignored. A tutorial on substantive dimensionality estimation with the R package EGAnet is offered, as well as practical guidelines for applied researchers., (© 2024. The Psychonomic Society, Inc.)
- Published
- 2024
- Full Text
- View/download PDF
15. Exploratory Bi-factor Analysis with Multiple General Factors.
- Author
-
Jiménez M, Abad FJ, Garcia-Garzon E, and Garrido LE
- Subjects
- Computer Simulation, Factor Analysis, Statistical, Monte Carlo Method, Psychometrics, Algorithms, Calcium Channels
- Abstract
Exploratory bi-factor analysis (EBFA) is a very popular approach to estimate models where specific factors are concomitant to a single, general dimension. However, the models typically encountered in fields like personality, intelligence, and psychopathology involve more than one general factor. To address this circumstance, we developed an algorithm (GSLiD) based on partially specified targets to perform exploratory bi-factor analysis with multiple general factors (EBFA-MGF). In EBFA-MGF, researchers do not need to conduct independent bi-factor analyses anymore because several bi-factor models are estimated simultaneously in an exploratory manner, guarding against biased estimates and model misspecification errors due to unexpected cross-loadings and factor correlations. The results from an exhaustive Monte Carlo simulation manipulating nine variables of interest suggested that GSLiD outperforms the Schmid-Leiman approximation and is robust to challenging conditions involving cross-loadings and pure items of the general factors. Thereby, we supply an R package (bifactor) to make EBFA-MGF readily available for substantive research. Finally, we use GSLiD to assess the hierarchical structure of a reduced version of the Personality Inventory for DSM-5 Short Form (PID-5-SF).
- Published
- 2023
- Full Text
- View/download PDF
16. Improving reliability estimation in cognitive diagnosis modeling.
- Author
-
Kreitchmann RS, de la Torre J, Sorrel MA, Nájera P, and Abad FJ
- Subjects
- Humans, Reproducibility of Results, Computer Simulation, Awareness
- Abstract
Cognitive diagnosis models (CDMs) are used in educational, clinical, or personnel selection settings to classify respondents with respect to discrete attributes, identifying strengths and needs, and thus allowing to provide tailored training/treatment. As in any assessment, an accurate reliability estimation is crucial for valid score interpretations. In this sense, most CDM reliability indices are based on the posterior probabilities of the estimated attribute profiles. These posteriors are traditionally computed using point estimates for the model parameters as approximations to their populational values. If the uncertainty around these parameters is unaccounted for, the posteriors may be overly peaked, deriving into overestimated reliabilities. This article presents a multiple imputation (MI) procedure to integrate out the model parameters in the estimation of the posterior distributions, thus correcting the reliability estimation. A simulation study was conducted to compare the MI procedure with the traditional reliability estimation. Five factors were manipulated: the attribute structure, the CDM model (DINA and G-DINA), test length, sample size, and item quality. Additionally, an illustration using the Examination for the Certificate of Proficiency in English data was analyzed. The effect of sample size was studied by sampling subsets of subjects from the complete data. In both studies, the traditional reliability estimation systematically provided overestimated reliabilities, whereas the MI procedure offered more accurate results. Accordingly, practitioners in small educational or clinical settings should be aware that the reliability estimation using model parameter point estimates may be positively biased. R codes for the MI procedure are made available., (© 2022. The Author(s).)
- Published
- 2023
- Full Text
- View/download PDF
17. Dimensionality assessment in bifactor structures with multiple general factors: A network psychometrics approach.
- Author
-
Jiménez M, Abad FJ, Garcia-Garzon E, Golino H, Christensen AP, and Garrido LE
- Abstract
The accuracy of factor retention methods for structures with one or more general factors, like the ones typically encountered in fields like intelligence, personality, and psychopathology, has often been overlooked in dimensionality research. To address this issue, we compared the performance of several factor retention methods in this context, including a network psychometrics approach developed in this study. For estimating the number of group factors, these methods were the Kaiser criterion, empirical Kaiser criterion, parallel analysis with principal components (PA
PCA ) or principal axis, and exploratory graph analysis with Louvain clustering (EGALV ). We then estimated the number of general factors using the factor scores of the first-order solution suggested by the best two methods, yielding a "second-order" version of PAPCA (PAPCA-FS ) and EGALV (EGALV-FS ). Additionally, we examined the direct multilevel solution provided by EGALV . All the methods were evaluated in an extensive simulation manipulating nine variables of interest, including population error. The results indicated that EGALV and PAPCA displayed the best overall performance in retrieving the true number of group factors, the former being more sensitive to high cross-loadings, and the latter to weak group factors and small samples. Regarding the estimation of the number of general factors, both PAPCA-FS and EGALV-FS showed a close to perfect accuracy across all the conditions, while EGALV was inaccurate. The methods based on EGA were robust to the conditions most likely to be encountered in practice. Therefore, we highlight the particular usefulness of EGALV (group factors) and EGALV-FS (general factors) for assessing bifactor structures with multiple general factors. (PsycInfo Database Record (c) 2023 APA, all rights reserved).- Published
- 2023
- Full Text
- View/download PDF
18. Is exploratory factor analysis always to be preferred? A systematic comparison of factor analytic techniques throughout the confirmatory-exploratory continuum.
- Author
-
Nájera P, Abad FJ, and Sorrel MA
- Abstract
The number of available factor analytic techniques has been increasing in the last decades. However, the lack of clear guidelines and exhaustive comparison studies between the techniques might hinder that these valuable methodological advances make their way to applied research. The present paper evaluates the performance of confirmatory factor analysis (CFA), CFA with sequential model modification using modification indices and the Saris procedure, exploratory factor analysis (EFA) with different rotation procedures (Geomin, target, and objectively refined target matrix), Bayesian structural equation modeling (BSEM), and a new set of procedures that, after fitting an unrestrictive model (i.e., EFA, BSEM), identify and retain only the relevant loadings to provide a parsimonious CFA solution (ECFA, BCFA). By means of an exhaustive Monte Carlo simulation study and a real data illustration, it is shown that CFA and BSEM are overly stiff and, consequently, do not appropriately recover the structure of slightly misspecified models. EFA usually provides the most accurate parameter estimates, although the rotation procedure choice is of major importance, especially depending on whether the latent factors are correlated or not. Finally, ECFA might be a sound option whenever an a priori structure cannot be hypothesized and the latent factors are correlated. Moreover, it is shown that the pattern of the results of a factor analytic technique can be somehow predicted based on its positioning in the confirmatory-exploratory continuum. Applied recommendations are given for the selection of the most appropriate technique under different representative scenarios by means of a detailed flowchart. (PsycInfo Database Record (c) 2023 APA, all rights reserved).
- Published
- 2023
- Full Text
- View/download PDF
19. On Bank Assembly and Block Selection in Multidimensional Forced-Choice Adaptive Assessments.
- Author
-
Kreitchmann RS, Sorrel MA, and Abad FJ
- Abstract
Multidimensional forced-choice (FC) questionnaires have been consistently found to reduce the effects of socially desirable responding and faking in noncognitive assessments. Although FC has been considered problematic for providing ipsative scores under the classical test theory, item response theory (IRT) models enable the estimation of nonipsative scores from FC responses. However, while some authors indicate that blocks composed of opposite-keyed items are necessary to retrieve normative scores, others suggest that these blocks may be less robust to faking, thus impairing the assessment validity. Accordingly, this article presents a simulation study to investigate whether it is possible to retrieve normative scores using only positively keyed items in pairwise FC computerized adaptive testing (CAT). Specifically, a simulation study addressed the effect of (a) different bank assembly (with a randomly assembled bank, an optimally assembled bank, and blocks assembled on-the-fly considering every possible pair of items), and (b) block selection rules (i.e., T , and Bayesian D and A -rules) over the estimate accuracy and ipsativity and overlap rates. Moreover, different questionnaire lengths (30 and 60) and trait structures (independent or positively correlated) were studied, and a nonadaptive questionnaire was included as baseline in each condition. In general, very good trait estimates were retrieved, despite using only positively keyed items. Although the best trait accuracy and lowest ipsativity were found using the Bayesian A -rule with questionnaires assembled on-the-fly , the T -rule under this method led to the worst results. This points out to the importance of considering both aspects when designing FC CAT., Competing Interests: The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article., (© The Author(s) 2022.)
- Published
- 2023
- Full Text
- View/download PDF
20. New Trends in Digital Technology-Based Psychological and Educational Assessment.
- Author
-
Elosua P, Aguado D, Fonseca-Pedrero E, Abad FJ, and Santamaría P
- Subjects
- Humans, Retrospective Studies, Psychometrics, Educational Measurement, Digital Technology, Software
- Abstract
Background: The emergence of digital technology in the field of psychological and educational measurement and assessment broadens the traditional concept of pencil and paper tests. New assessment models built on the proliferation of smartphones, social networks and software developments are opening up new horizons in the field., Method: This study is divided into four sections, each discussing the benefits and limitations of a specific type of technology-based assessment: ambulatory assessment, social networks, gamification and forced-choice testing., Results: The latest developments are clearly relevant in the field of psychological and educational measurement and assessment. Among other benefits, they bring greater ecological validity to the assessment process and eliminate the bias associated with retrospective assessment., Conclusions: Some of these new approaches point to a multidisciplinary scenario with a tradition which has yet to be created. Psychometrics must secure a place in this new world by contributing sound expertise in the measurement of psychological variables. The challenges and debates facing the field of psychology as it incorporates these new approaches are also discussed.
- Published
- 2023
- Full Text
- View/download PDF
21. When proxy-driven learning is no better than random: The consequences of representational incompleteness.
- Author
-
Zobel J, Vázquez-Abad FJ, and Lin P
- Subjects
- Advance Directives, Humans, Reinforcement, Psychology, Machine Learning, Proxy
- Abstract
Machine learning is widely used for personalisation, that is, to tune systems with the aim of adapting their behaviour to the responses of humans. This tuning relies on quantified features that capture the human actions, and also on objective functions-that is, proxies - that are intended to represent desirable outcomes. However, a learning system's representation of the world can be incomplete or insufficiently rich, for example if users' decisions are based on properties of which the system is unaware. Moreover, the incompleteness of proxies can be argued to be an intrinsic property of computational systems, as they are based on literal representations of human actions rather than on the human actions themselves; this problem is distinct from the usual aspects of bias that are examined in machine learning literature. We use mathematical analysis and simulations of a reinforcement-learning case study to demonstrate that incompleteness of representation can, first, lead to learning that is no better than random; and second, means that the learning system can be inherently unaware that it is failing. This result has implications for the limits and applications of machine learning systems in human domains., Competing Interests: The authors have declared that no competing interests exist.
- Published
- 2022
- Full Text
- View/download PDF
22. A genetic algorithm for optimal assembly of pairwise forced-choice questionnaires.
- Author
-
Kreitchmann RS, Abad FJ, and Sorrel MA
- Subjects
- Computer Simulation, Humans, Surveys and Questionnaires, Algorithms, Personality
- Abstract
The use of multidimensional forced-choice questionnaires has been proposed as a means of improving validity in the assessment of non-cognitive attributes in high-stakes scenarios. However, the reduced precision of trait estimates in this questionnaire format is an important drawback. Accordingly, this article presents an optimization procedure for assembling pairwise forced-choice questionnaires while maximizing posterior marginal reliabilities. This procedure is performed through the adaptation of a known genetic algorithm (GA) for combinatorial problems. In a simulation study, the efficiency of the proposed procedure was compared with a quasi-brute-force (BF) search. For this purpose, five-dimensional item pools were simulated to emulate the real problem of generating a forced-choice personality questionnaire under the five-factor model. Three factors were manipulated: (1) the length of the questionnaire, (2) the relative item pool size with respect to the questionnaire's length, and (3) the true correlations between traits. The recovery of the person parameters for each assembled questionnaire was evaluated through the squared correlation between estimated and true parameters, the root mean square error between the estimated and true parameters, the average difference between the estimated and true inter-trait correlations, and the average standard error for each trait level. The proposed GA offered more accurate trait estimates than the BF search within a reasonable computation time in every simulation condition. Such improvements were especially important when measuring correlated traits and when the relative item pool sizes were higher. A user-friendly online implementation of the algorithm was made available to the users., (© 2021. The Author(s).)
- Published
- 2022
- Full Text
- View/download PDF
23. Cross-Cultural Measurement Invariance in the Personality Inventory for DSM-5 ✰ .
- Author
-
Sorrel MA, García LF, Aluja A, Rolland JP, Rossier J, Roskam I, and Abad FJ
- Subjects
- Diagnostic and Statistical Manual of Mental Disorders, Humans, Personality Inventory, Psychometrics, Reproducibility of Results, Cross-Cultural Comparison, Personality Disorders diagnosis
- Abstract
The validity of cross-cultural comparisons of test scores requires that scores have the same meaning across cultures, which is usually tested by checking the invariance of the measurement model across groups. In the last decade, a large number of studies were conducted to verify the equivalence across cultures of the dimensional Alternative Model of Personality Disorders (DSM-5 Section III). These studies have provided information on configural invariance (i.e., the facets that compose the domains are the same) and metric invariance (i.e., facet-domain relationships are equal across groups), but not on the stricter scalar invariance (i.e., the baseline levels of the facets are the same), which is a prerequisite for meaningfully comparing group means. The present study aims to address this gap. The Personality Inventory for DSM-5 (PID-5) was administered to five samples differing on country and language (Belgium, Catalonia, France, Spain, and Switzerland), with a total of 4,380 participants. Configural and metric invariance were supported, denoting that the model structure was stable across samples. Partial scalar invariance was supported, being minimal the influence of non-invariant facets. This allowed cross-cultural mean comparisons. Results are discussed in light of the sample composition and a possible impact of culture on development of psychopathology., (Copyright © 2021. Published by Elsevier B.V.)
- Published
- 2021
- Full Text
- View/download PDF
24. Balancing fit and parsimony to improve Q-matrix validation.
- Author
-
Nájera P, Sorrel MA, de la Torre J, and Abad FJ
- Subjects
- Computer Simulation, Psychometrics, Models, Statistical, Research Design
- Abstract
The Q-matrix identifies the subset of attributes measured by each item in the cognitive diagnosis modelling framework. Usually constructed by domain experts, the Q-matrix might contain some misspecifications, disrupting classification accuracy. Empirical Q-matrix validation methods such as the general discrimination index (GDI) and Wald have shown promising results in addressing this problem. However, a cut-off point is used in both methods, which might be suboptimal. To address this limitation, the Hull method is proposed and evaluated in the present study. This method aims to find the optimal balance between fit and parsimony, and it is flexible enough to be used either with a measure of item discrimination (the proportion of variance accounted for, PVAF) or a coefficient of determination (pseudo-R
2 ). Results from a simulation study showed that the Hull method consistently showed the best performance and shortest computation time, especially when used with the PVAF. The Wald method also performed very well overall, while the GDI method obtained poor results when the number of attributes was high. The absence of a cut-off point provides greater flexibility to the Hull method, and it places it as a comprehensive solution to the Q-matrix specification problem in applied settings. This proposal is illustrated using real data., (© 2020 The British Psychological Society.)- Published
- 2021
- Full Text
- View/download PDF
25. Modeling Wording Effects Does Not Help in Recovering Uncontaminated Person Scores: A Systematic Evaluation With Random Intercept Item Factor Analysis.
- Author
-
Nieto MD, Garrido LE, Martínez-Molina A, and Abad FJ
- Abstract
The item wording (or keying) effect consists of logically inconsistent answers to positively and negatively worded items that tap into similar (but polarly opposite) content. Previous research has shown that this effect can be successfully modeled through the random intercept item factor analysis (RIIFA) model, as evidenced by the improvements in the model fit in comparison to models that only contain substantive factors. However, little is known regarding the capability of this model in recovering the uncontaminated person scores. To address this issue, the study analyzes the performance of the RIIFA approach across three types of wording effects proposed in the literature: carelessness, item verification difficulty, and acquiescence. In the context of unidimensional substantive models, four independent variables were manipulated, using Monte Carlo methods: type of wording effect, amount of wording effect, sample size, and test length. The results corroborated previous findings by showing that the RIIFA models were consistently able to account for the variance in the data, attaining an excellent fit regardless of the amount of bias. Conversely, the models without the RIIFA factor produced increasingly a poorer fit with greater amounts of wording effects. Surprisingly, however, the RIIFA models were not able to better estimate the uncontaminated person scores for any type of wording effect in comparison to the substantive unidimensional models. The simulation results were then corroborated with an empirical dataset, examining the relationship between learning strategies and personality with grade point average in undergraduate studies. The apparently paradoxical findings regarding the model fit and the recovery of the person scores are explained, considering the properties of the factor models examined., Competing Interests: The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest., (Copyright © 2021 Nieto, Garrido, Martínez-Molina and Abad.)
- Published
- 2021
- Full Text
- View/download PDF
26. Improving Accuracy and Usage by Correctly Selecting: The Effects of Model Selection in Cognitive Diagnosis Computerized Adaptive Testing.
- Author
-
Sorrel MA, Abad FJ, and Nájera P
- Abstract
Decisions on how to calibrate an item bank might have major implications in the subsequent performance of the adaptive algorithms. One of these decisions is model selection, which can become problematic in the context of cognitive diagnosis computerized adaptive testing, given the wide range of models available. This article aims to determine whether model selection indices can be used to improve the performance of adaptive tests. Three factors were considered in a simulation study, that is, calibration sample size, Q-matrix complexity, and item bank length. Results based on the true item parameters, and general and single reduced model estimates were compared to those of the combination of appropriate models. The results indicate that fitting a single reduced model or a general model will not generally provide optimal results. Results based on the combination of models selected by the fit index were always closer to those obtained with the true item parameters. The implications for practical settings include an improvement in terms of classification accuracy and, consequently, testing time, and a more balanced use of the item bank. An R package was developed, named cdcatR, to facilitate adaptive applications in this context., Competing Interests: Declaration of Conflicting Interests: The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article., (© The Author(s) 2020.)
- Published
- 2021
- Full Text
- View/download PDF
27. Determining the Number of Attributes in Cognitive Diagnosis Modeling.
- Author
-
Nájera P, Abad FJ, and Sorrel MA
- Abstract
Cognitive diagnosis models (CDMs) allow classifying respondents into a set of discrete attribute profiles. The internal structure of the test is determined in a Q-matrix, whose correct specification is necessary to achieve an accurate attribute profile classification. Several empirical Q-matrix estimation and validation methods have been proposed with the aim of providing well-specified Q-matrices. However, these methods require the number of attributes to be set in advance. No systematic studies about CDMs dimensionality assessment have been conducted, which contrasts with the vast existing literature for the factor analysis framework. To address this gap, the present study evaluates the performance of several dimensionality assessment methods from the factor analysis literature in determining the number of attributes in the context of CDMs. The explored methods were parallel analysis, minimum average partial, very simple structure, DETECT, empirical Kaiser criterion, exploratory graph analysis, and a machine learning factor forest model. Additionally, a model comparison approach was considered, which consists in comparing the model-fit of empirically estimated Q-matrices. The performance of these methods was assessed by means of a comprehensive simulation study that included different generating number of attributes, item qualities, sample sizes, ratios of the number of items to attribute, correlations among the attributes, attributes thresholds, and generating CDM. Results showed that parallel analysis (with Pearson correlations and mean eigenvalue criterion), factor forest model, and model comparison (with AIC) are suitable alternatives to determine the number of attributes in CDM applications, with an overall percentage of correct estimates above 76% of the conditions. The accuracy increased to 97% when these three methods agreed on the number of attributes. In short, the present study supports the use of three methods in assessing the dimensionality of CDMs. This will allow to test the assumption of correct dimensionality present in the Q-matrix estimation and validation methods, as well as to gather evidence of validity to support the use of the scores obtained with these models. The findings of this study are illustrated using real data from an intelligence test to provide guidelines for assessing the dimensionality of CDM data in applied settings., Competing Interests: The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest., (Copyright © 2021 Nájera, Abad and Sorrel.)
- Published
- 2021
- Full Text
- View/download PDF
28. On Omega Hierarchical Estimation: A Comparison of Exploratory Bi-Factor Analysis Algorithms.
- Author
-
Garcia-Garzon E, Abad FJ, and Garrido LE
- Subjects
- Factor Analysis, Statistical, Monte Carlo Method, Rotation, Algorithms, Judgment
- Abstract
As general factor modeling continues to grow in popularity, researchers have become interested in assessing how reliable general factor scores are. Even though omega hierarchical estimation has been suggested as a useful tool in this context, little is known about how to approximate it using modern bi-factor exploratory factor analysis methods. This study is the first to compare how omega hierarchical estimates were recovered by six alternative algorithms: Bi-quartimin, bi-geomin, Schmid-Leiman (SL), empirical iterative empirical target rotation based on an initial SL solution (SLiD), direct SL (DSL), and direct bi-factor (DBF). The algorithms were tested in three Monte-Carlo simulations including bi-factor and second-order structures and presenting complexities such as cross-loadings or pure indicators of the general factor and structures without a general factor. Results showed that SLiD provided the best approximation to omega hierarchical under most conditions. Overall, neither SL, bi-quartimin, nor bi-geomin produced an overall satisfactory recovery of omega hierarchical. Lastly, the performance of DSL and DBF depended upon the average discrepancy between the loadings of the general and the group factors. The re-analysis of eight classical datasets further illustrated how algorithm selection could influence judgments regarding omega hierarchical.
- Published
- 2021
- Full Text
- View/download PDF
29. Comparing Methods for Modeling Acquiescence in Multidimensional Partially Balanced Scales.
- Author
-
de la Fuente J and Abad FJ
- Subjects
- Bias, Computer Simulation, Factor Analysis, Statistical, Humans, Research Design
- Abstract
Background: The inclusion of direct and reversed items in scales is a commonly-used strategy to control acquiescence bias. However, this is not enough to avoid the distortions produced by this response style in the structure of covariances and means of the scale in question. This simulation study provides evidence on the performance of two different procedures for modelling the influence of acquiescence bias on partially balanced multidimensional scales: a method based on exploratory factor analysis (EFA) with target rotation, and a method based on random intercept factor analysis (RIFA)., Method: The independent variables analyzed in a simulation study were sample size, number of items per factor, balance of substantive loadings of direct and reversed items, size and heterogeneity of acquiescence loadings, and inter-factor correlation., Results: The RIFA method had better performance over most of the conditions, especially for the balanced conditions, although the variance of acquiescence factor loadings had a certain impact. In relation to the EFA method, it was severely affected by a low degree of balance., Conclusions: RIFA seems the most robust approach, but EFA also remains a good alternative for medium and fully balanced scales.
- Published
- 2020
- Full Text
- View/download PDF
30. Detecting Cheating Methods on Unproctored Internet Tests.
- Author
-
Sanz S, Luzardo M, García C, and Abad FJ
- Subjects
- Likelihood Functions, Deception, Internet
- Abstract
Background: Unproctored Internet Tests (UIT) are vulnerable to cheating attempts by candidates to obtain higher scores. To prevent this, subsequent procedures such as a verification test (VT) is carried out. This study compares five statistics used to detect cheating in Computerized Adaptive Tests (CATs): Guo and Drasgow's Z-test, the Adaptive Measure of Change (AMC), Likelihood Ratio Test (LRT), Score Test, and Modified Signed Likelihood Ratio Test (MSLRT)., Method: We simulated data from honest and cheating candidates to the UIT and the VT. Honest candidates responded to the UIT and the VT with their real ability level, while cheating candidates responded only to the VT, and different levels of cheating were simulated. We applied hypothesis tests, and obtained type I error and power rates., Results: Although we found differences in type I error rates between some of the procedures, all procedures reported quite accurate results with the exception of the Score Test. The power rates obtained point to MSLRT's superiority in detecting cheating., Conclusions: We consider the MSLRT to be the best test, as it has the highest power rate and a suitable type I error rate.
- Published
- 2020
- Full Text
- View/download PDF
31. Bi-factor Exploratory Structural Equation Modeling Done Right: Using the SLiDapp Application.
- Author
-
García-Garzón E, Nieto MD, Garrido LE, and Abad FJ
- Subjects
- Humans, Psychometrics, Factor Analysis, Statistical, Latent Class Analysis
- Abstract
Background: Due to its flexibility and statistical properties, bi-factor Exploratory Structural Equation Modeling (bi-factor ESEM) has become an often-recommended tool in psychometrics. Unfortunately, most recent methods for approximating these structures, such as the SLiD algorithm, are not available in the leading software for performing ESEM (i.e., Mplus). To resolve this issue, we present a novel, user-friendly Shiny application for integrating the SLiD algorithm in bi-factor ESEM estimation in Mplus. Thus, a two-stage framework for conducting SLiD-based bi-factor ESEM in Mplus was developed., Method: This approach was presented in a step-by-step guide for applied researchers, showing the utility of the developed SLiDApp application. Using data from the Open-Source Psychometrics Project (N = 2495), we conducted a bi-factor ESEM exploration of the Generic Conspiracist Beliefs Scale. We studied whether bi-factor modelling was appropriate and if both general and group factors were related to each personality trait., Results: The application of the SLiD algorithm provided unique information regarding this factor structure and its ESEM structural parameters., Conclusions: The results illustrated the usefulness and validity of SLiD-based bi-factor ESEM, and how the proposed Shiny app could make it eaiser for applied researchers to use these methods.
- Published
- 2020
- Full Text
- View/download PDF
32. Improving Robustness in Q-Matrix Validation Using an Iterative and Dynamic Procedure.
- Author
-
Nájera P, Sorrel MA, de la Torre J, and Abad FJ
- Abstract
In the context of cognitive diagnosis models (CDMs), a Q-matrix reflects the correspondence between attributes and items. The Q-matrix construction process is typically subjective in nature, which may lead to misspecifications. All this can negatively affect the attribute classification accuracy. In response, several methods of empirical Q-matrix validation have been developed. The general discrimination index (GDI) method has some relevant advantages such as the possibility of being applied to several CDMs. However, the estimation of the GDI relies on the estimation of the latent group sizes and success probabilities, which is made with the original (possibly misspecified) Q-matrix. This can be a problem, especially in those situations in which there is a great uncertainty about the Q-matrix specification. To address this, the present study investigates the iterative application of the GDI method, where only one item is modified at each step of the iterative procedure, and the required cutoff is updated considering the new parameter estimates. A simulation study was conducted to test the performance of the new procedure. Results showed that the performance of the GDI method improved when the application was iterative at the item level and an appropriate cutoff point was used. This was most notable when the original Q-matrix misspecification rate was high, where the proposed procedure performed better 96.5% of the times. The results are illustrated using Tatsuoka's fraction-subtraction data set., Competing Interests: Declaration of Conflicting Interests: The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article., (© The Author(s) 2020.)
- Published
- 2020
- Full Text
- View/download PDF
33. Adapting cognitive diagnosis computerized adaptive testing item selection rules to traditional item response theory.
- Author
-
Sorrel MA, Barrada JR, de la Torre J, and Abad FJ
- Subjects
- Algorithms, Bayes Theorem, Bias, Computer Simulation, Computers, Data Accuracy, Humans, Cognition, Educational Measurement methods, Psychometrics methods
- Abstract
Currently, there are two predominant approaches in adaptive testing. One, referred to as cognitive diagnosis computerized adaptive testing (CD-CAT), is based on cognitive diagnosis models, and the other, the traditional CAT, is based on item response theory. The present study evaluates the performance of two item selection rules (ISRs) originally developed in the CD-CAT framework, the double Kullback-Leibler information (DKL) and the generalized deterministic inputs, noisy "and" gate model discrimination index (GDI), in the context of traditional CAT. The accuracy and test security associated with these two ISRs are compared to those of the point Fisher information and weighted KL using a simulation study. The impact of the trait level estimation method is also investigated. The results show that the new ISRs, particularly DKL, could be used to improve the accuracy of CAT. Better accuracy for DKL is achieved at the expense of higher item overlap rate. Differences among the item selection rules become smaller as the test gets longer. The two CD-CAT ISRs select different types of items: items with the highest possible a parameter with DKL, and items with the lowest possible c parameter with GDI. Regarding the trait level estimator, expected a posteriori method is generally better in the first stages of the CAT, and converges with the maximum likelihood method when a medium to large number of items are involved. The use of DKL can be recommended in low-stakes settings where test security is less of a concern., Competing Interests: The authors have declared that no competing interests exist.
- Published
- 2020
- Full Text
- View/download PDF
34. Controlling for Response Biases in Self-Report Scales: Forced-Choice vs. Psychometric Modeling of Likert Items.
- Author
-
Kreitchmann RS, Abad FJ, Ponsoda V, Nieto MD, and Morillo D
- Abstract
One important problem in the measurement of non-cognitive characteristics such as personality traits and attitudes is that it has traditionally been made through Likert scales, which are susceptible to response biases such as social desirability (SDR) and acquiescent (ACQ) responding. Given the variability of these response styles in the population, ignoring their possible effects on the scores may compromise the fairness and the validity of the assessments. Also, response-style-induced errors of measurement can affect the reliability estimates and overestimate convergent validity by correlating higher with other Likert-scale-based measures. Conversely, it can attenuate the predictive power over non-Likert-based indicators, given that the scores contain more errors. This study compares the validity of the Big Five personality scores obtained: (1) ignoring the SDR and ACQ in graded-scale items (GSQ), (2) accounting for SDR and ACQ with a compensatory IRT model, and (3) using forced-choice blocks with a multi-unidimensional pairwise preference model (MUPP) variant for dominance items. The overall results suggest that ignoring SDR and ACQ offered the worst validity evidence, with a higher correlation between personality and SDR scores. The two remaining strategies have their own advantages and disadvantages. The results from the empirical reliability and the convergent validity analysis indicate that when modeling social desirability with graded-scale items, the SDR factor apparently captures part of the variance of the Agreeableness factor. On the other hand, the correlation between the corrected GSQ-based Openness to Experience scores, and the University Access Examination grades was higher than the one with the uncorrected GSQ-based scores, and considerably higher than that using the estimates from the forced-choice data. Conversely, the criterion-related validity of the Forced Choice Questionnaire (FCQ) scores was similar to the results found in meta-analytic studies, correlating higher with Conscientiousness. Nonetheless, the FCQ-scores had considerably lower reliabilities and would demand administering more blocks. Finally, the results are discussed, and some notes are provided for the treatment of SDR and ACQ in future studies., (Copyright © 2019 Kreitchmann, Abad, Ponsoda, Nieto and Morillo.)
- Published
- 2019
- Full Text
- View/download PDF
35. Reconsidering Cutoff Points in the General Method of Empirical Q-Matrix Validation.
- Author
-
Nájera P, Sorrel MA, and Abad FJ
- Abstract
Cognitive diagnosis models (CDMs) are latent class multidimensional statistical models that help classify people accurately by using a set of discrete latent variables, commonly referred to as attributes. These models require a Q-matrix that indicates the attributes involved in each item. A potential problem is that the Q-matrix construction process, typically performed by domain experts, is subjective in nature. This might lead to the existence of Q-matrix misspecifications that can lead to inaccurate classifications. For this reason, several empirical Q-matrix validation methods have been developed in the recent years. de la Torre and Chiu proposed one of the most popular methods, based on a discrimination index. However, some questions related to the usefulness of the method with empirical data remained open due the restricted number of conditions examined, and the use of a unique cutoff point ( EPS ) regardless of the data conditions. This article includes two simulation studies to test this validation method under a wider range of conditions, with the purpose of providing it with a higher generalization, and to empirically determine the most suitable EPS considering the data conditions. Results show a good overall performance of the method, the relevance of the different studied factors, and that using a single indiscriminate EPS is not acceptable. Specific guidelines for selecting an appropriate EPS are provided in the discussion., Competing Interests: Declaration of Conflicting Interests: The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article., (© The Author(s) 2019.)
- Published
- 2019
- Full Text
- View/download PDF
36. Searching for G: A New Evaluation of SPM-LS Dimensionality.
- Author
-
Garcia-Garzon E, Abad FJ, and Garrido LE
- Abstract
There has been increased interest in assessing the quality and usefulness of short versions of the Raven's Progressive Matrices. A recent proposal, composed of the last twelve matrices of the Standard Progressive Matrices (SPM-LS), has been depicted as a valid measure of g . Nonetheless, the results provided in the initial validation questioned the assumption of essential unidimensionality for SPM-LS scores. We tested this hypothesis through two different statistical techniques. Firstly, we applied exploratory graph analysis to assess SPM-LS dimensionality. Secondly, exploratory bi-factor modelling was employed to understand the extent that potential specific factors represent significant sources of variance after a general factor has been considered. Results evidenced that if modelled appropriately, SPM-LS scores are essentially unidimensional, and that constitute a reliable measure of g . However, an additional specific factor was systematically identified for the last six items of the test. The implications of such findings for future work on the SPM-LS are discussed., Competing Interests: “The authors declare no conflict of interest.”
- Published
- 2019
- Full Text
- View/download PDF
37. A Two-Dimensional Multiple-Choice Model Accounting for Omissions.
- Author
-
Kreitchmann RS, Abad FJ, and Ponsoda V
- Abstract
This paper presents a new two-dimensional Multiple-Choice Model accounting for Omissions (MCMO). Based on Thissen and Steinberg multiple-choice models, the MCMO defines omitted responses as the result of the respondent not knowing the correct answer and deciding to omit rather than to guess given a latent propensity to omit. Firstly, using a Monte Carlo simulation, the accuracy of the parameters estimated from data with different sample sizes (500, 1,000, and 2,000 subjects), test lengths (20, 40, and 80 items) and percentages of omissions (5, 10, and 15%) were investigated. Later, the appropriateness of the MCMO to the Trends in International Mathematics and Science Study (TIMSS) Advanced 2015 mathematics and physics multiple-choice items was analyzed and compared with the Holman and Glas' Between-item Multi-dimensional IRT model (B-MIRT) and with the three-parameter logistic (3PL) model with omissions treated as incorrect responses. The results of the simulation study showed a good recovery of scale and position parameters. Pseudo-guessing parameters ( d ) were less accurate, but this inaccuracy did not seem to have an important effect on the estimation of abilities. The precision of the propensity to omit strongly depended on the ability values (the higher the ability, the worse the estimate of the propensity to omit). In the empirical study, the empirical reliability for ability estimates was high in both physics and mathematics. As in the simulation study, the estimates of the propensity to omit were less reliable and their precision varied with ability. Regarding the absolute item fit, the MCMO fitted the data better than the other models. Also, the MCMO offered significant increments in convergent validity between scores from multiple-choice and constructed-response items, with an increase of around 0.02 to 0.04 in R
2 in comparison with the two other methods. Finally, the high correlation between the country means of the propensity to omit in mathematics and physics suggests that (1) the propensity to omit is somehow affected by the country of residence of the examinees, and (2) the propensity to omit is independent of the test contents.- Published
- 2018
- Full Text
- View/download PDF
38. Cheating on Unproctored Internet Test Applications: An Analysis of a Verification Test in a Real Personnel Selection Context.
- Author
-
Aguado D, Vidal A, Olea J, Ponsoda V, Barrada JR, and Abad FJ
- Subjects
- Adult, Female, Humans, Male, Deception, Educational Measurement standards, Internet, Personnel Selection standards
- Abstract
This study analyses the extent to which cheating occurs in a real selection setting. A two-stage, unproctored and proctored, test administration was considered. Test score inconsistencies were concluded by applying a verification test (Guo and Drasgow Z-test). An initial simulation study showed that the Z-test has adequate Type I error and power rates in the specific selection settings explored. A second study applied the Z-test statistic verification procedure to a sample of 954 employment candidates. Additional external evidence based on item time response to the verification items was gathered. The results revealed a good performance of the Z-test statistic and a relatively low, but non-negligible, number of suspected cheaters that showed higher distorted ability estimates. The study with real data provided additional information on the presence of suspected cheating in unproctored applications and the viability of using item response times as an additional evidence of cheating. In the verification test, suspected cheaters spent 5.78 seconds per item more than expected considering the item difficulty and their assumed ability in the unproctored stage. We found that the percentage of suspected cheaters in the empirical study could be estimated at 13.84%. In summary, the study provides evidence of the usefulness of the Z-test in the detection of cheating in a specific setting, in which a computerized adaptive test for assessing English grammar knowledge was used for personnel selection.
- Published
- 2018
- Full Text
- View/download PDF
39. Assessing the Big Five with bifactor computerized adaptive testing.
- Author
-
Nieto MD, Abad FJ, and Olea J
- Subjects
- Adolescent, Adult, Female, Humans, Male, Middle Aged, Reproducibility of Results, Students psychology, Young Adult, Diagnosis, Computer-Assisted methods, Personality Assessment, Personality Disorders diagnosis, Personality Disorders psychology
- Abstract
Multidimensional computerized adaptive testing based on the bifactor model (MCAT-B) can provide efficient assessments of multifaceted constructs. In this study, MCAT-B was compared with a short fixed-length scale and computerized adaptive testing based on unidimensional (UCAT) and multidimensional (correlated-factors) models (MCAT) to measure the Big Five model of personality. The sample comprised 826 respondents who completed a pool with 360 personality items measuring the Big Five domains and facets. The dimensionality of the Big Five domains was also tested. With only 12 items per domain, the MCAT and MCAT-B procedures were more efficient to assess highly multidimensional constructs (e.g., Agreeableness), whereas no differences were found with UCAT and the short scale with traits that were essentially unidimensional (e.g., Extraversion). Furthermore, the study showed that MCAT and MCAT-B provide better content-balance of the pool because, for each Big Five domain, items from all the facets are selected in similar proportions. (PsycINFO Database Record (c) 2018 APA, all rights reserved).
- Published
- 2018
- Full Text
- View/download PDF
40. Modeling General, Specific, and Method Variance in Personality Measures: Results for ZKA-PQ and NEO-PI-R.
- Author
-
Abad FJ, Sorrel MA, Garcia LF, and Aluja A
- Subjects
- Adolescent, Adult, Factor Analysis, Statistical, Female, Humans, Male, Middle Aged, Personality Disorders diagnosis, Psychometrics, Young Adult, Models, Psychological, Personality Inventory
- Abstract
Contemporary models of personality assume a hierarchical structure in which broader traits contain narrower traits. Individual differences in response styles also constitute a source of score variance. In this study, the bifactor model is applied to separate these sources of variance for personality subscores. The procedure is illustrated using data for two personality inventories-NEO Personality Inventory-Revised and Zuckerman-Kuhlman-Aluja Personality Questionnaire. The inclusion of the acquiescence method factor generally improved the fit to acceptable levels for the Zuckerman-Kuhlman-Aluja Personality Questionnaire, but not for the NEO Personality Inventory-Revised. This effect was higher in subscales where the number of direct and reverse items is not balanced. Loadings on the specific factors were usually smaller than the loadings on the general factor. In some cases, part of the variance was due to domains being different from the main one. This information is of particular interest to researchers as they can identify which subscale scores have more potential to increase predictive validity.
- Published
- 2018
- Full Text
- View/download PDF
41. Inferential Item-Fit Evaluation in Cognitive Diagnosis Modeling.
- Author
-
Sorrel MA, Abad FJ, Olea J, de la Torre J, and Barrada JR
- Abstract
Research related to the fit evaluation at the item level involving cognitive diagnosis models (CDMs) has been scarce. According to the parsimony principle, balancing goodness of fit against model complexity is necessary. General CDMs require a larger sample size to be estimated reliably, and can lead to worse attribute classification accuracy than the appropriate reduced models when the sample size is small and the item quality is poor, which is typically the case in many empirical applications. The main purpose of this study was to systematically examine the statistical properties of four inferential item-fit statistics: S - X 2 , the likelihood ratio (LR) test, the Wald (W) test, and the Lagrange multiplier (LM) test. To evaluate the performance of the statistics, a comprehensive set of factors, namely, sample size, correlational structure, test length, item quality, and generating model, is systematically manipulated using Monte Carlo methods. Results show that the S - X 2 statistic has unacceptable power. Type I error and power comparisons favor LR and W tests over the LM test. However, all the statistics are highly affected by the item quality. With a few exceptions, their performance is only acceptable when the item quality is high. In some cases, this effect can be ameliorated by an increase in sample size and test length. This implies that using the above statistics to assess item fit in practical settings when the item quality is low remains a challenge., Competing Interests: Declaration of Conflicting Interests: The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
- Published
- 2017
- Full Text
- View/download PDF
42. Comparison of methods for dealing with missing values in the EPV-R.
- Author
-
Paniagua D, Amor PJ, Echeburúa E, and Abad FJ
- Subjects
- Humans, Male, Models, Statistical, Risk Assessment, Self Report, Intimate Partner Violence statistics & numerical data
- Abstract
Background: The development of an effective instrument to assess the risk of partner violence is a topic of great social relevance. This study evaluates the scale of “Predicción del Riesgo de Violencia Grave Contra la Pareja” –Revisada– (EPV-R - Severe Intimate Partner Violence Risk Prediction Scale-Revised), a tool developed in Spain, which is facing the problem of how to treat the high rate of missing values, as is usual in this type of scale., Method: First, responses to the EPV-R in a sample of 1215 male abusers who were reported to the police were used to analyze the patterns of occurrence of missing values, as well as the factor structure. Second, we analyzed the performance of various imputation methods using simulated data that emulates the missing data mechanism found in the empirical database., Results: The imputation procedure originally proposed by the authors of the scale provides acceptable results, although the application of a method based on the Item Response Theory could provide greater accuracy and offers some additional advantages., Conclusions: Item Response Theory appears to be a useful tool for imputing missing data in this type of questionnaire.
- Published
- 2017
- Full Text
- View/download PDF
43. Calibrating a new item pool to adaptively assess the Big Five.
- Author
-
Nieto MD, Abad FJ, Hernández-Camacho A, Garrido LE, Barrada JR, Aguado D, and Olea J
- Subjects
- Adolescent, Adult, Female, Humans, Male, Middle Aged, Young Adult, Personality Tests
- Abstract
Background: Even though the Five Factor Model (FFM) has been the dominant paradigm in personality research for the past two decades, very few studies have measured the FFM adaptively. Thus, the purpose of this research was the building of a new item pool to develop a computerized adaptive test (CAT) for personality assessment., Method: A pool of 480 items that measured the FFM facets was developed and applied to 826 participants. Facets were calibrated separately and item selection was performed being mindful of the preservation of unidimensionality of each facet. Then, a post-hoc simulation study was carried out to test the performance of separate CATs to measure the facets., Results: The final item pool was composed of 360 items with good psychometric properties. Findings reveal that a CAT administration of four items per facet (total length of 120 items) provides accurate facets scores, while maintaining the factor structure of the FFM., Conclusions: An item pool with good psychometric properties was obtained and a CAT simulation study demonstrated that the FFM facets could be measured with precision using a third of the items in the pool.
- Published
- 2017
- Full Text
- View/download PDF
44. Iteration of Partially Specified Target Matrices: Application to the Bi-Factor Case.
- Author
-
Abad FJ, Garcia-Garzon E, Garrido LE, and Barrada JR
- Subjects
- Data Interpretation, Statistical, Humans, Monte Carlo Method, Multivariate Analysis, Quality of Life, Factor Analysis, Statistical, Models, Statistical
- Abstract
The current study proposes a new bi-factor rotation method, Schmid-Leiman with iterative target rotation (SLi), based on the iteration of partially specified target matrices and an initial target constructed from a Schmid-Leiman (SL) orthogonalization. SLi was expected to ameliorate some of the limitations of the previously presented SL bi-factor rotations, SL and SL with target rotation (SLt), when the factor structure either includes cross-loadings, near-zero loadings, or both. A Monte Carlo simulation was carried out to test the performance of SLi, SL, SLt, and the two analytic bi-factor rotations, bi-quartimin and bi-geomin. The results revealed that SLi accurately recovered the bi-factor structures across the majority of the conditions, and generally outperformed the other rotation methods. SLi provided the biggest improvements over SL and SLt when the bi-factor structures contained cross-loadings and pure indicators of the general factor. Additionally, SLi was superior to bi-quartimin and bi-geomin, which performed inconsistently across the types of factor structures evaluated. No method produced a good recovery of the bi-factor structures when small samples (N = 200) were combined with low factor loadings (0.30-0.50) in the specific factors. Thus, it is recommended that larger samples of at least 500 observations be obtained.
- Published
- 2017
- Full Text
- View/download PDF
45. Structural brain connectivity and cognitive ability differences: A multivariate distance matrix regression analysis.
- Author
-
Ponsoda V, Martínez K, Pineda-Pardo JA, Abad FJ, Olea J, Román FJ, Barbey AK, and Colom R
- Subjects
- Adolescent, Brain Mapping, Female, Humans, Imaging, Three-Dimensional, Magnetic Resonance Imaging, Male, Neural Pathways diagnostic imaging, Neural Pathways physiology, Neuropsychological Tests, Reproducibility of Results, Young Adult, Brain diagnostic imaging, Brain physiology, Cognition physiology, Multivariate Analysis, Regression Analysis
- Abstract
Neuroimaging research involves analyses of huge amounts of biological data that might or might not be related with cognition. This relationship is usually approached using univariate methods, and, therefore, correction methods are mandatory for reducing false positives. Nevertheless, the probability of false negatives is also increased. Multivariate frameworks have been proposed for helping to alleviate this balance. Here we apply multivariate distance matrix regression for the simultaneous analysis of biological and cognitive data, namely, structural connections among 82 brain regions and several latent factors estimating cognitive performance. We tested whether cognitive differences predict distances among individuals regarding their connectivity pattern. Beginning with 3,321 connections among regions, the 36 edges better predicted by the individuals' cognitive scores were selected. Cognitive scores were related to connectivity distances in both the full (3,321) and reduced (36) connectivity patterns. The selected edges connect regions distributed across the entire brain and the network defined by these edges supports high-order cognitive processes such as (a) (fluid) executive control, (b) (crystallized) recognition, learning, and language processing, and (c) visuospatial processing. This multivariate study suggests that one widespread, but limited number, of regions in the human brain, supports high-level cognitive ability differences. Hum Brain Mapp 38:803-816, 2017. © 2016 Wiley Periodicals, Inc., (© 2016 Wiley Periodicals, Inc.)
- Published
- 2017
- Full Text
- View/download PDF
46. A Dominance Variant Under the Multi-Unidimensional Pairwise-Preference Framework: Model Formulation and Markov Chain Monte Carlo Estimation.
- Author
-
Morillo D, Leenen I, Abad FJ, Hontangas P, de la Torre J, and Ponsoda V
- Abstract
Forced-choice questionnaires have been proposed as a way to control some response biases associated with traditional questionnaire formats (e.g., Likert-type scales). Whereas classical scoring methods have issues of ipsativity, item response theory (IRT) methods have been claimed to accurately account for the latent trait structure of these instruments. In this article, the authors propose the multi-unidimensional pairwise preference two-parameter logistic (MUPP-2PL) model, a variant within Stark, Chernyshenko, and Drasgow's MUPP framework for items that are assumed to fit a dominance model. They also introduce a Markov Chain Monte Carlo (MCMC) procedure for estimating the model's parameters. The authors present the results of a simulation study, which shows appropriate goodness of recovery in all studied conditions. A comparison of the newly proposed model with a Brown and Maydeu's Thurstonian IRT model led us to the conclusion that both models are theoretically very similar and that the Bayesian estimation procedure of the MUPP-2PL may provide a slightly better recovery of the latent space correlations and a more reliable assessment of the latent trait estimation errors. An application of the model to a real data set shows convergence between the two estimation procedures. However, there is also evidence that the MCMC may be advantageous regarding the item parameters and the latent trait correlations., Competing Interests: The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
- Published
- 2016
- Full Text
- View/download PDF
47. The relationships between WAIS-IV factor index scores and educational level: A bifactor model approach.
- Author
-
Abad FJ, Sorrel MA, Román FJ, and Colom R
- Subjects
- Adolescent, Adult, Aged, Female, Humans, Intelligence, Male, Middle Aged, Spain, Young Adult, Educational Status, Wechsler Scales
- Abstract
IQ summary scores may not involve equivalent psychological meaning for different educational levels. Ultimately, this relates to the distinction between constructs and measurements. Here, we explore this issue studying the standardization of the Wechsler Adult Intelligence Scale-Fourth Edition (WAIS-IV) for Spain. A representative sample of 743 individuals (374 females and 369 males) who completed the 15 subtests comprising this intelligence battery was considered. We analyzed (a) the best latent factor structure for modeling WAIS-IV subtest performance, (b) measurement invariance across educational levels, and (c) the relationships of educational level/attainment with latent factors, Full Scale IQ (FSIQ), and index factor scores. These were the main findings: (a) the bifactor model provides the best fit; (b) there is partial invariance, and therefore it is concluded that the battery is a proper measure of the constructs of interest for the educational levels analyzed (nevertheless, the relevance of g decreases at high educational levels); (c) at the latent level, g and, to a lesser extent, Verbal Comprehension and Processing Speed, are positively related to educational level/attainment; (d) despite the previous finding, we find that Verbal Comprehension and Processing Speed factor index scores have reduced incremental validity beyond FSIQ; and (e) FSIQ is a slightly biased measure of g. (PsycINFO Database Record, ((c) 2016 APA, all rights reserved).)
- Published
- 2016
- Full Text
- View/download PDF
48. Are fit indices really fit to estimate the number of factors with categorical variables? Some cautionary findings via Monte Carlo simulation.
- Author
-
Garrido LE, Abad FJ, and Ponsoda V
- Subjects
- Humans, Data Interpretation, Statistical, Models, Statistical, Monte Carlo Method
- Abstract
An early step in the process of construct validation consists of establishing the fit of an unrestricted "exploratory" factorial model for a prespecified number of common factors. For this initial unrestricted model, researchers have often recommended and used fit indices to estimate the number of factors to retain. Despite the logical appeal of this approach, little is known about the actual accuracy of fit indices in the estimation of data dimensionality. The present study aimed to reduce this gap by systematically evaluating the performance of 4 commonly used fit indices-the comparative fit index (CFI), the Tucker-Lewis index (TLI), the root mean square error of approximation (RMSEA), and the standardized root mean square residual (SRMR)-in the estimation of the number of factors with categorical variables, and comparing it with what is arguably the current golden rule, Horn's (1965) parallel analysis. The results indicate that the CFI and TLI provide nearly identical estimations and are the most accurate fit indices, followed at a step below by the RMSEA, and then by the SRMR, which gives notably poor dimensionality estimates. Difficulties in establishing optimal cutoff values for the fit indices and the general superiority of parallel analysis, however, suggest that applied researchers are better served by complementing their theoretical considerations regarding dimensionality with the estimates provided by the latter method., ((c) 2016 APA, all rights reserved).)
- Published
- 2016
- Full Text
- View/download PDF
49. Traditional scores versus IRT estimates on forced-choice tests based on a dominance model.
- Author
-
Hontangas PM, Leenen I, de la Torre J, Ponsoda V, Morillo D, and Abad FJ
- Subjects
- Humans, Psychometrics, Surveys and Questionnaires, Choice Behavior, Models, Psychological
- Abstract
Background: Forced-choice tests (FCTs) were proposed to minimize response biases associated with Likert format items. It remains unclear whether scores based on traditional methods for scoring FCTs are appropriate for between-subjects comparisons. Recently, Hontangas et al. (2015) explored the extent to which traditional scoring of FCTs relates to the true scores and IRT estimates. The authors found certain conditions under which traditional scores (TS) can be used with FCTs when the underlying IRT model was an unfolding model. In this study, we examine to what extent the results are preserved when the underlying process becomes a dominance model., Method: The independent variables analyzed in a simulation study are: forced-choice format, number of blocks, discrimination of items, polarity of items, variability of intra-block difficulty, range of difficulty, and correlation between dimensions., Results: A similar pattern of results was observed for both models; however, correlations between TS and true thetas are higher and the differences between TS and IRT estimates are less discrepant when a dominance model involved., Conclusions: A dominance model produces a linear relationship between TS and true scores, and the subjects with extreme thetas are better measured.
- Published
- 2016
- Full Text
- View/download PDF
50. Dealing with the high cost of biological therapies: developing and implementing a biological therapy prioritization protocol for ankylosing spondylitis patients in a tertiary hospital.
- Author
-
Borrás-Blasco J, Casterá E, Cortes X, Martín-Alonso J, Rosique-Robles JD, and Abad FJ
- Subjects
- Adult, Aged, Clinical Protocols, Female, Health Priorities, Humans, Male, Middle Aged, Tertiary Care Centers, Biological Therapy economics, Health Care Costs, Spondylitis, Ankylosing therapy
- Abstract
Objective: In January 2011, a biological therapies commission was created in our hospital to fully address the management of biological drugs. A biological therapy prioritization protocol was developed for ankylosing spondylitis (AS) patients. Here, we describe it and report on its economic impact to illustrate how we are optimizing the use of these expensive new drugs., Methods: The biological therapies commission established several procedures for the rational use of biological drugs such as cost-efficiency therapeutic protocols, pharmacovigilance, and therapeutic drug monitoring programs. The AS protocol was based on clinical and economic aspects. We estimated the economic impact of the protocol by comparing the cost of treating AS patients with biological drugs in the pre-commission (2009 - 2010) vs. post-commission period (2011 - 2013). AS patients treated with adalimumab (ADA), etanercept (ETN) or infliximab (IFX) for at least 6 months in the 2009 - 2013 period were included., Results: 107 patients were included. In the pre-commission period, total expenses increased by +30,944 Euro (+4%). After protocol implementation, total expenses decreased by 11,441 Euro (-1%) during 2011, and by an additional 36,781 Euro (-4%) and 53,872 Euro (-8%) in 2012 and 2013, respectively. In the 2010 - 2013 period the cost of biological therapy per patient-year decreased by 869 â¬, suggesting the positive effects of the biological therapy prioritization protocol instauration., Conclusion: We describe the establishment of a multidisciplinary biological therapy commission to optimize the use of biological therapies. We illustrate its work in developing a protocol for the management of AS patients with such therapies. We show that after 3-years of implementation, the biological therapy prioritization protocol allowed us to steadily decrease the direct cost of biological drug therapies per patient, up to 869 Euro.
- Published
- 2015
- Full Text
- View/download PDF
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.