Could we perhaps summarize (beyond the revisions I already agreed to make in my Word-file replies) what I must do next?
To be clear: I approve the paper, conditioned on the agreed up revisions.
Back to [Archive] Post-review discussions
Could we perhaps summarize (beyond the revisions I already agreed to make in my Word-file replies) what I must do next?
REPLY: Would it be ok to include “(but, see, <your link>)” immediately after the word “version” in our quote above? If not, we can delete this entire section.
No. Please see the discussion about your previous submission. The BDS does not have twice the g-loading of FDS. Dalliard and myself compiled several studies to show this, yet you made the same exact claim in your next submission. I find that odd.
This (the slightly higher mean IQ among the occupations) corresponds to the low correlation found between being out of a job and IQ at the individual level. You could back-estimate this correlation using this mean. Just a minor check.
You misunderstood. I proposed that you use the average of the BLS 2014 and 2012 values to remove some of the slight 'measurement error'.
I did this in my replication.
1)
Not controlling for known confounds (interests in this case) which you already have the data for is not really defensible. After all, you are interested in the effect of cognitive ability itself, not whatever it is that it happens to be correlated with. If you use correlations, you will get a confounded estimate of the influence of cognitive ability itself.
Your correlations are effect sizes, yes. However, I asked for "the effect sizes of the prior research so readers can see whether the effect sizes are similar".
You present some new results. What the readers need to know is whether they fit in size with the previous results. For instance, if you find r = .20 and previous studies have found r = .95, something is wrong somewhere.
The data variable (complexity) was correlated with mean IQ at .86 in your study. You cite:
Gottfredson, L. S. (1986). Occupational aptitude patterns map: Development and implications for a theory of job aptitude requirements (Monograph). Journal of Vocational Behavior, 29, 254-291.
Gottfredson, L. S. (2003). g, jobs, and life. In H. Nyborg (Ed.), The scientific study of general intelligence: Tribute to Arthur R. Jensen (pp. 293-342). New York: Pergamon.
However, I could not find any complexity x mean IQ correlation in these papers. She does give job mean IQs and presents factor analysis results of job attributes, but does not appear to actually correlate them. Maybe I missed the number somewhere?
I admit to not being a statistician, but to me, I found it more compelling to show that the demographic percentages were stable across two time periods than to get slightly more precise estimates by averaging them.
I misinterpreted what you meant last time, and see that you have done the analyses below. How shall I proceed?
I will look into this further. It’s possible Gottfredson didn’t report any.
White, Black or African American, Asian, American Indian and Alaska Native, and Native Hawaiian and Other Pacific Islander. In accordance with the Office of Management and Budget guidelines, these terms are used to describe the race of people. Beginning in 2003, people in these categories are those who selected that race group only. Those who identify multiple race groups are categorized as people of Two or More Races. (Previously, people identified a group as their main race.) People who identified themselves as Asian are further classified as Asian Indian, Chinese, Filipino, Japanese, Korean, Vietnamese, or Other Asian. The Other Asian category includes individuals of group not listed—such as Pakistani, Hmong, and Cambodian— and those who reported two or more Asian groups. Estimates for American Indians and Alaska Natives, Native Hawaiians and Other Pacific Islanders, and people of Two or More Races are not shown separately in all tables because the number of survey respondents is too small to develop estimates of sufficient quality. In the enumeration process, race is determined by the household respondent. More information on the 2003 changes to questions on race and Hispanic ethnicity is available on the BLS website at www.bls.gov/cps/rvcps03.pdf.
Hispanic or Latino ethnicity. This refers to people who identified themselves in the enumeration process as being of Hispanic, Latino or Spanish origin. These individuals are further classified by detailed Hispanic ethnicity. Previous versions of this report presented data for the following detailed Hispanic ethnicity categories: Mexican, Puerto Rican, Cuban, Central and South American, or Other Hispanic or Latino. The latter two categories were expanded in 2014 into additional categories: Central American, which includes the two subcategories of Salvadoran and Other Central American (excluding Salvadorans); South American; and Other Hispanic or Latino, which includes the two subcategories of Dominican and Other Hispanic or Latino (excluding Dominicans). People whose ethnicity is identified as Hispanic or Latino may be of any race. More information on the 2003 changes in questions on race and Hispanic ethnicity is available online at www.bls.gov/cps/rvcps03.pdf
Bryan,
So, I'd like you to use the average values of years 2012 and 2014 for the BLS data, and note their intercorrelation (as you already do).
...
Furthermore, I think that you should include the regression results. I.e. have each of the race% as outcome (dependent) variables, and use mean IQ, people, and things as predictors (independents).
....
Perhaps contact Linda Gottfredson to hear. Perhaps she is familiar with other literature on this topic.
....
As Fuerst notes, it is a good idea to note that the White, Black, Asian groups in this study include Hispanics.
....
One should also do the analyses for Hispanic%. It is not important, statistically, that this group overlaps with the others because you are not doing a simultaneous analyses of the race%'s.
Bryan,
So, I'd like you to use the average values of years 2012 and 2014 for the BLS data, and note their intercorrelation (as you already do).
...
Furthermore, I think that you should include the regression results. I.e. have each of the race% as outcome (dependent) variables, and use mean IQ, people, and things as predictors (independents).
....
Perhaps contact Linda Gottfredson to hear. Perhaps she is familiar with other literature on this topic.
....
As Fuerst notes, it is a good idea to note that the White, Black, Asian groups in this study include Hispanics.
....
One should also do the analyses for Hispanic%. It is not important, statistically, that this group overlaps with the others because you are not doing a simultaneous analyses of the race%'s.
Emil,
For sake of transparency -- I've asked this of other reviewers, too -- could you clarify which changes you absolutely require for approval versus which you recommend but do not insist upon? Thanks.
Of my comments, I require:
1. The regression analyses to control for the interest differences.
I recommend but do not require:
1. Using the average of BLS 2012 and BLS 2014 values.
2. Doing the Hispanic analyses.
3. Contact Linda Gottfredson to hear if someone else correlated job complexity with mean IQs by job/occupation.
After (1), I have no further objections and will approve of publication.
---
I'm curious as to why you don't want to do the Hispanic one? In fact, the Hispanic one is the only one that uses a clear definition! It's the White, Black and Asian which are confounded with Hispanics (mostly White).
Of my comments, I require:
1. The regression analyses to control for the interest differences.
I recommend but do not require:
1. Using the average of BLS 2012 and BLS 2014 values.
2. Doing the Hispanic analyses.
3. Contact Linda Gottfredson to hear if someone else correlated job complexity with mean IQs by job/occupation.
After (1), I have no further objections and will approve of publication.
---
I'm curious as to why you don't want to do the Hispanic one? In fact, the Hispanic one is the only one that uses a clear definition! It's the White, Black and Asian which are confounded with Hispanics (mostly White).
Hello,
I will do the regressions and all other changes suggested earlier. I just can't get past checking a box that says white, e.g., and then also a box that says Hispanic. That plus things summing to greater than 100% concerns me.
I will do the regressions and all other changes suggested earlier. I just can't get past checking a box that says white, e.g., and then also a box that says Hispanic. That plus things summing to greater than 100% concerns me.
Also, may I ask why you didn't include Data in with the regressions you want me to report?
> lm_white = lm("white ~ iq + people + things", data = d_jobdata) %>% MOD_summary(runs = 200)
> lm_white
$coefs
Beta SE CI.lower CI.upper
iq 0.39 0.09 0.20 0.57
people -0.19 0.10 -0.37 0.00
things -0.09 0.08 -0.26 0.07
$meta
N R2 R2 adj. R2 10-fold cv
124.00 0.26 0.24 0.19
> lm_white = lm("white ~ data + people + things", data = d_jobdata) %>% MOD_summary(runs = 200)
> lm_white
$coefs
Beta SE CI.lower CI.upper
data -0.44 0.09 -0.62 -0.27
people -0.16 0.09 -0.34 0.02
things -0.08 0.08 -0.24 0.08
$meta
N R2 R2 adj. R2 10-fold cv
124.00 0.29 0.28 0.21
> lm_white = lm("white ~ iq + data + people + things", data = d_jobdata) %>% MOD_summary(runs = 200)
> lm_white
$coefs
Beta SE CI.lower CI.upper
iq 0.07 0.15 -0.23 0.37
data -0.39 0.15 -0.69 -0.09
people -0.15 0.09 -0.34 0.03
things -0.08 0.08 -0.24 0.08
$meta
N R2 R2 adj. R2 10-fold cv
124.00 0.30 0.27 0.19
Any update?
Bryan
Jensen's method (correlated vectors)
When one has multiple, preferably many, indicators of a latent trait one can use Jensen's method to check whether this latent variable is related to a criterion variable or if its the other variance components (group factors or test/item specificity). This method requires that one has multiple indicators each of which have some measurement of how well they measure the latent trait. Usually factor loadings from factor analysis are used for this purpose, but if one has item-level data, one should use item response theory-based discrimination values instead.
It is true that jobs are basically mental tests of varying difficulty. SH for this kind of data would be the claim that the jobs where the g-job performance link is stronger would show larger group differences in job performance, assuming no other effects (such as selective hiring which is ubiquitous). However, hat is not the kind of data this study has. The data here are the racial proportions of each job and some information about the jobs. One cannot frame the current study as a test of SH.
However, I still think the present study is useful and I have no serious criticism of the methods used, but it's a study of something else. It's a study of what happens when one has groups with different mean ability levels and there are jobs that select for different levels of cognitive ability. In the absence of differential hiring, there should be more of the higher scoring groups in the jobs that recruit from higher up the cognitive ability distribution. Of course, we know that there is some differential hiring and using IQ tests without doing this may be illegal (but not in the military).
Preferably, one should try to model the racial proportions together based on demographic data and assumptions about the range each job recruits workers from. However, the authors prefer to take a simpler approach and check the simpler prediction that the jobs with higher means do have more persons from the higher ability groups. This is fine with me.
Write-up of regressions
The write-up of the additional regression analyses is not satisfactory. These regressions were not based on any pre-analysis hypothesizing as the writing says. They were entirely exploratory, post hoc regressions and should be clearly labeled as such. To do otherwise, is to HARK (http://psr.sagepub.com/content/2/3/196.abstract). I would write something like:
A reviewer* suggested using multiple regression by including the jobs' ratings for whether they involve working with people and things. This was done because there may be group differences in these preferences which may obscure the relationship to the mean levels of cognitive ability. A regression model was fit for each race% as the outcome and with the mean job IQ, people-interest rating and person-interest rating as the predictors. [3 tables of results] The mean IQ was a good predictor in all the models and, importantly, the small correlation seen for Asian% seemed to be due to a suppression effect from a confound with a relatively stronger preference among Asians for working with people.
I'm sorry, but the framing of the study has to be changed. This is not a test of Spearman's hypothesis.
* I usually name reviewers who give useful suggestions.
SH: The more x requires g, the larger the race difference on x.
In your example, x is job performance. The more job performance requires g, the larger the race difference. The dependent variable is job performance, measured quantitatively (e.g., units produced) or qualitatively (e.g., “excellent,” “poor”).
In my example, x is being employed in the job itself. The more being employed in a job or not depends on g, the larger the race difference. The dependent variable is “representation” (i.e., the relative percent of job holders who are White, Black, and Asian).
I do think this is a test of SH because the more IQ matters toward getting a job, the more Whites / the less Blacks there should be working that job. In other words, over/under representation in a job is partly determined by IQ, as predicted by SH, and indeed Blacks (e.g.) are more and more under-represented as job IQ goes up.
Obviously, the harked analyses weren’t in the original submission. We harked to “Hopefully Appease Reviewer Kirkegaard” (a double hark!?).
The analyses you suggested never occurred to us as we submitted the original paper here. So, adding the analyses had to be post hoc and exploratory. In fact we start discussion of this by stating “A reviewer recommended we explore….” I think that language makes it patently obvious that the analyses to follow are post hoc. We should add more disclaimers?
It occurs to me that “harking by reviewer” is interesting, and is likely an indictment that authors didn’t do something initially that they should have.
A reviewer recommended we explore further the relatively weak correlation between percent Asian and IQ.
We therefore conducted multiple regression analyses testing a hypothesis that Asians gravitate more toward “people jobs” (i.e., values for People might suppress the IQ / percent Asian correlation).
However, this analysis on percent Asian produced beta weights of .41 (IQ), .34 (People), and .00 (Things). It appears that values on People suppress the true correlation between Job IQ and percent Asian.
Also, (1) I don’t see what three regression tables add but length to the paper, and (2) I’ve always never acknowledged reviewers by name because it seems to diminish the significant service they do for our profession, for free, when they peer review.
Bryan,SH: The more x requires g, the larger the race difference on x.
In your example, x is job performance. The more job performance requires g, the larger the race difference. The dependent variable is job performance, measured quantitatively (e.g., units produced) or qualitatively (e.g., “excellent,” “poor”).
In my example, x is being employed in the job itself. The more being employed in a job or not depends on g, the larger the race difference. The dependent variable is “representation” (i.e., the relative percent of job holders who are White, Black, and Asian).
I do think this is a test of SH because the more IQ matters toward getting a job, the more Whites / the less Blacks there should be working that job. In other words, over/under representation in a job is partly determined by IQ, as predicted by SH, and indeed Blacks (e.g.) are more and more under-represented as job IQ goes up.
In the way you frame it here, yes, that would a kind of SH test. But note that your analysis do not look at group differences directly, and neither does it look at under or over-representation directly. It only looks at representation (proportion of workings from each SIRE). I don't see how it can be a test of SH without actual group difference data being analyzed.
Would it ever be possible to test SH by looking at group differences indirectly?
I'm not sure I'm correct here, and wonder what other readers think about this study being a test of SH?
I'll reply to your other comments after we resolve this issue, as it seems key to the paper's future.
They score six (“speaking-signaling”) on people, and two (“operating-controlling) on things.
I was suggesting that you might qualify your statement. For example, for precision, I might say: Spearman's hypothesis would predict that group differences are larger on more g-loaded tests, assuming no countervailing psychometric bias. Likewise: Spearman's hypothesis would predict that employment differences are larger for more g-loaded fields, assuming no countervailing societal bias e.g., affirmative action or defacto quotas. If you think that the qualification is obvious, don't bother.