Cognitive Metrics logo CognitiveMetrics

Otis Gamma / GET

Key Points
  • This article evaluates the psychometric properties of the Gifted Entry Test (GET), a modernized adaptation of the Otis Gamma (Form EM), using a large online sample (N = 76,009).
  • Results indicate excellent internal consistency (α ≈ .93) and strong measurement precision across a wide range of ability.
  • The GET demonstrated a substantial loading on general intelligence (λ = .81) and strong convergent validity with the AGCT (r ≈ .74), increasing to approximately r ≈ .82 after correction for range restriction.
  • Linear regression indicated strong normative alignment (β = 0.93, b₀ = 5.57, p < .001) with the AGCT, whose norms are derived from the general population.
  • IRT analyses supported a 2PL model and indicated broad coverage of ability, though with reduced ceiling at the upper range.
  • Despite limitations related to sample self-selection and ceiling effects, this article finds that the GET is a reliable, valid, and efficient measure of general intelligence suitable for large-scale online testing.

Introduction

Assessing general cognitive ability has long been central in educational and workplace selection contexts. Some of the most widely used intelligence tests in the early 20th century were developed by Arthur S. Otis, an American psychologist who pioneered the development of group-administered intelligence tests, making large-scale cognitive assessment more practical (Iqbal, 2024). First published in 1922, Otis’ tests focused on measuring general intelligence (g) through a variety of tasks covering verbal, quantitative, and fluid reasoning abilities. While the test was widely used in educational and selection contexts throughout the early to mid-twentieth century, it fell out of use over time, being replaced by more modern standardized cognitive and aptitude tests, such as the SAT and Wechsler scales. The Gifted Entry Test (GET), hosted on CognitiveMetrics, is a restoration and modernization of the Otis Gamma, Form EM. This report assesses the reliability and validity of CognitiveMetrics’ large-scale administration (N = 76,009) of the GET/Otis Gamma in a contemporary population via Item Response Theory and Confirmatory Factor Analysis, while also confirming the accuracy of the updated norms.

Background

Historical Background

During World War I, a significant demand for mass psychological testing arose from the need to efficiently classify large numbers of military recruits. Otis contributed to the development of intelligence tests (such as the Army Alpha) as part of the U.S. Army’s World War I testing program under the direction of Robert Yerkes. Following the war, Otis developed a series of tests collectively known as the Otis Group Intelligence Scales, which evolved through multiple revisions and formats across the 20th century. These included instruments tailored for educational, industrial, and business settings, emphasizing quick administration and easy scoring. Among these developments were the Otis Quick Scoring Mental Ability Test series, which includes three tests (Alpha, Beta, Gamma). The Alpha is intended for grades 1-4, Beta for grades 4-9, and the Gamma for high school/college students. The Otis Gamma has six unique forms: AM, BM, C, D, EM, and FM (Otis, 1937).

Test Description

The Otis Gamma was a multiple-choice written test designed to measure general intelligence through a combination of verbal, logical, and numerical reasoning tasks. The EM Form, adapted by the GET, was a Spanish adaptation distributed in Latin America and used in applied settings such as personnel selection and educational evaluation. According to its technical specifications, the Otis Gamma EM: Contains 80 items of progressive difficulty Has a 30-minute time limit Is intended for testees with high school education or above Can be taken individually or in group settings Demonstrates strong internal consistency (reported split-half reliability ≈ .88) (Gerardo Pérez & Vite San Pedro, 2012) The test samples multiple cognitive domains, including vocabulary, verbal analogies, sentence organization, logical reasoning, numerical series, and arithmetic problem solving. As such, it is best understood as a measure of general intelligence (g) rather than any specific, narrowly defined aptitude.

Adaptation for GET

The CognitiveMetrics GET adaptation of the Otis Gamma EM involved translating the Spanish form into English, with care taken to preserve the intended logic behind each item. Verbal knowledge and reasoning items were especially carefully reviewed, with the original semantic relationships and reasoning taken into account to ensure that the translated items were as faithful as possible to the originals. To evaluate the psychometric integrity of the adapted instrument, item response theory (IRT) analyses were conducted, providing evidence that the translated items function appropriately in measuring general cognitive ability (g) within an English-speaking contemporary population. As the sample size increased, the test’s norms were updated multiple times using IRT to improve interpretability and allow for comparison with other IQ scales. One aim of this report is to evaluate the validity of these norms.

Methods

Descriptive Statistics

The sample consists of 76,009 individuals who completed the GET on CognitiveMetrics, an online cognitive testing platform. Most participants likely encountered the test through online communities, such as the r/cognitiveTesting and r/iqtest subreddits, suggesting a self-selected sample with a pre-existing interest in IQ and cognitive testing. This is unlikely to be randomly distributed; individuals are generally more interested in domains they expect to perform well in, resulting in a selection bias toward higher-ability participants. For example, on the CORE FSIQ test, the number of subtests voluntarily completed (used here as a proxy for interest in IQ) was positively associated with IQ at r ≈ 0.25 (see Table 1). Testees who invested more time and effort to complete more subtests tended to have higher IQ.

Table 1. Selection Effect on IQ Scores by Number of CORE Subtests Completed
Subtests Taken Mean IQ SD n
1 113.2 13.65 12,297
2 115.7 11.98 4,234
3 117.0 12.21 1,923
4 118.5 12.12 1,173
5 119.1 11.60 967
6 120.3 11.72 761
7 119.7 12.00 641
8 119.5 12.05 612
9 121.4 11.77 544
10 122.9 12.08 586
11 123.1 11.99 590
12 122.7 11.05 469
13 123.6 11.18 337
14 122.3 12.32 110
15 122.3 10.94 118
16 124.0 10.25 130
17 127.8 10.45 144
Note. Interest in IQ (as measured by the number of subtests voluntarily taken) correlated with IQ itself r ≈ 0.245 in this sample.

Similarly, as shown in Table 2, participants who took the GET on CognitiveMetrics constitute a high-ability sample, averaging 117.3 IQ, a bit more than one standard deviation above the mean. A smaller standard deviation of 13.55 (1.45 points below the population standard deviation of 15) is further evidence of self-selection.

Table 2. GET Sample Descriptive Statistics
Statistic Value
N 76,909
Mean 117.1
SE Mean 0.0492
StDev 13.55

Visual inspection of the distribution plot in Figure 1 indicates a negative skew and evidence of a ceiling effect at higher IQ levels. The normal probability plot in Figure 2 shows scores clustering toward the upper end of the distribution as well, an expected pattern given the high-ability sample. The ceiling of the test is 148 IQ, meaning participants within this high ability sample who approach and exceed that value will be truncated. Accordingly, the observed non-normality should not be interpreted as a flaw of the test, but rather as a characteristic of the sampled population and testing context.

Figure 1. Distribution of IQ Scores on the GET
Figure 2. Normal Probability Plot of GET IQ Scores

Geographic Representativeness

Due to the test being in English, this sample is restricted to participants residing in the following English-speaking, Anglospheric countries: the United States, Canada, the United Kingdom, Australia, and New Zealand.

While self-selected, the GET sample still exhibited broad geographic diversity. To evaluate this, the number of participants from each of the five previously mentioned countries was compared to national population size, revealing a near-perfect linear relationship (r ≈ .997), indicating that participation scaled proportionally with population across countries.

This was then extended to the state level within the United States (as observed in Figure 4), where a similarly strong relationship was observed (r ≈ .986). The consistency of this pattern across both the country and state level suggests that the sample is broadly representative in its geographic distribution, with participation selecting from the overall population rather than specific regions.

Table 3. GET Sample by Country
Country N
United States 55,688
United Kingdom 7,940
Canada 7,751
Australia 3,910
New Zealand 720
Figure 3. Geographic Distribution of GET Test Takers Across the United States
Figure 4. Relationship Between US State Population and Number of GET Test Takers

Data Preparation

Out of the larger set of unsorted test attempts, a total of 76,009 were retained for analysis after cleaning the data. The data was first filtered to remove testees not from the aforementioned English-speaking countries. Next, all repeat attempts were removed, ensuring that only the first, unique attempt from each testee was kept. Finally, invalid test attempts (such as instances where participants did not engage meaningfully with the test or appeared to use external aid) were removed. The specific detection methodology is not described in detail in this report in order to preserve the integrity of the cheating-detection process. These steps guarantee that the final dataset contains valid, interpretable responses which measure our intended construct.

Reliability

Summary of Reliability Estimates

Internal consistency reliability was evaluated using classical test theory (CTT) and item response theory (IRT). As shown in Table 4, Cronbach’s alpha indicates excellent reliability (α = .93), agreed by an equivalent split-half estimate (λ₄ = .93). The average inter-item correlation was modest (r = .13), consistent with a broad measure of general intelligence, where diverse item content is expected and excessively high inter-item correlations would suggest a narrower, more specific ability. From an IRT perspective, conditional reliability at the general population mean (IQ = 100) was similarly high (rₓₓ = .94), indicating strong measurement precision at typical ability levels. For comparison, the Otis Gamma manual reports a lower split-half reliability of .88 (Gerardo Pérez & Vite San Pedro, 2012), suggesting that the GET’s adaptation improved internal consistency relative to the historical version.

Table 4. Summary of Reliability Estimates for the GET
Measure Estimate
Cronbach’s α 0.93
Split-half 0.93
Average inter-item correlation 0.13
IRT rₓₓ (IQ = 100) 0.94
Split-half (Otis Gamma, reported 1954) 0.88

Conditional Reliability and Test Information Curve

Figure 5 displays the conditional reliability and test information curve under the 2PL model across the IQ scale. Regardless of the high-ability nature of the sample, both curves indicate that measurement precision is highest at the average levels of ability, with both test information and conditional reliability peaking at approximately 99 IQ.

Figure 5. Conditional Reliability and Test Information Across Ability Levels

Item Response Theory

Model Comparison

Model comparisons using likelihood ratio tests indicated that the 2PL model provided a significantly better fit than the 1PL model, Δχ²(80) = 49,917.95, p < .001, supporting inclusion of the discrimination parameter.

In contrast, both the 3PL model and 4PL model failed to converge. Any estimates yielded by these models were not stable for interpretation. The 2PL model was therefore retained as the primary model.

Model fit for the 2PL model was evaluated using the M2 statistic (as can be seen in Table 5) and associated fit indices. Although the M2 test was statistically significant, M2(3080) = 191,184.90, p < .001, this was expected given the huge sample size. The practical fit indices indicated excellent model fit with RMSEA = .0283, SRMSR = .045, CFI = .948, and TLI = .947. These values imply that the 2PL model is a strong representation of the data.

Table 5. Model Fit Indices for the 2PL Model
Fit Index Value
M2 191184.9
df 3080
p < .001
RMSEA 0.0283
RMSEA 90% CI [0.0282, 0.0285]
SRMSR 0.045
TLI 0.947
CFI 0.948

Item Parameters

Appendix A presents the item parameter estimates and fit statistics for the 2PL model. Item discrimination parameters (a) varied substantially across items, ranging from very low values (e.g., Items 1-2, a ≈ 0.298, 0.331) to very high values (Items 22, a ≈ 1.802), indicating considerable variability in how well items differentiate between individuals across levels of the latent trait. A discrimination of roughly 0.5 to 2.0 is considered useful within 2PL models (Linacre, n.d.). With ~91% of items falling within this range, most items demonstrated moderate to high discrimination, suggesting that the test is generally effective at distinguishing between examinees of differing ability.

As shown in Figure 6, item difficulty parameters (b), expressed in IQ, were primarily concentrated around the average ability range, with most items covering approximately 70 and 120. Although the full range extended from very easy to very difficult items, very few items targeted the upper end (145+) of the ability continuum, justifying the current norms’ 148 IQ ceiling. Given the high ability of the online sample, this lack of highly difficult items likely contributed to the observed ceiling effects.

Figure 6. Distribution of Item Difficulties on the GET

Item-Level Fit

Item fit was evaluated using item-level χ² statistics, RMSEA, and infit/outfit stats. As expected in very large samples, almost every item exhibited statistically significant χ² values (p < .001). This value is highly sensitive to sample size, where even minor deviations from the model can result in statistical significance rather than true misfit. Accordingly, greater emphasis was placed on fit indices, such as RMSEA and infit/outfit.

Inspection of RMSEA indicated that all items demonstrated excellent model fit, with values ranging from .003 to .043. These values fall well within commonly accepted thresholds for close fit (≤ .05), suggesting that the 2PL model accurately represents the item-level data.

Infit and outfit values near 1.0 indicate good fit, with values between 0.5 and 1.5 generally considered “productive for measurement” (Linacre, 2002). However, these fit indices are intended for Rasch models and should be interpreted more cautiously for 2PL models (as discrimination can vary). For our model, infit and outfit statistics were all within the suggested range, with infit ranging from 0.919 to 1.025 and outfit ranging from 0.732 to 1.105, suggesting strong item fit.

It is also important to consider potential violations of the local independence assumption. Because items were administered sequentially within a single testing session, dependencies may arise. Missing data were treated as missing not at random (MNAR). Trailing missing responses (assumed to reflect items not reached due to time limits) were coded as “not reached” and excluded from estimation, whereas non-trailing skipped items were treated as incorrect responses. This approach reflects the testing context and mitigates bias in parameter estimation.

g-Loading

To evaluate the construct validity of the GET, a confirmatory factor analysis (CFA) was conducted jointly with CORE, an FSIQ test with 17 subtests covering six CHC domains. A latent general intelligence factor (g) was defined by the CORE subtests, and the standardized GET score was included as an additional indicator. Model fit indices indicated good overall fit (CFI = .97, TLI = .96, RMSEA = .014, SRMR = .055), despite a significant chi-square test (which is expected given the sample size).

Table 6. Confirmatory Factor Analysis Model Fit Indices
Fit Index Value
χ² (161) 610.88***
CFI 0.970
TLI 0.965
RMSEA 0.014
RMSEA (90% CI) [0.013, 0.016]
SRMR 0.055
Note. ***p < .001.

The GET loaded strongly on the general factor (λ = .811), indicating that it shares substantial variance with g as measured by an established IQ test. Loadings greater than 0.80 are considered to be strong measures of intelligence, supporting the psychometric validity of the GET and highlighting the enduring validity of the Otis Gamma many decades later.

Normative Alignment and Convergent Validity

Convergent Validity with the AGCT

To assess the GET’s norms, scores were compared with scores from the AGCT on CognitiveMetrics. The number of participants who had valid attempts on both tests is significantly smaller than the full sample, with N = 905. The AGCT was selected because it provides a comparison to a test with norms derived from a general population sample, which is especially important for anchoring given that the Otis Gamma’s contemporary norms are based on a high-ability, self-selected sample.

Convergent validity between the GET and the AGCT was evaluated using linear regression, as can be observed in Table 8. Results indicated a strong positive relationship, with GET scores significantly predicting AGCT scores (β = 0.93, p < .001). The model explained a substantial proportion of variance (R² = .55), corresponding to a correlation of approximately r ≈ .74. The regression slope was close to 1, suggesting strong agreement in scaling between the tests, with a small intercept. The standard error of estimate (≈ 8.6 IQ points) was consistent with typical differences observed between independent IQ tests. Overall, these findings provide strong evidence of convergent validity and support the comparability of GET scores with established IQ metrics.

It is also worth mentioning that the previously mentioned self-selection of interest in IQ selecting for IQ holds true within this sub-sample as well. Compared to the full sample, this sample is significantly smaller (1.19% of the original), the mean is higher (+11 points), and the standard deviation (-3.34 points) is restricted, mirroring the previously observed behavior that higher-ability individuals are more likely to have an interest in and engage further with cognitive testing.

Table 7. Descriptive Statistics for GET and AGCT Scores
Variable N Mean SD Min Max
GET 905 128.07 10.21 88 148
AGCT 905 124.88 12.84 90 151
Figure 7. Convergent Validity of GET with AGCT
Table 8. Linear Regression Results Predicting AGCT Scores from GET Scores
Statistic Value
β 0.932
Intercept (b0) 5.57
0.548
r 0.741
SEE 8.63
p-value < .001

Correcting for Range Restriction

Given the restricted ability range of the present sample, the observed correlation between the GET and AGCT (r ≈ .74) was corrected for range restriction using Thorndike’s Case II formula. Using the standard deviation of the overall sample (SD = 13.55) and the restricted subsample (SD = 10.21), the corrected correlation increased to r ≈ .79.

Notably, this correction is based on an already restricted sample (M ≈ 117.1, SD = 13.55), rather than a fully representative population. When extrapolated to a general population distribution (M = 100, SD = 15), the estimated corrected correlation rises further to r ≈ .82. This correlation is consistent with what is typically observed between well-established measures of general intelligence (typically between 0.8 and 0.9), confirming that the GET demonstrates strong convergent validity with g-loaded IQ tests.

Limitations

Several limitations should be considered when interpreting the results. Firstly, the sample is self-selected and drawn from an online population with an interest in cognitive testing. This introduces a selection bias toward individuals with above-average ability, as reflected in the higher mean IQ and reduced standard deviation of the sample. Interpretations of findings in this report should take the high-ability nature of the sample into account, especially when generalizing to the general population.

Relatedly, the restricted range of ability within the sample likely attenuated observed correlations between the GET and AGCT. Although statistical corrections for range restriction were applied, these corrections rely on assumptions (e.g., linearity and homoscedasticity) that may not be fully satisfied in practice. Corrected estimates should be taken as approximations rather than precise values.

A further limitation is regarding the administration format of the test. All items were presented sequentially within a single testing session, in alignment with how the original Otis Gamma was administered. However, this may violate the assumption of local independence underlying IRT models for some items by introducing dependencies between item responses. Nonetheless, both global and item-level model fit was strong, and parameter estimates were within expected and interpretable ranges, implying that any possible violations were likely minimal in practical impact.

Finally, the distribution of item difficulties was not uniform across ability. In relation to the sample’s ability, the upper-ceiling may be low, resulting in reduced measurement precision near the ceiling and a truncated distribution. Future revisions of the test would benefit from the addition of more difficult items to improve coverage and precision among higher-ability examinees.

Conclusion

This article provides comprehensive evidence supporting the psychometric validity of the GET as a modernized adaptation of the Otis Gamma (Form EM). Across multiple analytic approaches, the GET demonstrated excellent reliability, strong alignment with general intelligence (g), substantial convergence with established cognitive ability measures, and alignment with norms based on the general population. These findings indicate that the test is a quick way to accurately measure general intelligence for large-scale online testing.

While certain limitations were identified (most notably the self-selected, high-ability sample and reduced ceiling affecting the upper end of the ability distribution), these do not undermine the overall validity of the instrument. Further opportunities for improvement could be through the inclusion of more difficult items and further validation with more representative samples.

Overall, the GET’s strong psychometric properties, combined with its shorter duration and accessibility, support its use in research, large-scale assessment, and exploratory cognitive testing for online environments.

References

Iqbal, M. N. (2024). OTIS self-administering tests of mental ability. ResearchGate. https://www.researchgate.net/profile/Muhammad-Iqbal-657/publication/382334622_OTIS_Self-Administering_Tests_of_Mental_Ability/links/6698a4b94a172d2988aa2f6f/OTIS-Self-Administering-Tests-of-Mental-Ability.pdf

Gerardo Pérez, M. del C., & Vite San Pedro, S. (2012). Cuadernillo de prácticas 1: Pruebas de inteligencia: OTIS. Facultad de Psicología, Universidad Nacional Autónoma de México. https://www.psicologia.unam.mx/documentos/pdf/publicaciones/Cuadernillo_de_practicas_1.Pruebas_de_Inteligencia_OTIS_Gerardo_Perez_Vite_San_Pedro_TAD_6_y_7_sem.pdf

Linacre, J. M. (n.d.). Item discrimination or slope estimation. Winsteps. https://www.winsteps.com/winman/discriminationestimation.htm

Linacre, J. M. (2002). What do infit and outfit, mean-square and standardized mean? Rasch Measurement Transactions, 16(2), 878. http://www.rasch.org/rmt/rmt162f.htm

Otis, A. S. (1937). Leaflet with sheet, Otis quick-scoring mental ability tests: Manual of directions for Gamma test forms AM and BM. Smithsonian Institution. https://www.si.edu/object/leaflet-sheet-otis-quick-scoring-mental-ability-tests-manual-direction-gamma-test-forms-am-and-bm%3Anmah_692513

Appendix

Appendix A. Item Parameter Estimates and Fit Statistics
Item a b b (IQ) S-X² RMSEA Infit Outfit
1 0.298 -6.849 24 52.71 0.009 1.000 0.998
2 0.331 -4.517 56 32.00 0.006 0.999 0.994
3 1.329 -1.644 95 169.92 0.016 0.994 0.891
4 0.797 -5.027 49 24.80 0.005 1.020 0.971
5 1.032 -3.531 69 48.16 0.008 1.025 0.928
6 0.823 -2.807 79 61.08 0.009 1.005 0.959
7 0.656 -5.316 45 15.09 0.003 1.012 0.988
8 0.842 -4.566 55 28.52 0.006 1.022 0.955
9 0.905 -2.475 84 73.49 0.010 1.002 0.960
10 0.951 -3.436 71 59.48 0.009 1.016 0.972
11 0.807 -2.087 89 83.93 0.011 0.992 0.979
12 0.839 -2.634 81 62.69 0.009 1.003 0.962
13 0.783 -4.157 61 48.10 0.008 1.011 1.000
14 1.204 -1.790 93 110.14 0.013 0.993 0.927
15 0.915 -4.541 56 32.17 0.006 1.018 1.003
16 0.605 -2.862 78 102.91 0.012 0.995 0.994
17 0.747 -1.840 92 93.95 0.012 0.993 0.966
18 0.829 -3.185 74 38.69 0.007 1.005 0.979
19 1.008 -1.629 95 160.75 0.016 0.992 0.930
20 1.429 -2.626 81 101.50 0.012 1.014 0.945
21 1.382 -1.619 95 139.76 0.015 0.990 0.901
22 1.802 -3.336 72 128.27 0.014 0.919 0.732
23 0.654 -0.044 116 333.06 0.023 0.980 0.979
24 1.181 -1.086 102 323.85 0.023 0.979 0.911
25 0.762 -1.447 97 133.96 0.014 0.983 0.983
26 1.084 -1.612 95 96.28 0.012 0.988 0.937
27 0.540 -1.026 103 103.14 0.013 0.988 0.988
28 1.214 -2.020 90 85.26 0.011 0.989 0.971
29 0.592 -0.974 104 116.60 0.013 0.988 0.979
30 0.513 4.289 175 426.23 0.026 0.996 1.012
31 1.678 -1.672 94 169.86 0.016 0.970 0.997
32 0.478 -3.182 74 427.76 0.026 0.993 1.003
33 1.070 -1.425 98 113.38 0.013 0.983 0.939
34 0.705 -1.374 98 176.15 0.017 0.993 0.965
35 1.010 -2.481 83 731.62 0.034 0.978 1.105
36 0.706 -1.140 102 141.55 0.015 0.990 0.969
37 1.049 -1.733 94 282.36 0.021 0.977 0.999
38 0.782 -1.726 94 243.86 0.020 0.999 0.947
39 1.070 -1.783 93 96.26 0.012 0.985 0.964
40 0.857 -0.276 113 594.93 0.031 0.978 0.956
41 1.725 -1.689 94 122.62 0.014 0.981 0.902
42 0.816 0.859 129 610.80 0.031 0.965 0.985
43 1.535 -1.242 100 173.16 0.016 0.965 0.936
44 0.942 1.389 136 1133.14 0.043 0.947 1.033
45 0.590 0.213 120 206.45 0.018 0.983 0.982
46 0.780 -1.591 96 151.52 0.015 0.990 0.966
47 0.993 -1.657 95 87.51 0.011 0.986 0.959
48 0.984 -2.030 90 94.31 0.012 0.996 0.954
49 0.863 -1.251 100 171.19 0.016 0.979 0.975
50 0.984 -1.000 104 125.35 0.014 0.974 0.962
51 0.784 -3.954 63 338.72 0.023 1.000 1.075
52 1.133 -0.389 112 281.15 0.021 0.961 0.937
53 1.144 -1.464 97 156.62 0.016 0.977 0.970
54 1.124 -0.691 108 349.46 0.024 0.970 0.926
55 0.950 -2.330 86 205.29 0.018 0.990 1.020
56 0.864 0.363 122 294.98 0.022 0.967 0.973
57 0.823 -1.331 99 96.80 0.012 0.986 0.962
58 0.792 1.234 134 184.38 0.017 0.973 0.971
59 1.243 0.002 117 885.78 0.038 0.944 0.946
60 1.203 -1.828 92 105.16 0.013 0.989 0.970
61 0.908 -1.762 93 276.30 0.021 0.981 1.006
62 1.045 -0.019 117 217.23 0.019 0.959 0.951
63 1.361 -0.524 110 186.04 0.017 0.953 0.926
64 0.445 2.963 157 60.26 0.009 0.994 0.990
65 1.155 -0.273 113 138.14 0.015 0.959 0.938
66 1.100 -0.835 106 136.01 0.015 0.975 0.937
67 1.389 -0.351 112 513.17 0.029 0.954 0.904
68 1.087 -1.614 95 70.13 0.010 0.987 0.963
69 1.099 -0.357 112 142.58 0.015 0.963 0.949
70 0.476 0.201 120 61.77 0.009 0.989 0.988
71 0.783 -2.006 90 49.10 0.008 0.995 0.970
72 1.618 -1.496 97 97.46 0.012 0.994 0.879
73 0.302 -1.376 98 108.74 0.013 0.996 0.996
74 1.321 -1.021 103 286.32 0.021 0.985 0.884
75 1.149 -0.536 110 837.72 0.037 0.975 0.914
76 0.638 -0.018 117 61.24 0.009 0.982 0.980
77 1.708 -0.531 110 442.59 0.027 0.957 0.844
78 1.385 -1.086 102 144.05 0.015 0.984 0.888
79 1.199 0.276 121 1077.50 0.042 0.941 0.979
80 0.367 3.981 171 463.57 0.027 0.993 1.005