Have you ever experienced a situation where someone’s behaviour or actions have made you think “how can they live with themselves


Reliability, Validity and Fairness
Summary

TEA-Occ Reliability, Validity and Fairness Summary TalentLens is a trademark, in the U.S. and/or other countries, of Pearson Education, Inc. or its affiliate(s).
TEA-Occ Reliability, Validity and Fairness Summary

Reliability

The TEA-Occ standardisation for Rail Safety Standards Board (RSSB)

The TEA-Occ was completed by 134 candidates attending train driver assessment centres. Cronbach’s Alpha
was calculated to provide a measure of internal consistency for Lift Counting with Distraction, and a value of 0.84 was obtained. The TEA-Occ subtest Lift Counting with Distraction can therefore be considered to possess good internal consistency reliability. Due to the format of the Telephone Search and Telephone Search While Counting subtests, it is not possible to calculate internal consistency reliability estimates for
Test of Everyday Attention (TEA) original manual

In the original TEA manual test-retest reliability was assessed using alternate forms of the test. Three versions
of the test were developed, labelled Version A, Version B and Version C. The TEA-Occ was developed from
Table 1 shows test-retest reliability coefficients, taken from the original standardisation of the TEA. Table 1 shows the reliability coefficients for the one-week test-retest on versions A and B for 118 participants from the standardisation sample, and for 74 participants from a sample of stroke patients included in the original development and standardisation of the TEA. Results from the TEA subtests that were subsequently developed for the TEA-Occ are reported. Test-retest reliability figures are also given for a subsample of the standardisation sample who were given Version C of the test a further week after receiving Version B; these correlations are between versions B and C. The reliability of the original TEA is good for almost all subtests for both the control participants and stroke patients (see Table 1). The one exception is the dual-task decrement. The decreased reliability of this task in comparison with the others may be due to the large learning effects from one version to the other. It may be that participants differ in their ability to automate tasks, meaning that not all benefit equally from the experience of first taking the test which results in the lower test-retest reliability observed here. Table 1: Test-retest reliability from the original TEA standardisation Pearson correlations
Telephone Search While Counting – dual task decrement
TEA-Occ Reliability, Validity and Fairness Summary

It has been assumed for the TEA-Occ standardisation that, due to the high similarity between the subtest
administration, scoring and normative data of the TEA and the TEA-Occ, the test-retest properties of the TEA
Standard errors of measurement
Based on the test-retest reliability evidence available for the TEA, Table 2 shows the standard errors of
measurement (SEMs) for each of the three subtests. Adding and subtracting the SEM from an observed score gives a range that is typically referred to as the ‘confidence band’ or ‘confidence interval’. These figures are based on the test-retest between versions A and B of the TEA. The SEMs allow the accuracy of measurement, as evidenced through the test’s reliability, to be taken into account when interpreting scores. SEMs are given for both raw and T scores, at 68% and 95% levels of confidence. Table 2: Standard errors of measurement (SEM) for the TEA-Occ
Validity
Face validity

The subtests in the TEA-Occ make use of everyday materials and contexts, meaning they are realistic and
should be subject areas that respondents can readily relate to. The original authors cite this as one of the ‘strengths’ of the TEA. However, to ensure face validity it is important that test administrators clearly explain the purpose of the test, so allowing respondents to make the connection between the constructs being assessed by the test and the competencies required of a train driver.
Content validity

Development of the TEA was grounded in research on attention and established tasks which had been
empirically shown to be sensitive to individual differences in components of attention (e.g. Wilkins et al., 1987). The grounding of the subtests in applied problems that place demand on the attentional system for their successful completion, supports the content validity of the TEA-Occ.
TEA-Occ Reliability, Validity and Fairness Summary

Construct validity

This form of validity was used during the development of the TEA when examining the differential
performance between groups (e.g. stroke patients and a control group). Evidence of construct validity for the TEA-Occ comes primarily from the pilot studies where it has been used as part of a larger assessment of train drivers. Project T948 (RSSB, 2013) reported the correlation between the Lift Counting with Distraction and Telephone Search While Counting (dual task decrement) subtests to be -0.392 (n=177). This provides evidence of internal construct validity, showing that the association of the scores obtained on these two aspects of the TEA-Occ is modest and supporting the argument that these subtests are assessing distinct aspects of the attentional system. Squaring the correlation coefficient gives the degree of shared variance or ‘overlap’ between the measures. This value is just over 15 percent, meaning that performance on each of these subtests is relatively independent of performance on the other. The association between the TEA-Occ and a range of other psychometric assessments used as part of train driver selection is also reported as part of Project T948 (RSSB, 2013). These associations are summarised in Table 3. Detailed descriptions of the tests included in this analysis are given in the report for Project T948. Table 3: Associations between the TEA-Occ and other psychometric tests used as part of the train driver selection process Group Bourdon1
1The Group Bourdon is a paper-based psychometric test distributed by Southeastern. It is designed to measure attention. 2The
Determinations Gerat (DGT) test is designed to assess the operation of hand and foot controls for train driver selection. 3The
Tachistoscopic Traffic test (TAVTMB) is part of the computerised Vienna Test System battery. It assesses visual perception and
perceptive speed in traffic situations. 4The Trainability for Rules & Procedures Test (TRP) is a paper-based psychometric test
developed by OPC Assessment Ltd. It is designed to measure trainability, memory and reasoning.

TEA-Occ Reliability, Validity and Fairness Summary

The original TEA manual (Robertson et al., 1994) explored the association between the subtests included in
TEA-Occ and the National Adult Reading Test (NART), a measure of verbal intelligence, in the standardisation
sample. The results of this analysis are shown in Table 4, and indicate that verbal intelligence has very little effect on performance on any of the subtests. Table 4: Associations between TEA-Occ subtests and the NART Partial correlation with NART (age partialled out) Telephone Search While Counting (dual task decrement) The discriminant validity of the TEA was examined using two groups during its development: stroke patients and people with closed head injuries. In both cases the results from these groups were compared with controls. The results from these studies are summarised in Tables 5 and 6. T-tests (t) were used to establish the statistical significance of any differences and the probability (p) of these results are reported. Table 5: Performance on the TEA for stroke patients versus controls in two age groups Age 50-64
Age 65-80
5.65 (3.20) 3.45 <0.001 7.75 (2.90) 4.69 (3.50) 5.43 (2.60) 4.60 <0.001 4.22 (1.50) 7.38 (5.9)
TEA-Occ Reliability, Validity and Fairness Summary

Table 6: Performance on the TEA for people with closed head injuries and controls
Table 5 shows that there were clear and statistically significant differences between stroke patients and controls in almost all cases, with stroke patients showing reduced attentional functioning. A similar pattern of results was seen for the patients with head injuries, despite the limited numbers available for this study (n=15 per group).
Criterion validity
T948 (RSSB, 2013) reports concurrent criterion validity evidence for the TEA-Occ against the selection
criterion of ‘attention’. As the report describes “The selection criterion for attention is split into two sub- criteria: selective attention and divided attention. Selective attention is defined as the ability to differentiate between different sources of information and attend selectively to them. Divided attention is defined as the ability to switch attention between different sources of information” (RSSB, 2013, p 170-171). The results of the criterion validity study conducted as part of project T948 are shown in Table 7. Table 7: Criterion validity of the TEA-Occ against train driver selection criterion of attention Criterion: Selective attention
Criterion: Divided attention

TEA-Occ Reliability, Validity and Fairness Summary

As part of establishing the validity of the new set of assessments proposed for train driver selection, a range of
performance data was collected on participants. Ratings were made by driver managers on a number of
indicators of operational driving performance, using a scale ranging from 1 (very poor) to 5 (excellent). The key criteria in the validation of the TEA-Occ were ‘Train driving procedure-based work’ and ‘Preparation and disposal of trains’, selected on the basis of being the two most relevant for the TEA-Occ measures and the ones that showed associations in the predicted direction. The association of the TEA-Occ with these criteria is shown in Table 8, indicating the suitability of the TEA-Occ for assessment of these criteria. Table 8: Criterion validity of the TEA-Occ against ‘Train driving procedure-based work’ and ‘Preparation and disposal of TEA-Occ criterion validity
Operational performance data
Lift Counting with Distraction – total Telephone Search While Counting - Dual task decrement *p<0.5, **p<0.001 Copyright Pearson Education Ltd
TEA-Occ Reliability, Validity and Fairness Summary

Fairness

Information relevant to the fairness of the TEA-Occ is available from the original TEA manual, and the RSSB
T628 and T948 projects. T948 also examined the effect of pass rates according to group membership, if the recommended pass scores were applied. This data was evaluated against the ‘four-fifths rule’. The four-fifths rule compares the proportion of the majority group that is successful at any stage of a selection procedure to the proportion of the minority group that is successful. If the proportion of the minority group that is successful is less than four-fifths of the majority group, then adverse impact is said to be occurring. The four- fifths rule is not used in UK law, but is rather a guide to evaluating fairness.
Age: The original TEA manual (Robertson et al., 1994) reported that older people tended to do better on the
Telephone Search While Counting subtest. Decrement in performance was seen in people aged over 50, and particularly in those aged over 65. In the project T628 trials of the TEA-Occ (RSSB, 2010) it was found “there are no obvious differences on any test in the 21 - 50 age range but over 50s do less well on all the sub-tests” (p 35). However, project T948 concluded that the evidence for age effects on Lift Counting with Distraction was inconclusive (RSSB, 2013). When the pass scores were applied to the T948 project sample, 96 percent of those aged 50 and under were seen to pass Lift Counting with Distraction compared to only 76 percent of those aged 51 or older. The pass rates for Telephone Search While Counting (dual task decrement) were 92 percent for those aged 50 and under and 81 percent for those aged 51 or older. RSSB (2013) reports that if these figures remain stable, it would mean that Lift Counting with Distraction would fail to meet the four-fifths rule test for potential adverse impact, though Telephone Search While Counting would comply with this guide.
Gender: None of the research on the TEA or the TEA-Occ has found evidence of gender difference on any
subtest. When the pass scores were applied to the T948 project sample, 93 percent of males and 100 percent of females passed Lift Counting with Distraction, with the pass rates for Telephone Search While Counting (dual task decrement) being 91 percent and 100 percent. RSSB (2013) reports that these figures mean both subtests comply with the four-fifths rule. However, it should be noted that the number of females in this sample was very small (n=5), which means these results should be treated with caution.
Ethnic group: Although project T628 (RSSB, 2010) reported no evidence of ethnic group effects on subtest
performance, the application of the recommended pass scores was seen to result in differences in project T948 (RSSB, 2013). The pass rates for Lift Counting with Distraction were 94 percent for ‘Whites’ and 67 percent for ‘Others’ (‘Non-whites’). For Telephone Search While Counting (dual task decrement), the pass rates were 92 percent (‘Whites’) and 67 percent (‘Others’). These figures showed that both subtests failed to meet the four-fifths rule, suggesting that adverse impact may be occurring. However, interpretation of these results is hampered by the small number of people from ‘non-white’ backgrounds (n=6) and treating all ‘non- whites’ as a homogenous group for the purposes of analysis.
TEA-Occ Reliability, Validity and Fairness Summary

Overall, current evidence suggests that the TEA-Occ subtests do not have adverse effects according to group
membership. However, as the project T948 report notes “Every effort was made to collect information from females, older candidates and ethnic groups. However, people from these demographic groups are so poorly represented in the train driver population that it was only possible to obtain a very small sample” (RSSB, 2013, p213). Due to the small sample sizes, especially for females and non-whites, analysis of fairness needs to continue when larger sample sizes are available following implementation of the revised psychometric
References
Robertson, I. H., Ward, T., Ridgeway, V. and Nimmo-Smith, I. (1994). The Test of Everday Attention. London:
Pearson.

RSSB (2010). T628 Driver Selection: Development Phase – Updated selection criteria and validation study. London:
RSSB.
RSSB (2013). T948 Driver selection: Implementation phase – Final report and technical annexes. London: RSSB.




TEA-Occ Reliability, Validity and Fairness Summary

Copyright Pearson Education Ltd Pearson TalentLens January 2014 Contact Us
Customer Services
For all registration and order enquiries please contact our customer Services team using one of the
methods below.
Tel: 0845 630 88 88 (Monday to Friday, 8am to 5pm)
Fax: 0845 630 55 55
Post : Pearson Clinical and Talent Assessment
Oxford, OX2 8EJ
If you have a technical query about our products or services offered please email
UK Head Office
www.talentlens.co.uk

Source: http://www.talentlens.co.uk/assets/resources/TEA-Occ-Summary.pdf

e-pharm.kemkes.go.id

DIREKTORAT JENDERAL PELAYANAN KEFARMASIAN DAN ALAT KESEHATAN UNDANG – UNDANG REPUBLIK INDONESIA NOMOR 5 TAHUN 1997 P S I K O T R O P I K A DENGAN RAHMAT TUHAN YANG MAHA ESA PRESIDEN REPUBLIK INDONESIA, Menimbang : a. bahwa pembangunan nasional bertujuan untuk mewujudkan suatu masyarakat adil dan makmur yang merata materil dan spiritual berdasarkan Pancasila dan Undang

Newsletter 1-2006

SPAN – Società degli Psicologi dell’Area Neuropsicologica a cura della Commissione Editoriale N° 1/febbraio 2006 INFO DALLA SOCIETÀ’ Sono aperte le iscrizioni alla SPAN per il 2006. Per nuove adesioni o per ulteriori informazioni circa le modalità di rinnovo è possibile contattare il nostro segretario, Matteo Sozzi, all’indirizzo mail [email protected] È stata or

Copyright © 2013-2018 Pharmacy Abstracts