Item Critique
Reliability is a primary concept in any field of study. It helps ascertain the level of accuracy of results thus confirming scenarios in an empirically sound manner. Reliability, therefore, helps researchers to develop measures, which can best approximate the findings of a given scenario.
Items developed
Test-retest method
The test-retest method is one of the most elementary procedures for assessing reliability of data. It involves three major steps, firstly, the administration of a test such as IQ to a sample of people, administering the test again to the same sample group of people after some time. The final step is the correlation of the scores to analyze the extent of similarity. Higher similarity depicts reliability of the results. However, inconsistent results especially on IQ tests can be because of internal faultiness in the instruments of measurement. The effectiveness of this method mainly depends on the period between the first and the subsequent test (Nicholas 170).
The first step in parallel forms method is to administer the research form (e.g. Form 1) to a group of people. Subsequently one administers an equivalent or identical form (e.g. Form 2) of the test, comprising of related but different items to the same group. The final step is correlating the results to assess the similarity of the scores. Similar results depict reliability. However, the inability of accurately developing a form that is equivalent to the first can render the results of this method tenuous (Nicholas 171).
Inter-rater reliability approach
This process involves consensus between two or more observers or judges about the level of absence or presence of a particular trait. Moreover, it may involve observation of diagnosis of a particular condition. Inter-rater reliability is majorly used to examine the usefulness of a given diagnostic interview. This process is valuable and applicable in mental health studies (Kearney and Trull 79).
The Construct to measure
This study seeks to develop a test, which can validly measure IQ. According to Anastasi, validation begins by formulating construct definitions drawn from prior research, systematic analysis and observation of relevant behavior domain, and psychological theory. Preparation of test items should aim at fitting the construct definitions. Subsequently, empirical analyses follow coupled with the selection of the most useful items identified in the initial item pools (3).
Feedback to colleagues
Colleague 1: The split-half reliability method
The potential areas of revision
It is critical for the colleague to ensure consistency and reliability in the approach of selecting the groups. For instance, Barrick noted that a study on the corresponding personality and occupation showed that conscientiousness provided relations, which were consistent with all job performance approaches for all the professional groups. However, the estimated score for the other personality dimensions varied based on the type of approach and professional group. The study used a concept called Big Five dimensions of personality, namely, Emotional stability, conscientiousness, agreeableness, extraversion, and openness to experience(1)
Colleague 2: Internal consistency reliability
The potential areas of revision
It is important for the colleague using this method to note that this method uses indirect methods of computing consistency. Hence, it is important that he mitigate that using more direct methods such as parallel forms.
Works Cited
Anastasi, A. (1986) Evolving concepts of test validation. Annual Review of Psychology, 37(1), 1–15.Barrick, M. R., & Mount, M. K. (1991). The big five personality dimensions and job performance: A meta-analysis. Personnel Psychology, 44(1), 1–26.
Kearney, Chris, and Trull Timothy Abnormal psychology and Life: A dimensional Approach. Boston: Cengage Learning, 2014.Print
Nicholas, Lionel (Ed). Introduction to Psychology. Illinois: Juta and Company Ltd, 2009. Print.