SlideShare a Scribd company logo
1 of 30
Analysis of Test
Results
    Reliability, Validity, and Item
    Analysis
Learning Content
   Levels of Measurement
   Correlation Coefficient
   Reliability
   Validity
   Item Analysis
Objectives
   1. Determine the use of the different
    ways of establishing an assessment
    tools’ validity and reliability.
   2. Familiarize on the different methods
    of establishing an assessment tools’
    validity and reliability.
   3. Assess how good an assessment
    tool is by determining the index of
    validity, reliability, item
    discrimination, and item difficulty.
Levels of Measurement
   Nominal
   Ordinal
   Interval
   Ratio
Correlation Coefficient
   Relationship of two variables (X &
    Y)
   Direction
   Positive             Negative


X


         Y
Degree of Relationship
   0.80 – 1.00   Very High
                  relationship
   0.6 – 0.79    High Relationship
   0.40 – 0.59   Substantial/Marked
                  relationship
   0.20 – 0.39   Low relationship
   0.00 – 0.19   Negligible
                  relationship
Testing for Significance
   Nominal: Phi Coefficient
   Ordinal: Spearman rho
   Interval & Ratio: Pearson r
   Interval with nominal: Point biserial
   Decision rule:
   If p value < =.05: significant
    relationship
   If p value > =.05: no significant
    relationship
Variance
   R2
   Square the correlation coefficient
   Interpretation: percentage of time
    that the variability in X accounts for
    the variability in Y.
Reliability
   Consistency of scores Obtained by
    the same person when retested
    with the identical test or with an
    equivalent form of the test
Test-Retest Reliability
   Repeating the identical test on a second
    occasion
   Temporal stability
   When variables are stable ex: motor
    coordination, finger dexterity, aptitude,
    capacity to learn
   Correlate the scores from the first test
    and second test.· The higher the
    correlation the more reliable
Alternate Form/Parallel
Form
   Same person is tested with one form on
    the first occasion and with another
    equivalent form on the second
   Equivalence;
   Temporal stability and consistency of
    response
   Used for personality and mental ability
    tests
   Correlate scores on the first form and
    scores on the second form
Split half
   Two scores are obtained for each person
    by dividing the test into equivalent halves
   Internal consistency;
   Homogeneity of items
   Used for personality and mental ability
    tests
   The test should have many items
   Correlate scores of the odd and even
    numbered items
   Convert     the    obtained      correlation
    coefficient into a coefficient estimate
    using Spearman Brown

Kuder Richardson
(KR #20/KR #21)
   When       computing     for binary
    (e.g., true/false) items
   Consistency of responses to all
    items
   Used if there is a correct answer
    (right or wrong)
   Use KR #20 or KR #21 formula
Coefficient Alpha
   The reliability that would result if all
    values for each item were
    standardized (z transformed)
   Consistency of responses to all
    items
   Homogeneity of items
   Used for personality tests with
    multiple scored-items
   Use the cronbach’s alpha formula
Inter-item reliability
   Consistency of responses to all
    items
   Homogeneity of items
   Used for personality tests with
    multiple scored-items
   Each item is correlated with every
    item in the test
Scorer Reliability
   Having a sample of test papers
    independently scored by two examiners
   To decrease examiner or scorer variance
   Clinical   instruments    employed      in
    intensive individual tests ex. projective
    tests
   The two scores from the two raters
    obtained are correlated with each other
Validity
   Degree to which the test actually
    measures what it purports to
    measure
Content Validity
   Systematic examination of the test
    content to determine whether it covers a
    representative sample of the behavior
    domain to be measured.
   More appropriate for achievement tests
    & teacher made tests
   Items are based on instructional
    objectives, course syllabi & textbooks
   Consultation with experts
   Making test-specifications
Criterion-Prediction
Validity
   Prediction from the test to any
    criterion situation over time interval
   Hiring job applicants, selecting
    students      for    admission       to
    college,       assigning       military
    personnel to occupational training
    programs
   Test scores are correlated with
    other criterion measures ex:
    mechanical aptitude and job
    performance as a machinist
Concurrent validity
   Tests are administered to a group
    on whom criterion data are already
    available
   Diagnosing for existing status ex.
    entrance exam scores of students
    for college with their average grade
    for their senior year.
   Correlate the test score with the
    other existing measure
Construct Validity
   The extent to which the test may be said
    to measure a theoretical construct or
    trait.
   Used for personality tests. Measures that
    are multidimensional
           Correlate a new test with a similar
    earlier test as measured approximately
    the same general behavior
          Factor analysis
           Comparison of the upper and lower
    group
           Point-biserial correlation (pass and
    fail with total test score)
          Correlate subtest with the entire test
Convergent Validity

   The      test     should    correlate
    significantly from variables it is
    related to
   Commonly          for    personality
    measures
   Multitrait-multidimensional matrix
Divergent Validity
   The test should not correlate
    significantly from variables from
    which it should differ
   Commonly          for    personality
    measures
   Multitrait-multidimensional matrix
Item Analysis
   Item Difficulty – The percentage of
    respondents who answered an
    item correctly
   Item Discrimination – Degree to
    which an item differentiates
    correctly among test takers in the
    behavior that the test is designed
    to measure.
Difficulty Index
   Difficulty Index   Remark
   .76 or higher      Easy Item
   .25 to .75         Average Item
   .24 or lower       Difficult Item
Index Discrimination
   .40 and above - Very good item
   .30 - .39     - Good item
   .20 - .29     - Reasonably Good
                  item
   .10 - .19     - Marginal item
   Below .10     - Poor item

More Related Content

What's hot

Roles of Assessment in Classroom Instruction
Roles of Assessment in Classroom InstructionRoles of Assessment in Classroom Instruction
Roles of Assessment in Classroom InstructionJames Robert Villacorteza
 
Item Analysis - Discrimination and Difficulty Index
Item Analysis - Discrimination and Difficulty IndexItem Analysis - Discrimination and Difficulty Index
Item Analysis - Discrimination and Difficulty IndexMr. Ronald Quileste, PhD
 
Purposes of assessment
Purposes of assessmentPurposes of assessment
Purposes of assessmentsethezra
 
Assessment of Learning
Assessment of LearningAssessment of Learning
Assessment of LearningYaKu Loveria
 
CONSTRUCTING PAPER-AND-PENCIL TESTS
CONSTRUCTING PAPER-AND-PENCIL TESTSCONSTRUCTING PAPER-AND-PENCIL TESTS
CONSTRUCTING PAPER-AND-PENCIL TESTSJhenq Campo
 
Keys to Effective Testing
Keys to Effective TestingKeys to Effective Testing
Keys to Effective TestingVanz Justine
 
Reliability (assessment of student learning I)
Reliability (assessment of student learning I)Reliability (assessment of student learning I)
Reliability (assessment of student learning I)Rey-ra Mora
 
Assessment of Learning - Multiple Choice Test
Assessment of Learning - Multiple Choice TestAssessment of Learning - Multiple Choice Test
Assessment of Learning - Multiple Choice TestXiTian Miran
 
Constructing Objective Supply Type of Items
Constructing Objective Supply Type of ItemsConstructing Objective Supply Type of Items
Constructing Objective Supply Type of ItemsEzr Acelar
 
Guidelines for Constructing Effective Test Items
Guidelines for Constructing Effective Test ItemsGuidelines for Constructing Effective Test Items
Guidelines for Constructing Effective Test ItemsKimverly Torres
 
Development of classroom assessment tools
Development of classroom assessment toolsDevelopment of classroom assessment tools
Development of classroom assessment toolsEaicz12
 
12 Principles of High Quality Assessments (RE-UPLOADED)
12 Principles of High Quality Assessments (RE-UPLOADED)12 Principles of High Quality Assessments (RE-UPLOADED)
12 Principles of High Quality Assessments (RE-UPLOADED)Mr. Ronald Quileste, PhD
 

What's hot (20)

Roles of Assessment in Classroom Instruction
Roles of Assessment in Classroom InstructionRoles of Assessment in Classroom Instruction
Roles of Assessment in Classroom Instruction
 
Item Analysis - Discrimination and Difficulty Index
Item Analysis - Discrimination and Difficulty IndexItem Analysis - Discrimination and Difficulty Index
Item Analysis - Discrimination and Difficulty Index
 
Purposes of assessment
Purposes of assessmentPurposes of assessment
Purposes of assessment
 
Assessment of Learning
Assessment of LearningAssessment of Learning
Assessment of Learning
 
Improving the test items
Improving the test itemsImproving the test items
Improving the test items
 
Analyzing and using test item data
Analyzing and using test item dataAnalyzing and using test item data
Analyzing and using test item data
 
Test construction
Test constructionTest construction
Test construction
 
CONSTRUCTING PAPER-AND-PENCIL TESTS
CONSTRUCTING PAPER-AND-PENCIL TESTSCONSTRUCTING PAPER-AND-PENCIL TESTS
CONSTRUCTING PAPER-AND-PENCIL TESTS
 
Keys to Effective Testing
Keys to Effective TestingKeys to Effective Testing
Keys to Effective Testing
 
Reliability (assessment of student learning I)
Reliability (assessment of student learning I)Reliability (assessment of student learning I)
Reliability (assessment of student learning I)
 
Assessment of Learning - Multiple Choice Test
Assessment of Learning - Multiple Choice TestAssessment of Learning - Multiple Choice Test
Assessment of Learning - Multiple Choice Test
 
Improving test items
Improving test itemsImproving test items
Improving test items
 
Constructing Objective Supply Type of Items
Constructing Objective Supply Type of ItemsConstructing Objective Supply Type of Items
Constructing Objective Supply Type of Items
 
Item Analysis
Item AnalysisItem Analysis
Item Analysis
 
Teacher made tests
Teacher made testsTeacher made tests
Teacher made tests
 
Assessment of learning 1
Assessment of learning 1Assessment of learning 1
Assessment of learning 1
 
Guidelines for Constructing Effective Test Items
Guidelines for Constructing Effective Test ItemsGuidelines for Constructing Effective Test Items
Guidelines for Constructing Effective Test Items
 
Development of classroom assessment tools
Development of classroom assessment toolsDevelopment of classroom assessment tools
Development of classroom assessment tools
 
12 Principles of High Quality Assessments (RE-UPLOADED)
12 Principles of High Quality Assessments (RE-UPLOADED)12 Principles of High Quality Assessments (RE-UPLOADED)
12 Principles of High Quality Assessments (RE-UPLOADED)
 
Types of Test
Types of Test Types of Test
Types of Test
 

Viewers also liked

Item analysis and validation
Item analysis and validationItem analysis and validation
Item analysis and validationKEnkenken Tan
 
Insights and development of tests
Insights and development of testsInsights and development of tests
Insights and development of testsCarlo Magno
 
Guiding your child on their career decision making
Guiding your child on their career decision makingGuiding your child on their career decision making
Guiding your child on their career decision makingCarlo Magno
 
Analysis : Persuasive Discourse
Analysis : Persuasive DiscourseAnalysis : Persuasive Discourse
Analysis : Persuasive DiscourseAlwyn Lau
 
Assessment competencies literacies
Assessment competencies literaciesAssessment competencies literacies
Assessment competencies literaciesCarlo Magno
 
The practice of psychological assessment
The practice of psychological assessmentThe practice of psychological assessment
The practice of psychological assessmentCarlo Magno
 
11 adaptive testing-irt
11 adaptive testing-irt11 adaptive testing-irt
11 adaptive testing-irt宥均 林
 
Lesson 4: Evaluating Sources
Lesson 4: Evaluating SourcesLesson 4: Evaluating Sources
Lesson 4: Evaluating SourcesLeslie Lewis
 
Test interpretation and report writing
Test interpretation and report writingTest interpretation and report writing
Test interpretation and report writingCarlo Magno
 
Administering, analyzing, and improving the test or assessment
Administering, analyzing, and improving the test or assessmentAdministering, analyzing, and improving the test or assessment
Administering, analyzing, and improving the test or assessmentNema Grace Medillo
 
Teaching self-regulation
Teaching self-regulationTeaching self-regulation
Teaching self-regulationCarlo Magno
 
Chapter 6: Writing Objective Test Items
Chapter 6: Writing Objective Test ItemsChapter 6: Writing Objective Test Items
Chapter 6: Writing Objective Test ItemsSHELAMIE SANTILLAN
 

Viewers also liked (20)

Item analysis with spss software
Item analysis with spss softwareItem analysis with spss software
Item analysis with spss software
 
Chapter 6: Validity
Chapter 6: ValidityChapter 6: Validity
Chapter 6: Validity
 
Test validity
Test validityTest validity
Test validity
 
T est item analysis
T est item analysisT est item analysis
T est item analysis
 
Item analysis and validation
Item analysis and validationItem analysis and validation
Item analysis and validation
 
Item analysis
Item analysis Item analysis
Item analysis
 
Argumentative Discourse
Argumentative DiscourseArgumentative Discourse
Argumentative Discourse
 
Insights and development of tests
Insights and development of testsInsights and development of tests
Insights and development of tests
 
Guiding your child on their career decision making
Guiding your child on their career decision makingGuiding your child on their career decision making
Guiding your child on their career decision making
 
Analysis : Persuasive Discourse
Analysis : Persuasive DiscourseAnalysis : Persuasive Discourse
Analysis : Persuasive Discourse
 
Assessment competencies literacies
Assessment competencies literaciesAssessment competencies literacies
Assessment competencies literacies
 
Subjective & Objective Writing Skills
Subjective & Objective Writing SkillsSubjective & Objective Writing Skills
Subjective & Objective Writing Skills
 
The practice of psychological assessment
The practice of psychological assessmentThe practice of psychological assessment
The practice of psychological assessment
 
11 adaptive testing-irt
11 adaptive testing-irt11 adaptive testing-irt
11 adaptive testing-irt
 
Lesson 4: Evaluating Sources
Lesson 4: Evaluating SourcesLesson 4: Evaluating Sources
Lesson 4: Evaluating Sources
 
Item Analysis and Validation
Item Analysis and ValidationItem Analysis and Validation
Item Analysis and Validation
 
Test interpretation and report writing
Test interpretation and report writingTest interpretation and report writing
Test interpretation and report writing
 
Administering, analyzing, and improving the test or assessment
Administering, analyzing, and improving the test or assessmentAdministering, analyzing, and improving the test or assessment
Administering, analyzing, and improving the test or assessment
 
Teaching self-regulation
Teaching self-regulationTeaching self-regulation
Teaching self-regulation
 
Chapter 6: Writing Objective Test Items
Chapter 6: Writing Objective Test ItemsChapter 6: Writing Objective Test Items
Chapter 6: Writing Objective Test Items
 

Similar to Analyze Test Results Reliability Validity Item Analysis

Louzel Report - Reliability & validity
Louzel Report - Reliability & validity Louzel Report - Reliability & validity
Louzel Report - Reliability & validity Louzel Linejan
 
RELIABILITY AND VALIDITY
RELIABILITY AND VALIDITYRELIABILITY AND VALIDITY
RELIABILITY AND VALIDITYJoydeep Singh
 
Language testing and evaluation validity and reliability.
Language testing and evaluation validity and reliability.Language testing and evaluation validity and reliability.
Language testing and evaluation validity and reliability.Vadher Ankita
 
With-Hershey-Marie-Abarri-4hshzjzhzhzhzhzhz.pptx
With-Hershey-Marie-Abarri-4hshzjzhzhzhzhzhz.pptxWith-Hershey-Marie-Abarri-4hshzjzhzhzhzhzhz.pptx
With-Hershey-Marie-Abarri-4hshzjzhzhzhzhzhz.pptxJunrivRivera
 
Establishing Validity-and-Reliability-Test ppt.pptx
Establishing Validity-and-Reliability-Test ppt.pptxEstablishing Validity-and-Reliability-Test ppt.pptx
Establishing Validity-and-Reliability-Test ppt.pptxRayLorenzOrtega
 
Test standardization
Test standardizationTest standardization
Test standardizationKaye Batica
 
Reliability by Vartika Verma .pdf
Reliability by Vartika Verma .pdfReliability by Vartika Verma .pdf
Reliability by Vartika Verma .pdfVartika Verma
 
Validity & reliability seminar
Validity & reliability seminarValidity & reliability seminar
Validity & reliability seminarmrikara185
 
Testing in language programs (chapter 8)
Testing in language programs (chapter 8)Testing in language programs (chapter 8)
Testing in language programs (chapter 8)Tahere Bakhshi
 
Validity, Reliability ,Objective & Their Types
Validity, Reliability ,Objective & Their TypesValidity, Reliability ,Objective & Their Types
Validity, Reliability ,Objective & Their TypesMohammadRabbani18
 
Validity, reliability & practicality
Validity, reliability & practicalityValidity, reliability & practicality
Validity, reliability & practicalitySamcruz5
 
Characteristics of effective tests and hiring
Characteristics of effective tests and hiringCharacteristics of effective tests and hiring
Characteristics of effective tests and hiringBinibining Kalawakan
 
PHYSICS EDUCATION PRINCIPLE & EVALUATION TECHNIQUES (LARAS & NUR ASIAH)
PHYSICS EDUCATION PRINCIPLE & EVALUATION TECHNIQUES (LARAS & NUR ASIAH)PHYSICS EDUCATION PRINCIPLE & EVALUATION TECHNIQUES (LARAS & NUR ASIAH)
PHYSICS EDUCATION PRINCIPLE & EVALUATION TECHNIQUES (LARAS & NUR ASIAH)vina serevina
 
Test characteristics
Test characteristicsTest characteristics
Test characteristicsSamcruz5
 
Properties of-assessment-methods
Properties of-assessment-methodsProperties of-assessment-methods
Properties of-assessment-methodsCatherine Matias
 

Similar to Analyze Test Results Reliability Validity Item Analysis (20)

Louzel Report - Reliability & validity
Louzel Report - Reliability & validity Louzel Report - Reliability & validity
Louzel Report - Reliability & validity
 
RELIABILITY AND VALIDITY
RELIABILITY AND VALIDITYRELIABILITY AND VALIDITY
RELIABILITY AND VALIDITY
 
Language testing and evaluation validity and reliability.
Language testing and evaluation validity and reliability.Language testing and evaluation validity and reliability.
Language testing and evaluation validity and reliability.
 
Qualities of good evaluation tool (1)
Qualities of good evaluation  tool (1)Qualities of good evaluation  tool (1)
Qualities of good evaluation tool (1)
 
With-Hershey-Marie-Abarri-4hshzjzhzhzhzhzhz.pptx
With-Hershey-Marie-Abarri-4hshzjzhzhzhzhzhz.pptxWith-Hershey-Marie-Abarri-4hshzjzhzhzhzhzhz.pptx
With-Hershey-Marie-Abarri-4hshzjzhzhzhzhzhz.pptx
 
Establishing Validity-and-Reliability-Test ppt.pptx
Establishing Validity-and-Reliability-Test ppt.pptxEstablishing Validity-and-Reliability-Test ppt.pptx
Establishing Validity-and-Reliability-Test ppt.pptx
 
Test standardization
Test standardizationTest standardization
Test standardization
 
Reliability by Vartika Verma .pdf
Reliability by Vartika Verma .pdfReliability by Vartika Verma .pdf
Reliability by Vartika Verma .pdf
 
Validity & reliability seminar
Validity & reliability seminarValidity & reliability seminar
Validity & reliability seminar
 
Reliability and validity
Reliability and validityReliability and validity
Reliability and validity
 
Testing in language programs (chapter 8)
Testing in language programs (chapter 8)Testing in language programs (chapter 8)
Testing in language programs (chapter 8)
 
Validity and reliability
Validity and reliabilityValidity and reliability
Validity and reliability
 
Validity, Reliability ,Objective & Their Types
Validity, Reliability ,Objective & Their TypesValidity, Reliability ,Objective & Their Types
Validity, Reliability ,Objective & Their Types
 
Validity, reliability & practicality
Validity, reliability & practicalityValidity, reliability & practicality
Validity, reliability & practicality
 
Characteristics of effective tests and hiring
Characteristics of effective tests and hiringCharacteristics of effective tests and hiring
Characteristics of effective tests and hiring
 
PHYSICS EDUCATION PRINCIPLE & EVALUATION TECHNIQUES (LARAS & NUR ASIAH)
PHYSICS EDUCATION PRINCIPLE & EVALUATION TECHNIQUES (LARAS & NUR ASIAH)PHYSICS EDUCATION PRINCIPLE & EVALUATION TECHNIQUES (LARAS & NUR ASIAH)
PHYSICS EDUCATION PRINCIPLE & EVALUATION TECHNIQUES (LARAS & NUR ASIAH)
 
Quantitative Analysis
Quantitative AnalysisQuantitative Analysis
Quantitative Analysis
 
Test characteristics
Test characteristicsTest characteristics
Test characteristics
 
Properties of-assessment-methods
Properties of-assessment-methodsProperties of-assessment-methods
Properties of-assessment-methods
 
EM&E.pptx
EM&E.pptxEM&E.pptx
EM&E.pptx
 

More from Carlo Magno

Assessment Using the SOLO Framework.pptx
Assessment Using the SOLO Framework.pptxAssessment Using the SOLO Framework.pptx
Assessment Using the SOLO Framework.pptxCarlo Magno
 
Social and Emotional Learning
Social and Emotional LearningSocial and Emotional Learning
Social and Emotional LearningCarlo Magno
 
Educational assessment in the 4 ir
Educational assessment in the 4 irEducational assessment in the 4 ir
Educational assessment in the 4 irCarlo Magno
 
The process of research mentoring
The process of research mentoringThe process of research mentoring
The process of research mentoringCarlo Magno
 
Quality management services sustainability training
Quality management services sustainability trainingQuality management services sustainability training
Quality management services sustainability trainingCarlo Magno
 
Managing technology integration in schools
Managing technology integration in schoolsManaging technology integration in schools
Managing technology integration in schoolsCarlo Magno
 
Integrating technology in teaching
Integrating technology in teachingIntegrating technology in teaching
Integrating technology in teachingCarlo Magno
 
Empowering educators on technology integration
Empowering educators on technology integrationEmpowering educators on technology integration
Empowering educators on technology integrationCarlo Magno
 
Designing an online lesson
Designing an online lessonDesigning an online lesson
Designing an online lessonCarlo Magno
 
Curriculum integration
Curriculum integrationCurriculum integration
Curriculum integrationCarlo Magno
 
Accountability in Developing Student Learning
Accountability in Developing Student LearningAccountability in Developing Student Learning
Accountability in Developing Student LearningCarlo Magno
 
The Instructional leader: TOwards School Improvement
The Instructional leader: TOwards School ImprovementThe Instructional leader: TOwards School Improvement
The Instructional leader: TOwards School ImprovementCarlo Magno
 
Assessing Science Inquiry Skills
Assessing Science Inquiry SkillsAssessing Science Inquiry Skills
Assessing Science Inquiry SkillsCarlo Magno
 
Assessment in the Social Studies Curriculum
Assessment in the Social Studies CurriculumAssessment in the Social Studies Curriculum
Assessment in the Social Studies CurriculumCarlo Magno
 
Quantitative analysis in language research
Quantitative analysis in language researchQuantitative analysis in language research
Quantitative analysis in language researchCarlo Magno
 
Integrating technology in teaching
Integrating technology in teachingIntegrating technology in teaching
Integrating technology in teachingCarlo Magno
 
Hallmarks of textbook
Hallmarks of textbookHallmarks of textbook
Hallmarks of textbookCarlo Magno
 
managing the learner centered-classroom
managing the learner centered-classroommanaging the learner centered-classroom
managing the learner centered-classroomCarlo Magno
 
Assessing learning objectives
Assessing learning objectivesAssessing learning objectives
Assessing learning objectivesCarlo Magno
 
Accountability in developing assessment tools
Accountability in developing assessment toolsAccountability in developing assessment tools
Accountability in developing assessment toolsCarlo Magno
 

More from Carlo Magno (20)

Assessment Using the SOLO Framework.pptx
Assessment Using the SOLO Framework.pptxAssessment Using the SOLO Framework.pptx
Assessment Using the SOLO Framework.pptx
 
Social and Emotional Learning
Social and Emotional LearningSocial and Emotional Learning
Social and Emotional Learning
 
Educational assessment in the 4 ir
Educational assessment in the 4 irEducational assessment in the 4 ir
Educational assessment in the 4 ir
 
The process of research mentoring
The process of research mentoringThe process of research mentoring
The process of research mentoring
 
Quality management services sustainability training
Quality management services sustainability trainingQuality management services sustainability training
Quality management services sustainability training
 
Managing technology integration in schools
Managing technology integration in schoolsManaging technology integration in schools
Managing technology integration in schools
 
Integrating technology in teaching
Integrating technology in teachingIntegrating technology in teaching
Integrating technology in teaching
 
Empowering educators on technology integration
Empowering educators on technology integrationEmpowering educators on technology integration
Empowering educators on technology integration
 
Designing an online lesson
Designing an online lessonDesigning an online lesson
Designing an online lesson
 
Curriculum integration
Curriculum integrationCurriculum integration
Curriculum integration
 
Accountability in Developing Student Learning
Accountability in Developing Student LearningAccountability in Developing Student Learning
Accountability in Developing Student Learning
 
The Instructional leader: TOwards School Improvement
The Instructional leader: TOwards School ImprovementThe Instructional leader: TOwards School Improvement
The Instructional leader: TOwards School Improvement
 
Assessing Science Inquiry Skills
Assessing Science Inquiry SkillsAssessing Science Inquiry Skills
Assessing Science Inquiry Skills
 
Assessment in the Social Studies Curriculum
Assessment in the Social Studies CurriculumAssessment in the Social Studies Curriculum
Assessment in the Social Studies Curriculum
 
Quantitative analysis in language research
Quantitative analysis in language researchQuantitative analysis in language research
Quantitative analysis in language research
 
Integrating technology in teaching
Integrating technology in teachingIntegrating technology in teaching
Integrating technology in teaching
 
Hallmarks of textbook
Hallmarks of textbookHallmarks of textbook
Hallmarks of textbook
 
managing the learner centered-classroom
managing the learner centered-classroommanaging the learner centered-classroom
managing the learner centered-classroom
 
Assessing learning objectives
Assessing learning objectivesAssessing learning objectives
Assessing learning objectives
 
Accountability in developing assessment tools
Accountability in developing assessment toolsAccountability in developing assessment tools
Accountability in developing assessment tools
 

Analyze Test Results Reliability Validity Item Analysis

  • 1. Analysis of Test Results Reliability, Validity, and Item Analysis
  • 2. Learning Content  Levels of Measurement  Correlation Coefficient  Reliability  Validity  Item Analysis
  • 3. Objectives  1. Determine the use of the different ways of establishing an assessment tools’ validity and reliability.  2. Familiarize on the different methods of establishing an assessment tools’ validity and reliability.  3. Assess how good an assessment tool is by determining the index of validity, reliability, item discrimination, and item difficulty.
  • 4. Levels of Measurement  Nominal  Ordinal  Interval  Ratio
  • 5. Correlation Coefficient  Relationship of two variables (X & Y)  Direction  Positive Negative X Y
  • 6. Degree of Relationship  0.80 – 1.00 Very High relationship  0.6 – 0.79 High Relationship  0.40 – 0.59 Substantial/Marked relationship  0.20 – 0.39 Low relationship  0.00 – 0.19 Negligible relationship
  • 7. Testing for Significance  Nominal: Phi Coefficient  Ordinal: Spearman rho  Interval & Ratio: Pearson r  Interval with nominal: Point biserial  Decision rule:  If p value < =.05: significant relationship  If p value > =.05: no significant relationship
  • 8. Variance  R2  Square the correlation coefficient  Interpretation: percentage of time that the variability in X accounts for the variability in Y.
  • 9. Reliability  Consistency of scores Obtained by the same person when retested with the identical test or with an equivalent form of the test
  • 10. Test-Retest Reliability  Repeating the identical test on a second occasion  Temporal stability  When variables are stable ex: motor coordination, finger dexterity, aptitude, capacity to learn  Correlate the scores from the first test and second test.· The higher the correlation the more reliable
  • 11. Alternate Form/Parallel Form  Same person is tested with one form on the first occasion and with another equivalent form on the second  Equivalence;  Temporal stability and consistency of response  Used for personality and mental ability tests  Correlate scores on the first form and scores on the second form
  • 12. Split half  Two scores are obtained for each person by dividing the test into equivalent halves  Internal consistency;  Homogeneity of items  Used for personality and mental ability tests  The test should have many items  Correlate scores of the odd and even numbered items  Convert the obtained correlation coefficient into a coefficient estimate using Spearman Brown 
  • 13. Kuder Richardson (KR #20/KR #21)  When computing for binary (e.g., true/false) items  Consistency of responses to all items  Used if there is a correct answer (right or wrong)  Use KR #20 or KR #21 formula
  • 14. Coefficient Alpha  The reliability that would result if all values for each item were standardized (z transformed)  Consistency of responses to all items  Homogeneity of items  Used for personality tests with multiple scored-items  Use the cronbach’s alpha formula
  • 15. Inter-item reliability  Consistency of responses to all items  Homogeneity of items  Used for personality tests with multiple scored-items  Each item is correlated with every item in the test
  • 16. Scorer Reliability  Having a sample of test papers independently scored by two examiners  To decrease examiner or scorer variance  Clinical instruments employed in intensive individual tests ex. projective tests  The two scores from the two raters obtained are correlated with each other
  • 17. Validity  Degree to which the test actually measures what it purports to measure
  • 18. Content Validity  Systematic examination of the test content to determine whether it covers a representative sample of the behavior domain to be measured.  More appropriate for achievement tests & teacher made tests  Items are based on instructional objectives, course syllabi & textbooks  Consultation with experts  Making test-specifications
  • 19. Criterion-Prediction Validity  Prediction from the test to any criterion situation over time interval  Hiring job applicants, selecting students for admission to college, assigning military personnel to occupational training programs  Test scores are correlated with other criterion measures ex: mechanical aptitude and job performance as a machinist
  • 20. Concurrent validity  Tests are administered to a group on whom criterion data are already available  Diagnosing for existing status ex. entrance exam scores of students for college with their average grade for their senior year.  Correlate the test score with the other existing measure
  • 21. Construct Validity  The extent to which the test may be said to measure a theoretical construct or trait.  Used for personality tests. Measures that are multidimensional  Correlate a new test with a similar earlier test as measured approximately the same general behavior  Factor analysis  Comparison of the upper and lower group  Point-biserial correlation (pass and fail with total test score)  Correlate subtest with the entire test
  • 22. Convergent Validity  The test should correlate significantly from variables it is related to  Commonly for personality measures  Multitrait-multidimensional matrix
  • 23. Divergent Validity  The test should not correlate significantly from variables from which it should differ  Commonly for personality measures  Multitrait-multidimensional matrix
  • 24.
  • 25.
  • 26.
  • 27.
  • 28. Item Analysis  Item Difficulty – The percentage of respondents who answered an item correctly  Item Discrimination – Degree to which an item differentiates correctly among test takers in the behavior that the test is designed to measure.
  • 29. Difficulty Index  Difficulty Index Remark  .76 or higher Easy Item  .25 to .75 Average Item  .24 or lower Difficult Item
  • 30. Index Discrimination  .40 and above - Very good item  .30 - .39 - Good item  .20 - .29 - Reasonably Good item  .10 - .19 - Marginal item  Below .10 - Poor item