SlideShare a Scribd company logo
ADOPT, DEVELOP AND ADAPT?
ADOPT, DEVELOP AND ADAPT? Once a consensus has been built as to the purposes and types of test to employ in a program. A strategy must be worked out to maximize the quality and effectiveness of the test. In the best of all possible worlds each program would have a resident testing expert, whose entire job would be to develop tests especially created for and suited to that program. But even in the worst of all possible worlds rational decisions can be made when selecting commercially available tests from scratch or adopting them from commercial sources is the possibility of adapting existing tests so they are made to better fit with the purposes and objectives of the program.
Many language tests are, or should be, situation specific. That is to say, a test can be very effective in one situation with one particular group of students and be virtually useless in another situation or with another group of students. Teachers can not simply go out (or worse yet, illegally photocopy) a test and automatically expect it to work with their student. It may have been developed for completely different types of students (different in background, level of proficiency, gender, and so forth) and for entirely different purposes (that is, base on differing approaches, syllabuses, techniques or exercises).
Though all of these may seem like a great deal of work, remember that in most language programs, any rational approach to testing will be a vast improvement over the existing condition. The purpose in this section of the chapter will be to suggest systematically bases for getting started in adopting, developing, or adapting decent language test for a particular and very specific language program.
Adopting Language Tests The tests that are used for program decision are very often bought from commercial publishing houses. Tests are also sometimes adopted from other language programs or taken straight from the current textbook. Given the wide diversity and variation in the nationalities and levels involved in the various language programs around the world, it may turn out that any tests that is adopted is being applied to a population quite different from envisioned when the test were originally written. As a result, program decisions that can dramatically affect the lives of the student
May be irresponsibly base on tests consisting of test questions that are quite unrelated to the needs of the particular group of students or to the curriculum being taught in the specific program involved. Selecting good test that match the specific needs of a program is therefore important. Test reviews are one good place to start. Such reviews can be found in the review sections of some language teaching journals, right alongside the reviews of texts and professional books. Unfortunately, test reviews appear infrequently:
Language Testing  - is a journal that focuses on language tests and also provides reviews. For those in ESL/EEL, Alderson, Krahnke and Stanfield 1987 is a useful source of test reviews for most of the major test available at the time it was published. The  Mental Measurements Yearbook  also includes some reviews of language test.
Alternative ways to approach the task of selecting test for a program might include: 1. Taking a language testing course 2. Reading up on testing. 3. Hiring a person who already knows about  testing. 4. Giving one member of the staff release time ti  become informed all the topic.
Table 4.3 Test Evaluation Checklist  General Background Information 1. Title 2. Author 3. Publisher and Date of Publication 4. Published reviews available Theoretical Orientation 1. Test family (norm-referenced or criterion-referenced) 2. Purpose of decision (proficiency, placement, achievement or  diagnosis) 3. Language methodology orientation approach and  syllabuses Practical Orientation 1. Target population (age, level, nationality, language/dialect,  educational background and so forth) 2. Skills test (for instance reading, writing, listening, speaking,  structure, vocabulary and pronunciation)
3. Number of subtests and separate scores. 4. Type of items reflect appropriate techniques and exercises  (receptive true-false, multiple-choice, matching; productive:  fill-in, short response, essay, extended discourse task). Test Characteristics 1. Norms a. Standardization sample b. Type of standardized scores 2. Descriptive statistics (central tendency, dispersion, item,  characteristics) 3. Reliability a. Types of reliability procedures used b. Degree of reliability for each procedures  4. Validity a. Types of validity procedures use b. Do you buy the above validity arguments?
5. Practicality a. Cost of test booklets, cassette tapes, manual, answer sheets,    scoring templates, scoring services, any other necessary test    components. b. Quality of items listed immediately above (paper, printing,    audio clarity, durability and so forth) c. Ease of administration (time required, proctor/examinee,      ratio, proctor qualifications, equipment necessary,    availability and quality of direction for administration and so    forth) d. Ease of scoring (method of scoring, amount of training      necessary, time per test, score conversion information and    so forth) e. Ease of interpretation (quality of guidelines for the        interpretations of scores in terms of norms or other criteria.)
Developing Language Test In the best of all possible worlds, sufficient resources and expertise will be available in a program so that proficiency, placement, achievement, and diagnostic tests can be developed and fitted to the specific goals of the program and to the specific population studying in it. If this is the case, decision must be made about which types of tests must be developed first. That might mean first developing achievement and diagnosis tests, while temporarily adopting previously published proficiency and placement test.
A program specific placement test could be developed so that the reasons for separating students into levels in the program are related to the things that the students can learn while in those levels. It is rarely necessary or even useful to develop program specific proficiency tests because of their interprogrammatic nature. In other words, for purposes of reference to other programs elsewhere, an adopted test that is used by a wide variety of language programs will be most appropriate. Naturally, all of these decisions are up to the teachers, administrators, and curriculum developers in the program in question.
Adapting Language Test It may turn out that a pre-existing test that works fairly well, but not perfectly, can be adapted to the specific testing needs of a particular program.  The process of adapting a test to a specific situation will probably involve some variant of the following strategy: 1.  Administer the test to the students to the  program. 2. Select those items that appear to be doing a  good job of spreading.
3. Create a shorter, more efficient, revised version  of the test that fits the ability levels of the  specific population of students. 4. Create new item that function like those that  were working well in order to have a test of  sufficient length.
Organizing and Using Test Result Table 4.4 A Checklist for  Successful  Testing (adapted from Brown forthcoming) Purposes of Test 1.  Clearly defined (theoretical and practical orientation) 2. Understood and agreed upon by staff Test itself  Physical needs arranged 1.  Adequate and quiet space 2.  Enough time in that space for some flexibility 3.  Clear scheduling Preadministration arrangement 1.  Students properly notified
2.  Students signed up for level 3.  Students given precise information (where and when  test will be, as well as what they should do to prepare and  what they should bring with them especially identification  if required) Administration 1.  Adequate materials in hand (test booklets, answer  sheets, cassette tapes, pencils, scoring templates, and so  forth) plus extras 2.  All necessary equipment in hand and tested (cassette  players, microphones, public address system, videotape  players, blackboard, chalk, and so forth) with backups  where appropriate 3. Proctors trained in their duties 4. All necessary information distributed to proctors (test
directions, answers to obvious questions, schedule of who  is to be where and when, and so forth) Scoring 1. Adequate space for all scoring take place 2. Clear scheduling of scoring and notification of results 3. Sufficient qualified staff for all scoring activities 4. Staff trained in all scoring procedures Interpretation 1. Clearly defined uses for result 2. Provision for helping teachers interpret scores and explain  them to students 3. Provision for eventual systematic termination of records Ongoing 1. Results used to fall advantage for research 2. Results incorporated into overall program evaluation plan
 

More Related Content

PPTX
Assessment in language teaching
PPTX
Formal instruction and language learning
PPTX
Children VS Adult in Second-Language Learning
PPT
Chapter 3(designing classroom language tests)
PPTX
Principles of language assessment
PPTX
Testing grammar and vocabulary
PPTX
Chapter 10 style, context and register
Assessment in language teaching
Formal instruction and language learning
Children VS Adult in Second-Language Learning
Chapter 3(designing classroom language tests)
Principles of language assessment
Testing grammar and vocabulary
Chapter 10 style, context and register

What's hot (20)

PPT
Ug & sla
PPTX
The origins of esp
PDF
ASSESSMENT CONCEPTS AND ISSUES
PPTX
TBL (Task based learning)
PPTX
Testing Grammar and Vocabulary Skill
PPTX
Teaching across Age Level
PPTX
Principles of language assessment
PPTX
Structural Syllabus Design
PPT
Describing Teachers
PPTX
Testing reading
PPTX
Grammar translation method
PPTX
Task based syllabus
PPTX
Communicative Language Teaching (CLT)
PPTX
Syllabus design & materials development
PPTX
Materials design
PPT
Chapter 7(assessing speaking )
PPTX
English Intonation
PPTX
Spoken language (Discourse Analysis)
PPTX
Interaction hypothesis
Ug & sla
The origins of esp
ASSESSMENT CONCEPTS AND ISSUES
TBL (Task based learning)
Testing Grammar and Vocabulary Skill
Teaching across Age Level
Principles of language assessment
Structural Syllabus Design
Describing Teachers
Testing reading
Grammar translation method
Task based syllabus
Communicative Language Teaching (CLT)
Syllabus design & materials development
Materials design
Chapter 7(assessing speaking )
English Intonation
Spoken language (Discourse Analysis)
Interaction hypothesis
Ad

Viewers also liked (17)

PPT
Testing in language programs (chapter 8)
PPTX
Adopting and Adapting Materials
PPTX
Norm reference grading system.ppt
PPTX
ppt norm reference and criteration test
PPTX
Norm or criterion referenced grading
PPTX
Norm-referenced & Criterion-referenced Tests
PPTX
Norm-Referenced and Criterion-Referenced Interpretation
PPTX
Materials development for language learning and teaching
PPTX
Language teaching materials
PPTX
Brown, chapter 4 By Savaedi
PPTX
Testing language skills chapter one
PPTX
Language learning materials development
PPTX
Testing language areas and skills
PPTX
Fundamental concepts and principles in Language Testing
PPTX
Norm Reference Test
PPT
Global Marketing
PPTX
Test Usefulness
Testing in language programs (chapter 8)
Adopting and Adapting Materials
Norm reference grading system.ppt
ppt norm reference and criteration test
Norm or criterion referenced grading
Norm-referenced & Criterion-referenced Tests
Norm-Referenced and Criterion-Referenced Interpretation
Materials development for language learning and teaching
Language teaching materials
Brown, chapter 4 By Savaedi
Testing language skills chapter one
Language learning materials development
Testing language areas and skills
Fundamental concepts and principles in Language Testing
Norm Reference Test
Global Marketing
Test Usefulness
Ad

Similar to Chapter 4 testing aima (20)

PPTX
Teaching Interated skills of English at University.
PDF
Chapter-3 Dsigning classroom language Tests 1.pdf
PDF
Chapter-3 Dsigning classroom language Tests.pdf
PDF
2. presentation evaluation segunda part.pdf
PPTX
Testing and Evaluation Strategies in Second Language Teaching.pptx
PPT
Language testing final
PPT
Language Testing
PPTX
Principles of language assessment.pptx
PPT
Hoffmann magno testing_apiba_27_10_12
PPT
LING139 (MARTINEZ MAEd-Eng).ppt
PPTX
Uses of language by Brown 1990
PPTX
A1_Camila_Arias_Assessment_and_evaluation.pptx
PDF
Principles of language assessment
PPT
Languaje Testing, I Bimestr
PPTX
Designing classroom language tests
PPTX
Testing mc m
DOCX
What does language testing have to offier
PPTX
Learning_activity1_Navarro Luzuriaga_Joseph Andrés.pptx
PPT
A AND E.ppt
PPT
Communicative language testing
Teaching Interated skills of English at University.
Chapter-3 Dsigning classroom language Tests 1.pdf
Chapter-3 Dsigning classroom language Tests.pdf
2. presentation evaluation segunda part.pdf
Testing and Evaluation Strategies in Second Language Teaching.pptx
Language testing final
Language Testing
Principles of language assessment.pptx
Hoffmann magno testing_apiba_27_10_12
LING139 (MARTINEZ MAEd-Eng).ppt
Uses of language by Brown 1990
A1_Camila_Arias_Assessment_and_evaluation.pptx
Principles of language assessment
Languaje Testing, I Bimestr
Designing classroom language tests
Testing mc m
What does language testing have to offier
Learning_activity1_Navarro Luzuriaga_Joseph Andrés.pptx
A AND E.ppt
Communicative language testing

Recently uploaded (20)

PPTX
1st Inaugural Professorial Lecture held on 19th February 2020 (Governance and...
PPTX
GDM (1) (1).pptx small presentation for students
PDF
BÀI TẬP BỔ TRỢ 4 KỸ NĂNG TIẾNG ANH 9 GLOBAL SUCCESS - CẢ NĂM - BÁM SÁT FORM Đ...
PDF
Saundersa Comprehensive Review for the NCLEX-RN Examination.pdf
PPTX
Introduction_to_Human_Anatomy_and_Physiology_for_B.Pharm.pptx
PDF
The Lost Whites of Pakistan by Jahanzaib Mughal.pdf
PDF
Pre independence Education in Inndia.pdf
PPTX
Renaissance Architecture: A Journey from Faith to Humanism
PDF
2.FourierTransform-ShortQuestionswithAnswers.pdf
PPTX
school management -TNTEU- B.Ed., Semester II Unit 1.pptx
PDF
grade 11-chemistry_fetena_net_5883.pdf teacher guide for all student
PDF
TR - Agricultural Crops Production NC III.pdf
PDF
102 student loan defaulters named and shamed – Is someone you know on the list?
PDF
01-Introduction-to-Information-Management.pdf
PDF
VCE English Exam - Section C Student Revision Booklet
PDF
Anesthesia in Laparoscopic Surgery in India
PPTX
Pharma ospi slides which help in ospi learning
PDF
ANTIBIOTICS.pptx.pdf………………… xxxxxxxxxxxxx
PDF
Computing-Curriculum for Schools in Ghana
PDF
Abdominal Access Techniques with Prof. Dr. R K Mishra
1st Inaugural Professorial Lecture held on 19th February 2020 (Governance and...
GDM (1) (1).pptx small presentation for students
BÀI TẬP BỔ TRỢ 4 KỸ NĂNG TIẾNG ANH 9 GLOBAL SUCCESS - CẢ NĂM - BÁM SÁT FORM Đ...
Saundersa Comprehensive Review for the NCLEX-RN Examination.pdf
Introduction_to_Human_Anatomy_and_Physiology_for_B.Pharm.pptx
The Lost Whites of Pakistan by Jahanzaib Mughal.pdf
Pre independence Education in Inndia.pdf
Renaissance Architecture: A Journey from Faith to Humanism
2.FourierTransform-ShortQuestionswithAnswers.pdf
school management -TNTEU- B.Ed., Semester II Unit 1.pptx
grade 11-chemistry_fetena_net_5883.pdf teacher guide for all student
TR - Agricultural Crops Production NC III.pdf
102 student loan defaulters named and shamed – Is someone you know on the list?
01-Introduction-to-Information-Management.pdf
VCE English Exam - Section C Student Revision Booklet
Anesthesia in Laparoscopic Surgery in India
Pharma ospi slides which help in ospi learning
ANTIBIOTICS.pptx.pdf………………… xxxxxxxxxxxxx
Computing-Curriculum for Schools in Ghana
Abdominal Access Techniques with Prof. Dr. R K Mishra

Chapter 4 testing aima

  • 2. ADOPT, DEVELOP AND ADAPT? Once a consensus has been built as to the purposes and types of test to employ in a program. A strategy must be worked out to maximize the quality and effectiveness of the test. In the best of all possible worlds each program would have a resident testing expert, whose entire job would be to develop tests especially created for and suited to that program. But even in the worst of all possible worlds rational decisions can be made when selecting commercially available tests from scratch or adopting them from commercial sources is the possibility of adapting existing tests so they are made to better fit with the purposes and objectives of the program.
  • 3. Many language tests are, or should be, situation specific. That is to say, a test can be very effective in one situation with one particular group of students and be virtually useless in another situation or with another group of students. Teachers can not simply go out (or worse yet, illegally photocopy) a test and automatically expect it to work with their student. It may have been developed for completely different types of students (different in background, level of proficiency, gender, and so forth) and for entirely different purposes (that is, base on differing approaches, syllabuses, techniques or exercises).
  • 4. Though all of these may seem like a great deal of work, remember that in most language programs, any rational approach to testing will be a vast improvement over the existing condition. The purpose in this section of the chapter will be to suggest systematically bases for getting started in adopting, developing, or adapting decent language test for a particular and very specific language program.
  • 5. Adopting Language Tests The tests that are used for program decision are very often bought from commercial publishing houses. Tests are also sometimes adopted from other language programs or taken straight from the current textbook. Given the wide diversity and variation in the nationalities and levels involved in the various language programs around the world, it may turn out that any tests that is adopted is being applied to a population quite different from envisioned when the test were originally written. As a result, program decisions that can dramatically affect the lives of the student
  • 6. May be irresponsibly base on tests consisting of test questions that are quite unrelated to the needs of the particular group of students or to the curriculum being taught in the specific program involved. Selecting good test that match the specific needs of a program is therefore important. Test reviews are one good place to start. Such reviews can be found in the review sections of some language teaching journals, right alongside the reviews of texts and professional books. Unfortunately, test reviews appear infrequently:
  • 7. Language Testing - is a journal that focuses on language tests and also provides reviews. For those in ESL/EEL, Alderson, Krahnke and Stanfield 1987 is a useful source of test reviews for most of the major test available at the time it was published. The Mental Measurements Yearbook also includes some reviews of language test.
  • 8. Alternative ways to approach the task of selecting test for a program might include: 1. Taking a language testing course 2. Reading up on testing. 3. Hiring a person who already knows about testing. 4. Giving one member of the staff release time ti become informed all the topic.
  • 9. Table 4.3 Test Evaluation Checklist General Background Information 1. Title 2. Author 3. Publisher and Date of Publication 4. Published reviews available Theoretical Orientation 1. Test family (norm-referenced or criterion-referenced) 2. Purpose of decision (proficiency, placement, achievement or diagnosis) 3. Language methodology orientation approach and syllabuses Practical Orientation 1. Target population (age, level, nationality, language/dialect, educational background and so forth) 2. Skills test (for instance reading, writing, listening, speaking, structure, vocabulary and pronunciation)
  • 10. 3. Number of subtests and separate scores. 4. Type of items reflect appropriate techniques and exercises (receptive true-false, multiple-choice, matching; productive: fill-in, short response, essay, extended discourse task). Test Characteristics 1. Norms a. Standardization sample b. Type of standardized scores 2. Descriptive statistics (central tendency, dispersion, item, characteristics) 3. Reliability a. Types of reliability procedures used b. Degree of reliability for each procedures 4. Validity a. Types of validity procedures use b. Do you buy the above validity arguments?
  • 11. 5. Practicality a. Cost of test booklets, cassette tapes, manual, answer sheets, scoring templates, scoring services, any other necessary test components. b. Quality of items listed immediately above (paper, printing, audio clarity, durability and so forth) c. Ease of administration (time required, proctor/examinee, ratio, proctor qualifications, equipment necessary, availability and quality of direction for administration and so forth) d. Ease of scoring (method of scoring, amount of training necessary, time per test, score conversion information and so forth) e. Ease of interpretation (quality of guidelines for the interpretations of scores in terms of norms or other criteria.)
  • 12. Developing Language Test In the best of all possible worlds, sufficient resources and expertise will be available in a program so that proficiency, placement, achievement, and diagnostic tests can be developed and fitted to the specific goals of the program and to the specific population studying in it. If this is the case, decision must be made about which types of tests must be developed first. That might mean first developing achievement and diagnosis tests, while temporarily adopting previously published proficiency and placement test.
  • 13. A program specific placement test could be developed so that the reasons for separating students into levels in the program are related to the things that the students can learn while in those levels. It is rarely necessary or even useful to develop program specific proficiency tests because of their interprogrammatic nature. In other words, for purposes of reference to other programs elsewhere, an adopted test that is used by a wide variety of language programs will be most appropriate. Naturally, all of these decisions are up to the teachers, administrators, and curriculum developers in the program in question.
  • 14. Adapting Language Test It may turn out that a pre-existing test that works fairly well, but not perfectly, can be adapted to the specific testing needs of a particular program. The process of adapting a test to a specific situation will probably involve some variant of the following strategy: 1. Administer the test to the students to the program. 2. Select those items that appear to be doing a good job of spreading.
  • 15. 3. Create a shorter, more efficient, revised version of the test that fits the ability levels of the specific population of students. 4. Create new item that function like those that were working well in order to have a test of sufficient length.
  • 16. Organizing and Using Test Result Table 4.4 A Checklist for Successful Testing (adapted from Brown forthcoming) Purposes of Test 1. Clearly defined (theoretical and practical orientation) 2. Understood and agreed upon by staff Test itself Physical needs arranged 1. Adequate and quiet space 2. Enough time in that space for some flexibility 3. Clear scheduling Preadministration arrangement 1. Students properly notified
  • 17. 2. Students signed up for level 3. Students given precise information (where and when test will be, as well as what they should do to prepare and what they should bring with them especially identification if required) Administration 1. Adequate materials in hand (test booklets, answer sheets, cassette tapes, pencils, scoring templates, and so forth) plus extras 2. All necessary equipment in hand and tested (cassette players, microphones, public address system, videotape players, blackboard, chalk, and so forth) with backups where appropriate 3. Proctors trained in their duties 4. All necessary information distributed to proctors (test
  • 18. directions, answers to obvious questions, schedule of who is to be where and when, and so forth) Scoring 1. Adequate space for all scoring take place 2. Clear scheduling of scoring and notification of results 3. Sufficient qualified staff for all scoring activities 4. Staff trained in all scoring procedures Interpretation 1. Clearly defined uses for result 2. Provision for helping teachers interpret scores and explain them to students 3. Provision for eventual systematic termination of records Ongoing 1. Results used to fall advantage for research 2. Results incorporated into overall program evaluation plan
  • 19.