ABSTRACT
The purpose of this study was to investigate the psychometric properties (reliability, validity, difficulty, discrimination, guessing and differential item function) of the items of the Biology examinations conducted by the National Examination Council (NECO) and the West African Examination Council (WAEC) using the Latent Trait Theory (IRT). This was against the backdrop of persistent public outcry that NECO questions were too cheap to pass. 14 Research questions and
7 hypotheses were formulated, tested, and analyzed. The sample was made up of 1800 senior secondary year three students from 36 secondary schools in the urban and rural areas of Benue State. The multistage stratified sampling technique was used. The NECO and WAEC Biology examination questions from 2000–2002 were the instruments for data collection. Maximum Likelihood estimation technique (using BILOG MG computer programme) was used to analyze the data in order to answer the research questions, according to IRT procedures. The t-test was used to analyse the data in order to test the hypotheses. It was found that the Biology examination items from the two examination bodies were equally reliable and valid. Biology items in the NECO – conducted examination were more difficult than those of WAEC of same years. Therefore, NECO questions were really not cheap to pass. WAEC items were more prone to guessing than those of NECO, and boys performed better in
the WAEC items than in the NECO items. It was recommended that IRT procedures should be adopted by all examination bodies in Nigeria so that our measurement problems could be put to rest.
CHAPTER ONE
INTRODUCTION
Background to the Study
In Nigeria, today, there are three examination bodies charged with the responsibility of examining and awarding ordinary level certificates at the end of senior secondary school, namely West African Examinations Council (WAEC), the National Examinations Council (NECO) and National Business and Technical Education Board (NABTEB).
As reported by WAEC (2002), the origin of WAEC dates back to 1949 when the late Dr. G.G. Jeffery was invited by British Secretary of State for the colonies to visit West Africa to study and advise on a proposal that there should be instituted a West African School Examination Council. In 1950, Dr. Jeffery submitted a report strongly supporting the proposal for a West African Examination Council and making detailed recommendations on the composition and duties of the council. The report was adopted without reservation by the four West African governments (Nigerian, Ghana, Sierra Leone and the Gambia) and an ordinance establishing the Council as a corporate body was drafted by the West African Inter-Territorial Secretariat in consultation with the governments.
In 1953, the Nigerian Government made available a large block of
offices at the Technical Institute, Yaba which became the seat of the Deputy Registrar of the Council. The member countries of WAEC include Nigeria, Ghana, Sierra Leone and The Gambia.
The council conducts different categories of examinations:
i. National Examinations -These are restricted to the specific member countries for which they are developed and reflect their local policies, needs and aspirations.
ii. International Examinations – these are developed for candidates
in all the member countries. The West African Senior Schools Certificate Examination (WASSCE) is one of such examinations. It was introduced as part of the educational reform programmes in member countries.
The Federal Government of Nigeria in April, 1999 established the National Examinations Council (NECO) with its headquarters in Minna, Niger State. The aim of establishing this council is for the nation to have an independent/national examination outfit parallel but of the same standard as WAEC. The National Examinations Council conducts national examinations such as:
i. examinations into Unity Schools (federal government secondary colleges);
ii. examinations into schools for the gifted children;
iii. ordinary level school certificate examinations (facts about NECO,
2001).
As part of measures to uplift Technical and Business Education in the country, the Federal Government in June 1992 established the National Business and Technical Examinations Board (NABTEB) with its headquarters in Benin City. The board (NABTEB) examines similar subjects handled by
WAEC and NECO and they award the National Technical Certificate (NTC) and
National Business Certificate (NBC) (NABTEB, 1997).
Test is one of the devices for measurement in schools. In terms of given objectives, testing shows the achievement of students (Ogunniyi,
1984). It consists of a set of uniform questions or tasks to which a student is to respond independently and the result of which can be treated in such a way as to provide a quantitative comparison of the performance in different students (Nworgu, 1992). Testing is a fundamental part of the teaching- learning process used not only as a basis for ranking students at the end of the teaching-learning process but to guide teaching, and aid in the development of curriculum, as well as in the assessment of needs, learning difficulties, level of mastery and differences among students. Based on the level of performance criterion, there are three types of test which are the General Mental Ability Tests, Separate Ability Test and Achievement Test (Nkemakolam, 2003).
Achievement tests measure the present proficiency, mastery and
understanding of general and specific areas of knowledge (Kerlinger, 1973). According to Nkemakolam (2003), achievement tests are designed to measure the outcome or level of accomplishment in a specified programme of instruction in a subject area or occupation, which a student had undertaken in the recent past. Achievement tests may be classified into Teacher-made (classroom) Tests and Standardized Tests. Teacher-made tests are tests constructed and administered by the classroom teacher for the purpose of measuring the achievement of pupils. Standardized tests are more carefully
and accurately designed to cover many classes of a specified type in areas that are common to these classes. It is normally constructed by a team of teachers in the field with specialists in test development. On the basis of response, achievement test is classified into objective tests and subjective (essay) test.
Essay test allows students to express themselves freely in their answers to particular questions. In an objective test, however, students’ responses are restricted to a number of symbols, words, phrases or simple sentences, one of which is considered to be the best answer out of several plausible alternatives (Ogunniyi, 1984). One of the strengths of objective test over the essay test is its amenability to quantitative item analysis, the result of which could be used to improve the item itself, specifically, and classroom learning, in general (Nenty, 1992). Other advantages of the objective tests include adequate sampling of the content to be assessed, objectivity of scoring, ease and objectivity of item analysis and greater reliability (Aiken,
1987; Bennet & Ward, 1993).
Despite being convenient to use and having some desirable psychometric properties, objective test items have been criticized for their inability to assess beyond the level of rote memorization (Ndalichako & Rogers, 1997). Bennet & Ward (1993) argued that objective tests encourage teaching and learning of isolated facts and rote procedures at the expense of conceptual understanding and the development of problem-solving skills. This is right to a very large extent. However, Maguuire, Hattie & Haig (1994) indicated that objectivee items may also measure higher levels of thinking.
The NECO and WAEC Biology examinations are made up of three parts, namely: Paper 1 (practical), Paper 2 (objectives) and Paper 3 (easy questions). For this study, only the objective questions were used and analyzed. This is because psychometric properties of a test are easier to obtain using objective test items.
The practical relevance of these tests is largely dependent on their levels of reliability, validity, difficulty and discrimination. All these add up to the psychometric properties of a test. The development of achievement, ability, aptitude, interest and personality tests is generally a multi-step process that can follow one of two distinct measurement frameworks. These are usually called the classical Test Theory (CTT) and Item Response Theory (IRT) measurement strategies (Macdonald & Paunanen, 2002).
The CTT is a measurement framework used in the development of
achievement, aptitude, interest, ability, and personality tests. Measurement in Education and Psychology has, for a long time, based its theory and practice on the Classical Test Theory (CTT). Under the CTT framework, item analysis largely consists of calculating difficulty and discrimination indices for each item. According to Thorndike (1977), the classical test theory views the score X that a student receives on a test as the sum of two observable components (True score and Error score). In Psychometrics, item response theory (IRT) is a body of theory describing the application of mathematical models to data from questionnaires and tests as a basis for measuring things such as abilities and attitudes.
IRT models are mathematical functions that specify the probability of a discrete outcome, such as a correct response to an item, in terms of person and item parameters, Person parameters may, for example, represent the ability of a student or the strength of a person’s attitude. Item parameters include difficulty (location), discrimination (slope), and pseudoguessing (lower asymptote). Items may be questions that have incorrect and correct responses, or statements on questionnaires that allow respondents to indicate level of agreement.
Among other things, as a body of theory, IRT provides a basis for evaluating how well assessments work, in terms of being able to assess the individual’s ability and how well individual questions on assessments work. In education, psychometricians apply IRT in order to achieve tasks such as developing and refining examination items, maintaining banks of items for exams, and equating for the difficulties of successive versions of exams (for example, to allow comparisons between results over time).
Item Response Theory (IRT) deals with the characteristics of the items
in terms of difficulty, discrimination and response pattern of examinees. Its major interest is to determine what a particular examinee might do when confronted with a test item. Baker (2001) explained that a reasonable assumption is that each examinee responding to a test item possesses some amount of underlying ability. Thus, one can consider each examinee to have a numerical value, a score that places him or her somewhere on the ability scale. Responses to items are modeled as a function of a person’s performance level of the trait being measured and the characteristic of the
items completed. In other words, once an examinee’s ability level has been established, it is possible to determine the probability of a correct response to an item the examinee has never taken assuming that certain item parameters have already been determined (Lord, 1980; Wilcox 1987; Harris, 1989).
Nenty (1992) further explained that when an examinee encounters a test item, the intention is that he will respond to it with the best of his ability, that is, on the basis of how much of what the item measures he possesses. He added that during this person-by-item encounter, the person puts in all that he knows, his ability, in order to overcome the item. The ability of the examinee in what is being measured and the difficulty of the item are supposed to be the only determining factor in the person-by-item encounter. The difficulty of an item, according to Nenty (1992), could be viewed as the highest resistance it could withstand before it is overcome or answered correctly. If an examinee’s ability is higher than that necessary to overcome the item, he gets the item right, but if the ability demanded by the item is higher than that possessed by the examinee, then he cannot overcome the item. With this relationship, examinee’s ability and item difficulty can be measured on the same scale with the same unit such as means or percentages. The examinee’s ability is the amount of what the item is measuring which is possessed by the examinee, while item difficulty is the amount of what the item is measuring just necessary to overcome the item. In a testing situation, the examinee should be able to answer all the items which demand ability less than that which he possesses and should fail to
answer correctly all the items that demand ability higher than that which he possesses.
The West African Examinations Council and the National Examination
Council base the analysis of the psychometric properties of items on the classical test theory frame work. According to Korashy (1995); Smith (1996); as well as Ndalichako & Rogers (1997), the CTT is no longer valid for ensuring objectivity in measurement because it is seen as being too restrictive and inadequate in scope for the measurement require in evaluating the effectiveness of achievement test.
The CTT has some limitations that make its use in establishing the
psychometric properties of tests questionable. The major limitation is the variant nature of the indices used to describe the item parameter (Douglas,
1990; Gruiton & Ironson, 1983). This implies that item parameters (e.g item difficulty) vary from sample to sample of the same population of testees. The proportion of examinees in a sample who get an item correct changes from a sample whose mean ability is high to one whose mean ability is low Nkopne (2001). Weiss & Davison (1981) further explained that the same individual tested in two different samples may obtain two different errors of measurement and estimate of true score. For this limitation, Douglas (1990) says that item parameters or statistics which remain invariant or stable from one sample of examinee to another have been viewed as a desideratum for any psychometric measurement.
All estimation procedures used to establish reliability co-efficient under
the CTT frame work are sample-based in nature. This contributes to its
weakness. The reliability estimates are specifically a function of a particular set of items and sample of examinees on which the data were collected Nkopne (2001). It means that in CTT the contribution of each item to the test reliability and validity depends upon what other items are in the test. The CTT provides an overall reliability index for all items on the entire test. To this Gruiton & Ironson (1983) say that it is unreasonable to assume that scores throughout an entire test will have the same degree of precision of measurement or one overall index of dependability.
Scores obtained by an examinee in a test under the CTT are expressed as the total raw score the examinee obtained on the content areas concerned. It does not consider the patterns of response nor what a particular examinee might do when confronted with difficult test items. It does not take care of individual differences among examinees. These limitations have made the CTT framework unfit for proper establishment of the psychometric properties of tests (examinations). This calls for a better framework which the Item Response Theory (IRT) offers. To avoid the limitations of the CIT examining bodies should adopt IRT in establishing the psychometric properties of tests. The IRT is invariant in nature.
Item parameters are not described based on samples. The item parameters (e.g. item difficulty) do not vary from sample to sample as with the CTT. Gulliksen, (1969) and Lord & Novick (1988) identified invariant item parameters as an outstanding measurement issue and suggested that whosoever elucidated it would have made a significant contribution to item analysis theory. The invariant nature of IRT models makes the estimate of
the test item parameters and person’s ability independent of which subgroups the person belongs to or on the selection of the specific set of items provided the data fit the model (Wainer, Morgan & Gustafsson, 1980).
Another major advantage of IRT is that it provides a measure of
precision of ability estimate at each ability level. Thus instead of providing a single standard error of measurement that applies to all examinees, it provides a separate estimates of error for each examinee and each item. Based on this a precise and confident statement can be made than a conventional one overall standard error of measurement for the total test as in CTT (Guiton & Ironson, 1983).
Some of the major advantage of IRT are stated here:
1. In IRT a person’s score is based on a probability model for each item. This makes the score item-based. This is unlike the CTT where a person’s score is obtained by summing across all items in the scale which makes the score test-based.
2. IRT models consider the pattern of response by examinees.
According to Baker (1977), the knowledge of the patterns of response has implication for counseling both for students and teachers.
3. The objectivity of IRT gives it an edge over the CTT because in IRT comparison of ability of examinees does not depend on which particular group of examinees used to validate the item parameters or on a specific set of items (Tinsley & Dawis, 1975; Douglas 1980).
The CTT model is simpler and easier to use than the IRT models.
Although the IRT models are more complex and more challenging to learn
because they require some technical expertise to apply and requires the use of specialized soft wares; their practical advantages make their use a worthwhile venture.
The IRT models are three (3): the 1 – parameter model, 2 – parameter
model and 3 – parameter model.
The 1 – parameter model, also, known as the Rasch Model ascribes only the difficulty parameter of an item as the trait level required to correctly answer the question. Items that fit the model most differ only in difficulty.
The 2 – parameter model deals with the discrimination parameter of an item in addition to the items difficulty parameter.
The 3 – parameter model gives the probability of an individual with ability responding correctly to an item with a difficulty index, a discrimination index and a guessing index of. The model assumes that the three parameters (difficulty, discrimination and guessing) are necessary for an estimate of a valid relationship between the probability of a correct response of an item and the trait level (ability) of an individual. If the public is made to know the advantages of IRT, there will be no doubts about the quality of the items in the examinations.
All examinations conducted in Nigeria have been based on the CTT framework for years. Examination bodies have tended to rely on CTT in testing their candidates. The public perception about the examination bodies varies from time to time. NECO has been worst hit by negative perceptions over the quality of its examinations.
A comparison of results from three secondary schools based on the results of various subjects is presented in Appendix 1. A general observation is that candidates passed more in all subjects, with the exception of just a few, in examinations conducted by NECO than those by WAEC. For example, in 2000, the percentage passes in NECO examinations were less in only three subjects in one school, one subject in another school and three subjects in the third school, as shown in Appendix 1. Put differently, the failure rate was higher in examinations (including Biology) conducted by WAEC than those by NECO (details in Appendix 1). Something must be wrong with the CTT framework being adopted at present in the country, going by public concerns.
In terms of gender, boys (Government College) performed better than girls (Government Girls College) in 10 out of 12 subjects in the WAEC- conducted examination. This situation is similar in the rural schools (Appendix 2). In terms of school location, candidates of Government Secondary School, Ikpayongo (a rural-based school), performed better in most subjects than their counterparts in urban schools (Government Day Secondary School, Makurdi an urban centre), as shown in Appendix 2. Both schools are mixed (boys and girls).
Statement of the Problem
Some people (students and parents) believe that if a student wants to pass and make his ordinary level examinations at one sitting, he should write the NECO examinations because candidates who write the NECO examinations always passed most of their papers, for the reason that NECO examinations are cheap to pass (Newswatch, 2004, P.9). The public is of
opinion that WAEC examinations are much more difficult to pass than the NECO examinations. Alarm had been raised about “the high rate of failure in all public examinations”, particularly that of WAEC (The Guardian, 2002). An evidence of this is the record that as far back as June, 1994, out of the
524,294 candidates who sat for English Language, only 74.157 or 14.10 per cent had the effective grades of 1 to 6. A whopping 290,237 or 53.30 per cent had F9, and an additional 159,900 or 30.50 per cent made P7 or P8 in a subject where a credit pass is required for admission into any tertiary institution (The Guardian, 2000). That same year, 518,188 candidates sat for Mathematics and only 83,192 had the effective grades of 1 to 6; 219,818 had the F9 grade, while 215,108 had P7 or P8. This situation of high failure rate was attributed to the fact that the reading culture among students had drastically reduced over the years.
Similarly, in the 2000 examinations for Biology, the percentage failure was higher with WAEC examination than that of NECO in the schools sampled (see App. 1). In terms of gender the failure rate was higher among girls than the boys in the schools sampled. In terms of location, there was no significant difference in performance of students in urban and rural schools (App. 2).
The pioneer Registrar/Chief Executive of NECO, Professor Ojerinde, agrees that there are several examples of those who pass NECO and fail WAEC examinations and vice versa (Newswatch, 2004, p. 9). He agrees that some people say that “NECO is cheap to pass but that NECO papers are not cheap”. For this reason, some tertiary institutions tended to discriminate against NECO result by denying candidates placements during admissions.
Similarly, there were allegations of massive leakages in examinations conducted by NECO. To this, NECO (2001) suggested that examination malpractices were perpetrated with the connivance, active support or the neglect of some unscrupulous supervisors and invigilators. Both the WAEC and NECO examinations are based on the CTT framework in terms of analysis of their psychometric characteristics.
Obviously, the CTT has some serious disadvantages that make its use
in developing and establishing the psychometric properties of classroom achievement tests improper. For instance, the CTT model typically provides only one overall index of reliability for all items in a test. The CTT model itself does not ensure objectivity in measurement due to the fact that one technique of measuring a particular attribute is not interchangeable with an alternative measure of the same attribute. CTT fails to provide solutions to test score equating, test design and testing items for bias. Generally, the latent trait theory, unlike CTT provides invariant item parameters, equates different forms of the same test and presents test items tailored to individual test examinees. It tests items for bias and precisely determines test item fit. Accompanied by its standard error of measurement. Thus, the IRT model takes into account some qualitative information regarding the psychometric properties of each test item. Could the above reasons account for the observed differences in the performance of candidates who write examinations conducted by both WAEC and NECO? Could the latent trait theory (item response theory) offer a way out?
If the above observations are true, then IRT should be tried in developing and establishing the psychometric properties of tests. In both WAEC and NECO tests, what are the relative psychometric valules for candidates in urban and rural areas and between sexes (boys and girls), a check on this is, therefore, needed. What are the sources of performance differences by candidates who write examinations conducted by WAEC and NECO? Are the tests in one examination more valid and reliable than the other or more reliable and not valid? Having used/tried CTT through the years, can one confidently continue to accept and depend on the model for our future examinations? Can IRT offer acceptable and objective alternative to testing in Nigeria? When applied to tests conducted by examination bodies, would IRT give a better index of ability to students? Differences in performance of candidates continue to be observed? The application of IRT models to the said examinations would further confirm the levels of standardization of the tests. In other words, analyzing the tests using IRT framework would establish the nature of differences between the examinations in relation to specified psychometric properties. One way of investigating the validity of these claims is to find out how far the WAEC and NECO examinations have kept up high testing standards with particular reference to psychometric characteristics. Are the NECO test items of lower quality than those of the WAEC?
Purpose of the Study
The primary purpose of this study is to determine the relative strength of the psychometric characteristics of the test items used in Biology
examinations conducted by WAEC and NECO using IRT. Specifically, the study determined the:
1. reliability of the test items in the Biology examinations
conducted by NECO and WAEC, using the standard error of measurement.
2. validity of the test items in the Biology examinations conducted by NECO and WAEC; using the fit statistic technique.
3. difficulty parameter of the test items in the Biology examinations conducted by NECO and WAEC;
4. discrimination parameter of the tests items in Biology
examinations conducted by NECO and WAEC;
5. differential item functioning (DIF) of the test items administrated by WAEC and NECO; in terms of gender.
6. differential item functioning of the test items administered by
WAEC and NECO with reference to school location;
7. levels of guessing parameter of the test items in Biology examinations conducted by WAEC and NECO.
Significance of the Study
The study will provide input into test construction skill of test developers/examining bodies in determining the existence or non-existence of item differential functioning (DIF). Detecting item differential functioning or item bias as a problem in examinations will only be reasonably possible using the latent trait model because of its sample-invariant property.
The results from this study would encourage test developers to take pains to conduct rigorous item analysis before and after test administrations. The latent trait models ensure the equating of test scores using ability estimates and, thus, are useful in controlling the problem of non-equivalent groups associated with CTT. The understanding of the practical approach and procedure for estimating examinee ability even when some items might not have been responded to (that is total scores determined from partial scores) would solve the problem of using graded scoring model.
The study will be significant to classroom teachers, guidance counselors and educational institutions by providing clues to meaningful interpretation of examinees result through person-by-item encounter (latent trait model) during examinations.
The study will, also, serve as an efficient tool for diagnostic evaluation
by the teachers, guidance counselors and educational institutions, which involves not only the identification of errors but the identification of the misconceptions that underlie such errors. It will ensure improvement in the teacher’s teaching behaviour in terms of coverage, emphasis of curricula content and instructional practices. The strengths and weaknesses of candidates in examinations will be readily known because guidance counselors would be amply armed with the necessary information/data about each examinee’s performance item by item.
The results of this study would help to indicate the overall quality of the examinations conducted in Biology by WAEC and NECO. A confirmation of the reliability and validity of the examinations conducted by WAEC and NECO
would help to establish public confidence and acceptability of results from their examinations. The public needs to be convinced that the examinations conducted by both WAEC and NECO are of relatively equal standards and that no one is of inferior quality. Thus, the public will be enlightened on the interpretation of students’ results from the examinations conducted by these two examining bodies. They will not just assume high quality of either of the examining bodies. Presently, the performances of students in the examinations conducted by these two bodies are interpreted based on the sum of their total scores which is typical of CTT. The use of these scores just summed across all items to consider the performance of examinees hides the characteristics of both the examinee and the test. For objective and adequate decisions to be taken on the performance of students in examinations by WAEC and NECO, the psychometrics of the tests need to be determined. Examining bodies need to consider the psychometric properties of tests in taking decisions on the observable performance of candidates in order to improve upon test construction, administration and analysis. In addition, from this study, WAEC and NECO would have a clearer understanding of their performance in test construction and be appropriately guided from now on as they, hopefully, accept and adopt IRT evaluation framework.
Scope of the Study
The study covered all the secondary schools in the three education zones of Benue State, Nigeria. The study is limited to the May/June WAEC (2000 – 2002) Biology examinations and the August/September (2000) and (June/July, 2001 – 2002) NECO Biology examinations. This is because the
maiden examination conducted by NECO was in 2000, and so many candidates were said to have passed the examinations with good grades especially those who could not pass the WAEC-conducted Biology examination in that same year. The 2000 May/June WAEC Biology examination was the second year of WAEC administering the West African Senior School Certificate Examination (WASSCE). This study covers One, Two and Three – parameter models. As required in all IRT models, a relatively large sample size was used in order to ensure very high reliability of the findings.
Research Questions
The study was carried out through answering the following research questions:
1. What are the standard errors of measurement of the test items in the
Biology examinations conducted by NECO?
2. What are the standard errors of measurement of the test items in the
Biology examinations conducted by WAEC?
3. How valid are the test items of the Biology examinations conducted by
NECO?
4. How valid are the test items of the Biology examinations conducted by
WAEC?
5. What are the difficulty parameters of the test items in the Biology examinations conducted by NECO?
6. What are the difficulty parameters of the test items in the Biology examinations conducted by WAEC?
7. What are the discrimination parameters of the test items in the Biology examinations conducted by NECO?
8. What are the discrimination parameters of the test items in the Biology
examinations conducted by WAEC?
9. What are the guessing parameters of the test items in the Biology examinations conducted by NECO?
10. What are the guessing parameters of the test items in the Biology
examinations conducted by WAEC?
11. How do the test items of the Biology examinations conducted by NECO
function with respect to sex (boys and girls)?
12. How do the test items of the Biology examinations conducted by WAEC
function with respect to sex (boys and girls)?
13. How do the test items of the Biology examinations conducted by NECO
function with respect to location (urban and rural)?
14. How do the test items of the Biology examinations conducted by WAEC
function with respect to location (urban and rural)?
Hypotheses
1. There is no significant difference in the mean standard errors of measurement of the items in the Biology examinations conducted by NECO and WAEC from 2000 – 2002 based on One, Two and Three – parameter models of IRT.
2. There is no significant difference in the validity (fit statistic) of items of the Biology examinations conducted by NECO and WAEC from 2000-
2002 based on the One, Two and Three – parameter models of IRT.
3. There is no significant difference in the difficulty parameters of items in the NECO and WAEC conducted Biology examinations from 2000-2002 based on the One parameter model of IRT.
4. There is no significant difference in the discrimination parameters of
items in the Biology examinations conducted by NECO and WAEC from
2000-2002 based on Two – parameter model of IRT.
5. There is no significant difference in the guessing parameters of items of the Biology examinations conducted by NECO and WAEC from 2000-
2002 based on the Three-parameter model of IRT.
6. The test items of the Biology examinations conducted by NECO and
WAEC do not differ in function among examinees with respect to sex.
7. The test items of the Biology examinations conducted by NECO and WAEC do not differ in function among examinees with respect to location.
This material content is developed to serve as a GUIDE for students to conduct academic research
COMPARISON OF PSYCHOMETRIC PROPERTIES OF WEST AFRICAN EXAMINATIONS COUNCIL AND NATIONAL EXAMINATIONS COUNCIL TEST ITEMS UNDER ITEM RESPONSE THEORY>
PROJECTOPICS.com Support Team Are Always (24/7) Online To Help You With Your Project
Chat Us on WhatsApp » 07035244445
DO YOU NEED CLARIFICATION? CALL OUR HELP DESK:
07035244445 (Country Code: +234)YOU CAN REACH OUR SUPPORT TEAM VIA MAIL: [email protected]