Open Access Paper
28 December 2022 An objective method of measuring students’ achievement
Jin Li, Shiru Lan, Zhanglin Li, Wenjiu Du, Yong Wu, Xianyi Yao
Author Affiliations +
Proceedings Volume 12506, Third International Conference on Computer Science and Communication Technology (ICCSCT 2022); 1250622 (2022) https://doi.org/10.1117/12.2661817
Event: International Conference on Computer Science and Communication Technology (ICCSCT 2022), 2022, Beijing, China
Abstract
This paper introduces an objective method to measure students’ achievement. Among many popular methods for evaluating and testing students’ scores, different participants often obtain different results on subjective questions, which obviously lose the objective and fair standard for students’ evaluation. In order to overcome these defects, a new scoring method is recommended in this paper, so that different raters can obtain more objective and fair results in the test of subjective questions as much as possible. In this paper, this scoring method is divided into some main essential steps, which can be summarized as: find out the key points of subjective questions, and divide the questions into several different scoring grades according to different scoring steps.

1.

INTRODUCTION

The main defects of subjective test result evaluation are: different raters often get different results; The test results depend on the difficulty of the test questions, because the questions with different difficulty are not equivalent; Lack of scientific methods to evaluate errors. Therefore, using scores to evaluate subjective tests is not very scientific. This paper introduces a method of objectively evaluating students’ achievements, which can overcome the above defects.

2.

SCORING OF SUBJECTIVE QUESTIONS

An example is discussed below1.

Solving the equations:

00092_PSISDG12506_1250622_page_1_1.jpg
  • The solving steps: According to the inequality (1), solutions have to

  • Step (a): 00092_PSISDG12506_1250622_page_1_2.jpg 1 point;

  • Step (b): inequality (2), we can get x < 0 or x > 1 2 points;

  • Step (c): their common solution is 00092_PSISDG12506_1250622_page_1_3.jpg or 00092_PSISDG12506_1250622_page_1_4.jpg 4 points;

Because the examinee’s responses at each step have two results: right or wrong, the examinee’s response at each step is a random event.

Suppose00092_PSISDG12506_1250622_page_1_5.jpg; B : x < 0 or x > 1; 00092_PSISDG12506_1250622_page_1_6.jpg, through the analysis of the solution process of case 1, it is not difficult to find that no matter whether the subject’s answer to equation (1) is right or wrong, it will not affect the subject’s answer to equation (2). In other words, regardless of whether the examinee’s answer to equation (2) is right or wrong, it will not affect the examinee’s answer to Inequality (1). It can be seen that event A and event B are independent of each other. However, only after event A and event B occur at the same time, event C may occur. If event A or event B does not occur, event C must not occur, thus (AB) ⊃ C.

Through the above analysis, it is not difficult to find that the three events A, B, and C satisfy the following relationships: (1) event A, B are independent; 2) (AB) ⊃ C. According to the relationship between A, B, C, the following results may appear in the process of problem solving1:

  • Events A and B do not occur, that is, the examinee’s answers to both equations (1) and (2) are wrong, and the answer on C must be wrong. That is, an event occurred: 00092_PSISDG12506_1250622_page_2_1.jpg, denoted as 0 points;

  • Event A occurs but event B does not occur, that is, the subject answered correctly on equation (1) but incorrectly on equation (2), an event occurred: 00092_PSISDG12506_1250622_page_2_2.jpg, denoted as 1 points;

  • Event A did not occur, but event B occurred, that is, the subject answered incorrectly to equation (1), but the answer to equation (2) was correct, an event occurred: 00092_PSISDG12506_1250622_page_2_3.jpg, denoted as 2 points;

  • Events A and B both occur but C does not occur; event occurred: A · BC, denoted as 3 points;

  • Events A, B, and C all occur, denoted as 4 points.

Assume that 00092_PSISDG12506_1250622_page_2_4.jpg, the Ω contains all possible outcomes of the solution of case 1. Suppose that X = {0,1,2,3,4}, and define a 1 to 1 mapping function f : Ω → X, the function is specified such that each “score” in X corresponds to an event in the same position in Ω.

Through function f, events that may occur during problem solving are linked to a set of numbers2, this number set X can be used as the scoring step of the quiz item. In other words, case 1 can be divided into 4 different scoring steps, each step is scored as 1 point, 2 points, 3 points, 4 points. If the scenarios in which the subjects scored 0 are also counted, then case 1 can be divided into 5 different scoring levels. Its advantage is that the test score of the subjects has nothing to do with the reviewers. For the same topic, different reviewers can also score the same score, avoiding the error caused by the reviewers’ personal reasons.

3.

INVARIANTS IN THE EXAMINATION

In the test T. Let X1,X2,…,XN denote the scores (true scores) of N examinee’s on test T, respectively, Xi represents the score of the i-th subject in the test, Ni represents the total number of test scores not greater than Xi, and 00092_PSISDG12506_1250622_page_2_5.jpg represents the percentile of the examinee i in the test. When N → ∞, the limit value of Pi is the percentile of the examinee in the overall test3.

It is further assumed that the test T will preserve the order of the examiners, the meaning of preserve the order of the examiners can be explained that: suppose T1, T2 are any two tests in the same attribute set, A, B are any two examiners, the scores of examiner A on T1, T2 are X1 and X2 respectively, examiner B on T1, T2 are Y1 and Y2, then the corresponding relation of X2, Y2 is as follows: (1) 00092_PSISDG12506_1250622_page_2_6.jpg; (2) 00092_PSISDG12506_1250622_page_2_7.jpg; (3) 00092_PSISDG12506_1250622_page_2_8.jpg. Under the above assumptions, percentile Pi is an invariant parameter4.

Let 00092_PSISDG12506_1250622_page_2_9.jpg, θiR. From the definition of θі, we can know that θі is uniquely determined by Pi. Because Pi is an invariant parameter, θі is also an invariant parameter5. It is easy to know from the knowledge of mathematical statistics, the estimation of ability parameter θ has the following large sample properties: (1) The ability parameter θ is consistent estimation. (2) The ability parameter θ obey the normal distribution N(0, 1).

4.

INVARIANT FRACTION

Now we discuss how to convert the ability parameter into people’s habitual “score”6. Let 00092_PSISDG12506_1250622_page_3_1.jpg because θ ~ N(0, 1), according to the nature of normal distribution, it is not difficult to get 00092_PSISDG12506_1250622_page_3_2.jpg. Due to θ ~ N(0, 1), from the nature of normal distribution, the probability of falling within the interval (–2.5,2.5) is slightly 0.995, the probability of falling within interval (–∞,–2.5) or (2.5,+∞) is about 0.05. Utilize the property, θ = –2.5 or θ = 2.5 can be approximately regarded as an infinity point. When h = –2.5, that is, θ = –2.5 regarded as the zero point of X. Since the length of interval (–2.5,2.5) is 5, if the test score is set to 100 points, 00092_PSISDG12506_1250622_page_3_3.jpg can be taken. Take 00092_PSISDG12506_1250622_page_3_4.jpg, so under the above assumption, X ~ N(50,400), and the value range of X is [0,100]. Here X is called the ability score. The ability score obtained through the above method conversion also has all the properties of the ability parameter. Specifically, ability scores have the following characteristics7.

4.1

Ability score is consistent estimation

This is because the ability parameter is a consistent estimation, the ability score converted from the ability parameter must also be a consistent estimation. If the true value of an examiner’s ability score is X0. 00092_PSISDG12506_1250622_page_3_5.jpg is the ability score estimate of the examiner. Then, when the test sample size n → ∞, 00092_PSISDG12506_1250622_page_3_6.jpg will converge to the true value X0 in probability. This property provides a theoretical basis for accurate estimation of capacity fraction8.

4.2

Ability score is a “constant score”

Since the ability parameters are invariant, the ability scores obtained from the conversion of ability parameters also have the property of invariance. Because of this, the subjects’ ability score has nothing to do with the test. A participant can participate in high difficulty test or low difficulty test9. Except for the sampling error, the same ability score estimation will be obtained. Test scores in the general sense do not have such properties.

4.3

Normal distribution of ability scores

Since the capacity parameters are normally distributed, the ability scores must also be normally distributed. Because the ability parameters or ability scores obtained by the above method are invariant parameters, in the case of large sample random sampling, the ability scores from different tests can be directly compared because they are equivalent. And in the sense of ability score, the college entrance examination scores for different years will not fluctuate too much because of the change of test difficulty.

5.

APPLICATION EXAMPLES

As an application of the above method, we will discuss an example below1.

Table 1 is the estimated value of item parameters obtained from a practical test. There are 24 test questions in this test, including 8 selection questions, 8 filling questions, and 8 multi-level scoring questions. A total of 1226 participants participated in the test. The average score of the test is 60, and the standard deviation is 24.44.

Table 1.

Parameter estimates for a test item.

Item123456789101112
Difficulty0.840.770.350.640.810.810.660.540.860.700.620.70
Discrimination0.500.480.570.420.470.630.490.430.490.460.460.39
Item131415161718192021222324
Difficulty0.790.700.620.430.790.660.650.340.500.540.540.29
Discrimination0.450.510.520.320.470.650.690.590.700.690.690.68

From Figure 1, we can see that in this test, the distribution of test scores showed a gentle upward trend, with a certain proportion of examinee in each score range. This shows that the distribution of examinee with high or low scores in the test is relatively uniform, and the proportion of examinee with high scores is relatively large. Among them, examinee with scores higher than 60 accounted for 53% of the total, and examinee with scores higher than 80 accounted for 26.6% of the total. In the range of 80 to 90, the proportion of examinee reached the peak. On the whole, the distribution of scores showed a negative skew trend.

Figure 1.

The distribution of test scores for 1226 candidates in the test.

00092_PSISDG12506_1250622_page_4_1.jpg

From the distribution of ability scores shown in Figure 2, we can see that there is a great difference between the distribution of ability scores and test scores. Also, we can see that examinees with scores above 60 account for 24.6% of the total, which is 27.4% points lower than 53% of the test scores. The examinees who scored more than 80 accounted for 2.6% of the total. The difference from the test score of 26.6% is 24% points. Its peak value appears in the range of 50 to 60 points, and the distribution of ability score shows a significant normal distribution trend, with a variance of about 400.

Figure 2.

The ability score distribution diagram of 1226 examinees after converting the test score of each subject into ability score according to the method described in this paper.

00092_PSISDG12506_1250622_page_4_2.jpg

It is worth noting that the distribution of test scores is rarely normally distributed, because test scores are strictly dependent on the difficulty of the test, and the distribution of test scores is not the same for different difficulty tests.

Because the ability score has nothing to do with the difficulty of the test, it is always normal in the case of large sample random sampling10.

The ability score not only has the intuitive and easy to understand advantages of the test score, but also retains all the large sample properties of the ability parameters. Therefore, the ability score is a very ideal scoring method.

6.

TESTING ERROR

In the classical test theory, the test error is mainly described by reliability and test standard error. However, reliability is a general and rough indicator, which only roughly reflects the average difference between the real score and the test score of the subjects, and cannot reflect the test error of the examinees at different abilities. Therefore, reliability and test standard error are not a good way to characterize the test error. In this paper, we will use another method to characterize the test error.

Assuming that a test is composed of n items, the ability score estimates of subjects with ability score X0 on n items are X1, X2, L, Xn, 00092_PSISDG12506_1250622_page_5_1.jpg is the average score, according to mathematical statistics, when n → ∞, 00092_PSISDG12506_1250622_page_5_2.jpg, let 00092_PSISDG12506_1250622_page_5_3.jpg, Z ~ N(0, 1) can be obtained. Let confidence level α = 0.05, 00092_PSISDG12506_1250622_page_5_4.jpg can be obtained after calculation, therefore, in the 95% probability meaning, 00092_PSISDG12506_1250622_page_5_5.jpg, that is 00092_PSISDG12506_1250622_page_5_6.jpg. So the 95% confidence interval is 00092_PSISDG12506_1250622_page_5_7.jpg.Using the above formula, the confidence interval of the estimated ability score can be obtained. The confidence interval for the estimated capacity score shown in Figure 1 is shown in Figure 3.

Figure 3.

The confidence interval of the estimated capacity score.

00092_PSISDG12506_1250622_page_5_8.jpg

Note: The abscissa represents the estimated value of ability score, and the ordinate represents the true value of ability score. The lower curve represents the left-end curve of the confidence interval, and the upper curve represents the right-end curve of the confidence interval.

Figure 3 is obtained as follows11:

  • Firstly, the examinees’ ability scores are divided into 50 intervals on average. Because there are 1226 candidates in this test, there are 24 candidates in each interval on average.

  • Standard deviation of ability scores for each interval S;

  • Taking the middle value of the ability score of each interval as the abscissa and 00092_PSISDG12506_1250622_page_6_1.jpg as the ordinate to trace points in the rectangular coordinate system, where 00092_PSISDG12506_1250622_page_6_2.jpg is the average number of ability scores of subjects in each ability score range, and n is the sample number of test questions, in this example, n = 24, S is the sample standard deviation of the ability score of the subjects in each ability score interval;

  • Connect the trace points mentioned in the previous step with a smooth curve to obtain the left end curve of the confidence interval of the capability score. The right endpoint curve of the confidence interval can be obtained by imitating this.

In this case, due to the small number of samples in each ability interval, the error between the standard deviation of the sample S and the overall standard deviation of the interval is large, so the error between the confidence interval is also large. If you can guarantee enough samples in ability interval, the resulting confidence interval will be more accurate.

From the strip region of Figure 3, we can see that in this test, for the examinees whose ability scores are between 50 and 70, the estimation error is the smallest. It is noted that the difficulty of most examinees in this test is between 0.5 and 0.7, while the test error is the smallest between 50 and 70, indicating that the test error is closely related to the difficulty of test items.

In addition, from the expression of the left and right endpoints of the confidence interval, we can see that when the standard deviation is constant, the length of the confidence interval will become shorter with the increase of the number of test items n. When n approaches infinity, the length of the confidence interval will approach zero. This shows that appropriate increase in the number of test items can effectively improve the test accuracy. Thus, the confidence interval can not only better reflect the test error at different abilities, but also better reflect the relationship between the test error and the number of test items1.

In the classical measurement theory, the estimation of measurement error is mainly described by reliability and test standard error. From the approximate calculation formula of reliability 00092_PSISDG12506_1250622_page_6_3.jpg, the reliability of this test can be calculated as r = 0.85, and the test standard error is 00092_PSISDG12506_1250622_page_6_4.jpg.

Compared with the confidence interval, the method of using test reliability or test standard error to describe test error has the following defects: (1) Test reliability or test standard error only reflects the average difference between the test scores and the real scores of all examinees in the test. In this case, which is too rough and general, and is far less detailed than the confidence interval. (2) Test reliability or test standard error is based on strictly parallel test. In practice, strictly parallel test does not exist, so the practical significance of test reliability or test standard error is greatly reduced. (3) Test reliability or test standard error does not reflect the relationship between test number (even if strictly parallel tests do exist) and test item number and test error. That is to say, even if strictly parallel tests do exist, test reliability or test standard error basically remains unchanged no matter how many tests are conducted or how many test items are added. Therefore, test reliability or test standard error is a rigid indicator.

Due to the above defects in test reliability or test standard error, it is not a good way to describe test error with test reliability and test standard error. Relatively speaking, it is a more scientific and reliable method to describe test error with confidence interval.

7.

CONCLUSION

The characteristic of this paper is to give a new evaluation method under the guidance of item response theory (IRT). Since the ability parameter or ability score defined according to the method described in this paper is a constant parameter, in the case of large sample random sampling, the ability scores from different tests can be directly compared. For example, for the college entrance examination scores from different years, it can be directly compared in the sense of ability score, and the error is also controllable. Its popularization and application are of great significance to improve the scientificity and reliability of various tests.

ACKNOWLEDGEMENTS

This research was supported by Hunan provincial key project of educational information technology research (No. HNETR22042) and University project of Hunan Association of educational science researchers (No. XJKX22B206).

REFERENCES

[1] 

Du, W., “Scoring of subjective questions in mathematics test,” J. of Mathematics Education, 15 (3), 87 –88 (2006). Google Scholar

[2] 

Peng, J., “Research on the text similarity algorithms in automatic scoring of subjective questions,” in Journal of Physics Conference Series, 042039 (2021). Google Scholar

[3] 

Gomaa, W. H., and Fahmy, A. A., “Automatic scoring for answers to Arabic test questions,” Computer Speech and Language, 28 (4), 833 –857 (2014). https://doi.org/10.1016/j.csl.2013.10.005 Google Scholar

[4] 

Niemoller, J., and Washington, N., “Subjective perception scoring: Psychological interpretation of network usage metrics in order to predict user satisfaction,” Annals of Telecommunications, (2017). https://doi.org/10.1007/s12243-017-0575-6 Google Scholar

[5] 

Mai, F., Yue, X., and Zhao, Z., “Research on Chinese subjective questions scoring algorithm based on natural language processing,” in International Conference on Advanced Computer Control, IEEE, (2010). Google Scholar

[6] 

Zhou, Z., Hou, K. H., Yao, H. F., and Zhang, H., “Research on automatic scoring system of subjective questions based on TF-IDF and LSI model,” Computer Engineering and Software, (2019). Google Scholar

[7] 

Polat, M., Toraman, C., and Turhan, N. S., “Reliability analysis of PISA 2018 reading literacy student questionnaire based on item response theory (IRT): Turkey sample,” (2022). Google Scholar

[8] 

Zhu, G., Zhou, Y., Zhou, F., Wu, M., and Wang, J., “Proactive personality measurement using item response theory and social media text mining,” Frontiers in Psychology, 12 705005 (2021). https://doi.org/10.3389/fpsyg.2021.705005 Google Scholar

[9] 

Hambleton, R. K. and Jones, R. W., “Comparison of classical test theory and item response theory and their applications to test development,” Educational Measurement Issues and Practice, 12 (3), 38 –47 (2010). https://doi.org/10.1111/j.1745-3992.1993.tb00543.x Google Scholar

[10] 

Hambleton, R. K., and Jones, R. W., “An NCME instructional module on comparison of classical test theory and item response theory and their applications to test development,” Educational Measurement Issues and Practice, 12 (3), 38 –47 (2010). https://doi.org/10.1111/j.1745-3992.1993.tb00543.x Google Scholar

[11] 

Berg, J. L., Durant, J., Banks, S. J., and Miller, J. B., “Estimates of premorbid ability in a neurodegenerative disease clinic population: Comparing the test of premorbid functioning and the wide range achievement test,” Clinical Neuropsychologist, 1 –11 (2016). Google Scholar
© (2022) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Jin Li, Shiru Lan, Zhanglin Li, Wenjiu Du, Yong Wu, and Xianyi Yao "An objective method of measuring students’ achievement", Proc. SPIE 12506, Third International Conference on Computer Science and Communication Technology (ICCSCT 2022), 1250622 (28 December 2022); https://doi.org/10.1117/12.2661817
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Reliability

Statistical analysis

Error analysis

Lithium

Analytical research

Estimation theory

Information technology

Back to Top