• Title/Summary/Keyword: 서답형 답안

Search Result 11, Processing Time 0.02 seconds

Korean Automated Scoring System for Supply-Type Items using Semi-Supervised Learning (준지도학습 방법을 이용한 한국어 서답형 문항 자동채점 시스템)

  • Cheon, Min-Ah;Seo, Hyeong-Won;Kim, Jae-Hoon;Noh, Eun-Hee;Sung, Kyung-Hee;Lim, EunYoung
    • Annual Conference on Human and Language Technology
    • /
    • 2014.10a
    • /
    • pp.112-116
    • /
    • 2014
  • 서답형 문항은 학생들의 종합적인 사고능력을 판단하는데 매우 유용하지만 채점할 때, 시간과 비용이 매우 많이 소요되고 채점자의 공정성을 확보해야 하는 어려움이 있다. 이러한 문제를 개선하기 위해 본 논문에서는 서답형 문항에 대한 자동채점 시스템을 제안한다. 본 논문에서 제안하는 시스템은 크게 언어 처리 단계와 채점 단계로 나뉜다. 첫 번째로 언어 처리 단계에서는 형태소 분석과 같은 한국어 정보처리 시스템을 이용하여 학생들의 답안을 분석한다. 두 번째로 채점 단계를 진행하는데 이 단계는 아래와 같은 순서로 진행된다. 1) 첫 번째 단계에서 분석 결과가 완전히 일치하는 답안들을 하나의 유형으로 간주하여 각 유형에 속한 답안의 빈도수가 높은 순서대로 정렬하여 인간 채점자가 고빈도 학생 답안을 수동으로 채점한다. 2) 현재까지 채점된 결과와 모범답안을 학습말뭉치로 간주하여 자질 추출 및 자질 가중치 학습을 수행한다. 3) 2)의 학습 결과를 토대로 미채점 답안들을 군집화하여 분류한다. 4) 분류된 결과 중에서 신뢰성이 높은 채점 답안에 대해서 인간 채점자가 확인하고 학습말뭉치에 추가한다. 5) 이와 같은 방법으로 미채점 답안이 존재하지 않을 때까지 반복한다. 제안된 시스템을 평가하기 위해서 2013년 학업성취도 평가의 사회(중3) 및 국어(고2) 과목의 서답형 문항을 사용하였다. 각 과목에서 1000개의 학생 답안을 추출하여 채점시간과 정확률을 평가하였다. 채점시간을 전체적으로 약 80% 이상 줄일 수 있었고 채점 정확률은 사회 및 국어 과목에 대해 각각 98.7%와 97.2%로 나타났다. 앞으로 자동 채점 시스템의 성능을 개선하고 인간 채점자의 집중도를 높일 수 있도록 인터페이스를 개선한다면 국가수준의 대단위 평가에 충분히 활용할 수 있을 것으로 생각한다.

  • PDF

Concept-based Automatic Scoring System for Korean Free-text or Constructed Answers (개념 기반 한국어 서답형 답안의 자동채점 시스템)

  • Park, Il-Nam;Noh, Eun-Hee;Sim, Jae-Ho;Kim, Myung-Hwa;Kang, Seung-Shik
    • Annual Conference on Human and Language Technology
    • /
    • 2012.10a
    • /
    • pp.69-72
    • /
    • 2012
  • 본 논문은 한국어 서답형(단어, 구 수준) 문항 유형을 분석하고 실제 채점자가 채점 기준표를 보고 채점하는 방법을 컴퓨터가 인식할 수 있도록 정답 템플릿을 설계 및 개념 정의를 하여 한국어 서답형에 특화된 자동채점 시스템 방법을 제시한다. 본 시스템을 사용하여 1000개의 학생 답안지에 대한 유형 가지수 500개 이하의 2011년도 학업성취도 평가 과학 6개 문항에 대하여 채점 기준표 내용을 정답 템플릿으로 작성한 뒤 250개 학생 답안을 학습데이터로, 정답 템플릿을 업데이트로 사용, 750개 학생 답안에 대하여 자동채점한 결과, 평균 카파계수 0.84라는 수치로서 실제 사람 채점 결과와 거의 완벽히 일치라는 결과를 얻었다.

  • PDF

Answer Template Description for Automatic Scoring of Korean Free-text or Constructed Answers (한국어 서답형 자동채점을 위한 정답 템플릿 기술 방법)

  • Park, Il-Nam;Noh, Eun-Hee;Sim, Jae-Ho;Kim, Myung-Hwa;Kang, Seung-Shik
    • Annual Conference on Human and Language Technology
    • /
    • 2012.10a
    • /
    • pp.138-141
    • /
    • 2012
  • 한국어 서답형 문항의 자동채점 프로그램을 개발하기 위해서는 모범답안, 오답, 부분점수 부여를 위한 세부적인 내용을 채점 기준표로 기술해야 한다. 자동채점에 필요한 구체적인 사항들을 기술하기 위하여 XML 형식으로 정답 템플릿을 정의하였다. 채점에 필요한 내용을 단위 개념으로 정의하고 이를 컴퓨터가 엑세스 가능한 형태의 정답 템플릿을 설계하였다. 정답 템플릿 형식에 맞게 편리하게 템플릿을 작성할 수 있는 작성 도구를 이용하여 학업 성취도평가 각 문항에 대한 채점 기준표를 정답 템플릿으로 작성하여 채점기준표를 작성하는 실험을 수행하였다.

  • PDF

Semi-Automatic Scoring for Short Korean Free-Text Responses Using Semi-Supervised Learning (준지도학습 방법을 이용한 한국어 서답형 문항 반자동 채점)

  • Cheon, Min-Ah;Seo, Hyeong-Won;Kim, Jae-Hoon;Noh, Eun-Hee;Sung, Kyung-Hee;Lim, EunYoung
    • Korean Journal of Cognitive Science
    • /
    • v.26 no.2
    • /
    • pp.147-165
    • /
    • 2015
  • Through short-answer questions, we can reflect the depth of students' understanding and higher-order thinking skills. Scoring for short-answer questions may take long time and may be an issue on consistency of grading. To alleviate such the suffering, automated scoring systems are widely used in Europe and America, but are in the initial stage in research in Korea. In this paper, we propose a semi-automatic scoring system for short Korean free-text responses using semi-supervised learning. First of all, based on the similarity score between students' answers and model answers, the proposed system grades students' answers and the scored answers with high reliability have been included in the model answers through the thorough test. This process repeats until all answers are scored. The proposed system is used experimentally in Korean and social studies in Nationwide Scholastic Achievement Test. We have confirmed that the processing time and the consistency of grades are promisingly improved. Using the system, various assessment methods have got to be developed and comparative studies need to be performed before applying to school fields.

Development and Application of an Online Scoring System for Constructed Response Items (서답형 문항 온라인 채점 시스템의 개발과 적용)

  • Cho, Jimin;Kim, Kyunghoon
    • The Journal of Korean Association of Computer Education
    • /
    • v.17 no.2
    • /
    • pp.39-51
    • /
    • 2014
  • In high-stakes tests for large groups, the efficiency with which students' responses are distributed to raters and how systematic scoring procedures are managed is important to the overall success of the testing program. In the scoring of constructed response items, it is important to understand whether the raters themselves are making consistent judgments on the responses, and whether these judgments are similar across all raters in order to establish measures of rater reliability. The purpose of this study was to design, develop and carry out a pilot test of an online scoring system for constructed response items administered in a paper-and-pencil test to large groups, and to verify the system's reliability. In this study, we show that this online system provided information on the scoring process of individual raters, including intra-rater and inter-rater consistency, compared to conventional scoring methods. We found this system to be especially effective for obtaining reliable and valid scores for constructed response items.

  • PDF

A Study on the Classification of Constructed-Response Items in Geography Education: In Case of 4th Grade Items of the NAEP (지리과 서답형 문항의 주요 유형에 관한 연구 -NAEP의 지리과 4학년 문항을 사례로-)

  • Jang, Eui-Sun
    • Journal of the Korean Geographical Society
    • /
    • v.47 no.6
    • /
    • pp.934-954
    • /
    • 2012
  • Constructed-response items are useful to estimate a higher-order thinking abilities such as sense of place, graphicacy, and geographical imaginations which are aimed by geography education. This study aims to identify the conception of the constructed-response items, to classify those feasible items in geography education, and to propose some comments and suggestions for development and application in Korea. The author tries to classify the constructed-response items by analyzing the NAEP items in the formal aspect and the aspect of cognitive dimension respectively. In the formal aspect, they are classified as 'question-standalone' type and 'material-presenting' type. In the aspect of cognitive dimension, they are classified into three types as 'knowledge-requirement', 'understanding-requirement', and 'applying-requirement'. Some comments and suggestions are as follows. First, it is necessary to develope the constructed-response items that coherent to both content and cognitive dimensions. Second, it is required for students who take an examination to use a great variety of graphics, photos, thematic maps related to geography. Third, we need to inquire the rubric, grade, process to scoring because they are vital to success of the constructed-response items.

  • PDF

Exploring a Learning Progression for Eight Core Concepts of Middle School Science Using Constructed Response Items in the National Assessment of Educational Achievement (NAEA) (국가수준 학업성취도 평가의 서답형 문항을 이용한 중학교 과학 8개 핵심 개념에 대한 학습발달과정 탐색)

  • Lee, Kiyoung;Dong, Hyokwan;Choi, Wonho;Kwon, Gyeongpil;Lee, Inho;Kim, Yong-Jin
    • Journal of Science Education
    • /
    • v.41 no.3
    • /
    • pp.382-404
    • /
    • 2017
  • This study aims to explore a learning progression for eight core concepts of middle school science using constructed response items in the national assessment of educational achievement (NAEA). For this purpose, a total of 7,390 responses of the 9th graders on 8 constructed response items in NAEA conducted in 2015 were inputted by computer and the inputted answers were analyzed and classified into different levels. After completing this process, five levels were set for each construct and a construct map was created according to the levels. In this study, the construct map was written in a progression-based rubric format, which was used as a criterion for leveling the answers of all students. In addition, the Rasch model was applied to measure the scores and achievements of the respondents by means of statistical analysis to correlate the scores of the students with the achievement levels of the constructs. Lastly, a preliminary learning progression was created by revising the construct map reflecting the results of Rasch model application. This study was meaningful in that it explored the possibility of developing the learning process by using constructed response items in NAEA. However, the preliminary learning progression developed in this study is still hypothetical and inferential because it is not longitudinally traced to individual students. Therefore, it is necessary to continually revise and supplement through iterative research process.

Improving Automatic Scoring System for Korean Word or Phrase-level Free-text Answers (한국어 단어-구 수준 서답형 문항의 자동채점 프로그램의 개선 및 실용화)

  • Jang, Eun-Seo;Noh, Eun-Hee;Sung, Kyung-Hee;Lee, Sang-Ha;Kang, Seung-Shik
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2014.11a
    • /
    • pp.909-912
    • /
    • 2014
  • 최근 자연어 처리 분야에서는 한국어로 작성된 서답형 문항을 자동으로 채점하기 위한 시스템을 개발하기 위한 연구가 꾸준히 진행되고 있다. 2013년도에는 답안 정규화, 채점 기준 작성 및 편집, 자동 채점 및 채점 후처리 작업 등 자동 채점 프로세스를 수행할 수 있는 시스템을 설계 및 개발하였다. 본 논문은 2013년도에 개발된 자동 채점 시스템의 문제점을 분석하고 채점자가 보다 편리하게 채점을 수행할 수 있도록 추가 및 개선한 기능과 이를 구현하기 위해 이용한 자연어 처리 기술을 소개한다.

Automatic Scoring System for Korean Short Answers by Student Answer Analysis and Answer Template Construction (학생 답안 분석과 정답 템플릿 생성에 의한 한국어 서답형 문항의 자동채점 시스템)

  • Kang, SeungShik;Jang, EunSeo
    • KIISE Transactions on Computing Practices
    • /
    • v.22 no.5
    • /
    • pp.218-224
    • /
    • 2016
  • This paper proposes a computer-based practical automatic scoring system for Korean short answers through student answer analysis and natural language processing techniques. The proposed system reduces the overall scoring time and budget, while improving the ease-of-use to write answer templates from student answers as well as the accuracy and reliability of automatic scoring system. To evaluate the application of the automatic scoring system and compare to the human scoring process, we performed an experiment using the student answers of social science subject in 2014 National Assessment of Educational Achievement.

Automated Scoring System for Korean Short-Answer Questions Using Predictability and Unanimity (기계학습 분류기의 예측확률과 만장일치를 이용한 한국어 서답형 문항 자동채점 시스템)

  • Cheon, Min-Ah;Kim, Chang-Hyun;Kim, Jae-Hoon;Noh, Eun-Hee;Sung, Kyung-Hee;Song, Mi-Young
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.5 no.11
    • /
    • pp.527-534
    • /
    • 2016
  • The emergent information society requires the talent for creative thinking based on problem-solving skills and comprehensive thinking rather than simple memorization. Therefore, the Korean curriculum has also changed into the direction of the creative thinking through increasing short-answer questions that can determine the overall thinking of the students. However, their scoring results are a little bit inconsistency because scoring short-answer questions depends on the subjective scoring of human raters. In order to alleviate this point, an automated scoring system using a machine learning has been used as a scoring tool in overseas. Linguistically, Korean and English is totally different in the structure of the sentences. Thus, the automated scoring system used in English cannot be applied to Korean. In this paper, we introduce an automated scoring system for Korean short-answer questions using predictability and unanimity. We also verify the practicality of the automatic scoring system through the correlation coefficient between the results of the automated scoring system and those of human raters. In the experiment of this paper, the proposed system is evaluated for constructed-response items of Korean language, social studies, and science in the National Assessment of Educational Achievement. The analysis was used Pearson correlation coefficients and Kappa coefficient. Results of the experiment had showed a strong positive correlation with all the correlation coefficients at 0.7 or higher. Thus, the scoring results of the proposed scoring system are similar to those of human raters. Therefore, the automated scoring system should be found to be useful as a scoring tool.