• Title/Summary/Keyword: Machine System

Search Result 8,774, Processing Time 0.052 seconds

A Study on the Drug Classification Using Machine Learning Techniques (머신러닝 기법을 이용한 약물 분류 방법 연구)

  • Anmol Kumar Singh;Ayush Kumar;Adya Singh;Akashika Anshum;Pradeep Kumar Mallick
    • Advanced Industrial SCIence
    • /
    • v.3 no.2
    • /
    • pp.8-16
    • /
    • 2024
  • This paper shows the system of drug classification, the goal of this is to foretell the apt drug for the patients based on their demographic and physiological traits. The dataset consists of various attributes like Age, Sex, BP (Blood Pressure), Cholesterol Level, and Na_to_K (Sodium to Potassium ratio), with the objective to determine the kind of drug being given. The models used in this paper are K-Nearest Neighbors (KNN), Logistic Regression and Random Forest. Further to fine-tune hyper parameters using 5-fold cross-validation, GridSearchCV was used and each model was trained and tested on the dataset. To assess the performance of each model both with and without hyper parameter tuning evaluation metrics like accuracy, confusion matrices, and classification reports were used and the accuracy of the models without GridSearchCV was 0.7, 0.875, 0.975 and with GridSearchCV was 0.75, 1.0, 0.975. According to GridSearchCV Logistic Regression is the most suitable model for drug classification among the three-model used followed by the K-Nearest Neighbors. Also, Na_to_K is an essential feature in predicting the outcome.

Estimation of GARCH Models and Performance Analysis of Volatility Trading System using Support Vector Regression (Support Vector Regression을 이용한 GARCH 모형의 추정과 투자전략의 성과분석)

  • Kim, Sun Woong;Choi, Heung Sik
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.2
    • /
    • pp.107-122
    • /
    • 2017
  • Volatility in the stock market returns is a measure of investment risk. It plays a central role in portfolio optimization, asset pricing and risk management as well as most theoretical financial models. Engle(1982) presented a pioneering paper on the stock market volatility that explains the time-variant characteristics embedded in the stock market return volatility. His model, Autoregressive Conditional Heteroscedasticity (ARCH), was generalized by Bollerslev(1986) as GARCH models. Empirical studies have shown that GARCH models describes well the fat-tailed return distributions and volatility clustering phenomenon appearing in stock prices. The parameters of the GARCH models are generally estimated by the maximum likelihood estimation (MLE) based on the standard normal density. But, since 1987 Black Monday, the stock market prices have become very complex and shown a lot of noisy terms. Recent studies start to apply artificial intelligent approach in estimating the GARCH parameters as a substitute for the MLE. The paper presents SVR-based GARCH process and compares with MLE-based GARCH process to estimate the parameters of GARCH models which are known to well forecast stock market volatility. Kernel functions used in SVR estimation process are linear, polynomial and radial. We analyzed the suggested models with KOSPI 200 Index. This index is constituted by 200 blue chip stocks listed in the Korea Exchange. We sampled KOSPI 200 daily closing values from 2010 to 2015. Sample observations are 1487 days. We used 1187 days to train the suggested GARCH models and the remaining 300 days were used as testing data. First, symmetric and asymmetric GARCH models are estimated by MLE. We forecasted KOSPI 200 Index return volatility and the statistical metric MSE shows better results for the asymmetric GARCH models such as E-GARCH or GJR-GARCH. This is consistent with the documented non-normal return distribution characteristics with fat-tail and leptokurtosis. Compared with MLE estimation process, SVR-based GARCH models outperform the MLE methodology in KOSPI 200 Index return volatility forecasting. Polynomial kernel function shows exceptionally lower forecasting accuracy. We suggested Intelligent Volatility Trading System (IVTS) that utilizes the forecasted volatility results. IVTS entry rules are as follows. If forecasted tomorrow volatility will increase then buy volatility today. If forecasted tomorrow volatility will decrease then sell volatility today. If forecasted volatility direction does not change we hold the existing buy or sell positions. IVTS is assumed to buy and sell historical volatility values. This is somewhat unreal because we cannot trade historical volatility values themselves. But our simulation results are meaningful since the Korea Exchange introduced volatility futures contract that traders can trade since November 2014. The trading systems with SVR-based GARCH models show higher returns than MLE-based GARCH in the testing period. And trading profitable percentages of MLE-based GARCH IVTS models range from 47.5% to 50.0%, trading profitable percentages of SVR-based GARCH IVTS models range from 51.8% to 59.7%. MLE-based symmetric S-GARCH shows +150.2% return and SVR-based symmetric S-GARCH shows +526.4% return. MLE-based asymmetric E-GARCH shows -72% return and SVR-based asymmetric E-GARCH shows +245.6% return. MLE-based asymmetric GJR-GARCH shows -98.7% return and SVR-based asymmetric GJR-GARCH shows +126.3% return. Linear kernel function shows higher trading returns than radial kernel function. Best performance of SVR-based IVTS is +526.4% and that of MLE-based IVTS is +150.2%. SVR-based GARCH IVTS shows higher trading frequency. This study has some limitations. Our models are solely based on SVR. Other artificial intelligence models are needed to search for better performance. We do not consider costs incurred in the trading process including brokerage commissions and slippage costs. IVTS trading performance is unreal since we use historical volatility values as trading objects. The exact forecasting of stock market volatility is essential in the real trading as well as asset pricing models. Further studies on other machine learning-based GARCH models can give better information for the stock market investors.

Life Cylcle Assessment (LCA) on Rice Production Systems: Comparison of Greenhouse Gases (GHGs) Emission on Conventional, Without Agricultural Chemical and Organic Farming (쌀 생산체계에 대한 영농방법별 전과정평가: 관행농, 무농약, 유기농법별 탄소배출량 비교)

  • Ryu, Jong-Hee;Kwon, Young-Rip;Kim, Gun-Yeob;Lee, Jong-Sik;Kim, Kye-Hoon;So, Kyu-Ho
    • Korean Journal of Soil Science and Fertilizer
    • /
    • v.45 no.6
    • /
    • pp.1157-1163
    • /
    • 2012
  • This study was performed a comparative life cycle assessment (LCA) among three rice production systems in order to analyze the difference of greenhouse gases (GHGs) emissions and environment impacts. Its life cycle inventory (LCI) database (DB) was established using data obtained from interview with conventional, without agricultural chemical and organic farming at Gunsan and Iksan, Jeonbuk province in 2011. According to the result of LCI analysis, $CO_2$ was mostly emitted from fertilizer production process and rice cropping phase. $CH_4$ and $N_2O$ were almost emitted from rice cultivation phase. The value of carbon footprint to produce 1 kg rice (unhulled) on conventional rice production system was 1.01E+00 kg $CO_2$-eq. $kg^{-1}$ and it was the highest value among three rice production systems. The value of carbon footprints on without agricultural chemical and organic rice production systems were 5.37E-01 $CO_2$-eq. $kg^{-1}$ and 6.58E-01 $CO_2$-eq. $kg^{-1}$, respectively. Without agricultural chemical rice production system whose input amount was the smallest had the lowest value of carbon footprint. Although the yield of rice from organic farming was the lowest, its value of carbon footprint less than that of conventional farming. Because there is no compound fertilizer inputs in organic farming. Compound fertilizer production and methane emission during rice cultivation were the main factor to GHGs emission in conventional and without agricultural chemical rice production systems. In organic rice production system, the main factors to GHGs emission were using fossil fuel on machine operation and methane emission from rice paddy field.

A Study on the Neutron in Radiation Treatment System and Related Facility (방사선치료 장치 및 관련시설에서의 산란 중성자에 관한 연구)

  • Kim Dae-Sup;Kim Jeong-Man;Lee Hee-Seok;Lim Ra-Seung;Kim You-Hyun
    • The Journal of Korean Society for Radiation Therapy
    • /
    • v.17 no.2
    • /
    • pp.141-145
    • /
    • 2005
  • Purpose : It is known that the neutron is generally generated from the photon, its energy is larger than 10 MV. The neutron is leaked in the container inspection system installed at the customs though its energy is below 9 MV. It is needed that the spacial effect of the neutrons released from radiation treatment machine, linac, installed in the medical canter. Materials and Methods : The medical linear accelerator(Clinac 1800, varian, USA) was used in the experiment. Measuring neutron was used bubble detector(Bubble detector, BDPND type, BTI, Canada) which was created bubble by neutron. The bubble detector is located on the medical linear accelerator outskirt in three different distance, 30, 50, 120 cm and upper, lower four point from the iso-center. In addition, for effect on protect material we have measured eight points which are 50 cm distance from iso-center. The SAD(source-axis-distance), distance from photon source to iso-center, is adjusted to 100 cm and the field size is adjusted to $15{\times}15cm^2$. Irradiate 20 MU and calculate the dose rate in mrem/MU by measuring the number of bubble. Results : The neutron is more detected at 5 position in 30, 50 cm, 7 position in 120 cm and with wedge, and 2 position without mount. Conclusion : Though detection position is laid in the same distance in neutron measurement, the different value is shown in measuring results. Also, neutron dose is affected by the additional structure, the different value is obtained in each measurement positions. So, it is needed to measure and evaluate the neutron dose in the whole space considering the effect of the distance, angular distribution and additional structure.

  • PDF

Evaluation of the Accuracy for Respiratory-gated RapidArc (RapidArc를 이용한 호흡연동 회전세기조절방사선치료 할 때 전달선량의 정확성 평가)

  • Sung, Jiwon;Yoon, Myonggeun;Chung, Weon Kuu;Bae, Sun Hyun;Shin, Dong Oh;Kim, Dong Wook
    • Progress in Medical Physics
    • /
    • v.24 no.2
    • /
    • pp.127-132
    • /
    • 2013
  • The position of the internal organs can change continually and periodically inside the body due to the respiration. To reduce the respiration induced uncertainty of dose localization, one can use a respiratory gated radiotherapy where a radiation beam is exposed during the specific time of period. The main disadvantage of this method is that it usually requests a long treatment time, the massive effort during the treatment and the limitation of the patient selection. In this sense, the combination of the real-time position management (RPM) system and the volumetric intensity modulated radiotherapy (RapidArc) is promising since it provides a short treatment time compared with the conventional respiratory gated treatments. In this study, we evaluated the accuracy of the respiratory gated RapidArc treatment. Total sic patient cases were used for this study and each case was planned by RapidArc technique using varian ECLIPSE v8.6 planning machine. For the Quality Assurance (QA), a MatriXX detector and I'mRT software were used. The results show that more than 97% of area gives the gamma value less than one with 3% dose and 3 mm distance to agreement condition, which indicates the measured dose is well matched with the treatment plan's dose distribution for the gated RapidArc treatment cases.

A Study on the Dental Service Statifation of Cityizens in Deajeon (대전시 시민의 치과의료서비스에 관한 만족도 조사연구)

  • Sung, Bo-Kyun
    • Journal of Korean society of Dental Hygiene
    • /
    • v.8 no.4
    • /
    • pp.19-30
    • /
    • 2008
  • This study reached the following conclusions as a result of carrying out the questionnaire survey of self-descriptions for the satisfaction after the citizens of Daejon uses the dental clinics, in order to identify the factors of satisfaction to the medical services of such dental clinics to be utilized in the patient management by dental hygienists, provide the basic data to provide the medical services desired by patients. 1. 43.9% men responded to the facilities and 56.1% women to the atmosphere for the standards of selection of dental clinics by general characteristic, and the college graduates or more to the kindness (38.2%), high-school graduates (43.2%) and middle-school graduates (25.9%) or less to the close distance for the level of educational attainment (p=0.009), which was meant to have a statistical significance. 2. The execution of reservation system for the dental clinics showed 54.7%, the reserved time was observed upon the execution of such reservation system, the dental clinics where they practice such system were 40.6%, and the confirmation methods was done through the telephone with 62.5%. 3. The experience of fear upon the dental treatment showed 74.6%. The type of fear showed the machine sound (48.7%) for men and cry of others for women (70.8%) at the highest. 70% of those under 30 at the age responded to the sharp instruments at the highest. 83.3% of Yousung-gu showed the highest by responding to the cry of others for the residential areas. The statistically significant difference was shown in both the age and residential area (p=0.000). 4. Women showed higher in the distribution of gender for the sterilization of instruments for the external satisfaction of dental clinics(p=0.000) and those under 30 at the age showed the highest with 2.98${\pm}$0.95(p=0.001). Seo-gu (3.48${\pm}$0.77) was the highest for the residential area (p=0.000), and there was statistically significant differences in the gender, age and residential area. 5. Men showed higher satisfaction than women in the clean state and the statistically significant differences were shown (p=0.000) at the age as the high satisfaction was shown for those under 30 at the age (2.35${\pm}$0.79), those having the income not less than 10 million won and not more than 20 million won (2.43${\pm}$0.78), and Seo-gu (2.63 ${\pm}$0.69) for the residential area. 6. For the internal satisfaction of dental clinic by users for the medical services in the dental clinics, 61.1% women responded to no in the ability of solving the inconvenience in the service process, and showed low ability of solving the inconvenience from 30 at the age (26.2%) and by responding to Dong-gu (22.1%) for the residential area, showing statically significant differences(p=0.000). For the re-use of dental clinics, 46.6% men (p=0.043) for the gender, 24.3% under 30 at the age and 22.9% of Dong-gu for the residential area responded to the re-use, showing statistically significant differences for the gender and residential area (p=0.000). 7. The dissatisfaction showed a high rate of 69.5% for the satisfaction to the medical services of dental clinics. 46.2% men responded to the pain and women to the feeling of foreign substance for the reason of dissatisfaction while those under 30 at the age showed 55.6% for others, those between 50 and 59 41.7% for the feeling of foreign substance. 86.3% carried out the education for cautions after the treatments and most people turned out that they do not carry out the continuous health management of mouth as 20.5% responded to that they carry out such health management.

  • PDF

Shear bond strength of Universal bonding systems to Ni-Cr alloy (니켈-크롬 합금에 대한 다용도 접착 시스템의 전단결합강도)

  • Song, So-Yeon;Son, Byung-Wha;Kim, Jong-Yeob;Shin, Sang-Wan;Lee, Jeong-Yol
    • The Journal of Korean Academy of Prosthodontics
    • /
    • v.53 no.4
    • /
    • pp.295-300
    • /
    • 2015
  • Purpose: The aim of this study was to evaluate the shear bond strength between Ni-Cr alloy and composite resin using universal adhesive systems coMPared to conventional method using metal primers. Materials and methods: For this study, a total of 120 cast commercial Ni-Cr alloy (Vera Bond 2V) disks were embedded in acrylic resin, and their surfaces were smoothed with silicon carbide papers and airborne-particle abrasion. Specimens of each metal were divided into 6 groups based on the combination of metal primers (Metal primer II, Alloy primer, Metal & Zirconia primer, MKZ primer) and universal adhesive systems (Single Bond Universal, All Bond Universal). All specimens were stored in distilled water at $37^{\circ}C$ for 24 hours. Shear bond strength testing was performed with a universal testing machine at a cross head speed of 1 m/min. Data (MPa) were analyzed using one-way ANOVA and the post hoc Tukey's multiple comparison test (${\alpha}$=.05). Results: There were significant differences between Single Bond Universal, All Bond Universal, Metal Primer II and Alloy Primer, MKZ Primer, Metal & Zirconia Primer (P<.001). Conclusion: Universal Adhesive system groups indicated high shear bond strength value bonded to Ni-Cr alloy than that of conventional system groups using primers except Metal Primer II. Within the limitations of this study, improvement of universal adhesive systems which can be applied to all types of restorations is recommended especially non-precious metal alloy. More research is needed to evaluate the effect of silane inclusion or exclusion in universal adhesive systems.

Evaluation of static fracture resistances and patterns of pulpless tooth restored with poly-ether-ketone-ketone (PEKK) post (Poly-ether-ketone-ketone (PEKK) 포스트로 수복한 근관 치료 치아의 정적 파절 저항성 및 파절 형태에 관한 평가)

  • Park, Ha Eun;Lee, Cheol Won;Lee, Won Sup;Yang, Sung Eun;Lee, Su Young
    • The Journal of Korean Academy of Prosthodontics
    • /
    • v.57 no.2
    • /
    • pp.127-133
    • /
    • 2019
  • Purpose: The purpose of present study was to investigate fracture strength and mode of failure of endodontically treated teeth restored with metal cast post-core system, prefabricated fiber post system, and newly introduced polyetherketoneketone (PEKK) post-core system. Materials and methods: A total of 21 mandibular premolar were randomly grouped into 3 groups of 7 each according to the post material. Group A was for metal cast post core; Group B for prefabricated glass fiber post and resin core; and Group C for milled PEKK post cores. All specimens were restored with metal crown. The fracture strength of each specimen was measured by applying a static load of 135-degree to the tooth at 2 mm/min crosshead speed using a universal testing machine. After the fracture strength measurement, the mode of failure was observed. The results were analyzed using Kruscal-Wallis test and post hoc Mann-Whitney U test at confidence interval ${\alpha}=.05$. Results: Fracture resistance of PEKK post core was lower than those of cast metal post and fiber reinforced post with composite resin core. In the aspect of fracture mode most of the root fracture occurred in the metal post core, whereas the post detachment occurred mainly in the fiber reinforced post. In the case of PEKK post core, teeth and post were fractured together. Conclusion: It is necessary to select appropriate materials of post for extensively damaged teeth restoration and clinical application of the PEKK post seems to require more research on improvement of strength.

Video Analysis System for Action and Emotion Detection by Object with Hierarchical Clustering based Re-ID (계층적 군집화 기반 Re-ID를 활용한 객체별 행동 및 표정 검출용 영상 분석 시스템)

  • Lee, Sang-Hyun;Yang, Seong-Hun;Oh, Seung-Jin;Kang, Jinbeom
    • Journal of Intelligence and Information Systems
    • /
    • v.28 no.1
    • /
    • pp.89-106
    • /
    • 2022
  • Recently, the amount of video data collected from smartphones, CCTVs, black boxes, and high-definition cameras has increased rapidly. According to the increasing video data, the requirements for analysis and utilization are increasing. Due to the lack of skilled manpower to analyze videos in many industries, machine learning and artificial intelligence are actively used to assist manpower. In this situation, the demand for various computer vision technologies such as object detection and tracking, action detection, emotion detection, and Re-ID also increased rapidly. However, the object detection and tracking technology has many difficulties that degrade performance, such as re-appearance after the object's departure from the video recording location, and occlusion. Accordingly, action and emotion detection models based on object detection and tracking models also have difficulties in extracting data for each object. In addition, deep learning architectures consist of various models suffer from performance degradation due to bottlenects and lack of optimization. In this study, we propose an video analysis system consists of YOLOv5 based DeepSORT object tracking model, SlowFast based action recognition model, Torchreid based Re-ID model, and AWS Rekognition which is emotion recognition service. Proposed model uses single-linkage hierarchical clustering based Re-ID and some processing method which maximize hardware throughput. It has higher accuracy than the performance of the re-identification model using simple metrics, near real-time processing performance, and prevents tracking failure due to object departure and re-emergence, occlusion, etc. By continuously linking the action and facial emotion detection results of each object to the same object, it is possible to efficiently analyze videos. The re-identification model extracts a feature vector from the bounding box of object image detected by the object tracking model for each frame, and applies the single-linkage hierarchical clustering from the past frame using the extracted feature vectors to identify the same object that failed to track. Through the above process, it is possible to re-track the same object that has failed to tracking in the case of re-appearance or occlusion after leaving the video location. As a result, action and facial emotion detection results of the newly recognized object due to the tracking fails can be linked to those of the object that appeared in the past. On the other hand, as a way to improve processing performance, we introduce Bounding Box Queue by Object and Feature Queue method that can reduce RAM memory requirements while maximizing GPU memory throughput. Also we introduce the IoF(Intersection over Face) algorithm that allows facial emotion recognized through AWS Rekognition to be linked with object tracking information. The academic significance of this study is that the two-stage re-identification model can have real-time performance even in a high-cost environment that performs action and facial emotion detection according to processing techniques without reducing the accuracy by using simple metrics to achieve real-time performance. The practical implication of this study is that in various industrial fields that require action and facial emotion detection but have many difficulties due to the fails in object tracking can analyze videos effectively through proposed model. Proposed model which has high accuracy of retrace and processing performance can be used in various fields such as intelligent monitoring, observation services and behavioral or psychological analysis services where the integration of tracking information and extracted metadata creates greate industrial and business value. In the future, in order to measure the object tracking performance more precisely, there is a need to conduct an experiment using the MOT Challenge dataset, which is data used by many international conferences. We will investigate the problem that the IoF algorithm cannot solve to develop an additional complementary algorithm. In addition, we plan to conduct additional research to apply this model to various fields' dataset related to intelligent video analysis.

A New Approach to Automatic Keyword Generation Using Inverse Vector Space Model (키워드 자동 생성에 대한 새로운 접근법: 역 벡터공간모델을 이용한 키워드 할당 방법)

  • Cho, Won-Chin;Rho, Sang-Kyu;Yun, Ji-Young Agnes;Park, Jin-Soo
    • Asia pacific journal of information systems
    • /
    • v.21 no.1
    • /
    • pp.103-122
    • /
    • 2011
  • Recently, numerous documents have been made available electronically. Internet search engines and digital libraries commonly return query results containing hundreds or even thousands of documents. In this situation, it is virtually impossible for users to examine complete documents to determine whether they might be useful for them. For this reason, some on-line documents are accompanied by a list of keywords specified by the authors in an effort to guide the users by facilitating the filtering process. In this way, a set of keywords is often considered a condensed version of the whole document and therefore plays an important role for document retrieval, Web page retrieval, document clustering, summarization, text mining, and so on. Since many academic journals ask the authors to provide a list of five or six keywords on the first page of an article, keywords are most familiar in the context of journal articles. However, many other types of documents could not benefit from the use of keywords, including Web pages, email messages, news reports, magazine articles, and business papers. Although the potential benefit is large, the implementation itself is the obstacle; manually assigning keywords to all documents is a daunting task, or even impractical in that it is extremely tedious and time-consuming requiring a certain level of domain knowledge. Therefore, it is highly desirable to automate the keyword generation process. There are mainly two approaches to achieving this aim: keyword assignment approach and keyword extraction approach. Both approaches use machine learning methods and require, for training purposes, a set of documents with keywords already attached. In the former approach, there is a given set of vocabulary, and the aim is to match them to the texts. In other words, the keywords assignment approach seeks to select the words from a controlled vocabulary that best describes a document. Although this approach is domain dependent and is not easy to transfer and expand, it can generate implicit keywords that do not appear in a document. On the other hand, in the latter approach, the aim is to extract keywords with respect to their relevance in the text without prior vocabulary. In this approach, automatic keyword generation is treated as a classification task, and keywords are commonly extracted based on supervised learning techniques. Thus, keyword extraction algorithms classify candidate keywords in a document into positive or negative examples. Several systems such as Extractor and Kea were developed using keyword extraction approach. Most indicative words in a document are selected as keywords for that document and as a result, keywords extraction is limited to terms that appear in the document. Therefore, keywords extraction cannot generate implicit keywords that are not included in a document. According to the experiment results of Turney, about 64% to 90% of keywords assigned by the authors can be found in the full text of an article. Inversely, it also means that 10% to 36% of the keywords assigned by the authors do not appear in the article, which cannot be generated through keyword extraction algorithms. Our preliminary experiment result also shows that 37% of keywords assigned by the authors are not included in the full text. This is the reason why we have decided to adopt the keyword assignment approach. In this paper, we propose a new approach for automatic keyword assignment namely IVSM(Inverse Vector Space Model). The model is based on a vector space model. which is a conventional information retrieval model that represents documents and queries by vectors in a multidimensional space. IVSM generates an appropriate keyword set for a specific document by measuring the distance between the document and the keyword sets. The keyword assignment process of IVSM is as follows: (1) calculating the vector length of each keyword set based on each keyword weight; (2) preprocessing and parsing a target document that does not have keywords; (3) calculating the vector length of the target document based on the term frequency; (4) measuring the cosine similarity between each keyword set and the target document; and (5) generating keywords that have high similarity scores. Two keyword generation systems were implemented applying IVSM: IVSM system for Web-based community service and stand-alone IVSM system. Firstly, the IVSM system is implemented in a community service for sharing knowledge and opinions on current trends such as fashion, movies, social problems, and health information. The stand-alone IVSM system is dedicated to generating keywords for academic papers, and, indeed, it has been tested through a number of academic papers including those published by the Korean Association of Shipping and Logistics, the Korea Research Academy of Distribution Information, the Korea Logistics Society, the Korea Logistics Research Association, and the Korea Port Economic Association. We measured the performance of IVSM by the number of matches between the IVSM-generated keywords and the author-assigned keywords. According to our experiment, the precisions of IVSM applied to Web-based community service and academic journals were 0.75 and 0.71, respectively. The performance of both systems is much better than that of baseline systems that generate keywords based on simple probability. Also, IVSM shows comparable performance to Extractor that is a representative system of keyword extraction approach developed by Turney. As electronic documents increase, we expect that IVSM proposed in this paper can be applied to many electronic documents in Web-based community and digital library.