• Title/Summary/Keyword: 데이터 분석론

Search Result 1,379, Processing Time 0.03 seconds

Prediction of Correct Answer Rate and Identification of Significant Factors for CSAT English Test Based on Data Mining Techniques (데이터마이닝 기법을 활용한 대학수학능력시험 영어영역 정답률 예측 및 주요 요인 분석)

  • Park, Hee Jin;Jang, Kyoung Ye;Lee, Youn Ho;Kim, Woo Je;Kang, Pil Sung
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.4 no.11
    • /
    • pp.509-520
    • /
    • 2015
  • College Scholastic Ability Test(CSAT) is a primary test to evaluate the study achievement of high-school students and used by most universities for admission decision in South Korea. Because its level of difficulty is a significant issue to both students and universities, the government makes a huge effort to have a consistent difficulty level every year. However, the actual levels of difficulty have significantly fluctuated, which causes many problems with university admission. In this paper, we build two types of data-driven prediction models to predict correct answer rate and to identify significant factors for CSAT English test through accumulated test data of CSAT, unlike traditional methods depending on experts' judgments. Initially, we derive candidate question-specific factors that can influence the correct answer rate, such as the position, EBS-relation, readability, from the annual CSAT practices and CSAT for 10 years. In addition, we drive context-specific factors by employing topic modeling which identify the underlying topics over the text. Then, the correct answer rate is predicted by multiple linear regression and level of difficulty is predicted by classification tree. The experimental results show that 90% of accuracy can be achieved by the level of difficulty (difficult/easy) classification model, whereas the error rate for correct answer rate is below 16%. Points and problem category are found to be critical to predict the correct answer rate. In addition, the correct answer rate is also influenced by some of the topics discovered by topic modeling. Based on our study, it will be possible to predict the range of expected correct answer rate for both question-level and entire test-level, which will help CSAT examiners to control the level of difficulties.

An investigation Of IntraFraction Motion Correction For Lung Stereotactic Body Radiation Therapy By Using IntraFraction Cone Beam Computed Tomography (폐암 환자의 정위적 체부 방사선 치료 시 IntraFraction CBCT를 이용한 치료 중 자세 오차 교정에 대한 고찰)

  • Song, Hyeong Seok;Cho, Kang Chul;Park, Hyo Kuk;Yoon, Jong Won;Cho, Jung Hee
    • The Journal of Korean Society for Radiation Therapy
    • /
    • v.31 no.1
    • /
    • pp.67-74
    • /
    • 2019
  • Purpose: The purpose is to correct for position errors caused by long treatment times. By correcting the target motion that can occur during lung SBRT using IntraFraction CBCT. Methods and materials: We analyzed retrospectively the IFM data of 14 patients with two treatment arc in the treatment plan for lung cancer with stereotactic radiotherapy. An IntraFraction Motion was applied to the Arccheck phantom to acquire the Gamma index data. Results : IntraFraction Motion during the first treatment arc is in the left-right(LR), superiorinferior(SI), anterior-posterior(AP) directions were $0.16{\pm}0.05cm$, 0.72 cm(max error), $0.2{\pm}0.14cm$, 1.26 cm, $0.24{\pm}0.08cm$, 0.82 cm and rotational directions was $0.84{\pm}0.23^{\circ}$, $2.8^{\circ}$(pitch), $0.72{\pm}0.23^{\circ}$, $2.5^{\circ}$(yaw), $0.7{\pm}0.19^{\circ}$, $2^{\circ}$(roll). IntraFraction Motion during the second treatment arc is in the LR, SI, AP directions were $0.1{\pm}0.04cm$, 0.37 cm, $0.14{\pm}0.17cm$, 2 cm, $0.12{\pm}0.04cm$, 0.5 cm and rotational directions was $0.45{\pm}0.12^{\circ}$, $1.3^{\circ}$, $0.37{\pm}0.1^{\circ}$, $1^{\circ}$, $0.35{\pm}0.1^{\circ}$, $1.2^{\circ}$. Gamma index pass rates were $82.64{\pm}10.51%$, 48.4 %. Conclusions : In this study, we examined the validity of IntraFraction Motion correction in lung SBRT and the efficiency of IntraFraction CBCT. Due to the nature of SBRT treatment, IFM may increase due to the increased treatment time. It is believed that the increase in IFM with the increase in treatment time can be improved with the use of FFF Beam and additional position correction using CBCT during treatment.

Predicting stock movements based on financial news with systematic group identification (시스템적인 군집 확인과 뉴스를 이용한 주가 예측)

  • Seong, NohYoon;Nam, Kihwan
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.3
    • /
    • pp.1-17
    • /
    • 2019
  • Because stock price forecasting is an important issue both academically and practically, research in stock price prediction has been actively conducted. The stock price forecasting research is classified into using structured data and using unstructured data. With structured data such as historical stock price and financial statements, past studies usually used technical analysis approach and fundamental analysis. In the big data era, the amount of information has rapidly increased, and the artificial intelligence methodology that can find meaning by quantifying string information, which is an unstructured data that takes up a large amount of information, has developed rapidly. With these developments, many attempts with unstructured data are being made to predict stock prices through online news by applying text mining to stock price forecasts. The stock price prediction methodology adopted in many papers is to forecast stock prices with the news of the target companies to be forecasted. However, according to previous research, not only news of a target company affects its stock price, but news of companies that are related to the company can also affect the stock price. However, finding a highly relevant company is not easy because of the market-wide impact and random signs. Thus, existing studies have found highly relevant companies based primarily on pre-determined international industry classification standards. However, according to recent research, global industry classification standard has different homogeneity within the sectors, and it leads to a limitation that forecasting stock prices by taking them all together without considering only relevant companies can adversely affect predictive performance. To overcome the limitation, we first used random matrix theory with text mining for stock prediction. Wherever the dimension of data is large, the classical limit theorems are no longer suitable, because the statistical efficiency will be reduced. Therefore, a simple correlation analysis in the financial market does not mean the true correlation. To solve the issue, we adopt random matrix theory, which is mainly used in econophysics, to remove market-wide effects and random signals and find a true correlation between companies. With the true correlation, we perform cluster analysis to find relevant companies. Also, based on the clustering analysis, we used multiple kernel learning algorithm, which is an ensemble of support vector machine to incorporate the effects of the target firm and its relevant firms simultaneously. Each kernel was assigned to predict stock prices with features of financial news of the target firm and its relevant firms. The results of this study are as follows. The results of this paper are as follows. (1) Following the existing research flow, we confirmed that it is an effective way to forecast stock prices using news from relevant companies. (2) When looking for a relevant company, looking for it in the wrong way can lower AI prediction performance. (3) The proposed approach with random matrix theory shows better performance than previous studies if cluster analysis is performed based on the true correlation by removing market-wide effects and random signals. The contribution of this study is as follows. First, this study shows that random matrix theory, which is used mainly in economic physics, can be combined with artificial intelligence to produce good methodologies. This suggests that it is important not only to develop AI algorithms but also to adopt physics theory. This extends the existing research that presented the methodology by integrating artificial intelligence with complex system theory through transfer entropy. Second, this study stressed that finding the right companies in the stock market is an important issue. This suggests that it is not only important to study artificial intelligence algorithms, but how to theoretically adjust the input values. Third, we confirmed that firms classified as Global Industrial Classification Standard (GICS) might have low relevance and suggested it is necessary to theoretically define the relevance rather than simply finding it in the GICS.

The Study on the Effect of Target Volume in DQA based on MLC log file (MLC 로그 파일 기반 DQA에서 타깃 용적에 따른 영향 연구)

  • Shin, Dong Jin;Jung, Dong Min;Cho, Kang Chul;Kim, Ji Hoon;Yoon, Jong Won;Cho, Jeong Hee
    • The Journal of Korean Society for Radiation Therapy
    • /
    • v.32
    • /
    • pp.53-59
    • /
    • 2020
  • Purpose: The purpose of this study is to compare and analyze the difference between the MLC log file-based software (Mobius) and the conventional phantom-ionization chamber (ArcCheck) dose verification method according to the change of target volume. Material and method: Radius 0.25cm, 0.5cm, 1cm, 2cm, 3cm, 4cm, 5cm, 6cm, 7cm, 8cm, 9cm, 10cm with a Sphere-shaped target Twelve plans were created and dose verification using Mobius and ArcCheck was conducted three times each. The irradiated data were compared and analyzed using the point dose error value and the gamma passing rate (3%/3mm) as evaluation indicators. Result: Mobius point dose error values were -9.87% at a radius of 0.25cm and -4.39% at 0.5cm, and the error value was within 3% at the remaining target volume. The gamma passing rate was 95% at a radius of 9cm and 93.9% at 10cm, and a passing rate of more than 95% was shown in the remaining target volume. In ArcCheck, the average error value of the point dose was about 2% in all target volumes. The gamma passing rate also showed a pass rate of 98% or more in all target volumes. Conclusion: For small targets with a radius of 0.5cm or less or a large target with a radius of 9cm or more, considering the uncertainty of DQA based on MLC log files, phantom-ionized DQA is used in complementary ways to include point dose, gamma index, DVH, and target coverage. It is believed that it is desirable to verify the dose delivery through a comprehensive analysis.

Evaluation of the usefulness of the method according to changes in patient breathing during chest 4D CT imaging (흉부 4D CT에서 호흡 변화에 대한 일시 중지 및 재개 방법의 유용성 평가)

  • Heo, Sol;Shin, Chung Hun;Jeong, Hyun Sook;Yoo, Soon Mi;Kim, Jeong Mi;Yun, In Ha;Hong, Seung Mo;Back, Geum Mun
    • The Journal of Korean Society for Radiation Therapy
    • /
    • v.33
    • /
    • pp.47-54
    • /
    • 2021
  • Purpose : In order to evaluate the usefulness of clinical application of the Pause & Resume methods by comparing and analyzing the data stability and dose reduction effect when repeat scan assuming irregular breathing and using the Pause & Resume method during chest 4D CT using QuasarTM Phantom. Materials and Methods : Using the QuasarTM Phantom, set the breathing rate per minute to 15 BPM and 7.5 BPM, and set the S15 point as an irregular breathing section, and then placed OSLD to this point and use the Pause & Resume method to measure the dose of S15. CTDIvol, DLP, and ALARA-CT were used for comparative analysis of radiation dose between Pause & Resume method and Repeat-scan. In order to evaluate the stability and usability of the data applying the Pause & Resume method, the captured images were sorted by Advanced Workstation Volume Share7 and then sent to EclipseTM, the diameter and volume were analyzed by forming a contour on the iron ball in the QuasarTM Phantom Results : When using Pause & Resume, the dose of OSLD measurement increased by 1.97 times in the section of S15. As a result of image evaluation, the average value of all volumes measured with and without the Pause & Resume method at 15 BPM and 7.5 BPM was 15.2 cm3±0.5%.Allthemeasuredvaluesfor the radius of iron ball were 3.1 cm regardless of whether Pause & Resume method was used or not. In the case of using Pause & Resume, 33% decreased from the lowest DLP value and 38% decreased from the highest DLP value of repeat scan, and the effective dose also decreased 32.1% from the minimum value and 37.6% from the maximum value. Conclusion: Irradiation dose was increased by Pause & Resume method because of the repeat scan on the S15 site where assuming irregular breathing occurred, However Pause & Resume method led to a significant reduction in dose on overall scan range. It also proved the usefulness of clinical application of the Pause & Resume method as a result of similar diameters and volumes of iron ball measurement.

A Study on the Demand for Cultural Ecosystem Services in Urban Forests Using Topic Modeling (토픽모델링을 활용한 도시림의 문화서비스 수요 특성 분석)

  • Kim, Jee-Young;Son, Yong-Hoon
    • Journal of the Korean Institute of Landscape Architecture
    • /
    • v.50 no.4
    • /
    • pp.37-52
    • /
    • 2022
  • The purpose of this study is to analyze the demand for cultural ecosystem services in urban forests based on user perception and experience value by using Naver blog posts and LDA topic modeling. Bukhansan National Park was used to analyze and review the feasibility of spatial assessments. Based on the results of topic modeling from blog posts, a review process was conducted considering the relevance of Bukhansan National Park's cultural services and its suitability as a spatial assessment case, and finally, an index for the spatial assessment of urban forest's cultural service was derived. Specifically, 21 topics derived through topic analysis were interpreted, and 13 topics related to cultural ecosystem services were derived based on the MA(Millennium Ecosystem Assessment)'s classification system for ecosystem services. 72.7% of all documents reviewed had data deemed useful for this study. The contents of the topic fell into one of the seven types of cultural services related to "mountainous recreation activities" (23.7%), "indirect use value linked to tourism and convenience facilities" (12.4%), "inspirational activities" (11.2%), "seasonal recreation activities" (6.2%), "natural appreciation and static recreation activities" (3.7%). Next, for the 13 cultural service topics derived from data gathered about Bukhansan National Park, the possibility of spatial assessment of the characteristics of cultural ecosystem services provided by urban forests was reviewed, and a total of 8 cultural service indicators were derived. The MA's cultural service classification system for ecosystem services, which was widely used in previous studies, has limitations in that it does not reflect the actual user demand of urban forests, but it is meaningful in that it categorizes cultural service indicators suitable for domestic circumstances. In addition, the study is significant as it presented a methodology to interpret and derive the demand for cultural services using a large amount of user awareness and experience data.

The Measurement of Sensitivity and Comparative Analysis of Simplified Quantitation Methods to Measure Dopamine Transporters Using [I-123]IPT Pharmacokinetic Computer Simulations ([I-123]IPT 약역학 컴퓨터시뮬레이션을 이용한 민감도 측정 및 간편화된 운반체 정량분석 방법들의 비교분석 연구)

  • Son, Hye-Kyung;Nha, Sang-Kyun;Lee, Hee-Kyung;Kim, Hee-Joung
    • The Korean Journal of Nuclear Medicine
    • /
    • v.31 no.1
    • /
    • pp.19-29
    • /
    • 1997
  • Recently, [I-123]IPT SPECT has been used for early diagnosis of Parkinson's patients(PP) by imaging dopamine transporters. The dynamic time activity curves in basal ganglia(BG) and occipital cortex(OCC) without blood samples were obtained for 2 hours. These data were then used to measure dopamine transporters by operationally defined ratio methods of (BG-OCC)/OCC at 2 hrs, binding potential $R_v=k_3/k_4$ using graphic method or $R_A$= (ABBG-ABOCC)/ABOCC for 2 hrs, where ABBG represents accumulated binding activity in basal ganglia(${\int}^{120min}_0$ BG(t)dt) and ABOCC represents accumulated binding activity in occipital cortex(${\int}^{120min}_0$ OCC(t)dt). The purpose of this study was to examine the IPT pharmacokinetics and investigate the usefulness of simplified methods of (BG-OCC)/OCC, $R_A$, and $R_v$ which are often assumed that these values reflect the true values of $k_3/k_4$. The rate constants $K_1,\;k_2\;k_3$ and $k_4$ to be used for simulations were derived using [I-123]IPT SPECT and aterialized blood data with a standard three compartmental model. The sensitivities and time activity curves in BG and OCC were computed by changing $K_l$ and $k_3$(only BG) for every 5min over 2 hours. The values (BG-OCC)/OCC, $R_A$, and $R_v$ were then computed from the time activity curves and the linear regression analysis was used to measure the accuracies of these methods. The late constants $K_l,\;k_2\;k_3\;k_4$ at BG and OCC were $1.26{\pm}5.41%,\;0.044{\pm}19.58%,\;0.031{\pm}24.36%,\;0.008{\pm}22.78%$ and $1.36{\pm}4.76%,\;0.170{\pm}6.89%,\;0.007{\pm}23.89%,\;0.007{\pm}45.09%$, respectively. The Sensitivities for ((${\Delta}S/S$)/(${\Delta}k_3/k_3$)) and ((${\Delta}S/S$)/(${\Delta}K_l/K_l$)) at 30min and 120min were measured as (0.19, 0.50) and (0.61, 0,23), respectively. The correlation coefficients and slopes of ((BG-OCC)/OCC, $R_A$, and $R_v$) with $k_3/k_4$ were (0.98, 1.00, 0.99) and (1.76, 0.47, 1.25), respectively. These simulation results indicate that a late [I-123]IPT SPECT image may represent the distribution of the dopamine transporters. Good correlations were shown between (3G-OCC)/OCC, $R_A$ or $R_v$ and true $k_3/k_4$, although the slopes between them were not unity. Pharmacokinetic computer simulations may be a very useful technique in studying dopamine transporter systems.

  • PDF

An Expert System for the Estimation of the Growth Curve Parameters of New Markets (신규시장 성장모형의 모수 추정을 위한 전문가 시스템)

  • Lee, Dongwon;Jung, Yeojin;Jung, Jaekwon;Park, Dohyung
    • Journal of Intelligence and Information Systems
    • /
    • v.21 no.4
    • /
    • pp.17-35
    • /
    • 2015
  • Demand forecasting is the activity of estimating the quantity of a product or service that consumers will purchase for a certain period of time. Developing precise forecasting models are considered important since corporates can make strategic decisions on new markets based on future demand estimated by the models. Many studies have developed market growth curve models, such as Bass, Logistic, Gompertz models, which estimate future demand when a market is in its early stage. Among the models, Bass model, which explains the demand from two types of adopters, innovators and imitators, has been widely used in forecasting. Such models require sufficient demand observations to ensure qualified results. In the beginning of a new market, however, observations are not sufficient for the models to precisely estimate the market's future demand. For this reason, as an alternative, demands guessed from those of most adjacent markets are often used as references in such cases. Reference markets can be those whose products are developed with the same categorical technologies. A market's demand may be expected to have the similar pattern with that of a reference market in case the adoption pattern of a product in the market is determined mainly by the technology related to the product. However, such processes may not always ensure pleasing results because the similarity between markets depends on intuition and/or experience. There are two major drawbacks that human experts cannot effectively handle in this approach. One is the abundance of candidate reference markets to consider, and the other is the difficulty in calculating the similarity between markets. First, there can be too many markets to consider in selecting reference markets. Mostly, markets in the same category in an industrial hierarchy can be reference markets because they are usually based on the similar technologies. However, markets can be classified into different categories even if they are based on the same generic technologies. Therefore, markets in other categories also need to be considered as potential candidates. Next, even domain experts cannot consistently calculate the similarity between markets with their own qualitative standards. The inconsistency implies missing adjacent reference markets, which may lead to the imprecise estimation of future demand. Even though there are no missing reference markets, the new market's parameters can be hardly estimated from the reference markets without quantitative standards. For this reason, this study proposes a case-based expert system that helps experts overcome the drawbacks in discovering referential markets. First, this study proposes the use of Euclidean distance measure to calculate the similarity between markets. Based on their similarities, markets are grouped into clusters. Then, missing markets with the characteristics of the cluster are searched for. Potential candidate reference markets are extracted and recommended to users. After the iteration of these steps, definite reference markets are determined according to the user's selection among those candidates. Then, finally, the new market's parameters are estimated from the reference markets. For this procedure, two techniques are used in the model. One is clustering data mining technique, and the other content-based filtering of recommender systems. The proposed system implemented with those techniques can determine the most adjacent markets based on whether a user accepts candidate markets. Experiments were conducted to validate the usefulness of the system with five ICT experts involved. In the experiments, the experts were given the list of 16 ICT markets whose parameters to be estimated. For each of the markets, the experts estimated its parameters of growth curve models with intuition at first, and then with the system. The comparison of the experiments results show that the estimated parameters are closer when they use the system in comparison with the results when they guessed them without the system.

Radioimmunoassay Reagent Survey and Evaluation (검사별 radioimmunoassay시약 조사 및 비교실험)

  • Kim, Ji-Na;An, Jae-seok;Jeon, Young-woo;Yoon, Sang-hyuk;Kim, Yoon-cheol
    • The Korean Journal of Nuclear Medicine Technology
    • /
    • v.25 no.1
    • /
    • pp.34-40
    • /
    • 2021
  • Purpose If a new test is introduced or reagents are changed in the laboratory of a medical institution, the characteristics of the test should be analyzed according to the procedure and the assessment of reagents should be made. However, several necessary conditions must be met to perform all required comparative evaluations, first enough samples should be prepared for each test, and secondly, various reagents applicable to the comparative evaluations must be supplied. Even if enough comparative evaluations have been done, there is a limit to the fact that the data variation for the new reagent represents the overall patient data variation, The fact puts a burden on the laboratory to the change the reagent. Due to these various difficulties, reagent changes in the laboratory are limited. In order to introduce a competitive bid, the institute conducted a full investigation of Radioimmunoassay(RIA) reagents for each test and established the range of reagents available in the laboratory through comparative evaluations. We wanted to share this process. Materials and Methods There are 20 items of tests conducted in our laboratory except for consignment tests. For each test, RIA reagents that can be used were fully investigated with the reference to external quality control report. and the manuals for each reagent were obtained. Each reagent was checked for the manual to check the test method, Incubation time, sample volume needed for the test. After that, the primary selection was made according to whether it was available in this laboratory. The primary selected reagents were supplied with 2kits based on 100tests, and the data correlation test, sensitivity measurement, recovery rate measurement, and dilution test were conducted. The secondary selection was performed according to the results of the comparative evaluation. The reagents that passed the primary and secondary selections were submitted to the competitive bidding list. In the case of reagent is designated as a singular, we submitted a explanatory statement with the data obtained during the primary and secondary selection processes. Results Excluded from the primary selection was the case where TAT was expected to be delayed at the moment, and it was impossible to apply to our equipment due to the large volume of reagents used during the test. In the primary selection, there were five items which only one reagent was available.(squamous cell carcinoma Ag(SCC Ag), β-human chorionic gonadotropin(β-HCG), vitamin B12, folate, free testosterone), two reagents were available(CA19-9, CA125, CA72-4, ferritin, thyroglobulin antibody(TG Ab), microsomal antibody(Mic Ab), thyroid stimulating hormone-receptor-antibody(TSH-R-Ab), calcitonin), three reagents were available (triiodothyronine(T3), Tree T3, Free T4, TSH, intact parathyroid hormone(intact PTH)) and four reagents were available are carcinoembryonic antigen(CEA), TG. In the secondary selection, there were eight items which only one reagent was available.(ferritin, TG, CA19-9, SCC, β-HCG, vitaminB12, folate, free testosterone), two reagents were available(TG Ab, Mic Ab, TSH-R-Ab, CA125, CA72-4, intact PTH, calcitonin), three reagents were available(T3, Tree T3, Free T4, TSH, CEA). Reasons excluded from the secondary selection were the lack of reagent supply for comparative evaluations, the problems with data reproducibility, and the inability to accept data variations. The most problematic part of comparative evaluations was sample collection. It didn't matter if the number of samples requested was large and the capacity needed for the test was small. It was difficult to collect various concentration samples in the case of a small number of tests(100 cases per month or less), and it was difficult to conduct a recovery rate test in the case of a relatively large volume of samples required for a single test(more than 100 uL). In addition, the lack of dilution solution or standard zero material for sensitivity measurement or dilution tests was one of the problems. Conclusion Comparative evaluation for changing test reagents require appropriate preparation time to collect diverse and sufficient samples. In addition, setting the total sample volume and reagent volume range required for comparative evaluations, depending on the sample volume and reagent volume required for one test, will reduce the burden of sample collection and planning for each comparative evaluation.

Prediction of Sleep Disturbances in Korean Rural Elderly through Longitudinal Follow Up (추적 관찰을 통한 한국 농촌 노인의 수면 장애 예측)

  • Park, Kyung Mee;Kim, Woo Jung;Choi, Eun Chae;An, Suk Kyoon;Namkoong, Kee;Youm, Yoosik;Kim, Hyeon Chang;Lee, Eun
    • Sleep Medicine and Psychophysiology
    • /
    • v.24 no.1
    • /
    • pp.38-45
    • /
    • 2017
  • Objectives: Sleep disturbance is a very rapidly growing disease with aging. The purpose of this study was to investigate the prevalence of sleep disturbances and its predictive factors in a three-year cohort study of people aged 60 years and over in Korea. Methods: In 2012 and 2014, we obtained data from a survey of the Korean Social Life, Health, and Aging Project. We asked participants if they had been diagnosed with stroke, myocardial infarction, angina pectoris, arthritis, pulmonary tuberculosis, asthma, cataract, glaucoma, hepatitis B, urinary incontinence, prostate hypertrophy, cancer, osteoporosis, hypertension, diabetes, hyperlipidemia, or metabolic syndrome. Cognitive function was assessed using the Mini-Mental State Examination for dementia screening in 2012, and depression was assessed using the Center for Epidemiologic Studies Depression Scale in 2012 and 2014. In 2015, a structured clinical interview for Axis I psychiatric disorders was administered to 235 people, and sleep disturbance was assessed using the Pittsburgh Sleep Quality Index. The perceived stress scale and the State-trait Anger Expression Inventory were also administered. Logistic regression analysis was used to predict sleep disturbance by gender, age, education, depression score, number of coexisting diseases in 2012 and 2014, current anger score, and perceived stress score. Results: Twenty-seven percent of the participants had sleep disturbances. Logistic regression analysis showed that the number of medical diseases three years ago, the depression score one year ago, and the current perceived stress significantly predicted sleep disturbances. Conclusion: Comorbid medical disease three years previous and depressive symptoms evaluated one year previous were predictive of current sleep disturbances. Further studies are needed to determine whether treatment of medical disease and depressive symptoms can improve sleep disturbances.