• Title/Summary/Keyword: Model surgery

Search Result 1,532, Processing Time 0.025 seconds

Semi-Quantitative Analysis for Determining the Optimal Threshold Value on CT to Measure the Solid Portion of Pulmonary Subsolid Nodules (폐의 아고형결절에서 침습적 병소를 검출하기 위한 반-정량 분석을 통한 최적의 CT 임계 값 결정)

  • Sunyong Lee;Da Hyun Lee;Jae Ho Lee;Sungsoo Lee;Kyunghwa Han;Chul Hwan Park;Tae Hoon Kim
    • Journal of the Korean Society of Radiology
    • /
    • v.82 no.3
    • /
    • pp.670-681
    • /
    • 2021
  • Purpose This study aimed to investigate the optimal threshold value in Hounsfield units (HU) on CT to detect the solid components of pulmonary subsolid nodules using pathologic invasive foci as reference. Materials and Methods Thin-section non-enhanced chest CT scans of 25 patients with pathologically confirmed minimally invasive adenocarcinoma were retrospectively reviewed. On CT images, the solid portion was defined as the area with higher attenuation than various HU thresholds ranging from -600 to -100 HU in 50-HU intervals. The solid portion was measured as the largest diameter on axial images and as the maximum diameter on multiplanar reconstruction images. A linear mixed model was used to evaluate bias in each threshold by using the pathological size of invasive foci as reference. Results At a threshold of -400 HU, the biases were lowest between the largest/maximum diameter of the solid portion of subsolid nodule and the size of invasive foci of the pathological specimen, with 0.388 and -0.0176, respectively. They showed insignificant difference (p = 0.2682, p = 0.963, respectively) at a threshold of -400 HU. Conclusion For quantitative analysis, -400 HU may be the optimal threshold to define the solid portion of subsolid nodules as a surrogate marker of invasive foci.

A Study of 'Emotion Trigger' by Text Mining Techniques (텍스트 마이닝을 이용한 감정 유발 요인 'Emotion Trigger'에 관한 연구)

  • An, Juyoung;Bae, Junghwan;Han, Namgi;Song, Min
    • Journal of Intelligence and Information Systems
    • /
    • v.21 no.2
    • /
    • pp.69-92
    • /
    • 2015
  • The explosion of social media data has led to apply text-mining techniques to analyze big social media data in a more rigorous manner. Even if social media text analysis algorithms were improved, previous approaches to social media text analysis have some limitations. In the field of sentiment analysis of social media written in Korean, there are two typical approaches. One is the linguistic approach using machine learning, which is the most common approach. Some studies have been conducted by adding grammatical factors to feature sets for training classification model. The other approach adopts the semantic analysis method to sentiment analysis, but this approach is mainly applied to English texts. To overcome these limitations, this study applies the Word2Vec algorithm which is an extension of the neural network algorithms to deal with more extensive semantic features that were underestimated in existing sentiment analysis. The result from adopting the Word2Vec algorithm is compared to the result from co-occurrence analysis to identify the difference between two approaches. The results show that the distribution related word extracted by Word2Vec algorithm in that the words represent some emotion about the keyword used are three times more than extracted by co-occurrence analysis. The reason of the difference between two results comes from Word2Vec's semantic features vectorization. Therefore, it is possible to say that Word2Vec algorithm is able to catch the hidden related words which have not been found in traditional analysis. In addition, Part Of Speech (POS) tagging for Korean is used to detect adjective as "emotional word" in Korean. In addition, the emotion words extracted from the text are converted into word vector by the Word2Vec algorithm to find related words. Among these related words, noun words are selected because each word of them would have causal relationship with "emotional word" in the sentence. The process of extracting these trigger factor of emotional word is named "Emotion Trigger" in this study. As a case study, the datasets used in the study are collected by searching using three keywords: professor, prosecutor, and doctor in that these keywords contain rich public emotion and opinion. Advanced data collecting was conducted to select secondary keywords for data gathering. The secondary keywords for each keyword used to gather the data to be used in actual analysis are followed: Professor (sexual assault, misappropriation of research money, recruitment irregularities, polifessor), Doctor (Shin hae-chul sky hospital, drinking and plastic surgery, rebate) Prosecutor (lewd behavior, sponsor). The size of the text data is about to 100,000(Professor: 25720, Doctor: 35110, Prosecutor: 43225) and the data are gathered from news, blog, and twitter to reflect various level of public emotion into text data analysis. As a visualization method, Gephi (http://gephi.github.io) was used and every program used in text processing and analysis are java coding. The contributions of this study are as follows: First, different approaches for sentiment analysis are integrated to overcome the limitations of existing approaches. Secondly, finding Emotion Trigger can detect the hidden connections to public emotion which existing method cannot detect. Finally, the approach used in this study could be generalized regardless of types of text data. The limitation of this study is that it is hard to say the word extracted by Emotion Trigger processing has significantly causal relationship with emotional word in a sentence. The future study will be conducted to clarify the causal relationship between emotional words and the words extracted by Emotion Trigger by comparing with the relationships manually tagged. Furthermore, the text data used in Emotion Trigger are twitter, so the data have a number of distinct features which we did not deal with in this study. These features will be considered in further study.