• Title/Summary/Keyword: Term extraction

Search Result 342, Processing Time 0.023 seconds

Your Opinions Let us Know: Mining Social Network Sites to Evolve Software Product Lines

  • Ali, Nazakat;Hwang, Sangwon;Hong, Jang-Eui
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.13 no.8
    • /
    • pp.4191-4211
    • /
    • 2019
  • Software product lines (SPLs) are complex software systems by nature due to their common reference architecture and interdependencies. Therefore, any form of evolution can lead to a more complex situation than a single system. On the other hand, software product lines are developed keeping long-term perspectives in mind, which are expected to have a considerable lifespan and a long-term investment. SPL development organizations need to consider software evolution in a systematic way due to their complexity and size. Addressing new user requirements over time is one of the most crucial factors in the successful implementation SPL. Thus, the addition of new requirements or the rapid context change is common in SPL products. To cope with rapid change several researchers have discussed the evolution of software product lines. However, for the evolution of an SPL, the literature did not present a systematic process that would define activities in such a way that would lead to the rapid evolution of software. Our study aims to provide a requirements-driven process that speeds up the requirements engineering process using social network sites in order to achieve rapid software evolution. We used classification, topic modeling, and sentiment extraction to elicit user requirements. Lastly, we conducted a case study on the smartwatch domain to validate our proposed approach. Our results show that users' opinions can contain useful information which can be used by software SPL organizations to evolve their products. Furthermore, our investigation results demonstrate that machine learning algorithms have the capacity to identify relevant information automatically.

Arousal and Valence Classification Model Based on Long Short-Term Memory and DEAP Data for Mental Healthcare Management

  • Choi, Eun Jeong;Kim, Dong Keun
    • Healthcare Informatics Research
    • /
    • v.24 no.4
    • /
    • pp.309-316
    • /
    • 2018
  • Objectives: Both the valence and arousal components of affect are important considerations when managing mental healthcare because they are associated with affective and physiological responses. Research on arousal and valence analysis, which uses images, texts, and physiological signals that employ deep learning, is actively underway; research investigating how to improve the recognition rate is needed. The goal of this research was to design a deep learning framework and model to classify arousal and valence, indicating positive and negative degrees of emotion as high or low. Methods: The proposed arousal and valence classification model to analyze the affective state was tested using data from 40 channels provided by a dataset for emotion analysis using electrocardiography (EEG), physiological, and video signals (the DEAP dataset). Experiments were based on 10 selected featured central and peripheral nervous system data points, using long short-term memory (LSTM) as a deep learning method. Results: The arousal and valence were classified and visualized on a two-dimensional coordinate plane. Profiles were designed depending on the number of hidden layers, nodes, and hyperparameters according to the error rate. The experimental results show an arousal and valence classification model accuracy of 74.65 and 78%, respectively. The proposed model performed better than previous other models. Conclusions: The proposed model appears to be effective in analyzing arousal and valence; specifically, it is expected that affective analysis using physiological signals based on LSTM will be possible without manual feature extraction. In a future study, the classification model will be adopted in mental healthcare management systems.

A Study on Image Generation from Sentence Embedding Applying Self-Attention (Self-Attention을 적용한 문장 임베딩으로부터 이미지 생성 연구)

  • Yu, Kyungho;No, Juhyeon;Hong, Taekeun;Kim, Hyeong-Ju;Kim, Pankoo
    • Smart Media Journal
    • /
    • v.10 no.1
    • /
    • pp.63-69
    • /
    • 2021
  • When a person sees a sentence and understands the sentence, the person understands the sentence by reminiscent of the main word in the sentence as an image. Text-to-image is what allows computers to do this associative process. The previous deep learning-based text-to-image model extracts text features using Convolutional Neural Network (CNN)-Long Short Term Memory (LSTM) and bi-directional LSTM, and generates an image by inputting it to the GAN. The previous text-to-image model uses basic embedding in text feature extraction, and it takes a long time to train because images are generated using several modules. Therefore, in this research, we propose a method of extracting features by using the attention mechanism, which has improved performance in the natural language processing field, for sentence embedding, and generating an image by inputting the extracted features into the GAN. As a result of the experiment, the inception score was higher than that of the model used in the previous study, and when judged with the naked eye, an image that expresses the features well in the input sentence was created. In addition, even when a long sentence is input, an image that expresses the sentence well was created.

Power Quality Disturbances Detection and Classification using Fast Fourier Transform and Deep Neural Network (고속 푸리에 변환 및 심층 신경망을 사용한 전력 품질 외란 감지 및 분류)

  • Senfeng Cen;Chang-Gyoon Lim
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.18 no.1
    • /
    • pp.115-126
    • /
    • 2023
  • Due to the fluctuating random and periodical nature of renewable energy generation power quality disturbances occurred more frequently in power generation transformation transmission and distribution. Various power quality disturbances may lead to equipment damage or even power outages. Therefore it is essential to detect and classify different power quality disturbances in real time automatically. The traditional PQD identification method consists of three steps: feature extraction feature selection and classification. However, the handcrafted features are imprecise in the feature selection stage, resulting in low classification accuracy. This paper proposes a deep neural architecture based on Convolution Neural Network and Long Short Term Memory combining the time and frequency domain features to recognize 16 types of Power Quality signals. The frequency-domain data were obtained from the Fast Fourier Transform which could efficiently extract the frequency-domain features. The performance in synthetic data and real 6kV power system data indicate that our proposed method generalizes well compared with other deep learning methods.

Short-Term Water Quality Prediction of the Paldang Reservoir Using Recurrent Neural Network Models (순환신경망 모델을 활용한 팔당호의 단기 수질 예측)

  • Jiwoo Han;Yong-Chul Cho;Soyoung Lee;Sanghun Kim;Taegu Kang
    • Journal of Korean Society on Water Environment
    • /
    • v.39 no.1
    • /
    • pp.46-60
    • /
    • 2023
  • Climate change causes fluctuations in water quality in the aquatic environment, which can cause changes in water circulation patterns and severe adverse effects on aquatic ecosystems in the future. Therefore, research is needed to predict and respond to water quality changes caused by climate change in advance. In this study, we tried to predict the dissolved oxygen (DO), chlorophyll-a, and turbidity of the Paldang reservoir for about two weeks using long short-term memory (LSTM) and gated recurrent units (GRU), which are deep learning algorithms based on recurrent neural networks. The model was built based on real-time water quality data and meteorological data. The observation period was set from July to September in the summer of 2021 (Period 1) and from March to May in the spring of 2022 (Period 2). We tried to select an algorithm with optimal predictive power for each water quality parameter. In addition, to improve the predictive power of the model, an important variable extraction technique using random forest was used to select only the important variables as input variables. In both Periods 1 and 2, the predictive power after extracting important variables was further improved. Except for DO in Period 2, GRU was selected as the best model in all water quality parameters. This methodology can be useful for preventive water quality management by identifying the variability of water quality in advance and predicting water quality in a short period.

Development and Lessons Learned of Clinical Data Warehouse based on Common Data Model for Drug Surveillance (약물부작용 감시를 위한 공통데이터모델 기반 임상데이터웨어하우스 구축)

  • Mi Jung Rho
    • Korea Journal of Hospital Management
    • /
    • v.28 no.3
    • /
    • pp.1-14
    • /
    • 2023
  • Purposes: It is very important to establish a clinical data warehouse based on a common data model to offset the different data characteristics of each medical institution and for drug surveillance. This study attempted to establish a clinical data warehouse for Dankook university hospital for drug surveillance, and to derive the main items necessary for development. Methodology/Approach: This study extracted the electronic medical record data of Dankook university hospital tracked for 9 years from 2013 (2013.01.01. to 2021.12.31) to build a clinical data warehouse. The extracted data was converted into the Observational Medical Outcomes Partnership Common Data Model (Version 5.4). Data term mapping was performed using the electronic medical record data of Dankook university hospital and the standard term mapping guide. To verify the clinical data warehouse, the use of angiotensin receptor blockers and the incidence of liver toxicity were analyzed, and the results were compared with the analysis of hospital raw data. Findings: This study used a total of 670,933 data from electronic medical records for the Dankook university clinical data warehouse. Excluding the number of overlapping cases among the total number of cases, the target data was mapped into standard terms. Diagnosis (100% of total cases), drug (92.1%), and measurement (94.5%) were standardized. For treatment and surgery, the insurance EDI (electronic data interchange) code was used as it is. Extraction, conversion and loading were completed. R language-based conversion and loading software for the process was developed, and clinical data warehouse construction was completed through data verification. Practical Implications: In this study, a clinical data warehouse for Dankook university hospitals based on a common data model supporting drug surveillance research was established and verified. The results of this study provide guidelines for institutions that want to build a clinical data warehouse in the future by deriving key points necessary for building a clinical data warehouse.

  • PDF

Fabrication of implant-associated obturator after extraction of abutment teeth: a case report (지대치 발거 후 임플란트 연관 상악 폐색장치 제작 증례보고)

  • Ki-Yeol Jang;Gyeong-Je Lee
    • Journal of Dental Rehabilitation and Applied Science
    • /
    • v.39 no.4
    • /
    • pp.229-236
    • /
    • 2023
  • Maxillary bone defects may follow surgical treatment of benign and malignant tumors, trauma, and infection. Palatal defects often lead to problems with swallowing and pronunciation from the leakage of air into the nasal cavity and sinus. Obturators have been commonly used to solve these problems, but long-term use of the device may cause irritation of the oral mucosa or damage to the abutment teeth. Utilizing implants in the edentulous area for the fabrication of the obturators has gained attention. This case report describes a patient, who had undergone partial resection of the maxilla due to adenocarcinoma, in need of a new obturator after losing abutment teeth after long-term use of the previous obturator. Implants were placed in strategic locations, and an implant-retained maxillary obturator was fabricated, showing satisfactory results in the rehabilitation of multiple aspects, including palatal defect, masticatory function, swallowing, pronunciation, and aesthetics.

Enhancing anti-calcification efficacy in veterinary cardiovascular surgeries: evaluating short-term ethanol's role in glutaraldehyde fixed pericardial tissues in rats

  • Kyung-Min Kim;Won-Jong Lee;Woo-Jin Kim;Chang-Hwan Moon;Jaemin Jeong;Hae-Beom Lee;Seong-Mok Jeong;Dae-Hyun Kim
    • Korean Journal of Veterinary Research
    • /
    • v.64 no.2
    • /
    • pp.16.1-16.9
    • /
    • 2024
  • Autologous pericardial tissues are utilized in veterinary cardiovascular surgeries due to their accessibility and effectiveness. To enhance handling and biomechanical properties, glutaraldehyde (GA) fixation is applied. However, GA fixation can induce calcification, leading to tissue failure. This study aimed to establish an optimal rapid anti-calcification protocol by integrating ethanol treatment with the proven effective GA concentration and fixation time, facilitating application from collection to utilization. Pericardia were fixed with 0.625% GA for 20 min and subjected to ethanol treatment for 0 (group A, control), 20 (group B), and 30 minutes (group C). The treated tissues underwent mechanical test and were implanted subcutaneously in 3-week-old male rats for 7 weeks before extraction, followed by calcium analysis and histological examination via hematoxylin and eosin staining. No significant differences in mechanical properties were observed among the groups. The ethanol-treated groups (groups B and C; p < 0.05) exhibited significantly lower calcium levels than control (group A). Microscopy confirmed collagen and elastic fibers preservation, without significant immune cell variance. However, higher fibrocyte presence was noted in the ethanol-treated groups. This study presents a rapid anti-calcification protocol combining ethanol treatment with optimal GA fixation, suitable for direct surgical use of autologous tissues. Further research is necessary for long-term efficacy evaluation.

Hybrid LSTM and Deep Belief Networks with Attention Mechanism for Accurate Heart Attack Data Analytics

  • Mubarak Albathan
    • International Journal of Computer Science & Network Security
    • /
    • v.24 no.10
    • /
    • pp.1-16
    • /
    • 2024
  • Due to its complexity and high diagnosis and treatment costs, heart attack (HA) is the top cause of death globally. Heart failure's widespread effect and high morbidity and death rates make accurate and fast prognosis and diagnosis crucial. Due to the complexity of medical data, early and accurate prediction of HA is difficult. Healthcare providers must evaluate data quickly and accurately to intervene. This novel hybrid approach predicts HA using Long Short-Term Memory (LSTM) networks, Deep belief networks (DBNs) with attention mechanism, and robust data mining to fill this essential gap. HA is predicted using Kaggle, PhysioNet, and UCI datasets. Wearable sensor data, ECG signals, and demographic and clinical data provide a solid analytical base. To maintain consistency, ECG signals are normalized and segmented after thorough cleaning to remove missing values and noise. Feature extraction employs complex approaches like Principal Component Analysis (PCA) and Autoencoders to pick time-domain (MNN, SDNN, RMSSD, PNN50) and frequency-domain (PSD at VLF, LF, HF bands) characteristics. The hybrid model architecture uses LSTM networks for sequence learning and DBNs for feature representation and selection to create a robust and comprehensive prediction model. Accuracy, precision, recall, F1-score, and ROC-AUC are measured after cross-entropy loss and SGD optimization. The LSTM-DBN model outperforms predictive methods in accuracy, sensitivity, and specificity. The findings show that several data sources and powerful algorithms can improve heart attack predictions. The proposed architecture performed well on many datasets, with an accuracy rate of 96.00%, sensitivity of 98%, AUC of 0.98, and F1-score of 0.97. High performance proves this system's dependability. Moreover, the proposed approach is outperformed compared to state-of-the-art systems.

Knowledge Extraction Methodology and Framework from Wikipedia Articles for Construction of Knowledge-Base (지식베이스 구축을 위한 한국어 위키피디아의 학습 기반 지식추출 방법론 및 플랫폼 연구)

  • Kim, JaeHun;Lee, Myungjin
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.1
    • /
    • pp.43-61
    • /
    • 2019
  • Development of technologies in artificial intelligence has been rapidly increasing with the Fourth Industrial Revolution, and researches related to AI have been actively conducted in a variety of fields such as autonomous vehicles, natural language processing, and robotics. These researches have been focused on solving cognitive problems such as learning and problem solving related to human intelligence from the 1950s. The field of artificial intelligence has achieved more technological advance than ever, due to recent interest in technology and research on various algorithms. The knowledge-based system is a sub-domain of artificial intelligence, and it aims to enable artificial intelligence agents to make decisions by using machine-readable and processible knowledge constructed from complex and informal human knowledge and rules in various fields. A knowledge base is used to optimize information collection, organization, and retrieval, and recently it is used with statistical artificial intelligence such as machine learning. Recently, the purpose of the knowledge base is to express, publish, and share knowledge on the web by describing and connecting web resources such as pages and data. These knowledge bases are used for intelligent processing in various fields of artificial intelligence such as question answering system of the smart speaker. However, building a useful knowledge base is a time-consuming task and still requires a lot of effort of the experts. In recent years, many kinds of research and technologies of knowledge based artificial intelligence use DBpedia that is one of the biggest knowledge base aiming to extract structured content from the various information of Wikipedia. DBpedia contains various information extracted from Wikipedia such as a title, categories, and links, but the most useful knowledge is from infobox of Wikipedia that presents a summary of some unifying aspect created by users. These knowledge are created by the mapping rule between infobox structures and DBpedia ontology schema defined in DBpedia Extraction Framework. In this way, DBpedia can expect high reliability in terms of accuracy of knowledge by using the method of generating knowledge from semi-structured infobox data created by users. However, since only about 50% of all wiki pages contain infobox in Korean Wikipedia, DBpedia has limitations in term of knowledge scalability. This paper proposes a method to extract knowledge from text documents according to the ontology schema using machine learning. In order to demonstrate the appropriateness of this method, we explain a knowledge extraction model according to the DBpedia ontology schema by learning Wikipedia infoboxes. Our knowledge extraction model consists of three steps, document classification as ontology classes, proper sentence classification to extract triples, and value selection and transformation into RDF triple structure. The structure of Wikipedia infobox are defined as infobox templates that provide standardized information across related articles, and DBpedia ontology schema can be mapped these infobox templates. Based on these mapping relations, we classify the input document according to infobox categories which means ontology classes. After determining the classification of the input document, we classify the appropriate sentence according to attributes belonging to the classification. Finally, we extract knowledge from sentences that are classified as appropriate, and we convert knowledge into a form of triples. In order to train models, we generated training data set from Wikipedia dump using a method to add BIO tags to sentences, so we trained about 200 classes and about 2,500 relations for extracting knowledge. Furthermore, we evaluated comparative experiments of CRF and Bi-LSTM-CRF for the knowledge extraction process. Through this proposed process, it is possible to utilize structured knowledge by extracting knowledge according to the ontology schema from text documents. In addition, this methodology can significantly reduce the effort of the experts to construct instances according to the ontology schema.