• Title/Summary/Keyword: intelligent classification

Search Result 915, Processing Time 0.026 seconds

Method of Predicting Path Loss and Base Station Topography Classification using Artificial Intelligent in Mobile Communication Systems (이동통신 시스템에서 인공지능을 이용한 경로 손실 예측 및 기지국 지형 구분 방법)

  • Kim, Jaejeong;Lee, Heejun;Ji, Seunghwan
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.26 no.5
    • /
    • pp.703-713
    • /
    • 2022
  • Accurate and rapid establishment of mobile communication is important in mobile communication system. Currently, the base station parameters to establish a network are determined by cell planning tool. However, it is necessary to perform new cell planning for each new installation of the base station, and there may be a problem that parameters are not suitable for the actual environment are set, such as obstacle information that is not applied in the cell planning tool. In this paper, we proposed methods for path loss prediction using DNN and topographical division using CNN in SON server. After topography classification, a SON server configures the base station parameters according to topography, and update parameters for each topography. The proposed methods can configure the base station parameters automatically that are considered topography information and environmental changes.

A Study on the Quantitative Evaluation Method of Quality Control using Ultrasound Phantom in Ultrasound Imaging System based on Artificial Intelligence (인공지능을 활용한 초음파영상진단장치에서 초음파 팬텀 영상을 이용한 정도관리의 정량적 평가방법 연구)

  • Yeon Jin, Im;Ho Seong, Hwang;Dong Hyun, Kim;Ho Chul, Kim
    • Journal of Biomedical Engineering Research
    • /
    • v.43 no.6
    • /
    • pp.390-398
    • /
    • 2022
  • Ultrasound examination using ultrasound equipment is an ultrasound device that images human organs using sound waves and is used in various areas such as diagnosis, follow-up, and treatment of diseases. However, if the quality of ultrasound equipment is not guaranteed, the possibility of misdiagnosis increases, and the diagnosis rate decreases. Accordingly, The Korean Society of Radiology and Korea society of Ultrasound in Medicine presented guidelines for quality management of ultrasound equipment using ATS-539 phantom. The DenseNet201 classification algorithm shows 99.25% accuracy and 5.17% loss in the Dead Zone, 97.52% loss in Axial/Lateral Resolution, 96.98% accuracy and 20.64% loss in Sensitivity, 93.44% accuracy and 22.07% loss in the Gray scale and Dynamic Range. As a result, it is the best and is judged to be an algorithm that can be used for quantitative evaluation. Through this study, it can be seen that if quantitative evaluation using artificial intelligence is conducted in the qualitative evaluation item of ultrasonic equipment, the reliability of ultrasonic equipment can be increased with high accuracy.

An intelligent method for pregnancy diagnosis in breeding sows according to ultrasonography algorithms

  • Jung-woo Chae;Yo-han Choi;Jeong-nam Lee;Hyun-ju Park;Yong-dae Jeong;Eun-seok Cho;Young-sin, Kim;Tae-kyeong Kim;Soo-jin Sa;Hyun-chong Cho
    • Journal of Animal Science and Technology
    • /
    • v.65 no.2
    • /
    • pp.365-376
    • /
    • 2023
  • Pig breeding management directly contributes to the profitability of pig farms, and pregnancy diagnosis is an important factor in breeding management. Therefore, the need to diagnose pregnancy in sows is emphasized, and various studies have been conducted in this area. We propose a computer-aided diagnosis system to assist livestock farmers to diagnose sow pregnancy through ultrasound. Methods for diagnosing pregnancy in sows through ultrasound include the Doppler method, which measures the heart rate and pulse status, and the echo method, which diagnoses by amplitude depth technique. We propose a method that uses deep learning algorithms on ultrasonography, which is part of the echo method. As deep learning-based classification algorithms, Inception-v4, Xception, and EfficientNetV2 were used and compared to find the optimal algorithm for pregnancy diagnosis in sows. Gaussian and speckle noises were added to the ultrasound images according to the characteristics of the ultrasonography, which is easily affected by noise from the surrounding environments. Both the original and noise added ultrasound images of sows were tested together to determine the suitability of the proposed method on farms. The pregnancy diagnosis performance on the original ultrasound images achieved 0.99 in accuracy in the highest case and on the ultrasound images with noises, the performance achieved 0.98 in accuracy. The diagnosis performance achieved 0.96 in accuracy even when the intensity of noise was strong, proving its robustness against noise.

Research on data augmentation algorithm for time series based on deep learning

  • Shiyu Liu;Hongyan Qiao;Lianhong Yuan;Yuan Yuan;Jun Liu
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.17 no.6
    • /
    • pp.1530-1544
    • /
    • 2023
  • Data monitoring is an important foundation of modern science. In most cases, the monitoring data is time-series data, which has high application value. The deep learning algorithm has a strong nonlinear fitting capability, which enables the recognition of time series by capturing anomalous information in time series. At present, the research of time series recognition based on deep learning is especially important for data monitoring. Deep learning algorithms require a large amount of data for training. However, abnormal sample is a small sample in time series, which means the number of abnormal time series can seriously affect the accuracy of recognition algorithm because of class imbalance. In order to increase the number of abnormal sample, a data augmentation method called GANBATS (GAN-based Bi-LSTM and Attention for Time Series) is proposed. In GANBATS, Bi-LSTM is introduced to extract the timing features and then transfer features to the generator network of GANBATS.GANBATS also modifies the discriminator network by adding an attention mechanism to achieve global attention for time series. At the end of discriminator, GANBATS is adding averagepooling layer, which merges temporal features to boost the operational efficiency. In this paper, four time series datasets and five data augmentation algorithms are used for comparison experiments. The generated data are measured by PRD(Percent Root Mean Square Difference) and DTW(Dynamic Time Warping). The experimental results show that GANBATS reduces up to 26.22 in PRD metric and 9.45 in DTW metric. In addition, this paper uses different algorithms to reconstruct the datasets and compare them by classification accuracy. The classification accuracy is improved by 6.44%-12.96% on four time series datasets.

Students' Performance Prediction in Higher Education Using Multi-Agent Framework Based Distributed Data Mining Approach: A Review

  • M.Nazir;A.Noraziah;M.Rahmah
    • International Journal of Computer Science & Network Security
    • /
    • v.23 no.10
    • /
    • pp.135-146
    • /
    • 2023
  • An effective educational program warrants the inclusion of an innovative construction which enhances the higher education efficacy in such a way that accelerates the achievement of desired results and reduces the risk of failures. Educational Decision Support System (EDSS) has currently been a hot topic in educational systems, facilitating the pupil result monitoring and evaluation to be performed during their development. Insufficient information systems encounter trouble and hurdles in making the sufficient advantage from EDSS owing to the deficit of accuracy, incorrect analysis study of the characteristic, and inadequate database. DMTs (Data Mining Techniques) provide helpful tools in finding the models or forms of data and are extremely useful in the decision-making process. Several researchers have participated in the research involving distributed data mining with multi-agent technology. The rapid growth of network technology and IT use has led to the widespread use of distributed databases. This article explains the available data mining technology and the distributed data mining system framework. Distributed Data Mining approach is utilized for this work so that a classifier capable of predicting the success of students in the economic domain can be constructed. This research also discusses the Intelligent Knowledge Base Distributed Data Mining framework to assess the performance of the students through a mid-term exam and final-term exam employing Multi-agent system-based educational mining techniques. Using single and ensemble-based classifiers, this study intends to investigate the factors that influence student performance in higher education and construct a classification model that can predict academic achievement. We also discussed the importance of multi-agent systems and comparative machine learning approaches in EDSS development.

Development of the Artificial Intelligence Literacy Education Program for Preservice Secondary Teachers (예비 중등교사를 위한 인공지능 리터러시 교육 프로그램 개발)

  • Bong Seok Jang
    • Journal of Practical Engineering Education
    • /
    • v.16 no.1_spc
    • /
    • pp.65-70
    • /
    • 2024
  • As the interest in AI education grows, researchers have made efforts to implement AI education programs. However, research targeting pre-service teachers has been limited thus far. Therefore, this study was conducted to develop an AI literacy education program for preservice secondary teachers. The research results revealed that the weekly topics included the definition and applications of AI, analysis of intelligent agents, the importance of data, understanding machine learning, hands-on exercises on prediction and classification, hands-on exercises on clustering and classification, hands-on exercises on unstructured data, understanding deep learning, application of deep learning algorithms, fairness, transparency, accountability, safety, and social integration. Through this research, it is hoped that AI literacy education programs for preservice teachers will be expanded. In the future, it is anticipated that follow-up studies will be conducted to implement relevant education in teacher training institutions and analyze its effectiveness.

Knowledge Extraction Methodology and Framework from Wikipedia Articles for Construction of Knowledge-Base (지식베이스 구축을 위한 한국어 위키피디아의 학습 기반 지식추출 방법론 및 플랫폼 연구)

  • Kim, JaeHun;Lee, Myungjin
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.1
    • /
    • pp.43-61
    • /
    • 2019
  • Development of technologies in artificial intelligence has been rapidly increasing with the Fourth Industrial Revolution, and researches related to AI have been actively conducted in a variety of fields such as autonomous vehicles, natural language processing, and robotics. These researches have been focused on solving cognitive problems such as learning and problem solving related to human intelligence from the 1950s. The field of artificial intelligence has achieved more technological advance than ever, due to recent interest in technology and research on various algorithms. The knowledge-based system is a sub-domain of artificial intelligence, and it aims to enable artificial intelligence agents to make decisions by using machine-readable and processible knowledge constructed from complex and informal human knowledge and rules in various fields. A knowledge base is used to optimize information collection, organization, and retrieval, and recently it is used with statistical artificial intelligence such as machine learning. Recently, the purpose of the knowledge base is to express, publish, and share knowledge on the web by describing and connecting web resources such as pages and data. These knowledge bases are used for intelligent processing in various fields of artificial intelligence such as question answering system of the smart speaker. However, building a useful knowledge base is a time-consuming task and still requires a lot of effort of the experts. In recent years, many kinds of research and technologies of knowledge based artificial intelligence use DBpedia that is one of the biggest knowledge base aiming to extract structured content from the various information of Wikipedia. DBpedia contains various information extracted from Wikipedia such as a title, categories, and links, but the most useful knowledge is from infobox of Wikipedia that presents a summary of some unifying aspect created by users. These knowledge are created by the mapping rule between infobox structures and DBpedia ontology schema defined in DBpedia Extraction Framework. In this way, DBpedia can expect high reliability in terms of accuracy of knowledge by using the method of generating knowledge from semi-structured infobox data created by users. However, since only about 50% of all wiki pages contain infobox in Korean Wikipedia, DBpedia has limitations in term of knowledge scalability. This paper proposes a method to extract knowledge from text documents according to the ontology schema using machine learning. In order to demonstrate the appropriateness of this method, we explain a knowledge extraction model according to the DBpedia ontology schema by learning Wikipedia infoboxes. Our knowledge extraction model consists of three steps, document classification as ontology classes, proper sentence classification to extract triples, and value selection and transformation into RDF triple structure. The structure of Wikipedia infobox are defined as infobox templates that provide standardized information across related articles, and DBpedia ontology schema can be mapped these infobox templates. Based on these mapping relations, we classify the input document according to infobox categories which means ontology classes. After determining the classification of the input document, we classify the appropriate sentence according to attributes belonging to the classification. Finally, we extract knowledge from sentences that are classified as appropriate, and we convert knowledge into a form of triples. In order to train models, we generated training data set from Wikipedia dump using a method to add BIO tags to sentences, so we trained about 200 classes and about 2,500 relations for extracting knowledge. Furthermore, we evaluated comparative experiments of CRF and Bi-LSTM-CRF for the knowledge extraction process. Through this proposed process, it is possible to utilize structured knowledge by extracting knowledge according to the ontology schema from text documents. In addition, this methodology can significantly reduce the effort of the experts to construct instances according to the ontology schema.

The Role of Open Innovation for SME's R&D Success (중소기업 R&D 성공에 있어서 개방형 혁신의 효과에 관한 연구)

  • Yoo, In-Jin;Seo, Bong-Goon;Park, Do-Hyung
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.3
    • /
    • pp.89-117
    • /
    • 2018
  • The Korean companies are intensifying competition with not only domestic companies but also foreign companies in globalization. In this environment, it is essential activities not only for large companies but also Small and Medium Enterprises (SMEs) to get and develop the core competency. Particularly, SMEs that are inferior to resources of various aspects, such as financial resources etc., can make innovation through effective R&D investment. And then, SMEs can occupy a competency and can be survive at the environment. Conventionally, the method of "self-development" by using only the internal resources of the company has been dominant. Recently, however, R&D method through cooperation, also called "Open Innovation", is emerging. Especially SMEs are relatively short of available internal resources. Therefore, it is necessary to utilize technology and resources through cooperation with external companies(such as joint development or contract development etc.) rather than self-development R&D. In this context, we confirmed the effect of SMEs' factors on sales in Korea. Specifically, the factors that SMEs hold are classified as 'Technical characteristic', 'Company competency', and 'R&D activity' and analyzed how they influence the sales achieved as a result of R&D. The analysis was based on a two-year statistical survey conducted by the Korean government. In addition, we confirmed the influence of the factors on the sales according to the R&D method(Self-Development vs. Open Innovation), and also observed the influence change in 29 industrial categories. The results of the study are summarized as follows: First, regression analysis shows that twelve factors of SMEs have a significant effect on sales. Specifically, 15 factors included in the analysis, 12 factors excluding 3 factors were found to have significant influence. In the technical characteristic, 'imitation period' and 'product life cycle' of the technology were confirmed. In the company competency, 'R&D led person', 'researcher number', 'intellectual property registration status', 'number of R&D attempts', and 'ratio of success to trial' were confirmed. The R&D activity was found to have a significant impact on all included factors. Second, the influence of factors on the R&D method was confirmed, and the change was confirmed in four factors. In addition, these factors were found that have different effects on sales according to the R&D method. Specifically, 'researcher number', 'number of R&D attempts', 'performance compensation system', and 'R&D investment' were found to have significant moderate effects. In other words, the moderating effect of open innovation was confirmed for four factors. Third, on the industrial classification, it is confirmed that different factors have a significant influence on each industrial classification. At this point, it was confirmed that at least one factor, up to nine factors had a significant effect on the sales according to the industrial classification. Furthermore, different moderate effects have been confirmed in the industrial classification and R&D method. In the moderate effect, up to eight significant moderate effects were confirmed according to the industrial classification. In particular, 'R&D investment' and 'performance compensation system' were confirmed to be the most common moderating effect by each 12 times and 11 times in all industrial classification. This study provides the following suggestions: First, it is necessary for SMEs to determine the R&D method in consideration of the characteristics of the technology to be R&D as well as the enterprise competency and the R&D activity. In addition, there is a need to identify and concentrate on the factors that increase sales in R&D decisions, which are mainly affected by the industry classification to which the company belongs. Second, governments that support SMEs' R&D need to provide guidelines that are fit to their situation. It is necessary to differentiate the support for the company considering various factors such as technology and R&D purpose for their effective budget execution. Finally, based on the results of this study, we urge the need to reconsider the effectiveness of existing SME support policies.

UX Methodology Study by Data Analysis Focusing on deriving persona through customer segment classification (데이터 분석을 통한 UX 방법론 연구 고객 세그먼트 분류를 통한 페르소나 도출을 중심으로)

  • Lee, Seul-Yi;Park, Do-Hyung
    • Journal of Intelligence and Information Systems
    • /
    • v.27 no.1
    • /
    • pp.151-176
    • /
    • 2021
  • As the information technology industry develops, various kinds of data are being created, and it is now essential to process them and use them in the industry. Analyzing and utilizing various digital data collected online and offline is a necessary process to provide an appropriate experience for customers in the industry. In order to create new businesses, products, and services, it is essential to use customer data collected in various ways to deeply understand potential customers' needs and analyze behavior patterns to capture hidden signals of desire. However, it is true that research using data analysis and UX methodology, which should be conducted in parallel for effective service development, is being conducted separately and that there is a lack of examples of use in the industry. In thiswork, we construct a single process by applying data analysis methods and UX methodologies. This study is important in that it is highly likely to be used because it applies methodologies that are actively used in practice. We conducted a survey on the topic to identify and cluster the associations between factors to establish customer classification and target customers. The research methods are as follows. First, we first conduct a factor, regression analysis to determine the association between factors in the happiness data survey. Groups are grouped according to the survey results and identify the relationship between 34 questions of psychological stability, family life, relational satisfaction, health, economic satisfaction, work satisfaction, daily life satisfaction, and residential environment satisfaction. Second, we classify clusters based on factors affecting happiness and extract the optimal number of clusters. Based on the results, we cross-analyzed the characteristics of each cluster. Third, forservice definition, analysis was conducted by correlating with keywords related to happiness. We leverage keyword analysis of the thumb trend to derive ideas based on the interest and associations of the keyword. We also collected approximately 11,000 news articles based on the top three keywords that are highly related to happiness, then derived issues between keywords through text mining analysis in SAS, and utilized them in defining services after ideas were conceived. Fourth, based on the characteristics identified through data analysis, we selected segmentation and targetingappropriate for service discovery. To this end, the characteristics of the factors were grouped and selected into four groups, and the profile was drawn up and the main target customers were selected. Fifth, based on the characteristics of the main target customers, interviewers were selected and the In-depthinterviews were conducted to discover the causes of happiness, causes of unhappiness, and needs for services. Sixth, we derive customer behavior patterns based on segment results and detailed interviews, and specify the objectives associated with the characteristics. Seventh, a typical persona using qualitative surveys and a persona using data were produced to analyze each characteristic and pros and cons by comparing the two personas. Existing market segmentation classifies customers based on purchasing factors, and UX methodology measures users' behavior variables to establish criteria and redefine users' classification. Utilizing these segment classification methods, applying the process of producinguser classification and persona in UX methodology will be able to utilize them as more accurate customer classification schemes. The significance of this study is summarized in two ways: First, the idea of using data to create a variety of services was linked to the UX methodology used to plan IT services by applying it in the hot topic era. Second, we further enhance user classification by applying segment analysis methods that are not currently used well in UX methodologies. To provide a consistent experience in creating a single service, from large to small, it is necessary to define customers with common goals. To this end, it is necessary to derive persona and persuade various stakeholders. Under these circumstances, designing a consistent experience from beginning to end, through fast and concrete user descriptions, would be a very effective way to produce a successful service.

A Study on Web-based Technology Valuation System (웹기반 지능형 기술가치평가 시스템에 관한 연구)

  • Sung, Tae-Eung;Jun, Seung-Pyo;Kim, Sang-Gook;Park, Hyun-Woo
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.1
    • /
    • pp.23-46
    • /
    • 2017
  • Although there have been cases of evaluating the value of specific companies or projects which have centralized on developed countries in North America and Europe from the early 2000s, the system and methodology for estimating the economic value of individual technologies or patents has been activated on and on. Of course, there exist several online systems that qualitatively evaluate the technology's grade or the patent rating of the technology to be evaluated, as in 'KTRS' of the KIBO and 'SMART 3.1' of the Korea Invention Promotion Association. However, a web-based technology valuation system, referred to as 'STAR-Value system' that calculates the quantitative values of the subject technology for various purposes such as business feasibility analysis, investment attraction, tax/litigation, etc., has been officially opened and recently spreading. In this study, we introduce the type of methodology and evaluation model, reference information supporting these theories, and how database associated are utilized, focusing various modules and frameworks embedded in STAR-Value system. In particular, there are six valuation methods, including the discounted cash flow method (DCF), which is a representative one based on the income approach that anticipates future economic income to be valued at present, and the relief-from-royalty method, which calculates the present value of royalties' where we consider the contribution of the subject technology towards the business value created as the royalty rate. We look at how models and related support information (technology life, corporate (business) financial information, discount rate, industrial technology factors, etc.) can be used and linked in a intelligent manner. Based on the classification of information such as International Patent Classification (IPC) or Korea Standard Industry Classification (KSIC) for technology to be evaluated, the STAR-Value system automatically returns meta data such as technology cycle time (TCT), sales growth rate and profitability data of similar company or industry sector, weighted average cost of capital (WACC), indices of industrial technology factors, etc., and apply adjustment factors to them, so that the result of technology value calculation has high reliability and objectivity. Furthermore, if the information on the potential market size of the target technology and the market share of the commercialization subject refers to data-driven information, or if the estimated value range of similar technologies by industry sector is provided from the evaluation cases which are already completed and accumulated in database, the STAR-Value is anticipated that it will enable to present highly accurate value range in real time by intelligently linking various support modules. Including the explanation of the various valuation models and relevant primary variables as presented in this paper, the STAR-Value system intends to utilize more systematically and in a data-driven way by supporting the optimal model selection guideline module, intelligent technology value range reasoning module, and similar company selection based market share prediction module, etc. In addition, the research on the development and intelligence of the web-based STAR-Value system is significant in that it widely spread the web-based system that can be used in the validation and application to practices of the theoretical feasibility of the technology valuation field, and it is expected that it could be utilized in various fields of technology commercialization.