• Title/Summary/Keyword: ART Algorithm

Search Result 579, Processing Time 0.022 seconds

Complexity Estimation Based Work Load Balancing for a Parallel Lidar Waveform Decomposition Algorithm

  • Jung, Jin-Ha;Crawford, Melba M.;Lee, Sang-Hoon
    • Korean Journal of Remote Sensing
    • /
    • v.25 no.6
    • /
    • pp.547-557
    • /
    • 2009
  • LIDAR (LIght Detection And Ranging) is an active remote sensing technology which provides 3D coordinates of the Earth's surface by performing range measurements from the sensor. Early small footprint LIDAR systems recorded multiple discrete returns from the back-scattered energy. Recent advances in LIDAR hardware now make it possible to record full digital waveforms of the returned energy. LIDAR waveform decomposition involves separating the return waveform into a mixture of components which are then used to characterize the original data. The most common statistical mixture model used for this process is the Gaussian mixture. Waveform decomposition plays an important role in LIDAR waveform processing, since the resulting components are expected to represent reflection surfaces within waveform footprints. Hence the decomposition results ultimately affect the interpretation of LIDAR waveform data. Computational requirements in the waveform decomposition process result from two factors; (1) estimation of the number of components in a mixture and the resulting parameter estimates, which are inter-related and cannot be solved separately, and (2) parameter optimization does not have a closed form solution, and thus needs to be solved iteratively. The current state-of-the-art airborne LIDAR system acquires more than 50,000 waveforms per second, so decomposing the enormous number of waveforms is challenging using traditional single processor architecture. To tackle this issue, four parallel LIDAR waveform decomposition algorithms with different work load balancing schemes - (1) no weighting, (2) a decomposition results-based linear weighting, (3) a decomposition results-based squared weighting, and (4) a decomposition time-based linear weighting - were developed and tested with varying number of processors (8-256). The results were compared in terms of efficiency. Overall, the decomposition time-based linear weighting work load balancing approach yielded the best performance among four approaches.

Alzheimer's Disease Classification with Automated MRI Biomarker Detection Using Faster R-CNN for Alzheimer's Disease Diagnosis (치매 진단을 위한 Faster R-CNN 활용 MRI 바이오마커 자동 검출 연동 분류 기술 개발)

  • Son, Joo Hyung;Kim, Kyeong Tae;Choi, Jae Young
    • Journal of Korea Multimedia Society
    • /
    • v.22 no.10
    • /
    • pp.1168-1177
    • /
    • 2019
  • In order to diagnose and prevent Alzheimer's Disease (AD), it is becoming increasingly important to develop a CAD(Computer-aided Diagnosis) system for AD diagnosis, which provides effective treatment for patients by analyzing 3D MRI images. It is essential to apply powerful deep learning algorithms in order to automatically classify stages of Alzheimer's Disease and to develop a Alzheimer's Disease support diagnosis system that has the function of detecting hippocampus and CSF(Cerebrospinal fluid) which are important biomarkers in diagnosis of Alzheimer's Disease. In this paper, for AD diagnosis, we classify a given MRI data into three categories of AD, mild cognitive impairment, and normal control according by applying 3D brain MRI image to the Faster R-CNN model and detect hippocampus and CSF in MRI image. To do this, we use the 2D MRI slice images extracted from the 3D MRI data of the Faster R-CNN, and perform the widely used majority voting algorithm on the resulting bounding box labels for classification. To verify the proposed method, we used the public ADNI data set, which is the standard brain MRI database. Experimental results show that the proposed method achieves impressive classification performance compared with other state-of-the-art methods.

Study on Data Normalization and Representation for Quantitative Analysis of EEG Signals (뇌파 신호의 정량적 분석을 위한 데이터 정규화 및 표현기법 연구)

  • Hwang, Taehun;Kim, Jin Heon
    • Asia-pacific Journal of Multimedia Services Convergent with Art, Humanities, and Sociology
    • /
    • v.9 no.6
    • /
    • pp.729-738
    • /
    • 2019
  • Recently, we aim to improve the quality of virtual reality contents based on quantitative analysis results of emotions through combination of emotional recognition field and virtual reality field. Emotions are analyzed based on the participant's vital signs. Much research has been done in terms of signal analysis, but the methodology for quantifying emotions has not been fully discussed. In this paper, we propose a normalization function design and expression method to quantify the emotion between various bio - signals. Use the Brute force algorithm to find the optimal parameters of the normalization function and improve the confidence score of the parameters found using the true and false scores defined in this paper. As a result, it is possible to automate the parameter determination of the bio-signal normalization function depending on the experience, and the emotion can be analyzed quantitatively based on this.

Analysis of textbook contents according to the 2015 revised Elementary Software Education Achievement Standards (2015 개정 초등 소프트웨어교육 성취기준에 따른 교과서 내용 분석)

  • Kim, Jeongrang
    • Journal of The Korean Association of Information Education
    • /
    • v.23 no.1
    • /
    • pp.9-18
    • /
    • 2019
  • In this study, the contents of the Software Education area included in the Practical Art textbooks of the elementary school were analyzed based on the achievement standards of the Software Education in the 2015 revised curriculum for exploring the characteristics of the Software Education and the direction to go. According to the results of the analysis, the five achievement standards presented by the Ministry of Education were assigned to 17 or 18 times according to the textbooks. Unplugged activities related to Algorithm and unplugged activities related to Computer Science were included in textbooks. Contents of [6P04-09] and [6P04-10] standards were shown differently according to textbooks.Based on the results, it is suggested that Subsequent studies with quantitative and qualitative analysis for six kinds of textbooks should be continue.

The Development of the Recovery System of the Destroyed Epigraph - Focused on the Chinese standard script - (훼손된 금석문 판독시스템 개발 - 해서체를 중심으로 -)

  • Jang, Seon-Phil
    • Korean Journal of Heritage: History & Science
    • /
    • v.50 no.2
    • /
    • pp.80-93
    • /
    • 2017
  • This study proposes a new scientific measurement method for damaged epigraph. In this new method, the Chinese characters are converted and coordinates are created for this measurement. This method is then used to decipher partially damaged characters from the parts of the coordinated characters that are damaged and intact. The Chinese characters are divided into 9 square parts by the position of their Chinese Radicals. The unknown characters are then compared and deciphered dependent upon the character shape in 9 square parts that have been created. This method is more scientific, accurate, and makes it easier to find related characters than deciphering through contexts, which is current method. When creating a new software based on this algorithm, it will be especially useful in deciphering an old manuscript or a epigraph that made ancient Chinese characters which are not currently in use. This study will also be helpful in deciphering semi-cursive styled or cursive styled epigraph, as well as semi-cursive styled or cursive styled damaged characters during follow-up research.

A Batch Processing Algorithm for Moving k-Nearest Neighbor Queries in Dynamic Spatial Networks

  • Cho, Hyung-Ju
    • Journal of the Korea Society of Computer and Information
    • /
    • v.26 no.4
    • /
    • pp.63-74
    • /
    • 2021
  • Location-based services (LBSs) are expected to process a large number of spatial queries, such as shortest path and k-nearest neighbor queries that arrive simultaneously at peak periods. Deploying more LBS servers to process these simultaneous spatial queries is a potential solution. However, this significantly increases service operating costs. Recently, batch processing solutions have been proposed to process a set of queries using shareable computation. In this study, we investigate the problem of batch processing moving k-nearest neighbor (MkNN) queries in dynamic spatial networks, where the travel time of each road segment changes frequently based on the traffic conditions. LBS servers based on one-query-at-a-time processing often fail to process simultaneous MkNN queries because of the significant number of redundant computations. We aim to improve the efficiency algorithmically by processing MkNN queries in batches and reusing sharable computations. Extensive evaluation using real-world roadmaps shows the superiority of our solution compared with state-of-the-art methods.

Vision-based Food Shape Recognition and Its Positioning for Automated Production of Custom Cakes (주문형 케이크 제작 자동화를 위한 영상 기반 식품 모양 인식 및 측위)

  • Oh, Jang-Sub;Lee, Jaesung
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.24 no.10
    • /
    • pp.1280-1287
    • /
    • 2020
  • This paper proposes a vision-based food recognition method for automated production of custom cakes. A small camera module mounted on a food art printer recognizes objects' shape and estimates their center points through image processing. Through the perspective transformation, the top-view image is obtained from the original image taken at an oblique position. The line and circular hough transformations are applied to recognize square and circular shapes respectively. In addition, the center of gravity of each figure are accurately detected in units of pixels. The test results show that the shape recognition rate is more than 98.75% under 180 ~ 250 lux of light and the positioning error rate is less than 0.87% under 50 ~ 120 lux. These values sufficiently meet the needs of the corresponding market. In addition, the processing delay is also less than 0.5 seconds per frame, so the proposed algorithm is suitable for commercial purpose.

A Study on the Automated Payment System for Artificial Intelligence-Based Product Recognition in the Age of Contactless Services

  • Kim, Heeyoung;Hong, Hotak;Ryu, Gihwan;Kim, Dongmin
    • International Journal of Advanced Culture Technology
    • /
    • v.9 no.2
    • /
    • pp.100-105
    • /
    • 2021
  • Contactless service is rapidly emerging as a new growth strategy due to consumers who are reluctant to the face-to-face situation in the global pandemic of coronavirus disease 2019 (COVID-19), and various technologies are being developed to support the fast-growing contactless service market. In particular, the restaurant industry is one of the most desperate industrial fields requiring technologies for contactless service, and the representative technical case should be a kiosk, which has the advantage of reducing labor costs for the restaurant owners and provides psychological relaxation and satisfaction to the customer. In this paper, we propose a solution to the restaurant's store operation through the unmanned kiosk using a state-of-the-art artificial intelligence (AI) technology of image recognition. Especially, for the products that do not have barcodes in bakeries, fresh foods (fruits, vegetables, etc.), and autonomous restaurants on highways, which cause increased labor costs and many hassles, our proposed system should be very useful. The proposed system recognizes products without barcodes on the ground of image-based AI algorithm technology and makes automatic payments. To test the proposed system feasibility, we established an AI vision system using a commercial camera and conducted an image recognition test by training object detection AI models using donut images. The proposed system has a self-learning system with mismatched information in operation. The self-learning AI technology allows us to upgrade the recognition performance continuously. We proposed a fully automated payment system with AI vision technology and showed system feasibility by the performance test. The system realizes contactless service for self-checkout in the restaurant business area and improves the cost-saving in managing human resources.

EXECUTION TIME AND POWER CONSUMPTION OPTIMIZATION in FOG COMPUTING ENVIRONMENT

  • Alghamdi, Anwar;Alzahrani, Ahmed;Thayananthan, Vijey
    • International Journal of Computer Science & Network Security
    • /
    • v.21 no.1
    • /
    • pp.137-142
    • /
    • 2021
  • The Internet of Things (IoT) paradigm is at the forefront of present and future research activities. The huge amount of sensing data from IoT devices needing to be processed is increasing dramatically in volume, variety, and velocity. In response, cloud computing was involved in handling the challenges of collecting, storing, and processing jobs. The fog computing technology is a model that is used to support cloud computing by implementing pre-processing jobs close to the end-user for realizing low latency, less power consumption in the cloud side, and high scalability. However, it may be that some resources in fog computing networks are not suitable for some kind of jobs, or the number of requests increases outside capacity. So, it is more efficient to decrease sending jobs to the cloud. Hence some other fog resources are idle, and it is better to be federated rather than forwarding them to the cloud server. Obviously, this issue affects the performance of the fog environment when dealing with big data applications or applications that are sensitive to time processing. This research aims to build a fog topology job scheduling (FTJS) to schedule the incoming jobs which are generated from the IoT devices and discover all available fog nodes with their capabilities. Also, the fog topology job placement algorithm is introduced to deploy jobs into appropriate resources in the network effectively. Finally, by comparing our result with the state-of-art first come first serve (FCFS) scheduling technique, the overall execution time is reduced significantly by approximately 20%, the energy consumption in the cloud side is reduced by 18%.

Age and Gender Classification with Small Scale CNN (소규모 합성곱 신경망을 사용한 연령 및 성별 분류)

  • Jamoliddin, Uraimov;Yoo, Jae Hung
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.17 no.1
    • /
    • pp.99-104
    • /
    • 2022
  • Artificial intelligence is getting a crucial part of our lives with its incredible benefits. Machines outperform humans in recognizing objects in images, particularly in classifying people into correct age and gender groups. In this respect, age and gender classification has been one of the hot topics among computer vision researchers in recent decades. Deployment of deep Convolutional Neural Network(: CNN) models achieved state-of-the-art performance. However, the most of CNN based architectures are very complex with several dozens of training parameters so they require much computation time and resources. For this reason, we propose a new CNN-based classification algorithm with significantly fewer training parameters and training time compared to the existing methods. Despite its less complexity, our model shows better accuracy of age and gender classification on the UTKFace dataset.