• Title/Summary/Keyword: Intelligent Image Analysis System

Search Result 152, Processing Time 0.025 seconds

Emotion Recognition and Expression System of User using Multi-Modal Sensor Fusion Algorithm (다중 센서 융합 알고리즘을 이용한 사용자의 감정 인식 및 표현 시스템)

  • Yeom, Hong-Gi;Joo, Jong-Tae;Sim, Kwee-Bo
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.18 no.1
    • /
    • pp.20-26
    • /
    • 2008
  • As they have more and more intelligence robots or computers these days, so the interaction between intelligence robot(computer) - human is getting more and more important also the emotion recognition and expression are indispensable for interaction between intelligence robot(computer) - human. In this paper, firstly we extract emotional features at speech signal and facial image. Secondly we apply both BL(Bayesian Learning) and PCA(Principal Component Analysis), lastly we classify five emotions patterns(normal, happy, anger, surprise and sad) also, we experiment with decision fusion and feature fusion to enhance emotion recognition rate. The decision fusion method experiment on emotion recognition that result values of each recognition system apply Fuzzy membership function and the feature fusion method selects superior features through SFS(Sequential Forward Selection) method and superior features are applied to Neural Networks based on MLP(Multi Layer Perceptron) for classifying five emotions patterns. and recognized result apply to 2D facial shape for express emotion.

Development of Performance Evaluation Formula for Deep Learning Image Analysis System (딥러닝 영상분석 시스템의 성능평가 산정식 개발)

  • Hyun Ho Son;Yun Sang Kim;Choul Ki Lee
    • The Journal of The Korea Institute of Intelligent Transport Systems
    • /
    • v.22 no.4
    • /
    • pp.78-96
    • /
    • 2023
  • Urban traffic information is collected by various systems such as VDS, DSRC, and radar. Recently, with the development of deep learning technology, smart intersection systems are expanding, are more widely distributed, and it is possible to collect a variety of information such as traffic volume, and vehicle type and speed. However, as a result of reviewing related literature, the performance evaluation criteria so far are rbs-based evaluation systems that do not consider the deep learning area, and only consider the percent error of 'reference value-measured value'. Therefore, a new performance evaluation method is needed. Therefore, in this study, individual error, interval error, and overall error are calculated by using a formula that considers deep learning performance indicators such as precision and recall based on data ratio and weight. As a result, error rates for measurement value 1 were 3.99 and 3.54, and rates for measurement value 2 were 5.34 and 5.07.

Performance Analysis of Face Recognition by Face Image resolutions using CNN without Backpropergation and LDA (역전파가 제거된 CNN과 LDA를 이용한 얼굴 영상 해상도별 얼굴 인식률 분석)

  • Moon, Hae-Min;Park, Jin-Won;Pan, Sung Bum
    • Smart Media Journal
    • /
    • v.5 no.1
    • /
    • pp.24-29
    • /
    • 2016
  • To satisfy the needs of high-level intelligent surveillance system, it shall be able to extract objects and classify to identify precise information on the object. The representative method to identify one's identity is face recognition that is caused a change in the recognition rate according to environmental factors such as illumination, background and angle of camera. In this paper, we analyze the robust face recognition of face image by changing the distance through a variety of experiments. The experiment was conducted by real face images of 1m to 5m. The method of face recognition based on Linear Discriminant Analysis show the best performance in average 75.4% when a large number of face images per one person is used for training. However, face recognition based on Convolution Neural Network show the best performance in average 69.8% when the number of face images per one person is less than five. In addition, rate of low resolution face recognition decrease rapidly when the size of the face image is smaller than $15{\times}15$.

A Vehicular License Plate Recognition Framework For Skewed Images

  • Arafat, M.Y.;Khairuddin, A.S.M.;Paramesran, R.
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.12 no.11
    • /
    • pp.5522-5540
    • /
    • 2018
  • Vehicular license plate (LP) recognition system has risen as a significant field of research recently because various explorations are currently being conducted by the researchers to cope with the challenges of LPs which include different illumination and angular situations. This research focused on restricted conditions such as using image of only one vehicle, stationary background, no angular adjustment of the skewed images. A real time vehicular LP recognition scheme is proposed for the skewed images for detection, segmentation and recognition of LP. In this research, a polar co-ordinate transformation procedure is implemented to adjust the skewed vehicular images. Besides that, window scanning procedure is utilized for the candidate localization that is based on the texture characteristics of the image. Then, connected component analysis (CCA) is implemented to the binary image for character segmentation where the pixels get connected in an eight-point neighbourhood process. Finally, optical character recognition is implemented for the recognition of the characters. For measuring the performance of this experiment, 300 skewed images of different illumination conditions with various tilt angles have been tested. The results show that proposed method able to achieve accuracy of 96.3% in localizing, 95.4% in segmenting and 94.2% in recognizing the LPs with an average localization time of 0.52s.

Locating Text in Web Images Using Image Based Approaches (웹 이미지로부터 이미지기반 문자추출)

  • Chin, Seongah;Choo, Moonwon
    • Journal of Intelligence and Information Systems
    • /
    • v.8 no.1
    • /
    • pp.27-39
    • /
    • 2002
  • A locating text technique capable of locating and extracting text blocks in various Web images is presented here. Until now this area of work has been ignored by researchers even if this sort of text may be meaningful for internet users. The algorithms associated with the technique work without prior knowledge of the text orientation, size or font. In the work presented in this research, our text extraction algorithm utilizes useful edge detection followed by histogram analysis on the genuine characteristics of letters defined by text clustering region, to properly perform extraction of the text region that does not depend on font styles and sizes. By a number of experiments we have showed impressively acceptable results.

  • PDF

Escape Route Prediction and Tracking System using Artificial Intelligence (인공지능을 활용한 도주경로 예측 및 추적 시스템)

  • Yang, Bum-suk;Park, Dea-woo
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2022.05a
    • /
    • pp.225-227
    • /
    • 2022
  • Now In Seoul, about 75,000 CCTVs are installed in 25 district offices. Each ward office in Seoul has built a control center for CCTV control and is building information such as people, vehicle types, license plate recognition and color classification into big data through 24-hour artificial intelligence intelligent image analysis. Seoul Metropolitan Government has signed MOUs with the Ministry of Land, Infrastructure and Transport, the National Police Agency, the Fire Service, the Ministry of Justice, and the military base to enable rapid response to emergency/emergency situations. In other words, we are building a smart city that is safe and can prevent disasters by providing CCTV images of each ward office. In this paper, the CCTV image is designed to extract the characteristics of the vehicle and personnel when an incident occurs through artificial intelligence, and based on this, predict the escape route and enable continuous tracking. It is designed so that the AI automatically selects and displays the CCTV image of the route. It is designed to expand the smart city integration platform by providing image information and extracted information to the adjacent ward office when the escape route of a person or vehicle related to an incident is expected to an area other than the relevant jurisdiction. This paper will contribute as basic data to the development of smart city integrated platform research.

  • PDF

Autonomous Mobile Robot System Using Adaptive Spatial Coordinates Detection Scheme based on Stereo Camera (스테레오 카메라 기반의 적응적인 공간좌표 검출 기법을 이용한 자율 이동로봇 시스템)

  • Ko Jung-Hwan;Kim Sung-Il;Kim Eun-Soo
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.31 no.1C
    • /
    • pp.26-35
    • /
    • 2006
  • In this paper, an automatic mobile robot system for a intelligent path planning using the detection scheme of the spatial coordinates based on stereo camera is proposed. In the proposed system, face area of a moving person is detected from a left image among the stereo image pairs by using the YCbCr color model and its center coordinates are computed by using the centroid method and then using these data, the stereo camera embedded on the mobile robot can be controlled for tracking the moving target in real-time. Moreover, using the disparity map obtained from the left and right images captured by the tracking-controlled stereo camera system and the perspective transformation between a 3-D scene and an image plane, depth information can be detected. Finally, based-on the analysis of these calculated coordinates, a mobile robot system is derived as a intelligent path planning and a estimation. From some experiments on robot driving with 240 frames of the stereo images, it is analyzed that error ratio between the calculated and measured values of the distance between the mobile robot and the objects, and relative distance between the other objects is found to be very low value of $2.19\%$ and $1.52\%$ on average, respectably.

Intelligent Diagnosis Assistant System of Capsule Endoscopy Video Through Analysis of Video Frames (영상 프레임 분석을 통한 대용량 캡슐내시경 영상의 지능형 판독보조 시스템)

  • Lee, H.G.;Choi, H.K.;Lee, D.H.;Lee, S.C.
    • Journal of Intelligence and Information Systems
    • /
    • v.15 no.2
    • /
    • pp.33-48
    • /
    • 2009
  • Capsule endoscopy is one of the most remarkable inventions in last ten years. Causing less pain for patients, diagnosis for entire digestive system has been considered as a most convenience method over a normal endoscope. However, it is known that the diagnosis process typically requires very long inspection time for clinical experts because of considerably many duplicate images of same areas in human digestive system due to uncontrollable movement of a capsule endoscope. In this paper, we propose a method for clinical diagnosticians to get highly valuable information from capsule-endoscopy video. Our software system consists of three global maps, such as movement map, characteristic map, and brightness map, in temporal domain for entire sequence of the input video. The movement map can be used for effectively removing duplicated adjacent images. The characteristic and brightness maps provide frame content analyses that can be quickly used for segmenting regions or locating some features(such as blood) in the stream. Our experiments show the results of four patients having different health conditions. The result maps clearly capture the movements and characteristics from the image frames. Our method may help the diagnosticians quickly search the locations of lesion, bleeding, or some other interesting areas.

  • PDF

A Study on Intelligent Skin Image Identification From Social media big data

  • Kim, Hyung-Hoon;Cho, Jeong-Ran
    • Journal of the Korea Society of Computer and Information
    • /
    • v.27 no.9
    • /
    • pp.191-203
    • /
    • 2022
  • In this paper, we developed a system that intelligently identifies skin image data from big data collected from social media Instagram and extracts standardized skin sample data for skin condition diagnosis and management. The system proposed in this paper consists of big data collection and analysis stage, skin image analysis stage, training data preparation stage, artificial neural network training stage, and skin image identification stage. In the big data collection and analysis stage, big data is collected from Instagram and image information for skin condition diagnosis and management is stored as an analysis result. In the skin image analysis stage, the evaluation and analysis results of the skin image are obtained using a traditional image processing technique. In the training data preparation stage, the training data were prepared by extracting the skin sample data from the skin image analysis result. And in the artificial neural network training stage, an artificial neural network AnnSampleSkin that intelligently predicts the skin image type using this training data was built up, and the model was completed through training. In the skin image identification step, skin samples are extracted from images collected from social media, and the image type prediction results of the trained artificial neural network AnnSampleSkin are integrated to intelligently identify the final skin image type. The skin image identification method proposed in this paper shows explain high skin image identification accuracy of about 92% or more, and can provide standardized skin sample image big data. The extracted skin sample set is expected to be used as standardized skin image data that is very efficient and useful for diagnosing and managing skin conditions.

Evaluation of Drivers's Preference on Messages Delivered by VMS (VMS 메시지 이용자 선호도 평가)

  • Yeon, Ji-Youn;Kim, Tae-Hyung;Oh, Cheol
    • The Journal of The Korea Institute of Intelligent Transport Systems
    • /
    • v.7 no.4
    • /
    • pp.36-48
    • /
    • 2008
  • Variable Message Signs(VMS) is one of the important subsystems of Intelligent Transportation Systems (ITS), which is useful for providing drivers with real-time information on traffic, roadway or weather conditions. In order to maximize the effectiveness of the VMS system, it is required to reflect drivers's preference on designing and operating the system. In this context, this study was conducted to develop strategies to deliver the messages in an efficient manner while many other previous studies focused primarily on the contents of the messages, from drivers's perspectives. Drivers's preference on message expression formats and message display orders were investigated through image analysis in the perspectives of a total of 40 subjects. With respect to message expression formats, drivers preferred Gulim or Dodum in font style, middle arrangement of the letters, pictogram combination as opposed to letters only, blank time less than 0.5 sec, appearing message in animation effect, messages in single phase, non-flashing message. In the matter of message display orders, drivers preferred to obtain link or traffic information in the lust place. Then, they desired to be informed of roadway condition and instructions or recommendations for drivers to cope with unexpected events among various messages for traffic condition.

  • PDF