• Title/Summary/Keyword: 카메라 모델

Search Result 1,041, Processing Time 0.028 seconds

Mobile Augmented Reality Application for Early Childhood Language Education (유아 언어 교육을 위한 모바일 증강현실 어플리케이션)

  • Kang, Sanghoon;Shin, Minwoo;Kim, Minji;Park, Hanhoon
    • Journal of Broadcast Engineering
    • /
    • v.23 no.6
    • /
    • pp.914-924
    • /
    • 2018
  • In this paper, we implement an Android application for infant language education using marker-based augmented reality. Combining animal word markers (noun), size/color word markers (adjective), and action word markers (verb) in puzzle form to make a simple sentence, the application shows virtual contents related to the content of the sentence. For example, when an animal marker is showed up on a camera, the corresponding animal appears. Additionally, when the motion markers are combined, the animal's appearance changes into an animation in which it acts. When a user touched a marker, user can hear the sound of the word, which gives an auditory effect, and by adding the rotation function, user can see the animation in any direction. Our goal is to increase infants' interest in learning language and also increase the effectiveness of education on the meaning of words and the structure of simple sentences, by encouraging them to actively participate in language learning through visual and auditory stimuli.

User Customized Realization of Virtual Earthquakes based on Visual Intelligence and Dynamic Simulation (시각지능 및 동적 시뮬레이션 기반의 사용자 맞춤형 가상 지진 실감화)

  • Kwon, Jihoe;Ryu, Dongwoo;Lee, Sangho
    • Journal of the Korean Society of Mineral and Energy Resources Engineers
    • /
    • v.55 no.6
    • /
    • pp.614-623
    • /
    • 2018
  • The recent occurrence of consecutive large earthquakes in the southeastern part of the Korean peninsula has brought significant attention to the prevention of earthquake damage in Korea. This article aims to explore a technology-based approach for earthquake drills using state-of-the-art visual intelligence and virtual reality technologies. The technical process consists of several stages, including acquisition of image information in living spaces using a camera, recognition of objects from the acquired image information, extraction of three dimensional geometric information, simulation of virtual earthquakes using dynamic modelling techniques such as the discrete element method, and realization of the simulated earthquake in a virtual reality environment. This article provides a comprehensive analysis of the individual processes at each stage of the technical process, a survey on the current status of related technologies, and discussion of the technical challenges in its execution.

Remote Control System using Face and Gesture Recognition based on Deep Learning (딥러닝 기반의 얼굴과 제스처 인식을 활용한 원격 제어)

  • Hwang, Kitae;Lee, Jae-Moon;Jung, Inhwan
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.20 no.6
    • /
    • pp.115-121
    • /
    • 2020
  • With the spread of IoT technology, various IoT applications using facial recognition are emerging. This paper describes the design and implementation of a remote control system using deep learning-based face recognition and hand gesture recognition. In general, an application system using face recognition consists of a part that takes an image in real time from a camera, a part that recognizes a face from the image, and a part that utilizes the recognized result. Raspberry PI, a single board computer that can be mounted anywhere, has been used to shoot images in real time, and face recognition software has been developed using tensorflow's FaceNet model for server computers and hand gesture recognition software using OpenCV. We classified users into three groups: Known users, Danger users, and Unknown users, and designed and implemented an application that opens automatic door locks only for Known users who have passed both face recognition and hand gestures.

A Deep Learning Based Device-free Indoor People Counting Using CSI (CSI를 활용한 딥러닝 기반의 실내 사람 수 추정 기법)

  • An, Hyun-seong;Kim, Seungku
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.24 no.7
    • /
    • pp.935-941
    • /
    • 2020
  • People estimation is important to provide IoT services. Most people counting technologies use camera or sensor data. However, the conventional technologies have the disadvantages of invasion of privacy and the need to install extra infrastructure. This paper proposes a method for estimating the number of people using a Wi-Fi AP. We use channel state information of Wi-Fi and analyze that using deep learning technology. It can be achieved by pre-installed Wi-Fi infrastructure that reduce cost for people estimation and privacy infringement. The proposed algorithm uses a k-binding data for pre-processing process and a 1D-CNN learning model. Two APs were installed to analyze the estimation results of six people. The result of the accurate number estimation was 64.8%, but the result of classifying the number of people into classes showed a high result of 84.5%. This algorithm is expected to be applicable to estimate the density of people in a small space.

Loitering Behavior Detection Using Shadow Removal and Chromaticity Histogram Matching (그림자 제거와 색도 히스토그램 비교를 이용한 배회행위 검출)

  • Park, Eun-Soo;Lee, Hyung-Ho;Yun, Myoung-Kyu;Kim, Min-Gyu;Kwak, Jong-Hoon;Kim, Hak-Il
    • Journal of the Korea Institute of Information Security & Cryptology
    • /
    • v.21 no.6
    • /
    • pp.171-181
    • /
    • 2011
  • Proposed in this paper is the intelligent video surveillance system to effectively detect multiple loitering objects even that disappear from the out of camera's field of view and later return to a target zone. After the background and foreground are segmented using Gaussian mixture model and shadows are removed, the objects returning to the target zone is recognized using the chromaticity histogram and the duration of loitering is preserved. For more accurate measurement of the loitering behavior, the camera calibration is also applied to map the image plane to the real-world ground. Hence, the loitering behavior can be detected by considering the time duration of the object's existence in the real-world space. The experiment was performed using loitering video and all of the loitering behaviors are accurately detected.

Detection of Number and Character Area of License Plate Using Deep Learning and Semantic Image Segmentation (딥러닝과 의미론적 영상분할을 이용한 자동차 번호판의 숫자 및 문자영역 검출)

  • Lee, Jeong-Hwan
    • Journal of the Korea Convergence Society
    • /
    • v.12 no.1
    • /
    • pp.29-35
    • /
    • 2021
  • License plate recognition plays a key role in intelligent transportation systems. Therefore, it is a very important process to efficiently detect the number and character areas. In this paper, we propose a method to effectively detect license plate number area by applying deep learning and semantic image segmentation algorithm. The proposed method is an algorithm that detects number and text areas directly from the license plate without preprocessing such as pixel projection. The license plate image was acquired from a fixed camera installed on the road, and was used in various real situations taking into account both weather and lighting changes. The input images was normalized to reduce the color change, and the deep learning neural networks used in the experiment were Vgg16, Vgg19, ResNet18, and ResNet50. To examine the performance of the proposed method, we experimented with 500 license plate images. 300 sheets were used for learning and 200 sheets were used for testing. As a result of computer simulation, it was the best when using ResNet50, and 95.77% accuracy was obtained.

Extraction of Skin Regions through Filtering-based Noise Removal (필터링 기반의 잡음 제거를 통한 피부 영역의 추출)

  • Jang, Seok-Woo
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.21 no.12
    • /
    • pp.672-678
    • /
    • 2020
  • Ultra-high-speed images that accurately depict the minute movements of objects have become common as low-cost and high-performance cameras that can film at high speeds have emerged. In this paper, the proposed method removes unexpected noise contained in images after input at high speed, and then extracts an area of interest that can represent personal information, such as skin areas, from the image in which noise has been removed. In this paper, noise generated by abnormal electrical signals is removed by applying bilateral filters. A color model created through pre-learning is then used to extract the area of interest that represents the personal information contained within the image. Experimental results show that the introduced algorithms remove noise from high-speed images and then extract the area of interest robustly. The approach presented in this paper is expected to be useful in various applications related to computer vision, such as image preprocessing, noise elimination, tracking and monitoring of target areas, etc.

Acquisition of Region of Interest through Illumination Correction in Dynamic Image Data (동영상 데이터에서 조명 보정을 사용한 관심 영역의 획득)

  • Jang, Seok-Woo
    • Journal of the Korea Academia-Industrial cooperation Society
    • /
    • v.22 no.3
    • /
    • pp.439-445
    • /
    • 2021
  • Low-cost, ultra-high-speed cameras, made possible by the development of image sensors and small displays, can be very useful in image processing and pattern recognition. This paper introduces an algorithm that corrects irregular lighting from a high-speed image that is continuously input with a slight time interval, and which then obtains an exposed skin color region that is the area of interest in a person from the corrected image. In this study, the non-uniform lighting effect from a received high-speed image is first corrected using a frame blending technique. Then, the region of interest is robustly obtained from the input high-speed color image by applying an elliptical skin color distribution model generated from iterative learning in advance. Experimental results show that the approach presented in this paper corrects illumination in various types of color images, and then accurately acquires the region of interest. The algorithm proposed in this study is expected to be useful in various types of practical applications related to image recognition, such as face recognition and tracking, lighting correction, and video indexing and retrieval.

Development of An Interactive System Prototype Using Imitation Learning to Induce Positive Emotion (긍정감정을 유도하기 위한 모방학습을 이용한 상호작용 시스템 프로토타입 개발)

  • Oh, Chanhae;Kang, Changgu
    • The Journal of Korea Institute of Information, Electronics, and Communication Technology
    • /
    • v.14 no.4
    • /
    • pp.239-246
    • /
    • 2021
  • In the field of computer graphics and HCI, there are many studies on systems that create characters and interact naturally. Such studies have focused on the user's response to the user's behavior, and the study of the character's behavior to elicit positive emotions from the user remains a difficult problem. In this paper, we develop a prototype of an interaction system to elicit positive emotions from users according to the movement of virtual characters using artificial intelligence technology. The proposed system is divided into face recognition and motion generation of a virtual character. A depth camera is used for face recognition, and the recognized data is transferred to motion generation. We use imitation learning as a learning model. In motion generation, random actions are performed according to the first user's facial expression data, and actions that the user can elicit positive emotions are learned through continuous imitation learning.

Efficient Self-supervised Learning Techniques for Lightweight Depth Completion (경량 깊이완성기술을 위한 효율적인 자기지도학습 기법 연구)

  • Park, Jae-Hyuck;Min, Kyoung-Wook;Choi, Jeong Dan
    • The Journal of The Korea Institute of Intelligent Transport Systems
    • /
    • v.20 no.6
    • /
    • pp.313-330
    • /
    • 2021
  • In an autonomous driving system equipped with a camera and lidar, depth completion techniques enable dense depth estimation. In particular, using self-supervised learning it is possible to train the depth completion network even without ground truth. In actual autonomous driving, such depth completion should have very short latency as it is the input of other algorithms. So, rather than complicate the network structure to increase the accuracy like previous studies, this paper focuses on network latency. We design a U-Net type network with RegNet encoders optimized for GPU computation. Instead, this paper presents several techniques that can increase accuracy during the process of self-supervised learning. The proposed techniques increase the robustness to unreliable lidar inputs. Also, they improve the depth quality for edge and sky regions based on the semantic information extracted in advance. Our experiments confirm that our model is very lightweight (2.42 ms at 1280x480) but resistant to noise and has qualities close to the latest studies.