• Title/Summary/Keyword: ART2 Neural network

Search Result 136, Processing Time 0.031 seconds

Emotion Recognition and Expression System of Robot Based on 2D Facial Image (2D 얼굴 영상을 이용한 로봇의 감정인식 및 표현시스템)

  • Lee, Dong-Hoon;Sim, Kwee-Bo
    • Journal of Institute of Control, Robotics and Systems
    • /
    • v.13 no.4
    • /
    • pp.371-376
    • /
    • 2007
  • This paper presents an emotion recognition and its expression system of an intelligent robot like a home robot or a service robot. Emotion recognition method in the robot is used by a facial image. We use a motion and a position of many facial features. apply a tracking algorithm to recognize a moving user in the mobile robot and eliminate a skin color of a hand and a background without a facial region by using the facial region detecting algorithm in objecting user image. After normalizer operations are the image enlarge or reduction by distance of the detecting facial region and the image revolution transformation by an angel of a face, the mobile robot can object the facial image of a fixing size. And materialize a multi feature selection algorithm to enable robot to recognize an emotion of user. In this paper, used a multi layer perceptron of Artificial Neural Network(ANN) as a pattern recognition art, and a Back Propagation(BP) algorithm as a learning algorithm. Emotion of user that robot recognized is expressed as a graphic LCD. At this time, change two coordinates as the number of times of emotion expressed in ANN, and change a parameter of facial elements(eyes, eyebrows, mouth) as the change of two coordinates. By materializing the system, expressed the complex emotion of human as the avatar of LCD.

CCTV-Based Multi-Factor Authentication System

  • Kwon, Byoung-Wook;Sharma, Pradip Kumar;Park, Jong-Hyuk
    • Journal of Information Processing Systems
    • /
    • v.15 no.4
    • /
    • pp.904-919
    • /
    • 2019
  • Many security systems rely solely on solutions based on Artificial Intelligence, which are weak in nature. These security solutions can be easily manipulated by malicious users who can gain unlawful access. Some security systems suggest using fingerprint-based solutions, but they can be easily deceived by copying fingerprints with clay. Image-based security is undoubtedly easy to manipulate, but it is also a solution that does not require any special training on the part of the user. In this paper, we propose a multi-factor security framework that operates in a three-step process to authenticate the user. The motivation of the research lies in utilizing commonly available and inexpensive devices such as onsite CCTV cameras and smartphone camera and providing fully secure user authentication. We have used technologies such as Argon2 for hashing image features and physically unclonable identification for secure device-server communication. We also discuss the methodological workflow of the proposed multi-factor authentication framework. In addition, we present the service scenario of the proposed model. Finally, we analyze qualitatively the proposed model and compare it with state-of-the-art methods to evaluate the usability of the model in real-world applications.

Proposing a gamma radiation based intelligent system for simultaneous analyzing and detecting type and amount of petroleum by-products

  • Roshani, Mohammadmehdi;Phan, Giang;Faraj, Rezhna Hassan;Phan, Nhut-Huan;Roshani, Gholam Hossein;Nazemi, Behrooz;Corniani, Enrico;Nazemi, Ehsan
    • Nuclear Engineering and Technology
    • /
    • v.53 no.4
    • /
    • pp.1277-1283
    • /
    • 2021
  • It is important for operators of poly-pipelines in petroleum industry to continuously monitor characteristics of transferred fluid such as its type and amount. To achieve this aim, in this study a dual energy gamma attenuation technique in combination with artificial neural network (ANN) is proposed to simultaneously determine type and amount of four different petroleum by-products. The detection system is composed of a dual energy gamma source, including americium-241 and barium-133 radioisotopes, and one 2.54 cm × 2.54 cm sodium iodide detector for recording the transmitted photons. Two signals recorded in transmission detector, namely the counts under photo peak of Americium-241 with energy of 59.5 keV and the counts under photo peak of Barium-133 with energy of 356 keV, were applied to the ANN as the two inputs and volume percentages of petroleum by-products were assigned as the outputs.

Hybrid adaptive neuro-fuzzy inference system method for energy absorption of nano-composite reinforced beam with piezoelectric face-sheets

  • Lili Xiao
    • Advances in nano research
    • /
    • v.14 no.2
    • /
    • pp.141-154
    • /
    • 2023
  • Effects of viscoelastic foundation on vibration of curved-beam structure with clamped and simply-supported boundary conditions is investigated in this study. In doing so, a micro-scale laminate composite beam with two piezoelectric face layer with a carbon nanotube reinforces composite core is considered. The whole beam structure is laid on a viscoelastic substrate which normally occurred in actual conditions. Due to small scale of the structure non-classical elasticity theory provided more accurate results. Therefore, nonlocal strain gradient theory is employed here to capture both nano-scale effects on carbon nanotubes and microscale effects because of overall scale of the structure. Equivalent homogenous properties of the composite core is obtained using Halpin-Tsai equation. The equations of motion is derived considering energy terms of the beam and variational principle in minimizing total energy. The boundary condition is assumed to be clamped at one end and simply supported at the other end. Due to nonlinear terms in the equations of motion, semi-analytical method of general differential quadrature method is engaged to solve the equations. In addition, due to complexity in developing and solving equations of motion of arches, an artificial neural network is design and implemented to capture effects of different parameters on the inplane vibration of sandwich arches. At the end, effects of several parameters including nonlocal and gradient parameters, geometrical aspect ratios and substrate constants of the structure on the natural frequency and amplitude is derived. It is observed that increasing nonlocal and gradient parameters have contradictory effects of the amplitude and frequency of vibration of the laminate beam.

Hierarchical Recognition of English Calling Card by Using Multiresolution Images and Enhanced RBF Network (다해상도 영상과 개선된 RBF 네트워크를 이용한 계층적 영문 명함 인식)

  • Kim, Kwang-Baek;Kim, Young-Ju
    • The KIPS Transactions:PartB
    • /
    • v.10B no.4
    • /
    • pp.443-450
    • /
    • 2003
  • In this paper, we proposed the novel hierarchical algorithm for the recognition of English calling cards that processes multiresolution images of calling cards hierarchically to extract individual characters and recognizes the extracted characters by using the enhanced neural network method. The hierarchical recognition algorithm generates multiresolution images of calling cards, and each processing step in the algorithm selects and processes the image with suitable resolution for lower processing overhead and improved output. That is, first, the image of 1/3 times resolution, to which the horizontal smearing method is applied, is used to extract the areas including only characters from the calling card image, and next, by applying the vertical smearing and the contour tracking masking, the image of a half time resolution is used to extract individual characters from the character string areas. Lastly, the original image is used in the recognition step, because the image includes the morphological information of characters accurately. And for the recognition of characters with diverse font types and various sizes, the enhanced RBF network that improves the middle layer based on the ART1 was proposed and applied. The results of experiments on a large number of calling card images showed that the proposed algorithm is greatly improved in the performance of character extraction and recognition compared with the traditional recognition algorithms.

Ubiquitous healthcare model based on context recognition (상황인식에 기반한 유비쿼터스 헬스케어 모델)

  • Kim, Jeong-Won
    • Journal of the Korea Society of Computer and Information
    • /
    • v.15 no.9
    • /
    • pp.129-136
    • /
    • 2010
  • With mobile computing, wireless sensor network and sensor technologies, ubiquitous computing services are being realized and could satisfy the feasibility of ubiquitous healthcare to everyone. This u-Healthcare service can improve life quality of human since medical service can be provided to anyone, anytime, and anywhere. To confirm the vision of u-Healthcare service, we've implemented a healthcare system for heart disease patient which is composed of two components. Front-end collects various signals such as temperature, blood pressure, SpO2, and electrocardiogram, etc. As a backend, medical information server accumulates sensing data and performs back-end processing. To simply transfer these sensing values to a medical team may be too trivial. So, we've designed a model based on context awareness for more improved medical service which is based on artificial neural network. Through rigid experiments, we could confirm that the proposed system can provide improved medical service.

Dual-stream Co-enhanced Network for Unsupervised Video Object Segmentation

  • Hongliang Zhu;Hui Yin;Yanting Liu;Ning Chen
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.18 no.4
    • /
    • pp.938-958
    • /
    • 2024
  • Unsupervised Video Object Segmentation (UVOS) is a highly challenging problem in computer vision as the annotation of the target object in the testing video is unknown at all. The main difficulty is to effectively handle the complicated and changeable motion state of the target object and the confusion of similar background objects in video sequence. In this paper, we propose a novel deep Dual-stream Co-enhanced Network (DC-Net) for UVOS via bidirectional motion cues refinement and multi-level feature aggregation, which can fully take advantage of motion cues and effectively integrate different level features to produce high-quality segmentation mask. DC-Net is a dual-stream architecture where the two streams are co-enhanced by each other. One is a motion stream with a Motion-cues Refine Module (MRM), which learns from bidirectional optical flow images and produces fine-grained and complete distinctive motion saliency map, and the other is an appearance stream with a Multi-level Feature Aggregation Module (MFAM) and a Context Attention Module (CAM) which are designed to integrate the different level features effectively. Specifically, the motion saliency map obtained by the motion stream is fused with each stage of the decoder in the appearance stream to improve the segmentation, and in turn the segmentation loss in the appearance stream feeds back into the motion stream to enhance the motion refinement. Experimental results on three datasets (Davis2016, VideoSD, SegTrack-v2) demonstrate that DC-Net has achieved comparable results with some state-of-the-art methods.

Super Resolution Fusion Scheme for General- and Face Dataset (범용 데이터 셋과 얼굴 데이터 셋에 대한 초해상도 융합 기법)

  • Mun, Jun Won;Kim, Jae Seok
    • Journal of Korea Multimedia Society
    • /
    • v.22 no.11
    • /
    • pp.1242-1250
    • /
    • 2019
  • Super resolution technique aims to convert a low-resolution image with coarse details to a corresponding high-resolution image with refined details. In the past decades, the performance is greatly improved due to progress of deep learning models. However, universal solution for various objects is a still challenging issue. We observe that learning super resolution with a general dataset has poor performance on faces. In this paper, we propose a super resolution fusion scheme that works well for both general- and face datasets to achieve more universal solution. In addition, object-specific feature extractor is employed for better reconstruction performance. In our experiments, we compare our fusion image and super-resolved images from one- of the state-of-the-art deep learning models trained with DIV2K and FFHQ datasets. Quantitative and qualitative evaluates show that our fusion scheme successfully works well for both datasets. We expect our fusion scheme to be effective on other objects with poor performance and this will lead to universal solutions.

Robust Deep Age Estimation Method Using Artificially Generated Image Set

  • Jang, Jaeyoon;Jeon, Seung-Hyuk;Kim, Jaehong;Yoon, Hosub
    • ETRI Journal
    • /
    • v.39 no.5
    • /
    • pp.643-651
    • /
    • 2017
  • Human age estimation is one of the key factors in the field of Human-Robot Interaction/Human-Computer Interaction (HRI/HCI). Owing to the development of deep-learning technologies, age recognition has recently been attempted. In general, however, deep learning techniques require a large-scale database, and for age learning with variations, a conventional database is insufficient. For this reason, we propose an age estimation method using artificially generated data. Image data are artificially generated through 3D information, thus solving the problem of shortage of training data, and helping with the training of the deep-learning technique. Augmentation using 3D has advantages over 2D because it creates new images with more information. We use a deep architecture as a pre-trained model, and improve the estimation capacity using artificially augmented training images. The deep architecture can outperform traditional estimation methods, and the improved method showed increased reliability. We have achieved state-of-the-art performance using the proposed method in the Morph-II dataset and have proven that the proposed method can be used effectively using the Adience dataset.

EER-ASSL: Combining Rollback Learning and Deep Learning for Rapid Adaptive Object Detection

  • Ahmed, Minhaz Uddin;Kim, Yeong Hyeon;Rhee, Phill Kyu
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.14 no.12
    • /
    • pp.4776-4794
    • /
    • 2020
  • We propose a rapid adaptive learning framework for streaming object detection, called EER-ASSL. The method combines the expected error reduction (EER) dependent rollback learning and the active semi-supervised learning (ASSL) for a rapid adaptive CNN detector. Most CNN object detectors are built on the assumption of static data distribution. However, images are often noisy and biased, and the data distribution is imbalanced in a real world environment. The proposed method consists of collaborative sampling and EER-ASSL. The EER-ASSL utilizes the active learning (AL) and rollback based semi-supervised learning (SSL). The AL allows us to select more informative and representative samples measuring uncertainty and diversity. The SSL divides the selected streaming image samples into the bins and each bin repeatedly transfers the discriminative knowledge of the EER and CNN models to the next bin until convergence and incorporation with the EER rollback learning algorithm is achieved. The EER models provide a rapid short-term myopic adaptation and the CNN models an incremental long-term performance improvement. EER-ASSL can overcome noisy and biased labels in varying data distribution. Extensive experiments shows that EER-ASSL obtained 70.9 mAP compared to state-of-the-art technology such as Faster RCNN, SSD300, and YOLOv2.