• Title/Summary/Keyword: Computer Vision Technology

Search Result 673, Processing Time 0.037 seconds

The Aesthetic Transformation of Shadow Images and the Extended Imagination (그림자 이미지의 미학적 변용과 확장된 상상력 :디지털 실루엣 애니메이션과 최근 미디어 아트의 흐름을 중심으로)

  • Kim, Young-Ok
    • Cartoon and Animation Studies
    • /
    • s.49
    • /
    • pp.651-676
    • /
    • 2017
  • Shadow images are a representative medium and means of expression for the imagination that exists between consciousness and unconsciousness for thousands of years. Wherever light exists, people create play with their own shadows without special skills, and have made a fantasy at once. Shadow images have long been used as subjects and materials of literacy, art, philosophy, and popular culture. Especially in the field of art, people have been experimenting with visual stimulation through the uniqueness of simple silhouettes images. In the field of animation, it became to be recognized as a form of non - mainstream areas that are difficult to make. However, shadow images have been used more actively in the field of digital arts and media art. In this Environment with technologies, Various formative imaginations are being expressed more with shadow images in a new dimension. This study is to introduce and analyze these trends, the aesthetic transformations and extended methods focusing on digital silhouette animation and recent media art works using shadow images. Screen-based silhouette animation combines digital technology and new approaches that have escaped conventional methods have removed most of the elements that have been considered limitations, and these factors have become a matter of choice for the directors. Especially, in the display environment using various light sources, projection, and camera technology, shadow images were expressed with multiple-layered virtual spaces, and it becomes possible to imagine a new extended imagination. Through the computer vision, it became possible to find new gaze and spatial images and use it more flexibly. These changes have given new possibility to the use shadow images in a different way.

A Novel Fast and High-Performance Image Quality Assessment Metric using a Simple Laplace Operator (단순 라플라스 연산자를 사용한 새로운 고속 및 고성능 영상 화질 측정 척도)

  • Bae, Sung-Ho;Kim, Munchurl
    • Journal of Broadcast Engineering
    • /
    • v.21 no.2
    • /
    • pp.157-168
    • /
    • 2016
  • In image processing and computer vision fields, mean squared error (MSE) has popularly been used as an objective metric in image quality optimization problems due to its desirable mathematical properties such as metricability, differentiability and convexity. However, as known that MSE is not highly correlated with perceived visual quality, much effort has been made to develop new image quality assessment (IQA) metrics having both the desirable mathematical properties aforementioned and high prediction performances for subjective visual quality scores. Although recent IQA metrics having the desirable mathematical properties have shown to give some promising results in prediction performance for visual quality scores, they also have high computation complexities. In order to alleviate this problem, we propose a new fast IQA metric using a simple Laplace operator. Since the Laplace operator used in our IQA metric can not only effectively mimic operations of receptive fields in retina for luminance stimulus but also be simply computed, our IQA metric can yield both very fast processing speed and high prediction performance. In order to verify the effectiveness of the proposed IQA metric, our method is compared to some state-of-the-art IQA metrics. The experimental results showed that the proposed IQA metric has the fastest running speed compared the IQA methods except MSE under comparison. Moreover, our IQA metric achieves the best prediction performance for subjective image quality scores among the state-of-the-art IQA metrics under test.

Non-face-to-face online home training application study using deep learning-based image processing technique and standard exercise program (딥러닝 기반 영상처리 기법 및 표준 운동 프로그램을 활용한 비대면 온라인 홈트레이닝 어플리케이션 연구)

  • Shin, Youn-ji;Lee, Hyun-ju;Kim, Jun-hee;Kwon, Da-young;Lee, Seon-ae;Choo, Yun-jin;Park, Ji-hye;Jung, Ja-hyun;Lee, Hyoung-suk;Kim, Joon-ho
    • The Journal of the Convergence on Culture Technology
    • /
    • v.7 no.3
    • /
    • pp.577-582
    • /
    • 2021
  • Recently, with the development of AR, VR, and smart device technologies, the demand for services based on non-face-to-face environments is also increasing in the fitness industry. The non-face-to-face online home training service has the advantage of not being limited by time and place compared to the existing offline service. However, there are disadvantages including the absence of exercise equipment, difficulty in measuring the amount of exercise and chekcing whether the user maintains an accurate exercise posture or not. In this study, we develop a standard exercise program that can compensate for these shortcomings and propose a new non-face-to-face home training application by using a deep learning-based body posture estimation image processing algorithm. This application allows the user to directly watch and follow the trainer of the standard exercise program video, correct the user's own posture, and perform an accurate exercise. Furthermore, if the results of this study are customized according to their purpose, it will be possible to apply them to performances, films, club activities, and conferences

A Study on the Application of Object Detection Method in Construction Site through Real Case Analysis (사례분석을 통한 객체검출 기술의 건설현장 적용 방안에 관한 연구)

  • Lee, Kiseok;Kang, Sungwon;Shin, Yoonseok
    • Journal of the Society of Disaster Information
    • /
    • v.18 no.2
    • /
    • pp.269-279
    • /
    • 2022
  • Purpose: The purpose of this study is to develop a deep learning-based personal protective equipment detection model for disaster prevention at construction sites, and to apply it to actual construction sites and to analyze the results. Method: In the method of conducting this study, the dataset on the real environment was constructed and the developed personal protective equipment(PPE) detection model was applied. The PPE detection model mainly consists of worker detection and PPE classification model.The worker detection model uses a deep learning-based algorithm to build a dataset obtained from the actual field to learn and detect workers, and the PPE classification model applies the PPE detection algorithm learned from the worker detection area extracted from the work detection model. For verification of the proposed model, experimental results were derived from data obtained from three construction sites. Results: The application of the PPE recognition model to construction site brings up the problems related to mis-recognition and non-recognition. Conclusions: The analysis outcomes were produced to apply the object recognition technology to a construction site, and the need for follow-up research was suggested through representative cases of worker recognition and non-recognition, and mis-recognition of personal protective equipment.

CycleGAN Based Translation Method between Asphalt and Concrete Crack Images for Data Augmentation (데이터 증강을 위한 순환 생성적 적대 신경망 기반의 아스팔트와 콘크리트 균열 영상 간의 변환 기법)

  • Shim, Seungbo
    • The Journal of The Korea Institute of Intelligent Transport Systems
    • /
    • v.21 no.5
    • /
    • pp.171-182
    • /
    • 2022
  • The safe use of a structure requires it to be maintained in an undamaged state. Thus, a typical factor that determines the safety of a structure is a crack in it. In addition, cracks are caused by various reasons, damage the structure in various ways, and exist in different shapes. Making matters worse, if these cracks are unattended, the risk of structural failure increases and proceeds to a catastrophe. Hence, recently, methods of checking structural damage using deep learning and computer vision technology have been introduced. These methods usually have the premise that there should be a large amount of training image data. However, the amount of training image data is always insufficient. Particularly, this insufficiency negatively affects the performance of deep learning crack detection algorithms. Hence, in this study, a method of augmenting crack image data based on the image translation technique was developed. In particular, this method obtained the crack image data for training a deep learning neural network model by transforming a specific case of a asphalt crack image into a concrete crack image or vice versa . Eventually, this method expected that a robust crack detection algorithm could be developed by increasing the diversity of its training data.

Mean Teacher Learning Structure Optimization for Semantic Segmentation of Crack Detection (균열 탐지의 의미론적 분할을 위한 Mean Teacher 학습 구조 최적화 )

  • Seungbo Shim
    • Journal of the Korea institute for structural maintenance and inspection
    • /
    • v.27 no.5
    • /
    • pp.113-119
    • /
    • 2023
  • Most infrastructure structures were completed during periods of economic growth. The number of infrastructure structures reaching their lifespan is increasing, and the proportion of old structures is gradually increasing. The functions and performance of these structures at the time of design may deteriorate and may even lead to safety accidents. To prevent this repercussion, accurate inspection and appropriate repair are requisite. To this end, demand is increasing for computer vision and deep learning technology to accurately detect even minute cracks. However, deep learning algorithms require a large number of training data. In particular, label images indicating the location of cracks in the image are required. To secure a large number of those label images, a lot of labor and time are consumed. To reduce these costs as well as increase detection accuracy, this study proposed a learning structure based on mean teacher method. This learning structure was trained on a dataset of 900 labeled image dataset and 3000 unlabeled image dataset. The crack detection network model was evaluated on over 300 labeled image dataset, and the detection accuracy recorded a mean intersection over union of 89.23% and an F1 score of 89.12%. Through this experiment, it was confirmed that detection performance was improved compared to supervised learning. It is expected that this proposed method will be used in the future to reduce the cost required to secure label images.

Damage estimation for structural safety evaluation using dynamic displace measurement (구조안전도 평가를 위한 동적변위 기반 손상도 추정 기법 개발)

  • Shin, Yoon-Soo;Kim, Junhee
    • Journal of the Korea institute for structural maintenance and inspection
    • /
    • v.23 no.7
    • /
    • pp.87-94
    • /
    • 2019
  • Recently, the advance of accurate dynamic displacement measurement devices, such as GPS, computer vision, and optic laser sensor, has enhanced the structural monitoring technology. In this study, the dynamic displacement data was used to verify the applicability of the structural physical parameter estimation method through subspace system identification. The subspace system identification theory for estimating state-space model from measured data and physics-based interpretation for deriving the physical parameter of the estimated system are presented. Three-degree-freedom steel structures were fabricated for the experimental verification of the theory in this study. Laser displacement sensor and accelerometer were used to measure the displacement data of each floor and the acceleration data of the shaking table. Discrete state-space model generated from measured data was verified for precision. The discrete state-space model generated from the measured data extracted the floor stiffness of the building after accuracy verification. In addition, based on the story stiffness extracted from the state space model, five column stiffening and damage samples were set up to extract the change rate of story stiffness for each sample. As a result, in case of reinforcement and damage under the same condition, the stiffness change showed a high matching rate.

Images Grouping Technology based on Camera Sensors for Efficient Stitching of Multiple Images (다수의 영상간 효율적인 스티칭을 위한 카메라 센서 정보 기반 영상 그룹핑 기술)

  • Im, Jiheon;Lee, Euisang;Kim, Hoejung;Kim, Kyuheon
    • Journal of Broadcast Engineering
    • /
    • v.22 no.6
    • /
    • pp.713-723
    • /
    • 2017
  • Since the panoramic image can overcome the limitation of the viewing angle of the camera and have a wide field of view, it has been studied effectively in the fields of computer vision and stereo camera. In order to generate a panoramic image, stitching images taken by a plurality of general cameras instead of using a wide-angle camera, which is distorted, is widely used because it can reduce image distortion. The image stitching technique creates descriptors of feature points extracted from multiple images, compares the similarities of feature points, and links them together into one image. Each feature point has several hundreds of dimensions of information, and data processing time increases as more images are stitched. In particular, when a panorama is generated on the basis of an image photographed by a plurality of unspecified cameras with respect to an object, the extraction processing time of the overlapping feature points for similar images becomes longer. In this paper, we propose a preprocessing process to efficiently process stitching based on an image obtained from a number of unspecified cameras for one object or environment. In this way, the data processing time can be reduced by pre-grouping images based on camera sensor information and reducing the number of images to be stitched at one time. Later, stitching is done hierarchically to create one large panorama. Through the grouping preprocessing proposed in this paper, we confirmed that the stitching time for a large number of images is greatly reduced by experimental results.

The raise the efficiency of game graphics design education using game engine : In focus of Unity3D and Torque (게임엔진 활용으로 게임 그래픽 교육 효율성 제고: 유니티3D(Unity3D)와 토크(Torque) 엔진을 중심으로)

  • Kim, Chee-Hoon;Park, Sung-Il
    • Cartoon and Animation Studies
    • /
    • s.29
    • /
    • pp.151-172
    • /
    • 2012
  • Game production education in a university is very important because it is the stage for completing a game education course as well as for determining a future of game industry. In order to perform various experience and creative learning, it should be able to effectively use a computer infrastructure representing the knowledge and information society for the purpose of obtaining and re-processing information necessary for game production through prediction of directions of game industry as well information technology. This research is focused on an effective game engine education for students whom want to become game graphics designers. The purpose of this study is to draw a lesson of game production utilizing game engines and it enables practice-focused class for game production. It also allows the class participant to manufacture prototypes without support from game programmers for their outcomes of works planned during the game production class. The theoretical background of game production compared and analyzed exemplary game engines. Based on the result, the study selected Unity 3D engine and conducted the research on the background where the Unity engine has been selected and its characteristics. In addition, this study provided an example of game production utilizing a game engine, and also described the details of actual realization. This study selected Torque3D with the Unity in order to identify the purpose of this study and efficiency of learning. Thus, the previous situation is that the class remained in making a game plan during the course of game production project and, students whose major is not game programming. Now, it is necessary for students to make many efforts to make a game in an active and positive attitude by utilizing a game engine beyond the previous method of class.

A Study on Person Re-Identification System using Enhanced RNN (확장된 RNN을 활용한 사람재인식 시스템에 관한 연구)

  • Choi, Seok-Gyu;Xu, Wenjie
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.17 no.2
    • /
    • pp.15-23
    • /
    • 2017
  • The person Re-identification is the most challenging part of computer vision due to the significant changes in human pose and background clutter with occlusions. The picture from non-overlapping cameras enhance the difficulty to distinguish some person from the other. To reach a better performance match, most methods use feature selection and distance metrics separately to get discriminative representations and proper distance to describe the similarity between person and kind of ignoring some significant features. This situation has encouraged us to consider a novel method to deal with this problem. In this paper, we proposed an enhanced recurrent neural network with three-tier hierarchical network for person re-identification. Specifically, the proposed recurrent neural network (RNN) model contain an iterative expectation maximum (EM) algorithm and three-tier Hierarchical network to jointly learn both the discriminative features and metrics distance. The iterative EM algorithm can fully use of the feature extraction ability of convolutional neural network (CNN) which is in series before the RNN. By unsupervised learning, the EM framework can change the labels of the patches and train larger datasets. Through the three-tier hierarchical network, the convolutional neural network, recurrent network and pooling layer can jointly be a feature extractor to better train the network. The experimental result shows that comparing with other researchers' approaches in this field, this method also can get a competitive accuracy. The influence of different component of this method will be analyzed and evaluated in the future research.