• Title/Summary/Keyword: camera image

Search Result 4,918, Processing Time 0.04 seconds

Visualization of Khitan Scripts in Ancient Documents using Active Infrared Thermography (고문서 거란문자의 능동형 적외선 열영상 가시화)

  • Kim, Nohyu;Chung, Jaeyoung
    • The Journal of the Convergence on Culture Technology
    • /
    • v.8 no.2
    • /
    • pp.329-336
    • /
    • 2022
  • Unreadable Khitan scripts of ancient documents written by indian ink on parchment(sheepskin) are visualized by active infrared thermography without contacting and damaging the document which are deteriorated and aged presumably over many years. Sinusoidal infrared thermal wave using Halogen lamp is applied to the surface of the document in order to selectively magnify and record the thermal response of indian ink. The infrared image of the document captured in real time by infrared camera under the active external excitation shows the better sharpness and readability of Khitan characters than the optical image, from which many Khitan letters like ' ' and ' ' sounding as 'd' and 'ri' in English alphabet are detected and deciphered. It is concluded from the experiment that the active infrared thermography can be used as a promising method for digital reconstruction and preservation of ancient documents in the future.

Deep Learning-based Depth Map Estimation: A Review

  • Abdullah, Jan;Safran, Khan;Suyoung, Seo
    • Korean Journal of Remote Sensing
    • /
    • v.39 no.1
    • /
    • pp.1-21
    • /
    • 2023
  • In this technically advanced era, we are surrounded by smartphones, computers, and cameras, which help us to store visual information in 2D image planes. However, such images lack 3D spatial information about the scene, which is very useful for scientists, surveyors, engineers, and even robots. To tackle such problems, depth maps are generated for respective image planes. Depth maps or depth images are single image metric which carries the information in three-dimensional axes, i.e., xyz coordinates, where z is the object's distance from camera axes. For many applications, including augmented reality, object tracking, segmentation, scene reconstruction, distance measurement, autonomous navigation, and autonomous driving, depth estimation is a fundamental task. Much of the work has been done to calculate depth maps. We reviewed the status of depth map estimation using different techniques from several papers, study areas, and models applied over the last 20 years. We surveyed different depth-mapping techniques based on traditional ways and newly developed deep-learning methods. The primary purpose of this study is to present a detailed review of the state-of-the-art traditional depth mapping techniques and recent deep learning methodologies. This study encompasses the critical points of each method from different perspectives, like datasets, procedures performed, types of algorithms, loss functions, and well-known evaluation metrics. Similarly, this paper also discusses the subdomains in each method, like supervised, unsupervised, and semi-supervised methods. We also elaborate on the challenges of different methods. At the conclusion of this study, we discussed new ideas for future research and studies in depth map research.

An Improved ViBe Algorithm of Moving Target Extraction for Night Infrared Surveillance Video

  • Feng, Zhiqiang;Wang, Xiaogang;Yang, Zhongfan;Guo, Shaojie;Xiong, Xingzhong
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.15 no.12
    • /
    • pp.4292-4307
    • /
    • 2021
  • For the research field of night infrared surveillance video, the target imaging in the video is easily affected by the light due to the characteristics of the active infrared camera and the classical ViBe algorithm has some problems for moving target extraction because of background misjudgment, noise interference, ghost shadow and so on. Therefore, an improved ViBe algorithm (I-ViBe) for moving target extraction in night infrared surveillance video is proposed in this paper. Firstly, the video frames are sampled and judged by the degree of light influence, and the video frame is divided into three situations: no light change, small light change, and severe light change. Secondly, the ViBe algorithm is extracted the moving target when there is no light change. The segmentation factor of the ViBe algorithm is adaptively changed to reduce the impact of the light on the ViBe algorithm when the light change is small. The moving target is extracted using the region growing algorithm improved by the image entropy in the differential image of the current frame and the background model when the illumination changes drastically. Based on the results of the simulation, the I-ViBe algorithm proposed has better robustness to the influence of illumination. When extracting moving targets at night the I-ViBe algorithm can make target extraction more accurate and provide more effective data for further night behavior recognition and target tracking.

A Remote Control of 6 d.o.f. Robot Arm Based on 2D Vision Sensor (2D 영상센서 기반 6축 로봇 팔 원격제어)

  • Hyun, Woong-Keun
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.17 no.5
    • /
    • pp.933-940
    • /
    • 2022
  • In this paper, the algorithm was developed to recognize hand 3D position through 2D image sensor and implemented a system to remotely control the 6 d.o.f. robot arm by using it. The system consists of a camera that acquires hand position in 2D, a computer that controls robot arm that performs movement by hand position recognition. The image sensor recognizes the specific color of the glove putting on operator's hand and outputs the recognized range and position by including the color area of the glove as a shape of rectangle. We recognize the velocity vector of end effector and control the robot arm by the output data of the position and size of the detected rectangle. Through the several experiments using developed 6 axis robot, it was confirmed that the 6 d.o.f. robot arm remote control was successfully performed.

Application Development to provide General Phrase Medication Guidance for Visually Impaired (시각장애인을 위한 일반의약품 복용 방법 제공 애플리케이션 연구)

  • Min-Seok Cho;Min-Ki Yoon;Min-Su Seo;Young-Hoon Hwang;Won-Whoi Huh
    • The Journal of the Institute of Internet, Broadcasting and Communication
    • /
    • v.23 no.2
    • /
    • pp.201-208
    • /
    • 2023
  • Visually impaired people have difficulty accessing medical services and drug information, and it is not easy for them to take the correct drugs. A health law has been established to deal with this, but the plans are not unified and not applicable to all over-the-counter medicines. Therefore, we planned an application that provides instructions for taking general medicines using image recognition technology, bar code, and QR recognition technology for visually impaired people who are in the blind spot of taking general medicines. In this research, we will optimize the camera shooting environment standards and UI and UX screens for image recognition for the visually impaired, and improve the accessibility for the visually impaired to make it more convenient. If you can apply the results of your research to an application and provide it, it will help people with visual impairments learn the correct way to take over-the-counter medicines.

Biometric identification of Black Bengal goat: unique iris pattern matching system vs deep learning approach

  • Menalsh Laishram;Satyendra Nath Mandal;Avijit Haldar;Shubhajyoti Das;Santanu Bera;Rajarshi Samanta
    • Animal Bioscience
    • /
    • v.36 no.6
    • /
    • pp.980-989
    • /
    • 2023
  • Objective: Iris pattern recognition system is well developed and practiced in human, however, there is a scarcity of information on application of iris recognition system in animals at the field conditions where the major challenge is to capture a high-quality iris image from a constantly moving non-cooperative animal even when restrained properly. The aim of the study was to validate and identify Black Bengal goat biometrically to improve animal management in its traceability system. Methods: Forty-nine healthy, disease free, 3 months±6 days old female Black Bengal goats were randomly selected at the farmer's field. Eye images were captured from the left eye of an individual goat at 3, 6, 9, and 12 months of age using a specialized camera made for human iris scanning. iGoat software was used for matching the same individual goats at 3, 6, 9, and 12 months of ages. Resnet152V2 deep learning algorithm was further applied on same image sets to predict matching percentages using only captured eye images without extracting their iris features. Results: The matching threshold computed within and between goats was 55%. The accuracies of template matching of goats at 3, 6, 9, and 12 months of ages were recorded as 81.63%, 90.24%, 44.44%, and 16.66%, respectively. As the accuracies of matching the goats at 9 and 12 months of ages were low and below the minimum threshold matching percentage, this process of iris pattern matching was not acceptable. The validation accuracies of resnet152V2 deep learning model were found 82.49%, 92.68%, 77.17%, and 87.76% for identification of goat at 3, 6, 9, and 12 months of ages, respectively after training the model. Conclusion: This study strongly supported that deep learning method using eye images could be used as a signature for biometric identification of an individual goat.

Multi-Region based Radial GCN algorithm for Human action Recognition (행동인식을 위한 다중 영역 기반 방사형 GCN 알고리즘)

  • Jang, Han Byul;Lee, Chil Woo
    • Smart Media Journal
    • /
    • v.11 no.1
    • /
    • pp.46-57
    • /
    • 2022
  • In this paper, multi-region based Radial Graph Convolutional Network (MRGCN) algorithm which can perform end-to-end action recognition using the optical flow and gradient of input image is described. Because this method does not use information of skeleton that is difficult to acquire and complicated to estimate, it can be used in general CCTV environment in which only video camera is used. The novelty of MRGCN is that it expresses the optical flow and gradient of the input image as directional histograms and then converts it into six feature vectors to reduce the amount of computational load and uses a newly developed radial type network model to hierarchically propagate the deformation and shape change of the human body in spatio-temporal space. Another important feature is that the data input areas are arranged being overlapped each other, so that information is not spatially disconnected among input nodes. As a result of performing MRGCN's action recognition performance evaluation experiment for 30 actions, it was possible to obtain Top-1 accuracy of 84.78%, which is superior to the existing GCN-based action recognition method using skeleton data as an input.

Characterization of stacked geotextile tube structure using digital image correlation

  • Dong-Ju Kim;Dong Geon Son;Jong-Sub Lee;Thomas H.-K. Kang;Tae Sup Yun;Yong-Hoon Byun
    • Computers and Concrete
    • /
    • v.31 no.5
    • /
    • pp.385-394
    • /
    • 2023
  • Displacement is an important element for evaluating the stability and failure mechanism of hydraulic structures. Digital image correlation (DIC) is a useful technique to measure a three-dimensional displacement field using two cameras without any contact with test material. The objective of this study is to evaluate the behavior of stacked geotextile tubes using the DIC technique. Geotextile tubes are stacked to build a small-scale temporary dam model to exclude water from a specific area. The horizontal and vertical displacements of four stacked geotextile tubes are monitored using a dual camera system according to the upstream water level. The geotextile tubes are prepared with two different fill materials. For each dam model, the interface layers between upper and lower geotextile tubes are either unreinforced or reinforced with a cementitious binder. The displacement of stacked geotextile tubes is measured to analyze the behavior of geotextile tubes. Experimental results show that as upstream water level increases, horizontal and vertical displacements at each layer of geotextile tubes initially increase with water level, and then remain almost constant until the subsequent water level. The displacement of stacked geotextile tubes depends on the type of fill material and interfacial reinforcement with a cementitious binder. Thus, the proposed DIC technique can be effectively used to evaluate the behavior of a hydraulic structure, which consists of geotextile tubes.

Estimation of Heading Date of Paddy Rice from Slanted View Images Using Deep Learning Classification Model

  • Hyeokjin Bak;Hoyoung Ban;SeongryulChang;Dongwon Gwon;Jae-Kyeong Baek;Jeong-Il Cho;Wan-Gyu Sang
    • Proceedings of the Korean Society of Crop Science Conference
    • /
    • 2022.10a
    • /
    • pp.80-80
    • /
    • 2022
  • Estimation of heading date of paddy rice is laborious and time consuming. Therefore, automatic estimation of heading date of paddy rice is highly essential. In this experiment, deep learning classification models were used to classify two difference categories of rice (vegetative and reproductive stage) based on the panicle initiation of paddy field. Specifically, the dataset includes 444 slanted view images belonging to two categories and was then expanded to include 1,497 images via IMGAUG data augmentation technique. We adopt two transfer learning strategies: (First, used transferring model weights already trained on ImageNet to six classification network models: VGGNet, ResNet, DenseNet, InceptionV3, Xception and MobileNet, Second, fine-tuned some layers of the network according to our dataset). After training the CNN model, we used several evaluation metrics commonly used for classification tasks, including Accuracy, Precision, Recall, and F1-score. In addition, GradCAM was used to generate visual explanations for each image patch. Experimental results showed that the InceptionV3 is the best performing model in terms of the accuracy, average recall, precision, and F1-score. The fine-tuned InceptionV3 model achieved an overall classification accuracy of 0.95 with a high F1-score of 0.95. Our CNN model also represented the change of rice heading date under different date of transplanting. This study demonstrated that image based deep learning model can reliably be used as an automatic monitoring system to detect the heading date of rice crops using CCTV camera.

  • PDF

Distance and Entropy Based Image Viewpoint Selection for Accurate 3D Reconstruction with NeRF (NeRF의 정확한 3차원 복원을 위한 거리-엔트로피 기반 영상 시점 선택 기술)

  • Jinwon Choi;Chanho Seo;Junhyeok Choi;Sunglok Choi
    • The Journal of Korea Robotics Society
    • /
    • v.19 no.1
    • /
    • pp.98-105
    • /
    • 2024
  • This paper proposes a new approach with a distance-based regularization to the entropy applied to the NBV (Next-Best-View) selection with NeRF (Neural Radiance Fields). 3D reconstruction requires images from various viewpoints, and selecting where to capture these images is a highly complex problem. In a recent work, image acquisition was derived using NeRF's ray-based uncertainty. While this work was effective for evaluating candidate viewpoints at fixed distances from a camera to an object, it is limited when dealing with a range of candidate viewpoints at various distances, because it tends to favor selecting viewpoints at closer distances. Acquiring images from nearby viewpoints is beneficial for capturing surface details. However, with the limited number of images, its image selection is less overlapped and less frequently observed, so its reconstructed result is sensitive to noise and contains undesired artifacts. We propose a method that incorporates distance-based regularization into entropy, allowing us to acquire images at distances conducive to capturing both surface details without undesired noise and artifacts. Our experiments with synthetic images demonstrated that NeRF models with the proposed distance and entropy-based criteria achieved around 50 percent fewer reconstruction errors than the recent work.