• Title/Summary/Keyword: HRNet

Search Result 11, Processing Time 0.02 seconds

Pedestrian Recognition of Crosswalks Using Foot Estimation Techniques Based on HigherHRNet (HigherHRNet 기반의 발추정 기법을 통한 횡단보도 보행자 인식)

  • Jung, Kyung-Min;Han, Joo-Hoon;Lee, Hyun
    • IEMEK Journal of Embedded Systems and Applications
    • /
    • v.16 no.5
    • /
    • pp.171-177
    • /
    • 2021
  • It is difficult to accurately extract features of pedestrian because the pedestrian is photographed at a crosswalk using a camera positioned higher than the pedestrian. In addition, it is more difficult to extract features when a part of the pedestrian's body is covered by an umbrella or parasol or when the pedestrian is holding an object. Representative methods to solve this problem include Object Detection, Instance Segmentation, and Pose Estimation. Among them, this study intends to use the Pose Estimation method. In particular, we intend to increase the recognition rate of pedestrians in crosswalks by maintaining the image resolution through HigherHRNet and applying the foot estimation technique. Finally, we show the superiority of the proposed method by applying and analyzing several data sets covered by body parts to the existing method and the proposed method.

Design and Implementation of HRNet Model Combined with Spatial Information Attention Module of Polarized Self-attention (편광 셀프어텐션의 공간정보 강조 모듈을 결합한 HRNet 모델 설계 및 구현)

  • Jin-Seong Kim;Jun Park;Se-Hoon Jung;Chun-Bo Sim
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2023.05a
    • /
    • pp.485-487
    • /
    • 2023
  • 컴퓨터 비전의 하위 태스크(Task)인 의미론적 분할(Semantic Segmentation)은 자율주행, 해상에서 선박찾기 등 다양한 분야에서 연구되고 있다. 기존 FCN(Fully Conovlutional Networks) 기반 의미론적 분할 모델은 다운샘플링(Dowsnsampling)과정에서 공간정보의 손실이 발생하여 정확도가 하락했다. 본 논문에서는 공간정보 손실을 완화하고자 PSA(Polarized Self-attention)의 공간정보 강조 모듈을 HRNet(High-resolution Networks)의 합성곱 블록 사이에 추가한다. 실험결과 파라미터는 3.1M, GFLOPs는 3.2G 증가했으나 mIoU는 0.26% 증가했다. 공간정보가 의미론적 분할 정확도에 영향이 미치는 것을 확인했다.

Semantic Segmentation of the Habitats of Ecklonia Cava and Sargassum in Undersea Images Using HRNet-OCR and Swin-L Models (HRNet-OCR과 Swin-L 모델을 이용한 조식동물 서식지 수중영상의 의미론적 분할)

  • Kim, Hyungwoo;Jang, Seonwoong;Bak, Suho;Gong, Shinwoo;Kwak, Jiwoo;Kim, Jinsoo;Lee, Yangwon
    • Korean Journal of Remote Sensing
    • /
    • v.38 no.5_3
    • /
    • pp.913-924
    • /
    • 2022
  • In this paper, we presented a database construction of undersea images for the Habitats of Ecklonia cava and Sargassum and conducted an experiment for semantic segmentation using state-of-the-art (SOTA) models such as High Resolution Network-Object Contextual Representation (HRNet-OCR) and Shifted Windows-L (Swin-L). The result showed that our segmentation models were superior to the existing experiments in terms of the 29% increased mean intersection over union (mIOU). Swin-L model produced better performance for every class. In particular, the information of the Ecklonia cava class that had small data were also appropriately extracted by Swin-L model. Target objects and the backgrounds were well distinguished owing to the Transformer backbone better than the legacy models. A bigger database under construction will ensure more accuracy improvement and can be utilized as deep learning database for undersea images.

Comparative evaluation of deep learning-based building extraction techniques using aerial images (항공영상을 이용한 딥러닝 기반 건물객체 추출 기법들의 비교평가)

  • Mo, Jun Sang;Seong, Seon Kyeong;Choi, Jae Wan
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.39 no.3
    • /
    • pp.157-165
    • /
    • 2021
  • Recently, as the spatial resolution of satellite and aerial images has improved, various studies using remotely sensed data with high spatial resolution have been conducted. In particular, since the building extraction is essential for creating digital thematic maps, high accuracy of building extraction result is required. In this manuscript, building extraction models were generated using SegNet, U-Net, FC-DenseNet, and HRNetV2, which are representative semantic segmentation models in deep learning techniques, and then the evaluation of building extraction results was performed. Training dataset for building extraction were generated by using aerial orthophotos including various buildings, and evaluation was conducted in three areas. First, the model performance was evaluated through the region adjacent to the training dataset. In addition, the applicability of the model was evaluated through the region different from the training dataset. As a result, the f1-score of HRNetV2 represented the best values in terms of model performance and applicability. Through this study, the possibility of creating and modifying the building layer in the digital map was confirmed.

Thoracic Spine Segmentation of X-ray Images Using a Modified HRNet (수정된 HRNet을 이용한 X-ray 영상의 흉추 분할 기법)

  • Lee, Ye-Eun;Lee, Dong-Gyu;Jeong, Ji-Hoon;Kim, Hyung-Kyu;Kim, Ho-Joon
    • Proceedings of the Korea Information Processing Society Conference
    • /
    • 2022.05a
    • /
    • pp.705-707
    • /
    • 2022
  • 인체의 흉부 X-ray 영상으로부터 척추질환과 관련된 의료 진단지표를 자동으로 추출하는 과정을 위하여 흉추조직의 정확한 분할이 필요하다. 본 연구에서는 HRNet 기반의 학습을 통하여 흉추조직을 분할하는 방법을 고찰한다. 분할 과정에서 영상 내의 상대적인 위치 정보가 효과적으로 반영될 수 있도록, 계층별로 영상의 고해상도의 표현이 그대로 유지되는 구조와 저해상도의 특징 지도로 변환되는 구조가 병렬적으로 연결되는 형태의 심층 신경망 모델을 채택하였다. 흉부 X-ray 영상에서 콥각도(Cobb's angle)를 산출하는 문제를 대상으로 흉추 분할을 위한 학습 방법, 진단지표 추출 방법 등을 소개하며, 부수적으로 피사체의 위치 변화 및 크기 변화 등에 강인한 성능을 제공하기 위하여 학습 데이터를 증강하는 방법론을 제시하였다. 총 145개의 영상을 사용한 실험을 통하여 제안된 이론의 타당성을 평가하였다.

Semantic Segmentation of the Submerged Marine Debris in Undersea Images Using HRNet Model (HRNet 기반 해양침적쓰레기 수중영상의 의미론적 분할)

  • Kim, Daesun;Kim, Jinsoo;Jang, Seonwoong;Bak, Suho;Gong, Shinwoo;Kwak, Jiwoo;Bae, Jaegu
    • Korean Journal of Remote Sensing
    • /
    • v.38 no.6_1
    • /
    • pp.1329-1341
    • /
    • 2022
  • Destroying the marine environment and marine ecosystem and causing marine accidents, marine debris is generated every year, and among them, submerged marine debris is difficult to identify and collect because it is on the seabed. Therefore, deep-learning-based semantic segmentation was experimented on waste fish nets and waste ropes using underwater images to identify efficient collection and distribution. For segmentation, a high-resolution network (HRNet), a state-of-the-art deep learning technique, was used, and the performance of each optimizer was compared. In the segmentation result fish net, F1 score=(86.46%, 86.20%, 85.29%), IoU=(76.15%, 75.74%, 74.36%), For the rope F1 score=(80.49%, 80.48%, 77.86%), IoU=(67.35%, 67.33%, 63.75%) in the order of adaptive moment estimation (Adam), Momentum, and stochastic gradient descent (SGD). Adam's results were the highest in both fish net and rope. Through the research results, the evaluation of segmentation performance for each optimizer and the possibility of segmentation of marine debris in the latest deep learning technique were confirmed. Accordingly, it is judged that by applying the latest deep learning technique to the identification of submerged marine debris through underwater images, it will be helpful in estimating the distribution of marine sedimentation debris through more accurate and efficient identification than identification through the naked eye.

Image Matching for Orthophotos by Using HRNet Model (HRNet 모델을 이용한 항공정사영상간 영상 매칭)

  • Seong, Seonkyeong;Choi, Jaewan
    • Korean Journal of Remote Sensing
    • /
    • v.38 no.5_1
    • /
    • pp.597-608
    • /
    • 2022
  • Remotely sensed data have been used in various fields, such as disasters, agriculture, urban planning, and the military. Recently, the demand for the multitemporal dataset with the high-spatial-resolution has increased. This manuscript proposed an automatic image matching algorithm using a deep learning technique to utilize a multitemporal remotely sensed dataset. The proposed deep learning model was based on High Resolution Net (HRNet), widely used in image segmentation. In this manuscript, denseblock was added to calculate the correlation map between images effectively and to increase learning efficiency. The training of the proposed model was performed using the multitemporal orthophotos of the National Geographic Information Institute (NGII). In order to evaluate the performance of image matching using a deep learning model, a comparative evaluation was performed. As a result of the experiment, the average horizontal error of the proposed algorithm based on 80% of the image matching rate was 3 pixels. At the same time, that of the Zero Normalized Cross-Correlation (ZNCC) was 25 pixels. In particular, it was confirmed that the proposed method is effective even in mountainous and farmland areas where the image changes according to vegetation growth. Therefore, it is expected that the proposed deep learning algorithm can perform relative image registration and image matching of a multitemporal remote sensed dataset.

Detection of Wildfire Burned Areas in California Using Deep Learning and Landsat 8 Images (딥러닝과 Landsat 8 영상을 이용한 캘리포니아 산불 피해지 탐지)

  • Youngmin Seo;Youjeong Youn;Seoyeon Kim;Jonggu Kang;Yemin Jeong;Soyeon Choi;Yungyo Im;Yangwon Lee
    • Korean Journal of Remote Sensing
    • /
    • v.39 no.6_1
    • /
    • pp.1413-1425
    • /
    • 2023
  • The increasing frequency of wildfires due to climate change is causing extreme loss of life and property. They cause loss of vegetation and affect ecosystem changes depending on their intensity and occurrence. Ecosystem changes, in turn, affect wildfire occurrence, causing secondary damage. Thus, accurate estimation of the areas affected by wildfires is fundamental. Satellite remote sensing is used for forest fire detection because it can rapidly acquire topographic and meteorological information about the affected area after forest fires. In addition, deep learning algorithms such as convolutional neural networks (CNN) and transformer models show high performance for more accurate monitoring of fire-burnt regions. To date, the application of deep learning models has been limited, and there is a scarcity of reports providing quantitative performance evaluations for practical field utilization. Hence, this study emphasizes a comparative analysis, exploring performance enhancements achieved through both model selection and data design. This study examined deep learning models for detecting wildfire-damaged areas using Landsat 8 satellite images in California. Also, we conducted a comprehensive comparison and analysis of the detection performance of multiple models, such as U-Net and High-Resolution Network-Object Contextual Representation (HRNet-OCR). Wildfire-related spectral indices such as normalized difference vegetation index (NDVI) and normalized burn ratio (NBR) were used as input channels for the deep learning models to reflect the degree of vegetation cover and surface moisture content. As a result, the mean intersection over union (mIoU) was 0.831 for U-Net and 0.848 for HRNet-OCR, showing high segmentation performance. The inclusion of spectral indices alongside the base wavelength bands resulted in increased metric values for all combinations, affirming that the augmentation of input data with spectral indices contributes to the refinement of pixels. This study can be applied to other satellite images to build a recovery strategy for fire-burnt areas.

Change Detection of Building Objects in Urban Area by Using Transfer Learning (전이학습을 활용한 도시지역 건물객체의 변화탐지)

  • Mo, Jun-sang;Seong, Seon-kyeong;Choi, Jae-wan
    • Korean Journal of Remote Sensing
    • /
    • v.37 no.6_1
    • /
    • pp.1685-1695
    • /
    • 2021
  • To generate a deep learning model with high performance, a large training dataset should be required. However, it requires a lot of time and cost to generate a large training dataset in remote sensing. Therefore, the importance of transfer learning of deep learning model using a small dataset have been increased. In this paper, we performed transfer learning of trained model based on open datasets by using orthoimages and digital maps to detect changes of building objects in multitemporal orthoimages. For this, an initial training was performed on open dataset for change detection through the HRNet-v2 model, and transfer learning was performed on dataset by orthoimages and digital maps. To analyze the effect of transfer learning, change detection results of various deep learning models including deep learning model by transfer learning were evaluated at two test sites. In the experiments, results by transfer learning represented best accuracy, compared to those by other deep learning models. Therefore, it was confirmed that the problem of insufficient training dataset could be solved by using transfer learning, and the change detection algorithm could be effectively applied to various remote sensed imagery.

Class Classification and Validation of a Musculoskeletal Risk Factor Dataset for Manufacturing Workers (제조업 노동자 근골격계 부담요인 데이터셋 클래스 분류와 유효성 검증)

  • Young-Jin Kang;;;Jeong, Seok Chan
    • The Journal of Bigdata
    • /
    • v.8 no.1
    • /
    • pp.49-59
    • /
    • 2023
  • There are various items in the safety and health standards of the manufacturing industry, but they can be divided into work-related diseases and musculoskeletal diseases according to the standards for sickness and accident victims. Musculoskeletal diseases occur frequently in manufacturing and can lead to a decrease in labor productivity and a weakening of competitiveness in manufacturing. In this paper, to detect the musculoskeletal harmful factors of manufacturing workers, we defined the musculoskeletal load work factor analysis, harmful load working postures, and key points matching, and constructed data for Artificial Intelligence(AI) learning. To check the effectiveness of the suggested dataset, AI algorithms such as YOLO, Lite-HRNet, and EfficientNet were used to train and verify. Our experimental results the human detection accuracy is 99%, the key points matching accuracy of the detected person is @AP0.5 88%, and the accuracy of working postures evaluation by integrating the inferred matching positions is LEGS 72.2%, NECT 85.7%, TRUNK 81.9%, UPPERARM 79.8%, and LOWERARM 92.7%, and considered the necessity for research that can prevent deep learning-based musculoskeletal diseases.