• Title/Summary/Keyword: automated object recognition

Search Result 40, Processing Time 0.027 seconds

Object Pose Estimation and Motion Planning for Service Automation System (서비스 자동화 시스템을 위한 물체 자세 인식 및 동작 계획)

  • Youngwoo Kwon;Dongyoung Lee;Hosun Kang;Jiwook Choi;Inho Lee
    • The Journal of Korea Robotics Society
    • /
    • v.19 no.2
    • /
    • pp.176-187
    • /
    • 2024
  • Recently, automated solutions using collaborative robots have been emerging in various industries. Their primary functions include Pick & Place, Peg in the Hole, fastening and assembly, welding, and more, which are being utilized and researched in various fields. The application of these robots varies depending on the characteristics of the grippers attached to the end of the collaborative robots. To grasp a variety of objects, a gripper with a high degree of freedom is required. In this paper, we propose a service automation system using a multi-degree-of-freedom gripper, collaborative robots, and vision sensors. Assuming various products are placed at a checkout counter, we use three cameras to recognize the objects, estimate their pose, and create grasping points for grasping. The grasping points are grasped by the multi-degree-of-freedom gripper, and experiments are conducted to recognize barcodes, a key task in service automation. To recognize objects, we used a CNN (Convolutional Neural Network) based algorithm and point cloud to estimate the object's 6D pose. Using the recognized object's 6d pose information, we create grasping points for the multi-degree-of-freedom gripper and perform re-grasping in a direction that facilitates barcode scanning. The experiment was conducted with four selected objects, progressing through identification, 6D pose estimation, and grasping, recording the success and failure of barcode recognition to prove the effectiveness of the proposed system.

Agricultural Autonomous Robots System for Automatic Transfer of Agricultural Harvests (수확물 자동 이송을 위한 농업용 자율주행 로봇 시스템)

  • Kim, Jong-Sil;Kim, Eung-Kon
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.16 no.4
    • /
    • pp.749-754
    • /
    • 2021
  • In order to solve problems such as a decrease in the agricultural population and an aging population, research on agricultural robots is being actively conducted for the purpose of automating various agricultural tasks. The harvesting process is the most labor-intensive process among farm work and this process consumes about 2-3 times more compared to other processes. Since the transport of agricultural crops requires the most labor costs and there is a risk of injury during the operation, automating the transport operation through an agricultural robot can improve safety and significantly improve productivity. Therefore, this paper proposes an agricultural robot that is optimized for farm worksites and capable of autonomous driving.

Hazy Particle Map-based Automated Fog Removal Method with Haziness Degree Evaluator Applied (Haziness Degree Evaluator를 적용한 Hazy Particle Map 기반 자동화 안개 제거 방법)

  • Sim, Hwi Bo;Kang, Bong Soon
    • Journal of Korea Multimedia Society
    • /
    • v.25 no.9
    • /
    • pp.1266-1272
    • /
    • 2022
  • With the recent development of computer vision technology, image processing-based mechanical devices are being developed to realize autonomous driving. The camera-taken images of image processing-based machines are invisible due to scattering and absorption of light in foggy conditions. This lowers the object recognition rate and causes malfunction. The safety of the technology is very important because the malfunction of autonomous driving leads to human casualties. In order to increase the stability of the technology, it is necessary to apply an efficient haze removal algorithm to the camera. In the conventional haze removal method, since the haze removal operation is performed regardless of the haze concentration of the input image, excessive haze is removed and the quality of the resulting image is deteriorated. In this paper, we propose an automatic haze removal method that removes haze according to the haze density of the input image by applying Ngo's Haziness Degree Evaluator (HDE) to Kim's haze removal algorithm using Hazy Particle Map. The proposed haze removal method removes the haze according to the haze concentration of the input image, thereby preventing the quality degradation of the input image that does not require haze removal and solving the problem of excessive haze removal. The superiority of the proposed haze removal method is verified through qualitative and quantitative evaluation.

A Filter Algorithm based on Partial Mask and Lagrange Interpolation for Impulse Noise Removal (임펄스 잡음 제거를 위한 부분 마스크와 라그랑지 보간법에 기반한 필터 알고리즘)

  • Cheon, Bong-Won;Kim, Nam-Ho
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.26 no.5
    • /
    • pp.675-681
    • /
    • 2022
  • Recently, with the development of IoT technology and AI, unmanned and automated in various fields, interest in video processing, which is the basis for automation such as object recognition and object classification, is increasing. Various studies have been conducted on noise removal in the video processing process, which has a significant impact on image quality and system accuracy and reliability, but there is a problem that it is difficult to restore images for areas with high impulse noise density. In this paper proposes a filter algorithm based on partial mask and Lagrange interpolation to restore the damaged area of impulse noise in the image. In the proposed algorithm, the filtering process was switched by comparing the filtering mask with the noise estimate and the purge weight was calculated based on the low frequency component and the high frequency component of the image to restore the image.

A Collaborative Video Annotation and Browsing System using Linked Data (링크드 데이터를 이용한 협업적 비디오 어노테이션 및 브라우징 시스템)

  • Lee, Yeon-Ho;Oh, Kyeong-Jin;Sean, Vi-Sal;Jo, Geun-Sik
    • Journal of Intelligence and Information Systems
    • /
    • v.17 no.3
    • /
    • pp.203-219
    • /
    • 2011
  • Previously common users just want to watch the video contents without any specific requirements or purposes. However, in today's life while watching video user attempts to know and discover more about things that appear on the video. Therefore, the requirements for finding multimedia or browsing information of objects that users want, are spreading with the increasing use of multimedia such as videos which are not only available on the internet-capable devices such as computers but also on smart TV and smart phone. In order to meet the users. requirements, labor-intensive annotation of objects in video contents is inevitable. For this reason, many researchers have actively studied about methods of annotating the object that appear on the video. In keyword-based annotation related information of the object that appeared on the video content is immediately added and annotation data including all related information about the object must be individually managed. Users will have to directly input all related information to the object. Consequently, when a user browses for information that related to the object, user can only find and get limited resources that solely exists in annotated data. Also, in order to place annotation for objects user's huge workload is required. To cope with reducing user's workload and to minimize the work involved in annotation, in existing object-based annotation automatic annotation is being attempted using computer vision techniques like object detection, recognition and tracking. By using such computer vision techniques a wide variety of objects that appears on the video content must be all detected and recognized. But until now it is still a problem facing some difficulties which have to deal with automated annotation. To overcome these difficulties, we propose a system which consists of two modules. The first module is the annotation module that enables many annotators to collaboratively annotate the objects in the video content in order to access the semantic data using Linked Data. Annotation data managed by annotation server is represented using ontology so that the information can easily be shared and extended. Since annotation data does not include all the relevant information of the object, existing objects in Linked Data and objects that appear in the video content simply connect with each other to get all the related information of the object. In other words, annotation data which contains only URI and metadata like position, time and size are stored on the annotation sever. So when user needs other related information about the object, all of that information is retrieved from Linked Data through its relevant URI. The second module enables viewers to browse interesting information about the object using annotation data which is collaboratively generated by many users while watching video. With this system, through simple user interaction the query is automatically generated and all the related information is retrieved from Linked Data and finally all the additional information of the object is offered to the user. With this study, in the future of Semantic Web environment our proposed system is expected to establish a better video content service environment by offering users relevant information about the objects that appear on the screen of any internet-capable devices such as PC, smart TV or smart phone.

Development of a Web-based Presentation Attitude Correction Program Centered on Analyzing Facial Features of Videos through Coordinate Calculation (좌표계산을 통해 동영상의 안면 특징점 분석을 중심으로 한 웹 기반 발표 태도 교정 프로그램 개발)

  • Kwon, Kihyeon;An, Suho;Park, Chan Jung
    • The Journal of the Korea Contents Association
    • /
    • v.22 no.2
    • /
    • pp.10-21
    • /
    • 2022
  • In order to improve formal presentation attitudes such as presentation of job interviews and presentation of project results at the company, there are few automated methods other than observation by colleagues or professors. In previous studies, it was reported that the speaker's stable speech and gaze processing affect the delivery power in the presentation. Also, there are studies that show that proper feedback on one's presentation has the effect of increasing the presenter's ability to present. In this paper, considering the positive aspects of correction, we developed a program that intelligently corrects the wrong presentation habits and attitudes of college students through facial analysis of videos and analyzed the proposed program's performance. The proposed program was developed through web-based verification of the use of redundant words and facial recognition and textualization of the presentation contents. To this end, an artificial intelligence model for classification was developed, and after extracting the video object, facial feature points were recognized based on the coordinates. Then, using 4000 facial data, the performance of the algorithm in this paper was compared and analyzed with the case of facial recognition using a Teachable Machine. Use the program to help presenters by correcting their presentation attitude.

Intelligent Collision Prevention Technique for Construction Equipment using Ultrasound Scanning (초음파 스캐닝을 활용한 지능형 건설기계 충돌방지 기술)

  • Lee, Jaehoon;Hwang, Yeongseo;Yang, Kanghyeok
    • Korean Journal of Construction Engineering and Management
    • /
    • v.22 no.5
    • /
    • pp.48-54
    • /
    • 2021
  • According to the Ministry of Employment and Labor's statistics on occupational fatalities in South Korea, more than half of the fatalities in the past five years have occurred in the construction industry. The stuck-by and caught-in-between accidents associated with construction equipment is the major source of fatalities from construction sites. In order to prevent such accidents in construction sites, the government has spent lots of efforts including proposing the "special law on construction safety" and encouraging the implementation of new technology for accident prevention. However, numerous accidents are still occurred at construction sites and further efforts are still required. In this manner, this study developed a collision prevention technique that can prevent collision between equipment and worker by recognizing location and type of the nearby objects through ultrasound scanning. The study conducted a pilot experiment and the analysis results demonstrate the feasibility of achieving high performance in both object recognition and location estimation. The developed technique will contribute to prevent collision accidents at construction sites and provide the supplemental knowledge on developing automated collision prevention system for construction equipment.

Modified Center Weight Filter Algorithm using Pixel Segmentation of Local Area in AWGN Environments (AWGN 환경에서 국부영역의 화소분할을 사용한 변형된 중심 가중치 필터 알고리즘)

  • Cheon, Bong-Won;Kim, Nam-Ho
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2022.10a
    • /
    • pp.250-252
    • /
    • 2022
  • Recently, with the development of IoT technology and AI, unmanned and automated systems are progressing in various fields, and various application technologies are being studied in systems using algorithms such as object detection, recognition, and tracking. In the case of a system operating based on an image, noise removal is performed as a pre-processing process, and precise noise removal is sometimes required depending on the environment of the system. In this paper, we propose a modified central weight filter algorithm using pixel division of local regions to minimize the blurring that tends to occur in the filtering process and to emphasize the details of the resulting image. In the proposed algorithm, when a pixel of a local area is divided into two areas, the center of the dominant area among the divided areas is set as a criterion for the weight filter algorithm. The resulting image is calculated by convolving the transformed center weight with the pixel value inside the filtering mask.

  • PDF

Method of Biological Information Analysis Based-on Object Contextual (대상객체 맥락 기반 생체정보 분석방법)

  • Kim, Kyung-jun;Kim, Ju-yeon
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2022.05a
    • /
    • pp.41-43
    • /
    • 2022
  • In order to prevent and block infectious diseases caused by the recent COVID-19 pandemic, non-contact biometric information acquisition and analysis technology is attracting attention. The invasive and attached biometric information acquisition method accurately has the advantage of measuring biometric information, but has a risk of increasing contagious diseases due to the close contact. To solve these problems, the non-contact method of extracting biometric information such as human fingerprints, faces, iris, veins, voice, and signatures with automated devices is increasing in various industries as data processing speed increases and recognition accuracy increases. However, although the accuracy of the non-contact biometric data acquisition technology is improved, the non-contact method is greatly influenced by the surrounding environment of the object to be measured, which is resulting in distortion of measurement information and poor accuracy. In this paper, we propose a context-based bio-signal modeling technique for the interpretation of personalized information (image, signal, etc.) for bio-information analysis. Context-based biometric information modeling techniques present a model that considers contextual and user information in biometric information measurement in order to improve performance. The proposed model analyzes signal information based on the feature probability distribution through context-based signal analysis that can maximize the predicted value probability.

  • PDF

Ontology Modeling and Rule-based Reasoning for Automatic Classification of Personal Media (미디어 영상 자동 분류를 위한 온톨로지 모델링 및 규칙 기반 추론)

  • Park, Hyun-Kyu;So, Chi-Seung;Park, Young-Tack
    • Journal of KIISE
    • /
    • v.43 no.3
    • /
    • pp.370-379
    • /
    • 2016
  • Recently personal media were produced in a variety of ways as a lot of smart devices have been spread and services using these data have been desired. Therefore, research has been actively conducted for the media analysis and recognition technology and we can recognize the meaningful object from the media. The system using the media ontology has the disadvantage that can't classify the media appearing in the video because of the use of a video title, tags, and script information. In this paper, we propose a system to automatically classify video using the objects shown in the media data. To do this, we use a description logic-based reasoning and a rule-based inference for event processing which may vary in order. Description logic-based reasoning system proposed in this paper represents the relation of the objects in the media as activity ontology. We describe how to another rule-based reasoning system defines an event according to the order of the inference activity and order based reasoning system automatically classify the appropriate event to the category. To evaluate the efficiency of the proposed approach, we conducted an experiment using the media data classified as a valid category by the analysis of the Youtube video.