• Title/Summary/Keyword: Videos

Search Result 1,523, Processing Time 0.03 seconds

An Automatic Summarization System of Baseball Game Video Using the Caption Information (자막 정보를 이용한 야구경기 비디오의 자동요약 시스템)

  • 유기원;허영식
    • Journal of Broadcast Engineering
    • /
    • v.7 no.2
    • /
    • pp.107-113
    • /
    • 2002
  • In this paper, we propose a method and a software system for automatic summarization of baseball game videos. The proposed system pursues fast execution and high accuracy of summarization. To satisfy the requirement, the detection of important events in baseball video is performed through DC-based shot boundary detection algorithm and simple caption recognition method. Furthermore, the proposed system supports a hierarchical description so that users can browse and navigate videos in several levels of summarization. In this paper, we propose a method and a software system for automatic summarization of baseball game videos. The proposed system pursues fast execution and high accuracy of summarization. To satisfy the requirement, the detection of important events in baseball video is performed through DC-based shot boundary detection algorithm and simple caption recognition method. Furthermore, the proposed system supports a hierarchical description so that users can browse and navigate videos in several levels of summarization.

Scenario-Based Design of The Next Generation Information Appliances (시나리오 기반 차세대 정보가전 신제품 개발)

  • 박지수
    • Archives of design research
    • /
    • v.16 no.2
    • /
    • pp.35-48
    • /
    • 2003
  • Home networking technology connects individual home appliances through a wired or wireless network and makes possible new functions that were impossible when they were used independently. However, the new functions must not simply be confusing arrays of functions that are possible to implement, but those absolutely necessary to the users. To develop innovative information appliances with such functions, scenarios were used and played guiding roles in suggesting new product ideas, making design mockups, and producing videos to show natural situations where the products would be used in home of the future. In the phase of suggesting new product ideas, user action scenarios in the home, generated by a team consisting of experts in the fields of cognitive engineering, user interface, computer science, cultural anthropology, interaction design, and product design, helped the team identify user needs and design factors necessary to fulfill those needs and suggest new product ideas from the design factors. In the phase of making design mockups, the procedures of using the products were described in the scenario format. Based on the scenarios the s쇼les and the user interfaces of them were designed. In the phase of producing videos, the interactions between the user and the product were embodied in the course of professional writers'arranging the scenarios of using the products for the scripts of the videos. Videos were produced to show the actual situations where the design mockups would be used in home of the future and the dynamic aspects of interaction design.

  • PDF

Enhancement Method of CCTV Video Quality Based on SRGAN (SRGAN 기반의 CCTV 영상 화질 개선 기법)

  • Ha, Hyunsoo;Hwang, Byung-Yeon
    • Journal of Korea Multimedia Society
    • /
    • v.21 no.9
    • /
    • pp.1027-1034
    • /
    • 2018
  • CCTV has been known to possess high level of objectivity and utility. Hence, the government has recently focused on replacing low quality CCTV with higher quality ones or even by adding high resolution CCTV. However, converting all existing low-quality CCTV to high quality can be extremely costly. Furthermore, low quality videos prior to CCTV replacement are likely to be of poor quality and thus not utilized correctly. In order to solve these problems, this paper proposes a method to improve videos quality of images using SRGAN(Super Resolution Generative Advisory Networks). Through experiments, we have proven that it is possible to improve low quality CCTV videos clearly. For this experiment, a total of 4 types of CCTV videos were used and 10,000 images were sampled from each type. Those images could then be used for machine learning. The fact that the pre-process for machine learning has been done manually and the long time that required for machine learning seems to be complementary.

Spatio-Temporal Residual Networks for Slide Transition Detection in Lecture Videos

  • Liu, Zhijin;Li, Kai;Shen, Liquan;Ma, Ran;An, Ping
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.13 no.8
    • /
    • pp.4026-4040
    • /
    • 2019
  • In this paper, we present an approach for detecting slide transitions in lecture videos by introducing the spatio-temporal residual networks. Given a lecture video which records the digital slides, the speaker, and the audience by multiple cameras, our goal is to find keyframes where slide content changes. Since temporal dependency among video frames is important for detecting slide changes, 3D Convolutional Networks has been regarded as an efficient approach to learn the spatio-temporal features in videos. However, 3D ConvNet will cost much training time and need lots of memory. Hence, we utilize ResNet to ease the training of network, which is easy to optimize. Consequently, we present a novel ConvNet architecture based on 3D ConvNet and ResNet for slide transition detection in lecture videos. Experimental results show that the proposed novel ConvNet architecture achieves the better accuracy than other slide progression detection approaches.

Video-Induced Visit to a Destination: The Psychological Mechanism and the Role of Video Sharing Sources

  • Wang, Erin Yirun;Luo, Sara Yushan;Fong, Lawrence Hoc Nang;Law, Rob
    • Journal of Smart Tourism
    • /
    • v.1 no.3
    • /
    • pp.7-16
    • /
    • 2021
  • Promotional videos have been increasingly adopted in social media marketing. Given that video production incurs high costs, evaluating the effectiveness of promotional videos is necessary. This research examines the effect of promotional video on visit intention through affective and cognitive destination images, respectively. Furthermore, grounded in the theoretical reasoning of social ties and persuasive intent, the effects of promotional videos disseminated by different sources (i.e., unknown users, friends, and marketers) are compared. Two experimental studies were carried out with 200 participants in Study 1 and 243 participants in Study 2. Study 1 revealed that the presence of promotional video induces visit intention through affective and cognitive destination images. Study 2 found that promotional videos disseminated by friends (versus unknown users) are more effective to induce visit intention because of positive cognitive destination image, while the video effects are indifferent between marketers and users. Meaningful implications are provided for destination marketers.

Using CNN- VGG 16 to detect the tennis motion tracking by information entropy and unascertained measurement theory

  • Zhong, Yongfeng;Liang, Xiaojun
    • Advances in nano research
    • /
    • v.12 no.2
    • /
    • pp.223-239
    • /
    • 2022
  • Object detection has always been to pursue objects with particular properties or representations and to predict details on objects including the positions, sizes and angle of rotation in the current picture. This was a very important subject of computer vision science. While vision-based object tracking strategies for the analysis of competitive videos have been developed, it is still difficult to accurately identify and position a speedy small ball. In this study, deep learning (DP) network was developed to face these obstacles in the study of tennis motion tracking from a complex perspective to understand the performance of athletes. This research has used CNN-VGG 16 to tracking the tennis ball from broadcasting videos while their images are distorted, thin and often invisible not only to identify the image of the ball from a single frame, but also to learn patterns from consecutive frames, then VGG 16 takes images with 640 to 360 sizes to locate the ball and obtain high accuracy in public videos. VGG 16 tests 99.6%, 96.63%, and 99.5%, respectively, of accuracy. In order to avoid overfitting, 9 additional videos and a subset of the previous dataset are partly labelled for the 10-fold cross-validation. The results show that CNN-VGG 16 outperforms the standard approach by a wide margin and provides excellent ball tracking performance.

Research on the Discourse of Libraries During COVID-19 in YouTube Videos Using Topic Modeling and Social Network Analysis

  • Euikyung Oh;Ok Nam Park
    • Journal of Information Science Theory and Practice
    • /
    • v.11 no.3
    • /
    • pp.29-42
    • /
    • 2023
  • This study explored issues related to the library in the COVID-19 era in YouTube videos in Korea. This study performed social network analysis and topic modeling analysis by collecting 479 YouTube videos, 20,545 words, and 8,379 channels related to COVID-19 and the library from 2019 to 2020. The study results confirmed that YouTube, a social media platform, was used as an important medium to connect users and physical libraries and provide/promote online library services. In the study, major topics and keywords such as quarantine, vlog, and library identity during the COVID-19 pandemic, library services and functions, and introductions and user guides of libraries were derived. Additionally, it was identified that videos about COVID-19 and the library are being produced by various actors (news and media channels, libraries, government agencies, librarians, and individual users). However, the study also identified that the actor network is fragmented through the channel network, showing a low density or weak linkage, and that the centrality of the library in the actor network is weak.

Classification of Education Video by Subtitle Analysis (자막 분석을 통한 교육 영상의 카테고리 분류 방안)

  • Lee, Ji-Hoon;Lee, Hyeon Sup;Kim, Jin-Deog
    • Proceedings of the Korean Institute of Information and Commucation Sciences Conference
    • /
    • 2021.05a
    • /
    • pp.88-90
    • /
    • 2021
  • This paper introduces a method for extracting subtitles from lecture videos through a Korean morpheme analyzer and classifying video categories according to the extracted morpheme information. In some cases incorrect information is entered due to human error and reflected in the characteristics of the items, affecting the accuracy of the recommendation system. To prevent this, we generate a keyword table for each category using morpheme information extracted from pre-classified videos, and compare the similarity of morpheme in each category keyword table to classify categories of Lecture videos using the most similar keyword table. These human intervention reduction systems directly classify videos and aim to increase the accuracy of the system.

  • PDF

Movement Detection Using Keyframes in Video Surveillance System

  • Kim, Kyutae;Jia, Qiong;Dong, Tianyu;Jang, Euee S.
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2022.06a
    • /
    • pp.1249-1252
    • /
    • 2022
  • In this paper, we propose a conceptual framework that identifies video frames in motion containing the movement of people and vehicles in traffic videos. The automatic selection of video frames in motion is an important topic in security and surveillance video because the number of videos to be monitored simultaneously is simply too large due to limited human resources. The conventional method to identify the areas in motion is to compute the differences over consecutive video frames, which has been costly because of its high computational complexity. In this paper, we reduced the overall complexity by examining only the keyframes (or I-frames). The basic assumption is that the time period between I-frames is rather shorter (e.g., 1/10 ~ 3 secs) than the usual length of objects in motion in video (i.e., pedestrian walking, automobile passing, etc.). The proposed method estimates the possibility of videos containing motion between I-frames by evaluating the difference of consecutive I-frames with the long-time statistics of the previously decoded I-frames of the same video. The experimental results showed that the proposed method showed more than 80% accuracy in short surveillance videos obtained from different locations while keeping the computational complexity as low as 20 % compared to the HM decoder.

  • PDF

Two-Stage Deep Learning Based Algorithm for Cosmetic Object Recognition (화장품 물체 인식을 위한 Two-Stage 딥러닝 기반 알고리즘)

  • Jongmin Kim;Daeho Seo
    • Journal of Korean Society of Industrial and Systems Engineering
    • /
    • v.46 no.4
    • /
    • pp.101-106
    • /
    • 2023
  • With the recent surge in YouTube usage, there has been a proliferation of user-generated videos where individuals evaluate cosmetics. Consequently, many companies are increasingly utilizing evaluation videos for their product marketing and market research. However, a notable drawback is the manual classification of these product review videos incurring significant costs and time. Therefore, this paper proposes a deep learning-based cosmetics search algorithm to automate this task. The algorithm consists of two networks: One for detecting candidates in images using shape features such as circles, rectangles, etc and Another for filtering and categorizing these candidates. The reason for choosing a Two-Stage architecture over One-Stage is that, in videos containing background scenes, it is more robust to first detect cosmetic candidates before classifying them as specific objects. Although Two-Stage structures are generally known to outperform One-Stage structures in terms of model architecture, this study opts for Two-Stage to address issues related to the acquisition of training and validation data that arise when using One-Stage. Acquiring data for the algorithm that detects cosmetic candidates based on shape and the algorithm that classifies candidates into specific objects is cost-effective, ensuring the overall robustness of the algorithm.