• Title/Summary/Keyword: Feature mapping

Search Result 334, Processing Time 0.027 seconds

A Numerical Speech Recognition by Parameters Estimated from the Data on the Estimated Plane and a Neural Network (추정평면에서 평가한 데이터와 인공신경망에 의한 숫자음 인식)

  • Choi, Il-Hong;Jang, Seung-Kwan;Cha, Tae-Hoo;Choi, Ung-Se;Kim, Chang-Seok
    • The Journal of the Acoustical Society of Korea
    • /
    • v.15 no.4
    • /
    • pp.58-64
    • /
    • 1996
  • This paper was proposed the recognition method by using parameters which was estimated from the data on the estimated plane and a neural network. After the LPC estimated in each frame algorithm was mapped to the estimated plane by the optimum feature mapping function, we estimated the C-LPC and the maximum and minimum value and 3 divided power from the mapping data on the estimated plane. As a result of the experiment of the speech recognition that those parameters were applied to the input of a neural network, it was found that those parameters estimated from the estimated plane have the features of the original speech for a change in the time scale and that the recongnition rate by the proposed methods was 96.3 percent.

  • PDF

2D-MELPP: A two dimensional matrix exponential based extension of locality preserving projections for dimensional reduction

  • Xiong, Zixun;Wan, Minghua;Xue, Rui;Yang, Guowei
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.16 no.9
    • /
    • pp.2991-3007
    • /
    • 2022
  • Two dimensional locality preserving projections (2D-LPP) is an improved algorithm of 2D image to solve the small sample size (SSS) problems which locality preserving projections (LPP) meets. It's able to find the low dimension manifold mapping that not only preserves local information but also detects manifold embedded in original data spaces. However, 2D-LPP is simple and elegant. So, inspired by the comparison experiments between two dimensional linear discriminant analysis (2D-LDA) and linear discriminant analysis (LDA) which indicated that matrix based methods don't always perform better even when training samples are limited, we surmise 2D-LPP may meet the same limitation as 2D-LDA and propose a novel matrix exponential method to enhance the performance of 2D-LPP. 2D-MELPP is equivalent to employing distance diffusion mapping to transform original images into a new space, and margins between labels are broadened, which is beneficial for solving classification problems. Nonetheless, the computational time complexity of 2D-MELPP is extremely high. In this paper, we replace some of matrix multiplications with multiple multiplications to save the memory cost and provide an efficient way for solving 2D-MELPP. We test it on public databases: random 3D data set, ORL, AR face database and Polyu Palmprint database and compare it with other 2D methods like 2D-LDA, 2D-LPP and 1D methods like LPP and exponential locality preserving projections (ELPP), finding it outperforms than others in recognition accuracy. We also compare different dimensions of projection vector and record the cost time on the ORL, AR face database and Polyu Palmprint database. The experiment results above proves that our advanced algorithm has a better performance on 3 independent public databases.

The Role of Animation Technical Director of Disney's 3D Feature Animation (디즈니 극장용 3D 애니메이션에서 애니메이션 테크니컬 디렉터의 역할)

  • Paik, Jiwon;Kim, Jae-Woong
    • Cartoon and Animation Studies
    • /
    • s.37
    • /
    • pp.491-508
    • /
    • 2014
  • As number of making 3D feature animation films is increasing, 3D production pipeline become more complicated and more artists are needed than before. Major studios in foreign countries, in burden of producing high quality films with limited amount of budget and time, have been handling such difficulties by hiring technical directors in each department such as animation, rigging, cloth hair, and effect. Technical director is new occupation which appears after trend of producing animation is changed from 2D to 3D. Importance of technical director is increasing in respect to studios' needs which are related to complication in production time, manpower, budget, and production pipeline. This research is based on the researcher's work experience as an animation TD at Walt Diseny Animation Studio and Sony Pictures Imageworks, interview with working professionals, and related books and thesis. It focuses on the role of animation technical director in Disney's 3D feature animation film from two perspectives, 'Designing Production Pipeline' and 'Analyzing Problem of Shot'. Animation technical directors design and test production pipeline so that they can detect and solve problems that may arise in production process as early as possible. They not only analyze numerous problems of characters or shots limited to animation department but also in other departments such as modeling, mapping, character rigging, cloth, hair, lighting, rendering, software development in order to support artists to complete their shots according to the production schedule. In accordance with recent trend of increasing number of 3D feature animation film production in South Korea and collaboration with foreign studios outside of South Korea, it is vital to train animation technical directors who can develop production pipeline, analyze various problems of shots and characters to escalate efficiency in production.

Analysis of Metadata Standards of Record Management for Metadata Interoperability From the viewpoint of the Task model and 5W1H (메타데이터 상호운용성을 위한 기록관리 메타데이터 표준 분석 5W1H와 태스크 모델의 관점에서)

  • Baek, Jae-Eun;Sugimoto, Shigeo
    • The Korean Journal of Archival Studies
    • /
    • no.32
    • /
    • pp.127-176
    • /
    • 2012
  • Metadata is well recognized as one of the foundational factors in archiving and long-term preservation of digital resources. There are several metadata standards for records management, archives and preservation, e.g. ISAD(G), EAD, AGRkMs, PREMIS, and OAIS. Consideration is important in selecting appropriate metadata standards in order to design metadata schema that meet the requirements of a particular archival system. Interoperability of metadata with other systems should be considered in schema design. In our previous research, we have presented a feature analysis of metadata standards by identifying the primary resource lifecycle stages where each standard is applied. We have clarified that any single metadata standard cannot cover the whole records lifecycle for archiving and preservation. Through this feature analysis, we analyzed the features of metadata in the whole records lifecycle, and we clarified the relationships between the metadata standards and the stages of the lifecycle. In the previous study, more detailed analysis was left for future study. This paper proposes to analyze the metadata schemas from the viewpoint of tasks performed in the lifecycle. Metadata schemas are primarily defined to describe properties of a resource in accordance with the purposes of description, e.g. finding aids, records management, preservation and so forth. In other words, the metadata standards are resource- and purpose-centric, and the resource lifecycle is not explicitly reflected in the standards. There are no systematic methods for mapping between different metadata standards in accordance with the lifecycle. This paper proposes a method for mapping between metadata standards based on the tasks contained in the resource lifecycle. We first propose a Task Model to clarify tasks applied to resources in each stage of the lifecycle. This model is created as a task-centric model to identify features of metadata standards and to create mappings among elements of those standards. It is important to categorize the elements in order to limit the semantic scope of mapping among elements and decrease the number of combinations of elements for mapping. This paper proposes to use 5W1H (Who, What, Why, When, Where, How) model to categorize the elements. 5W1H categories are generally used for describing events, e.g. news articles. As performing a task on a resource causes an event and metadata elements are used in the event, we consider that the 5W1H categories are adequate to categorize the elements. By using these categories, we determine the features of every element of metadata standards which are AGLS, AGRkMS, PREMIS, EAD, OAIS and an attribute set extracted from DPC decision flow. Then, we perform the element mapping between the standards, and find the relationships between the standards. In this study, we defined a set of terms for each of 5W1H categories, which typically appear in the definition of an element, and used those terms to categorize the elements. For example, if the definition of an element includes the terms such as person and organization that mean a subject which contribute to create, modify a resource the element is categorized into the Who category. A single element can be categorized into one or more 5W1H categories. Thus, we categorized every element of the metadata standards using the 5W1H model, and then, we carried out mapping among the elements in each category. We conclude that the Task Model provides a new viewpoint for metadata schemas and is useful to help us understand the features of metadata standards for records management and archives. The 5W1H model, which is defined based on the Task Model, provides us a core set of categories to semantically classify metadata elements from the viewpoint of an event caused by a task.

Performance Improvement of Radial Basis Function Neural Networks Using Adaptive Feature Extraction (적응적 특징추출을 이용한 Radial Basis Function 신경망의 성능개선)

  • 조용현
    • Journal of Korea Multimedia Society
    • /
    • v.3 no.3
    • /
    • pp.253-262
    • /
    • 2000
  • This paper proposes a new RBF neural network that determines the number and the center of hidden neurons based on the adaptive feature extraction for the input data. The principal component analysis is applied for extracting adaptively the features by reducing the dimension of the given input data. It can simultaneously achieve a superior property of both the principal component analysis by mapping input data into set of statistically independent features and the RBF neural networks. The proposed neural networks has been applied to classify the 200 breast cancer databases by 2-class. The simulation results shows that the proposed neural networks has better performances of the learning time and the classification for test data, in comparison with those using the k-means clustering algorithm. And it is affected less than the k-means clustering algorithm by the initial weight setting and the scope of the smoothing factor.

  • PDF

A Study on the Determination of Grain Size of Heat-treated Stainless Steel Using Digital Ultrasonic Signal Processing Techniques. (디지털 초음파 신호처리 기법을 이용한 열처리된 스테인레스 스틸의 그레인 크기 결정에 관한 연구)

  • 임내묵;이영석;김성환
    • The Journal of the Acoustical Society of Korea
    • /
    • v.18 no.8
    • /
    • pp.84-93
    • /
    • 1999
  • Determination of grain size of heat-treated stainless steel based fm digital ultrasonic signal processing technique is presented. This techniques consist in evidence accumulation with multiple feature parameters, difference absolute mean value(DAMV), variance(VAR), mean frequency (MEANF), auto regressive model coefficient(ARC) and linear cepstrum coefficient(LCC). Feature parameters were extracted from ultrasonic echo signal of heat-treated metals. It was found that a few parameters might not be sufficient to exactly evaluate the grain size of heat-treated metals. The determination of grain size of heat-treated metals was carried out through the evidence accumulation procedure using the distances measured with reference parameters. A fuzzy mapping function is designed to transform the distances for the application of the evidence accumulation method. In the work presented, heat-treated stainless steel samples with various grain sizes are examined. The processed experimental results supports the feasibility of the grain size determination technique presented.

  • PDF

Mobile Robot Localization and Mapping using Scale-Invariant Features (스케일 불변 특징을 이용한 이동 로봇의 위치 추정 및 매핑)

  • Lee, Jong-Shill;Shen, Dong-Fan;Kwon, Oh-Sang;Lee, Eung-Hyuk;Hong, Seung-Hong
    • Journal of IKEEE
    • /
    • v.9 no.1 s.16
    • /
    • pp.7-18
    • /
    • 2005
  • A key component of an autonomous mobile robot is to localize itself accurately and build a map of the environment simultaneously. In this paper, we propose a vision-based mobile robot localization and mapping algorithm using scale-invariant features. A camera with fisheye lens facing toward to ceiling is attached to the robot to acquire high-level features with scale invariance. These features are used in map building and localization process. As pre-processing, input images from fisheye lens are calibrated to remove radial distortion then labeling and convex hull techniques are used to segment ceiling region from wall region. At initial map building process, features are calculated for segmented regions and stored in map database. Features are continuously calculated from sequential input images and matched against existing map until map building process is finished. If features are not matched, they are added to the existing map. Localization is done simultaneously with feature matching at map building process. Localization. is performed when features are matched with existing map and map building database is updated at same time. The proposed method can perform a map building in 2 minutes on $50m^2$ area. The positioning accuracy is ${\pm}13cm$, the average error on robot angle with the positioning is ${\pm}3$ degree.

  • PDF

Improved SIM Algorithm for Contents-based Image Retrieval (내용 기반 이미지 검색을 위한 개선된 SIM 방법)

  • Kim, Kwang-Baek
    • Journal of Intelligence and Information Systems
    • /
    • v.15 no.2
    • /
    • pp.49-59
    • /
    • 2009
  • Contents-based image retrieval methods are in general more objective and effective than text-based image retrieval algorithms since they use color and texture in search and avoid annotating all images for search. SIM(Self-organizing Image browsing Map) is one of contents-based image retrieval algorithms that uses only browsable mapping results obtained by SOM(Self Organizing Map). However, SOM may have an error in selecting the right BMU in learning phase if there are similar nodes with distorted color information due to the intensity of light or objects' movements in the image. Such images may be mapped into other grouping nodes thus the search rate could be decreased by this effect. In this paper, we propose an improved SIM that uses HSV color model in extracting image features with color quantization. In order to avoid unexpected learning error mentioned above, our SOM consists of two layers. In learning phase, SOM layer 1 has the color feature vectors as input. After learning SOM Layer 1, the connection weights of this layer become the input of SOM Layer 2 and re-learning occurs. With this multi-layered SOM learning, we can avoid mapping errors among similar nodes of different color information. In search, we put the query image vector into SOM layer 2 and select nodes of SOM layer 1 that connects with chosen BMU of SOM layer 2. In experiment, we verified that the proposed SIM was better than the original SIM and avoid mapping error effectively.

  • PDF

Brain MRI Template-Driven Medical Images Mapping Method Based on Semantic Features for Ischemic Stroke (허혈성 뇌졸중을 위한 뇌 자기공명영상의 의미적 특징 기반 템플릿 중심 의료 영상 매핑 기법)

  • Park, Ye-Seul;Lee, Meeyeon;Lee, Jung-Won
    • KIPS Transactions on Software and Data Engineering
    • /
    • v.5 no.2
    • /
    • pp.69-78
    • /
    • 2016
  • Ischemic stroke is a disease that the brain tissues cannot function by reducing blood flow due to thrombosis or embolisms. Due to the nature of the disease, it is most important to identify the status of cerebral vessel and the medical images are necessarily used for its diagnosis. Among many indicators, brain MRI is most widely utilized because experts can effectively obtain the semantic information such as cerebral anatomy aiding the diagnosis with it. However, in case of emergency diseases like ischemic stroke, even though a intelligent system is required for supporting the prompt diagnosis and treatment, the current systems have some difficulties to provide the information of medical images intuitively. In other words, as the current systems have managed the medical images based on the basic meta-data such as image name, ID and so on, they cannot consider semantic information inherent in medical images. Therefore, in this paper, to provide core information like cerebral anatomy contained in brain MRI, we suggest a template-driven medical images mapping method. The key idea of the method is defining the mapping characteristics between anatomic feature and representative images by using template images that can be representative of the whole brain MRI image set and revealing the semantic relations that only medical experts can check between images. With our method, it will be possible to manage the medical images based on semantic.

Correction of Erroneous Model Key Points Extracted from Segmented Laser Scanner Data and Accuracy Evaluation

  • Yoo, Eun Jin;Park, So Young;Yom, Jae-Hong;Lee, Dong-Cheon
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.31 no.6_2
    • /
    • pp.611-623
    • /
    • 2013
  • Point cloud data (i.e., LiDAR; Light Detection and Ranging) collected by Airborne Laser Scanner (ALS) system is one of the major sources for surface reconstruction including DEM generation, topographic mapping and object modeling. Recently, demand and requirement of the accurate and realistic Digital Building Model (DBM) increase for geospatial platforms and spatial data infrastructure. The main issues in the object modeling such as building and city modeling are efficiency of the methodology and quality of the final products. Efficiency and quality are associated with automation and accuracy, respectively. However, these two factors are often opposite each other. This paper aims to introduce correction scheme of incorrectly determined Model Key Points (MKPs) regardless of the segmentation method. Planimetric and height locations of the MKPs were refined by surface patch fitting based on the Least-Squares Solution (LESS). The proposed methods were applied to the synthetic and real LiDAR data. Finally, the results were analyzed by comparing adjusted MKPs with the true building model data.