• Title/Summary/Keyword: data quality

Search Result 20,604, Processing Time 0.046 seconds

A Data Quality Management Maturity Model

  • Ryu, Kyung-Seok;Park, Joo-Seok;Park, Jae-Hong
    • ETRI Journal
    • /
    • v.28 no.2
    • /
    • pp.191-204
    • /
    • 2006
  • Many previous studies of data quality have focused on the realization and evaluation of both data value quality and data service quality. These studies revealed that poor data value quality and poor data service quality were caused by poor data structure. In this study we focus on metadata management, namely, data structure quality and introduce the data quality management maturity model as a preferred maturity model. We empirically show that data quality improves as data management matures.

  • PDF

A propose of Big data quality elements (빅 데이터의 품질 요소 제안)

  • Choi, Sang-Kyoon;Jeon, Soon-Cheon
    • Journal of Advanced Navigation Technology
    • /
    • v.17 no.1
    • /
    • pp.9-15
    • /
    • 2013
  • Big data has a key engine of the new value creation and troubleshooting are becoming more data-centric era begins in earnest. This paper takes advantage of the big data, big data in order to secure the quality of the quality elements for ensuring the quality of Justice and quality per-element strategy argue against. To achieve this, big data, case studies, resources of the big data plan and the elements of knowledge, analytical skills and big data processing technology, and more. This defines the quality of big data and quality, quality strategy. The quality of the data is secured by big companies from the large amounts of data through the data reinterpreted in big corporate competitiveness and to extract data for various strategies.

Proposal of Process Model for Research Data Quality Management (연구데이터 품질관리를 위한 프로세스 모델 제안)

  • Na-eun Han
    • Journal of the Korean Society for information Management
    • /
    • v.40 no.1
    • /
    • pp.51-71
    • /
    • 2023
  • This study analyzed the government data quality management model, big data quality management model, and data lifecycle model for research data management, and analyzed the components common to each data quality management model. Those data quality management models are designed and proposed according to the lifecycle or based on the PDCA model according to the characteristics of target data, which is the object that performs quality management. And commonly, the components of planning, collection and construction, operation and utilization, and preservation and disposal are included. Based on this, the study proposed a process model for research data quality management, in particular, the research data quality management to be performed in a series of processes from collecting to servicing on a research data platform that provides services using research data as target data was discussed in the stages of planning, construction and operation, and utilization. This study has significance in providing knowledge based for research data quality management implementation methods.

Saliency Score-Based Visualization for Data Quality Evaluation

  • Kim, Yong Ki;Lee, Keon Myung
    • International Journal of Fuzzy Logic and Intelligent Systems
    • /
    • v.15 no.4
    • /
    • pp.289-294
    • /
    • 2015
  • Data analysts explore collections of data to search for valuable information using various techniques and tricks. Garbage in, garbage out is a well-recognized idiom that emphasizes the importance of the quality of data in data analysis. It is therefore crucial to validate the data quality in the early stage of data analysis, and an effective method of evaluating the quality of data is hence required. In this paper, a method to visually characterize the quality of data using the notion of a saliency score is introduced. The saliency score is a measure comprising five indexes that captures certain aspects of data quality. Some experiment results are presented to show the applicability of proposed method.

An Empirical Study on the Effects of Source Data Quality on the Usefulness and Utilization of Big Data Analytics Results (원천 데이터 품질이 빅데이터 분석결과의 유용성과 활용도에 미치는 영향)

  • Park, Sohyun;Lee, Kukhie;Lee, Ayeon
    • Journal of Information Technology Applications and Management
    • /
    • v.24 no.4
    • /
    • pp.197-214
    • /
    • 2017
  • This study sheds light on the source data quality in big data systems. Previous studies about big data success have called for future research and further examination of the quality factors and the importance of source data. This study extracted the quality factors of source data from the user's viewpoint and empirically tested the effects of source data quality on the usefulness and utilization of big data analytics results. Based on the previous researches and focus group evaluation, four quality factors have been established such as accuracy, completeness, timeliness and consistency. After setting up 11 hypotheses on how the quality of the source data contributes to the usefulness, utilization, and ongoing use of the big data analytics results, e-mail survey was conducted at a level of independent department using big data in domestic firms. The results of the hypothetical review identified the characteristics and impact of the source data quality in the big data systems and drew some meaningful findings about big data characteristics.

Method for improving video/image data quality for AI learning of unstructured data (비정형데이터의 AI학습을 위한 영상/이미지 데이터 품질 향상 방법)

  • Kim Seung Hee;Dongju Ryu
    • Convergence Security Journal
    • /
    • v.23 no.2
    • /
    • pp.55-66
    • /
    • 2023
  • Recently, there is an increasing movement to increase the value of AI learning data and to secure high-quality data based on previous research on AI learning data in all areas of society. Therefore, quality management is very important in construction projects to secure high-quality data. In this paper, quality management to secure high-quality data when building AI learning data and improvement plans for each construction process are presented. In particular, more than 80% of the data quality of unstructured data built for AI learning is determined during the construction process. In this paper, we performed quality inspection of image/video data. In addition, we identified inspection procedures and problem elements that occurred in the construction phases of acquisition, data cleaning, labeling, and models, and suggested ways to secure high-quality data by solving them. Through this, it is expected that it will be an alternative to overcome the quality deviation of data for research groups and operators participating in the construction of AI learning data.

A Study on the Derivation of Items for Development of Data Quality Standard for 3D Building Data in National Digital Twin (디지털 트윈국토 건물 데이터 품질 표준 개발을 위한 항목 도출에 관한 연구)

  • Kim, Byeongsun;Lee, Heeseok;Hong, Sangki
    • Journal of Cadastre & Land InformatiX
    • /
    • v.52 no.1
    • /
    • pp.37-55
    • /
    • 2022
  • This study presents the plans to derive quality items for develop the data quality standard for ensuring the quality of 3D building geospatial data in NDT(National Digital Twin). This paper is organized as follows. The first section briefly examines various factors that impact the quality of 3D geospatial data, and proposes the role and necessity of the data quality standard as a means of addressing the data errors properly and also meeting the minimum requirements of stakeholders. The second section analyzes the relationship between the standards - building data model for NDT and ISO 19157: Geospatial data quality - in order to consider directly relevant standards. Finally, we suggest three plans on developing NDT data quality standard: (1) the scope for evaluating data quality, (2) additional quality elements(geometric integrity, geometric fidelity, positional accuracy and semantic classification accuracy), and (3) NDT data quality items model based on ISO 19157. The plans reveled through the study would contribute to establish a way for the national standard on NDT data quality as well as the other standards associated with NDT over the coming years.

Development of Automated Tools for Data Quality Diagnostics (데이터 품질진단을 위한 자동화도구 개발)

  • Ko, Jae-Hwan;Kim, Dong-Soo;Han, Ki-Joon
    • Journal of Information Technology Services
    • /
    • v.11 no.4
    • /
    • pp.153-170
    • /
    • 2012
  • When companies or institutes manage data, in order to utilize it as useful resources for decision-making, it is essential to offer precise and reliable data. While most small and medium-sized enterprises and public institutes have been investing a great amount of money in management and maintenance of their data systems, the investment in data management has been inadequate. When public institutions establish their data systems, inspection has been constantly carried out on the data systems in order to improve safety and effectiveness. However, their capabilities in improving the quality of data have been insufficient. This study develops an automatic tool to diagnose the quality of data in a way to diagnose the data quality condition of the inspected institute quantitatively at the stage of design and closure by inspecting the data system and proves its practicality by applying the automatic tool to inspection. As a means to diagnose the quality, this study categorizes, in the aspect of quality characteristics, the items that may be improved through diagnosis at the stage of design, the early stage of establishing the data system and the measurement items by the quality index regarding measurable data values at the stage of establishment and operation. The study presents a way of quantitative measurement regarding the data structures and data values by concretizing the measurement items by quality index in a function of the automatic tool program. Also, the practicality of the tool is proved by applying the tool in the inspection field. As a result, the areas which the institute should improve are reported objectively through a complete enumeration survey on the diagnosed items and the indicators for quality improvement are presented quantitatively by presenting the quality condition quantitatively.

A Data Quality Measuring Tool (데이타 품질 측정 도구)

  • 양자영;최병주
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.9 no.3
    • /
    • pp.278-288
    • /
    • 2003
  • Quality of the software is affected by quality of data required for operating the actual software. Especially, it is important that assure the quality of data in a knowledge-engineering system that extracts the meaningful knowledge from stored data. In this paper, we developed DAQUM tool that can measure quality of data. This paper shows: 1) main contents for implement of DAQUM tool; 2) detection of dirty data via DAQUM tool through case study and measurement of data quality which is quantifiable from end-user's point of view. DAQUM tool will greatly contribute to improving quality of software product that processes mainly the data through control and measurement of data quality.

Quality Characteristics of Public Open Data (공공개방데이터 품질 특성에 관한 연구)

  • Park, Go-Eun;Kim, Chang-Jae
    • Journal of Digital Convergence
    • /
    • v.13 no.10
    • /
    • pp.135-146
    • /
    • 2015
  • Public data open is one of the important tasks of Korea Government 3.0. By making open data available to the private sector, the goal is to create jobs, increase innovation and improve quality of life. Public data open is a policy that emphasized its importance worldwide. Open data should have adequate quality in order to achieve the object of the public. However, there are open data's quality problems due to the lack of data quality management and standardization. The purpose of this study is to derive data characteristics of public open data from existing researches. In addition, the model was modified and verified through a survey targeting the experts on public open data. The study indicates that public open data's quality characteristics as publicity, usability, reliability, suitability. This study is significant in that it suggests quality characteristics to improve the data quality and promote utilization of the open data.