• Title/Summary/Keyword: Statistical Methodology

Search Result 1,300, Processing Time 0.026 seconds

Optimization of the Reaction Conditions and the Effect of Surfactants on the Kinetic Resolution of [R,S]-Naoroxen 2,2,2-Trifluoroethyl Thioester by Using Lipse (리파아제를 이용한 라세믹 나프록센 2,2,2-트리플로로에틸 씨오에스터의 Kinetic Resolution에서 반응조건 죄적화와 계면활성제 영향)

  • Song, Yoon-Seok;Lee, Jung-Ho;Cho, Sang-Won;Kang, Seong-Woo;Kim, Seung-Wook
    • KSBB Journal
    • /
    • v.23 no.3
    • /
    • pp.257-262
    • /
    • 2008
  • In this study, the reaction conditions for lipase-catalyzed resolution of racemic naproxen 2,2,2-trilfluoroethyl thioester were optimized, and the effect of surfactants was investigated. Among the organic solvents tested, the isooctane showed the highest conversion (92.19%) in a hydrolytic reaction of (S)-naproxen 2,2,2-trifluoroethyl thioester. In addition, the isooctane induced the highest initial reaction rate of (S)-naproxen 2,2,2-trifluoroethyl thioester ($V_s=2.34{\times}10^{-2}mM/h$), the highest enantioselectivity (E = 36.12) and the highest specific activity ($V_s/(E_t)=7.80{\times}10^{-4}mmol/h{\cdot}g$) of lipase. Furthermore, reaction conditions such as temperature, concentration of the substrate and enzyme, and agitation speed were optimized using response surface methodology (RSM), and the statistical analysis indicated that the optimal conditions were $48.2^{\circ}C$, 3.51 mM, 30.11 mg/mL and 180 rpm, respectively. When the optimal reaction conditions were used, the conversion of (S)-naproxen 2,2,2-trifluoroethyl thioester was 96.5%, which is similar to the conversion (94.6%) that was predicted by the model. After optimization of reaction conditions, the initial reaction rate, lipase specific activity and conversion of (S)-naproxen 2,2,2-trifluoroethyl thioester increased by approximately 19.54%, 19.12% and 4.05%, respectively. The effect of surfactants such as Triton X-100 and NP-10 was also studied and NP-10 showed the highest conversion (89.43%), final reaction rate of (S)-naproxen 2,2,2-trifluoroethyl thioester ($V_s=1.175{\times}10^{-2}mM/h$) and enantioselectivity (E = 59.24) of lipase.

Development of Market Growth Pattern Map Based on Growth Model and Self-organizing Map Algorithm: Focusing on ICT products (자기조직화 지도를 활용한 성장모형 기반의 시장 성장패턴 지도 구축: ICT제품을 중심으로)

  • Park, Do-Hyung;Chung, Jaekwon;Chung, Yeo Jin;Lee, Dongwon
    • Journal of Intelligence and Information Systems
    • /
    • v.20 no.4
    • /
    • pp.1-23
    • /
    • 2014
  • Market forecasting aims to estimate the sales volume of a product or service that is sold to consumers for a specific selling period. From the perspective of the enterprise, accurate market forecasting assists in determining the timing of new product introduction, product design, and establishing production plans and marketing strategies that enable a more efficient decision-making process. Moreover, accurate market forecasting enables governments to efficiently establish a national budget organization. This study aims to generate a market growth curve for ICT (information and communication technology) goods using past time series data; categorize products showing similar growth patterns; understand markets in the industry; and forecast the future outlook of such products. This study suggests the useful and meaningful process (or methodology) to identify the market growth pattern with quantitative growth model and data mining algorithm. The study employs the following methodology. At the first stage, past time series data are collected based on the target products or services of categorized industry. The data, such as the volume of sales and domestic consumption for a specific product or service, are collected from the relevant government ministry, the National Statistical Office, and other relevant government organizations. For collected data that may not be analyzed due to the lack of past data and the alteration of code names, data pre-processing work should be performed. At the second stage of this process, an optimal model for market forecasting should be selected. This model can be varied on the basis of the characteristics of each categorized industry. As this study is focused on the ICT industry, which has more frequent new technology appearances resulting in changes of the market structure, Logistic model, Gompertz model, and Bass model are selected. A hybrid model that combines different models can also be considered. The hybrid model considered for use in this study analyzes the size of the market potential through the Logistic and Gompertz models, and then the figures are used for the Bass model. The third stage of this process is to evaluate which model most accurately explains the data. In order to do this, the parameter should be estimated on the basis of the collected past time series data to generate the models' predictive value and calculate the root-mean squared error (RMSE). The model that shows the lowest average RMSE value for every product type is considered as the best model. At the fourth stage of this process, based on the estimated parameter value generated by the best model, a market growth pattern map is constructed with self-organizing map algorithm. A self-organizing map is learning with market pattern parameters for all products or services as input data, and the products or services are organized into an $N{\times}N$ map. The number of clusters increase from 2 to M, depending on the characteristics of the nodes on the map. The clusters are divided into zones, and the clusters with the ability to provide the most meaningful explanation are selected. Based on the final selection of clusters, the boundaries between the nodes are selected and, ultimately, the market growth pattern map is completed. The last step is to determine the final characteristics of the clusters as well as the market growth curve. The average of the market growth pattern parameters in the clusters is taken to be a representative figure. Using this figure, a growth curve is drawn for each cluster, and their characteristics are analyzed. Also, taking into consideration the product types in each cluster, their characteristics can be qualitatively generated. We expect that the process and system that this paper suggests can be used as a tool for forecasting demand in the ICT and other industries.

A Study on the Recognition of Modern Cultural Heritage Value of Japanese-style Building Groups Using Q Methodology - Focusing on Huam-dong, Seoul - (Q 방법론을 이용한 일본식 건물군의 근대문화유산 가치에 관한 인식 연구 - 서울시 후암동을 중심으로 -)

  • Park, Han-Sol;Sung, Jong-Sang
    • Journal of the Korean Institute of Landscape Architecture
    • /
    • v.47 no.6
    • /
    • pp.115-128
    • /
    • 2019
  • Huam-dong is a representative area from the Japanese colonial period and is the space where most Japanese-style buildings remain in Seoul. Interest in modern cultural heritage continues to increase, including the registration of cultural properties in 2001, building assets in 2015, and the registration of cultural property units in 2018. As the debate continues over the necessity of preserving cultural heritage that reminds us of the Japanese colonial, there is a need for research to grasp the perceptions of stakeholders along with the perceived value of such spaces. This study identified the subjective perception types of the stakeholders concerned with the Japanese-style building group in Huam-dong, analyzed characteristics by types, and debated the issues. For this purpose, Q methodology, which is a statistical technique for measuring human self-subjectivity and extracting common human perspectives, was used. A literature study on the values of Huam-dong and modern cultural heritage was conducted, and a Q questionnaire based on five aspects of modern cultural heritage values (historical, architectural, sociocultural, landscape, and economic) was applied. The results of the study depicted three types of cognition and showed different attitudes toward the Japanese building group. This study found a conflict comparing the perceptional differences between the types of cognition. This study is meaningful in that it provides an in-depth approach to the perspectives of the stakeholders concerned with the Japanese-style buildings clustered in central Seoul. It is also meant to present a theoretical framework that can be applied to the use area as sustainable cultural heritage through the establishment of preservation and utilization of Japanese-style areas and conflict resolution.

Understanding the Protox Inhibition Activity of Novel 1-(5-methyl-3-phenylisoxazolin-5-yl)methoxy-2-chloro-4-fluorobenzene Derivatives Using Comparative Molecular Similarity Indices Analysis (CoMSIA) Methodology (비교 분자 유사성 지수분석(CoMSIA) 방법에 따른 1-(5-methyl-3-phenylisoxazolin-5-yl)methoxy-2-chlore-4-fluorobenzene 유도체들의 Protox 저해 활성에 관한 이해)

  • Song, Jong-Hwan;Park, Kyung-Yong;Sung, Nack-Do
    • Applied Biological Chemistry
    • /
    • v.47 no.4
    • /
    • pp.414-421
    • /
    • 2004
  • 3D QSAR studies for protox inhibition activities against root and shoot of the rice plant (Orysa sativa L.) and barnyardgrass (Echinochloa crus-galli) by a series of new 1-(5-methyl-3-phenylisoxazolin-5-yl)methoxy-2-chloro-4-fluorobenzene derivatives were conducted based on the results (Sung, N. D. et al.'s, (2004) J. Korean Soc. Appl. Biol. Chem. 47(3), 351-356) using comparative molecular similarity indices analysis (CoMSIA) methodology. Four CoMSIA models, without hydrogen bond donor field for the protox inhibition activities against root and shoot of the two plants, were derived from the combination of several fields using steric field, hydrophobic field, hydrogen bond acceptor field, LUMO molecular orbital field, dipole moment (DM) and molar refractivity (MR) as additional descriptors. The predictabilities and fitness of CoMSIA models for protox inhibition activities against barnyard-grass were higher than that of rice plant. The statistical results of these models showed the best predictability of the protox inhibition activities against barnyard-grass based on the cross-validated value $r^2\;_{cv}\;(q^2=0.635{\sim}0.924)$, non cross-validated, conventional coefficient $r^2\;_{ncv.}$ value $(r^2=0.928{\sim}0.977)$ and PRESS value $(0.255{\sim}0.273)$. The protox inhibition activities exhibited a strong correlation with the steric $(5.4{\sim}15.7%)$ and hydrophobic $(68.0{\sim}84.3%)$ factors of the molecules. Particularly, the CoMSIA models indicated that the groups of increasing steric bulk at ortho-position on the C-phenyl ring will enhance the protox inhibition activities against barnyard-grass and subsequently increase the selectivity.

Understanding the protox inhibition activity of novel 1-(5-methyl-3-phenylisoxazolin-5-yl)methoxy-2-chloro-4-fluorobenzene derivatives using comparative molecular field analysis (CoMFA) methodology (비교 분자장 분석 (CoMFA) 방법에 따른 1-(5-methyl-3-phenylisoxazolin-5-yl)methoxy-2-chloro-4-fluoro-benzene 유도체들의 Protox 저해 활성에 관한 이해)

  • Sung, Nack-Do;Song, Jong-Hwan;Yang, Sook-Young;Park, Kyeng-Yong
    • The Korean Journal of Pesticide Science
    • /
    • v.8 no.3
    • /
    • pp.151-161
    • /
    • 2004
  • Three dimensional quantitative structure-activity relationships (3D-QSAR) studies for the protox inhibition activities against root and shoot of rice plant (Orysa sativa L.) and barnyardgrass (Echinochloa crus-galli) by a series of new A=3,4,5,6-tetrahydrophthalimino, B=3-chloro-4,5,6,7-tetrahydro-2H-indazolyl and C=3,4-dimethylmaleimino group, and R-group substituted on the phenyl ring in 1-(5-methyl-3-phenylisoxazolin-5-yl)methoxy-2chloro-4-fluorobenzene derivatives were performed using comparative molecular field analyses (CoMFA) methodology with Gasteiger-Huckel charge. Four CoMFA models for the protox inhibition activities against root and shoot of the two plants were generated using 46 molecules as training set and the predictive ability of the each models was evaluated against a test set of 8 molecules. And the statistical results of these models with combination (SIH) of standard field, indicator field and H-bond field showed the best predictability of the protox inhibition activities based on the cross-validated value $r^2_{cv.}$ $(q^2=0.635\sim0.924)$, conventional coefficient $(r^2_{ncv.}=0.928\sim0.977)$ and PRESS value $(0.091\sim0.156)$, respectively. The activities exhibited a strong correlation with steric $(74.3\sim87.4%)$, electrostatic $(10.10\sim18.5%)$ and hydrophobic $(1.10\sim8.30%)$ factors of the molecules. The steric feature of molecule may be an important factor for the activities. We founded that an novel selective and higher protox inhibitors between the two plants may be designed by modification of X-subsitutents for barnyardgrass based upon the results obtained from CoMFA analyses.

Optimization of Extraction Conditions to Obtain Functional Components from Buckwheat (Fagopyrum esculentum M.) Sprouts, using Response Surface Methodology (반응표면분석법에 의한 메밀(Fagopyrum esculentum M.) 새싹 기능성분의 추출 조건 최적화)

  • Park, Kee-Jai;Lim, Jeong-Ho;Kim, Bum-Keun;Jeong, Jin-Woong;Kim, Jong-Chan;Lee, Myung-Heon;Cho, Young-Sim;Jung, Hee-Yong
    • Food Science and Preservation
    • /
    • v.16 no.5
    • /
    • pp.734-741
    • /
    • 2009
  • Response surface methodology (RSM) was used to optimize extraction conditions for functional components of buckwheat (Fagopyrum esculentum). A central composite design was applied to investigate the effects of three independent variables, namelyextraction temperature (X1), extraction time (X2), and ethanol concentration (X3), on responses including extraction yield (Y1), total phenolic content in the extract (Y2), $\alpha$-glucosidase inhibition activity (Y3), and acetylcholine esterase (ACE) inhibition activity (Y4). Data were analyzed using an expert design strategy and statistical software. The maximum yield was 24.95% (w/w) at $55.75^{\circ}C$ extraction temperature, 8.75 hextraction time, and 15.65% (v/v) ethanol. The maximum total phenolic yield was 222.45 mg/100 g under the conditions of $28.11^{\circ}C$ extraction temperature, 8.65 h extraction time, and 81.72% (v/v) ethanol. The maximum $\alpha$-glucosidase inhibition activity was 85.38% at $9.62^{\circ}C$, 7.86 h, and 57.58% (v/v) ethanol. The maximum ACE inhibition activity was 86.91% under extraction conditions of $10.12^{\circ}C$, 4.86 h, and 44.44% (v/v) ethanol. Based on superimposition of a four-dimensional RSM with respect to levels of total phenolics, $\alpha$-glucosidase inhibition activity, and ACE inhibition activity, obtained under various extraction conditions, the optimum ranges of conditions were an extraction temperature of $0-70^{\circ}C$, an extraction time of 2-8 h, and an ethanol concentration of 30-80% (v/v).

The Need for Paradigm Shift in Semantic Similarity and Semantic Relatedness : From Cognitive Semantics Perspective (의미간의 유사도 연구의 패러다임 변화의 필요성-인지 의미론적 관점에서의 고찰)

  • Choi, Youngseok;Park, Jinsoo
    • Journal of Intelligence and Information Systems
    • /
    • v.19 no.1
    • /
    • pp.111-123
    • /
    • 2013
  • Semantic similarity/relatedness measure between two concepts plays an important role in research on system integration and database integration. Moreover, current research on keyword recommendation or tag clustering strongly depends on this kind of semantic measure. For this reason, many researchers in various fields including computer science and computational linguistics have tried to improve methods to calculating semantic similarity/relatedness measure. This study of similarity between concepts is meant to discover how a computational process can model the action of a human to determine the relationship between two concepts. Most research on calculating semantic similarity usually uses ready-made reference knowledge such as semantic network and dictionary to measure concept similarity. The topological method is used to calculated relatedness or similarity between concepts based on various forms of a semantic network including a hierarchical taxonomy. This approach assumes that the semantic network reflects the human knowledge well. The nodes in a network represent concepts, and way to measure the conceptual similarity between two nodes are also regarded as ways to determine the conceptual similarity of two words(i.e,. two nodes in a network). Topological method can be categorized as node-based or edge-based, which are also called the information content approach and the conceptual distance approach, respectively. The node-based approach is used to calculate similarity between concepts based on how much information the two concepts share in terms of a semantic network or taxonomy while edge-based approach estimates the distance between the nodes that correspond to the concepts being compared. Both of two approaches have assumed that the semantic network is static. That means topological approach has not considered the change of semantic relation between concepts in semantic network. However, as information communication technologies make advantage in sharing knowledge among people, semantic relation between concepts in semantic network may change. To explain the change in semantic relation, we adopt the cognitive semantics. The basic assumption of cognitive semantics is that humans judge the semantic relation based on their cognition and understanding of concepts. This cognition and understanding is called 'World Knowledge.' World knowledge can be categorized as personal knowledge and cultural knowledge. Personal knowledge means the knowledge from personal experience. Everyone can have different Personal Knowledge of same concept. Cultural Knowledge is the knowledge shared by people who are living in the same culture or using the same language. People in the same culture have common understanding of specific concepts. Cultural knowledge can be the starting point of discussion about the change of semantic relation. If the culture shared by people changes for some reasons, the human's cultural knowledge may also change. Today's society and culture are changing at a past face, and the change of cultural knowledge is not negligible issues in the research on semantic relationship between concepts. In this paper, we propose the future directions of research on semantic similarity. In other words, we discuss that how the research on semantic similarity can reflect the change of semantic relation caused by the change of cultural knowledge. We suggest three direction of future research on semantic similarity. First, the research should include the versioning and update methodology for semantic network. Second, semantic network which is dynamically generated can be used for the calculation of semantic similarity between concepts. If the researcher can develop the methodology to extract the semantic network from given knowledge base in real time, this approach can solve many problems related to the change of semantic relation. Third, the statistical approach based on corpus analysis can be an alternative for the method using semantic network. We believe that these proposed research direction can be the milestone of the research on semantic relation.

The effect of Big-data investment on the Market value of Firm (기업의 빅데이터 투자가 기업가치에 미치는 영향 연구)

  • Kwon, Young jin;Jung, Woo-Jin
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.2
    • /
    • pp.99-122
    • /
    • 2019
  • According to the recent IDC (International Data Corporation) report, as from 2025, the total volume of data is estimated to reach ten times higher than that of 2016, corresponding to 163 zettabytes. then the main body of generating information is moving more toward corporations than consumers. So-called "the wave of Big-data" is arriving, and the following aftermath affects entire industries and firms, respectively and collectively. Therefore, effective management of vast amounts of data is more important than ever in terms of the firm. However, there have been no previous studies that measure the effects of big data investment, even though there are number of previous studies that quantitatively the effects of IT investment. Therefore, we quantitatively analyze the Big-data investment effects, which assists firm's investment decision making. This study applied the Event Study Methodology, which is based on the efficient market hypothesis as the theoretical basis, to measure the effect of the big data investment of firms on the response of market investors. In addition, five sub-variables were set to analyze this effect in more depth: the contents are firm size classification, industry classification (finance and ICT), investment completion classification, and vendor existence classification. To measure the impact of Big data investment announcements, Data from 91 announcements from 2010 to 2017 were used as data, and the effect of investment was more empirically observed by observing changes in corporate value immediately after the disclosure. This study collected data on Big Data Investment related to Naver 's' News' category, the largest portal site in Korea. In addition, when selecting the target companies, we extracted the disclosures of listed companies in the KOSPI and KOSDAQ market. During the collection process, the search keywords were searched through the keywords 'Big data construction', 'Big data introduction', 'Big data investment', 'Big data order', and 'Big data development'. The results of the empirically proved analysis are as follows. First, we found that the market value of 91 publicly listed firms, who announced Big-data investment, increased by 0.92%. In particular, we can see that the market value of finance firms, non-ICT firms, small-cap firms are significantly increased. This result can be interpreted as the market investors perceive positively the big data investment of the enterprise, allowing market investors to better understand the company's big data investment. Second, statistical demonstration that the market value of financial firms and non - ICT firms increases after Big data investment announcement is proved statistically. Third, this study measured the effect of big data investment by dividing by company size and classified it into the top 30% and the bottom 30% of company size standard (market capitalization) without measuring the median value. To maximize the difference. The analysis showed that the investment effect of small sample companies was greater, and the difference between the two groups was also clear. Fourth, one of the most significant features of this study is that the Big Data Investment announcements are classified and structured according to vendor status. We have shown that the investment effect of a group with vendor involvement (with or without a vendor) is very large, indicating that market investors are very positive about the involvement of big data specialist vendors. Lastly but not least, it is also interesting that market investors are evaluating investment more positively at the time of the Big data Investment announcement, which is scheduled to be built rather than completed. Applying this to the industry, it would be effective for a company to make a disclosure when it decided to invest in big data in terms of increasing the market value. Our study has an academic implication, as prior research looked for the impact of Big-data investment has been nonexistent. This study also has a practical implication in that it can be a practical reference material for business decision makers considering big data investment.

Knowledge Extraction Methodology and Framework from Wikipedia Articles for Construction of Knowledge-Base (지식베이스 구축을 위한 한국어 위키피디아의 학습 기반 지식추출 방법론 및 플랫폼 연구)

  • Kim, JaeHun;Lee, Myungjin
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.1
    • /
    • pp.43-61
    • /
    • 2019
  • Development of technologies in artificial intelligence has been rapidly increasing with the Fourth Industrial Revolution, and researches related to AI have been actively conducted in a variety of fields such as autonomous vehicles, natural language processing, and robotics. These researches have been focused on solving cognitive problems such as learning and problem solving related to human intelligence from the 1950s. The field of artificial intelligence has achieved more technological advance than ever, due to recent interest in technology and research on various algorithms. The knowledge-based system is a sub-domain of artificial intelligence, and it aims to enable artificial intelligence agents to make decisions by using machine-readable and processible knowledge constructed from complex and informal human knowledge and rules in various fields. A knowledge base is used to optimize information collection, organization, and retrieval, and recently it is used with statistical artificial intelligence such as machine learning. Recently, the purpose of the knowledge base is to express, publish, and share knowledge on the web by describing and connecting web resources such as pages and data. These knowledge bases are used for intelligent processing in various fields of artificial intelligence such as question answering system of the smart speaker. However, building a useful knowledge base is a time-consuming task and still requires a lot of effort of the experts. In recent years, many kinds of research and technologies of knowledge based artificial intelligence use DBpedia that is one of the biggest knowledge base aiming to extract structured content from the various information of Wikipedia. DBpedia contains various information extracted from Wikipedia such as a title, categories, and links, but the most useful knowledge is from infobox of Wikipedia that presents a summary of some unifying aspect created by users. These knowledge are created by the mapping rule between infobox structures and DBpedia ontology schema defined in DBpedia Extraction Framework. In this way, DBpedia can expect high reliability in terms of accuracy of knowledge by using the method of generating knowledge from semi-structured infobox data created by users. However, since only about 50% of all wiki pages contain infobox in Korean Wikipedia, DBpedia has limitations in term of knowledge scalability. This paper proposes a method to extract knowledge from text documents according to the ontology schema using machine learning. In order to demonstrate the appropriateness of this method, we explain a knowledge extraction model according to the DBpedia ontology schema by learning Wikipedia infoboxes. Our knowledge extraction model consists of three steps, document classification as ontology classes, proper sentence classification to extract triples, and value selection and transformation into RDF triple structure. The structure of Wikipedia infobox are defined as infobox templates that provide standardized information across related articles, and DBpedia ontology schema can be mapped these infobox templates. Based on these mapping relations, we classify the input document according to infobox categories which means ontology classes. After determining the classification of the input document, we classify the appropriate sentence according to attributes belonging to the classification. Finally, we extract knowledge from sentences that are classified as appropriate, and we convert knowledge into a form of triples. In order to train models, we generated training data set from Wikipedia dump using a method to add BIO tags to sentences, so we trained about 200 classes and about 2,500 relations for extracting knowledge. Furthermore, we evaluated comparative experiments of CRF and Bi-LSTM-CRF for the knowledge extraction process. Through this proposed process, it is possible to utilize structured knowledge by extracting knowledge according to the ontology schema from text documents. In addition, this methodology can significantly reduce the effort of the experts to construct instances according to the ontology schema.

Development of Systematic Process for Estimating Commercialization Duration and Cost of R&D Performance (기술가치 평가를 위한 기술사업화 기간 및 비용 추정체계 개발)

  • Jun, Seoung-Pyo;Choi, Daeheon;Park, Hyun-Woo;Seo, Bong-Goon;Park, Do-Hyung
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.2
    • /
    • pp.139-160
    • /
    • 2017
  • Technology commercialization creates effective economic value by linking the company's R & D processes and outputs to the market. This technology commercialization is important in that a company can retain and maintain a sustained competitive advantage. In order for a specific technology to be commercialized, it goes through the stage of technical planning, technology research and development, and commercialization. This process involves a lot of time and money. Therefore, the duration and cost of technology commercialization are important decision information for determining the market entry strategy. In addition, it is more important information for a technology investor to rationally evaluate the technology value. In this way, it is very important to scientifically estimate the duration and cost of the technology commercialization. However, research on technology commercialization is insufficient and related methodology are lacking. In this study, we propose an evaluation model that can estimate the duration and cost of R & D technology commercialization for small and medium-sized enterprises. To accomplish this, this study collected the public data of the National Science & Technology Information Service (NTIS) and the survey data provided by the Small and Medium Business Administration. Also this study will develop the estimation model of commercialization duration and cost of R&D performance on using these data based on the market approach, one of the technology valuation methods. Specifically, this study defined the process of commercialization as consisting of development planning, development progress, and commercialization. We collected the data from the NTIS database and the survey of SMEs technical statistics of the Small and Medium Business Administration. We derived the key variables such as stage-wise R&D costs and duration, the factors of the technology itself, the factors of the technology development, and the environmental factors. At first, given data, we estimates the costs and duration in each technology readiness level (basic research, applied research, development research, prototype production, commercialization), for each industry classification. Then, we developed and verified the research model of each industry classification. The results of this study can be summarized as follows. Firstly, it is reflected in the technology valuation model and can be used to estimate the objective economic value of technology. The duration and the cost from the technology development stage to the commercialization stage is a critical factor that has a great influence on the amount of money to discount the future sales from the technology. The results of this study can contribute to more reliable technology valuation because it estimates the commercialization duration and cost scientifically based on past data. Secondly, we have verified models of various fields such as statistical model and data mining model. The statistical model helps us to find the important factors to estimate the duration and cost of technology Commercialization, and the data mining model gives us the rules or algorithms to be applied to an advanced technology valuation system. Finally, this study reaffirms the importance of commercialization costs and durations, which has not been actively studied in previous studies. The results confirm the significant factors to affect the commercialization costs and duration, furthermore the factors are different depending on industry classification. Practically, the results of this study can be reflected in the technology valuation system, which can be provided by national research institutes and R & D staff to provide sophisticated technology valuation. The relevant logic or algorithm of the research result can be implemented independently so that it can be directly reflected in the system, so researchers can use it practically immediately. In conclusion, the results of this study can be a great contribution not only to the theoretical contributions but also to the practical ones.