• Title/Summary/Keyword: FINANCE

Search Result 4,585, Processing Time 0.028 seconds

A Study on Risk Parity Asset Allocation Model with XGBoos (XGBoost를 활용한 리스크패리티 자산배분 모형에 관한 연구)

  • Kim, Younghoon;Choi, HeungSik;Kim, SunWoong
    • Journal of Intelligence and Information Systems
    • /
    • v.26 no.1
    • /
    • pp.135-149
    • /
    • 2020
  • Artificial intelligences are changing world. Financial market is also not an exception. Robo-Advisor is actively being developed, making up the weakness of traditional asset allocation methods and replacing the parts that are difficult for the traditional methods. It makes automated investment decisions with artificial intelligence algorithms and is used with various asset allocation models such as mean-variance model, Black-Litterman model and risk parity model. Risk parity model is a typical risk-based asset allocation model which is focused on the volatility of assets. It avoids investment risk structurally. So it has stability in the management of large size fund and it has been widely used in financial field. XGBoost model is a parallel tree-boosting method. It is an optimized gradient boosting model designed to be highly efficient and flexible. It not only makes billions of examples in limited memory environments but is also very fast to learn compared to traditional boosting methods. It is frequently used in various fields of data analysis and has a lot of advantages. So in this study, we propose a new asset allocation model that combines risk parity model and XGBoost machine learning model. This model uses XGBoost to predict the risk of assets and applies the predictive risk to the process of covariance estimation. There are estimated errors between the estimation period and the actual investment period because the optimized asset allocation model estimates the proportion of investments based on historical data. these estimated errors adversely affect the optimized portfolio performance. This study aims to improve the stability and portfolio performance of the model by predicting the volatility of the next investment period and reducing estimated errors of optimized asset allocation model. As a result, it narrows the gap between theory and practice and proposes a more advanced asset allocation model. In this study, we used the Korean stock market price data for a total of 17 years from 2003 to 2019 for the empirical test of the suggested model. The data sets are specifically composed of energy, finance, IT, industrial, material, telecommunication, utility, consumer, health care and staple sectors. We accumulated the value of prediction using moving-window method by 1,000 in-sample and 20 out-of-sample, so we produced a total of 154 rebalancing back-testing results. We analyzed portfolio performance in terms of cumulative rate of return and got a lot of sample data because of long period results. Comparing with traditional risk parity model, this experiment recorded improvements in both cumulative yield and reduction of estimated errors. The total cumulative return is 45.748%, about 5% higher than that of risk parity model and also the estimated errors are reduced in 9 out of 10 industry sectors. The reduction of estimated errors increases stability of the model and makes it easy to apply in practical investment. The results of the experiment showed improvement of portfolio performance by reducing the estimated errors of the optimized asset allocation model. Many financial models and asset allocation models are limited in practical investment because of the most fundamental question of whether the past characteristics of assets will continue into the future in the changing financial market. However, this study not only takes advantage of traditional asset allocation models, but also supplements the limitations of traditional methods and increases stability by predicting the risks of assets with the latest algorithm. There are various studies on parametric estimation methods to reduce the estimated errors in the portfolio optimization. We also suggested a new method to reduce estimated errors in optimized asset allocation model using machine learning. So this study is meaningful in that it proposes an advanced artificial intelligence asset allocation model for the fast-developing financial markets.

Adaptive RFID anti-collision scheme using collision information and m-bit identification (충돌 정보와 m-bit인식을 이용한 적응형 RFID 충돌 방지 기법)

  • Lee, Je-Yul;Shin, Jongmin;Yang, Dongmin
    • Journal of Internet Computing and Services
    • /
    • v.14 no.5
    • /
    • pp.1-10
    • /
    • 2013
  • RFID(Radio Frequency Identification) system is non-contact identification technology. A basic RFID system consists of a reader, and a set of tags. RFID tags can be divided into active and passive tags. Active tags with power source allows their own operation execution and passive tags are small and low-cost. So passive tags are more suitable for distribution industry than active tags. A reader processes the information receiving from tags. RFID system achieves a fast identification of multiple tags using radio frequency. RFID systems has been applied into a variety of fields such as distribution, logistics, transportation, inventory management, access control, finance and etc. To encourage the introduction of RFID systems, several problems (price, size, power consumption, security) should be resolved. In this paper, we proposed an algorithm to significantly alleviate the collision problem caused by simultaneous responses of multiple tags. In the RFID systems, in anti-collision schemes, there are three methods: probabilistic, deterministic, and hybrid. In this paper, we introduce ALOHA-based protocol as a probabilistic method, and Tree-based protocol as a deterministic one. In Aloha-based protocols, time is divided into multiple slots. Tags randomly select their own IDs and transmit it. But Aloha-based protocol cannot guarantee that all tags are identified because they are probabilistic methods. In contrast, Tree-based protocols guarantee that a reader identifies all tags within the transmission range of the reader. In Tree-based protocols, a reader sends a query, and tags respond it with their own IDs. When a reader sends a query and two or more tags respond, a collision occurs. Then the reader makes and sends a new query. Frequent collisions make the identification performance degrade. Therefore, to identify tags quickly, it is necessary to reduce collisions efficiently. Each RFID tag has an ID of 96bit EPC(Electronic Product Code). The tags in a company or manufacturer have similar tag IDs with the same prefix. Unnecessary collisions occur while identifying multiple tags using Query Tree protocol. It results in growth of query-responses and idle time, which the identification time significantly increases. To solve this problem, Collision Tree protocol and M-ary Query Tree protocol have been proposed. However, in Collision Tree protocol and Query Tree protocol, only one bit is identified during one query-response. And, when similar tag IDs exist, M-ary Query Tree Protocol generates unnecessary query-responses. In this paper, we propose Adaptive M-ary Query Tree protocol that improves the identification performance using m-bit recognition, collision information of tag IDs, and prediction technique. We compare our proposed scheme with other Tree-based protocols under the same conditions. We show that our proposed scheme outperforms others in terms of identification time and identification efficiency.

Jang(Fermented Soybean) in Official and Royal Documents in Chosun Dynasty Period (조선조의 공문서 및 왕실자료에 나타난 장류)

  • Ann, Yong-Geun
    • The Korean Journal of Food And Nutrition
    • /
    • v.25 no.2
    • /
    • pp.368-382
    • /
    • 2012
  • This paper investigated the system that is relevant to Jang(fermented soybean paste or solution), the relief of hunger-stricken people by Jang, 33 kinds of Jang, and its consumption in the documents, such as the annals of the Chosun Dynasty, Ihlseong-document, Seungjeongwon daily, Uigwe(record of national ceremony), official documents on the basis of Kyujanggak institute for the Korean studies and data base of Korean classics. There are lots of Jang named after the place of particular soybean's production from the ancient times. Jang, soybean, salt and Meju(source of Jang), during the Dynasty, were collected as taxation or tribute. In the 5th year of Hyeonjong(1664), the storage amount of soybean in Hojo(ministry of finance) was 16,200 $k{\ell}$, and its consumption was 7,694 $k{\ell}$ a year. In the 32nd year of Yongjo(1756), the 1,800 $k{\ell}$ of soybean was distributed to the people at the time of disaster, and in his 36th year(1756), the 15,426 $k{\ell}$ of soybean was reduced from the soybean taxation nationwide. The offices managing Jang are Naejashi, Saseonseo, Sadoshi, Yebinshi and Bongsangshi. Chongyoongcheong(Gyeonggi military headquarters) stored the 175.14 $k{\ell}$ of Jang, and the 198 $k{\ell}$ of Jang in Yebinshi. There are such posts managing Jang as Jangsaek, Jangdoo, and Saseonsikjang. In the year of Jeongjong(1777~1800), the royal family distributed the 3.6 $k{\ell}$ of Meju to Gasoon-court, Hygyeong-court, queen's mother-court, queen's court, royal palace. The 13.41 $k{\ell}$ of Gamjang(fermented soybean solution) was distributed to the Gasoon-court, 17.23 $k{\ell}$ to Hegyeong-court, 17.09 $k{\ell}$ to the queen's mother-court, and the 17.17 $k{\ell}$ to the queen's court each. There are 112 Jang-storing pots in the royal storages, and the 690 are in Namhan-hill, where the 2.7 $k{\ell}$ of fermented Jang was made and brought back by them each year. At the time of starvation, Jang relieved the starving people. There are 20 occasions of big reliefs, according to the annals of the Chosun Dynasty. In the 5th year of Sejong(1423), the 360 $k{\ell}$ of Jang was given to the hunger-stricken people. In his 6th year(1424), the 8,512.92 $k{\ell}$ of rice, bean, and Jang was provided and in the 28th year(1446), the 8,322.68 $k{\ell}$ of Jang was also provided to them. In the Dynasty, Jang was given as a salary. In case that when they were bereaved, they didn't eat Jang patiently for its preservation. They were awarded for their filial piety. In the annals of the Chosun Dynasty, there are 19 kinds of Jang. They are listed in the order of Jang(108), Yeomjang(90), Maljang(11), Yookjang(5), Gamjang(4), and etc.,. In Seungjeongwon daily, there are 11 kinds of Jang. Jang(6), Cheongjang (5), Maljang(5), and Tojang(3) are listed in order. In the Ihlseong-document, there are 5 kinds of Jang. They are listed in Jang(15), Maljang(2), Gamjang(2), and etc.,. There are 13 kinds of Jang in Uigwe, and the official documents, in the order of Gamjang(59), Ganjang(37), Jang(28), Yeomjang(7), Maljang(6), and Cheongjang(5). In addition, shi are Jeonshi(7), and Dooshi(4). All these are made of only soybean except, for Yookjang. The most-frequently recorded Jang among anthology, cookbook, the annals of the Chosun Dynasty, Ihlseong-document, Seoungjeongwon daily, Uigwe, or official document is Jang(372), and then Yeomjang(194), Gamjang(73), Cheongjang(46), Ganjang(46), Soojang(33), and Maljang(26), which were made of soybean. Jang from China in cookbook is not in anthology and royal palace documents. Thus, traditional Jang made of soybean was used in the daily food life in the royal court, and in the public during the Chosun period.

Intelligent Brand Positioning Visualization System Based on Web Search Traffic Information : Focusing on Tablet PC (웹검색 트래픽 정보를 활용한 지능형 브랜드 포지셔닝 시스템 : 태블릿 PC 사례를 중심으로)

  • Jun, Seung-Pyo;Park, Do-Hyung
    • Journal of Intelligence and Information Systems
    • /
    • v.19 no.3
    • /
    • pp.93-111
    • /
    • 2013
  • As Internet and information technology (IT) continues to develop and evolve, the issue of big data has emerged at the foreground of scholarly and industrial attention. Big data is generally defined as data that exceed the range that can be collected, stored, managed and analyzed by existing conventional information systems and it also refers to the new technologies designed to effectively extract values from such data. With the widespread dissemination of IT systems, continual efforts have been made in various fields of industry such as R&D, manufacturing, and finance to collect and analyze immense quantities of data in order to extract meaningful information and to use this information to solve various problems. Since IT has converged with various industries in many aspects, digital data are now being generated at a remarkably accelerating rate while developments in state-of-the-art technology have led to continual enhancements in system performance. The types of big data that are currently receiving the most attention include information available within companies, such as information on consumer characteristics, information on purchase records, logistics information and log information indicating the usage of products and services by consumers, as well as information accumulated outside companies, such as information on the web search traffic of online users, social network information, and patent information. Among these various types of big data, web searches performed by online users constitute one of the most effective and important sources of information for marketing purposes because consumers search for information on the internet in order to make efficient and rational choices. Recently, Google has provided public access to its information on the web search traffic of online users through a service named Google Trends. Research that uses this web search traffic information to analyze the information search behavior of online users is now receiving much attention in academia and in fields of industry. Studies using web search traffic information can be broadly classified into two fields. The first field consists of empirical demonstrations that show how web search information can be used to forecast social phenomena, the purchasing power of consumers, the outcomes of political elections, etc. The other field focuses on using web search traffic information to observe consumer behavior, identifying the attributes of a product that consumers regard as important or tracking changes on consumers' expectations, for example, but relatively less research has been completed in this field. In particular, to the extent of our knowledge, hardly any studies related to brands have yet attempted to use web search traffic information to analyze the factors that influence consumers' purchasing activities. This study aims to demonstrate that consumers' web search traffic information can be used to derive the relations among brands and the relations between an individual brand and product attributes. When consumers input their search words on the web, they may use a single keyword for the search, but they also often input multiple keywords to seek related information (this is referred to as simultaneous searching). A consumer performs a simultaneous search either to simultaneously compare two product brands to obtain information on their similarities and differences, or to acquire more in-depth information about a specific attribute in a specific brand. Web search traffic information shows that the quantity of simultaneous searches using certain keywords increases when the relation is closer in the consumer's mind and it will be possible to derive the relations between each of the keywords by collecting this relational data and subjecting it to network analysis. Accordingly, this study proposes a method of analyzing how brands are positioned by consumers and what relationships exist between product attributes and an individual brand, using simultaneous search traffic information. It also presents case studies demonstrating the actual application of this method, with a focus on tablets, belonging to innovative product groups.

Bankruptcy Forecasting Model using AdaBoost: A Focus on Construction Companies (적응형 부스팅을 이용한 파산 예측 모형: 건설업을 중심으로)

  • Heo, Junyoung;Yang, Jin Yong
    • Journal of Intelligence and Information Systems
    • /
    • v.20 no.1
    • /
    • pp.35-48
    • /
    • 2014
  • According to the 2013 construction market outlook report, the liquidation of construction companies is expected to continue due to the ongoing residential construction recession. Bankruptcies of construction companies have a greater social impact compared to other industries. However, due to the different nature of the capital structure and debt-to-equity ratio, it is more difficult to forecast construction companies' bankruptcies than that of companies in other industries. The construction industry operates on greater leverage, with high debt-to-equity ratios, and project cash flow focused on the second half. The economic cycle greatly influences construction companies. Therefore, downturns tend to rapidly increase the bankruptcy rates of construction companies. High leverage, coupled with increased bankruptcy rates, could lead to greater burdens on banks providing loans to construction companies. Nevertheless, the bankruptcy prediction model concentrated mainly on financial institutions, with rare construction-specific studies. The bankruptcy prediction model based on corporate finance data has been studied for some time in various ways. However, the model is intended for all companies in general, and it may not be appropriate for forecasting bankruptcies of construction companies, who typically have high liquidity risks. The construction industry is capital-intensive, operates on long timelines with large-scale investment projects, and has comparatively longer payback periods than in other industries. With its unique capital structure, it can be difficult to apply a model used to judge the financial risk of companies in general to those in the construction industry. Diverse studies of bankruptcy forecasting models based on a company's financial statements have been conducted for many years. The subjects of the model, however, were general firms, and the models may not be proper for accurately forecasting companies with disproportionately large liquidity risks, such as construction companies. The construction industry is capital-intensive, requiring significant investments in long-term projects, therefore to realize returns from the investment. The unique capital structure means that the same criteria used for other industries cannot be applied to effectively evaluate financial risk for construction firms. Altman Z-score was first published in 1968, and is commonly used as a bankruptcy forecasting model. It forecasts the likelihood of a company going bankrupt by using a simple formula, classifying the results into three categories, and evaluating the corporate status as dangerous, moderate, or safe. When a company falls into the "dangerous" category, it has a high likelihood of bankruptcy within two years, while those in the "safe" category have a low likelihood of bankruptcy. For companies in the "moderate" category, it is difficult to forecast the risk. Many of the construction firm cases in this study fell in the "moderate" category, which made it difficult to forecast their risk. Along with the development of machine learning using computers, recent studies of corporate bankruptcy forecasting have used this technology. Pattern recognition, a representative application area in machine learning, is applied to forecasting corporate bankruptcy, with patterns analyzed based on a company's financial information, and then judged as to whether the pattern belongs to the bankruptcy risk group or the safe group. The representative machine learning models previously used in bankruptcy forecasting are Artificial Neural Networks, Adaptive Boosting (AdaBoost) and, the Support Vector Machine (SVM). There are also many hybrid studies combining these models. Existing studies using the traditional Z-Score technique or bankruptcy prediction using machine learning focus on companies in non-specific industries. Therefore, the industry-specific characteristics of companies are not considered. In this paper, we confirm that adaptive boosting (AdaBoost) is the most appropriate forecasting model for construction companies by based on company size. We classified construction companies into three groups - large, medium, and small based on the company's capital. We analyzed the predictive ability of AdaBoost for each group of companies. The experimental results showed that AdaBoost has more predictive ability than the other models, especially for the group of large companies with capital of more than 50 billion won.

Analysis on Factors Influencing Welfare Spending of Local Authority : Implementing the Detailed Data Extracted from the Social Security Information System (지방자치단체 자체 복지사업 지출 영향요인 분석 : 사회보장정보시스템을 통한 접근)

  • Kim, Kyoung-June;Ham, Young-Jin;Lee, Ki-Dong
    • Journal of Intelligence and Information Systems
    • /
    • v.19 no.2
    • /
    • pp.141-156
    • /
    • 2013
  • Researchers in welfare services of local government in Korea have rather been on isolated issues as disables, childcare, aging phenomenon, etc. (Kang, 2004; Jung et al., 2009). Lately, local officials, yet, realize that they need more comprehensive welfare services for all residents, not just for above-mentioned focused groups. Still cases dealt with focused group approach have been a main research stream due to various reason(Jung et al., 2009; Lee, 2009; Jang, 2011). Social Security Information System is an information system that comprehensively manages 292 welfare benefits provided by 17 ministries and 40 thousand welfare services provided by 230 local authorities in Korea. The purpose of the system is to improve efficiency of social welfare delivery process. The study of local government expenditure has been on the rise over the last few decades after the restarting the local autonomy, but these studies have limitations on data collection. Measurement of a local government's welfare efforts(spending) has been primarily on expenditures or budget for an individual, set aside for welfare. This practice of using monetary value for an individual as a "proxy value" for welfare effort(spending) is based on the assumption that expenditure is directly linked to welfare efforts(Lee et al., 2007). This expenditure/budget approach commonly uses total welfare amount or percentage figure as dependent variables (Wildavsky, 1985; Lee et al., 2007; Kang, 2000). However, current practice of using actual amount being used or percentage figure as a dependent variable may have some limitation; since budget or expenditure is greatly influenced by the total budget of a local government, relying on such monetary value may create inflate or deflate the true "welfare effort" (Jang, 2012). In addition, government budget usually contain a large amount of administrative cost, i.e., salary, for local officials, which is highly unrelated to the actual welfare expenditure (Jang, 2011). This paper used local government welfare service data from the detailed data sets linked to the Social Security Information System. The purpose of this paper is to analyze the factors that affect social welfare spending of 230 local authorities in 2012. The paper applied multiple regression based model to analyze the pooled financial data from the system. Based on the regression analysis, the following factors affecting self-funded welfare spending were identified. In our research model, we use the welfare budget/total budget(%) of a local government as a true measurement for a local government's welfare effort(spending). Doing so, we exclude central government subsidies or support being used for local welfare service. It is because central government welfare support does not truly reflect the welfare efforts(spending) of a local. The dependent variable of this paper is the volume of the welfare spending and the independent variables of the model are comprised of three categories, in terms of socio-demographic perspectives, the local economy and the financial capacity of local government. This paper categorized local authorities into 3 groups, districts, and cities and suburb areas. The model used a dummy variable as the control variable (local political factor). This paper demonstrated that the volume of the welfare spending for the welfare services is commonly influenced by the ratio of welfare budget to total local budget, the population of infants, self-reliance ratio and the level of unemployment factor. Interestingly, the influential factors are different by the size of local government. Analysis of determinants of local government self-welfare spending, we found a significant effect of local Gov. Finance characteristic in degree of the local government's financial independence, financial independence rate, rate of social welfare budget, and regional economic in opening-to-application ratio, and sociology of population in rate of infants. The result means that local authorities should have differentiated welfare strategies according to their conditions and circumstances. There is a meaning that this paper has successfully proven the significant factors influencing welfare spending of local government in Korea.

Trends of Study and Classification of Reference on Occupational Health Management in Korea after Liberation (해방 이후 우리나라 산업보건관리에 관한 문헌분류 및 연구동향)

  • Ha, Eun-Hee;Park, Hye-Sook;Kim, Young-Bok;Song, Hyun-Jong
    • Journal of Preventive Medicine and Public Health
    • /
    • v.28 no.4 s.51
    • /
    • pp.809-844
    • /
    • 1995
  • The purposes of this study are to define the scope of occupational health management and to classify occupational management by review of related journals from 1945 to 1994 in Korea. The steps of this study were as follows: (1) Search of secondary reference; (2) Collection and review of primary reference; (3) Survey; and (4) Analysis and discussion. The results were as follows ; 1. Most of the respondents majored in occupational health(71.6%), and were working in university (68.3%), males and over the age 40. Seventy percent of the respondents agreed with the idea that classification of occupational health management is necessary, and 10% disagreed. 2. After integration of the idea of respondents, we reclassified the scope of occupational health management. It was defined 3 parts, that is , occupational health system, occupational health service and others (such as assessment, epidemiology, cost-effectiveness analysis and so on). 3. The number of journals on occupational health management was 510. It was sightly increased from 1986 and abruptly increased after 1991. The kinds of journals related to occupational health management were The Korean Journal of Occupational Medicine(18.2%), Several Kinds of Medical Colloge Journal(17.0%), The Korean Journal Occupational Health(15.1%), The Korean Journal of Preventive Medicine(15.1%) and others(34.6%). As for the contents, the number of journals on occupational health management systems was 33(6.5%) and occupational health services 477(93.5%). Of the journals on occupational health management systems, the number of journals on the occupational health resource system was 15(45.5%), occupational finance system 8(24.2%), occupational health management system 6(18.2%), occupational organization 3(9.1%) and occupational health delivery system 1 (3.0%). Of the journals on occupational health services, the number of journals on disease management was 269(57.2%), health management 116(24.7%), working environmental management 85(18.1%). As for the subjects, the number of journals on general workers was 185(71.1%), followed by women worker, white coiler workers and so on. 4. Respondents made occupational health service(such as health management, working environmental management and health education) the first priority of occupational health management. Tied for the second are quality analysis(such as education, training and job contents of occupational health manager) and occupational health systems(such as the recommendation of systems of occupational and general disease and occupational health organization). 5. Thirty seven respondents suggested 48 ideas about the future research of occupational health management. The results were as follows: (1) Study of occupational health service 40.5%; (2) Study of organization system 27.1%; (3) Study of occupational health system (e.g. information network) 8.3%; (4) Study of working condition 6.2%; and (5) Study of occupational health service analysis 4.2%.

  • PDF

A Study on Knowledge Entity Extraction Method for Individual Stocks Based on Neural Tensor Network (뉴럴 텐서 네트워크 기반 주식 개별종목 지식개체명 추출 방법에 관한 연구)

  • Yang, Yunseok;Lee, Hyun Jun;Oh, Kyong Joo
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.2
    • /
    • pp.25-38
    • /
    • 2019
  • Selecting high-quality information that meets the interests and needs of users among the overflowing contents is becoming more important as the generation continues. In the flood of information, efforts to reflect the intention of the user in the search result better are being tried, rather than recognizing the information request as a simple string. Also, large IT companies such as Google and Microsoft focus on developing knowledge-based technologies including search engines which provide users with satisfaction and convenience. Especially, the finance is one of the fields expected to have the usefulness and potential of text data analysis because it's constantly generating new information, and the earlier the information is, the more valuable it is. Automatic knowledge extraction can be effective in areas where information flow is vast, such as financial sector, and new information continues to emerge. However, there are several practical difficulties faced by automatic knowledge extraction. First, there are difficulties in making corpus from different fields with same algorithm, and it is difficult to extract good quality triple. Second, it becomes more difficult to produce labeled text data by people if the extent and scope of knowledge increases and patterns are constantly updated. Third, performance evaluation is difficult due to the characteristics of unsupervised learning. Finally, problem definition for automatic knowledge extraction is not easy because of ambiguous conceptual characteristics of knowledge. So, in order to overcome limits described above and improve the semantic performance of stock-related information searching, this study attempts to extract the knowledge entity by using neural tensor network and evaluate the performance of them. Different from other references, the purpose of this study is to extract knowledge entity which is related to individual stock items. Various but relatively simple data processing methods are applied in the presented model to solve the problems of previous researches and to enhance the effectiveness of the model. From these processes, this study has the following three significances. First, A practical and simple automatic knowledge extraction method that can be applied. Second, the possibility of performance evaluation is presented through simple problem definition. Finally, the expressiveness of the knowledge increased by generating input data on a sentence basis without complex morphological analysis. The results of the empirical analysis and objective performance evaluation method are also presented. The empirical study to confirm the usefulness of the presented model, experts' reports about individual 30 stocks which are top 30 items based on frequency of publication from May 30, 2017 to May 21, 2018 are used. the total number of reports are 5,600, and 3,074 reports, which accounts about 55% of the total, is designated as a training set, and other 45% of reports are designated as a testing set. Before constructing the model, all reports of a training set are classified by stocks, and their entities are extracted using named entity recognition tool which is the KKMA. for each stocks, top 100 entities based on appearance frequency are selected, and become vectorized using one-hot encoding. After that, by using neural tensor network, the same number of score functions as stocks are trained. Thus, if a new entity from a testing set appears, we can try to calculate the score by putting it into every single score function, and the stock of the function with the highest score is predicted as the related item with the entity. To evaluate presented models, we confirm prediction power and determining whether the score functions are well constructed by calculating hit ratio for all reports of testing set. As a result of the empirical study, the presented model shows 69.3% hit accuracy for testing set which consists of 2,526 reports. this hit ratio is meaningfully high despite of some constraints for conducting research. Looking at the prediction performance of the model for each stocks, only 3 stocks, which are LG ELECTRONICS, KiaMtr, and Mando, show extremely low performance than average. this result maybe due to the interference effect with other similar items and generation of new knowledge. In this paper, we propose a methodology to find out key entities or their combinations which are necessary to search related information in accordance with the user's investment intention. Graph data is generated by using only the named entity recognition tool and applied to the neural tensor network without learning corpus or word vectors for the field. From the empirical test, we confirm the effectiveness of the presented model as described above. However, there also exist some limits and things to complement. Representatively, the phenomenon that the model performance is especially bad for only some stocks shows the need for further researches. Finally, through the empirical study, we confirmed that the learning method presented in this study can be used for the purpose of matching the new text information semantically with the related stocks.

Color-related Query Processing for Intelligent E-Commerce Search (지능형 검색엔진을 위한 색상 질의 처리 방안)

  • Hong, Jung A;Koo, Kyo Jung;Cha, Ji Won;Seo, Ah Jeong;Yeo, Un Yeong;Kim, Jong Woo
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.1
    • /
    • pp.109-125
    • /
    • 2019
  • As interest on intelligent search engines increases, various studies have been conducted to extract and utilize the features related to products intelligencely. In particular, when users search for goods in e-commerce search engines, the 'color' of a product is an important feature that describes the product. Therefore, it is necessary to deal with the synonyms of color terms in order to produce accurate results to user's color-related queries. Previous studies have suggested dictionary-based approach to process synonyms for color features. However, the dictionary-based approach has a limitation that it cannot handle unregistered color-related terms in user queries. In order to overcome the limitation of the conventional methods, this research proposes a model which extracts RGB values from an internet search engine in real time, and outputs similar color names based on designated color information. At first, a color term dictionary was constructed which includes color names and R, G, B values of each color from Korean color standard digital palette program and the Wikipedia color list for the basic color search. The dictionary has been made more robust by adding 138 color names converted from English color names to foreign words in Korean, and with corresponding RGB values. Therefore, the fininal color dictionary includes a total of 671 color names and corresponding RGB values. The method proposed in this research starts by searching for a specific color which a user searched for. Then, the presence of the searched color in the built-in color dictionary is checked. If there exists the color in the dictionary, the RGB values of the color in the dictioanry are used as reference values of the retrieved color. If the searched color does not exist in the dictionary, the top-5 Google image search results of the searched color are crawled and average RGB values are extracted in certain middle area of each image. To extract the RGB values in images, a variety of different ways was attempted since there are limits to simply obtain the average of the RGB values of the center area of images. As a result, clustering RGB values in image's certain area and making average value of the cluster with the highest density as the reference values showed the best performance. Based on the reference RGB values of the searched color, the RGB values of all the colors in the color dictionary constructed aforetime are compared. Then a color list is created with colors within the range of ${\pm}50$ for each R value, G value, and B value. Finally, using the Euclidean distance between the above results and the reference RGB values of the searched color, the color with the highest similarity from up to five colors becomes the final outcome. In order to evaluate the usefulness of the proposed method, we performed an experiment. In the experiment, 300 color names and corresponding color RGB values by the questionnaires were obtained. They are used to compare the RGB values obtained from four different methods including the proposed method. The average euclidean distance of CIE-Lab using our method was about 13.85, which showed a relatively low distance compared to 3088 for the case using synonym dictionary only and 30.38 for the case using the dictionary with Korean synonym website WordNet. The case which didn't use clustering method of the proposed method showed 13.88 of average euclidean distance, which implies the DBSCAN clustering of the proposed method can reduce the Euclidean distance. This research suggests a new color synonym processing method based on RGB values that combines the dictionary method with the real time synonym processing method for new color names. This method enables to get rid of the limit of the dictionary-based approach which is a conventional synonym processing method. This research can contribute to improve the intelligence of e-commerce search systems especially on the color searching feature.

Self-Regulatory Mode Effects on Emotion and Customer's Response in Failed Services - Focusing on the moderate effect of attribution processing - (고객의 자기조절성향이 서비스 실패에 따른 부정적 감정과 고객반응에 미치는 영향 - 귀인과정에 따른 조정적 역할을 중심으로 -)

  • Sung, Hyung-Suk;Han, Sang-Lin
    • Asia Marketing Journal
    • /
    • v.12 no.2
    • /
    • pp.83-110
    • /
    • 2010
  • Dissatisfied customers may express their dissatisfaction behaviorally. These behavioral responses may impact the firms' profitability. How do we model the impact of self regulatory orientation on emotions and subsequent customer behaviors? Obviously, the positive and negative emotions experienced in these situations will influence the overall degree of satisfaction or dissatisfaction with the service(Zeelenberg and Pieters 1999). Most likely, these specific emotions will also partly determine the subsequent behavior in relation to the service and service provider, such as the likelihood of complaining, the degree to which customers will switch or repurchase, and the extent of word of mouth communication they will engage in(Zeelenberg and Pieters 2004). This study investigates the antecedents, consequences of negative consumption emotion and the moderate effect of attribution processing in an integrated model(self regulatory mode → specific emotions → behavioral responses). We focused on the fact that regret and disappointment have effects on consumer behavior. Especially, There are essentially two approaches in this research: the valence based approach and the specific emotions approach. The authors indicate theoretically and show empirically that it matters to distinguish these approaches in services research. and The present studies examined the influence of two regulatory mode concerns(Locomotion orientation and Assessment orientation) with making comparisons on experiencing post decisional regret and disappointment(Pierro, Kruglanski, and Higgins 2006; Pierro et al. 2008). When contemplating a decision with a negative outcome, it was predicted that high (vs low) locomotion would induce more disappointment than regret, whereas high (vs low) assessment would induce more regret than disappointment. The validity of the measurement scales was also confirmed by evaluations provided by the participating respondents and an independent advisory panel; samples provided recommendations throughout the primary, exploratory phases of the study. The resulting goodness of fit statistics were RMR or RMSEA of 0.05, GFI and AGFI greater than 0.9, and a chi-square with a 175.11. The indicators of the each constructs were very good measures of variables and had high convergent validity as evidenced by the reliability with a more than 0.9. Some items were deleted leaving those that reflected the cognitive dimension of importance rather than the dimension. The indicators were very good measures and had convergent validity as evidenced by the reliability of 0.9. These results for all constructs indicate the measurement fits the sample data well and is adequate for use. The scale for each factor was set by fixing the factor loading to one of its indicator variables and then applying the maximum likelihood estimation method. The results of the analysis showed that directions of the effects in the model are ultimately supported by the theory underpinning the causal linkages of the model. This research proposed 6 hypotheses on 6 latent variables and tested through structural equation modeling. 6 alternative measurements were compared through statistical significance test of the paths of research model and the overall fitting level of structural equation model and the result was successful. Also, Locomotion orientation more positively influences disappointment when internal attribution is high than low and Assessment orientation more positively influences regret when external attribution is high than low. In sum, The results of our studies suggest that assessment and locomotion concerns, both as chronic individual predispositions and as situationally induced states, influence the amount of people's experienced regret and disappointment. These findings contribute to our understanding of regulatory mode, regret, and disappointment. In previous studies of regulatory mode, relatively little attention has been paid to the post actional evaluative phase of self regulation. The present findings indicate that assessment concerns and locomotion concerns are clearly distinct in this phase, with individuals higher in assessment delving more into possible alternatives to past actions and individuals higher in locomotion engaging less in such reflective thought. What this suggests is that, separate from decreasing the amount of counterfactual thinking per se, individuals with locomotion concerns want to move on, to get on with it. Regret is about the past and not the future. Thus, individuals with locomotion concerns are less likely to experience regret. The results supported our predictions. We discuss the implications of these findings for the nature of regret and disappointment from the perspective of their relation to regulatory mode. Also, self regulatory mode and the specific emotions(disappointment and regret) were assessed and their influence on customers' behavioral responses(inaction, word of mouth) was examined, using a sample of 275 customers. It was found that emotions have a direct impact on behavior over and above the effects of negative emotions and customer behavior. Hence, We argue against incorporating emotions such as regret and disappointment into a specific response measure and in favor of a specific emotions approach on self regulation. Implications for services marketing practice and theory are discussed.

  • PDF