• Title/Summary/Keyword: Initial data

Search Result 4,975, Processing Time 0.045 seconds

Overview of Research Trends in Estimation of Forest Carbon Stocks Based on Remote Sensing and GIS (원격탐사와 GIS 기반의 산림탄소저장량 추정에 관한 주요국 연구동향 개관)

  • Kim, Kyoung-Min;Lee, Jung-Bin;Kim, Eun-Sook;Park, Hyun-Ju;Roh, Young-Hee;Lee, Seung-Ho;Park, Key-Ho;Shin, Hyu-Seok
    • Journal of the Korean Association of Geographic Information Studies
    • /
    • v.14 no.3
    • /
    • pp.236-256
    • /
    • 2011
  • Forest carbon stocks change due to land use change is an important data required by UNFCCC(United Nations framework convention on climate change). Spatially explicit estimation of forest carbon stocks based on IPCC GPG(intergovernmental panel on climate change good practice guidance) tier 3 gives high reliability. But a current estimation which was aggregated from NFI data doesn't have detail forest carbon stocks by polygon or cell. In order to improve an estimation remote sensing and GIS have been used especially in Europe and North America. We divided research trends in main countries into 4 categories such as remote sensing, GIS, geostatistics and environmental modeling considering spatial heterogeneity. The easiest way to apply is combination NFI data with forest type map based on GIS. Considering especially complicated forest structure of Korea, geostatistics is useful to estimate local variation of forest carbon. In addition, fine scale image is good for verification of forest carbon stocks and determination of CDM site. Related domestic researches are still on initial status and forest carbon stocks are mainly estimated using k-nearest neighbor(k-NN). In order to select suitable method for forest in Korea, an applicability of diverse spatial data and algorithm must be considered. Also the comparison between methods is required.

Preliminary Study on the Application of Remote Sensing to Mineral Exploration Using Landsat and ASTER Data (Landsat과 ASTER 위성영상 자료를 이용한 광물자원탐사로의 적용 가능성을 위한 예비연구)

  • Lee, Hong-Jin;Park, Maeng-Eon;Kim, Eui-Jun
    • Economic and Environmental Geology
    • /
    • v.43 no.5
    • /
    • pp.467-475
    • /
    • 2010
  • The Landsat and ASTER data have been used in mineralogical and lithological studies, and they have also proved to be useful tool in the initial steps for mineral exploration throughout Nevada mining district, US. Huge pyrophyllite quarry mines, including Jungang, Samsung, Kyeongju, and Naenam located in the southeastern part of Gyeongsang Basin. The geology of study area consists mainly of Cretaceous volcanic rocks, which belong into Cretaceous Hayang and Jindong Group. They were intruded by Bulgugsa granites, so called Sannae-Eonyang granites. To extraction of Ratio model for pyrophyllite deposits, tuffaceous rock and pyrophyllite ores from the Jungang mine used in reflectance spectral analysis and these results were re-sampled to Landsat and ASTER bandpass. As a result of these processes, the pyrophyllite ores spectral features show strong reflectance at band 5, whereas strong absorption at band 7 in Landsat data. In the ASTER data, the pyrophyllite ores spectral features show strong absorption at band 5 and 8, whereas strong reflectance at band 4 and 7. Based on these spectral features, as a result of application of $Py_{Landsat}$ model to hydrothermal alteration zone and other exposed sites, the DN values of two different areas are 1.94 and 1.19 to 1.49, respectively. The differences values between pyrophyllite deposits and concrete-barren area are 0.472 and 0.399 for $Py_{ASTER}$ model, 0.452 and 0.371 for OHIb model, 0.365 and 0.311 for PAK model, respectively. Thus, $Py_{ASTER}$ and $Py_{Landsat}$ model proposed from this study proved to be more useful tool for the extraction of pyrophyllite deposits relative to previous models.

Applicability of Theoretical Adsorption Models for Studies on Adsorption Properties of Adsorbents(III) (흡착제의 흡착특성 규명을 위한 흡착모델의 적용성 평가(III) - 열역학적 특성을 중심으로)

  • Na, Choon-Ki;Jeong, Jin-Hwa;Park, Hyun-Ju
    • Journal of Korean Society of Environmental Engineers
    • /
    • v.34 no.4
    • /
    • pp.260-269
    • /
    • 2012
  • The aim of this study is to evaluate the applicability of adsorption models for understanding the thermodynamic properties of adsorption process. For this study, the adsorption isotherm data of $NO_3$-N ion onto a commercial anion exchange resin obtained at various experimental conditions, i.e. different initial concentrations of adsorbate, different dosages of adsorbent, and different temperatures, were used in calculating the thermodynamic parameters and the adsorption energy of adsorption process. The Gibbs free energy change (${\Delta}G^0$) of adsorption process could be calculated using the Langmuir constant $b_M$ as well as the Sips constant, even though the results were significantly dependant on the experimental conditions. The thermodynamic parameters such as standard enthalpy change (${\Delta}H^0$), standard entropy change (${\Delta}S^0$) and ${\Delta}G^0$ could be calculated by using the experimental data obtained at different temperatures, if the adsorption data well fitted to the Langmuir isotherm model and the plot of ln b versus 1/T gives a straight line. As an alternative, the empirical equilibrium constant(K) defined as $q_e/C_e$ could be used for evaluating the thermodynamic parameters instead of the Langmuir constant. The results from the applications of D-R model and Temkin model to evaluate the adsorption energy suggest that the D-R model is better than Temkin model for describing the experimental data, and the availability of Temkin model is highly limited by the experimental conditions. Although adsorption energies determined using D-R model show significantly different values depending on the experimental conditions, they were sufficient to show that the adsorption of $NO_3$-N onto anion exchange resin is an endothermic process and an ion-exchange process.

Koreanized Analysis System Development for Groundwater Flow Interpretation (지하수유동해석을 위한 한국형 분석시스템의 개발)

  • Choi, Yun-Yeong
    • Journal of the Korean Society of Hazard Mitigation
    • /
    • v.3 no.3 s.10
    • /
    • pp.151-163
    • /
    • 2003
  • In this study, the algorithm of groundwater flow process was established for koreanized groundwater program development dealing with the geographic and geologic conditions of the aquifer have dynamic behaviour in groundwater flow system. All the input data settings of the 3-DFM model which is developed in this study are organized in Korean, and the model contains help function for each input data. Thus, it is designed to get detailed information about each input parameter when the mouse pointer is placed on the corresponding input parameter. This model also is designed to easily specify the geologic boundary condition for each stratum or initial head data in the work sheet. In addition, this model is designed to display boxes for input parameter writing for each analysis condition so that the setting for each parameter is not so complicated as existing MODFLOW is when steady and unsteady flow analysis are performed as well as the analysis for the characteristics of each stratum. Descriptions for input data are displayed on the right side of the window while the analysis results are displayed on the left side as well as the TXT file for this results is available to see. The model developed in this study is a numerical model using finite differential method, and the applicability of the model was examined by comparing and analyzing observed and simulated groundwater heads computed by the application of real recharge amount and the estimation of parameters. The 3-DFM model is applied in this study to Sehwa-ri, and Songdang-ri area, Jeju, Korea for analysis of groundwater flow system according to pumping, and obtained the results that the observed and computed groundwater head were almost in accordance with each other showing the range of 0.03 - 0.07 error percent. It is analyzed that the groundwater flow distributed evenly from Nopen-orum and Munseogi-orum to Wolang-bong, Yongnuni-orum, and Songja-bong through the computation of equipotentials and velocity vector using the analysis result of simulation which was performed before the pumping started in the study area. These analysis results show the accordance with MODFLOW's.

The Adoption and Diffusion of Semantic Web Technology Innovation: Qualitative Research Approach (시맨틱 웹 기술혁신의 채택과 확산: 질적연구접근법)

  • Joo, Jae-Hun
    • Asia pacific journal of information systems
    • /
    • v.19 no.1
    • /
    • pp.33-62
    • /
    • 2009
  • Internet computing is a disruptive IT innovation. Semantic Web can be considered as an IT innovation because the Semantic Web technology possesses the potential to reduce information overload and enable semantic integration, using capabilities such as semantics and machine-processability. How should organizations adopt the Semantic Web? What factors affect the adoption and diffusion of Semantic Web innovation? Most studies on adoption and diffusion of innovation use empirical analysis as a quantitative research methodology in the post-implementation stage. There is criticism that the positivist requiring theoretical rigor can sacrifice relevance to practice. Rapid advances in technology require studies relevant to practice. In particular, it is realistically impossible to conduct quantitative approach for factors affecting adoption of the Semantic Web because the Semantic Web is in its infancy. However, in an early stage of introduction of the Semantic Web, it is necessary to give a model and some guidelines and for adoption and diffusion of the technology innovation to practitioners and researchers. Thus, the purpose of this study is to present a model of adoption and diffusion of the Semantic Web and to offer propositions as guidelines for successful adoption through a qualitative research method including multiple case studies and in-depth interviews. The researcher conducted interviews with 15 people based on face-to face and 2 interviews by telephone and e-mail to collect data to saturate the categories. Nine interviews including 2 telephone interviews were from nine user organizations adopting the technology innovation and the others were from three supply organizations. Semi-structured interviews were used to collect data. The interviews were recorded on digital voice recorder memory and subsequently transcribed verbatim. 196 pages of transcripts were obtained from about 12 hours interviews. Triangulation of evidence was achieved by examining each organization website and various documents, such as brochures and white papers. The researcher read the transcripts several times and underlined core words, phrases, or sentences. Then, data analysis used the procedure of open coding, in which the researcher forms initial categories of information about the phenomenon being studied by segmenting information. QSR NVivo version 8.0 was used to categorize sentences including similar concepts. 47 categories derived from interview data were grouped into 21 categories from which six factors were named. Five factors affecting adoption of the Semantic Web were identified. The first factor is demand pull including requirements for improving search and integration services of the existing systems and for creating new services. Second, environmental conduciveness, reference models, uncertainty, technology maturity, potential business value, government sponsorship programs, promising prospects for technology demand, complexity and trialability affect the adoption of the Semantic Web from the perspective of technology push. Third, absorptive capacity is an important role of the adoption. Fourth, suppler's competence includes communication with and training for users, and absorptive capacity of supply organization. Fifth, over-expectance which results in the gap between user's expectation level and perceived benefits has a negative impact on the adoption of the Semantic Web. Finally, the factor including critical mass of ontology, budget. visible effects is identified as a determinant affecting routinization and infusion. The researcher suggested a model of adoption and diffusion of the Semantic Web, representing relationships between six factors and adoption/diffusion as dependent variables. Six propositions are derived from the adoption/diffusion model to offer some guidelines to practitioners and a research model to further studies. Proposition 1 : Demand pull has an influence on the adoption of the Semantic Web. Proposition 1-1 : The stronger the degree of requirements for improving existing services, the more successfully the Semantic Web is adopted. Proposition 1-2 : The stronger the degree of requirements for new services, the more successfully the Semantic Web is adopted. Proposition 2 : Technology push has an influence on the adoption of the Semantic Web. Proposition 2-1 : From the perceptive of user organizations, the technology push forces such as environmental conduciveness, reference models, potential business value, and government sponsorship programs have a positive impact on the adoption of the Semantic Web while uncertainty and lower technology maturity have a negative impact on its adoption. Proposition 2-2 : From the perceptive of suppliers, the technology push forces such as environmental conduciveness, reference models, potential business value, government sponsorship programs, and promising prospects for technology demand have a positive impact on the adoption of the Semantic Web while uncertainty, lower technology maturity, complexity and lower trialability have a negative impact on its adoption. Proposition 3 : The absorptive capacities such as organizational formal support systems, officer's or manager's competency analyzing technology characteristics, their passion or willingness, and top management support are positively associated with successful adoption of the Semantic Web innovation from the perceptive of user organizations. Proposition 4 : Supplier's competence has a positive impact on the absorptive capacities of user organizations and technology push forces. Proposition 5 : The greater the gap of expectation between users and suppliers, the later the Semantic Web is adopted. Proposition 6 : The post-adoption activities such as budget allocation, reaching critical mass, and sharing ontology to offer sustainable services are positively associated with successful routinization and infusion of the Semantic Web innovation from the perceptive of user organizations.

Multiple SL-AVS(Small size & Low power Around View System) Synchronization Maintenance Method (다중 SL-AVS 동기화 유지기법)

  • Park, Hyun-Moon;Park, Soo-Huyn;Seo, Hae-Moon;Park, Woo-Chool
    • Journal of the Korea Society for Simulation
    • /
    • v.18 no.3
    • /
    • pp.73-82
    • /
    • 2009
  • Due to the many advantages including low price, low power consumption, and miniaturization, the CMOS camera has been utilized in many applications, including mobile phones, the automotive industry, medical sciences and sensoring, robotic controls, and research in the security field. In particular, the 360 degree omni-directional camera when utilized in multi-camera applications has displayed issues of software nature, interface communication management, delays, and a complicated image display control. Other issues include energy management problems, and miniaturization of a multi-camera in the hardware field. Traditional CMOS camera systems are comprised of an embedded system that consists of a high-performance MCU enabling a camera to send and receive images and a multi-layer system similar to an individual control system that consists of the camera's high performance Micro Controller Unit. We proposed the SL-AVS (Small Size/Low power Around-View System) to be able to control a camera while collecting image data using a high speed synchronization technique on the foundation of a single layer low performance MCU. It is an initial model of the omni-directional camera that takes images from a 360 view drawing from several CMOS camera utilizing a 110 degree view. We then connected a single MCU with four low-power CMOS cameras and implemented controls that include synchronization, controlling, and transmit/receive functions of individual camera compared with the traditional system. The synchronization of the respective cameras were controlled and then memorized by handling each interrupt through the MCU. We were able to improve the efficiency of data transmission that minimizes re-synchronization amongst a target, the CMOS camera, and the MCU. Further, depending on the choice of users, respective or groups of images divided into 4 domains were then provided with a target. We finally analyzed and compared the performance of the developed camera system including the synchronization and time of data transfer and image data loss, etc.

Current feeding practices and maternal nutritional knowledge on complementary feeding in Korea (이유기 보충식 현황과 어머니 인식 조사)

  • Yom, Hye Won;Seo, Jeong Wan;Park, Hyesook;Choi, Kwang Hae;Chang, Ju Young;Ryoo, Eell;Yang, Hye Ran;Kim, Jae Young;Seo, Ji Hyun;Kim, Yong Joo;Moon, Kyung Rye;Kang, Ki Soo;Park, Kie Young;Lee, Seong Soo;Shim, Jeong Ok
    • Clinical and Experimental Pediatrics
    • /
    • v.52 no.10
    • /
    • pp.1090-1102
    • /
    • 2009
  • Purpose:To evaluate current feeding practices and maternal nutritional knowledge on complementary feeding. Methods:Mothers of babies aged 9-15 months who visited pediatric clinics of 14 general hospitals between September and December 2008 were asked to fill questionnaires. Data from 1,078 questionnaires were analyzed. Results:Complementary food was introduced at 4-7 months in 89% of babies. Home-made rice gruel was the first complementary food in 93% cases. Spoons were used for initial feeding in 97% cases. At 6-7 months, <50% of babies were fed meat (beef, 43%). Less than 12-month-old babies were fed salty foods such as salted laver (35%) or bean-paste soup (51%) and cow's milk (11%). The following were the maternal sources of information on complementary feeding: books/magazines (58%), friends (30%), internet web sites (29%), relatives (14%), and hospitals (4%). Compared to the 1993 survey, the incidence of complementary food introduction before 4 months (0.4% vs. 21%) and initial use of commercial food (7% vs. 39%) had decreased. Moreover, spoons were increasingly used for initial feeding (97% vs. 57%). The average maternal nutritional knowledge score was 7.5/10. Less percentage of mothers agreed with the following suggestions: bottle formula weaning before 15-18 months (68%), no commercial baby drinks as complementary food (67%), considering formula (or cow's milk) better than soy milk (65%), and feeding minced meat from 6-7 months (57%). Conclusion:Complementary feeding practices have considerably improved since the last decade. Pediatricians should advise timely introduction of appropriate complementary foods and monitor diverse information sources on complementary feeding.

Derivation of the Synthetic Unit Hydrograph Based on the Watershed Characteristics (유역특성에 의한 합성단위도의 유도에 관한 연구)

  • 서승덕
    • Magazine of the Korean Society of Agricultural Engineers
    • /
    • v.17 no.1
    • /
    • pp.3642-3654
    • /
    • 1975
  • The purpose of this thesis is to derive a unit hydrograph which may be applied to the ungaged watershed area from the relations between directly measurable unitgraph properties such as peak discharge(qp), time to peak discharge (Tp), and lag time (Lg) and watershed characteristics such as river length(L) from the given station to the upstream limits of the watershed area in km, river length from station to centroid of gravity of the watershed area in km (Lca), and main stream slope in meter per km (S). Other procedure based on routing a time-area diagram through catchment storage named Instantaneous Unit Hydrograph(IUH). Dimensionless unitgraph also analysed in brief. The basic data (1969 to 1973) used in these studies are 9 recording level gages and rating curves, 41 rain gages and pluviographs, and 40 observed unitgraphs through the 9 sub watersheds in Nak Oong River basin. The results summarized in these studies are as follows; 1. Time in hour from start of rise to peak rate (Tp) generally occured at the position of 0.3Tb (time base of hydrograph) with some indication of higher values for larger watershed. The base flow is comparelatively higher than the other small watershed area. 2. Te losses from rainfall were divided into initial loss and continuing loss. Initial loss may be defined as that portion of storm rainfall which is intercepted by vegetation, held in deppression storage or infiltrated at a high rate early in the storm and continuing loss is defined as the loss which continues at a constant rate throughout the duration of the storm after the initial loss has been satisfied. Tis continuing loss approximates the nearly constant rate of infiltration (${\Phi}$-index method). The loss rate from this analysis was estimated 50 Per cent to the rainfall excess approximately during the surface runoff occured. 3. Stream slope seems approximate, as is usual, to consider the mainstreamonly, not giving any specific consideration to tributary. It is desirable to develop a single measure of slope that is representative of the who1e stream. The mean slope of channel increment in 1 meter per 200 meters and 1 meter per 1400 meters were defined at Gazang and Jindong respectively. It is considered that the slopes are low slightly in the light of other river studies. Flood concentration rate might slightly be low in the Nak Dong river basin. 4. It found that the watershed lag (Lg, hrs) could be expressed by Lg=0.253 (L.Lca)0.4171 The product L.Lca is a measure of the size and shape of the watershed. For the logarithms, the correlation coefficient for Lg was 0.97 which defined that Lg is closely related with the watershed characteristics, L and Lca. 5. Expression for basin might be expected to take form containing theslope as {{{{ { L}_{g }=0.545 {( { L. { L}_{ca } } over { SQRT {s} } ) }^{0.346 } }}}} For the logarithms, the correlation coefficient for Lg was 0.97 which defined that Lg is closely related with the basin characteristics too. It should be needed to take care of analysis which relating to the mean slopes 6. Peak discharge per unit area of unitgraph for standard duration tr, ㎥/sec/$\textrm{km}^2$, was given by qp=10-0.52-0.0184Lg with a indication of lower values for watershed contrary to the higher lag time. For the logarithms, the correlation coefficient qp was 0.998 which defined high sign ificance. The peak discharge of the unitgraph for an area could therefore be expected to take the from Qp=qp. A(㎥/sec). 7. Using the unitgraph parameter Lg, the base length of the unitgraph, in days, was adopted as {{{{ {T}_{b } =0.73+2.073( { { L}_{g } } over {24 } )}}}} with high significant correlation coefficient, 0.92. The constant of the above equation are fixed by the procedure used to separate base flow from direct runoff. 8. The width W75 of the unitgraph at discharge equal to 75 per cent of the peak discharge, in hours and the width W50 at discharge equal to 50 Per cent of the peak discharge in hours, can be estimated from {{{{ { W}_{75 }= { 1.61} over { { q}_{b } ^{1.05 } } }}}} and {{{{ { W}_{50 }= { 2.5} over { { q}_{b } ^{1.05 } } }}}} respectively. This provides supplementary guide for sketching the unitgraph. 9. Above equations define the three factors necessary to construct the unitgraph for duration tr. For the duration tR, the lag is LgR=Lg+0.2(tR-tr) and this modified lag, LgRis used in qp and Tb It the tr happens to be equal to or close to tR, further assume qpR=qp. 10. Triangular hydrograph is a dimensionless unitgraph prepared from the 40 unitgraphs. The equation is shown as {{{{ { q}_{p } = { K.A.Q} over { { T}_{p } } }}}} or {{{{ { q}_{p } = { 0.21A.Q} over { { T}_{p } } }}}} The constant 0.21 is defined to Nak Dong River basin. 11. The base length of the time-area diagram for the IUH routing is {{{{C=0.9 {( { L. { L}_{ca } } over { SQRT { s} } ) }^{1/3 } }}}}. Correlation coefficient for C was 0.983 which defined a high significance. The base length of the T-AD was set to equal the time from the midpoint of rain fall excess to the point of contraflexure. The constant K, derived in this studies is K=8.32+0.0213 {{{{ { L} over { SQRT { s} } }}}} with correlation coefficient, 0.964. 12. In the light of the results analysed in these studies, average errors in the peak discharge of the Synthetic unitgraph, Triangular unitgraph, and IUH were estimated as 2.2, 7.7 and 6.4 per cent respectively to the peak of observed average unitgraph. Each ordinate of the Synthetic unitgraph was approached closely to the observed one.

  • PDF

APPROXIMATE ESTIMATION OF RECRUITMENT IN FISH POPULATION UTILIZING STOCK DENSITY AND CATCH (밀도지수와 어획량으로서 수산자원의 가입량을 근사적으로 추정하는 방법)

  • KIM Kee Ju
    • Korean Journal of Fisheries and Aquatic Sciences
    • /
    • v.8 no.2
    • /
    • pp.47-60
    • /
    • 1975
  • For the calculation of population parameter and estimation of recruitment of a fish population, an application of multiple regression method was used with some statistical inferences. Then, the differences between the calculated values and the true parameters were discussed. In addition, this method criticized by applying it to the statistical data of a population of bigeye tuna, Thunnus obesus of the Indian Ocean. The method was also applied to the available data of a population of Pacific saury, Cololabis saira, to estimate its recuitments. A stock at t year and t+1 year is, $N_{0,\;t+1}=N_{0,\;t}(1-m_t)-C_t+R_{t+1}$ where $N_0$ is the initial number of fish in a given year; C, number o: fish caught; R, number of recruitment; and M, rate of natural mortality. The foregoing equation is $$\phi_{t+1}=\frac{(1-\varrho^{-z}{t+1})Z_t}{(1-\varrho^{-z}t)Z_{t+1}}-\frac{1-\varrho^{-z}t+1}{Z_{t+1}}\phi_t-a'\frac{1-\varrho^{-z}t+1}{Z_{t+1}}C_t+a'\frac{1-\varrho^{-z}t+1}{Z_{t+1}}R_{t+1}......(1)$$ where $\phi$ is CPUE; a', CPUE $(\phi)$ to average stock $(\bar{N})$ in number; Z, total mortality coefficient; and M, natural mortality coefficient. In the equation (1) , the term $(1-\varrho^{-z}t+1)/Z_{t+1}$s almost constant to the variation of effort (X) there fore coefficients $\phi$ and $C_t$, can be calculated, when R is a constant, by applying the method of multiple regression, where $\phi_{t+1}$ is a dependent variable; $\phi_t$ and $C_t$ are independent variables. The values of Mand a' are calculated from the coefficients of $\phi_t$ and $C_t$; and total mortality coefficient (Z), where Z is a'X+M. By substituting M, a', $Z_t$, and $Z_{t+1}$ to the equation (1) recruitment $(R_{t+1})$ can be calculated. In this precess $\phi$ can be substituted by index of stock in number (N'). This operational procedures of the method of multiple regression can be applicable to the data which satisfy the above assumptions, even though the data were collected from any chosen year with similar recruitments, though it were not collected from the consecutive years. Under the condition of varying effort the data with such variation can be treated effectively by this method. The calculated values of M and a' include some deviation from the population parameters. Therefore, the estimated recruitment (R) is a relative value instead of all absolute one. This method of multiple regression is also applicable to the stock density and yield in weight instead of in number. For the data of the bigeye tuna of the Indian Ocean, the values of estimated recruitment (R) calculated from the parameter which is obtained by the present multiple regression method is proportional with an identical fluctuation pattern to the values of those derived from the parameters M and a', which were calculated by Suda (1970) for the same data. Estimated recruitments of Pacific saury of the eastern coast of Korea were calculated by the present multiple regression method. Not only spring recruitment $(1965\~1974)$ but also fall recruitment $(1964\~1973)$ was found to fluctuate in accordance with the fluctuations of stock densities (CPUE) of the same spring and fall, respectively.

  • PDF

Visualizing the Results of Opinion Mining from Social Media Contents: Case Study of a Noodle Company (소셜미디어 콘텐츠의 오피니언 마이닝결과 시각화: N라면 사례 분석 연구)

  • Kim, Yoosin;Kwon, Do Young;Jeong, Seung Ryul
    • Journal of Intelligence and Information Systems
    • /
    • v.20 no.4
    • /
    • pp.89-105
    • /
    • 2014
  • After emergence of Internet, social media with highly interactive Web 2.0 applications has provided very user friendly means for consumers and companies to communicate with each other. Users have routinely published contents involving their opinions and interests in social media such as blogs, forums, chatting rooms, and discussion boards, and the contents are released real-time in the Internet. For that reason, many researchers and marketers regard social media contents as the source of information for business analytics to develop business insights, and many studies have reported results on mining business intelligence from Social media content. In particular, opinion mining and sentiment analysis, as a technique to extract, classify, understand, and assess the opinions implicit in text contents, are frequently applied into social media content analysis because it emphasizes determining sentiment polarity and extracting authors' opinions. A number of frameworks, methods, techniques and tools have been presented by these researchers. However, we have found some weaknesses from their methods which are often technically complicated and are not sufficiently user-friendly for helping business decisions and planning. In this study, we attempted to formulate a more comprehensive and practical approach to conduct opinion mining with visual deliverables. First, we described the entire cycle of practical opinion mining using Social media content from the initial data gathering stage to the final presentation session. Our proposed approach to opinion mining consists of four phases: collecting, qualifying, analyzing, and visualizing. In the first phase, analysts have to choose target social media. Each target media requires different ways for analysts to gain access. There are open-API, searching tools, DB2DB interface, purchasing contents, and so son. Second phase is pre-processing to generate useful materials for meaningful analysis. If we do not remove garbage data, results of social media analysis will not provide meaningful and useful business insights. To clean social media data, natural language processing techniques should be applied. The next step is the opinion mining phase where the cleansed social media content set is to be analyzed. The qualified data set includes not only user-generated contents but also content identification information such as creation date, author name, user id, content id, hit counts, review or reply, favorite, etc. Depending on the purpose of the analysis, researchers or data analysts can select a suitable mining tool. Topic extraction and buzz analysis are usually related to market trends analysis, while sentiment analysis is utilized to conduct reputation analysis. There are also various applications, such as stock prediction, product recommendation, sales forecasting, and so on. The last phase is visualization and presentation of analysis results. The major focus and purpose of this phase are to explain results of analysis and help users to comprehend its meaning. Therefore, to the extent possible, deliverables from this phase should be made simple, clear and easy to understand, rather than complex and flashy. To illustrate our approach, we conducted a case study on a leading Korean instant noodle company. We targeted the leading company, NS Food, with 66.5% of market share; the firm has kept No. 1 position in the Korean "Ramen" business for several decades. We collected a total of 11,869 pieces of contents including blogs, forum contents and news articles. After collecting social media content data, we generated instant noodle business specific language resources for data manipulation and analysis using natural language processing. In addition, we tried to classify contents in more detail categories such as marketing features, environment, reputation, etc. In those phase, we used free ware software programs such as TM, KoNLP, ggplot2 and plyr packages in R project. As the result, we presented several useful visualization outputs like domain specific lexicons, volume and sentiment graphs, topic word cloud, heat maps, valence tree map, and other visualized images to provide vivid, full-colored examples using open library software packages of the R project. Business actors can quickly detect areas by a swift glance that are weak, strong, positive, negative, quiet or loud. Heat map is able to explain movement of sentiment or volume in categories and time matrix which shows density of color on time periods. Valence tree map, one of the most comprehensive and holistic visualization models, should be very helpful for analysts and decision makers to quickly understand the "big picture" business situation with a hierarchical structure since tree-map can present buzz volume and sentiment with a visualized result in a certain period. This case study offers real-world business insights from market sensing which would demonstrate to practical-minded business users how they can use these types of results for timely decision making in response to on-going changes in the market. We believe our approach can provide practical and reliable guide to opinion mining with visualized results that are immediately useful, not just in food industry but in other industries as well.