• Title/Summary/Keyword: Web data

Search Result 5,605, Processing Time 0.04 seconds

Analysis of Internet Biology Study Sites and Guidelines for Constructing Educational Homepages (인터넷상의 고등학교 생물 학습사이트 비교분석 및 웹사이트 구축방안에 관한 연구)

  • Kim, Joo-Hyun;Sung, Jung-Hee
    • Journal of The Korean Association For Science Education
    • /
    • v.22 no.4
    • /
    • pp.779-795
    • /
    • 2002
  • Internet, a world wide network of computers, is considered as a sea of information because it allows people to share information beyond the barriors of time and space. However, in spite of the unmeasurable potential applications of the internet, its use in the field of biology education has been extremely limited mainly due to the scarcity of good biology-related sites. In order to provide useful guidelines for constructing user-friendly study sites, which can help high school students with different intellectual levels to study biology, comparative studies were performed on selected educational sites. Initially, hundreds of related sites were examined, and, subsequently, four distinct sites were selected not only because they are well organized, but also because each is unique in its contents. Also, a survey was carried out against the users of each site. The survey results indicated that the high school students regard the web-based biology study tools as effective teaching methods although there might be some bias in criteria for selecting target sites. In addition to the detailed biology topics and the related biology informations, multimedia data including pictures, animations and movies are found to be one of the important ingredients for desirable biology study sites. Thus, the inclusion of multimedia components should also be considered when developing a systematic biology study site. Overall, the role of the cyber space is expected to become more and more important. Since the development of the user-satisfied and self-guided sites require interdisciplinary collaborational efforts which should be made to promote extensive communication among teachers, education professionals, and computer engineers. Furthermore, the introduction of good biology study sites to the students by their teachers is also important factor for the successful web-based education.

Optimal supervised LSA method using selective feature dimension reduction (선택적 자질 차원 축소를 이용한 최적의 지도적 LSA 방법)

  • Kim, Jung-Ho;Kim, Myung-Kyu;Cha, Myung-Hoon;In, Joo-Ho;Chae, Soo-Hoan
    • Science of Emotion and Sensibility
    • /
    • v.13 no.1
    • /
    • pp.47-60
    • /
    • 2010
  • Most of the researches about classification usually have used kNN(k-Nearest Neighbor), SVM(Support Vector Machine), which are known as learn-based model, and Bayesian classifier, NNA(Neural Network Algorithm), which are known as statistics-based methods. However, there are some limitations of space and time when classifying so many web pages in recent internet. Moreover, most studies of classification are using uni-gram feature representation which is not good to represent real meaning of words. In case of Korean web page classification, there are some problems because of korean words property that the words have multiple meanings(polysemy). For these reasons, LSA(Latent Semantic Analysis) is proposed to classify well in these environment(large data set and words' polysemy). LSA uses SVD(Singular Value Decomposition) which decomposes the original term-document matrix to three different matrices and reduces their dimension. From this SVD's work, it is possible to create new low-level semantic space for representing vectors, which can make classification efficient and analyze latent meaning of words or document(or web pages). Although LSA is good at classification, it has some drawbacks in classification. As SVD reduces dimensions of matrix and creates new semantic space, it doesn't consider which dimensions discriminate vectors well but it does consider which dimensions represent vectors well. It is a reason why LSA doesn't improve performance of classification as expectation. In this paper, we propose new LSA which selects optimal dimensions to discriminate and represent vectors well as minimizing drawbacks and improving performance. This method that we propose shows better and more stable performance than other LSAs' in low-dimension space. In addition, we derive more improvement in classification as creating and selecting features by reducing stopwords and weighting specific values to them statistically.

  • PDF

Consumers Perceptions on Monosodium L-glutamate in Social Media (소셜미디어 분석을 통한 소비자들의 L-글루타민산나트륨에 대한 인식 조사)

  • Lee, Sooyeon;Lee, Wonsung;Moon, Il-Chul;Kwon, Hoonjeong
    • Journal of Food Hygiene and Safety
    • /
    • v.31 no.3
    • /
    • pp.153-166
    • /
    • 2016
  • The purpose of this study was to investigate consumers' perceptions on monosodium L-glutamate (MSG) in social media. Data were collected from Naver blogs and Naver web communities (Korean representative portal web-site), and media reports including comment sections on a Yonhap news website (Korean largest news agency). The results from Naver blogs and Naver web communities showed that it was primarily mentioned MSG-use restaurant reviews, 'MSG-no added' products, its safety, and methods of reducing MSG in food. When TV shows on current affairs, newspaper, or TV news reported uses and side effects of MSG, search volume for MSG has increased in both PC and mobile search engines. Search volume has increased especially when TV shows on current affairs reported it. There are more periods with increased search volume for Mobile than PC. Also, it was mainly commented about safety of MSG, criticism of low-quality foods, abuse of MSG, and distrust of government below the news on the Yonhap news site. The label of MSG-no added products in market emphasized "MSG-free" even though it is allocated as an acceptable daily intake (ADI) not-specified by the Joint FAO/WHO Expert Committee on Food Additives (JECFA). When consumers search for MSG (monosodium L-glutamate) or purchase food on market, they might perceive that 'MSG-no added' products are better. Competent authorities, offices of education and local government provide guidelines based on no added MSG principle and these policies might affect consumers' perceptions. TV program or news program could be a powerful and effective consumer communication channel about MSG through Mobile rather than PC. Therefore media including TV should report item on monosodium L-glutamate with responsibility and information based on scientific background for consumers to get reliable information.

A Study on the Product Design Process in I-Business Environment Focusing on Development of the Internet-based Design Process - (e-비지니스환경에서의 제품디자인 프로세스에 관한 기초연구-인터넷기반의 디자인 프로세스 개발을 중심으로-)

  • 이수봉;이돈희
    • Archives of design research
    • /
    • v.16 no.1
    • /
    • pp.181-198
    • /
    • 2003
  • The purpose of this study is to develop a on-line design tool for effectively coping with e-Business environment, or product design process into a Cyber model for traditional manufacturers which attempts new product development under such environment. It was finally developed as a model named $\ulcorner$Design Vortal Site; e-BVDS) that was based on the structure and style of internet web site. Results of the study can be described as follows ; \circled1 e-Business is based on the Internet. All processes in the context of e-Business require models whose structure and method of use are on-line styles. \circled2 In case that a traditional manufacturing business is converted into e-Business, it is better to first consider Hybrid Model that combines resources and advantages of both such traditional and digital businesses. \circled3 The product design process appropriate for e-Business environment has to have a structure and style that ensure utilization of the process as an Internet web site, active participation by product developers and interactive communication between participants in designing and designers. \circled4 $\ulcorner$e-BDVS) makes possible the use of designers around the wend like in-house designers, overcoming lack in creativity, ideas and human resources traditional business organizations face. However, the operation of $\ulcorner$e-BDVS$\lrcorner$ requires time and budget investments in securing related elements and conditions. \circled5 Cyber designers under $\ulcorner$e-BDVS$\lrcorner$ can easily perform all design projects in cyber space. But they have some limits in playing a role as designers and they have difficulty in getting rewards if such projects completed by them are not finally accepted. \circled6 $\ulcorner$e-BDVS) ensures the rapid use of a wide range of design information and data, reception of a variety of solutions and ideas and effective design development, all of which are not possible through traditional processes. However, this process may not be suitable to be used routine process or tool. \circled7 $\ulcorner$e-BDVS$\lrcorner$ makes it possible for out-sourcing or partners businesses to overcome restrictions in time and space and improve productivity and effectiveness. But such they may have to continue off-line works that can not be treated on-line.

  • PDF

A Methodology for Extracting Shopping-Related Keywords by Analyzing Internet Navigation Patterns (인터넷 검색기록 분석을 통한 쇼핑의도 포함 키워드 자동 추출 기법)

  • Kim, Mingyu;Kim, Namgyu;Jung, Inhwan
    • Journal of Intelligence and Information Systems
    • /
    • v.20 no.2
    • /
    • pp.123-136
    • /
    • 2014
  • Recently, online shopping has further developed as the use of the Internet and a variety of smart mobile devices becomes more prevalent. The increase in the scale of such shopping has led to the creation of many Internet shopping malls. Consequently, there is a tendency for increasingly fierce competition among online retailers, and as a result, many Internet shopping malls are making significant attempts to attract online users to their sites. One such attempt is keyword marketing, whereby a retail site pays a fee to expose its link to potential customers when they insert a specific keyword on an Internet portal site. The price related to each keyword is generally estimated by the keyword's frequency of appearance. However, it is widely accepted that the price of keywords cannot be based solely on their frequency because many keywords may appear frequently but have little relationship to shopping. This implies that it is unreasonable for an online shopping mall to spend a great deal on some keywords simply because people frequently use them. Therefore, from the perspective of shopping malls, a specialized process is required to extract meaningful keywords. Further, the demand for automating this extraction process is increasing because of the drive to improve online sales performance. In this study, we propose a methodology that can automatically extract only shopping-related keywords from the entire set of search keywords used on portal sites. We define a shopping-related keyword as a keyword that is used directly before shopping behaviors. In other words, only search keywords that direct the search results page to shopping-related pages are extracted from among the entire set of search keywords. A comparison is then made between the extracted keywords' rankings and the rankings of the entire set of search keywords. Two types of data are used in our study's experiment: web browsing history from July 1, 2012 to June 30, 2013, and site information. The experimental dataset was from a web site ranking site, and the biggest portal site in Korea. The original sample dataset contains 150 million transaction logs. First, portal sites are selected, and search keywords in those sites are extracted. Search keywords can be easily extracted by simple parsing. The extracted keywords are ranked according to their frequency. The experiment uses approximately 3.9 million search results from Korea's largest search portal site. As a result, a total of 344,822 search keywords were extracted. Next, by using web browsing history and site information, the shopping-related keywords were taken from the entire set of search keywords. As a result, we obtained 4,709 shopping-related keywords. For performance evaluation, we compared the hit ratios of all the search keywords with the shopping-related keywords. To achieve this, we extracted 80,298 search keywords from several Internet shopping malls and then chose the top 1,000 keywords as a set of true shopping keywords. We measured precision, recall, and F-scores of the entire amount of keywords and the shopping-related keywords. The F-Score was formulated by calculating the harmonic mean of precision and recall. The precision, recall, and F-score of shopping-related keywords derived by the proposed methodology were revealed to be higher than those of the entire number of keywords. This study proposes a scheme that is able to obtain shopping-related keywords in a relatively simple manner. We could easily extract shopping-related keywords simply by examining transactions whose next visit is a shopping mall. The resultant shopping-related keyword set is expected to be a useful asset for many shopping malls that participate in keyword marketing. Moreover, the proposed methodology can be easily applied to the construction of special area-related keywords as well as shopping-related ones.

An Analysis of IT Trends Using Tweet Data (트윗 데이터를 활용한 IT 트렌드 분석)

  • Yi, Jin Baek;Lee, Choong Kwon;Cha, Kyung Jin
    • Journal of Intelligence and Information Systems
    • /
    • v.21 no.1
    • /
    • pp.143-159
    • /
    • 2015
  • Predicting IT trends has been a long and important subject for information systems research. IT trend prediction makes it possible to acknowledge emerging eras of innovation and allocate budgets to prepare against rapidly changing technological trends. Towards the end of each year, various domestic and global organizations predict and announce IT trends for the following year. For example, Gartner Predicts 10 top IT trend during the next year, and these predictions affect IT and industry leaders and organization's basic assumptions about technology and the future of IT, but the accuracy of these reports are difficult to verify. Social media data can be useful tool to verify the accuracy. As social media services have gained in popularity, it is used in a variety of ways, from posting about personal daily life to keeping up to date with news and trends. In the recent years, rates of social media activity in Korea have reached unprecedented levels. Hundreds of millions of users now participate in online social networks and communicate with colleague and friends their opinions and thoughts. In particular, Twitter is currently the major micro blog service, it has an important function named 'tweets' which is to report their current thoughts and actions, comments on news and engage in discussions. For an analysis on IT trends, we chose Tweet data because not only it produces massive unstructured textual data in real time but also it serves as an influential channel for opinion leading on technology. Previous studies found that the tweet data provides useful information and detects the trend of society effectively, these studies also identifies that Twitter can track the issue faster than the other media, newspapers. Therefore, this study investigates how frequently the predicted IT trends for the following year announced by public organizations are mentioned on social network services like Twitter. IT trend predictions for 2013, announced near the end of 2012 from two domestic organizations, the National IT Industry Promotion Agency (NIPA) and the National Information Society Agency (NIA), were used as a basis for this research. The present study analyzes the Twitter data generated from Seoul (Korea) compared with the predictions of the two organizations to analyze the differences. Thus, Twitter data analysis requires various natural language processing techniques, including the removal of stop words, and noun extraction for processing various unrefined forms of unstructured data. To overcome these challenges, we used SAS IRS (Information Retrieval Studio) developed by SAS to capture the trend in real-time processing big stream datasets of Twitter. The system offers a framework for crawling, normalizing, analyzing, indexing and searching tweet data. As a result, we have crawled the entire Twitter sphere in Seoul area and obtained 21,589 tweets in 2013 to review how frequently the IT trend topics announced by the two organizations were mentioned by the people in Seoul. The results shows that most IT trend predicted by NIPA and NIA were all frequently mentioned in Twitter except some topics such as 'new types of security threat', 'green IT', 'next generation semiconductor' since these topics non generalized compound words so they can be mentioned in Twitter with other words. To answer whether the IT trend tweets from Korea is related to the following year's IT trends in real world, we compared Twitter's trending topics with those in Nara Market, Korea's online e-Procurement system which is a nationwide web-based procurement system, dealing with whole procurement process of all public organizations in Korea. The correlation analysis show that Tweet frequencies on IT trending topics predicted by NIPA and NIA are significantly correlated with frequencies on IT topics mentioned in project announcements by Nara market in 2012 and 2013. The main contribution of our research can be found in the following aspects: i) the IT topic predictions announced by NIPA and NIA can provide an effective guideline to IT professionals and researchers in Korea who are looking for verified IT topic trends in the following topic, ii) researchers can use Twitter to get some useful ideas to detect and predict dynamic trends of technological and social issues.

AKARI FAR-INFRARED ALL-SKY SURVEY MAPS

  • Doi, Yasuo;Komugi, Shinya;Kawada, Mitsunobu;Takita, Satoshi;Arimatsu, Ko;Ikeda, Norio;Kato, Daisuke;Kitamura, Yoshimi;Nakagawa, Takao;Ootsubo, Takafumi;Morishima, Takahiro;Hattori, Makoto;Tanaka, Masahiro;White, Glenn J.;Etxaluze, Mireya;Shibai, Hiroshi
    • Publications of The Korean Astronomical Society
    • /
    • v.27 no.4
    • /
    • pp.111-116
    • /
    • 2012
  • Far-infrared observations provide crucial data for the investigation and characterisation of the properties of dusty material in the Interstellar Medium (ISM), since most of its energy is emitted between ~ 100 and $200{\mu}m$. We present the first all-sky image from a sensitive all-sky survey using the Japanese AKARI satellite, in the wavelength range $50-180{\mu}m$. Covering > 99% of the sky in four photometric bands with four filters centred at $65{\mu}m$, $90{\mu}m$, $140{\mu}m$, and $160{\mu}m$ wavelengths, this achieved spatial resolutions from 1 to 2 arcmin and a detection limit of < 10 MJy $sr^{-1}$, with absolute and relative photometric accuracies of < 20%. All-sky images of the Galactic dust continuum emission enable astronomers to map the large-scale distribution of the diffuse ISM cirrus, to study its thermal dust temperature, emissivity and column density, and to measure the interaction of the Galactic radiation field and embedded objects with the surrounding ISM. In addition to the point source population of stars, protostars, star-forming regions, and galaxies, the high Galactic latitude sky is shown to be covered with a diffuse filamentary-web of dusty emission that traces the potential sites of high latitude star formation. We show that the temperature of dust particles in thermal equilibrium with the ambient interstellar radiation field can be estimated by using $90{\mu}m$, $140{\mu}m$, and $160{\mu}m$ data. The FIR AKARI full-sky maps provide a rich new data set within which astronomers can investigate the distribution of interstellar matter throughout our Galaxy, and beyond.

A 2D / 3D Map Modeling of Indoor Environment (실내환경에서의 2 차원/ 3 차원 Map Modeling 제작기법)

  • Jo, Sang-Woo;Park, Jin-Woo;Kwon, Yong-Moo;Ahn, Sang-Chul
    • 한국HCI학회:학술대회논문집
    • /
    • 2006.02a
    • /
    • pp.355-361
    • /
    • 2006
  • In large scale environments like airport, museum, large warehouse and department store, autonomous mobile robots will play an important role in security and surveillance tasks. Robotic security guards will give the surveyed information of large scale environments and communicate with human operator with that kind of data such as if there is an object or not and a window is open. Both for visualization of information and as human machine interface for remote control, a 3D model can give much more useful information than the typical 2D maps used in many robotic applications today. It is easier to understandable and makes user feel like being in a location of robot so that user could interact with robot more naturally in a remote circumstance and see structures such as windows and doors that cannot be seen in a 2D model. In this paper we present our simple and easy to use method to obtain a 3D textured model. For expression of reality, we need to integrate the 3D models and real scenes. Most of other cases of 3D modeling method consist of two data acquisition devices. One for getting a 3D model and another for obtaining realistic textures. In this case, the former device would be 2D laser range-finder and the latter device would be common camera. Our algorithm consists of building a measurement-based 2D metric map which is acquired by laser range-finder, texture acquisition/stitching and texture-mapping to corresponding 3D model. The algorithm is implemented with laser sensor for obtaining 2D/3D metric map and two cameras for gathering texture. Our geometric 3D model consists of planes that model the floor and walls. The geometry of the planes is extracted from the 2D metric map data. Textures for the floor and walls are generated from the images captured by two 1394 cameras which have wide Field of View angle. Image stitching and image cutting process is used to generate textured images for corresponding with a 3D model. The algorithm is applied to 2 cases which are corridor and space that has the four wall like room of building. The generated 3D map model of indoor environment is shown with VRML format and can be viewed in a web browser with a VRML plug-in. The proposed algorithm can be applied to 3D model-based remote surveillance system through WWW.

  • PDF

Benchmarking the Regional Patients Using DEA : Focused on A Oriental Medicine Hospital (자료포락분석방법을 이용한 내원환자의 지역별 벤치마킹분석 : 일개 한방병원을 중심으로)

  • Moon, Kyeong-Jun;Lee, Kwang-Soo;Kwon, Hyuk-Jun
    • The Journal of Society for e-Business Studies
    • /
    • v.19 no.3
    • /
    • pp.91-105
    • /
    • 2014
  • This study purposed to benchmark the number of patients who visited an oriental medicine hospital from its surrounding regions using data envelopment analysis (DEA) model, and to analyze the relationships between regional characteristics and efficiency scores from DEA. Study data was collected from one oriental medicine hospital operated in a metropolitan city in Korea. Patient locations were identified at the smallest administrative district, Dong, and number of patients was calculated at the Dong level based on the address of patients in hospital information system. Socio-demographic variables of each Dong were identified from the Statistics of Korea web-sites. DEA was used to benchmark the number of patients between Dongs and to compute the efficiency scores. Tobit regression analysis model was applied to analyze the relationship between efficiency scores and regional variables. 6 Dongs were identified as efficient after DEA. In Tobit analysis, number of medical aid recipients and number of total population in each Dong was significant in explaining the differences of efficiency scores. The study model introduced the application of DEA model in benchmarking the patients between regions. It can be applied to identify the number of patients in each region which a hospital needs to improve their performances.

Development of decision support system for water resources management using GloSea5 long-term rainfall forecasts and K-DRUM rainfall-runoff model (GloSea5 장기예측 강수량과 K-DRUM 강우-유출모형을 활용한 물관리 의사결정지원시스템 개발)

  • Song, Junghyun;Cho, Younghyun;Kim, Ilseok;Yi, Jonghyuk
    • Journal of Satellite, Information and Communications
    • /
    • v.12 no.3
    • /
    • pp.22-34
    • /
    • 2017
  • The K-DRUM(K-water hydrologic & hydraulic Distributed RUnoff Model), a distributed rainfall-runoff model of K-water, calculates predicted runoff and water surface level of a dam using precipitation data. In order to obtain long-term hydrometeorological information, K-DRUM requires long-term weather forecast. In this study, we built a system providing long-term hydrometeorological information using predicted rainfall ensemble of GloSea5(Global Seasonal Forecast System version 5), which is the seasonal meteorological forecasting system of KMA introduced in 2014. This system produces K-DRUM input data by automatic pre-processing and bias-correcting GloSea5 data, then derives long-term inflow predictions via K-DRUM. Web-based UI was developed for users to monitor the hydrometeorological information such as rainfall, runoff, and water surface level of dams. Through this UI, users can also test various dam management scenarios by adjusting discharge amount for decision-making.