• Title/Summary/Keyword: 처리시스템

Search Result 27,767, Processing Time 0.059 seconds

Image Watermarking for Copyright Protection of Images on Shopping Mall (쇼핑몰 이미지 저작권보호를 위한 영상 워터마킹)

  • Bae, Kyoung-Yul
    • Journal of Intelligence and Information Systems
    • /
    • v.19 no.4
    • /
    • pp.147-157
    • /
    • 2013
  • With the advent of the digital environment that can be accessed anytime, anywhere with the introduction of high-speed network, the free distribution and use of digital content were made possible. Ironically this environment is raising a variety of copyright infringement, and product images used in the online shopping mall are pirated frequently. There are many controversial issues whether shopping mall images are creative works or not. According to Supreme Court's decision in 2001, to ad pictures taken with ham products is simply a clone of the appearance of objects to deliver nothing but the decision was not only creative expression. But for the photographer's losses recognized in the advertising photo shoot takes the typical cost was estimated damages. According to Seoul District Court precedents in 2003, if there are the photographer's personality and creativity in the selection of the subject, the composition of the set, the direction and amount of light control, set the angle of the camera, shutter speed, shutter chance, other shooting methods for capturing, developing and printing process, the works should be protected by copyright law by the Court's sentence. In order to receive copyright protection of the shopping mall images by the law, it is simply not to convey the status of the product, the photographer's personality and creativity can be recognized that it requires effort. Accordingly, the cost of making the mall image increases, and the necessity for copyright protection becomes higher. The product images of the online shopping mall have a very unique configuration unlike the general pictures such as portraits and landscape photos and, therefore, the general image watermarking technique can not satisfy the requirements of the image watermarking. Because background of product images commonly used in shopping malls is white or black, or gray scale (gradient) color, it is difficult to utilize the space to embed a watermark and the area is very sensitive even a slight change. In this paper, the characteristics of images used in shopping malls are analyzed and a watermarking technology which is suitable to the shopping mall images is proposed. The proposed image watermarking technology divide a product image into smaller blocks, and the corresponding blocks are transformed by DCT (Discrete Cosine Transform), and then the watermark information was inserted into images using quantization of DCT coefficients. Because uniform treatment of the DCT coefficients for quantization cause visual blocking artifacts, the proposed algorithm used weighted mask which quantizes finely the coefficients located block boundaries and coarsely the coefficients located center area of the block. This mask improves subjective visual quality as well as the objective quality of the images. In addition, in order to improve the safety of the algorithm, the blocks which is embedded the watermark are randomly selected and the turbo code is used to reduce the BER when extracting the watermark. The PSNR(Peak Signal to Noise Ratio) of the shopping mall image watermarked by the proposed algorithm is 40.7~48.5[dB] and BER(Bit Error Rate) after JPEG with QF = 70 is 0. This means the watermarked image is high quality and the algorithm is robust to JPEG compression that is used generally at the online shopping malls. Also, for 40% change in size and 40 degrees of rotation, the BER is 0. In general, the shopping malls are used compressed images with QF which is higher than 90. Because the pirated image is used to replicate from original image, the proposed algorithm can identify the copyright infringement in the most cases. As shown the experimental results, the proposed algorithm is suitable to the shopping mall images with simple background. However, the future study should be carried out to enhance the robustness of the proposed algorithm because the robustness loss is occurred after mask process.

Comparison of marginal fit before and after porcelain build-up of two kinds of CAD/CAM zirconia all-ceramic restorations (두 종류의 CAD/CAM 지르코니아 전부도재관의 도재 축성 전후의 변연적합도 비교)

  • Shin, Ho-Sik;Kim, Seok-Gyu
    • The Journal of Korean Academy of Prosthodontics
    • /
    • v.46 no.5
    • /
    • pp.528-534
    • /
    • 2008
  • Purpose: Marginal fit is one of the important components for the successful prosthodontic restoration. Poor fitting margin of the restoration causes hypersensitivity, secondary caries, and plaque accumulation, which later result in prosthodontic failure. CAD/CAM zirconia all-ceramic restorations, such as $LAVA^{(R)}$ (3M ESPE, St.Paul, MN) and $EVEREST^{(R)}$ (KaVo Dental GmbH, Biberach, Germany) systems were recently introduced in Korea. It is clinically meaningful to evaluate the changes of the marginal fit of the CAD/CAM zirconia systems before and after build-up. The purposes of this study are to compare the marginal fit of the two CAD/CAM all-ceramic systems with that of the ceramometal restoration, before and after porcelain build-up Material and methods: A maxillary first premolar dentiform tooth was prepared with 2.0 mm occlusal reduction, 1.0 mm axial reduction, chamfer margin, and 6 degree taperness in the axial wall. The prepared dentiform die was duplicated into the metal abutment die. The metal die was placed in the dental study model, and the full arch impressions of the model were made. Twenty four copings of 3 groups which were $LAVA^{(R)}$, $EVEREST^{(R)}$, and ceramometal restorations were fabricated. Each coping was cemented on the metal die with color-mixed Fit-checker $II^{(R)}$ (GC Cor., Tokyo, Japan). The marginal opening of each coping was measured with $Microhiscope^{(R)}$ system (HIROX KH-1000 ING-Plus, Seoul, Korea. X300 magnification). After porcelain build-up, the marginal openings of $LAVA^{(R)}$, $EVEREST^{(R)}$,and ceramometal restorations were also evaluated in the same method. Statistical analysis was done with paired t-test and one-way ANOVA test. Results: In coping states, the mean marginal opening for $EVEREST^{(R)}$ restorations was $52.00{\pm}11.94\;{\mu}m$ for $LAVA^{(R)}$ restorations $56.97{\pm}10.00\;{\mu}m$, and for ceramometal restorations $97.38{\pm}18.54\;{\mu}m$. After porcelain build-up, the mean marginal opening for $EVEREST^{(R)}$ restorations was $61.69{\pm}19.33\;{\mu}m$, for $LAVA^{(R)}$ restorations $70.81{\pm}12.99\;{\mu}m$, and for ceramometal restorations $1115.25{\pm}23.86\;{\mu}m$. Conclusion: 1. $LAVA^{(R)}$ and $EVEREST^{(R)}$ restorations in comparison with ceramometal restorations showed better marginal fit, which had significant differences (P < 0.05) in coping state and also after porcelain build-up . 2. The mean marginal opening values between $LAVA^{(R)}$ and $EVEREST^{(R)}$ restorations did not showed significant differences after porcelain build-up as well as in coping state (P > .05). 3. $EVEREST^{(R)}$, $LAVA^{(R)}$ and ceramometal restorations showed a little increased marginal opening after porcelain build-up, but did not show any statistical significance (P > .05).

A Study on Differences of Contents and Tones of Arguments among Newspapers Using Text Mining Analysis (텍스트 마이닝을 활용한 신문사에 따른 내용 및 논조 차이점 분석)

  • Kam, Miah;Song, Min
    • Journal of Intelligence and Information Systems
    • /
    • v.18 no.3
    • /
    • pp.53-77
    • /
    • 2012
  • This study analyses the difference of contents and tones of arguments among three Korean major newspapers, the Kyunghyang Shinmoon, the HanKyoreh, and the Dong-A Ilbo. It is commonly accepted that newspapers in Korea explicitly deliver their own tone of arguments when they talk about some sensitive issues and topics. It could be controversial if readers of newspapers read the news without being aware of the type of tones of arguments because the contents and the tones of arguments can affect readers easily. Thus it is very desirable to have a new tool that can inform the readers of what tone of argument a newspaper has. This study presents the results of clustering and classification techniques as part of text mining analysis. We focus on six main subjects such as Culture, Politics, International, Editorial-opinion, Eco-business and National issues in newspapers, and attempt to identify differences and similarities among the newspapers. The basic unit of text mining analysis is a paragraph of news articles. This study uses a keyword-network analysis tool and visualizes relationships among keywords to make it easier to see the differences. Newspaper articles were gathered from KINDS, the Korean integrated news database system. KINDS preserves news articles of the Kyunghyang Shinmun, the HanKyoreh and the Dong-A Ilbo and these are open to the public. This study used these three Korean major newspapers from KINDS. About 3,030 articles from 2008 to 2012 were used. International, national issues and politics sections were gathered with some specific issues. The International section was collected with the keyword of 'Nuclear weapon of North Korea.' The National issues section was collected with the keyword of '4-major-river.' The Politics section was collected with the keyword of 'Tonghap-Jinbo Dang.' All of the articles from April 2012 to May 2012 of Eco-business, Culture and Editorial-opinion sections were also collected. All of the collected data were handled and edited into paragraphs. We got rid of stop-words using the Lucene Korean Module. We calculated keyword co-occurrence counts from the paired co-occurrence list of keywords in a paragraph. We made a co-occurrence matrix from the list. Once the co-occurrence matrix was built, we used the Cosine coefficient matrix as input for PFNet(Pathfinder Network). In order to analyze these three newspapers and find out the significant keywords in each paper, we analyzed the list of 10 highest frequency keywords and keyword-networks of 20 highest ranking frequency keywords to closely examine the relationships and show the detailed network map among keywords. We used NodeXL software to visualize the PFNet. After drawing all the networks, we compared the results with the classification results. Classification was firstly handled to identify how the tone of argument of a newspaper is different from others. Then, to analyze tones of arguments, all the paragraphs were divided into two types of tones, Positive tone and Negative tone. To identify and classify all of the tones of paragraphs and articles we had collected, supervised learning technique was used. The Na$\ddot{i}$ve Bayesian classifier algorithm provided in the MALLET package was used to classify all the paragraphs in articles. After classification, Precision, Recall and F-value were used to evaluate the results of classification. Based on the results of this study, three subjects such as Culture, Eco-business and Politics showed some differences in contents and tones of arguments among these three newspapers. In addition, for the National issues, tones of arguments on 4-major-rivers project were different from each other. It seems three newspapers have their own specific tone of argument in those sections. And keyword-networks showed different shapes with each other in the same period in the same section. It means that frequently appeared keywords in articles are different and their contents are comprised with different keywords. And the Positive-Negative classification showed the possibility of classifying newspapers' tones of arguments compared to others. These results indicate that the approach in this study is promising to be extended as a new tool to identify the different tones of arguments of newspapers.

Research Trend Analysis Using Bibliographic Information and Citations of Cloud Computing Articles: Application of Social Network Analysis (클라우드 컴퓨팅 관련 논문의 서지정보 및 인용정보를 활용한 연구 동향 분석: 사회 네트워크 분석의 활용)

  • Kim, Dongsung;Kim, Jongwoo
    • Journal of Intelligence and Information Systems
    • /
    • v.20 no.1
    • /
    • pp.195-211
    • /
    • 2014
  • Cloud computing services provide IT resources as services on demand. This is considered a key concept, which will lead a shift from an ownership-based paradigm to a new pay-for-use paradigm, which can reduce the fixed cost for IT resources, and improve flexibility and scalability. As IT services, cloud services have evolved from early similar computing concepts such as network computing, utility computing, server-based computing, and grid computing. So research into cloud computing is highly related to and combined with various relevant computing research areas. To seek promising research issues and topics in cloud computing, it is necessary to understand the research trends in cloud computing more comprehensively. In this study, we collect bibliographic information and citation information for cloud computing related research papers published in major international journals from 1994 to 2012, and analyzes macroscopic trends and network changes to citation relationships among papers and the co-occurrence relationships of key words by utilizing social network analysis measures. Through the analysis, we can identify the relationships and connections among research topics in cloud computing related areas, and highlight new potential research topics. In addition, we visualize dynamic changes of research topics relating to cloud computing using a proposed cloud computing "research trend map." A research trend map visualizes positions of research topics in two-dimensional space. Frequencies of key words (X-axis) and the rates of increase in the degree centrality of key words (Y-axis) are used as the two dimensions of the research trend map. Based on the values of the two dimensions, the two dimensional space of a research map is divided into four areas: maturation, growth, promising, and decline. An area with high keyword frequency, but low rates of increase of degree centrality is defined as a mature technology area; the area where both keyword frequency and the increase rate of degree centrality are high is defined as a growth technology area; the area where the keyword frequency is low, but the rate of increase in the degree centrality is high is defined as a promising technology area; and the area where both keyword frequency and the rate of degree centrality are low is defined as a declining technology area. Based on this method, cloud computing research trend maps make it possible to easily grasp the main research trends in cloud computing, and to explain the evolution of research topics. According to the results of an analysis of citation relationships, research papers on security, distributed processing, and optical networking for cloud computing are on the top based on the page-rank measure. From the analysis of key words in research papers, cloud computing and grid computing showed high centrality in 2009, and key words dealing with main elemental technologies such as data outsourcing, error detection methods, and infrastructure construction showed high centrality in 2010~2011. In 2012, security, virtualization, and resource management showed high centrality. Moreover, it was found that the interest in the technical issues of cloud computing increases gradually. From annual cloud computing research trend maps, it was verified that security is located in the promising area, virtualization has moved from the promising area to the growth area, and grid computing and distributed system has moved to the declining area. The study results indicate that distributed systems and grid computing received a lot of attention as similar computing paradigms in the early stage of cloud computing research. The early stage of cloud computing was a period focused on understanding and investigating cloud computing as an emergent technology, linking to relevant established computing concepts. After the early stage, security and virtualization technologies became main issues in cloud computing, which is reflected in the movement of security and virtualization technologies from the promising area to the growth area in the cloud computing research trend maps. Moreover, this study revealed that current research in cloud computing has rapidly transferred from a focus on technical issues to for a focus on application issues, such as SLAs (Service Level Agreements).

Individual Thinking Style leads its Emotional Perception: Development of Web-style Design Evaluation Model and Recommendation Algorithm Depending on Consumer Regulatory Focus (사고가 시각을 바꾼다: 조절 초점에 따른 소비자 감성 기반 웹 스타일 평가 모형 및 추천 알고리즘 개발)

  • Kim, Keon-Woo;Park, Do-Hyung
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.4
    • /
    • pp.171-196
    • /
    • 2018
  • With the development of the web, two-way communication and evaluation became possible and marketing paradigms shifted. In order to meet the needs of consumers, web design trends are continuously responding to consumer feedback. As the web becomes more and more important, both academics and businesses are studying consumer emotions and satisfaction on the web. However, some consumer characteristics are not well considered. Demographic characteristics such as age and sex have been studied extensively, but few studies consider psychological characteristics such as regulatory focus (i.e., emotional regulation). In this study, we analyze the effect of web style on consumer emotion. Many studies analyze the relationship between the web and regulatory focus, but most concentrate on the purpose of web use, particularly motivation and information search, rather than on web style and design. The web communicates with users through visual elements. Because the human brain is influenced by all five senses, both design factors and emotional responses are important in the web environment. Therefore, in this study, we examine the relationship between consumer emotion and satisfaction and web style and design. Previous studies have considered the effects of web layout, structure, and color on emotions. In this study, however, we excluded these web components, in contrast to earlier studies, and analyzed the relationship between consumer satisfaction and emotional indexes of web-style only. To perform this analysis, we collected consumer surveys presenting 40 web style themes to 204 consumers. Each consumer evaluated four themes. The emotional adjectives evaluated by consumers were composed of 18 contrast pairs, and the upper emotional indexes were extracted through factor analysis. The emotional indexes were 'softness,' 'modernity,' 'clearness,' and 'jam.' Hypotheses were established based on the assumption that emotional indexes have different effects on consumer satisfaction. After the analysis, hypotheses 1, 2, and 3 were accepted and hypothesis 4 was rejected. While hypothesis 4 was rejected, its effect on consumer satisfaction was negative, not positive. This means that emotional indexes such as 'softness,' 'modernity,' and 'clearness' have a positive effect on consumer satisfaction. In other words, consumers prefer emotions that are soft, emotional, natural, rounded, dynamic, modern, elaborate, unique, bright, pure, and clear. 'Jam' has a negative effect on consumer satisfaction. It means, consumer prefer the emotion which is empty, plain, and simple. Regulatory focus shows differences in motivation and propensity in various domains. It is important to consider organizational behavior and decision making according to the regulatory focus tendency, and it affects not only political, cultural, ethical judgments and behavior but also broad psychological problems. Regulatory focus also differs from emotional response. Promotion focus responds more strongly to positive emotional responses. On the other hand, prevention focus has a strong response to negative emotions. Web style is a type of service, and consumer satisfaction is affected not only by cognitive evaluation but also by emotion. This emotional response depends on whether the consumer will benefit or harm himself. Therefore, it is necessary to confirm the difference of the consumer's emotional response according to the regulatory focus which is one of the characteristics and viewpoint of the consumers about the web style. After MMR analysis result, hypothesis 5.3 was accepted, and hypothesis 5.4 was rejected. But hypothesis 5.4 supported in the opposite direction to the hypothesis. After validation, we confirmed the mechanism of emotional response according to the tendency of regulatory focus. Using the results, we developed the structure of web-style recommendation system and recommend methods through regulatory focus. We classified the regulatory focus group in to three categories that promotion, grey, prevention. Then, we suggest web-style recommend method along the group. If we further develop this study, we expect that the existing regulatory focus theory can be extended not only to the motivational part but also to the emotional behavioral response according to the regulatory focus tendency. Moreover, we believe that it is possible to recommend web-style according to regulatory focus and emotional desire which consumers most prefer.

A Study of Anomaly Detection for ICT Infrastructure using Conditional Multimodal Autoencoder (ICT 인프라 이상탐지를 위한 조건부 멀티모달 오토인코더에 관한 연구)

  • Shin, Byungjin;Lee, Jonghoon;Han, Sangjin;Park, Choong-Shik
    • Journal of Intelligence and Information Systems
    • /
    • v.27 no.3
    • /
    • pp.57-73
    • /
    • 2021
  • Maintenance and prevention of failure through anomaly detection of ICT infrastructure is becoming important. System monitoring data is multidimensional time series data. When we deal with multidimensional time series data, we have difficulty in considering both characteristics of multidimensional data and characteristics of time series data. When dealing with multidimensional data, correlation between variables should be considered. Existing methods such as probability and linear base, distance base, etc. are degraded due to limitations called the curse of dimensions. In addition, time series data is preprocessed by applying sliding window technique and time series decomposition for self-correlation analysis. These techniques are the cause of increasing the dimension of data, so it is necessary to supplement them. The anomaly detection field is an old research field, and statistical methods and regression analysis were used in the early days. Currently, there are active studies to apply machine learning and artificial neural network technology to this field. Statistically based methods are difficult to apply when data is non-homogeneous, and do not detect local outliers well. The regression analysis method compares the predictive value and the actual value after learning the regression formula based on the parametric statistics and it detects abnormality. Anomaly detection using regression analysis has the disadvantage that the performance is lowered when the model is not solid and the noise or outliers of the data are included. There is a restriction that learning data with noise or outliers should be used. The autoencoder using artificial neural networks is learned to output as similar as possible to input data. It has many advantages compared to existing probability and linear model, cluster analysis, and map learning. It can be applied to data that does not satisfy probability distribution or linear assumption. In addition, it is possible to learn non-mapping without label data for teaching. However, there is a limitation of local outlier identification of multidimensional data in anomaly detection, and there is a problem that the dimension of data is greatly increased due to the characteristics of time series data. In this study, we propose a CMAE (Conditional Multimodal Autoencoder) that enhances the performance of anomaly detection by considering local outliers and time series characteristics. First, we applied Multimodal Autoencoder (MAE) to improve the limitations of local outlier identification of multidimensional data. Multimodals are commonly used to learn different types of inputs, such as voice and image. The different modal shares the bottleneck effect of Autoencoder and it learns correlation. In addition, CAE (Conditional Autoencoder) was used to learn the characteristics of time series data effectively without increasing the dimension of data. In general, conditional input mainly uses category variables, but in this study, time was used as a condition to learn periodicity. The CMAE model proposed in this paper was verified by comparing with the Unimodal Autoencoder (UAE) and Multi-modal Autoencoder (MAE). The restoration performance of Autoencoder for 41 variables was confirmed in the proposed model and the comparison model. The restoration performance is different by variables, and the restoration is normally well operated because the loss value is small for Memory, Disk, and Network modals in all three Autoencoder models. The process modal did not show a significant difference in all three models, and the CPU modal showed excellent performance in CMAE. ROC curve was prepared for the evaluation of anomaly detection performance in the proposed model and the comparison model, and AUC, accuracy, precision, recall, and F1-score were compared. In all indicators, the performance was shown in the order of CMAE, MAE, and AE. Especially, the reproduction rate was 0.9828 for CMAE, which can be confirmed to detect almost most of the abnormalities. The accuracy of the model was also improved and 87.12%, and the F1-score was 0.8883, which is considered to be suitable for anomaly detection. In practical aspect, the proposed model has an additional advantage in addition to performance improvement. The use of techniques such as time series decomposition and sliding windows has the disadvantage of managing unnecessary procedures; and their dimensional increase can cause a decrease in the computational speed in inference.The proposed model has characteristics that are easy to apply to practical tasks such as inference speed and model management.

Application and Analysis of Ocean Remote-Sensing Reflectance Quality Assurance Algorithm for GOCI-II (천리안해양위성 2호(GOCI-II) 원격반사도 품질 검증 시스템 적용 및 결과)

  • Sujung Bae;Eunkyung Lee;Jianwei Wei;Kyeong-sang Lee;Minsang Kim;Jong-kuk Choi;Jae Hyun Ahn
    • Korean Journal of Remote Sensing
    • /
    • v.39 no.6_2
    • /
    • pp.1565-1576
    • /
    • 2023
  • An atmospheric correction algorithm based on the radiative transfer model is required to obtain remote-sensing reflectance (Rrs) from the Geostationary Ocean Color Imager-II (GOCI-II) observed at the top-of-atmosphere. This Rrs derived from the atmospheric correction is utilized to estimate various marine environmental parameters such as chlorophyll-a concentration, total suspended materials concentration, and absorption of dissolved organic matter. Therefore, an atmospheric correction is a fundamental algorithm as it significantly impacts the reliability of all other color products. However, in clear waters, for example, atmospheric path radiance exceeds more than ten times higher than the water-leaving radiance in the blue wavelengths. This implies atmospheric correction is a highly error-sensitive process with a 1% error in estimating atmospheric radiance in the atmospheric correction process can cause more than 10% errors. Therefore, the quality assessment of Rrs after the atmospheric correction is essential for ensuring reliable ocean environment analysis using ocean color satellite data. In this study, a Quality Assurance (QA) algorithm based on in-situ Rrs data, which has been archived into a database using Sea-viewing Wide Field-of-view Sensor (SeaWiFS) Bio-optical Archive and Storage System (SeaBASS), was applied and modified to consider the different spectral characteristics of GOCI-II. This method is officially employed in the National Oceanic and Atmospheric Administration (NOAA)'s ocean color satellite data processing system. It provides quality analysis scores for Rrs ranging from 0 to 1 and classifies the water types into 23 categories. When the QA algorithm is applied to the initial phase of GOCI-II data with less calibration, it shows the highest frequency at a relatively low score of 0.625. However, when the algorithm is applied to the improved GOCI-II atmospheric correction results with updated calibrations, it shows the highest frequency at a higher score of 0.875 compared to the previous results. The water types analysis using the QA algorithm indicated that parts of the East Sea, South Sea, and the Northwest Pacific Ocean are primarily characterized as relatively clear case-I waters, while the coastal areas of the Yellow Sea and the East China Sea are mainly classified as highly turbid case-II waters. We expect that the QA algorithm will support GOCI-II users in terms of not only statistically identifying Rrs resulted with significant errors but also more reliable calibration with quality assured data. The algorithm will be included in the level-2 flag data provided with GOCI-II atmospheric correction.

Smoking Adolescents' Acquisition of Cigarettes and Status of Proof of Age (흡연을 하는 청소년의 담배 구입 경로 및 신분 확인의 유무)

  • Kim, Hee Ra;Kim, Ji Young;Lee, Gee Hyung;Choung, Ji Tae;Park, Sang Hee
    • Clinical and Experimental Pediatrics
    • /
    • v.48 no.4
    • /
    • pp.363-368
    • /
    • 2005
  • Purpose : The aim of this study is to identify where and how adolescents acquire cigarettes and how many were asked for identification while purchasing cigarettes. Methods : This study was conducted in 2003; participants were 2,200 students in middle and high schools, aged from 13 to 18 years old(males 1,098; females 1,102) in Ansan, Korea. The questionnaire assured them of anonymity, and self-administered in school. The data was analyzed with chisquare test for trends. Results : The prevalence of smoking was about 20 percent among respondents, was higher in males than in females, and in older students than in younger students(P<0.001). The most frequent source of cigarettes was purchased from a store(36.3 percent). About 29.2 percent of the students borrowed from friends or family members. By sex, the main sources of cigarettes were purchase from a store and borrowing. Younger students were borrowed more cigarettes; older students purchased more cigarettes from stores. Only 48.8 percent were asked for proof of age during their purchase. Of those asked for proof of age, about 73.3 percent answered that this made it difficult to buy cigarettes(P<0.001), and they thought that it was more difficult when asked for a photo ID than simply being asked their age(P=0.019). Conclusion : So far, there has been no systemic prevention of adolescents' smoking. It is difficult for minors to purchase cigarettes if asked for proof of age, but most minors go to stores to purchase cigarettes. Therefore, prevention efforts should include educating retailers not to sell cigarettes to minors and enforcing existing laws requiring youth to provide proof of age when attempting to buy cigarettes.

Estimation of Optimal Size of the Treatment Facility for Nonpoint Source Pollution due to Watershed Development (비점오염원의 정량화방안에 따른 적정 설계용량결정)

  • Kim, Jin-Kwan
    • Journal of the Korean Society of Hazard Mitigation
    • /
    • v.8 no.6
    • /
    • pp.149-153
    • /
    • 2008
  • The pollutant capacity occurred before and after the development of a watershed should be quantitatively estimated and controlled for the minimization of water contamination. The Ministry of Environment suggested a guideline for the legal management of nonpoint source from 2006. However, the rational method for the determination of treatment capacity from nonpoint source proposed in the guideline has the problem in the field application because it does not reflect the project based cases and overestimates the pollutant load to be reduced. So, we perform the standard rainfall analysis by analytical probabilistic method for the estimation of an additional pollutant load occurred by a project and suggest a methodology for the estimation of contaminant capacity instead of a simple rational method. The suggested methodology in this study could determine the reasonable capacity and efficiency of a treatment facility through the estimation of pollutant load from nonpoint source and from this we can manage the watershed appropriately. We applied a suggested methodology to the projects of housing land development and a dam construction in the watersheds. When we determine the treatment capacity by a rational method without consideration of the types of projects we should treat the 90% of pollutant capacity occurred by the development and to do so, about 30% of the total cost for the development should be invested for the treatment facility. This requires too big cost and is not realistic. If we use the suggested method the target pollutant capacity to be reduced will be 10 to 30% of the capacity occurred by the development and about 5 to 10% of the total cost can be used. The control of nonpoint source must be performed for the water resources management. However it is not possible to treat the 90% of pollutant load occurred by the development. The proper pollutant capacity from nonpoint source should be estimated and controlled based on various project types and in reality, this is very important for the watershed management. Therefore the results of this study might be more reasonable than the rational method proposed in the Ministry of Environment.

Investigation of the Rice Plant Transfer and the Leaching Characteristics of Copper and Lead for the Stabilization Process with a Pilot Scale Test (논토양 안정화 현장 실증 시험을 통한 납, 구리의 용출 저감 및 벼로의 식물전이 특성 규명)

  • Lee, Ha-Jung;Lee, Min-Hee
    • Economic and Environmental Geology
    • /
    • v.45 no.3
    • /
    • pp.255-264
    • /
    • 2012
  • The stabilization using limestone ($CaCO_3$) and steel making slag as the immobilization amendments for Cu and Pb contaminated farmland soils was investigated by batch tests, continuous column experiments and the pilot scale feasibility study with 4 testing grounds at the contaminated site. From the results of batch experiment, the amendment with the mixture of 3% of limestone and 2% of steel making slag reduced more than 85% of Cu and Pb compared with the soil without amendment. The acryl column (1 m in length and 15 cm in diameter) equipped with valves, tubes and a sprinkler was used for the continuous column experiments. Without the amendment, the Pb concentration of the leachate from the column maintained higher than 0.1 mg/L (groundwater tolerance limit). However, the amendment with 3% limestone and 2% steel making slag reduced more than 60% of Pb leaching concentration within 1 year and the Pb concentration of leachate maintained below 0.04 mg/L. For the testing ground without the amendment, the Pb and Cu concentrations of soil water after 60 days incubation were 0.38 mg/L and 0.69 mg/l, respectively, suggesting that the continuous leaching of Cu and Pb may occur from the site. For the testing ground amended with mixture of 3% of limestone + 2% of steel making slag, no water soluble Pb and Cu were detected after 20 days incubation. For all testing grounds, the ratio of Pb and Cu transfer to plant showed as following: root > leaves(including stem) > rice grain. The amendment with limestone and steel making slag reduced more than 75% Pb and Cu transfer to plant comparing with no amendment. The results of this study showed that the amendment with mixture of limestone and steel making slag decreases not only the leaching of heavy metals but also the plant transfer from the soil.