• Title/Summary/Keyword: Multi-Target

Search Result 1,399, Processing Time 0.035 seconds

A study on improving the accuracy of machine learning models through the use of non-financial information in predicting the Closure of operator using electronic payment service (전자결제서비스 이용 사업자 폐업 예측에서 비재무정보 활용을 통한 머신러닝 모델의 정확도 향상에 관한 연구)

  • Hyunjeong Gong;Eugene Hwang;Sunghyuk Park
    • Journal of Intelligence and Information Systems
    • /
    • v.29 no.3
    • /
    • pp.361-381
    • /
    • 2023
  • Research on corporate bankruptcy prediction has been focused on financial information. Since the company's financial information is updated quarterly, there is a problem that timeliness is insufficient in predicting the possibility of a company's business closure in real time. Evaluated companies that want to improve this need a method of judging the soundness of a company that uses information other than financial information to judge the soundness of a target company. To this end, as information technology has made it easier to collect non-financial information about companies, research has been conducted to apply additional variables and various methodologies other than financial information to predict corporate bankruptcy. It has become an important research task to determine whether it has an effect. In this study, we examined the impact of electronic payment-related information, which constitutes non-financial information, when predicting the closure of business operators using electronic payment service and examined the difference in closure prediction accuracy according to the combination of financial and non-financial information. Specifically, three research models consisting of a financial information model, a non-financial information model, and a combined model were designed, and the closure prediction accuracy was confirmed with six algorithms including the Multi Layer Perceptron (MLP) algorithm. The model combining financial and non-financial information showed the highest prediction accuracy, followed by the non-financial information model and the financial information model in order. As for the prediction accuracy of business closure by algorithm, XGBoost showed the highest prediction accuracy among the six algorithms. As a result of examining the relative importance of a total of 87 variables used to predict business closure, it was confirmed that more than 70% of the top 20 variables that had a significant impact on the prediction of business closure were non-financial information. Through this, it was confirmed that electronic payment-related information of non-financial information is an important variable in predicting business closure, and the possibility of using non-financial information as an alternative to financial information was also examined. Based on this study, the importance of collecting and utilizing non-financial information as information that can predict business closure is recognized, and a plan to utilize it for corporate decision-making is also proposed.

Off-pump Coronary Artery Bypass Surgery Versus Drug Eluting Stent for Multi-vessel Coronary Artery Disease (다혈관 관상동맥질환에서의 심폐바이패스를 사용하지 않은 관상동맥우회술과 약물용출 스텐트시술)

  • Lee, Jae-Hang;Kim, Ki-Bong;Cho, Kwang-Ree;Park, Jin-Shik;Kang, Hyun-Jae;Koo, Bon-Kwon;Kim, Hyo-Soo;Sohn, Dae-Won;Oh, Byung-Hee;Park, Young-Bae
    • Journal of Chest Surgery
    • /
    • v.41 no.2
    • /
    • pp.202-209
    • /
    • 2008
  • Background: The introduction of Drug Eluting Stents (DES) decreased the number of patients referred for coronary artery bypass grafting (CABG). The impact of DES on CABG (Step 1) was studied and compared with the 1-year outcome after CABG with DES (Step 2). Material and Method: Surgical results for patients who underwent off-pump CABG (OPCAB) before the introduction of DES(n=298) were compared with those who underwent OPCAB after the introduction of DES (n=288) (Step 1). Postoperative 30-day and 1-year results were also compared between the patients who underwent percutaneous coronary intervention (PCI) using DES (n=220) and those who underwent OPCAB (n=255) (Step 2). Result: Since the introduction of DES, the ratio of CABG versus PCI decreased. In the CABG group, the number of high risk patients such as elderly patients (age 62 vs. 64, p=0.023), those with chronic renal failure (4% vs. 9%, p=0.021), calcification of the ascending aorta (9% vs. 15%, p=0.043), or frequency of urgent or emergent operations (12% vs. 22%, p=0.002) increased. However, there were no differences in the cardiac death and graft patency rates between the two groups (step 1). During the one-year follow up period, the rate of target vessel revascularization (12.3% vs. 2.4%, p<0.001) and major adverse cardiac events (MACE: death, myocardial infarct, TVR) were higher in the DES than the CABG group (13.6% vs 4.3%) (stage 2). Conclusion: Introduction of DES decreased the number of patients referred for surgery, and increased the comorbidity in patients who underwent CABG. DES increased the rate of target vessel revascularization, and the occurrence of MACE during the 1-year follow-up. However, there was no difference in the incidence of myocardial infarction and cardiac death between the two groups.

Feasibility of Deep Learning Algorithms for Binary Classification Problems (이진 분류문제에서의 딥러닝 알고리즘의 활용 가능성 평가)

  • Kim, Kitae;Lee, Bomi;Kim, Jong Woo
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.1
    • /
    • pp.95-108
    • /
    • 2017
  • Recently, AlphaGo which is Bakuk (Go) artificial intelligence program by Google DeepMind, had a huge victory against Lee Sedol. Many people thought that machines would not be able to win a man in Go games because the number of paths to make a one move is more than the number of atoms in the universe unlike chess, but the result was the opposite to what people predicted. After the match, artificial intelligence technology was focused as a core technology of the fourth industrial revolution and attracted attentions from various application domains. Especially, deep learning technique have been attracted as a core artificial intelligence technology used in the AlphaGo algorithm. The deep learning technique is already being applied to many problems. Especially, it shows good performance in image recognition field. In addition, it shows good performance in high dimensional data area such as voice, image and natural language, which was difficult to get good performance using existing machine learning techniques. However, in contrast, it is difficult to find deep leaning researches on traditional business data and structured data analysis. In this study, we tried to find out whether the deep learning techniques have been studied so far can be used not only for the recognition of high dimensional data but also for the binary classification problem of traditional business data analysis such as customer churn analysis, marketing response prediction, and default prediction. And we compare the performance of the deep learning techniques with that of traditional artificial neural network models. The experimental data in the paper is the telemarketing response data of a bank in Portugal. It has input variables such as age, occupation, loan status, and the number of previous telemarketing and has a binary target variable that records whether the customer intends to open an account or not. In this study, to evaluate the possibility of utilization of deep learning algorithms and techniques in binary classification problem, we compared the performance of various models using CNN, LSTM algorithm and dropout, which are widely used algorithms and techniques in deep learning, with that of MLP models which is a traditional artificial neural network model. However, since all the network design alternatives can not be tested due to the nature of the artificial neural network, the experiment was conducted based on restricted settings on the number of hidden layers, the number of neurons in the hidden layer, the number of output data (filters), and the application conditions of the dropout technique. The F1 Score was used to evaluate the performance of models to show how well the models work to classify the interesting class instead of the overall accuracy. The detail methods for applying each deep learning technique in the experiment is as follows. The CNN algorithm is a method that reads adjacent values from a specific value and recognizes the features, but it does not matter how close the distance of each business data field is because each field is usually independent. In this experiment, we set the filter size of the CNN algorithm as the number of fields to learn the whole characteristics of the data at once, and added a hidden layer to make decision based on the additional features. For the model having two LSTM layers, the input direction of the second layer is put in reversed position with first layer in order to reduce the influence from the position of each field. In the case of the dropout technique, we set the neurons to disappear with a probability of 0.5 for each hidden layer. The experimental results show that the predicted model with the highest F1 score was the CNN model using the dropout technique, and the next best model was the MLP model with two hidden layers using the dropout technique. In this study, we were able to get some findings as the experiment had proceeded. First, models using dropout techniques have a slightly more conservative prediction than those without dropout techniques, and it generally shows better performance in classification. Second, CNN models show better classification performance than MLP models. This is interesting because it has shown good performance in binary classification problems which it rarely have been applied to, as well as in the fields where it's effectiveness has been proven. Third, the LSTM algorithm seems to be unsuitable for binary classification problems because the training time is too long compared to the performance improvement. From these results, we can confirm that some of the deep learning algorithms can be applied to solve business binary classification problems.

Customer Behavior Prediction of Binary Classification Model Using Unstructured Information and Convolution Neural Network: The Case of Online Storefront (비정형 정보와 CNN 기법을 활용한 이진 분류 모델의 고객 행태 예측: 전자상거래 사례를 중심으로)

  • Kim, Seungsoo;Kim, Jongwoo
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.2
    • /
    • pp.221-241
    • /
    • 2018
  • Deep learning is getting attention recently. The deep learning technique which had been applied in competitions of the International Conference on Image Recognition Technology(ILSVR) and AlphaGo is Convolution Neural Network(CNN). CNN is characterized in that the input image is divided into small sections to recognize the partial features and combine them to recognize as a whole. Deep learning technologies are expected to bring a lot of changes in our lives, but until now, its applications have been limited to image recognition and natural language processing. The use of deep learning techniques for business problems is still an early research stage. If their performance is proved, they can be applied to traditional business problems such as future marketing response prediction, fraud transaction detection, bankruptcy prediction, and so on. So, it is a very meaningful experiment to diagnose the possibility of solving business problems using deep learning technologies based on the case of online shopping companies which have big data, are relatively easy to identify customer behavior and has high utilization values. Especially, in online shopping companies, the competition environment is rapidly changing and becoming more intense. Therefore, analysis of customer behavior for maximizing profit is becoming more and more important for online shopping companies. In this study, we propose 'CNN model of Heterogeneous Information Integration' using CNN as a way to improve the predictive power of customer behavior in online shopping enterprises. In order to propose a model that optimizes the performance, which is a model that learns from the convolution neural network of the multi-layer perceptron structure by combining structured and unstructured information, this model uses 'heterogeneous information integration', 'unstructured information vector conversion', 'multi-layer perceptron design', and evaluate the performance of each architecture, and confirm the proposed model based on the results. In addition, the target variables for predicting customer behavior are defined as six binary classification problems: re-purchaser, churn, frequent shopper, frequent refund shopper, high amount shopper, high discount shopper. In order to verify the usefulness of the proposed model, we conducted experiments using actual data of domestic specific online shopping company. This experiment uses actual transactions, customers, and VOC data of specific online shopping company in Korea. Data extraction criteria are defined for 47,947 customers who registered at least one VOC in January 2011 (1 month). The customer profiles of these customers, as well as a total of 19 months of trading data from September 2010 to March 2012, and VOCs posted for a month are used. The experiment of this study is divided into two stages. In the first step, we evaluate three architectures that affect the performance of the proposed model and select optimal parameters. We evaluate the performance with the proposed model. Experimental results show that the proposed model, which combines both structured and unstructured information, is superior compared to NBC(Naïve Bayes classification), SVM(Support vector machine), and ANN(Artificial neural network). Therefore, it is significant that the use of unstructured information contributes to predict customer behavior, and that CNN can be applied to solve business problems as well as image recognition and natural language processing problems. It can be confirmed through experiments that CNN is more effective in understanding and interpreting the meaning of context in text VOC data. And it is significant that the empirical research based on the actual data of the e-commerce company can extract very meaningful information from the VOC data written in the text format directly by the customer in the prediction of the customer behavior. Finally, through various experiments, it is possible to say that the proposed model provides useful information for the future research related to the parameter selection and its performance.

Development of New Variables Affecting Movie Success and Prediction of Weekly Box Office Using Them Based on Machine Learning (영화 흥행에 영향을 미치는 새로운 변수 개발과 이를 이용한 머신러닝 기반의 주간 박스오피스 예측)

  • Song, Junga;Choi, Keunho;Kim, Gunwoo
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.4
    • /
    • pp.67-83
    • /
    • 2018
  • The Korean film industry with significant increase every year exceeded the number of cumulative audiences of 200 million people in 2013 finally. However, starting from 2015 the Korean film industry entered a period of low growth and experienced a negative growth after all in 2016. To overcome such difficulty, stakeholders like production company, distribution company, multiplex have attempted to maximize the market returns using strategies of predicting change of market and of responding to such market change immediately. Since a film is classified as one of experiential products, it is not easy to predict a box office record and the initial number of audiences before the film is released. And also, the number of audiences fluctuates with a variety of factors after the film is released. So, the production company and distribution company try to be guaranteed the number of screens at the opining time of a newly released by multiplex chains. However, the multiplex chains tend to open the screening schedule during only a week and then determine the number of screening of the forthcoming week based on the box office record and the evaluation of audiences. Many previous researches have conducted to deal with the prediction of box office records of films. In the early stage, the researches attempted to identify factors affecting the box office record. And nowadays, many studies have tried to apply various analytic techniques to the factors identified previously in order to improve the accuracy of prediction and to explain the effect of each factor instead of identifying new factors affecting the box office record. However, most of previous researches have limitations in that they used the total number of audiences from the opening to the end as a target variable, and this makes it difficult to predict and respond to the demand of market which changes dynamically. Therefore, the purpose of this study is to predict the weekly number of audiences of a newly released film so that the stakeholder can flexibly and elastically respond to the change of the number of audiences in the film. To that end, we considered the factors used in the previous studies affecting box office and developed new factors not used in previous studies such as the order of opening of movies, dynamics of sales. Along with the comprehensive factors, we used the machine learning method such as Random Forest, Multi Layer Perception, Support Vector Machine, and Naive Bays, to predict the number of cumulative visitors from the first week after a film release to the third week. At the point of the first and the second week, we predicted the cumulative number of visitors of the forthcoming week for a released film. And at the point of the third week, we predict the total number of visitors of the film. In addition, we predicted the total number of cumulative visitors also at the point of the both first week and second week using the same factors. As a result, we found the accuracy of predicting the number of visitors at the forthcoming week was higher than that of predicting the total number of them in all of three weeks, and also the accuracy of the Random Forest was the highest among the machine learning methods we used. This study has implications in that this study 1) considered various factors comprehensively which affect the box office record and merely addressed by other previous researches such as the weekly rating of audiences after release, the weekly rank of the film after release, and the weekly sales share after release, and 2) tried to predict and respond to the demand of market which changes dynamically by suggesting models which predicts the weekly number of audiences of newly released films so that the stakeholders can flexibly and elastically respond to the change of the number of audiences in the film.

Biotope Mapping of Pinus densiflora Based on Growth Environment of Tricholoma matsutake - A Case Study of Yangyang-gun, Kang Won-do - (송이 생육환경 특성을 고려한 소나무비오톱지도 작성 연구 - 강원도 양양군을 사례로 -)

  • Han, Bong-Ho;Park, Seok-Cheol;Kwak, Jeong-In;Kim, Bo-Hyun;Lee, Kyong-Jae
    • Korean Journal of Environment and Ecology
    • /
    • v.25 no.2
    • /
    • pp.211-226
    • /
    • 2011
  • The purpose of this paper was to ensure the basis for effective management of Tricholoma matsutake mountain province, to perform biotope mapping of Pinus densiflora based on growth environment of Tricholoma matsutake, target a cluster of Yangyang-gun, Kang Won-do. Study Methods were to review on growth and environmental characteristics of Tricholoma matsutake through internal and external documents and to identify vegetational structure and soil characteristics. This paper studied growth structure and soil environment of Pinus densiflora forest where a farm of production area for Tricholoma matsutake of in order to set the standard of Pinus densiflora biotope. Mapping standards were derived by separating of landform conditions, soil conditions, vegetation conditions. Biotope types were divided into possible production area for Tricholoma matsutake and potential production area for Tricholoma matsutake, possible production area for Tricholoma matsutake were Pinus densiflora biotope in landform and soil structure that enables Tricholoma matsutake production and Single-layered Pinus densiflora biotope of less than 30cm(DBH)-Tree species that other shrub is dominant in shrub layer, Multi-layered Pinus densiflora biotope that Pinus densiflora forest was predominant in understrory layer. Potential production area for Tricholoma matsutake were single-layered Pinus densiflora biotope of more than 30cm(DBH) in landform that enables Tricholoma matsutake production, Pinus densiflora biotope with Quercus predominant in the understrory layer, single-layered Pinus densiflora biotope with Quercus predominant in shrub layer, inappropriate vegetation structure area that the induction of production of Tricholoma matsutake was possible through future vegetation management. According to the research results, Pinus densiflora forest were divided into 16 types; 6 types of possible Tricholoma matsutake production areas, 9 potential Tricholoma matsutake production areas and 16 types of areas where Tricholoma matsutake production was impossible. Possible production areas account for 15.48%, or $9.8km^2$ out of the total Pinus densiflora forest while potential production areas take up 32.42%, or $20.52km^2$, and areas where Tricholoma matsutake production was impossible was 52.10%, or $32.97km^2$.

Estimation of Jaw and MLC Transmission Factor Obtained by the Auto-modeling Process in the Pinnacle3 Treatment Planning System (피나클치료계획시스템에서 자동모델화과정으로 얻은 Jaw와 다엽콜리메이터의 투과 계수 평가)

  • Hwang, Tae-Jin;Kang, Sei-Kwon;Cheong, Kwang-Ho;Park, So-Ah;Lee, Me-Yeon;Kim, Kyoung-Ju;Oh, Do-Hoon;Bae, Hoon-Sik;Suh, Tae-Suk
    • Progress in Medical Physics
    • /
    • v.20 no.4
    • /
    • pp.269-276
    • /
    • 2009
  • Radiation treatment techniques using photon beam such as three-dimensional conformal radiation therapy (3D-CRT) as well as intensity modulated radiotherapy treatment (IMRT) demand accurate dose calculation in order to increase target coverage and spare healthy tissue. Both jaw collimator and multi-leaf collimators (MLCs) for photon beams have been used to achieve such goals. In the Pinnacle3 treatment planning system (TPS), which we are using in our clinics, a set of model parameters like jaw collimator transmission factor (JTF) and MLC transmission factor (MLCTF) are determined from the measured data because it is using a model-based photon dose algorithm. However, model parameters obtained by this auto-modeling process can be different from those by direct measurement, which can have a dosimetric effect on the dose distribution. In this paper we estimated JTF and MLCTF obtained by the auto-modeling process in the Pinnacle3 TPS. At first, we obtained JTF and MLCTF by direct measurement, which were the ratio of the output at the reference depth under the closed jaw collimator (MLCs for MLCTF) to that at the same depth with the field size $10{\times}10\;cm^2$ in the water phantom. And then JTF and MLCTF were also obtained by auto-modeling process. And we evaluated the dose difference through phantom and patient study in the 3D-CRT plan. For direct measurement, JTF was 0.001966 for 6 MV and 0.002971 for 10 MV, and MLCTF was 0.01657 for 6 MV and 0.01925 for 10 MV. On the other hand, for auto-modeling process, JTF was 0.001983 for 6 MV and 0.010431 for 10 MV, and MLCTF was 0.00188 for 6 MV and 0.00453 for 10 MV. JTF and MLCTF by direct measurement were very different from those by auto-modeling process and even more reasonable considering each beam quality of 6 MV and 10 MV. These different parameters affect the dose in the low-dose region. Since the wrong estimation of JTF and MLCTF can lead some dosimetric error, comparison of direct measurement and auto-modeling of JTF and MLCTF would be helpful during the beam commissioning.

  • PDF

IMAGING SIMULATIONS FOR THE KOREAN VLBI NETWORK(KVN) (한국우주전파관측망(KVN)의 영상모의실험)

  • Jung, Tae-Hyun;Rhee, Myung-Hyun;Roh, Duk-Gyoo;Kim, Hyun-Goo;Sohn, Bong-Won
    • Journal of Astronomy and Space Sciences
    • /
    • v.22 no.1
    • /
    • pp.1-12
    • /
    • 2005
  • The Korean VLBI Network (KVN) will open a new field of research in astronomy, geodesy and earth science using the newest three Elm radio telescopes. This will expand our ability to look at the Universe in the millimeter regime. Imaging capability of radio interferometry is highly dependent upon the antenna configuration, source size, declination and the shape of target. In this paper, imaging simulations are carried out with the KVN system configuration. Five test images were used which were a point source, multi-point sources, a uniform sphere with two different sizes compared to the synthesis beam of the KVN and a Very Large Array (VLA) image of Cygnus A. The declination for the full time simulation was set as +60 degrees and the observation time range was -6 to +6 hours around transit. Simulations have been done at 22GHz, one of the KVN observation frequency. All these simulations and data reductions have been run with the Astronomical Image Processing System (AIPS) software package. As the KVN array has a resolution of about 6 mas (milli arcsecond) at 220Hz, in case of model source being approximately the beam size or smaller, the ratio of peak intensity over RMS shows about 10000:1 and 5000:1. The other case in which model source is larger than the beam size, this ratio shows very low range of about 115:1 and 34:1. This is due to the lack of short baselines and the small number of antenna. We compare the coordinates of the model images with those of the cleaned images. The result shows mostly perfect correspondence except in the case of the 12mas uniform sphere. Therefore, the main astronomical targets for the KVN will be the compact sources and the KVN will have an excellent performance in the astrometry for these sources.

Selection of Insecticide Resistance Markers in Field-collected Populations of Myzus persicae (복숭아혹진딧물 야외개체군의 살충제 저항성 마커 선발)

  • Kim, Ju Il;Kwon, Min;Shim, Jae Dong;Kim, Jeom Soon;Lee, Yeong Gyu;Jee, Sam Nyu;Lee, Jeong Tae;Ryu, Jong Soo;Yoo, Dong Lim;Lee, Gye Jun
    • Korean journal of applied entomology
    • /
    • v.53 no.2
    • /
    • pp.149-156
    • /
    • 2014
  • The resistance levels of the green peach aphid, Myzus persicae (Sulzer), against 10 insecticides was checked and selected the applicable insecticide resistance markers. We conducted our study in 5 cabbage cultivation regions (Pyeongchang, Hongcheon, Bongwha, Muju, and Jeju) of Korea, over 3 successive years (2009-2011). We selected a multi-resistant (MR) strain from among the 5 field-collected populations. We analyzed esterase over-expression and mutation(s) in the target sites, by using native isoelectric focusing (IEF) and quantitative sequencing (QS). We detected esterase over-expression and StoF mutation in the acetylcholinesterase 1 gene (ace1) in all of the field-collected populations, including the MR strain. We did not detect the LtoF mutation, which is a well-known knockdown resistance (kdr) mutation in the para-type sodium channel gene (para), in the MR strain; however, the value of the MR strain for bifenthrin was 3,461-fold higher than that of the susceptible strain. Our results indicate that insecticide resistance is more effectively evaluated using molecular markers than by conducting a bioassay. The molecular markers StoF in ace1 and MtoL in para can easily be applied in diagnostic methods such as QS or PCR amplification of specific alleles (PASA). These methods may be extended to management of M. persicae resistance in the field.

Growth and yield components of rice under different NPK rates in Prateah Lang soil type in Cambodia

  • Kea, Kong;Sarom, Men;Vang, Seng;Kato, Yoichiro;Yamauchi, Akira;Ehara, Hiroshi
    • Proceedings of the Korean Society of Crop Science Conference
    • /
    • 2017.06a
    • /
    • pp.361-361
    • /
    • 2017
  • The NPK are known as macro elements that affect crop growth and yield. In 1989, Cambodia Agricultural Research and Development Institute (CARDI) gave a recommendation rate of fertilizer on rice production based on soil types. This recommended rate of NPK seems however relatively low as compared to farmers' practices nowadays and the amount in the neighboring countries. The CARDI recommended rate for Prateah Lang soil type is 50kg N, $25kg\;P_2O_5$, $25kg\;K_2O\;ha^{-1}$ while recent farmers' practice rates are 55 - 64kg N, 24 - 46kg $P_2O_5$, $30kg\;K_2O\;ha^{-1}$. However, the overuse of chemical fertilizer will lead to un-preferable plant growth, insect pest, disease and economic yield. Thus, we examined the effect of different NPK application rates on the growth and yield components in Prateah Lang soil type in Takeo province to investigate appropriate rates for improving rice productivity with economic efficiency. This study was conducted from July to November during wet season in 2013. A multi-locational trial with 6 treatments (T0 - T5) of NPK rates in 5 locations (trial 1 - 5) with 3 replications was conducted. The different combinations of NPK application were employed from 0, 50, 60, 80, 100, $120kg\;N\;ha^{-1}$, 0, 25, 30 45, $60kg\;P_2O_5\;ha^{-1}$ and 0, 15, 25, 30, $45kg\;K_2O\;ha^{-1}$. Urea, DAP and KCl were used for fertilization. Split application was employed [basal: 20% of N, 100% of P and K, top dressing-1st: 40% of N (30DAT), 2nd: 40% of N (PI stage)]. Three-week-old seedlings of var. Phka Rumdoul were transplanted with 2 - 3 seedlings $hill^{-1}$ with $20cm{\times}20cm$ spacing. Plant length, tiller number at the maximum tillering stage and yield components were measured. The different rates of NPK application affected some yield components. The panicle number per hill was the most important key component followed by the spikelet number per panicle. However, the other parameters such as the filled grain percentage and 1000 grains weight had small effect or weak relation with the yield. Although the panicle number per hill had a significantly positive correlation with the stem number per hill, it was not correlated with the percentage of productive culms. The variation in the grain yield among the 5 trials was small and the difference was not significant. Although the yield tended to be higher at higher N and P application, there was no significant difference above 60kg N and $30kg\;P_2O_5$. The yield was the highest at 15, 30 and $45kg\;K_2O$ followed by $25kg\;K_2O$. The relationships between N, P and the stem number per hill were significantly linear positive, though it was not linear between K and the stem number. From these results, to increase rice productivity in the target area, farmers' effort to increase N and P input rather than CARDI recommendation up to 60kg N and $30kg\;P_2O_5$ will be sufficient considering economic efficiency. Besides, the amount of K application should be reconsidered.

  • PDF