• Title/Summary/Keyword: basis vectors

Search Result 149, Processing Time 0.024 seconds

Cutaneous Microflora from Geographically Isolated Groups of Bradysia agrestis, an Insect Vector of Diverse Plant Pathogens

  • Park, Jong Myong;You, Young-Hyun;Park, Jong-Han;Kim, Hyeong-Hwan;Ghim, Sa-Youl;Back, Chang-Gi
    • Mycobiology
    • /
    • v.45 no.3
    • /
    • pp.160-171
    • /
    • 2017
  • Larvae of Bradysia agrestis, an insect vector that transports plant pathogens, were sampled from geographically isolated regions in Korea to identify their cutaneous fungal and bacterial flora. Sampled areas were chosen within the distribution range of B. agrestis; each site was more than 91 km apart to ensure geographical segregation. We isolated 76 microbial (fungi and bacteria) strains (site 1, 29; site 2, 29; site 3, 18 strains) that were identified on the basis of morphological differences. Species identification was molecularly confirmed by determination of universal fungal internal transcribed spacer and bacterial 16S rRNA gene sequences in comparison to sequences in the EzTaxon database and the NCBI GenBank database, and their phylogenetic relationships were determined. The fungal isolates belonged to 2 phyla, 5 classes, and 7 genera; bacterial species belonged to 23 genera and 32 species. Microbial diversity differed significantly among the geographical groups with respect to Margalef's richness (3.9, 3.6, and 4.5), Menhinick's index (2.65, 2.46, and 3.30), Simpson's index (0.06, 0.12, and 0.01), and Shannon's index (2.50, 2.17, and 2.58). Although the microbial genera distribution or diversity values clearly varied among geographical groups, common genera were identified in all groups, including the fungal genus Cladosporium, and the bacterial genera Bacillus and Rhodococcus. According to classic principles of co-evolutionary relationship, these genera might have a closer association with their host insect vector B. agrestis than other genera identified. Some cutaneous bacterial genera (e.g., Pseudomonas) displaying weak interdependency with insect vectors may be hazardous to agricultural environments via mechanical transmission via B. agrestis. This study provides comprehensive information regarding the cutaneous microflora of B. agrestis, which can help in the control of such pests for crop management.

P Element-Mediated Transformation with the rosy Gene in Drosophila melanogaster (D. melanogaster에 있어서 P Element를 이용한 rosy 유전자의 형질전환)

  • Kim, Wook;Kidwell, Margaret G.
    • The Korean Journal of Zoology
    • /
    • v.38 no.3
    • /
    • pp.340-347
    • /
    • 1995
  • We have used two kinds of P element constructs, Pc[(ry+)B] and p[(ry+)$\Delta$SX9], for genetic transformation by microinjection of D. melanogaster. Pc[(ry+)B] construct carrying the rosy gene within an autonomous P element was injected into a true M strain caring the ry506. mutation. The source of transposase for microinjection and transformation was provided by a P element helper plasmid designated p-$\Delta$2-3hs$\pi$, which was co-injected with nonautonomous P[(ry+)$\Delta$SX9] construct into same ry506 M strains. A dechorination method was adopted and 35 independent transformed lines were obtained froin 1143 G0 Injected (35/1143). About 20% of the injected embryos eclosed as adults. Among G0 eclosed flies, approximately 40% exhibited eye color that was similar to wild-type (ry+), but about 60% of fertile G0 transformed lines appeared to have no G1 transformants. Therefore it is unlikely that G0 expression requires integration of the rosy transposon into chromosomes. Pc[(ry+)B] and P[(ry+)$\Delta$SX9] constructs were found to be nearly same in the frequency of element-mediated transformation. On the basis of these results, nonautonomous P elements constructs could he used as same effective vectors in P element-mediated transformation for introducing and fixing genes in insect populations.

  • PDF

Regeneration of the Retarded Time Vector for Enhancing the Precision of Acoustic Pyrometry (온도장 측정 정밀도 향상을 위한 시간 지연 벡터의 재형성)

  • Kim, Tae-Kyoon;Ih, Jeong-Guon
    • The Journal of the Acoustical Society of Korea
    • /
    • v.33 no.2
    • /
    • pp.118-125
    • /
    • 2014
  • An approximation of speed of sound in the measurement plane is essential for the inverse estimation of temperature. To this end, an inverse problem relating the measured retarded time data in between set of sensors and actuators array located on the wall is formulated. The involved transfer matrix and its coefficient vectors approximate speed of sound of the measurement plane by using the radial basis function with finite number of interpolation points deployed inside the target field. Then, the temperature field can be reconstructed by using spatial interpolation technique, which can achieve high spatial resolution with proper number of interpolation points. A large number of retarded time data of acoustic paths in between sensors and arrays are needed to obtain accurate reconstruction result. However, the shortage of interpolation points due to practical limitations can cause the decrease of spatial resolution and deterioration of the reconstruction result. In this works, a regeneration for obtaining the additional retarded time data for an arbitrary acoustic path is suggested to overcome the shortage of interpolation points. By applying the regeneration technique, many interpolation points can be deployed inside the field by increasing the number of retarded time data. As a simulation example, two rectangular duct sections having arbitrary temperature distribution are reconstructed by two different data set: measured data only, combination of measured and regenerated data. The result shows a decrease in reconstruction error by 15 % by combining the original and regenerated retarded time data.

A study of using quality for Radial Basis Function based score-level fusion in multimodal biometrics (RBF 기반 유사도 단계 융합 다중 생체 인식에서의 품질 활용 방안 연구)

  • Choi, Hyun-Soek;Shin, Mi-Young
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.45 no.5
    • /
    • pp.192-200
    • /
    • 2008
  • Multimodal biometrics is a method for personal authentication and verification using more than two types of biometrics data. RBF based score-level fusion uses pattern recognition algorithm for multimodal biometrics, seeking the optimal decision boundary to classify score feature vectors each of which consists of matching scores obtained from several unimodal biometrics system for each sample. In this case, all matching scores are assumed to have the same reliability. However, in recent research it is reported that the quality of input sample affects the result of biometrics. Currently the matching scores having low reliability caused by low quality of samples are not currently considered for pattern recognition modelling in multimodal biometrics. To solve this problem, in this paper, we proposed the RBF based score-level fusion approach which employs quality information of input biometrics data to adjust decision boundary. As a result the proposed method with Qualify information showed better recognition performance than both the unimodal biometrics and the usual RBF based score-level fusion without using quality information.

A Feature Map Compression Method for Multi-resolution Feature Map with PCA-based Transformation (PCA 기반 변환을 통한 다해상도 피처 맵 압축 방법)

  • Park, Seungjin;Lee, Minhun;Choi, Hansol;Kim, Minsub;Oh, Seoung-Jun;Kim, Younhee;Do, Jihoon;Jeong, Se Yoon;Sim, Donggyu
    • Journal of Broadcast Engineering
    • /
    • v.27 no.1
    • /
    • pp.56-68
    • /
    • 2022
  • In this paper, we propose a compression method for multi-resolution feature maps for VCM. The proposed compression method removes the redundancy between the channels and resolution levels of the multi-resolution feature map through PCA-based transformation. According to each characteristic, the basis vectors and mean vector used for transformation, and the transformation coefficient obtained through the transformation are compressed using a VVC-based coder and DeepCABAC. In order to evaluate performance of the proposed method, the object detection performance was measured for the OpenImageV6 and COCO 2017 validation set, and the BD-rate of MPEG-VCM anchor and feature map compression anchor proposed in this paper was compared using bpp and mAP. As a result of the experiment, the proposed method shows a 25.71% BD-rate performance improvement compared to feature map compression anchor in OpenImageV6. Furthermore, for large objects of the COCO 2017 validation set, the BD-rate performance is improved by up to 43.72% compared to the MPEG-VCM anchor.

The Prediction of DEA based Efficiency Rating for Venture Business Using Multi-class SVM (다분류 SVM을 이용한 DEA기반 벤처기업 효율성등급 예측모형)

  • Park, Ji-Young;Hong, Tae-Ho
    • Asia pacific journal of information systems
    • /
    • v.19 no.2
    • /
    • pp.139-155
    • /
    • 2009
  • For the last few decades, many studies have tried to explore and unveil venture companies' success factors and unique features in order to identify the sources of such companies' competitive advantages over their rivals. Such venture companies have shown tendency to give high returns for investors generally making the best use of information technology. For this reason, many venture companies are keen on attracting avid investors' attention. Investors generally make their investment decisions by carefully examining the evaluation criteria of the alternatives. To them, credit rating information provided by international rating agencies, such as Standard and Poor's, Moody's and Fitch is crucial source as to such pivotal concerns as companies stability, growth, and risk status. But these types of information are generated only for the companies issuing corporate bonds, not venture companies. Therefore, this study proposes a method for evaluating venture businesses by presenting our recent empirical results using financial data of Korean venture companies listed on KOSDAQ in Korea exchange. In addition, this paper used multi-class SVM for the prediction of DEA-based efficiency rating for venture businesses, which was derived from our proposed method. Our approach sheds light on ways to locate efficient companies generating high level of profits. Above all, in determining effective ways to evaluate a venture firm's efficiency, it is important to understand the major contributing factors of such efficiency. Therefore, this paper is constructed on the basis of following two ideas to classify which companies are more efficient venture companies: i) making DEA based multi-class rating for sample companies and ii) developing multi-class SVM-based efficiency prediction model for classifying all companies. First, the Data Envelopment Analysis(DEA) is a non-parametric multiple input-output efficiency technique that measures the relative efficiency of decision making units(DMUs) using a linear programming based model. It is non-parametric because it requires no assumption on the shape or parameters of the underlying production function. DEA has been already widely applied for evaluating the relative efficiency of DMUs. Recently, a number of DEA based studies have evaluated the efficiency of various types of companies, such as internet companies and venture companies. It has been also applied to corporate credit ratings. In this study we utilized DEA for sorting venture companies by efficiency based ratings. The Support Vector Machine(SVM), on the other hand, is a popular technique for solving data classification problems. In this paper, we employed SVM to classify the efficiency ratings in IT venture companies according to the results of DEA. The SVM method was first developed by Vapnik (1995). As one of many machine learning techniques, SVM is based on a statistical theory. Thus far, the method has shown good performances especially in generalizing capacity in classification tasks, resulting in numerous applications in many areas of business, SVM is basically the algorithm that finds the maximum margin hyperplane, which is the maximum separation between classes. According to this method, support vectors are the closest to the maximum margin hyperplane. If it is impossible to classify, we can use the kernel function. In the case of nonlinear class boundaries, we can transform the inputs into a high-dimensional feature space, This is the original input space and is mapped into a high-dimensional dot-product space. Many studies applied SVM to the prediction of bankruptcy, the forecast a financial time series, and the problem of estimating credit rating, In this study we employed SVM for developing data mining-based efficiency prediction model. We used the Gaussian radial function as a kernel function of SVM. In multi-class SVM, we adopted one-against-one approach between binary classification method and two all-together methods, proposed by Weston and Watkins(1999) and Crammer and Singer(2000), respectively. In this research, we used corporate information of 154 companies listed on KOSDAQ market in Korea exchange. We obtained companies' financial information of 2005 from the KIS(Korea Information Service, Inc.). Using this data, we made multi-class rating with DEA efficiency and built multi-class prediction model based data mining. Among three manners of multi-classification, the hit ratio of the Weston and Watkins method is the best in the test data set. In multi classification problems as efficiency ratings of venture business, it is very useful for investors to know the class with errors, one class difference, when it is difficult to find out the accurate class in the actual market. So we presented accuracy results within 1-class errors, and the Weston and Watkins method showed 85.7% accuracy in our test samples. We conclude that the DEA based multi-class approach in venture business generates more information than the binary classification problem, notwithstanding its efficiency level. We believe this model can help investors in decision making as it provides a reliably tool to evaluate venture companies in the financial domain. For the future research, we perceive the need to enhance such areas as the variable selection process, the parameter selection of kernel function, the generalization, and the sample size of multi-class.

Enhanced production of monacolin-K through supplement of monacolin-K precursors into production medium and cloning of SAM synthetase gene (metK) (Precursor제공 및 생합성 관련 유전자의 cloning을 통한 Monacolin-K 생산성 향상)

  • Lee, Mi-Jin;Jeong, Yong-Seob;Chun, Gie-Taek
    • KSBB Journal
    • /
    • v.23 no.6
    • /
    • pp.519-524
    • /
    • 2008
  • Monacolin-K is a strong anti-hypercholesterolemic agent produced by Monascus sp. via polyketide pathway. High-yielding mutants of monacolin-K were developed through rational screening strategies adopted based on understanding of monacolin-K biosynthetic pathway. Through the experiments for investigating various amino acids as putative precursors for the monacolin-K biosynthesis, it was found that production level of monacolin-K was remarkably increased when optimum amount of cysteine was supplemented into the production medium. We suggested that these phenomena might be related to the special roles of SAM (S-adenosyl methionine), a putative methyl group donor in the biosynthetic pathway of monacolin-K, demonstrating close interrelationship between SAM-synthesizing primary metabolism and monacolin-K synthesizing secondary metabolism. Namely, increase in the intracellular amount of SAM derived from the putative precursor, cysteine which was extracellularly supplemented into the production medium might contribute to the significant enhancement in the monacolin-K biosynthetic capability of the highly mutated producers. On the basis of these assumptions derived from the above fermentation results, we decided to construct efficient expression vectors harboring SAM synthetase gene (metK) cloned from A. nidulans, with the hope that increased intracellular level of SAM could lead to further enhancement in the monacolin-K production through overcoming a rate-limiting step associated with monacolin-K biosynthesis. Hence, in order to overcome the plausible rate-limiting step associated with monacolin-K biosynthesis by increasing intracellular level of SAM, we transformed the producer mutants with an efficient expression vector harboring gpdA promoter of the producer microorganism, and metK gene. Notably, from the resulting various transformants, we were able to screen a very high-yielding transformant which showed approximately 3.3 fold higher monacolin-K productivity than the parallel nontransformed mutants in shake flask cultures performed under the identical fermentation conditions.

Korean Sentence Generation Using Phoneme-Level LSTM Language Model (한국어 음소 단위 LSTM 언어모델을 이용한 문장 생성)

  • Ahn, SungMahn;Chung, Yeojin;Lee, Jaejoon;Yang, Jiheon
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.2
    • /
    • pp.71-88
    • /
    • 2017
  • Language models were originally developed for speech recognition and language processing. Using a set of example sentences, a language model predicts the next word or character based on sequential input data. N-gram models have been widely used but this model cannot model the correlation between the input units efficiently since it is a probabilistic model which are based on the frequency of each unit in the training set. Recently, as the deep learning algorithm has been developed, a recurrent neural network (RNN) model and a long short-term memory (LSTM) model have been widely used for the neural language model (Ahn, 2016; Kim et al., 2016; Lee et al., 2016). These models can reflect dependency between the objects that are entered sequentially into the model (Gers and Schmidhuber, 2001; Mikolov et al., 2010; Sundermeyer et al., 2012). In order to learning the neural language model, texts need to be decomposed into words or morphemes. Since, however, a training set of sentences includes a huge number of words or morphemes in general, the size of dictionary is very large and so it increases model complexity. In addition, word-level or morpheme-level models are able to generate vocabularies only which are contained in the training set. Furthermore, with highly morphological languages such as Turkish, Hungarian, Russian, Finnish or Korean, morpheme analyzers have more chance to cause errors in decomposition process (Lankinen et al., 2016). Therefore, this paper proposes a phoneme-level language model for Korean language based on LSTM models. A phoneme such as a vowel or a consonant is the smallest unit that comprises Korean texts. We construct the language model using three or four LSTM layers. Each model was trained using Stochastic Gradient Algorithm and more advanced optimization algorithms such as Adagrad, RMSprop, Adadelta, Adam, Adamax, and Nadam. Simulation study was done with Old Testament texts using a deep learning package Keras based the Theano. After pre-processing the texts, the dataset included 74 of unique characters including vowels, consonants, and punctuation marks. Then we constructed an input vector with 20 consecutive characters and an output with a following 21st character. Finally, total 1,023,411 sets of input-output vectors were included in the dataset and we divided them into training, validation, testsets with proportion 70:15:15. All the simulation were conducted on a system equipped with an Intel Xeon CPU (16 cores) and a NVIDIA GeForce GTX 1080 GPU. We compared the loss function evaluated for the validation set, the perplexity evaluated for the test set, and the time to be taken for training each model. As a result, all the optimization algorithms but the stochastic gradient algorithm showed similar validation loss and perplexity, which are clearly superior to those of the stochastic gradient algorithm. The stochastic gradient algorithm took the longest time to be trained for both 3- and 4-LSTM models. On average, the 4-LSTM layer model took 69% longer training time than the 3-LSTM layer model. However, the validation loss and perplexity were not improved significantly or became even worse for specific conditions. On the other hand, when comparing the automatically generated sentences, the 4-LSTM layer model tended to generate the sentences which are closer to the natural language than the 3-LSTM model. Although there were slight differences in the completeness of the generated sentences between the models, the sentence generation performance was quite satisfactory in any simulation conditions: they generated only legitimate Korean letters and the use of postposition and the conjugation of verbs were almost perfect in the sense of grammar. The results of this study are expected to be widely used for the processing of Korean language in the field of language processing and speech recognition, which are the basis of artificial intelligence systems.

A Study on Knowledge Entity Extraction Method for Individual Stocks Based on Neural Tensor Network (뉴럴 텐서 네트워크 기반 주식 개별종목 지식개체명 추출 방법에 관한 연구)

  • Yang, Yunseok;Lee, Hyun Jun;Oh, Kyong Joo
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.2
    • /
    • pp.25-38
    • /
    • 2019
  • Selecting high-quality information that meets the interests and needs of users among the overflowing contents is becoming more important as the generation continues. In the flood of information, efforts to reflect the intention of the user in the search result better are being tried, rather than recognizing the information request as a simple string. Also, large IT companies such as Google and Microsoft focus on developing knowledge-based technologies including search engines which provide users with satisfaction and convenience. Especially, the finance is one of the fields expected to have the usefulness and potential of text data analysis because it's constantly generating new information, and the earlier the information is, the more valuable it is. Automatic knowledge extraction can be effective in areas where information flow is vast, such as financial sector, and new information continues to emerge. However, there are several practical difficulties faced by automatic knowledge extraction. First, there are difficulties in making corpus from different fields with same algorithm, and it is difficult to extract good quality triple. Second, it becomes more difficult to produce labeled text data by people if the extent and scope of knowledge increases and patterns are constantly updated. Third, performance evaluation is difficult due to the characteristics of unsupervised learning. Finally, problem definition for automatic knowledge extraction is not easy because of ambiguous conceptual characteristics of knowledge. So, in order to overcome limits described above and improve the semantic performance of stock-related information searching, this study attempts to extract the knowledge entity by using neural tensor network and evaluate the performance of them. Different from other references, the purpose of this study is to extract knowledge entity which is related to individual stock items. Various but relatively simple data processing methods are applied in the presented model to solve the problems of previous researches and to enhance the effectiveness of the model. From these processes, this study has the following three significances. First, A practical and simple automatic knowledge extraction method that can be applied. Second, the possibility of performance evaluation is presented through simple problem definition. Finally, the expressiveness of the knowledge increased by generating input data on a sentence basis without complex morphological analysis. The results of the empirical analysis and objective performance evaluation method are also presented. The empirical study to confirm the usefulness of the presented model, experts' reports about individual 30 stocks which are top 30 items based on frequency of publication from May 30, 2017 to May 21, 2018 are used. the total number of reports are 5,600, and 3,074 reports, which accounts about 55% of the total, is designated as a training set, and other 45% of reports are designated as a testing set. Before constructing the model, all reports of a training set are classified by stocks, and their entities are extracted using named entity recognition tool which is the KKMA. for each stocks, top 100 entities based on appearance frequency are selected, and become vectorized using one-hot encoding. After that, by using neural tensor network, the same number of score functions as stocks are trained. Thus, if a new entity from a testing set appears, we can try to calculate the score by putting it into every single score function, and the stock of the function with the highest score is predicted as the related item with the entity. To evaluate presented models, we confirm prediction power and determining whether the score functions are well constructed by calculating hit ratio for all reports of testing set. As a result of the empirical study, the presented model shows 69.3% hit accuracy for testing set which consists of 2,526 reports. this hit ratio is meaningfully high despite of some constraints for conducting research. Looking at the prediction performance of the model for each stocks, only 3 stocks, which are LG ELECTRONICS, KiaMtr, and Mando, show extremely low performance than average. this result maybe due to the interference effect with other similar items and generation of new knowledge. In this paper, we propose a methodology to find out key entities or their combinations which are necessary to search related information in accordance with the user's investment intention. Graph data is generated by using only the named entity recognition tool and applied to the neural tensor network without learning corpus or word vectors for the field. From the empirical test, we confirm the effectiveness of the presented model as described above. However, there also exist some limits and things to complement. Representatively, the phenomenon that the model performance is especially bad for only some stocks shows the need for further researches. Finally, through the empirical study, we confirmed that the learning method presented in this study can be used for the purpose of matching the new text information semantically with the related stocks.