• Title/Summary/Keyword: Complex vector

Search Result 622, Processing Time 0.026 seconds

A modified U-net for crack segmentation by Self-Attention-Self-Adaption neuron and random elastic deformation

  • Zhao, Jin;Hu, Fangqiao;Qiao, Weidong;Zhai, Weida;Xu, Yang;Bao, Yuequan;Li, Hui
    • Smart Structures and Systems
    • /
    • v.29 no.1
    • /
    • pp.1-16
    • /
    • 2022
  • Despite recent breakthroughs in deep learning and computer vision fields, the pixel-wise identification of tiny objects in high-resolution images with complex disturbances remains challenging. This study proposes a modified U-net for tiny crack segmentation in real-world steel-box-girder bridges. The modified U-net adopts the common U-net framework and a novel Self-Attention-Self-Adaption (SASA) neuron as the fundamental computing element. The Self-Attention module applies softmax and gate operations to obtain the attention vector. It enables the neuron to focus on the most significant receptive fields when processing large-scale feature maps. The Self-Adaption module consists of a multiplayer perceptron subnet and achieves deeper feature extraction inside a single neuron. For data augmentation, a grid-based crack random elastic deformation (CRED) algorithm is designed to enrich the diversities and irregular shapes of distributed cracks. Grid-based uniform control nodes are first set on both input images and binary labels, random offsets are then employed on these control nodes, and bilinear interpolation is performed for the rest pixels. The proposed SASA neuron and CRED algorithm are simultaneously deployed to train the modified U-net. 200 raw images with a high resolution of 4928 × 3264 are collected, 160 for training and the rest 40 for the test. 512 × 512 patches are generated from the original images by a sliding window with an overlap of 256 as inputs. Results show that the average IoU between the recognized and ground-truth cracks reaches 0.409, which is 29.8% higher than the regular U-net. A five-fold cross-validation study is performed to verify that the proposed method is robust to different training and test images. Ablation experiments further demonstrate the effectiveness of the proposed SASA neuron and CRED algorithm. Promotions of the average IoU individually utilizing the SASA and CRED module add up to the final promotion of the full model, indicating that the SASA and CRED modules contribute to the different stages of model and data in the training process.

Research Trends in Record Management Using Unstructured Text Data Analysis (비정형 텍스트 데이터 분석을 활용한 기록관리 분야 연구동향)

  • Deokyong Hong;Junseok Heo
    • Journal of Korean Society of Archives and Records Management
    • /
    • v.23 no.4
    • /
    • pp.73-89
    • /
    • 2023
  • This study aims to analyze the frequency of keywords used in Korean abstracts, which are unstructured text data in the domestic record management research field, using text mining techniques to identify domestic record management research trends through distance analysis between keywords. To this end, 1,157 keywords of 77,578 journals were visualized by extracting 1,157 articles from 7 journal types (28 types) searched by major category (complex study) and middle category (literature informatics) from the institutional statistics (registered site, candidate site) of the Korean Citation Index (KCI). Analysis of t-Distributed Stochastic Neighbor Embedding (t-SNE) and Scattertext using Word2vec was performed. As a result of the analysis, first, it was confirmed that keywords such as "record management" (889 times), "analysis" (888 times), "archive" (742 times), "record" (562 times), and "utilization" (449 times) were treated as significant topics by researchers. Second, Word2vec analysis generated vector representations between keywords, and similarity distances were investigated and visualized using t-SNE and Scattertext. In the visualization results, the research area for record management was divided into two groups, with keywords such as "archiving," "national record management," "standardization," "official documents," and "record management systems" occurring frequently in the first group (past). On the other hand, keywords such as "community," "data," "record information service," "online," and "digital archives" in the second group (current) were garnering substantial focus.

Comparison of the Vertical Data between Eulerian and Lagrangian Method (오일러와 라그랑주 관측방식의 연직 자료 비교)

  • Hyeok-Jin Bae;Byung Hyuk Kwon;Sang Jin Kim;Kyung-Hun Lee;Geon-Myeong Lee;Yu-Jin Kim;Ji-Woo Seo;Yu-Jung Koo
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.18 no.6
    • /
    • pp.1009-1014
    • /
    • 2023
  • Comprehensive observations of the Euler method and the Lagrangian method were performed in order to obtain high-resolution observation data in space and time for the complex environment of new city. The two radiosondes, which measure meteorological parameters using Lagrangian methods, produced air pressure, wind speed and wind direction. They were generally consistent with each other even if the observation points or times were different. The temperature measured by the sensor exposed to the air during the day was relatively high as the altitude increased due to the influence of solar radiation. The temporal difference in wind direction and speed was found in the comparison of Euler's wind profiler data with radiosonde data. When the wind field is horizontally in homogeneous, this result implies the need to consider the advection component to compare the data of the two observation methods. In this study, a method of using observation data at different times for each altitude section depending on the observation period of the Euler method is proposed to effectively compare the data of the two observation methods.

A Study of Textured Image Segmentation using Phase Information (페이즈 정보를 이용한 텍스처 영상 분할 연구)

  • Oh, Suk
    • Journal of the Korea Society of Computer and Information
    • /
    • v.16 no.2
    • /
    • pp.249-256
    • /
    • 2011
  • Finding a new set of features representing textured images is one of the most important studies in textured image analysis. This is because it is impossible to construct a perfect set of features representing every textured image, and it is inevitable to choose some relevant features which are efficient to on-going image processing jobs. This paper intends to find relevant features which are efficient to textured image segmentation. In this regards, this paper presents a different method for the segmentation of textured images based on the Gabor filter. Gabor filter is known to be a very efficient and effective tool which represents human visual system for texture analysis. Filtering a real-valued input image by the Gabor filter results in complex-valued output data defined in the spatial frequency domain. This complex value, as usual, gives the module and the phase. This paper focused its attention on the phase information, rather than the module information. In fact, the module information is considered very useful at region analysis in texture, while the phase information was considered almost of no use. But this paper shows that the phase information can also be fully useful and effective at region analysis in texture, once a good method introduced. We now propose "phase derivated method", which is an efficient and effective way to compute the useful phase information directly from the filtered value. This new method reduces effectively computing burden and widen applicable textured images.

Estimation of GARCH Models and Performance Analysis of Volatility Trading System using Support Vector Regression (Support Vector Regression을 이용한 GARCH 모형의 추정과 투자전략의 성과분석)

  • Kim, Sun Woong;Choi, Heung Sik
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.2
    • /
    • pp.107-122
    • /
    • 2017
  • Volatility in the stock market returns is a measure of investment risk. It plays a central role in portfolio optimization, asset pricing and risk management as well as most theoretical financial models. Engle(1982) presented a pioneering paper on the stock market volatility that explains the time-variant characteristics embedded in the stock market return volatility. His model, Autoregressive Conditional Heteroscedasticity (ARCH), was generalized by Bollerslev(1986) as GARCH models. Empirical studies have shown that GARCH models describes well the fat-tailed return distributions and volatility clustering phenomenon appearing in stock prices. The parameters of the GARCH models are generally estimated by the maximum likelihood estimation (MLE) based on the standard normal density. But, since 1987 Black Monday, the stock market prices have become very complex and shown a lot of noisy terms. Recent studies start to apply artificial intelligent approach in estimating the GARCH parameters as a substitute for the MLE. The paper presents SVR-based GARCH process and compares with MLE-based GARCH process to estimate the parameters of GARCH models which are known to well forecast stock market volatility. Kernel functions used in SVR estimation process are linear, polynomial and radial. We analyzed the suggested models with KOSPI 200 Index. This index is constituted by 200 blue chip stocks listed in the Korea Exchange. We sampled KOSPI 200 daily closing values from 2010 to 2015. Sample observations are 1487 days. We used 1187 days to train the suggested GARCH models and the remaining 300 days were used as testing data. First, symmetric and asymmetric GARCH models are estimated by MLE. We forecasted KOSPI 200 Index return volatility and the statistical metric MSE shows better results for the asymmetric GARCH models such as E-GARCH or GJR-GARCH. This is consistent with the documented non-normal return distribution characteristics with fat-tail and leptokurtosis. Compared with MLE estimation process, SVR-based GARCH models outperform the MLE methodology in KOSPI 200 Index return volatility forecasting. Polynomial kernel function shows exceptionally lower forecasting accuracy. We suggested Intelligent Volatility Trading System (IVTS) that utilizes the forecasted volatility results. IVTS entry rules are as follows. If forecasted tomorrow volatility will increase then buy volatility today. If forecasted tomorrow volatility will decrease then sell volatility today. If forecasted volatility direction does not change we hold the existing buy or sell positions. IVTS is assumed to buy and sell historical volatility values. This is somewhat unreal because we cannot trade historical volatility values themselves. But our simulation results are meaningful since the Korea Exchange introduced volatility futures contract that traders can trade since November 2014. The trading systems with SVR-based GARCH models show higher returns than MLE-based GARCH in the testing period. And trading profitable percentages of MLE-based GARCH IVTS models range from 47.5% to 50.0%, trading profitable percentages of SVR-based GARCH IVTS models range from 51.8% to 59.7%. MLE-based symmetric S-GARCH shows +150.2% return and SVR-based symmetric S-GARCH shows +526.4% return. MLE-based asymmetric E-GARCH shows -72% return and SVR-based asymmetric E-GARCH shows +245.6% return. MLE-based asymmetric GJR-GARCH shows -98.7% return and SVR-based asymmetric GJR-GARCH shows +126.3% return. Linear kernel function shows higher trading returns than radial kernel function. Best performance of SVR-based IVTS is +526.4% and that of MLE-based IVTS is +150.2%. SVR-based GARCH IVTS shows higher trading frequency. This study has some limitations. Our models are solely based on SVR. Other artificial intelligence models are needed to search for better performance. We do not consider costs incurred in the trading process including brokerage commissions and slippage costs. IVTS trading performance is unreal since we use historical volatility values as trading objects. The exact forecasting of stock market volatility is essential in the real trading as well as asset pricing models. Further studies on other machine learning-based GARCH models can give better information for the stock market investors.

Sentiment Analysis of Movie Review Using Integrated CNN-LSTM Mode (CNN-LSTM 조합모델을 이용한 영화리뷰 감성분석)

  • Park, Ho-yeon;Kim, Kyoung-jae
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.4
    • /
    • pp.141-154
    • /
    • 2019
  • Rapid growth of internet technology and social media is progressing. Data mining technology has evolved to enable unstructured document representations in a variety of applications. Sentiment analysis is an important technology that can distinguish poor or high-quality content through text data of products, and it has proliferated during text mining. Sentiment analysis mainly analyzes people's opinions in text data by assigning predefined data categories as positive and negative. This has been studied in various directions in terms of accuracy from simple rule-based to dictionary-based approaches using predefined labels. In fact, sentiment analysis is one of the most active researches in natural language processing and is widely studied in text mining. When real online reviews aren't available for others, it's not only easy to openly collect information, but it also affects your business. In marketing, real-world information from customers is gathered on websites, not surveys. Depending on whether the website's posts are positive or negative, the customer response is reflected in the sales and tries to identify the information. However, many reviews on a website are not always good, and difficult to identify. The earlier studies in this research area used the reviews data of the Amazon.com shopping mal, but the research data used in the recent studies uses the data for stock market trends, blogs, news articles, weather forecasts, IMDB, and facebook etc. However, the lack of accuracy is recognized because sentiment calculations are changed according to the subject, paragraph, sentiment lexicon direction, and sentence strength. This study aims to classify the polarity analysis of sentiment analysis into positive and negative categories and increase the prediction accuracy of the polarity analysis using the pretrained IMDB review data set. First, the text classification algorithm related to sentiment analysis adopts the popular machine learning algorithms such as NB (naive bayes), SVM (support vector machines), XGboost, RF (random forests), and Gradient Boost as comparative models. Second, deep learning has demonstrated discriminative features that can extract complex features of data. Representative algorithms are CNN (convolution neural networks), RNN (recurrent neural networks), LSTM (long-short term memory). CNN can be used similarly to BoW when processing a sentence in vector format, but does not consider sequential data attributes. RNN can handle well in order because it takes into account the time information of the data, but there is a long-term dependency on memory. To solve the problem of long-term dependence, LSTM is used. For the comparison, CNN and LSTM were chosen as simple deep learning models. In addition to classical machine learning algorithms, CNN, LSTM, and the integrated models were analyzed. Although there are many parameters for the algorithms, we examined the relationship between numerical value and precision to find the optimal combination. And, we tried to figure out how the models work well for sentiment analysis and how these models work. This study proposes integrated CNN and LSTM algorithms to extract the positive and negative features of text analysis. The reasons for mixing these two algorithms are as follows. CNN can extract features for the classification automatically by applying convolution layer and massively parallel processing. LSTM is not capable of highly parallel processing. Like faucets, the LSTM has input, output, and forget gates that can be moved and controlled at a desired time. These gates have the advantage of placing memory blocks on hidden nodes. The memory block of the LSTM may not store all the data, but it can solve the CNN's long-term dependency problem. Furthermore, when LSTM is used in CNN's pooling layer, it has an end-to-end structure, so that spatial and temporal features can be designed simultaneously. In combination with CNN-LSTM, 90.33% accuracy was measured. This is slower than CNN, but faster than LSTM. The presented model was more accurate than other models. In addition, each word embedding layer can be improved when training the kernel step by step. CNN-LSTM can improve the weakness of each model, and there is an advantage of improving the learning by layer using the end-to-end structure of LSTM. Based on these reasons, this study tries to enhance the classification accuracy of movie reviews using the integrated CNN-LSTM model.

Overexpression of TMP21 Could Induce not only Downregulation of TrkA/ERK Phosphorylation but also Upregulation of p75NTR/RhoA Expression on NGF Receptor Signaling Pathway (γ-Secretase 활성억제단백질인 TMP21의 과발현이 신경세포주에서 NGF 수용체 신호전달과정에 미치는 영향)

  • Choi, Sun-Il;Jee, Seung-Wan;Her, Youn-Kyung;Kim, Ji-Eun;Nam, So-Hee;Hwang, In-Sik;Lee, Hye-Ryun;Goo, Jun-Seo;Lee, Young-Ju;Lee, Eon-Pil;Choi, Hae-Wook;Kim, Hong-Sung;Lee, Jae-Ho;Jung, Young-Jin;Lee, Su-Hae;Shim, Sun-Bo;Hwang, Dae-Youn
    • Journal of Life Science
    • /
    • v.21 no.8
    • /
    • pp.1134-1141
    • /
    • 2011
  • Transmembrane protein 21 (TMP21) is a member of the p24 cargo protein family and has been shown to modulate ${\alpha}$-secretase-mediated A${\beta}$ production which was specifically observed in the brains of subjects with Alzheimer's disease (AD). In order to investigate whether TMP21 could affect nerve growth factor (NGF) receptor signaling pathway, the alteration of NGF receptors and their downstream proteins were detected in TMP21 over-expressed cells. CMV/hTMP21 vector used in this study was successfully expressed into TMP21 proteins in B35 cells after lipofectamin transfection. Expressed TMP21 proteins induced the down-regulation of ${\gamma}$-secretase complex components including Presenlin-1 (PS-1), PS-2, Nicastrin (NST), Pen-2 and APH-1. Also, the expression level of NGF receptor $p75^{NTR}$ and RhoA were significantly higher in CMV/hTMP21 transfectants than vehicle transfectants, while their levels returned to vehicle levels after NGF treatment. However, the phosphorylation of NGF receptor TrkA was dramtically decreased in NGF No-treated CMV/hTMP21 transfectants compared with vehicle transfectants, and increased in NGF treated CMV/hTMP21 transfectants. In TrkA downstream signaling pathway, the phosphorylation level of ERK was also decreased in CMV/hTMP21 transfectants, while the phosphorylation of Akt was increased in the same transfectants. Furthermore, NGF treatment induced the increase of phosphorylation level of Akt and ERK in CMV/hTMP21 transfectants. Therefore, these results suggested that over-expression of TMP21may simultaneously induce the up-regulation of $p75^{NTR}$/RhoA expression and the down-regulation of TrkA/ERK phosphorylation through the inhibition of ${\gamma}$-secretase activity.

On Method for LBS Multi-media Services using GML 3.0 (GML 3.0을 이용한 LBS 멀티미디어 서비스에 관한 연구)

  • Jung, Kee-Joong;Lee, Jun-Woo;Kim, Nam-Gyun;Hong, Seong-Hak;Choi, Beyung-Nam
    • 한국공간정보시스템학회:학술대회논문집
    • /
    • 2004.12a
    • /
    • pp.169-181
    • /
    • 2004
  • SK Telecom has already constructed GIMS system as the base common framework of LBS/GIS service system based on OGC(OpenGIS Consortium)'s international standard for the first mobile vector map service in 2002, But as service content appears more complex, renovation has been needed to satisfy multi-purpose, multi-function and maximum efficiency as requirements have been increased. This research is for preparation ion of GML3-based platform to upgrade service from GML2 based GIMS system. And with this, it will be possible for variety of application services to provide location and geographic data easily and freely. In GML 3.0, it has been selected animation, event handling, resource for style mapping, topology specification for 3D and telematics services for mobile LBS multimedia service. And the schema and transfer protocol has been developed and organized to optimize data transfer to MS(Mobile Stat ion) Upgrade to GML 3.0-based GIMS system has provided innovative framework in the view of not only construction but also service which has been implemented and applied to previous research and system. Also GIMS channel interface has been implemented to simplify access to GIMS system, and service component of GIMS internals, WFS and WMS, has gotten enhanded and expanded function.

  • PDF

Amino Acid Biosynthesis and Gene Regulation in Seed (종자내 아미노산 합성 조절 유전자에 관한 연구)

  • ;;;;;Fumio Takaiwa
    • Proceedings of the Botanical Society of Korea Conference
    • /
    • 1996.07a
    • /
    • pp.61-74
    • /
    • 1996
  • Human and monogastric animals can not synthesize 10 out of the 20 amino asids and therefor need to obtain these from their diet. The plant seed is a major source of dietary protein. It is particular important in their study to increase nutritional quality of the seed storage proteins. The low contents of lysine, asparagine and threonenein various cereal seeds and of cystein and methionine. In legume seeds is due to the low proportions of these amino acids in the major storage proteins, we have tried to apply the three strategies; (1) mutagenesis and selection of specific amino acid analogue resistance, (2) cloning and expression study of lysine biosynthesis related gene, (3) transfomation of lysine rich soybean glycinin gene. The 5-methyltryptophan (5MT) resistant cell lines, SAR1, SAR2 and SAR3 were selected from anther derived callus of rice (Oryza sativa L. "Sasanishiki"). Among these selected cell lines, two (SAR1 and SAR3) were able to grow stably at 200 mg/L of 5MT. Analysis of the freed amino acids in callus shows that 5MT resistant cells (SAR3) accumulated free tryptophan at least up to 50 times higher than those that of the higher than of SAS. These results indicated that the 5MT resistant cell lines are useful in studies of amino acid biosynthesis. Tr75, a rice (Oryza sativa L., var. Sasanishiki) mutant resistant to 5MT was segregated from the progenies of its initial mutant line, TR1. The 5MT resistant of TR75 was inherited in the M8 generations as a single dominant nuclear gene. The content of free amino acids in the TR75 homozygous seeds increased approximately 1.5 to 2.0 fold compared to wild-type seeds. Especially, the contents of tryptophan, phenylalanine and aspartic acid were 5.0, 5.3 and 2.7 times higher than those of wild-type seeds, respectively. The content of lysine is significantly low in rice. The lysine is synthesized by a complex pathway that is predominantly regulated by feedback inhibition of several enzymes including asparginase, aspatate kinase, dihydrodipicolinat synthase, etc. For understanding the regulation mechanism of lysine synthesis in rice, we try to clone the lysine biosynthetic metabolism related gene, DHPS and asparaginase, from rice. We have isolated a rice DHPS genomic clone which contains an ORF of 1044 nucleotides (347 amino acids, Mr. 38, 381 daltons), an intron of 587 nucleotides and 5'and 3'-flanking regions by screening of rice genomic DNA library. Deduced amino acid sequence of mature peptide domain of GDHPS clone is highly conserved in monocot and dicot plants whereas that of transit peptide domain is extremely different depending on plant specie. Southern blot analysis indicated that GDHPS is located two copy gene in rice genome. The transcripts of a rice GDHPS were expressed in leaves and roots but not detected in callus tissues. The transcription level of GDHPS is much higher in leaves indicating enormous chloroplast development than roots. Genomic DNA clones for asparaginase genes were screened from the rice genomic library by using plaque hybridization technique. Twelve different genomic clones were isolated from first and second screening, and 8 of 12 clones were analyzed by restriction patterns and identified by Southern Blotting, Restriction enzyme digestion patterns and Southern blot analysis of 8 clones show the different pattern for asparaginase gene. Genomic Southern blot analysis from rice were done. It is estimated that rice has at least 2-3 copy of asparaginase gene. One of 8 positive clones was subcloned into the pBluescript SK(+) vector, and was constructed the physical map. For transformation of lysine rich storage protein into tobacco, soybean glycinin genes are transformed into tobacco. To examine whether glycinin could be stably accumulated in endosperm tissue, the glycinin cDNA was transcriptionally fused to an endosperm-specific promotor of the rice storage protein glutelin gene and then introduced into tobacco genomic via Agrobacterium-mediated transformation. Consequently the glycinin gene was expressed in a seed-and developmentally-specific manner in transgenic tobacco seeds. Glycinin were targeted to vacuole-derived protein bodies in the endosperm tissue and highly accumulated in the matrix region of many transgenic plant (1-4% of total seed proteins). Synthesized glycinin was processed into mature form, and assembled into a hexamer in a similar manner as the glycinin in soybean seed. Modified glycinin, in which 4 contiguous methionine residues were inserted at the variable regions corresponding to the C - teminal regions of the acidic and basic polypeptides, were also found to be accumulated similarly as in the normal glycinin. There was no apparent difference in the expression level, processing and targeting to protein bodies, or accumulation level between normal and modified glycinin. glycinin.

  • PDF

Construction of Web-Based Database for Anisakis Research (고래회충 연구를 위한 웹기반 데이터베이스 구축)

  • Lee, Yong-Seok;Baek, Moon-Ki;Jo, Yong-Hun;Kang, Se-Won;Lee, Jae-Bong;Han, Yeon-Soo;Cha, Hee-Jae;Yu, Hak-Sun;Ock, Mee-Sun
    • Journal of Life Science
    • /
    • v.20 no.3
    • /
    • pp.411-415
    • /
    • 2010
  • Anisakis simplex is one of the parasitic nematodes, and has a complex life cycle in crustaceans, fish, squid or whale. When people eat under-processed or raw fish, it causes anisakidosis and also plays a critical role in inducing serious allergic reactions in humans. However, no web-based database on A. simplex at the level of DNA or protein has been so far reported. In this context, we constructed a web-based database for Anisakis research. To build up the web-based database for Anisakis research, we proceeded with the following measures: First, sequences of order Ascaridida were downloaded and translated into the multifasta format which was stored as database for stand-alone BLAST. Second, all of the nucleotide and EST sequences were clustered and assembled. And EST sequences were translated into amino acid sequences for Nuclear Localization Signal prediction. In addition, we added the vector, E. coli, and repeat sequences into the database to confirm a potential contamination. The web-based database gave us several advantages. Only data that agrees with the nucleotide sequences directly related with the order Ascaridida can be found and retrieved when searching BLAST. It is also very convenient to confirm contamination when making the cDNA or genomic library from Anisakis. Furthermore, BLAST results on the Anisakis sequence information can be quickly accessed. Taken together, the Web-based database on A. simplex will be valuable in developing species specific PCR markers and in studying SNP in A. simplex-related researches in the future.