• Title/Summary/Keyword: seed point generation

Search Result 9, Processing Time 0.025 seconds

MODIFIED DOUBLE SNAKE ALGORITHM FOR ROAD FEATURE UPDATING OF DIGITAL MAPS USING QUICKBIRD IMAGERY

  • Choi, Jae-Wan;Kim, Hye-Jin;Byun, Young-Gi;Han, You-Kyung;Kim, Yong-Il
    • Proceedings of the KSRS Conference
    • /
    • 2007.10a
    • /
    • pp.234-237
    • /
    • 2007
  • Road networks are important geospatial databases for various GIS (Geographic Information System) applications. Road digital maps may contain geometric spatial errors due to human and scanning errors, but manually updating roads information is time consuming. In this paper, we developed a new road features updating methodology using from multispectral high-resolution satellite image and pre-existing vector map. The approach is based on initial seed point generation using line segment matching and a modified double snake algorithm. Firstly, we conducted line segment matching between the road vector data and the edges of image obtained by Canny operator. Then, the translated road data was used to initialize the seed points of the double snake model in order to refine the updating of road features. The double snake algorithm is composed of two open snake models which are evolving jointly to keep a parallel between them. In the proposed algorithm, a new energy term was added which behaved as a constraint. It forced the snake nodes not to be out of potential road pixels in multispectral image. The experiment was accomplished using a QuickBird pan-sharpened multispectral image and 1:5,000 digital road maps of Daejeon. We showed the feasibility of the approach by presenting results in this urban area.

  • PDF

Interior Surface Design by Fractal Pattern (프랙탈 패턴에 의한 인테리어 표면디자인)

  • Kim, Joo-Mi
    • Proceedings of the Korean Institute of Interior Design Conference
    • /
    • 2008.05a
    • /
    • pp.36-38
    • /
    • 2008
  • The purpose of the study is to propose a new surface design concepts within fractal pattern. In this study, I am offering the fractal concepts drawn from science, as a new anchoring point for surface design. Fractal Patterns are generated by transforming a seed slab into a number of constituent elements through fractal operations of rotation, scaling and linear transformations. These elements are bound together as a second generation seed shape which is reiterated according to the same transformations. This process continues for as many generations as desired. In conclusion, this study places a great emphasis on the natural pattern order to the surface generation, which I hope will contribute to generating a number of creative possibilities for interior design.

  • PDF

A new clustering algorithm based on the connected region generation

  • Feng, Liuwei;Chang, Dongxia;Zhao, Yao
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.12 no.6
    • /
    • pp.2619-2643
    • /
    • 2018
  • In this paper, a new clustering algorithm based on the connected region generation (CRG-clustering) is proposed. It is an effective and robust approach to clustering on the basis of the connectivity of the points and their neighbors. In the new algorithm, a connected region generating (CRG) algorithm is developed to obtain the connected regions and an isolated point set. Each connected region corresponds to a homogeneous cluster and this ensures the separability of an arbitrary data set theoretically. Then, a region expansion strategy and a consensus criterion are used to deal with the points in the isolated point set. Experimental results on the synthetic datasets and the real world datasets show that the proposed algorithm has high performance and is insensitive to noise.

Design and Analysis of Technical Management System of Personal Information Security using Web Crawer (웹 크롤러를 이용한 개인정보보호의 기술적 관리 체계 설계와 해석)

  • Park, In-pyo;Jeon, Sang-june;Kim, Jeong-ho
    • Journal of Platform Technology
    • /
    • v.6 no.4
    • /
    • pp.69-77
    • /
    • 2018
  • In the case of personal information files containing personal information, there is insufficient awareness of personal information protection in end-point areas such as personal computers, smart terminals, and personal storage devices. In this study, we use Diffie-Hellman method to securely retrieve personal information files generated by web crawler. We designed SEED and ARIA using hybrid slicing to protect against attack on personal information file. The encryption performance of the personal information file collected by the Web crawling method is compared with the encryption decryption rate according to the key generation and the encryption decryption sharing according to the user key level. The simulation was performed on the personal information file delivered to the external agency transmission process. As a result, we compared the performance of existing methods and found that the detection rate is improved by 4.64 times and the information protection rate is improved by 18.3%.

Evolutionary Multi-Objective Optimization Algorithms for Uniform Distributed Pareto Optimal Solutions (균일분포의 파레토 최적해 생성을 위한 다목적 최적화 진화 알고리즘)

  • Jang Su-Hyun;Yoon Byungjoo
    • The KIPS Transactions:PartB
    • /
    • v.11B no.7 s.96
    • /
    • pp.841-848
    • /
    • 2004
  • Evolutionary a1gorithms are well-suited for multi-objective optimization problems involving several, often conflicting objectives. Pareto-based evolutionary algorithms, in particular, have shown better performance than other multi-objective evolutionary algorithms in comparison. However, generalized evolutionary multi-objective optimization algorithms have a weak point, in which the distribution of solutions are not uni-formly distributed onto Pareto optimal front. In this paper, we propose an evolutionary a1gorithm for multi-objective optimization which uses seed individuals in order to overcome weakness of algorithms Published. Seed individual means a solution which is not located in the crowded region on Pareto front. And the idea of our algorithm uses seed individuals for reproducing individuals for next generation. Thus, proposed a1go-rithm takes advantage of local searching effect because new individuals are produced near the seed individual with high probability, and is able to produce comparatively uniform distributed pareto optimal solutions. Simulation results on five testbed problems show that the proposed algo-rithm could produce uniform distributed solutions onto pareto optimal front, and is able to show better convergence compared to NSGA-II on all testbed problems except multi-modal problem.

A Technique of Watermark Generation and Similarity Embedding for Still Images Based on Cross Reference Points (교차참조점에 기반한 정지영상의 워터마크 생성 및 유사성 삽입 기법)

  • Lee, Hang-Chan
    • The Transactions of The Korean Institute of Electrical Engineers
    • /
    • v.56 no.8
    • /
    • pp.1484-1490
    • /
    • 2007
  • The Cross Reference Point(CRP) is a robust method for finding salient points in watermarking systems because it is based on the geometrical structure of a normalized image in order to avoid pointing error caused by digital attacks. After normalization of an image, the 100 CRPs are calculated. Next, the 100 blocks centered by CRPS are formed. These 100 blocks are arranged using a secrete key. Each boundary of 50 out of 100 blocks is surrounded by 8 blocks which are selected by the ordered number of a preceding block. This number is a seed of random number generator for selecting 8 out of 50 blocks. The search area of a center block is formed by a secrete key. The pixels of a center block are quantized to 10 levels by predefined thresholds. The watermarks are generated by the 50 quantized center blocks. These watermarks are embedded directly in the remaining 50 blocks. In other words, 50 out of 100 blocks are utilized to generate watermarks and the remaining 50 blocks are used to watermark embedding. Because the watermarks are generated in the given images, we can successfully detect watermarks after several digital attacks. The reason is that the blocks for the generation and detection of watermarks are equally affected by digital attacks except for the case of local distortion such as cropping.

Area based image matching with MOC-NA imagery (MOC-NA 영상의 영역기준 영상정합)

  • Youn, Jun-Hee;Park, Choung-Hwan
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.28 no.4
    • /
    • pp.463-469
    • /
    • 2010
  • Since MOLA(Mars Orbiter Laser Altimeter) data, which provides altimetry data for Mars, does not cover the whole Mars area, image matching with MOC imagery should be implemented for the generation of DEM. However, automatic image matching is difficult because of insufficient features and low contrast. In this paper, we present the area based semi-automatic image matching algorithm with MOC-NA(Mars Orbiter Camera ? Narrow Angle) imagery. To accomplish this, seed points describing conjugate points are manually added for the stereo imagery, and interesting points are automatically produced by using such seed points. Produced interesting points being used as initial conjugate points, area based image matching is implemented. For the points which fail to match, the locations of initial conjugate points are recalculated by using matched six points and image matching process is re-implemented. The quality assessment by reversing the role of target and search image shows 97.5 % of points were laid within one pixel absolute difference.

Normalized Digital Surface Model Extraction and Slope Parameter Determination through Region Growing of UAV Data (무인항공기 데이터의 영역 확장법 적용을 통한 정규수치표면모델 추출 및 경사도 파라미터 설정)

  • Yeom, Junho;Lee, Wonhee;Kim, Taeheon;Han, Youkyung
    • Journal of the Korean Society of Surveying, Geodesy, Photogrammetry and Cartography
    • /
    • v.37 no.6
    • /
    • pp.499-506
    • /
    • 2019
  • NDSM (Normalized Digital Surface Model) is key information for the detailed analysis of remote sensing data. Although NDSM can be simply obtained by subtracting a DTM (Digital Terrain Model) from a DSM (Digital Surface Model), in case of UAV (Unmanned Aerial Vehicle) data, it is difficult to get an accurate DTM due to high resolution characteristics of UAV data containing a large number of complex objects on the ground such as vegetation and urban structures. In this study, RGB-based UAV vegetation index, ExG (Excess Green) was used to extract initial seed points having low ExG values for region growing such that a DTM can be generated cost-effectively based on high resolution UAV data. For this process, local window analysis was applied to resolve the problem of erroneous seed point extraction from local low ExG points. Using the DSM values of seed points, region growing was applied to merge neighboring terrain pixels. Slope criteria were adopted for the region growing process and the seed points were determined as terrain points in case the size of segments is larger than 0.25 ㎡. Various slope criteria were tested to derive the optimized value for UAV data-based NDSM generation. Finally, the extracted terrain points were evaluated and interpolation was performed using the terrain points to generate an NDSM. The proposed method was applied to agricultural area in order to extract the above ground heights of crops and check feasibility of agricultural monitoring.

Subject-Balanced Intelligent Text Summarization Scheme (주제 균형 지능형 텍스트 요약 기법)

  • Yun, Yeoil;Ko, Eunjung;Kim, Namgyu
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.2
    • /
    • pp.141-166
    • /
    • 2019
  • Recently, channels like social media and SNS create enormous amount of data. In all kinds of data, portions of unstructured data which represented as text data has increased geometrically. But there are some difficulties to check all text data, so it is important to access those data rapidly and grasp key points of text. Due to needs of efficient understanding, many studies about text summarization for handling and using tremendous amounts of text data have been proposed. Especially, a lot of summarization methods using machine learning and artificial intelligence algorithms have been proposed lately to generate summary objectively and effectively which called "automatic summarization". However almost text summarization methods proposed up to date construct summary focused on frequency of contents in original documents. Those summaries have a limitation for contain small-weight subjects that mentioned less in original text. If summaries include contents with only major subject, bias occurs and it causes loss of information so that it is hard to ascertain every subject documents have. To avoid those bias, it is possible to summarize in point of balance between topics document have so all subject in document can be ascertained, but still unbalance of distribution between those subjects remains. To retain balance of subjects in summary, it is necessary to consider proportion of every subject documents originally have and also allocate the portion of subjects equally so that even sentences of minor subjects can be included in summary sufficiently. In this study, we propose "subject-balanced" text summarization method that procure balance between all subjects and minimize omission of low-frequency subjects. For subject-balanced summary, we use two concept of summary evaluation metrics "completeness" and "succinctness". Completeness is the feature that summary should include contents of original documents fully and succinctness means summary has minimum duplication with contents in itself. Proposed method has 3-phases for summarization. First phase is constructing subject term dictionaries. Topic modeling is used for calculating topic-term weight which indicates degrees that each terms are related to each topic. From derived weight, it is possible to figure out highly related terms for every topic and subjects of documents can be found from various topic composed similar meaning terms. And then, few terms are selected which represent subject well. In this method, it is called "seed terms". However, those terms are too small to explain each subject enough, so sufficient similar terms with seed terms are needed for well-constructed subject dictionary. Word2Vec is used for word expansion, finds similar terms with seed terms. Word vectors are created after Word2Vec modeling, and from those vectors, similarity between all terms can be derived by using cosine-similarity. Higher cosine similarity between two terms calculated, higher relationship between two terms defined. So terms that have high similarity values with seed terms for each subjects are selected and filtering those expanded terms subject dictionary is finally constructed. Next phase is allocating subjects to every sentences which original documents have. To grasp contents of all sentences first, frequency analysis is conducted with specific terms that subject dictionaries compose. TF-IDF weight of each subjects are calculated after frequency analysis, and it is possible to figure out how much sentences are explaining about each subjects. However, TF-IDF weight has limitation that the weight can be increased infinitely, so by normalizing TF-IDF weights for every subject sentences have, all values are changed to 0 to 1 values. Then allocating subject for every sentences with maximum TF-IDF weight between all subjects, sentence group are constructed for each subjects finally. Last phase is summary generation parts. Sen2Vec is used to figure out similarity between subject-sentences, and similarity matrix can be formed. By repetitive sentences selecting, it is possible to generate summary that include contents of original documents fully and minimize duplication in summary itself. For evaluation of proposed method, 50,000 reviews of TripAdvisor are used for constructing subject dictionaries and 23,087 reviews are used for generating summary. Also comparison between proposed method summary and frequency-based summary is performed and as a result, it is verified that summary from proposed method can retain balance of all subject more which documents originally have.