• Title/Summary/Keyword: Search algorithms

Search Result 1,328, Processing Time 0.032 seconds

An Empirical Study on Statistical Optimization Model for the Portfolio Construction of Sponsored Search Advertising(SSA) (키워드검색광고 포트폴리오 구성을 위한 통계적 최적화 모델에 대한 실증분석)

  • Yang, Hognkyu;Hong, Juneseok;Kim, Wooju
    • Journal of Intelligence and Information Systems
    • /
    • v.25 no.2
    • /
    • pp.167-194
    • /
    • 2019
  • This research starts from the four basic concepts of incentive incompatibility, limited information, myopia and decision variable which are confronted when making decisions in keyword bidding. In order to make these concept concrete, four framework approaches are designed as follows; Strategic approach for the incentive incompatibility, Statistical approach for the limited information, Alternative optimization for myopia, and New model approach for decision variable. The purpose of this research is to propose the statistical optimization model in constructing the portfolio of Sponsored Search Advertising (SSA) in the Sponsor's perspective through empirical tests which can be used in portfolio decision making. Previous research up to date formulates the CTR estimation model using CPC, Rank, Impression, CVR, etc., individually or collectively as the independent variables. However, many of the variables are not controllable in keyword bidding. Only CPC and Rank can be used as decision variables in the bidding system. Classical SSA model is designed on the basic assumption that the CPC is the decision variable and CTR is the response variable. However, this classical model has so many huddles in the estimation of CTR. The main problem is the uncertainty between CPC and Rank. In keyword bid, CPC is continuously fluctuating even at the same Rank. This uncertainty usually raises questions about the credibility of CTR, along with the practical management problems. Sponsors make decisions in keyword bids under the limited information, and the strategic portfolio approach based on statistical models is necessary. In order to solve the problem in Classical SSA model, the New SSA model frame is designed on the basic assumption that Rank is the decision variable. Rank is proposed as the best decision variable in predicting the CTR in many papers. Further, most of the search engine platforms provide the options and algorithms to make it possible to bid with Rank. Sponsors can participate in the keyword bidding with Rank. Therefore, this paper tries to test the validity of this new SSA model and the applicability to construct the optimal portfolio in keyword bidding. Research process is as follows; In order to perform the optimization analysis in constructing the keyword portfolio under the New SSA model, this study proposes the criteria for categorizing the keywords, selects the representing keywords for each category, shows the non-linearity relationship, screens the scenarios for CTR and CPC estimation, selects the best fit model through Goodness-of-Fit (GOF) test, formulates the optimization models, confirms the Spillover effects, and suggests the modified optimization model reflecting Spillover and some strategic recommendations. Tests of Optimization models using these CTR/CPC estimation models are empirically performed with the objective functions of (1) maximizing CTR (CTR optimization model) and of (2) maximizing expected profit reflecting CVR (namely, CVR optimization model). Both of the CTR and CVR optimization test result show that the suggested SSA model confirms the significant improvements and this model is valid in constructing the keyword portfolio using the CTR/CPC estimation models suggested in this study. However, one critical problem is found in the CVR optimization model. Important keywords are excluded from the keyword portfolio due to the myopia of the immediate low profit at present. In order to solve this problem, Markov Chain analysis is carried out and the concept of Core Transit Keyword (CTK) and Expected Opportunity Profit (EOP) are introduced. The Revised CVR Optimization model is proposed and is tested and shows validity in constructing the portfolio. Strategic guidelines and insights are as follows; Brand keywords are usually dominant in almost every aspects of CTR, CVR, the expected profit, etc. Now, it is found that the Generic keywords are the CTK and have the spillover potentials which might increase consumers awareness and lead them to Brand keyword. That's why the Generic keyword should be focused in the keyword bidding. The contribution of the thesis is to propose the novel SSA model based on Rank as decision variable, to propose to manage the keyword portfolio by categories according to the characteristics of keywords, to propose the statistical modelling and managing based on the Rank in constructing the keyword portfolio, and to perform empirical tests and propose a new strategic guidelines to focus on the CTK and to propose the modified CVR optimization objective function reflecting the spillover effect in stead of the previous expected profit models.

AGREEMENT BETWEEN DSM-IV AND ICD-10 DIAGNOSIS ON ATTENTION-DEFICIT/HYPERACTIVITY DISORDER (주의력결핍과잉행동장애에 대한 DSM-IV와 ICD-10 진단의 합의)

  • Lee Soyoung Irene;Jung Han-Yone
    • Journal of the Korean Academy of Child and Adolescent Psychiatry
    • /
    • v.16 no.2
    • /
    • pp.192-198
    • /
    • 2005
  • Objectives : This review examines the extent of harmonization between the Diagnostic and Statistical Manual (fourth edition) (DSM-IV) attention deficit hyperactivity disorder (ADHD) and International Classification of Diseases (tenth edition) (ICD-10) hyperkinetic disorder (HKD). Methods : PubMed review of the clinical literatures from 1980 to March 2005 was undertaken using the search term of ADHD, HKD, DSM, ICD, diagnostic criteria, algorithm, and validity. Results : There continues to be a substantial difference in the prevalence of ADHD and HKD. Based on the limited studies that directly compare the most recent two algorithms of ADHD and HKD, several implications for clinical practice and research are discussed. Conclusion : To understand further if the criteria for DSM-IV and ICD-10 for ADHD and HKD are converged or not, more researches on the validity of ADHD and HKD subtypes should be undertaken.

  • PDF

Finding the time sensitive frequent itemsets based on data mining technique in data streams (데이터 스트림에서 데이터 마이닝 기법 기반의 시간을 고려한 상대적인 빈발항목 탐색)

  • Park, Tae-Su;Chun, Seok-Ju;Lee, Ju-Hong;Kang, Yun-Hee;Choi, Bum-Ghi
    • Journal of The Korean Association of Information Education
    • /
    • v.9 no.3
    • /
    • pp.453-462
    • /
    • 2005
  • Recently, due to technical improvements of storage devices and networks, the amount of data increase rapidly. In addition, it is required to find the knowledge embedded in a data stream as fast as possible. Huge data in a data stream are created continuously and changed fast. Various algorithms for finding frequent itemsets in a data stream are actively proposed. Current researches do not offer appropriate method to find frequent itemsets in which flow of time is reflected but provide only frequent items using total aggregation values. In this paper we proposes a novel algorithm for finding the relative frequent itemsets according to the time in a data stream. We also propose the method to save frequent items and sub-frequent items in order to take limited memory into account and the method to update time variant frequent items. The performance of the proposed method is analyzed through a series of experiments. The proposed method can search both frequent itemsets and relative frequent itemsets only using the action patterns of the students at each time slot. Thus, our method can enhance the effectiveness of learning and make the best plan for individual learning.

  • PDF

Optimal Cost Design of Pipe Network Systems Using Genetic Algorithms (遺傳子 알고리즘을 이용한 管網시스템의 最適費用 設計)

  • Park, Yeong-Su;Kim, Jong-U;Kim, Tae-Gyun;Kim, Jung-Hun
    • Journal of Korea Water Resources Association
    • /
    • v.32 no.1
    • /
    • pp.71-81
    • /
    • 1999
  • The objective of this study is to develop a model which can design an optimal pipe network system of least cost while satisfying all the design constraints including hydraulic constraints using a genetic algorithm technique. Hydraulic constraints interfaced with the simulation program(KYPIPE) checked feasible solution region. Genetic algorithm(GA) technique is a relatively new optimization technique. The GA is known as a very powerful search and optimization technique especially when solving nonlinear programming problems. The model developed in this study selects optimal pipe diameters in the form of commercial discrete sizes using the pipe diameters and the pumping powers as decision variables. The model not only determines the optimal diameters and pumping powers of pipe network system but also satisfies the discharge and pressure requirements at demanding nodes. The model has been applied to an imaginary and an existing pipe network systems. One system is adopted from journal papers which has been used as an example network by many other researchers. Comparison of the results shows compatibility of the model developed in this study. The model is also applied to a system in Goyang city in order to check the model applicability to finding of optimal pumping powers. It has been found that the developed model can be successfully applied to optimal design of pipe network systems in a relatively simple manner.

  • PDF

A QoS-based Inter-Domain Routing Scheme for Distributed Multimedia Applications in a High Wide Area Network (분산 멀티미디어 응용을 위한 대규모 고속 통신망에서의 QoS-근거 계층적 도메인간 라우팅 방식)

  • 김승훈;김치하
    • The Journal of Korean Institute of Communications and Information Sciences
    • /
    • v.24 no.7B
    • /
    • pp.1239-1251
    • /
    • 1999
  • In this paper a scalable QoS-based hierarchical inter-domain routing scheme for distributed multimedia applications in a high speed wide area network. The problem of QoS-based routing is formulated as a multicriteria shortest path problem, known as NP-complete[21,30]. Our routing scheme consists of two phases. In Phase 1, two graph construction algorithms are performed to model the network under consideration as a graph. The graph contains a part of the network topology which is completely neglected or partially considered by existing routing schemes, thus maintaining more accurate topology information. In Phase 2, a heuristic call-by-call algorithm is performed for selecting a feasible path efficiently in depth first search-like manner on the graph and tailoring to each application's QoS requirements, beginning at a vertex that represents the source node. In this paper, a simple rule is also produced, by which the visiting order of outgoing edges at each vertex on the graph is determined. The rule is based on each edge's the minimum normalized slackness to the QoS requested. The proposed routing scheme extends the PNNI-type hierarchical routing framework. Note that our routing scheme is one of a few QoS-based hierarchical routing schemes that address explicitly the issue of selecting a path with multiple metrics.

  • PDF

Video Matching Algorithm of Content-Based Video Copy Detection for Copyright Protection (저작권보호를 위한 내용기반 비디오 복사검출의 비디오 정합 알고리즘)

  • Hyun, Ki-Ho
    • Journal of Korea Multimedia Society
    • /
    • v.11 no.3
    • /
    • pp.315-322
    • /
    • 2008
  • Searching a location of the copied video in video database, signatures should be robust to video reediting, channel noise, time variation of frame rate. Several kinds of signatures has been proposed. Ordinal signature, one of them, is difficult to describe the spatial characteristics of frame due to the site of fixed window, $N{\times}N$, which is compute the average gray value. In this paper, I studied an algorithm of sequence matching in video copy detection for the copyright protection, employing the R-tree index method for retrieval and suggesting a robust ordinal signatures for the original video clips and the same signatures of the pirated video. Robust ordinal has a 2-dimensional vector structures that has a strong to the noise and the variation of the frame rate. Also, it express as MBR form in search space of R-tree. Moreover, I focus on building a video copy detection method into which content publishers register their valuable digital content. The video copy detection algorithms compares the web content to the registered content and notifies the content owners of illegal copies. Experimental results show the proposed method is improve the video matching rate and it has a characteristics of signature suitable to the large video databases.

  • PDF

An Approach of Scalable SHIF Ontology Reasoning using Spark Framework (Spark 프레임워크를 적용한 대용량 SHIF 온톨로지 추론 기법)

  • Kim, Je-Min;Park, Young-Tack
    • Journal of KIISE
    • /
    • v.42 no.10
    • /
    • pp.1195-1206
    • /
    • 2015
  • For the management of a knowledge system, systems that automatically infer and manage scalable knowledge are required. Most of these systems use ontologies in order to exchange knowledge between machines and infer new knowledge. Therefore, approaches are needed that infer new knowledge for scalable ontology. In this paper, we propose an approach to perform rule based reasoning for scalable SHIF ontologies in a spark framework which works similarly to MapReduce in distributed memories on a cluster. For performing efficient reasoning in distributed memories, we focus on three areas. First, we define a data structure for splitting scalable ontology triples into small sets according to each reasoning rule and loading these triple sets in distributed memories. Second, a rule execution order and iteration conditions based on dependencies and correlations among the SHIF rules are defined. Finally, we explain the operations that are adapted to execute the rules, and these operations are based on reasoning algorithms. In order to evaluate the suggested methods in this paper, we perform an experiment with WebPie, which is a representative ontology reasoner based on a cluster using the LUBM set, which is formal data used to evaluate ontology inference and search speed. Consequently, the proposed approach shows that the throughput is improved by 28,400% (157k/sec) from WebPie(553/sec) with LUBM.

Finding the One-to-One Optimum Path Considering User's Route Perception Characteristics of Origin and Destination (Focused on the Origin-Based Formulation and Algorithm) (출발지와 도착지의 경로인지특성을 반영한 One-to-One 최적경로탐색 (출발지기반 수식 및 알고리즘을 중심으로))

  • Shin, Seong-Il;Sohn, Kee-Min;Cho, Chong-Suk;Cho, Tcheol-Woong;Kim, Won-Keun
    • Journal of Korean Society of Transportation
    • /
    • v.23 no.7 s.85
    • /
    • pp.99-110
    • /
    • 2005
  • Total travel cost of route which connects origin with destination (O-D) is consist of the total sum of link travel cost and route perception cost. If the link perception cost is different according to the origin and destination, optimal route search has limitation to reflect the actual condition by route enumeration problem. The purpose of this study is to propose optimal route searching formulation and algorithm which is enable to reflect different link perception cost by each route, not only avoid the enumeration problem between origin and destination. This method defines minimum unit of route as a link and finally compares routes using link unit costs. The proposed method considers the perception travel cost at both origin and destination in optimal route searching process, while conventional models refect the perception cost only at origin. However this two-way searching algorithm is still not able to guarantee optimum solution. To overcome this problem, this study proposed an orign based optimal route searching method which was developed based on destination based optimal perception route tree. This study investigates whether proposed numerical formulas and algorithms are able to reflect route perception behavior reflected the feature of origin and destination in a real traffic network by the example research including the diversity of route information for the surrounding area and the perception cost for the road hierarchy.

A Semantic-Based Mashup Development Tool Supporting Various Open API Types (다양한 Open API 타입들을 지원하는 시맨틱 기반 매쉬업 개발 툴)

  • Lee, Yong-Ju
    • Journal of Internet Computing and Services
    • /
    • v.13 no.3
    • /
    • pp.115-126
    • /
    • 2012
  • Mashups have become very popular over the last few years, and their use also varies for IT convergency services. In spite of their popularity, there are several challenging issues when combining Open APIs into mashups, First, since portal sites may have a large number of APIs available for mashups, manually searching and finding compatible APIs can be a tedious and time-consuming task. Second, none of the existing portal sites provides a way to leverage semantic techniques that have been developed to assist users in locating and integrating APIs like those seen in traditional SOAP-based web services. Third, although suitable APIs have been discovered, the integration of these APIs is required for in-depth programming knowledge. To solve these issues, we first show that existing techniques and algorithms used for finding and matching SOAP-based web services can be reused, with only minor changes. Next, we show how the characteristics of APIs can be syntactically defined and semantically described, and how to use the syntactic and semantic descriptions to aid the easy discovery and composition of Open APIs. Finally, we propose a goal-directed interactive approach for the dynamic composition of APIs, where the final mashup is gradually generated by a forward chaining of APIs. At each step, a new API is added to the composition.

A Study on Integrated Binding Service Strategy Based on Name/property in Wide-Area Object Computing Environments (광역 객체 컴퓨팅 환경에서 이름/속성기반의 통합 바이딩 서비스 방안)

  • Jeong, Chang-Won;Oh, Sung-Kwun;Joo, Su-Chong
    • The KIPS Transactions:PartA
    • /
    • v.9A no.2
    • /
    • pp.241-248
    • /
    • 2002
  • With the structure of tilde-area computing system which Is specified by a researching team in Vrije University, Netherlands, lots of researchers and developers have been progressing the studies of global location and interconnection services of distributed objects existing in global sites. Most of them halve focused on binding services of only non-duplicated computational objects existing wide-area computing sites without any consideration of duplication problems. But all of objects existing on the earth rave the duplicated characteristics according to how to categorize their own names or properties. These objects with the same property can define as duplicated computational objects. Up to now, the existing naming or trading mechanism has not supported the binding services of duplicated objects, because of deficiency of independent location service. For this reason, we suggest a new model that can not only manages locations of duplicated objects In wide-area computing environments, but also provide minimum binding time by considering both the optimal selection of one of duplicated objects and load balance among distributed systems. Our model is functionally divided into 2 parts, one part to obtain an unique object handle of duplicated objects with same property as a naming and trading service, and the other to search one or more contact addresses by a node manager using a liven object handle, as a location service For location transparency, these services are independently executing each other. Based on our model, we described structure of wide-area integrated tree and algorithms for searching and updating contact address of distributed object on this tree. finally, we showed a federation structure that can globally bind distributed objects located on different regions from an arbitrary client object.