• Title/Summary/Keyword: 개미 집단 시스템

Search Result 27, Processing Time 0.031 seconds

The Effect of Multiagent Interaction Strategy on the Performance of Ant Model (개미 모델 성능에서 다중 에이전트 상호작용 전략의 효과)

  • Lee Seung-Gwan
    • The Journal of the Korea Contents Association
    • /
    • v.5 no.3
    • /
    • pp.193-199
    • /
    • 2005
  • One of the important fields for heuristics algorithm is how to balance between Intensificationand Diversification. Ant Colony System(ACS) is a new meta heuristics algorithm to solve hard combinatorial optimization problem. It is a population based approach that uses exploitation of positive feedback as well as greedy search. It was first proposed for tackling the well known Traveling Salesman Problem(TSP). In this paper, we propose Multi Colony Interaction Ant Model that achieves positive negative interaction through elite strategy divided by intensification strategy and diversification strategy to improve the performance of original ACS. And, we apply multi colony interaction ant model by this proposed elite strategy to TSP and compares with original ACS method for the performance.

  • PDF

Multi Colony Ant Model using Positive.Negative Interaction between Colonies (집단간 긍정적.부정적 상호작용을 이용한 다중 집단 개미 모델)

  • Lee, Seung-Gwan;Chung, Tae-Choong
    • The KIPS Transactions:PartB
    • /
    • v.10B no.7
    • /
    • pp.751-756
    • /
    • 2003
  • Ant Colony Optimization (ACO) is new meta heuristics method to solve hard combinatorial optimization problem. It is a population based approach that uses exploitation of positive feedback as well as greedy search. It was firstly proposed for tackling the well known Traveling Salesman Problem (TSP) . In this paper, we introduce Multi Colony Ant Model that achieve positive interaction and negative interaction through Intensification and Diversification to improve original ACS performance. This algorithm is a method to solve problem through interaction between ACS groups that consist of some agent colonies to solve TSP problem. In this paper, we apply this proposed method to TSP problem and evaluates previous method and comparison for the performance and we wish to certify that qualitative level of problem solution is excellent.

Evolvable Hardware Using Ant Colony System (개미 집단 시스템을 이용한 진화 하드웨어)

  • 황금성;조성배
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2002.10d
    • /
    • pp.244-246
    • /
    • 2002
  • 진화 하드웨어(Evolvable Hardware)는 환경 적응력이 강하고 최적의 상태를 유연하게 유지하는 하드웨어 설계 기법이나 회로가 복잡해질수록 진화가 어려워지는 문제로 인해 활용이 늦어지고 있다. 본 논문에서는 이를 해결하기 위한 많은 연구 중 회로 진화 과정 분석을 위한 방법으로 개미집단 시스템을 제안한다. 경로 최적화 알고리즘인 개미집단 시스템을 적절히 변형하여 진화 하드웨어에 적용시키는 방법을 제안하고 이를 실험으로 확인하였으며, 실험 결과 하드웨어의 진화 과정을 관찰할 수 있었고, 목표 하드웨어의 해공간 특성이 페로몬으로 분포하고 있음도 관찰할 수 있었다.

  • PDF

Multi Colony Intensification.Diversification Interaction Ant Reinforcement Learning Using Temporal Difference Learning (Temporal Difference 학습을 이용한 다중 집단 강화.다양화 상호작용 개미 강화학습)

  • Lee Seung-Gwan
    • The Journal of the Korea Contents Association
    • /
    • v.5 no.5
    • /
    • pp.1-9
    • /
    • 2005
  • In this paper, we suggest multi colony interaction ant reinforcement learning model. This method is a hybrid of multi colony interaction by elite strategy and reinforcement teaming applying Temporal Difference(TD) learning to Ant-Q loaming. Proposed model is consisted of some independent AS colonies, and interaction achieves search according to elite strategy(Intensification, Diversification strategy) between the colonies. Intensification strategy enables to select of good path to use heuristic information of other agent colony. This makes to select the high frequency of the visit of a edge by agents through positive interaction of between the colonies. Diversification strategy makes to escape selection of the high frequency of the visit of a edge by agents achieve negative interaction by search information of other agent colony. Through this strategies, we could know that proposed reinforcement loaming method converges faster to optimal solution than original ACS and Ant-Q.

  • PDF

Balance between Intensification and Diversification in Ant Colony Optimization (개미 집단 최적화에서 강화와 다양화의 조화)

  • Lee, Seung-Gwan;Choi, Jin-Hyuk
    • The Journal of the Korea Contents Association
    • /
    • v.11 no.3
    • /
    • pp.100-107
    • /
    • 2011
  • One of the important fields for heuristic algorithm is how to balance between Intensification and Diversification. In this paper, we deal with the performance improvement techniques through balance the intensification and diversification in Ant Colony System(ACS) which is one of Ant Colony Optimization(ACO). In this paper, we propose the hybrid searching method between intensification strategy and diversification strategy. First, the length of the global optimal path does not improved within the limited iterations, we evaluates this state that fall into the local optimum and selects the next node using changed parameters in the state transition rule. And then we consider the overlapping edge of the global best path of the previous and the current, and, to enhance the pheromone for the overlapping edges increases the probability that the optimal path is configured. Finally, the performance of Best and Average-Best of proposed algorithm outperforms ACS-3-opt, ACS-Subpath, ACS-Iter and ACS-Global-Ovelap algorithms.

Ant Colony System Considering the Iteration Search Frequency that the Global Optimal Path does not Improved (전역 최적 경로가 향상되지 않는 반복 탐색 횟수를 고려한 개미 집단 시스템)

  • Lee, Seung-Gwan;Lee, Dae-Ho
    • Journal of the Korea Society of Computer and Information
    • /
    • v.14 no.1
    • /
    • pp.9-15
    • /
    • 2009
  • Ant Colony System is new meta heuristic for hard combinatorial optimization problem. The original ant colony system accomplishes a pheromone updating about only the global optimal path using global updating rule. But, If the global optimal path is not searched until the end condition is satisfied, only pheromone evaporation happens to no matter how a lot of iteration accomplishment. In this paper, the length of the global optimal path does not improved within the limited iterations, we evaluates this state that fall into the local optimum and selects the next node using changed parameters in the state transition rule. This method has effectiveness of the search for a path through diversifications is enhanced by decreasing the value of parameter of the state transition rules for the select of next node, and escape from the local optima is possible. Finally, the performance of Best and Average_Best of proposed algorithm outperforms original ACS.

Optimal solution search method by using modified local updating rule in Ant Colony System (개미군락시스템에서 수정된 지역 갱신 규칙을 이용한 최적해 탐색 기법)

  • Hong, Seok-Mi;Chung, Tae-Choong
    • Journal of the Korean Institute of Intelligent Systems
    • /
    • v.14 no.1
    • /
    • pp.15-19
    • /
    • 2004
  • Ant Colony System(ACS) is a meta heuristic approach based on biology in order to solve combinatorial optimization problem. It is based on the tracing action of real ants which accumulate pheromone on the passed path and uses as communication medium. In order to search the optimal path, ACS requires to explore various edges. In existing ACS, the local updating rule assigns the same pheromone to visited edge. In this paper, our local updating rule gives the pheromone according to the number of visiting times and the distance between visited cities. Our approach can have less local optima than existing ACS and find better solution by taking advantage of more informations during searching.

개미 시스템을 기반으로 한 Ad hoc 네트워크 멀티캐스팅

  • 이세영;김중항;장형수
    • Proceedings of the Korean Information Science Society Conference
    • /
    • 2004.10c
    • /
    • pp.1-3
    • /
    • 2004
  • 본 논문에서는 Core Based Tree(CBT) 알고리즘과 개미 집단 알고리즘의 특성을 융합하여 Mobile Ad hoc Network(MANET)에 맞는 멀티캐스팅 알고리즘, Ad hoc network Multicasting with Ant System (ANMAS)을 제안한다. ANMAS는 개미 알고리즘의 간접적 정보 전달 및 평가 방법을 통해 멀티캐스팅에 필요한 위상정보를 수집하여 견고한 멀티캐스팅 그룹을 형성함으로서 기존의 알고리즘에 비해 효율적이며 실험결과를 통해 이를 확인할 수 있다.

  • PDF

Performance Improvement of Cooperating Agents through Balance between Intensification and Diversification (강화와 다양화의 조화를 통한 협력 에이전트 성능 개선에 관한 연구)

  • 이승관;정태충
    • Journal of the Institute of Electronics Engineers of Korea CI
    • /
    • v.40 no.6
    • /
    • pp.87-94
    • /
    • 2003
  • One of the important fields for heuristic algorithm is how to balance between Intensification and Diversification. Ant Colony Optimization(ACO) is a new meta heuristic algorithm to solve hard combinatorial optimization problem. It is a population based approach that uses exploitation of positive feedback as well as Breedy search It was first Proposed for tackling the well known Traveling Salesman Problem(TSP). In this paper, we deal with the performance improvement techniques through balance the Intensification and Diversification in Ant Colony System(ACS). First State Transition considering the number of times that agents visit about each edge makes agents search more variously and widen search area. After setting up criteria which divide elite tour that receive Positive Intensification about each tour, we propose a method to do addition Intensification by the criteria. Implemetation of the algorithm to solve TSP and the performance results under various conditions are conducted, and the comparision between the original An and the proposed method is shown. It turns out that our proposed method can compete with the original ACS in terms of solution quality and computation speed to these problem.

A Reinforcement Loaming Method using TD-Error in Ant Colony System (개미 집단 시스템에서 TD-오류를 이용한 강화학습 기법)

  • Lee, Seung-Gwan;Chung, Tae-Choong
    • The KIPS Transactions:PartB
    • /
    • v.11B no.1
    • /
    • pp.77-82
    • /
    • 2004
  • Reinforcement learning takes reward about selecting action when agent chooses some action and did state transition in Present state. this can be the important subject in reinforcement learning as temporal-credit assignment problems. In this paper, by new meta heuristic method to solve hard combinational optimization problem, examine Ant-Q learning method that is proposed to solve Traveling Salesman Problem (TSP) to approach that is based for population that use positive feedback as well as greedy search. And, suggest Ant-TD reinforcement learning method that apply state transition through diversification strategy to this method and TD-error. We can show through experiments that the reinforcement learning method proposed in this Paper can find out an optimal solution faster than other reinforcement learning method like ACS and Ant-Q learning.