• Title/Summary/Keyword: Cognitive Interface

Search Result 203, Processing Time 0.019 seconds

Improved Resource Allocation Model for Reducing Interference among Secondary Users in TV White Space for Broadband Services

  • Marco P. Mwaimu;Mike Majham;Ronoh Kennedy;Kisangiri Michael;Ramadhani Sinde
    • International Journal of Computer Science & Network Security
    • /
    • v.23 no.4
    • /
    • pp.55-68
    • /
    • 2023
  • In recent years, the Television White Space (TVWS) has attracted the interest of many researchers due to its propagation characteristics obtainable between 470MHz and 790MHz spectrum bands. The plenty of unused channels in the TV spectrum allows the secondary users (SUs) to use the channels for broadband services especially in rural areas. However, when the number of SUs increases in the TVWS wireless network the aggregate interference also increases. Aggregate interferences are the combined harmful interferences that can include both co-channel and adjacent interferences. The aggregate interference on the side of Primary Users (PUs) has been extensively scrutinized. Therefore, resource allocation (power and spectrum) is crucial when designing the TVWS network to avoid interferences from Secondary Users (SUs) to PUs and among SUs themselves. This paper proposes a model to improve the resource allocation for reducing the aggregate interface among SUs for broadband services in rural areas. The proposed model uses joint power and spectrum hybrid Firefly algorithm (FA), Genetic algorithm (GA), and Particle Swarm Optimization algorithm (PSO) which is considered the Co-channel interference (CCI) and Adjacent Channel Interference (ACI). The algorithm is integrated with the admission control algorithm so that; there is a possibility to remove some of the SUs in the TVWS network whenever the SINR threshold for SUs and PU are not met. We considered the infeasible system whereby all SUs and PU may not be supported simultaneously. Therefore, we proposed a joint spectrum and power allocation with an admission control algorithm whose better complexity and performance than the ones which have been proposed in the existing algorithms in the literature. The performance of the proposed algorithm is compared using the metrics such as sum throughput, PU SINR, algorithm running time and SU SINR less than threshold and the results show that the PSOFAGA with ELGR admission control algorithm has best performance compared to GA, PSO, FA, and FAGAPSO algorithms.

Comparison of Deep Learning Frameworks: About Theano, Tensorflow, and Cognitive Toolkit (딥러닝 프레임워크의 비교: 티아노, 텐서플로, CNTK를 중심으로)

  • Chung, Yeojin;Ahn, SungMahn;Yang, Jiheon;Lee, Jaejoon
    • Journal of Intelligence and Information Systems
    • /
    • v.23 no.2
    • /
    • pp.1-17
    • /
    • 2017
  • The deep learning framework is software designed to help develop deep learning models. Some of its important functions include "automatic differentiation" and "utilization of GPU". The list of popular deep learning framework includes Caffe (BVLC) and Theano (University of Montreal). And recently, Microsoft's deep learning framework, Microsoft Cognitive Toolkit, was released as open-source license, following Google's Tensorflow a year earlier. The early deep learning frameworks have been developed mainly for research at universities. Beginning with the inception of Tensorflow, however, it seems that companies such as Microsoft and Facebook have started to join the competition of framework development. Given the trend, Google and other companies are expected to continue investing in the deep learning framework to bring forward the initiative in the artificial intelligence business. From this point of view, we think it is a good time to compare some of deep learning frameworks. So we compare three deep learning frameworks which can be used as a Python library. Those are Google's Tensorflow, Microsoft's CNTK, and Theano which is sort of a predecessor of the preceding two. The most common and important function of deep learning frameworks is the ability to perform automatic differentiation. Basically all the mathematical expressions of deep learning models can be represented as computational graphs, which consist of nodes and edges. Partial derivatives on each edge of a computational graph can then be obtained. With the partial derivatives, we can let software compute differentiation of any node with respect to any variable by utilizing chain rule of Calculus. First of all, the convenience of coding is in the order of CNTK, Tensorflow, and Theano. The criterion is simply based on the lengths of the codes and the learning curve and the ease of coding are not the main concern. According to the criteria, Theano was the most difficult to implement with, and CNTK and Tensorflow were somewhat easier. With Tensorflow, we need to define weight variables and biases explicitly. The reason that CNTK and Tensorflow are easier to implement with is that those frameworks provide us with more abstraction than Theano. We, however, need to mention that low-level coding is not always bad. It gives us flexibility of coding. With the low-level coding such as in Theano, we can implement and test any new deep learning models or any new search methods that we can think of. The assessment of the execution speed of each framework is that there is not meaningful difference. According to the experiment, execution speeds of Theano and Tensorflow are very similar, although the experiment was limited to a CNN model. In the case of CNTK, the experimental environment was not maintained as the same. The code written in CNTK has to be run in PC environment without GPU where codes execute as much as 50 times slower than with GPU. But we concluded that the difference of execution speed was within the range of variation caused by the different hardware setup. In this study, we compared three types of deep learning framework: Theano, Tensorflow, and CNTK. According to Wikipedia, there are 12 available deep learning frameworks. And 15 different attributes differentiate each framework. Some of the important attributes would include interface language (Python, C ++, Java, etc.) and the availability of libraries on various deep learning models such as CNN, RNN, DBN, and etc. And if a user implements a large scale deep learning model, it will also be important to support multiple GPU or multiple servers. Also, if you are learning the deep learning model, it would also be important if there are enough examples and references.

The role of the middle term in the integration of the two premises in linear syllogistic reasoning (선형 삼단 논법의 두 전제 통합 과정에서 중간 항목의 역할)

  • 정혜선;조명한
    • Korean Journal of Cognitive Science
    • /
    • v.12 no.3
    • /
    • pp.29-46
    • /
    • 2001
  • This study attempted to demonstrate that the integration of the two premises in linear syllogism is mediated by the middle term the term that is repeated in the two premises. In Experiment 1. we examined whether representing the middle term is more important than representing the end terms. We asked a question to each premise. Depending on the order of the questions either the two end terms or the middle term became the answer in both premises. Participants solved the problems better when the middle term became the answer suggesting that it is more important to represent the middle term than the end terms. In Experiment 2 we examined whether additional processing is needed for the integration beyond establishing co-referential link through the middle term. We pronominalized the middle term in the second premise and provided two kinds of information to disambiguate the pronoun. In the direct information condition we provided information about who the pronoun is whereas in the indirect information condition we provided information about the relative location of the pronoun. Participants solved the problems more quickly in the indirect information condition than in the direct information condition indicating that mere co-referential link was not enough and that the relative location of the middle term needs to be computed for the integration of the two premises.

  • PDF