scispace - formally typeset
Search or ask a question
Journal ArticleDOI

Knowledge Engineering on Internet of Things through Reinforcement Learning

17 Mar 2020-International Journal of Computer Applications (Foundation of Computer Science)-Vol. 177, Iss: 44, pp 1-7
About: This article is published in International Journal of Computer Applications.The article was published on 2020-03-17 and is currently open access. It has received 11 citations till now. The article focuses on the topics: Knowledge engineering & Reinforcement learning.

Content maybe subject to copyright    Report

Citations
More filters
DOI
01 Jun 2020
TL;DR: This paper presents a first fast semi-online 3-Dimensional machine learning algorithm suitable for proper beam selection as is emitted from UAVs and presents a detailed step by step approach involved in the multi-armed bandit approach in solving UAV solving selection exploration to exploitation dilemmas.
Abstract: Unmanned Aerial vehicles (UAVs) emerged into a promising research trend applied in several disciplines based on the benefits, including efficient communication, on-time search, and rescue operations, appreciate customer deliveries among more. The current technologies are using fixed base stations (BS) to operate onsite and off-site in the fixed position with its associated problems like poor connectivity. These open gates for the UAVs technology to be used as a mobile alternative to increase accessibility in beam selection with a fifth-generation (5G) connectivity that focuses on increased availability and connectivity. This paper presents a first fast semi-online 3-Dimensional machine learning algorithm suitable for proper beam selection as is emitted from UAVs. Secondly, it presents a detailed step by step approach that is involved in the multi-armed bandit approach in solving UAV solving selection exploration to exploitation dilemmas. The obtained results depicted that a multi-armed bandit problem approach can be applied in optimizing the performance of any mobile networked devices issue based on bandit samples like Thompson sampling, Bayesian algorithm, and e-Greedy Algorithm. The results further illustrated that the 3-Dimensional algorithm optimizes utilization of technological resources compared to the existing single and the 2-Dimensional algorithms thus close optimal performance on the average period through machine learning of realistic UAV communication situations.

15 citations

Journal ArticleDOI
TL;DR: A study on deep learning to medical field application in general and detailed steps that are involved in the multiarmed bandit (MAB) approach in solving the UAV biomedical engineering technology device and medical exploration to exploitation dilemma are provided.
Abstract: The unmanned aerial vehicles (UAVs) emerged into a promising research trend within the recurrent year where current and future networks are to use enhanced connectivity in these digital immigrations in different fields like medical, communication, and search and rescue operations among others. The current technologies are using fixed base stations to operate onsite and off-site in the fixed position with its associated problems like poor connectivity. This open gate for the UAV technology is to be used as a mobile alternative to increase accessibility with fifth-generation (5G) connectivity that focuses on increased availability and connectivity. There has been less usage of wireless technologies in the medical field. This paper first presents a study on deep learning to medical field application in general and provides detailed steps that are involved in the multiarmed bandit (MAB) approach in solving the UAV biomedical engineering technology device and medical exploration to exploitation dilemma. The paper further presents a detailed description of the bandit network applicability to achieve close optimal performance and efficiency of medical engineered devices. The simulated results depicted that a multiarmed bandit problem approach can be applied in optimizing the performance of any medical networked device issue compared to the Thompson sampling, Bayesian algorithm, and e-greedy algorithm. The results obtained further illustrated the optimized utilization of biomedical engineering technology systems achieving thus close optimal performance on the average period through deep learning of realistic medical situations.

14 citations


Cites background from "Knowledge Engineering on Internet o..."

  • ...[45] which is the individual periphery recognition due to the fact of the significance to accomplish precise entity mining of Chinas’ electronic medical records....

    [...]

Journal ArticleDOI
TL;DR: This paper modeled beam selection with environmental responsiveness in millimeter Wave UAV to accomplish close optimum assessments on the regular period through learning from the available situation.
Abstract: The 5G technology is predicted to achieve the unoptimized millimeter Wave (mmWave) of 30-300 GHz bands. This unoptimized band because of the loss of mm-Wave bands, like path attenuation and propagation losses. Nonetheless, because of: (i) directional transmission paving way for beamforming to recompense for the path attenuation, and (ii) sophisticated placement concreteness of the base stations (BS) is the best alternative for array wireless communications in mmWave bands (that is to say 100-150 m). The advance in technology and innovation of unmanned aerial vehicles (UAVs) necessitates many opportunities and uncertainties. UAVs are agile and can fly all complexities if the terrains making ground robots unsuitable. The UAV may be managed either independently through aboard computers or distant controlled of a flight attendant on pulverized wireless communication links in our case 5G. Although a fast algorithm solved the problematic aspect of beam selection for 2-dimensional scenarios. This paper presents 3-dimensional scenarios for UAV. We modeled beam selection with environmental responsiveness in millimeter Wave UAV to accomplish close optimum assessments on the regular period through learning from the available situation.

11 citations


Cites background from "Knowledge Engineering on Internet o..."

  • ...INTRODUCTION Machine Learning (ML) being a part of the artificial intelligence, it has influenced some communication like in broadcast communication well detailed in where ML can be used in extending parameters during connectivity remote areas either positively or negatively including information retrievals, internet of Things that entails UAVs, media and web communication, online making, monetization, mining, quality of services in web qualities, web security and privacy issues in social networks among others [1-11]....

    [...]

Journal ArticleDOI
30 Sep 2020
TL;DR: An in-depth assessment of RL techniques in IoT systems focusing on the main known RL techniques like artificial neural network (ANN), Q-learning, Markov Decision Process (MDP), Learning Automata (LA) is surveyed.
Abstract: Reinforcement learning (RL) is a new propitious research space that is well-known nowadays on the internet of things (IoT), media and social sensing computing are addressing a broad and pertinent task through making decisions sequentially by deterministic and stochastic evolutions. The IoTs extend world connectivity to physical devices like electronic devices network by use interconnect with others over the Internet with the possibility of remotely being supervised and meticulous. In this paper, we comprehensively survey an in-depth assessment of RL techniques in IoT systems focusing on the main known RL techniques like artificial neural network (ANN), Q-learning, Markov Decision Process (MDP), Learning Automata (LA). This study examines and analyses learning technique with focusing on challenges, models performance, similarities and the differences in IoTs accomplish with most correlated proposed state of the art models. The results obtained can be used as a foundation for designing, a model implementation based on the bottlenecks currently assessed with an evaluation of the most fashionable hands-on utility of current methods for reinforcement learning.

8 citations


Cites background from "Knowledge Engineering on Internet o..."

  • ...It is based on a collection of connected nodes called artificial neurons that loosely model the neurons in a biological brain [15], [117]....

    [...]

Journal ArticleDOI
14 May 2021
TL;DR: A new model to address mainly energy efficiency about response time and the service delays in IoT-ESC is introduced and simulated results demonstrated that the proposed model minimized service delay and reduced energy consumption during computation.
Abstract: In recent years, the IoT) Internet of Things (IoT) allows devices to connect to the Internet that has become a promising research area mainly due to the constant emerging of the dynamic improvement of technologies and their associated challenges. In an approach to solve these challenges, fog computing came to play since it closely manages IoT connectivity. Fog-Enabled Smart Cities (IoT-ESC) portrays equitable energy consumption of a 7% reduction from 18.2% renewable energy contribution, which extends resource computation as a great advantage. The initialization of IoT-Enabled Smart Grids including (FESC) like fog nodes in fog computing, reduced workload in Terminal Nodes services (TNs) that are the sensors and actuators of the Internet of Things (IoT) set up. This paper proposes an integrated energy-efficiency model computation about the response time and delays service minimization delay in FESC. The FESC gives an impression of an auspicious computing model for location, time, and delay-sensitive applications supporting vertically -isolated, service delay, sensitive solicitations by providing abundant, ascendable, and scattered figuring stowage and system associativity. We first reviewed the persisting challenges in the proposed state-of-the models and based on them. We introduce a new model to address mainly energy efficiency about response time and the service delays in IoT-ESC. The iFogsim simulated results demonstrated that the proposed model minimized service delay and reduced energy consumption during computation. We employed IoT-ESC to decide autonomously or semi-autonomously whether the computation is to be made on Fog nodes or its transfer to the cloud.

6 citations


Cites background from "Knowledge Engineering on Internet o..."

  • ...During the last era of technology, highly intensive research activities took place in IoMT [7]....

    [...]

References
More filters
Journal ArticleDOI
01 Nov 2012
TL;DR: The workings of the natural gradient is described, which has made its way into many actor-critic algorithms over the past few years, and a review of several standard and natural actor-Critic algorithms is given.
Abstract: Policy-gradient-based actor-critic algorithms are amongst the most popular algorithms in the reinforcement learning framework. Their advantage of being able to search for optimal policies using low-variance gradient estimates has made them useful in several real-life applications, such as robotics, power control, and finance. Although general surveys on reinforcement learning techniques already exist, no survey is specifically dedicated to actor-critic algorithms in particular. This paper, therefore, describes the state of the art of actor-critic algorithms, with a focus on methods that can work in an online setting and use function approximation in order to deal with continuous state and action spaces. After starting with a discussion on the concepts of reinforcement learning and the origins of actor-critic algorithms, this paper describes the workings of the natural gradient, which has made its way into many actor-critic algorithms over the past few years. A review of several standard and natural actor-critic algorithms is given, and the paper concludes with an overview of application areas and a discussion on open issues.

764 citations

Journal ArticleDOI
TL;DR: An extensive literature review over the period 2002-2013 of machine learning methods that were used to address common issues in WSNs is presented and a comparative guide is provided to aid WSN designers in developing suitable machine learning solutions for their specific application challenges.
Abstract: Wireless sensor networks (WSNs) monitor dynamic environments that change rapidly over time. This dynamic behavior is either caused by external factors or initiated by the system designers themselves. To adapt to such conditions, sensor networks often adopt machine learning techniques to eliminate the need for unnecessary redesign. Machine learning also inspires many practical solutions that maximize resource utilization and prolong the lifespan of the network. In this paper, we present an extensive literature review over the period 2002–2013 of machine learning methods that were used to address common issues in WSNs. The advantages and disadvantages of each proposed algorithm are evaluated against the corresponding problem. We also provide a comparative guide to aid WSN designers in developing suitable machine learning solutions for their specific application challenges.

704 citations

Journal ArticleDOI
01 Jul 1974
TL;DR: Attention has been focused on the norms of behavior of learning automata, issues in the design of updating schemes, convergence of the action probabilities, and interaction of several automata.
Abstract: Stochastic automata operating in an unknown random environment have been proposed earlier as models of learning. These automata update their action probabilities in accordance with the inputs received from the environment and can improve their own performance during operation. In this context they are referred to as learning automata. A survey of the available results in the area of learning automata has been attempted in this paper. Attention has been focused on the norms of behavior of learning automata, issues in the design of updating schemes, convergence of the action probabilities, and interaction of several automata. Utilization of learning automata in parameter optimization and hypothesis testing is discussed, and potential areas of application are suggested.

688 citations

Journal ArticleDOI
Junfei Qiu1, Qihui Wu1, Guoru Ding1, Yuhua Xu1, Shuo Feng1 
TL;DR: A literature survey of the latest advances in researches on machine learning for big data processing finds some promising learning methods in recent studies, such as representation learning, deep learning, distributed and parallel learning, transfer learning, active learning, and kernel-based learning.
Abstract: There is no doubt that big data are now rapidly expanding in all science and engineering domains. While the potential of these massive data is undoubtedly significant, fully making sense of them requires new ways of thinking and novel learning techniques to address the various challenges. In this paper, we present a literature survey of the latest advances in researches on machine learning for big data processing. First, we review the machine learning techniques and highlight some promising learning methods in recent studies, such as representation learning, deep learning, distributed and parallel learning, transfer learning, active learning, and kernel-based learning. Next, we focus on the analysis and discussions about the challenges and possible solutions of machine learning for big data. Following that, we investigate the close connections of machine learning with signal processing techniques for big data processing. Finally, we outline several open issues and research trends.

636 citations

Journal ArticleDOI
TL;DR: This model of reinforcement learning among cognitive strategies (RELACS) captures the 3 deviations, the learning curves, and the effect of information on uncertainty avoidance and outperforms other models in fitting the data and in predicting behavior in other experiments.
Abstract: Analysis of binary choice behavior in iterated tasks with immediate feedback reveals robust deviations from maximization that can be described as indications of 3 effects: (a) a payoff variability effect, in which high payoff variability seems to move choice behavior toward random choice; (b) underweighting of rare events, in which alternatives that yield the best payoffs most of the time are attractive even when they are associated with a lower expected return; and (c) loss aversion, in which alternatives that minimize the probability of losses can be more attractive than those that maximize expected payoffs. The results are closer to probability matching than to maximization. Best approximation is provided with a model of reinforcement learning among cognitive strategies (RELACS). This model captures the 3 deviations, the learning curves, and the effect of information on uncertainty avoidance. It outperforms other models in fitting the data and in predicting behavior in other experiments.

446 citations