liu.seSearch for publications in DiVA
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Multiagent Reinforcement Learning Meets Random Access in Massive Cellular Internet of Things
Linköping University, Department of Electrical Engineering, Communication Systems. Linköping University, Faculty of Science & Engineering. Virgnia Tech, VA 24060 USA.
Virginia Tech, VA 24060 USA.
Virginia Tech, VA 24060 USA.
Virginia Tech, VA 24060 USA.
2021 (English)In: IEEE Internet of Things Journal, ISSN 2327-4662, Vol. 8, no 24, p. 17417-17428Article in journal (Refereed) Published
Abstract [en]

Internet of Things (IoT) has attracted considerable attention in recent years due to its potential of interconnecting a large number of heterogeneous wireless devices. However, it is usually challenging to provide reliable and efficient random access control when massive IoT devices are trying to access the network simultaneously. In this article, we investigate methods to introduce intelligent random access management for a massive cellular IoT network to reduce access latency and access failures. Toward this end, we introduce two novel frameworks, namely, local device selection (LDS) and intelligent preamble selection (IPS). LDS enables local communication between neighboring devices to provide cluster-wide cooperative congestion control, which leads to a better distribution of the access intensity under bursty traffics. Taking advantage of the capability of reinforcement learning in developing cooperative multiagent policies, IPS is introduced to enable the optimization of the preamble selection policy in each IoT clusters. To handle the exponentially growing action space in IPS, we design a novel reinforcement learning structure, named branching actor-critic, to ensure that the output size of the underlying neural networks only grows linearly with the number of action dimensions. Simulation results indicate that the introduced mechanism achieves much lower access delays with fewer access failures in various realistic scenarios of interests.

Place, publisher, year, edition, pages
IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC , 2021. Vol. 8, no 24, p. 17417-17428
Keywords [en]
Internet of Things; Delays; Reinforcement learning; Quality of service; IP networks; Access control; Wireless communication; Internet of Things (IoT); massive connectivity; multiagent reinforcement learning (MARL); random access
National Category
Communication Systems
Identifiers
URN: urn:nbn:se:liu:diva-181906DOI: 10.1109/JIOT.2021.3081692ISI: 000728152700019OAI: oai:DiVA.org:liu-181906DiVA, id: diva2:1622075
Note

Funding Agencies|U.S. National Science Foundation (NSF)National Science Foundation (NSF) [ECCS-1811497, CCF-1937487, CNS-2003059]

Available from: 2021-12-21 Created: 2021-12-21 Last updated: 2021-12-21

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full text

Search in DiVA

By author/editor
Bai, Jianan
By organisation
Communication SystemsFaculty of Science & Engineering
In the same journal
IEEE Internet of Things Journal
Communication Systems

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 261 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • oxford
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf