Index of /publis/slides/2019_04__Presentation_IEEE_WCNC__Demo_ICT_2018

[ICO]NameLast modifiedSizeDescription
[PARENTDIR]Parent Directory  -  
[DIR]plots/2021-03-04 16:23 -  
[   ].gitignore2021-03-04 16:23 7  
[   ]Makefile2021-03-04 16:23 1.6K 
[TXT]README.md2021-03-04 16:23 8.1K 
[TXT]preprocess_tex.sh2021-03-04 16:23 119  
[TXT]slides.md2021-03-04 16:23 8.1K 
[   ]slides.pdf2021-03-04 16:23 2.2M 
[   ]slides.pdfpc2021-03-04 16:23 32  
[   ]slides_169.pdf2021-03-04 16:23 2.2M 
[   ]slides_169.pdfpc2021-03-04 16:23 36  
[   ]slides_pandoc.pdfpc2021-03-04 16:23 39  

title: GNU Radio Implementation of Multi-Armed bandits Learning for Internet-of-things Networks subtitle: IEEE WCNC 2019 author: Lilian Besson institute: SCEE Team, IETR, CentraleSupélec, Rennes date: Wednesday 17th of April, 2019

lang: english

IEEE WCNC 2019: "GNU Radio Implementation of Multi-Armed bandits Learning for Internet-of-things Networks"

Christophe Moy
@ IETR, Rennes
Emilie Kaufmann
@ CNRS & Inria, Lille
8% 14% 12% 16%

See our paper at HAL.Inria.fr/hal-02006825


Introduction

Joint work by R. Bonnefoi, L. Besson and C. Moy.


:timer_clock: Outline

1. Motivations

2. System Model

3. Multi-Armed Bandit (MAB) Model and Algorithms

4. GNU Radio Implementation

5. Results

Please :pray:

Ask questions at the end if you want!


1. Motivations

But...


2. System Model

Wireless network - In ISM band, centered at $433.5$ MHz (in Europe) - $K=4$ (or more) orthogonal channels

Gateway - One gateway, handling different objects - Communications with ALOHA protocol (without retransmission) - Objects send data for $1$s in one channel, wait for an acknowledgement for $1$s in same channel, use Ack as feedback: success / failure - Each object: communicate from time to time (e.g., every $10$ s) - Goal: max successful communications $\Longleftrightarrow$ max nb of received Ack


2. System Model

80%


Hypotheses

  1. We focus on one gateway

  2. Different IoT objects using the same standard are able to run a low-cost learning algorithm on their embedded CPU

  3. The spectrum occupancy generated by the rest of the environment is assumed to be stationary

  4. And non uniform traffic: some channels are more occupied than others.


3. Multi-Armed Bandits (MAB)

3.1. Model

3.2. Algorithms


3.1. Multi-Armed Bandits Model

Why is it famous?

Simple but good model for exploration/exploitation dilemma.


3.2. Multi-Armed Bandits Algorithms

Often "index based"

Example: "Follow the Leader"


Upper Confidence Bounds algorithm (UCB)

Parameter $\alpha$: tradeoff exploration vs exploitation


4. GNU Radio Implementation

4.1. Physical layer and protocol

4.2. Equipment

4.3. Implementation

4.4. User interface


4.1. Physical layer and protocol

Very simple ALOHA-based protocol

An uplink message $\,\nearrow\,$ is made of... - a preamble (for phase synchronization) - an ID of the IoT object, made of QPSK symbols $1\pm1j \in \mathbb{C}$ - then arbitrary data, made of QPSK symbols $1\pm1j \in \mathbb{C}$

A downlink (Ack) message $\,\swarrow\,$ is then... - same preamble - the same ID (so a device knows if the Ack was sent for itself or not)


4.2. Equipment

$\geq3$ USRP boards

1: gateway 2: traffic generator 3: IoT dynamic objects (as much as we want)

55%


4.3. Implementation

(examples...)


Flowchart of the random traffic generator

95%


Flowchart of the IoT gateway

95%


Flowchart of the IoT dynamic object

95%


4.4. User interface of our demonstration

→ See video of the demo: YouTu.be/HospLNQhcMk

30%


5. Example of simulation and results

On an example of a small IoT network: - with $K=4$ channels, - and non uniform "background" traffic (other networks), with a repartition of $15\%$, $10\%$, $2\%$, $1\%$

  1. $\Longrightarrow$ the uniform access strategy obtains a successful communication rate of about $40\%$.

  2. About $400$ communication slots are enough for the learning IoT objects to reach a successful communication rate close to $80\%$, using UCB algorithm or another one (Thompson Sampling).

Note: similar gains of performance were obtained in other scenarios.


Illustration

70%


6. Summary

We showed

  1. The system model and PHY/MAC layers of our demo
  2. The Multi-Armed Bandits model and algorithms
  3. Our demonstration written in C++ and GNU Radio Companion
  4. Empirical results: the proposed approach works fine, is simple to set-up in existing networks, and give impressive results!

Take home message

Dynamically reconfigurable IoT objects can learn on their own to favor certain channels, if the environment traffic is not uniform between the $K$ channels, and greatly improve their succesful communication rates!


6. Future works

We are exploring these directions


6. Conclusion

→ See our paper: HAL.Inria.fr/hal-02006825

→ See video of the demo: YouTu.be/HospLNQhcMk

→ See the code of our demo:

Under GPL open-source license, for GNU Radio: bitbucket.org/scee_ietr/malin-multi-arm-bandit-learning-for-iot-networks-with-grc

Thanks for listening !