4.3 Article

Deep reinforcement learning for the olfactory search POMDP: a quantitative benchmark

Journal

EUROPEAN PHYSICAL JOURNAL E
Volume 46, Issue 3, Pages -

Publisher

SPRINGER
DOI: 10.1140/epje/s10189-023-00277-8

Keywords

-

Ask authors/readers for more resources

This article introduces a sequential decision-making problem that simulates the task of insects searching for odor sources in turbulence and explores the application of its solutions to sniffer robots. Due to the infeasibility of exact solutions, the challenge lies in finding the best possible approximate solutions with reasonable computational cost. The researchers provide a quantitative benchmarking of a deep reinforcement learning-based solver against traditional POMDP approximate solvers, demonstrating that deep reinforcement learning is a competitive alternative for generating lightweight policies suitable for robots.
The olfactory search POMDP (partially observable Markov decision process) is a sequential decision-making problem designed to mimic the task faced by insects searching for a source of odor in turbulence, and its solutions have applications to sniffer robots. As exact solutions are out of reach, the challenge consists in finding the best possible approximate solutions while keeping the computational cost reasonable. We provide a quantitative benchmarking of a solver based on deep reinforcement learning against traditional POMDP approximate solvers. We show that deep reinforcement learning is a competitive alternative to standard methods, in particular to generate lightweight policies suitable for robots.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.3
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available