4.8 Article

Privacy-Cost Management in Smart Meters With Mutual-Information-Based Reinforcement Learning

Journal

IEEE INTERNET OF THINGS JOURNAL
Volume 9, Issue 22, Pages 22389-22398

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/JIOT.2021.3128488

Keywords

Cyber-physical system (CPS); deep double Q-learning (DDQL); deep reinforcement learning (DRL); Internet of Things (IoT); mutual information (MI); smart meters (SMs) privacy

Funding

  1. Hydro-Quebec
  2. Natural Sciences and Engineering Research Council of Canada
  3. McGill University [IRCPJ406021-14]
  4. European Union's Horizon 2020 Research and Innovation Programme through the Marie Sklodowska-Curie Grant [792464]
  5. Marie Curie Actions (MSCA) [792464] Funding Source: Marie Curie Actions (MSCA)

Ask authors/readers for more resources

This article presents a novel method to learn the privacy-cost management unit (PCMU) policy using deep reinforcement learning (DRL). The mutual information (MI) between the user's demand load and the masked load seen by the power grid is adopted as a reliable and general privacy measure. The method combines a neural network to estimate the MI-based reward signal with a model-free DRL algorithm known as the deep double Q-learning (DDQL) method. Empirical assessment using actual smart meters (SMs) data set shows significant improvements over simpler privacy measures.
The rapid development and expansion of the Internet of Things (IoT) paradigm has drastically increased the collection and exchange of data between sensors and systems, a phenomenon that raises serious privacy concerns in some domains. In particular, smart meters (SMs) share fine-grained electricity consumption of households with utility providers that can potentially violate users' privacy as sensitive information is leaked through the data. In order to enhance privacy, electricity consumers can exploit the availability of physical resources such as a rechargeable battery (RB) to shape their power demand as dictated by a privacy-cost management unit (PCMU). In this article, we present a novel method to learn the PCMU policy using deep reinforcement learning (DRL). We adopt the mutual information (MI) between the user's demand load and the masked load seen by the power grid as a reliable and general privacy measure. Unlike previous studies, we model the whole temporal correlation in the data to learn the MI in its general form and use a neural network to estimate the MI-based reward signal to guide the PCMU learning process. This approach is combined with a model-free DRL algorithm known as the deep double Q-learning (DDQL) method. The performance of the complete DDQL-MI algorithm is assessed empirically using an actual SMs data set and compared with simpler privacy measures. Our results show significant improvements over state-of-the-art privacy-aware demand shaping methods.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.8
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available