Abstract
It is known that successful load disaggregation via deep learning relies on a large number of labeled data to train the deep neural networks. However, it is hard and expensive to acquire a large amount of appliance-level power data or ON/OFF labels. For overcoming such weakness, in this paper, unsupervised pre-training is applied to the state-of-the-art sequence-to-point (s2p) deep learning approach for NILM, where labeling is not required in pre-training. In the proposed method, the s2p deep neural network is initially pre-trained on unlabeled aggregate power readings for other houses, and then fine-tuned on a small set of aggregate power for the target house labeled by individual appliance monitoring. Finally, the generated network is applied to testing with the aggregate for the target house as input and outputting the power signal for the target load. The proposed method is validated on the UK REFIT dataset1, benchmark with s2p in two popular evaluation metrics. Experimental results show that the proposed unsupervised pre-training effectively improves NILM performance of the deep neural network with a lack of labeled training data.
Original language | English |
---|---|
Title of host publication | 5th IEEE Conference on Energy Internet and Energy System Integration |
Subtitle of host publication | Energy Internet for Carbon Neutrality, EI2 2021 |
Publisher | Institute of Electrical and Electronics Engineers Inc. |
Pages | 3224-3229 |
Number of pages | 6 |
ISBN (Electronic) | 9781665434256 |
DOIs | |
Publication status | Published - 2021 |
Event | 5th IEEE Conference on Energy Internet and Energy System Integration, EI2 2021 - Taiyuan, China Duration: 22 Oct 2021 → 25 Oct 2021 |
Publication series
Name | 5th IEEE Conference on Energy Internet and Energy System Integration: Energy Internet for Carbon Neutrality, EI2 2021 |
---|
Conference
Conference | 5th IEEE Conference on Energy Internet and Energy System Integration, EI2 2021 |
---|---|
Country/Territory | China |
City | Taiyuan |
Period | 22/10/21 → 25/10/21 |
Bibliographical note
Funding Information:This work is supported in part by the National Natural Science Foundation of China (Key Program), NSFC-SGCC (U2066207).
Keywords
- deep neural network
- non-intrusive load monitoring
- sequence-to-point learning
- unsupervised pre-training