Open Access
Open access
Eurasip Journal on Audio, Speech, and Music Processing, volume 2024, issue 1, publication number 56

Point neuron learning: a new physics-informed neural network architecture

Hanwen Bi 1
Thushara D. Abhayapala 1
1
 
Audio & Acoustics Signal Processing Group, School of Engineering, The Australian National University, Canberra, Australia
Publication typeJournal Article
Publication date2024-11-04
scimago Q2
SJR0.414
CiteScore4.1
Impact factor1.7
ISSN16874714, 16874722
Abstract
Machine learning and neural networks have advanced numerous research domains, but challenges such as large training data requirements and inconsistent model performance hinder their application in certain scientific problems. To overcome these challenges, researchers have investigated integrating physics principles into machine learning models, mainly through (i) physics-guided loss functions, generally termed as physics-informed neural networks, and (ii) physics-guided architectural design. While both approaches have demonstrated success across multiple scientific disciplines, they have limitations including being trapped to a local minimum, poor interpretability, and restricted generalizability beyond sampled data range. This paper proposes a new physics-informed neural network (PINN) architecture that combines the strengths of both approaches by embedding the fundamental solution of the wave equation into the network architecture, enabling the learned model to strictly satisfy the wave equation. The proposed point neuron learning method can model an arbitrary sound field based on microphone observations without any dataset. Compared to other PINN methods, our approach directly processes complex numbers, offers better interpretability, and can be generalized to out-of-sample scenarios. We evaluate the versatility of the proposed architecture by a sound field reconstruction problem in a reverberant environment. Results indicate that the point neuron method outperforms two competing methods and can efficiently handle noisy environments with sparse microphone observations.
Found 

Are you a researcher?

Create a profile to get free access to personal recommendations for colleagues and new articles.
Share
Cite this
GOST | RIS | BibTex
Found error?