CENTRAL AND NONCENTRAL LIMIT THEOREMS ARISING FROM THE SCATTERING TRANSFORM AND ITS NEURAL ACTIVATION GENERALIZATION

Gi Ren Liu, Yuan Chung Sheu, Hau Tieng Wu

Research output: Contribution to journalArticlepeer-review

Abstract

Motivated by the analysis of complicated time series, we examine a generalization of the scattering transform that includes broad neural activation functions. This generalization is the neural activation scattering transform (NAST). NAST comprises a sequence of "neural processing units," each of which applies a high pass filter to the input from the previous layer followed by a composition with a nonlinear function as the output to the next neuron. Here, the nonlinear function models how a neuron gets excited by the input signal. In addition to showing properties like nonexpansion, horizontal translational invariability, and insensitivity to local deformation, we explore the statistical properties of the second-order NAST of a Gaussian process with various dependence structures and its interaction with the chosen wavelets and activation functions. We also provide central limit theorem (CLT) and non-CLT results. Numerical simulations demonstrate the developed theorems. Our results explain how NAST processes complicated time series, paving a way toward statistical inference based on NAST for real-world applications.

Original languageEnglish (US)
Pages (from-to)1170-1213
Number of pages44
JournalSIAM Journal on Mathematical Analysis
Volume55
Issue number2
DOIs
StatePublished - Apr 2023

Keywords

  • deformation insensitive
  • dependent processes
  • large scaling limits
  • long range dependence
  • neural activation scattering transform
  • scattering moments
  • translation invariant

ASJC Scopus subject areas

  • Analysis
  • Computational Mathematics
  • Applied Mathematics

Fingerprint

Dive into the research topics of 'CENTRAL AND NONCENTRAL LIMIT THEOREMS ARISING FROM THE SCATTERING TRANSFORM AND ITS NEURAL ACTIVATION GENERALIZATION'. Together they form a unique fingerprint.

Cite this