4.5 Article

Error bounds for approximations with deep ReLU neural networks in Ws,p norms

期刊

ANALYSIS AND APPLICATIONS
卷 18, 期 5, 页码 803-859

出版社

WORLD SCIENTIFIC PUBL CO PTE LTD
DOI: 10.1142/S0219530519410021

关键词

Deep neural networks; approximation rates; Sobolev spaces; PDEs; curse of dimension

资金

  1. Research Training Group Differential Equationand Data-driven Models in Life Sciences and Fluid Dynamics: An Interdisciplinary Research Training Group (DAEDALUS) - German Research Foundation (DFG) [GRK 2433]
  2. Bundesministerium fur Bildung und Forschung (BMBF) through the Berliner Zentrum for Machine Learning (BZML)
  3. Deutsche Forschungsgemeinschaft (DFG) [CRC 1114, CRC/TR 109, RTG 2433, RTG 2260, SPP 1798]
  4. Berlin Mathematics Research Center MATH+ [EF1-1, EF1-4]
  5. Einstein Foundation Berlin
  6. DFG Research Fellowship Shearlet-based energy functionals for anisotropic phase-field methods

向作者/读者索取更多资源

We analyze to what extent deep Rectified Linear Unit (ReLU) neural networks can efficiently approximate Sobolev regular functions if the approximation error is measured with respect to weaker Sobolev norms. In this context, we first establish upper approximation bounds by ReLU neural networks for Sobolev regular functions by explicitly constructing the approximate ReLU neural networks. Then, we establish lower approximation bounds for the same type of function classes. A trade-off between the regularity used in the approximation norm and the complexity of the neural network can be observed in upper and lower bounds. Our results extend recent advances in the approximation theory of ReLU networks to the regime that is most relevant for applications in the numerical analysis of partial differential equations.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.5
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

Article Computer Science, Theory & Methods

Topological Properties of the Set of Functions Generated by Neural Networks of Fixed Size

Philipp Petersen, Mones Raslan, Felix Voigtlaender

Summary: The set of functions that can be implemented by neural networks of a fixed size has many undesirable properties, such as being highly non-convex and not closed with respect to Lp-norms. This may lead to issues in the training procedure of deep learning, such as no guaranteed convergence, explosion of parameters, and slow convergence.

FOUNDATIONS OF COMPUTATIONAL MATHEMATICS (2021)

Editorial Material Mathematics, Applied

Preface to the Special Issue of ILAS 2019: Linear Algebra without Borders, The 22ndILAS Conference, Rio de Janeiro, Brazil, July 08-12, 2019

Nair Abreu, Christoph Helmberg, Gitta Kutyniok, Vilmar Trevisan

LINEAR ALGEBRA AND ITS APPLICATIONS (2021)

Article Mathematics, Applied

Efficient approximation of solutions of parametric linear transport equations by ReLU DNNs

Fabian Laakmann, Philipp Petersen

Summary: Deep neural networks using ReLU activation function can efficiently approximate solutions for various types of parametric linear transport equations, even for high-dimensional and non-smooth cases. Through their inherent compositionality, these networks can resolve the characteristic flow underlying the transport equations and achieve approximation rates independent of the parameter dimension.

ADVANCES IN COMPUTATIONAL MATHEMATICS (2021)

Article Mathematics, Applied

Analysis of simultaneous inpainting and geometric separation based on sparse decomposition

Van Tiep Do, Ron Levie, Gitta Kutyniok

Summary: This paper discusses the problem of superposition of different geometric characteristics in natural images and proposes a method for simultaneously decomposing and inpainting the image. The method utilizes an l(1) minimization approach with two dictionaries to sparsify and inpaint the two parts of the image.

ANALYSIS AND APPLICATIONS (2022)

Article Mathematics

Approximation Spaces of Deep Neural Networks

Remi Gribonval, Gitta Kutyniok, Morten Nielsen, Felix Voigtlaender

Summary: This study focuses on the expressivity of deep neural networks, analyzing the complexity of networks based on the number of connections or neurons. The research investigates a class of functions whose error of best approximation decays at a certain rate with increasing complexity budget. The findings suggest that certain types of skip connections do not affect the resulting approximation spaces, and discuss the impact of network nonlinearity and depth on these spaces.

CONSTRUCTIVE APPROXIMATION (2022)

Article Mathematics

A Theoretical Analysis of Deep Neural Networks and Parametric PDEs

Gitta Kutyniok, Philipp Petersen, Mones Raslan, Reinhold Schneider

Summary: By utilizing the low dimensionality of the solution manifold, we have constructed neural networks for approximating the solution maps of parametric partial differential equations with significantly superior approximation rates compared to classical results.

CONSTRUCTIVE APPROXIMATION (2022)

Article Engineering, Electrical & Electronic

RadioUNet: Fast Radio Map Estimation With Convolutional Neural Networks

Ron Levie, Cagkan Yapar, Gitta Kutyniok, Giuseppe Caire

Summary: This paper proposes a deep learning method for accurately estimating the pathloss function in urban environments, showing significant improvement over previous methods.

IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS (2021)

Article Mathematics, Applied

Deep microlocal reconstruction for limited-angle tomography

Hector Andrade-Loarca, Gitta Kutyniok, Ozan Oektem, Philipp Petersen

Summary: In this paper, we propose a deep-learning-based algorithm that tackles both the reconstruction and wavefront set extraction problems in tomographic imaging. By using the wavefront set information of X-ray data, the algorithm requires the neural networks to extract the correct ground truth wavefront set and image simultaneously, leading to improved reconstruction results.

APPLIED AND COMPUTATIONAL HARMONIC ANALYSIS (2022)

Article Mathematics, Applied

Transferability of graph neural networks: An extended graphon approach

Sohir Maskey, Ron Levie, Gitta Kutyniok

Summary: We study the transferability of spectral graph convolutional neural networks (GCNNs) and provide several contributions. Firstly, we prove that any fixed GCNN with continuous filters is transferable based on graphon analysis. Secondly, we establish the transferability for graphs that approximate unbounded graphon shift operators. Lastly, we obtain non-asymptotic approximation results and prove the linear stability of GCNNs. These findings extend the current state-of-the-art results in transferability of GCNNs.

APPLIED AND COMPUTATIONAL HARMONIC ANALYSIS (2023)

Editorial Material Psychology, Educational

ChatGPT for good? On opportunities and challenges of large language models for education

Enkelejda Kasneci, Kathrin Sessler, Stefan Kuechemann, Maria Bannert, Daryna Dementieva, Frank Fischer, Urs Gasser, Georg Groh, Stephan Guennemann, Eyke Huellermeier, Stepha Krusche, Gitta Kutyniok, Tilman Michaeli, Claudia Nerdel, Juergen Pfeffer, Oleksandra Poquet, Michael Sailer, Albrecht Schmidt, Tina Seidel, Matthias Stadler, Jochen Weller, Jochen Kuhn, Gjergji Kasneci

Summary: Large language models are a significant advancement in AI, and despite criticism and bans, they are here to stay. This commentary discusses the potential benefits and challenges of using these models in education, emphasizing the need for competencies, critical thinking, and strategies for fact checking. Teachers and learners must understand the technology and its limitations, and educational systems need clear strategies and pedagogical approaches to maximize the benefits. Challenges such as bias, human oversight, and potential misuse can provide insights and opportunities for early education on societal biases and risks of AI applications.

LEARNING AND INDIVIDUAL DIFFERENCES (2023)

Proceedings Paper Computer Science, Artificial Intelligence

A Rate-Distortion Framework for Explaining Black-Box Model Decisions

Stefan Kolek, Duc Anh Nguyen, Ron Levie, Joan Bruna, Gitta Kutyniok

Summary: This paper presents a framework called Rate-Distortion Explanation (RDE) for explaining black-box model decisions. The framework is based on perturbations of the target input signal and can be applied to any differentiable pre-trained model, such as neural networks. Experiments demonstrate the adaptability of the framework to diverse data modalities, including images, audio, and physical simulations of urban environments.

XXAI - BEYOND EXPLAINABLE AI: International Workshop, Held in Conjunction with ICML 2020, July 18, 2020, Vienna, Austria, Revised and Extended Papers (2022)

Proceedings Paper Computer Science, Artificial Intelligence

Neural Tangent Kernel Beyond the Infinite-Width Limit: Effects of Depth and Initialization

Mariia Seleznova, Gitta Kutyniok

Summary: In this paper, the NTK of fully-connected ReLU networks with depth comparable to width is studied. It is found that the properties of NTK depend significantly on the depth-to-width ratio and the distribution of parameters at initialization. The importance of the ordered phase in the hyperparameter space for the stability of NTK is also highlighted.

INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162 (2022)

Proceedings Paper Computer Science, Artificial Intelligence

Cartoon Explanations of Image Classifiers

Stefan Kolek, Duc Anh Nguyen, Ron Levie, Joan Bruna, Gitta Kutyniok

Summary: CartoonX is a novel explanation method tailored towards image classifiers that extracts relevant piecewise smooth parts of an image in the wavelet domain to reveal valuable explanatory information, particularly for misclassifications, achieving lower distortion with fewer coefficients than state-of-the-art methods.

COMPUTER VISION, ECCV 2022, PT XII (2022)

Proceedings Paper Acoustics

LOCUNET: FAST URBAN POSITIONING USING RADIO MAPS AND DEEP LEARNING

Cagkan Yapar, Ron Levie, Gitta Kutyniok, Giuseppe Caire

Summary: This paper presents a solution to the problem of localization in dense urban scenarios. It introduces a deep learning-based method called LocUNet, which relies solely on the received signal strength from base stations for localization, offering high accuracy and robustness. By utilizing estimated pathloss radio maps and a neural network-based radio map estimator, the proposed method is suitable for real-time applications.

2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP) (2022)

暂无数据