Deep Neural Networks for Estimation and Inference

S-Tier
Journal: Econometrica
Year: 2021
Volume: 89
Issue: 1
Pages: 181-213

Authors (3)

Max H. Farrell (University of California-Santa...) Tengyuan Liang (not in RePEc) Sanjog Misra (not in RePEc)

Score contribution per author:

2.681 = (α=2.01 / 3 authors) × 4.0x S-tier

α: calibrated so average coauthorship-adjusted count equals average raw count

Abstract

We study deep neural networks and their use in semiparametric inference. We establish novel nonasymptotic high probability bounds for deep feedforward neural nets. These deliver rates of convergence that are sufficiently fast (in some cases minimax optimal) to allow us to establish valid second‐step inference after first‐step estimation with deep learning, a result also new to the literature. Our nonasymptotic high probability bounds, and the subsequent semiparametric inference, treat the current standard architecture: fully connected feedforward neural networks (multilayer perceptrons), with the now‐common rectified linear unit activation function, unbounded weights, and a depth explicitly diverging with the sample size. We discuss other architectures as well, including fixed‐width, very deep networks. We establish the nonasymptotic bounds for these deep nets for a general class of nonparametric regression‐type loss functions, which includes as special cases least squares, logistic regression, and other generalized linear models. We then apply our theory to develop semiparametric inference, focusing on causal parameters for concreteness, and demonstrate the effectiveness of deep learning with an empirical application to direct mail marketing.

Technical Details

RePEc Handle
repec:wly:emetrp:v:89:y:2021:i:1:p:181-213
Journal Field
General
Author Count
3
Added to Database
2026-01-25