2021
DOI: 10.48550/arxiv.2107.14417
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Creating Powerful and Interpretable Models with Regression Networks

Abstract: As the discipline has evolved, research in machine learning has been focused more and more on creating more powerful neural networks, without regard for the interpretability of these networks. Such "black-box models" yield state-of-the-art results, but we cannot understand why they make a particular decision or prediction. Sometimes this is acceptable, but often it is not. We propose a novel architecture, Regression Networks, which combines the power of neural networks with the understandability of regression … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(4 citation statements)
references
References 12 publications
(19 reference statements)
0
4
0
Order By: Relevance
“…HNN-F could be seen to be on the fringe of it, with its multiplicative effects that would certainly be an odd modeling choice without a time-varying unobserved components regression in mind. Closely related, Agarwal et al (2020), O'Neill et al (2021), and Rügamer et al (2020 all develop an architecture inspired from generalized additive models to enhance interpretability in deep networks for generic tasks. While these articles certainly tackle some of the opacity issues coming from nonparametric nonlinear estimation with deep learning, none address those that are inherent to any non-sparse high-dimensional (even linear) regression-i.e., that analyzing partial derivatives of 200 things that typically co-move together unfortunately borders on the meaningless.…”
Section: Neural Network and Macroeconomic Forecastingmentioning
confidence: 99%
“…HNN-F could be seen to be on the fringe of it, with its multiplicative effects that would certainly be an odd modeling choice without a time-varying unobserved components regression in mind. Closely related, Agarwal et al (2020), O'Neill et al (2021), and Rügamer et al (2020 all develop an architecture inspired from generalized additive models to enhance interpretability in deep networks for generic tasks. While these articles certainly tackle some of the opacity issues coming from nonparametric nonlinear estimation with deep learning, none address those that are inherent to any non-sparse high-dimensional (even linear) regression-i.e., that analyzing partial derivatives of 200 things that typically co-move together unfortunately borders on the meaningless.…”
Section: Neural Network and Macroeconomic Forecastingmentioning
confidence: 99%
“…Similar approaches using neural networks to construct GAMs and to perform shape functions are the basis of methods called GAMI-Net [16] and AxNNs [17]. An architecture called the regression network which can be also regarded as a modification of NAM is proposed by O'Neill et al [61].…”
Section: Related Workmentioning
confidence: 99%
“…It is important to note that the number of neural networks to take into account pairwise interactions does not increase in the proposed approach. This is a very important difference of the proposed approach from NAM [15] and its modifications, for example, regression networks [61]. Another advantage of the above approach in comparison with other approaches is that we get shape functions of pairs of features instead of some coefficients.…”
Section: Identifying Pairwise Interactionsmentioning
confidence: 99%
See 1 more Smart Citation