2020
DOI: 10.48550/arxiv.2006.05103
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

The Curious Case of Convex Neural Networks

Abstract: In this paper, we investigate a constrained formulation of neural networks where the output is a convex function of the input. We show that the convexity constraints can be enforced on both fully connected and convolutional layers, making them applicable to most architectures. The convexity constraints include restricting the weights (for all but the first layer) to be non-negative and using a non-decreasing convex activation function. Albeit simple, these constraints have profound implications on the generali… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(3 citation statements)
references
References 15 publications
0
3
0
Order By: Relevance
“…Proposed recently in [62], they have known a growing interest thanks to their original properties. They have found applications for classification problems [63], for new control strategies [64] and to approximate the convex functions space [65]. For a given input x ∈ R d , the output z L of the network with L layers of D hidden neurons is defined recursively by the following expression for i ∈ [0, 1, .., L] :…”
Section: Minimax Formulation Of the Dual Monge-kantorovich Problemmentioning
confidence: 99%
“…Proposed recently in [62], they have known a growing interest thanks to their original properties. They have found applications for classification problems [63], for new control strategies [64] and to approximate the convex functions space [65]. For a given input x ∈ R d , the output z L of the network with L layers of D hidden neurons is defined recursively by the following expression for i ∈ [0, 1, .., L] :…”
Section: Minimax Formulation Of the Dual Monge-kantorovich Problemmentioning
confidence: 99%
“…In another work, [16] describe an ICNN architecture for binary classification, where the output is a two dimensional vector. Each element of this vector is a convex function of the inputs.…”
Section: Cdinn-convex Difference Neural Networkmentioning
confidence: 99%
“…In particular, when biases are ignored, even identity mapping cannot be learnt by ICNN without pass-through layers. One approach to address this problem is suggested in [16], where it is proposed to use Leaky ReLU or ELU as activation function. ELU activation function is given by output = ( ⋅( −1) if < 0, and if >= 0) and Leaky ReLU is given by output = ( if < 0, and if >= 0), where is the input and is a hyperparameter.…”
Section: Pass-through Layersmentioning
confidence: 99%