Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) 2018
DOI: 10.18653/v1/p18-1028
|View full text |Cite
|
Sign up to set email alerts
|

Bridging CNNs, RNNs, and Weighted Finite-State Machines

Abstract: Recurrent and convolutional neural networks comprise two distinct families of models that have proven to be useful for encoding natural language utterances. In this paper we present SoPa, a new model that aims to bridge these two approaches. SoPa combines neural representation learning with weighted finite-state automata (WFSAs) to learn a soft version of traditional surface patterns. We show that SoPa is an extension of a one-layer CNN, and that such CNNs are equivalent to a restricted version of SoPa, and ac… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
19
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
3
3

Relationship

1
5

Authors

Journals

citations
Cited by 16 publications
(19 citation statements)
references
References 53 publications
0
19
0
Order By: Relevance
“…To apply their method, one has to modify the structure of an RNN before training, while our method does not need any special structure to RNNs and can be applied to already trained RNNs. Schwartz, Thomson, and Smith (2018) introduce a neural network architecture that can represent (restricted forms of) CNNs and RNNs. WFAs could also be expressed by their architecture, but extraction of automata is out of their interest.…”
Section: Related Workmentioning
confidence: 99%
“…To apply their method, one has to modify the structure of an RNN before training, while our method does not need any special structure to RNNs and can be applied to already trained RNNs. Schwartz, Thomson, and Smith (2018) introduce a neural network architecture that can represent (restricted forms of) CNNs and RNNs. WFAs could also be expressed by their architecture, but extraction of automata is out of their interest.…”
Section: Related Workmentioning
confidence: 99%
“…Our definition is equivalent, giving the weight functions value0 wherever they were undefined. 3 ε-transitions can be handled with a slight modification (Schwartz et al, 2018). Note though that if A contains a cycle of ε-transitions, then either K must follow the star semiring laws (Kuich and Salomaa, 1986), or the number of consecutive ε-transitions allowed must be capped.…”
Section: ε /mentioning
confidence: 99%
“…So far our discussion has centered on B, a twostate WFSA capturing unigram patterns (Example 5). In the same spirit as going from unigram to n-gram features, one can use WFSAs with more states to capture longer patterns (Schwartz et al, 2018). In this section we augment B by introducing more states, and explore its relationship to some neural architectures motivated by n-gram features.…”
Section: More Than Two Statesmentioning
confidence: 99%
See 2 more Smart Citations