2020
DOI: 10.1609/aaai.v34i04.5977
|View full text |Cite
|
Sign up to set email alerts
|

Weighted Automata Extraction from Recurrent Neural Networks via Regression on State Spaces

Abstract: We present a method to extract a weighted finite automaton (WFA) from a recurrent neural network (RNN). Our method is based on the WFA learning algorithm by Balle and Mohri, which is in turn an extension of Angluin's classic L* algorithm. Our technical novelty is in the use of regression methods for the so-called equivalence queries, thus exploiting the internal state space of an RNN to prioritize counterexample candidates. This way we achieve a quantitative/weighted extension of the recent work by Weiss, Gold… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
23
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 33 publications
(25 citation statements)
references
References 16 publications
0
23
0
Order By: Relevance
“…Recurrent neural network (RNN) [29] is an extension of traditional feedforward neural network. It has the memory function of inputting information.…”
Section: Deep Learningmentioning
confidence: 99%
“…Recurrent neural network (RNN) [29] is an extension of traditional feedforward neural network. It has the memory function of inputting information.…”
Section: Deep Learningmentioning
confidence: 99%
“…Eisner (2001) describes an algorithm for estimating probabilities in a finite-state transducer from data using EM-based methods. Weiss, Goldberg, and Yahav (2019) and Okudono et al (2020) provide adaptations to the Weiss, Goldberg, and Yahav (2018) DFA extraction algorithm to yield weighted automata.…”
Section: Related Workmentioning
confidence: 99%
“…Recent work [ 22 ] also shows the equivalence between a 2-RNN with linear hidden activation and weighted automata. Related work [ 23 , 24 ] used weighted automata to explore the relationships between deep learning and grammatical inference. Since a 2-RNN has a 3D tensor weight, computation is more intensive.…”
Section: Preliminariesmentioning
confidence: 99%