2022
DOI: 10.1007/s10994-022-06163-2
|View full text |Cite
|
Sign up to set email alerts
|

Extracting automata from recurrent neural networks using queries and counterexamples (extended version)

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
15
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 16 publications
(23 citation statements)
references
References 41 publications
0
15
0
Order By: Relevance
“…To generate a state diagram from an RNN model, we develop a method that clusters semantically related hidden states of the RNN model into an abstract state. Our work is inspired by the modelbased analysis of stateful RNNs [15,16,47,51,61]. These works apply various techniques to extract interpretable state transition…”
Section: Design and Implementation 51 State Abstractionmentioning
confidence: 99%
“…To generate a state diagram from an RNN model, we develop a method that clusters semantically related hidden states of the RNN model into an abstract state. Our work is inspired by the modelbased analysis of stateful RNNs [15,16,47,51,61]. These works apply various techniques to extract interpretable state transition…”
Section: Design and Implementation 51 State Abstractionmentioning
confidence: 99%
“…On a more applied note, the MQ+EQ model has recently been used for recurrent and binarized neural networks (Weiss et al, 2018(Weiss et al, , 2019Okudono et al, 2020;Shih et al, 2019), and interpretability (Camacho and McIlraith, 2019). It is also worth noting that the MQ learning model has been criticized by the applied machine learning community, as labels can be queried in the whole input space, irrespective of the distribution that generates the data.…”
Section: Learning With Queriesmentioning
confidence: 99%
“…Another possibility is a hypothesis class again consisting of a large key-value memory, but now coupled with a small RASP program [WGY21] or other textual program. Yet another possibility is a target class consisting of a large logical circuit where a significant fraction of the nodes correspond to "human-understandable" concepts (e.g., words from a dictionary, or concepts encoded by a more trusted LLM).…”
Section: Extensions and Future Directionsmentioning
confidence: 99%