2020 6th International Conference on Robotics and Artificial Intelligence 2020
DOI: 10.1145/3449301.3449332
|View full text |Cite
|
Sign up to set email alerts
|

Entropy Targets for Adaptive Distillation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
0
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(2 citation statements)
references
References 6 publications
0
0
0
Order By: Relevance
“…Additionally, models like BERT have shown improved performance with more parameters, leading to gradual increases in model size. Thus, research is being conducted to reduce both the number of parameters and the computational complexity of such models [5][6][7][8].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Additionally, models like BERT have shown improved performance with more parameters, leading to gradual increases in model size. Thus, research is being conducted to reduce both the number of parameters and the computational complexity of such models [5][6][7][8].…”
Section: Introductionmentioning
confidence: 99%
“…Feature Representations Transferring features in a processed rather than simple form can lead to more effective results [8,[21][22][23]. Examples include transformation methods such as cosine representation or Euclidean distance.…”
mentioning
confidence: 99%