Efficient Long-Range Transformers: You Need to Attend More, but Not Necessarily at Every Layer
- Award ID(s):
- 2226152
- PAR ID:
- 10529633
- Publisher / Repository:
- Association for Computational Linguistics
- Date Published:
- Page Range / eLocation ID:
- 2775 to 2786
- Format(s):
- Medium: X
- Location:
- Singapore
- Sponsoring Org:
- National Science Foundation
More Like this
No document suggestions found