babilong / info /RMT.md
booydar's picture
add new results and model description tab
95cd467 verified
|
raw
history blame
No virus
508 Bytes

RMT is a memory-augmented segment-level recurrent Transformer. We implement our memory mechanism as a wrapper for any Hugging Face model by adding special memory tokens to the input sequence. The model is trained to control both memory operations and sequence representations processing.

See: [ paper ] and [ code ] for Recurrent Memory Transformer implementation and training examples.