Skip to content

Latest commit

 

History

History
13 lines (8 loc) · 431 Bytes

File metadata and controls

13 lines (8 loc) · 431 Bytes

Mask Attention Networks

This repo contains the codes and pretrained models for our paper:

Mask Attention Networks: Rethinking and Strengthen Transformer

The two sub-directories includes reproducible codes and instructions for the machine translation and abstractive summarization. Please find the READMEs in the sub-directories for the detailed instructions for reproduction.