Skip to content

LividWo/Revisit-MMT

Repository files navigation

Good for Misconceived Reasons: An Empirical Revisiting on the Need for Visual Context in Multimodal Machine Translation

This repo contains code needed to replicate our findings in the ACL’2021 paper as titled. Our implementation is based on FairSeq.

Setup conda environment (recommanded)

  • conda create --name revisit python=3.7.6
  • conda activate revisit
  • pip install transformers==3.0.2
  • conda install pytorch==1.5.0 torchvision==0.6.0 cudatoolkit=10.1 -c pytorch
  • cd Revisit-MMT/
  • pip install --editable ./

Resources

File Name Description Download
resnet50-avgpool.npy pre-extracted image features, each image is represented as a 2048-dimensional vector. Link
retriever.bin pre-trained text->image retriever Link
checkpoints pre-trained Transformer/Gated Fusion/RMMT on Multi30K En-De for quick reproduction Link

Example Usage

# We have included pre-processed (BPE,re-index image) raw data from Multi30K En-De/Fr in the repo, 
# with the following format (take train set as an example):

train.en # source sentence
train.de # target sentences
train.vision.en # image id associated with each source sentence
# We re-index image id from 0 to #number of images for convinence
train.bert.en # the source sentence (without bpe) used for retrieval

Transformer basleine on Multi30K En-De

Gated Fusion on Multi30K En-De

PS: you need to download pre-extracted visual features to train a Gated Fusion model.

RMMT on Multi30K En-De

PS: you need to download pre-extracted visual features and a pre-trained image retriever to tran RMMT.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages