You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
@InProceedings{tanDIDAN2020,
author={Reuben Tan and Bryan A. Plummer and Kate Saenko},
title={Detecting Cross-Modal Inconsistency to Defend Against Neural Fake News},
booktitle={Empirical Methods in Natural Language Processing (EMNLP)},
year={2020} }
For each image, we extract 36 region features using a Faster-RCNN model (https://github.com/peteanderson80/bottom-up-attention) that is pretrained on Visual Genome. The region features for each image is stored separately as a .npy file.
We use the SpaCY python library to parse the articles and captions to detect named entities. We store this information as dictionary where the keys are the article names and the values are sets of detected name entities.
Required Arguments
captioning_dataset_path: Path to GoodNews captioning dataset json file
fake_articles: Path to generated articles
image_representations_dir: Directory which contains the object representations of images
real_articles_dir: Directory which contains the preprocessed Torch text files for real articles
fake_articles_dir: Directory which contains the preprocessed Torch text files for generated articles
real_captions_dir: Directory which contains the preprocessed Torch text files for real captions
ner_dir: Directory which contains a dictionary of named entities for each article and caption