Home

Awesome

Logic2Text

Data

In the dataset folder, we have the full dataset (all_data.json), and the train test split (train.json, valid.json, test.json). Each example is in a dictionary of the following format:

    "topic": table caption,
    "wiki": link to the wikipedia page,
    "action": logic type,
    "sent": description,
    "annotation": raw annotation data from the mturk,
    "logic": logic form,
    "logic_str": linearized logic form,
    "nid": number of fuction nodes,
    "table_header": table header, in list format
    "table_cont": table content, list of lists
  

Logic form execution

In the execution folder, run

python execute.py

It will execute all the logic forms in all_data.json. All the function definitions are in APIs.py

This site is under construction, and we will release other codes in the future.

Model

The pre-trained GPT-2 can be downloaded via Dropbox.

requirements

python 3.6 tensorflow 1.12

data pre-process

In addition to the original data folder, this script will create another folder to contain the preprocessed data, which will be used for train and test.

python preprocess.py data_folder GPT_folder

train

Modify the data paths and parameters in the Main.py file. Then run:

CUDA_VISIBLE_DEVICES=0,1 python Main.py --mode train

test

CUDA_VISIBLE_DEVICES=0,1 python Main.py --mode test --saved_model_path load_model_path

AMT annotation interface

In the folder amt, we have all the html scripts for AMT annotations.

folder composition: html files for statements composition
folder logic_form: html files for logic form annotation
folder quality_check: html files for checking logic form annotation quality