Skip to content

Code and data for the paper "Pretrained Language Models are Symbolic Mathematics Solvers too!", arXiv:2110.03501

License

Notifications You must be signed in to change notification settings

softsys4ai/differentiable-proving

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Pretrained Language Models are Symbolic Mathematics Solvers too

This is a repository containing code and data for the paper:

K. Noorbakhsh, M. Sulaiman, M. Sharifi, K. Roy and P. Jamshidi. Pretrained Language Models are Symbolic Mathematics Solvers too!

Project Website

Pre-requisites

This code depends on the following packages:

  1. Torch
  2. NumPy
  3. SymPy
  4. Transformers
  5. Apex

Structure

  • trainer.py contains code for fine-tuning the pre-trained language models. Please modify the following parameters for running:

    1. language: the pre-trained language.
    2. Model_Type: mbart or Marian.
    3. path1 and path2: the path of the training and the validation data.
    4. max_input_length and max_output_length: 1024 for the mBART model and 512 for the Marian-MT model.
    5. model_name: name of the model you wish to save.
  • evaluator.py contains code for evaluting the fine-tuned language model on the symbolic math data. Please modify the parameter 1-4 same as the trainer section and also modify the following parameter:

    1. path: the path of the test dataset.
    2. saved_model: the path of the saved fine-tuned model.
  • src/hf_utils.py contains code for reading the datasets and some utilities for evaluation.

The rest of the code is adopted from Deep learning for symbolic mathematics (Lample et al.).

Datasets

The datasets are available here.

  1. train, valid, and test files contain the training, validation and test datasets for the mBART model.
  2. language_data contains data for the training, validation and test datasets of the Marian-MT model.
  3. distribution_test contains the test files for the distribution shift section (polynomial, trgonometric and logarithmic).

Citation

Please cite us if you use our work in your research.

@article{noorbakhsh2021pretrained,
    title={Pretrained Language Models are Symbolic Mathematics Solvers too!}, 
    author={Kimia Noorbakhsh and Modar Sulaiman and Mahdi Sharifi and Kallol Roy and Pooyan Jamshidi},
    journal={arXiv preprint arXiv:2110.03501},
    year={2021}
 }

About

Code and data for the paper "Pretrained Language Models are Symbolic Mathematics Solvers too!", arXiv:2110.03501

Topics

Resources

License

Stars

Watchers

Forks

Languages