Skip to content

Files

Latest commit

1bfb577 · Jul 10, 2020

History

History
This branch is 4590 commits behind tensorflow/models:master.

nlp

Folders and files

NameName
Last commit message
Last commit date

parent directory

..
Jul 10, 2020
Jul 10, 2020
Jul 10, 2020
Jul 9, 2020
Jul 8, 2020
Jun 22, 2020
Jul 10, 2020
Jul 8, 2020
Jun 1, 2020
Jun 3, 2020
Jul 19, 2019
May 28, 2020

TensorFlow NLP Modelling Toolkit

This codebase provides a Natrual Language Processing modeling toolkit written in TF2. It allows researchers and developers to reproduce state-of-the-art model results and train custom models to experiment new research ideas.

Features

  • Reusable and modularized modeling building blocks
  • State-of-the-art reproducible
  • Easy to customize and extend
  • End-to-end training
  • Distributed trainable on both GPUs and TPUs

Major components

Libraries

We provide modeling library to allow users to train custom models for new research ideas. Detailed intructions can be found in READMEs in each folder.

  • modeling/: modeling library that provides building blocks (e.g., Layers, Networks, and Models) that can be assembled into transformer-based achitectures .
  • data/: binaries and utils for input preprocessing, tokenization, etc.

State-of-the-Art models and examples

We provide SoTA model implementations, pre-trained models, training and evaluation examples, and command lines. Detail instructions can be found in the READMEs for specific papers.

  1. BERT: BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding by Devlin et al., 2018
  2. ALBERT: A Lite BERT for Self-supervised Learning of Language Representations by Lan et al., 2019
  3. XLNet: XLNet: Generalized Autoregressive Pretraining for Language Understanding by Yang et al., 2019
  4. Transformer for translation: Attention Is All You Need by Vaswani et al., 2017
  5. NHNet: Generating Representative Headlines for News Stories by Gu et al, 2020