Skip to content
UniLM - Unified Language Model Pre-training / Pre-training for NLP and Beyond
Python
Branch: master
Clone or download

Latest commit

Latest commit 769852e May 16, 2020

Files

Permalink
Type Name Latest commit message Commit time
Failed to load latest commit information.
layoutlm Update README.md May 16, 2020
minilm Update README.md Apr 16, 2020
s2s-ft s2s CPU Apr 2, 2020
storage Create unilm-base-cased-vocab.txt Dec 16, 2019
unilm-v1 Update README.md Apr 26, 2020
unilm Update README.md Apr 30, 2020
.gitignore Initial commit Jul 23, 2019
CODE_OF_CONDUCT.md Initial commit Jul 23, 2019
CONTRIBUTING.md Create CONTRIBUTING.md Sep 19, 2019
LICENSE init Sep 30, 2019
NOTICE.md init Sep 30, 2019
README.md Update README.md Apr 5, 2020

README.md

UniLM

We develop pre-trained models for natural language understanding (NLU) and generation (NLG) tasks

The family of UniLM:

UniLM: unified pre-training for language understanding and generation

MiniLM (new): small pre-trained models for language understanding and generation

LayoutLM (new): multimodal (text + layout/format + image) pre-training for document understanding (e.g. scanned documents, PDF, etc.)

s2s-ft (new): sequence-to-sequence fine-tuning toolkit

News

Release

***** New February, 2020: UniLM v2 | MiniLM v1 | LayoutLM v1 | s2s-ft v1 release *****

***** October 1st, 2019: UniLM v1 release *****

License

This project is licensed under the license found in the LICENSE file in the root directory of this source tree. Portions of the source code are based on the transformers project.

Microsoft Open Source Code of Conduct

Contact Information

For help or issues using UniLM, please submit a GitHub issue.

For other communications related to UniLM, please contact Li Dong (lidong1@microsoft.com), Furu Wei (fuwei@microsoft.com).

You can’t perform that action at this time.