Official repository of OFA (ICML 2022). Paper: OFA: Unifying Architectures, Tasks, and Modalities Through a Simple Sequence-to-Sequence Learning Framework
-
Updated
Jan 11, 2023 - Python
Official repository of OFA (ICML 2022). Paper: OFA: Unifying Architectures, Tasks, and Modalities Through a Simple Sequence-to-Sequence Learning Framework
Papers about pretraining and self-supervised learning on Graph Neural Networks (GNN).
Recent Advances in Vision and Language PreTrained Models (VL-PTMs)
X-modaler is a versatile and high-performance codebase for cross-modal analytics(e.g., image captioning, video captioning, vision-language pre-training, visual question answering, visual commonsense reasoning, and cross-modal retrieval).
Official Pytorch Implementation of: "ImageNet-21K Pretraining for the Masses"(NeurIPS, 2021) paper
EntitySeg Toolbox: Towards Open-World and High-Quality Image Segmentation
End-to-End recipes for pre-training and fine-tuning BERT using Azure Machine Learning Service
PITI: Pretraining is All You Need for Image-to-Image Translation
Paddle Distributed Training Examples. 飞桨分布式训练示例 Resnet Bert GPT MOE DataParallel ModelParallel PipelineParallel HybridParallel AutoParallel Zero Sharding Recompute GradientMerge Offload AMP DGC LocalSGD Wide&Deep
BigDetection: A Large-scale Benchmark for Improved Object Detector Pre-training
PyTorch code for "Unifying Vision-and-Language Tasks via Text Generation" (ICML 2021)
An official implementation for " UniVL: A Unified Video and Language Pre-Training Model for Multimodal Understanding and Generation"
[ACL 2022] LinkBERT: A Knowledgeable Language Model
Transformer with Untied Positional Encoding (TUPE). Code of paper "Rethinking Positional Encoding in Language Pre-training". Improve existing models like BERT.
OpenAI GPT2 pre-training and sequence prediction implementation in Tensorflow 2.0
Research code for EMNLP 2020 paper "HERO: Hierarchical Encoder for Video+Language Omni-representation Pre-training"
A Universal 3D Molecular Representation Learning Framework
code for TCL: Vision-Language Pre-Training with Triple Contrastive Learning, CVPR 2022
Parameter-Efficient Transfer from Sequential Behaviors for User Modeling and Recommendation
Add a description, image, and links to the pretraining topic page so that developers can more easily learn about it.
To associate your repository with the pretraining topic, visit your repo's landing page and select "manage topics."