Skip to content
#

gpt

Here are 96 public repositories matching this topic...

transformers
patrickvonplaten
patrickvonplaten commented Dec 11, 2020

🚀 Feature request

Bart is a seq2seq model, but there might be applications where one would like to use only the pre-trained BartDecoder in an EncoderDecoder setting with a "long" encoder, such as

from transformers import EncoderDecoderModel

model = EncoderDecoderModel("allenai/longformer-large-4096", "facebook/bart-large")

# fine-tune model ...

This is already p

transfer-learning-conv-ai
jb33k
jb33k commented Jun 4, 2019

I'm playing around with this wonderful code but I'm running into a curious issue when I try to train the model with my own data.

I replicated the personachat_self_original.json file structure and added my own data. I deleted dataset_cache_OpenAIGPTTokenizer file but when I try to train, I get this error:

INFO:train.py:Pad inputs and convert to Tensor
Traceback (most recent call last)

Improve this page

Add a description, image, and links to the gpt topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the gpt topic, visit your repo's landing page and select "manage topics."

Learn more

You can’t perform that action at this time.