github huggingface/transformers v0.5.1
Bug fix update to load the pretrained `TransfoXLModel` from s3, added fallback for OpenAIGPTTokenizer when SpaCy is not installed

latest releases: v4.45.1, v4.45.0, v4.44.2...
5 years ago

Mostly a bug fix update for loading the TransfoXLModel from s3:

  • Fixes a bug in the loading of the pretrained TransfoXLModel from the s3 dump (which is a converted TransfoXLLMHeadModel) in which the weights were not loaded.
  • Added a fallback of OpenAIGPTTokenizer on BERT's BasicTokenizer when SpaCy and ftfy are not installed. Using BERT's BasicTokenizer instead of SpaCy should be fine in most cases as long as you have a relatively clean input (SpaCy+ftfy were included to exactly reproduce the paper's pre-processing steps on the Toronto Book Corpus) and this also let us use the never_split option to avoid splitting special tokens like [CLS], [SEP]... which is easier than adding the tokens after tokenization.
  • Updated the README on the tokenizers options and methods which was lagging behind a bit.

Don't miss a new transformers release

NewReleases is sending notifications on new releases.