'AttributeError: 'GPT2TokenizerFast' object has no attribute 'max_len'
I am just using the huggingface transformer library and get the following message when running run_lm_finetuning.py: AttributeError: 'GPT2TokenizerFast' object has no attribute 'max_len'. Anyone else with this problem or an idea how to fix it? Thanks!
My full experiment run: mkdir experiments
for epoch in 5
do
python run_lm_finetuning.py
--model_name_or_path distilgpt2
--model_type gpt2
--train_data_file small_dataset_train_preprocessed.txt
--output_dir experiments/epochs_$epoch
--do_train
--overwrite_output_dir
--per_device_train_batch_size 4
--num_train_epochs $epoch
done
Solution 1:[1]
The "AttributeError: 'BertTokenizerFast' object has no attribute 'max_len'" Github issue contains the fix:
The
run_language_modeling.py
script is deprecated in favor oflanguage-modeling/run_{clm, plm, mlm}.py
.If not, the fix is to change
max_len
tomodel_max_length
.
Solution 2:[2]
I use this command to solve it.
pip install transformers==3.0.2
Sources
This article follows the attribution requirements of Stack Overflow and is licensed under CC BY-SA 3.0.
Source: Stack Overflow
Solution | Source |
---|---|
Solution 1 | Wiktor Stribiżew |
Solution 2 | white |