Fine-tune BERTForMaskedLM

See original GitHub issue

Hello,

I am doing a project on spelling correction. I used pre-trained “bert-base-cased” model. However, the results are not that accurate. Therefore, I planned to fine-tune the BERT for Masked LM task. I couldn’t find any examples for fine-tuning BERT model for Masked LM. I tried to use “run_language_modeling.py” for fine-tuning. But, I came across with the following error:

C:\Users\ravida6d\spell_correction\transformers\examples\language-modeling>python run_language_modeling.py --output_dir ="C:\\Users\\ravida6d\\spell_correction\\contextualSpellCheck\\fine_tune\\" --model_type = bert --model_name_or_path = bert-base-cased --do_train --train_data_file =$TRAIN_FILE --do_eval --eval_data_file =$TEST_FILE –mlm

C:\Users\ravida6d\AppData\Local\Continuum\anaconda3\envs\contextualSpellCheck\lib\site-packages\transformers\training_args.py:291: FutureWarning: The `evaluate_during_training` argument is deprecated in favor of `evaluation_strategy` (which has more options)
  FutureWarning,

Traceback (most recent call last):
  File "run_language_modeling.py", line 313, in <module>
    main()
  File "run_language_modeling.py", line 153, in main
    model_args, data_args, training_args = parser.parse_args_into_dataclasses()
  File "C:\Users\ravida6d\AppData\Local\Continuum\anaconda3\envs\contextualSpellCheck\lib\site-packages\transformers\hf_argparser.py", line 151, in parse_args_into_dataclasses

    raise ValueError(f"Some specified arguments are not used by the HfArgumentParser: {remaining_args}")
ValueError: Some specified arguments are not used by the HfArgumentParser: ['bert', 'bert-base-cased']

I am not understanding how to use this script. Can anyone give some information for understanding the fine-tuning of BERT Masked LM.

Issue Analytics

  • State:closed
  • Created 3 years ago
  • Comments:7 (1 by maintainers)

github_iconTop GitHub Comments

1reaction
naturecreatorcommented, Sep 29, 2020

While fine-tuning, we can only see loss and perplexity which is useful. Is it also possible to see the accuracy of the model and also the tensorboard when using the “run_language_modeling.py” script? It would be really helpful if anyone could explain how the “loss” is calculated for BERTForMaskedLM task (as there are no labels provided while fine-tuning).

0reactions
ucas010commented, Nov 28, 2022

hi,dear how to use Spelling Error Correction with this rp? could you pls help me ?

Read more comments on GitHub >

github_iconTop Results From Across the Web

Fine-tuning BERT Model on domain specific language and for ...
Hi guys First of all, what I am trying to do: I want to fine-tune a BERT Model ... We are using BertForMaskedLM...
Read more >
Finetune and generate text with BertForMaskedLM · Issue #2119
Questions & Help I am trying to fine-tune and generate text using BertForMaskedLM. Although my script works I am not getting the output...
Read more >
Fine Tuning BERT using Masked Language Modelling
Hello, In this tutorial, we are going to fine-tune or pre-train our BERT model (from the huggingface transformers) using a famous ...
Read more >
Fine-tune a BERT model for context specific embeddigns
The model that is used is one of the BERTForLM familly. The idea is to create a dataset using the TextDataset that tokenizes...
Read more >
How to fine-tune BERT using HuggingFace - Joe Cummings
HuggingFace offers several versions of the BERT model including a base BertModel , BertLMHeadMoel , BertForPretraining , BertForMaskedLM ...
Read more >

github_iconTop Related Medium Post

No results found

github_iconTop Related StackOverflow Question

No results found

github_iconTroubleshoot Live Code

Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free

github_iconTop Related Reddit Thread

No results found

github_iconTop Related Hackernoon Post

No results found

github_iconTop Related Tweet

No results found

github_iconTop Related Dev.to Post

No results found

github_iconTop Related Hashnode Post

No results found