WillHeld/hinglish_top
Viewer • Updated • 10.9k • 65 • 3
How to use SRDdev/HingMaskedLM with Transformers:
# Use a pipeline as a high-level helper
from transformers import pipeline
pipe = pipeline("fill-mask", model="SRDdev/HingMaskedLM") # Load model directly
from transformers import AutoTokenizer, AutoModelForMaskedLM
tokenizer = AutoTokenizer.from_pretrained("SRDdev/HingMaskedLM")
model = AutoModelForMaskedLM.from_pretrained("SRDdev/HingMaskedLM")MaskedLM is a pre-training technique used in Natural Language Processing (NLP) for deep-learning models like Transformers. It is a variant of language modeling where a portion of the input text is masked, and the model is trained to predict the masked tokens based on the context provided by the unmasked tokens. This model is trained for Masked Language Modeling for Hinglish Data.
Hinglish-Top Dataset columns
| Epoch | Loss |
|---|---|
| 1 | 0.0465 |
| 2 | 0.0262 |
| 3 | 0.0116 |
| 4 | 0.00385 |
| 5 | 0.0103 |
| 6 | 0.00738 |
| 7 | 0.00892 |
| 8 | 0.00379 |
| 9 | 0.00126 |
| 10 | 0.000684 |
from transformers import AutoTokenizer, AutoModelForMaskedLM, pipeline
tokenizer = AutoTokenizer.from_pretrained("SRDdev/HingMaskedLM")
model = AutoModelForMaskedLM.from_pretrained("SRDdev/HingMaskedLM")
fill = pipeline('fill-mask', model=model, tokenizer=tokenizer)
fill(f'please {fill.tokenizer.mask_token} ko cancel kardo')
Author: @SRDdev
Name: Shreyas Dixit
framework: Pytorch
Year: Jan 2023
Pipeline: fill-mask
Github: https://github.com/SRDdev
LinkedIn: https://www.linkedin.com/in/srddev/