About 133,000 results
Open links in new tab
  1. python - How to use DistilBERT Huggingface NLP model to …

    Oct 24, 2021 · I am using DistilBERT to do sentiment analysis on my dataset. The dataset contains text and a label for each row which identifies whether the text is a positive or negative …

  2. Download pre-trained BERT model locally - Stack Overflow

    After instantiating the SentenceTransformer via download, you can then save it to any path of your choosing with the 'save ()' method.

  3. How to Fine-tune HuggingFace BERT model for Text Classification

    Sep 2, 2021 · Base Model - TFDistilBertModel Hugging Face Transformers: Fine-tuning DistilBERT for Binary Classification Tasks TFDistilBertModel class to instantiate the base …

  4. nlp - Fine-Tuning DistilBertForSequenceClassification: Is not …

    Aug 2, 2020 · This Stack Overflow question discusses issues with fine-tuning DistilBertForSequenceClassification, including unchanged loss and weights not updating …

  5. No model was supplied, defaulted to - Stack Overflow

    Apr 23, 2023 · It tells you, that the pipeline is using distilbert-base-uncased-finetuned-sst-2-english because you haven't specified a model_id. In other words, it might not yield the best …

  6. python - Problem loading transformers ... - Stack Overflow

    Sep 27, 2024 · I want to use some of the models available through huggingface. I am having the hardest time even getting started. Can anyone help me identify and solve this problem? I am …

  7. huggingface - Error ("bus error") running the simplest example on ...

    Nov 13, 2024 · No model was supplied, defaulted to distilbert/distilbert-base-uncased-finetuned-sst-2-english and revision 714eb0f (https://huggingface.co/distilbert/distilbert-base-uncased …

  8. why take the first hidden state for sequence classification ...

    Feb 6, 2020 · In the last few layers of sequence classification by HuggingFace, they took the first hidden state of the sequence length of the transformer output to be used for classification. …

  9. How can I train distilBERT more efficiently on my large text ...

    Jul 25, 2023 · Model Training: How can I make DistilBERT handle this beast of a dataset more efficiently? Anyone got experience tweaking the training strategy, batch size, learning rate, …

  10. How to get output_attentions of a pretrained Distilbert Model?

    Apr 19, 2022 · I got the solution to anyone looking: modified the code as ` dbert_model = TFDistilBertModel.from_pretrained ('distilbert-base-uncased',config, output_attentions=True)` …