Questions tagged [huggingface]
The huggingface tag has no usage guidance.
109
questions
0
votes
0
answers
15
views
Hugging Face Real Time Object Detection Deployment
I'm developing a live object detection app using Streamlit and the YOLOv8 model. The app runs smoothly with real-time inference on my local machine. However, when I deploy it to Hugging Face Spaces, ...
0
votes
0
answers
14
views
How can I make my Hugging Face fine-tuned model's config.json file reference a specific revision/commit from the original pretrained model?
I uploaded this model: https://huggingface.co/pamessina/CXRFE, which is a fine-tuned version of this model: https://huggingface.co/microsoft/BiomedVLP-CXR-BERT-specialized
Unfortunately, CXR-BERT-...
0
votes
0
answers
18
views
Does it common for LM (hundreds million parameters) beat LLM (billion parameters) for binary classification task?
Preface
I am trying to fine-tune the transformer-based model (LM and LLM). The LM that I used is DEBERTA, and the LLM is LLaMA 3. The task is to classify whether a text contains condescending language ...
0
votes
0
answers
20
views
NLP: how to handle bad tokenization
I get nonsense when trying to translate the following german sentence to swedish using google/madlad400-3b-mt:
a. Natürliche Personen: BundID mit ELSTER-Zertifikat oder nPA/eID/eAT-Authentifizierung
...
0
votes
1
answer
31
views
attentions not returned from transformers ViT model when using output_attentions=True
I'm using this code snippet from the docs of HuggingFace ViT classification model - with one addition: I'm using the output_attentions=True parameter. Nevertheless, ...
0
votes
0
answers
31
views
Instruction LLM - extract data from text wrongly continues
I'm trying to fine-tune open sourced LLMs, for now let's stick with Mistral-7b-instruct model.
My task is a follow: I have emails, that represents "price requests" for shipments sends by our ...
2
votes
1
answer
42
views
Since LoRA parameters are randomly initialized, shouldn't that mean that initially breaks a models output?
I have just tried using LoRA on Llama 3 8B and I found without doing any fine tuning it performed pretty well on my dataset. But then I realized that surely the LoRA parameters are randomly ...
1
vote
0
answers
402
views
"No sentence-transformers model found with name" on huggingface even though it exists
I am trying to use infgrad/stella-base-en-v2 on hugging to generate embeddings using langchain
The model exists on the huggingface hub
The model is listed on the MTEB leaderboard
The model has ...
0
votes
0
answers
28
views
Not able to use huggingface inference API to get text embeddings
From the tutorials I am using the example that is provided
...
1
vote
1
answer
142
views
How do I get model.generate() to omit the input sequence from the generation?
I'm using Huggingface to do inference on llama-3-B. Here is my model:
...
0
votes
0
answers
28
views
Does Fine Tuning with Custom Label Build Upon the Capability of Zero Shot Classification or Does It Train from Scratch?
The task is to classify email text bodies into exclusive categories like feedback, complaint etc. I have a labelled dataset available having about 350 samples.
I have tried the ...
1
vote
1
answer
26
views
Public Email Classification Dataset but not Spam vs Ham
Context
Working to deliver a POC on automated email classification (in customer service context) to tag emails as related to feedback, complain, lost and found etc. The tags are not entirely exclusive,...
0
votes
1
answer
60
views
Generate VTT file from speech to text
I was able to generate text from an audio file using huggingface, using this code
...
0
votes
0
answers
13
views
Training split generation - Extremely Slow
I am fine-tuning mbert on wikipedia dataset, loaded with Datasets (Hugging face)
...
0
votes
0
answers
9
views
Will hypermeters tuned on sampled dataset work for the whole dataset?
I'm doing multi-label classification on text data using BERT model. Since the dataset is huge, around 50 thousand rows, I was thinking to use stratify sampling on dataset to reduce it to around 2-4 ...