Skip to main content
Filter by
Sorted by
Tagged with
1 vote
0 answers
68 views

My proxy goal is to change LoRA from h = (W +BA)x to h = (W + BAP)x. Preliminary code attached for your reference My actual goal is to train a model with the following loss: 〖Θ ̃=(arg min)┬Δ ̂ 〗⁡〖‖𝑓_(...
Jason Rich Darmawan's user avatar
1 vote
0 answers
53 views

I'm trying to fine-tune Hugging Face BLIP (Bootstrapped Language-Image Pretraining) to classify pizza boxes as either recyclable (clean) or non-recyclable (contaminated) by generating captions that ...
Wow Wow's user avatar
  • 11
0 votes
0 answers
75 views

Intro to the problem I am trying to train Llama-3.1 8B on an H100 but I keep running into the following error when trying to resume training ... File "/home/jovyan/folder/training/.venv/lib/...
Praanto's user avatar
  • 369
-2 votes
1 answer
56 views

I am using hugging face Trainer API. transformers version==4.31.0 torch==2.0.1 accelerate==0.27.0 I'm trying to fine-tune a TimeSformer model for video classification using the Hugging Face ...
Hamza Azhar's user avatar
2 votes
2 answers
4k views

I'm trying to fine-tune a model using SFTTrainer from trl, but I'm facing multiple TypeError issues related to unexpected keyword arguments. from transformers import TrainingArguments from trl import ...
trougc's user avatar
  • 415
0 votes
1 answer
715 views

I’m trying to train a language model using google/gemma-2-2b with the Hugging Face Transformers Trainer. The same training script works fine for other models like gpt2 and meta-llama/Meta-Llama-3-8B, ...
Charlie Parker's user avatar
2 votes
1 answer
101 views

I am using Huggingface Trainer to train a cumstom model subclassing a Llama llm. After tokenized by the tokenizer, my dataset has these fields 'input_ids', 'labels' and so on, and I additionally add 2 ...
Janji Avicii's user avatar
0 votes
1 answer
630 views

I have a dataset I want to fine-tune a huggingface LLM with. This dataset is quite simple. It has two columns: one column has DNA sequences (each in the form of a string 5000 letters long). Another ...
youtube's user avatar
  • 514
0 votes
0 answers
811 views

I'm trying to download the TheBloke/falcon-40b-instruct-GPTQ model using the Hugging Face Transformers CLI in PowerShell on Windows 10, but I consistently encounter an SSL certificate error. It ...
bbartling's user avatar
  • 3,564
0 votes
0 answers
64 views

Please consider the following code: from datasets import load_dataset_builder, load_dataset import numpy as np import os import torch from transformers import AutoModelForSeq2SeqLM, AutoTokenizer, ...
Antonio Sesto's user avatar
0 votes
1 answer
587 views

I'm using the Hugging Face Trainer (or SFTTrainer) for fine-tuning, and I want to log the training loss at step 0 (before any training steps are executed). I know there's an eval_on_start option for ...
Charlie Parker's user avatar
3 votes
0 answers
213 views

I'm working on a sentence regression task using Hugging Face’s Trainer. Each sample consists of: input_ids: The tokenized sentence. labels: A numerical scalar target (for regression). metadata: A ...
enter_thevoid's user avatar
1 vote
1 answer
312 views

I'm a little puzzled where (and if) EOS tokens are being added when using Huggignface's trainer classes to train a T5 (LongT5 actually) model. The data set contains pairs of text like this: from to ...
gphilip's user avatar
  • 718
0 votes
1 answer
62 views

I am trying to do a machine translation from Hindi to Sanskrit using NLLB model. But I keep getting the error: IndexError: Invalid key: 39463 is out of bounds for size 0. The error is coming when ...
user27310271's user avatar
1 vote
1 answer
534 views

Keep getting the error in my terminal: ConnectionRefusedError: [Errno 111] Connection refused I got the above error by trying to add in this command: accelerate launch --num_processes=1 --...
Aaron's user avatar
  • 31
0 votes
1 answer
237 views

I use a BertTokenizer and add my custom tokens using add_tokens() function. Minimal sample code here: checkpoint = 'fnlp/bart-base-chinese' tokenizer = BertTokenizer.from_pretrained(checkpoint) ...
Raptor's user avatar
  • 54.4k
0 votes
0 answers
91 views

I'm receiving this error from HuggingFace's Trainer: Tried to track the number of tokens seen, however the current model is not configured properly to know what item is the input. To fix this, add a `...
Rylan Schaeffer's user avatar
2 votes
0 answers
306 views

I am trying to fine-tune using this notebook: GLiNER/examples/finetune.ipynb at main · urchade/GLiNER (github.com) However, the logs only show 'loss' , which I assume is the training data set loss, ...
andream's user avatar
  • 53
1 vote
1 answer
367 views

The usual steps to use the Trainer from huggingface requires that: Load the data Tokenize the data Pass tokenized data to Trainer MWE: data = generate_random_data(10000) # Generate 10,000 samples ...
Kirk Walla's user avatar
2 votes
0 answers
1k views

Currently you can let SFTTrainer teach your models to learn to predict every token in your dataset, or you can let it train on "completions only", using the DataCollatorForCompletionOnlyLM ...
Jelle De Loecker's user avatar
0 votes
0 answers
40 views

I tried to use the following code to train my model but I get the following issue Here's the code: import torch import pyarrow as pa import pandas as pd from transformers import ...
MarMarhoun's user avatar
1 vote
1 answer
1k views

I am trying to use an Accelerator with a Trainer using the code bellow: tokenizer = AutoTokenizer.from_pretrained(model_args.model_name_or_path) config = AutoConfig.from_pretrained(model_args....
Evelin Amorim's user avatar
1 vote
1 answer
256 views

Background I am finetuning a mistral-7B-instruct-v01 model using the same workflow as is outlined in these two blogposts (using Sagemaker): How to Fine-Tune LLMs in 2024 with Hugging Face Train and ...
Axel Sjöberg's user avatar
1 vote
1 answer
2k views

There are several ways to get metrics for transformers.Trainer but only for the evaluation and not for the training. I read and found answers scattered in different posts such as this post. But ...
Wassim Jaoui's user avatar
0 votes
1 answer
310 views

I am facing an issue whilst using Trainer class with Pytorch on Google Colab as it demands accelarate>=0.21.0 even though I have updated all the requirements, is there any alternative to it? "...
ishaan's user avatar
  • 1