Skip to main content
Filter by
Sorted by
Tagged with
0 votes
0 answers
42 views

I am currently working on a Python based Gen AI project that requires the efficient deployment and serving of multiple LLMs specifically models with different parameter counts ( Llama-2 7B and Mistral ...
Amira Yassin's user avatar
0 votes
0 answers
30 views

Background Knowledge According to the Hugging Face documentation, now it's supported to run a GGUF model directly using Ollama with ollama run hf.co/bartowski/Llama-3.2-1B-Instruct-GGUF something like ...
Gorun's user avatar
  • 118
0 votes
1 answer
143 views

Getting issue with use_auth_token keyword while implementing a pipeline from pyannote.audio. I already used:- pipeline = Pipeline.from_pretrained("pyannote/speaker-diarization-3.1", ...
phantomguild's user avatar
-1 votes
0 answers
24 views

I am trying to use the following model Emotion Llama and try to understand how to download the models and place them in the right dir from huggingface. It actually suggests to donwload three models in ...
Jose Ramon's user avatar
  • 5,364
1 vote
0 answers
159 views

I need to to run a series of pre-trained fine-tuned models from Hugging Face to Jupyter notebook. I have updated to the latest version of both PyTorch and Transformers, but when I run the code from ...
Alex Colville's user avatar
1 vote
1 answer
78 views

I'm trying to implement Speech-to-Text transcription in my Swift app using Hugging Face's swift-transformers package to run Whisper models locally. I've added the package to my Xcode project, but when ...
Zaid's user avatar
  • 451
1 vote
0 answers
68 views

My proxy goal is to change LoRA from h = (W +BA)x to h = (W + BAP)x. Preliminary code attached for your reference My actual goal is to train a model with the following loss: 〖Θ ̃=(arg min)┬Δ ̂ 〗⁡〖‖𝑓_(...
Jason Rich Darmawan's user avatar
-1 votes
2 answers
95 views

I’m trying to use LangChain’s Hugging Face integration to chat with the model TinyLlama/TinyLlama-1.1B-Chat-v1.0 for the very first time, but I’m getting a StopIteration error when calling .invoke(). ...
forstudy's user avatar
0 votes
0 answers
61 views

has anyone managed to translate something using Helsinki-NLP and ONNX Runtime in Java? Using a Python script, I generated these files: ├── encoder_model.onnx ├── decoder_model.onnx ├── ...
minizibi's user avatar
  • 393
2 votes
1 answer
80 views

I'm creating a conversation dataset for an image classification task where the system message should contain only text, and the user message contains both text and an image. However, after mapping my ...
GauravGiri's user avatar
0 votes
0 answers
96 views

When fine tuning a model, using the HuggingFace inference hub, the error below was encountered: The tokenizer has new PAD/BOS/EOS tokens that differ from the model config and generation config. The ...
Kingsley Uchunor's user avatar
0 votes
0 answers
89 views

After failing to make the QwenImageEditPlus run (https://huggingface.co/spaces/discord-community/README/discussions/9#68d260e32053323e6bfab30c), I tried a different approach (thanks to all the example ...
Siladittya's user avatar
  • 1,215
1 vote
0 answers
61 views

I encounterd this error while trying to run hugging face trainer on a multi-gpu. RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cuda:0 and cuda:1! I use a ...
Dwi Rezky Fahlan's user avatar
0 votes
1 answer
379 views

I have been stepping into GenAI and currently I am working with Hugging face's open source models. However, I am not able to receive any response from the API. I have created access token on hugging ...
Apoorva Walia's user avatar
0 votes
0 answers
61 views

I trained a Qwen model on my own dataset. Now I need to evaluate my trained model using the loss function, but I don’t know how to do it. I saw examples for other metrics such as accuracy and ...
Kathi Meyer's user avatar
0 votes
0 answers
25 views

I was trying to use Gradio in Huggingface Spaces. I added an app.py file in my VScode, and VScode told me that the push was all right. However, Huggingface Spaces declared "No application file&...
Alex YAN's user avatar
1 vote
0 answers
37 views

I have an Azure ML studio notebook. I want to use the HuggingFace "cross-encoder-nli-deberta-v3-base" model to do zero-shot classification. This code instantiates the endpoint without error: ...
msand's user avatar
  • 11
0 votes
0 answers
218 views

Just trying to use an Text-Gen LLM from HuggingFace Inference Provider using LangChain in Node.js, I chose Model Qwen/Qwen2.5-1.5B-Instruct, trying out other models did not seem to work, I couldn't ...
Basel_Dev's user avatar
0 votes
0 answers
57 views

the Smolagents CodeAgent is given a task to convert a string into markdown table format. It successfully captures the related part of the string and writes the code for markdown table formatting. ...
aearslan's user avatar
  • 176
0 votes
1 answer
117 views

I am trying to load a training dataset in my Google Colab notebook but keep getting an error. Here is the code snippet which returns the error: from datasets import load_dataset ds = load_dataset(&...
AlecArk's user avatar
1 vote
2 answers
171 views

I am reading about Text embeddings in LLM from the book Hands-On Large Language Models. It is mentioned that as follows: from sentence_transformers.evaluation import EmbeddingSimilarityEvaluator from ...
venkysmarty's user avatar
  • 11.5k
1 vote
0 answers
799 views

I’m trying to load gpt-oss-20b locally using Hugging Face transformers with CPU only. Minimal code: from transformers import pipeline model_path = "/mnt/d/Projects/models/gpt-oss-20b" pipe = ...
mindlesscoding's user avatar
0 votes
1 answer
161 views

Discussion HuggingFace accelerate's init_empty_weights() properly loads all text encoders I tested to the PyTorch meta device and consumes no apparent memory or disk space while loaded. However, it ...
Matthew Ross's user avatar
0 votes
0 answers
232 views

I am trying to initialize a PPO_trainer but have issues. from trl import PPOTrainer, PPOConfig ppo_config = PPOConfig( batch_size=4, learning_rate=1e-5, mini_batch_size=2, use_cpu=...
m0ss's user avatar
  • 472
1 vote
0 answers
53 views

I'm trying to fine-tune Hugging Face BLIP (Bootstrapped Language-Image Pretraining) to classify pizza boxes as either recyclable (clean) or non-recyclable (contaminated) by generating captions that ...
Wow Wow's user avatar
  • 11
0 votes
0 answers
56 views

I'm using LeRobot to train a SO101 arm policy with 3 video streams (front, above, gripper) and a state vector. The dataset can be found at this link. I created a custom JSON config (the train_config....
Aaron Serpilin's user avatar
0 votes
0 answers
47 views

I'm trying to use the TinyLlama/TinyLlama-1.1B-Chat-v1.0 model from Hugging Face with LangChain using the langchain_huggingface integration. My goal is to get a simple response from the model using ...
Simran Dalvi's user avatar
1 vote
1 answer
322 views

I'm trying to use the langchain_huggingface.HuggingFaceEndpoint integration to call the "google/flan-t5-large" model from Hugging Face in a LangChain pipeline. Here's my code: from langchain....
coderbhai's user avatar
0 votes
1 answer
239 views

I'm trying to load the Qwen2.5-VL-7B-Instruct model from hugging face with 4-bit weight-only quantization using TorchAoConfig (similar to how its mentioned in the documentation here), but I'm getting ...
Sankalp Dhupar's user avatar
0 votes
0 answers
63 views

I am trying to apply below transformation for preparing my datasets for fine tuning using unsloth huggingface. It requires the dataset to be in following format. def convert_to_conversation(sample): ...
SoraHeart's user avatar
  • 428
0 votes
0 answers
92 views

Problem I have two nearly identical Python applications using LlamaIndex + Ollama for document Q&A: Online version: ~5 seconds response time Offline version: ~18 seconds response time FYI i am ...
sai's user avatar
  • 1
0 votes
1 answer
240 views

I am creating a benchmark to evaluate a language model. First, I generated the dataset that I'm gonna prompt the Language model with. Subsequently, I tried to evaluate any tiny language model just to ...
Mahmoud Hanouneh's user avatar
1 vote
1 answer
677 views

I upgraded my Python trl package to version 0.18.1. I use the SFTTrainer of the package to finetune a Qwen2.5 LLM neural net. Previously, I used the TrainingArgument class to set additional params. I ...
soosmann's user avatar
  • 119
0 votes
1 answer
142 views

There is a simple way to download a model from hugging face, # Load model directly from transformers import AutoTokenizer, AutoModel tokenizer = AutoTokenizer.from_pretrained("sentence-...
Uwe.Schneider's user avatar
3 votes
0 answers
208 views

I am new to this. I have been trying but could not make the the model answer on images. from llama_cpp import Llama import torch from PIL import Image import base64 llm = Llama( model_path='Holo1-...
Abhash Rai's user avatar
0 votes
0 answers
44 views

I'm having relatively good results with HelsinkiNlp models for translation, except for one thing: some special characters are omitted from the translation. If I decode without skipping the special ...
gooopil's user avatar
  • 17
1 vote
0 answers
94 views

Wanted to use the pipeline api from @huggingface/transformers js for sentence-similarity - but I do not see a specific pipeline for it. The closest thing is text classification and feature extractions ...
Edv Beq's user avatar
  • 1,020
0 votes
0 answers
56 views

I wanted to make a web app that uses llama-index to answer queries using RAG from specific documents. I have locally set up Llama3.2-1B-instruct llm and using that locally to create indexes of the ...
Utkarsh's user avatar
2 votes
1 answer
210 views

I’m trying to generate embeddings using the Hugging Face Inference API with LangChain in Python, but I’m running into issues. My goal is to use the API (not local models) to generate embeddings for ...
Jeevan's user avatar
  • 11
0 votes
0 answers
42 views

In my .NET project, I am configuring the Huggingface library as follows: builder.Services .AddKernel() .AddHuggingFaceChatCompletion( model: "deepseek-ai/DeepSeek-R1", ...
Murat Öztürk's user avatar
0 votes
0 answers
30 views

I'm trying to run a sentiment analysis function using the @xenova/transformers package in a NextJS project with Webpack, but I'm encountering the following error: Module parse failed: Unexpected ...
Santhosh's user avatar
1 vote
1 answer
87 views

I have installed docling successfully, but when doing the following: from langchain_docling import DoclingLoader source_path = "shared\abc.pdf" loader = DoclingLoader(file_path=source_path) ...
chaos24's user avatar
  • 11
0 votes
1 answer
227 views

a huggingface model, like Qwen32B-GGUF, contains some quantization-related files which are large. Perhaps, only use one quantization-related file and the rest is not used. By huggingface-cli, it ...
Hobin C.'s user avatar
  • 793
0 votes
0 answers
288 views

I am using Colab and HuggingFace Token is added in Colab secrets. from langchain_huggingface import ChatHuggingFace, HuggingFaceEndpoint from dotenv import load_dotenv from google.colab import ...
Sam2021's user avatar
  • 21
1 vote
2 answers
675 views

I am trying to create a simple langchain app on text-generation using API to communicate with models on HuggingFace servers. I created a “.env” file and stored by KEY in the variable: “...
Sri2110's user avatar
  • 345
0 votes
0 answers
48 views

I’m preparing a text dataset for next-token language-model pre-training. Using HF datasets with batched=True, I wrote a helper that: prepends a BOS token (if the tokenizer has one), appends an EOS ...
Charlie Parker's user avatar
0 votes
0 answers
109 views

I'm attempting to serve a pretrained Stable Diffusion model with LoRA weights applied using TorchServe. However, the LoRA weights don't seem to load properly, and I'm not sure why. Could anyone help ...
박연수's user avatar
0 votes
1 answer
583 views

I'm trying to run a Python script that uses the ollama library to generate responses from a custom LLM model. My code attempts to call ollama.generate() using the following model name: chosen_model = '...
JaS's user avatar
  • 45
0 votes
0 answers
137 views

import { DataAPIClient } from "@datastax/astra-db-ts"; import { PuppeteerWebBaseLoader } from "langchain/document_loaders/web/puppeteer"; import axios from "axios"; ...
Rakib islam's user avatar
0 votes
1 answer
107 views

from sentence_transformers import SentenceTransformer model = SentenceTransformer("BAAI/bge-small-en-v1.5") sentences = [ "The weather is lovely today.", "It's so ...
SM9595's user avatar
  • 39

1
2 3 4 5
19