Skip to main content
Filter by
Sorted by
Tagged with
Advice
0 votes
0 replies
15 views

I’m building a tool that generates new mathematics exam problems using an internal database of past problems. My current setup uses a RAG pipeline, Pinecone as the vector database, and GPT-5 as the ...
Marc-Loïc Abena's user avatar
Best practices
0 votes
1 replies
105 views

I'm building a voice-based calling system where users can create AI agents that make outbound phone calls. The agent uses Deepgram for real-time transcription and ElevenLabs/Cartesia for speech ...
Sarthak Sahu's user avatar
Advice
0 votes
1 replies
58 views

I have a large set of phrases obtained via Azure Fast Transcription, and I need to group them into coherent semantic chunks (to use later in a RAG pipeline). Initially, I tried grouping phrases based ...
Daniel's user avatar
  • 13
0 votes
0 answers
24 views

I'm using LlamaIndex 0.14.7. I would like to embed document text without concatenating metadata, because I put a long text in metadata. Here's my code: table_vec_store: SimpleVectorStore = ...
Trams's user avatar
  • 421
0 votes
0 answers
54 views

This is my embedding code, which I run once only: embeddings = OpenAIEmbeddings(model="text-embedding-3-large") vector_store = MongoDBAtlasVectorSearch.from_connection_string( ...
Mingruifu Lin's user avatar
1 vote
1 answer
122 views

I’m trying to evaluate my Retrieval-Augmented Generation (RAG) pipeline using Ragas. . Here’s a complete version of my code: """# RAG Evaluation""" from datasets import ...
Chandima's user avatar
0 votes
1 answer
64 views

My objective is to do keyword filtering in Chroma. I have a field called keywords with a list of strings and I want to filter with it, but chroma won't let me add lists as a field. I checked my Chroma ...
Elena López-Negrete Burón's user avatar
1 vote
0 answers
51 views

I built a RAG chatbot using LangChain + ChromaDB + OpenAI embeddings. The pipeline works, but sometimes the chatbot doesn’t return the most relevant PDF content, even though it exists in the vector DB....
Naitik Mittal's user avatar
1 vote
0 answers
72 views

I built a RAG chatbot in python,langchain, and FAISS for the vectorstore. And the data is stored as JSON. The chatbot sometimes refuses to answer when a question is rephrased. Here are two ...
SoftwareEngineer's user avatar
0 votes
0 answers
28 views

Question: I'm building a memory-augmented AI system using RAG with persistent vector storage, but facing memory leaks and context contamination between sessions. Problem: Vector embeddings aren't ...
TensorMind's user avatar
0 votes
1 answer
55 views

i am trying to create a small starter llm RAG project using haystack. my project packages are below (I use UV): [project] name = "llm-project" version = "0.1.0" description = "...
femi's user avatar
  • 984
0 votes
0 answers
49 views

I am trying to use lancedb to perform FTS, but getting spurious results. Here is a minimal example: # Data generation import lancedb import polars as pl from string import ascii_lowercase words = [...
MKWL's user avatar
  • 41
0 votes
0 answers
146 views

On the ingestion part to the graph db, I pass a json file, as an episode, custom entities (and edges), using gemini api, but I get some discrepancy on the structured output, like so: LLM generation ...
George Petropoulos's user avatar
0 votes
0 answers
56 views

I am using RAGFlow connected to a Spring Boot MCP server. My agent flow is simple: Begin node → collects inputs (auth_token, tenant_id, x_request_status) Agent (gpt-4o) → connected to MCP Tool (server)...
Ishan Garg's user avatar
1 vote
0 answers
71 views

I am using the python package ragas with the goal of generating a testset for a RAG application. I am defining my BaseRagasLLM as: from langchain_ollama import OllamaLLM from ragas.llms import ...
oyster's user avatar
  • 21
0 votes
0 answers
75 views

I have the corpus.jsonl which has 6.5gb storage.And i use the one h100 gpu to embedding the corpus to the chromadb,but it seems very slowly.I want to find how can i accelerate the progress(gpu,cpu,io)....
YiJun Sachs's user avatar
1 vote
1 answer
260 views

I set up a self-hosted Firecrawl instance and I want to crawl my internal intranet site (e.g. https://intranet.xxx.gov.tr/). I can access the site directly both from the host machine and from inside ...
birdalugur's user avatar
2 votes
1 answer
186 views

I'm building a document Q&A system using FAISS for vector search on an AWS EC2 t3.micro instance (1 vCPU, 1GB RAM). My FAISS index is relatively small (8.4MB .faiss + 1.4MB .pkl files), but I'm ...
user29255210's user avatar
0 votes
0 answers
131 views

I'm building a RAG (Retrieval-Augmented Generation) chatbot using LangChain, Gemini API, and Qdrant, with a Streamlit frontend. I want to write unit tests for the app using pytest, and I’m trying to ...
Krishna Suthar's user avatar
0 votes
1 answer
120 views

When using rag and memory, multiple identical copies of the same information is sent to the ai, when asking related questions. I have import java.util.ArrayList; import java.util.List; import dev....
MTilsted's user avatar
  • 5,535
0 votes
1 answer
268 views

I am trying to delete all the data points that are associated with a particular email Id, but I am encountering the following error. source code: app.get('/cleanUpResources', async (req, res) => { ...
Abhishek Prasad's user avatar
-1 votes
1 answer
256 views

The problem with this piece of code is that I am unable to import Client from the pinecone library. I tried to uninstalling and reinstalling different versions none of them worked. I also tried it ...
ACR's user avatar
  • 11
-1 votes
1 answer
54 views

I'm building a LangChain RAG pipeline using the FAISS vector store. I'm merging multiple FAISS indexes — each representing one document — and then querying them to generate summaries or answers via ...
Musab's user avatar
  • 54
1 vote
0 answers
193 views

I'm building a web application using Spring Boot 3.4.5 and Spring AI 1.0.0 with Llama3.2(Ollama) model integration. I've implemented tool calling, and because I have many tools in the application, I'm ...
Sarath Molathoti's user avatar
0 votes
0 answers
101 views

I have been recently trying to do a multiagent project that to summarize, consists on: Through an user input (often a query), the first agent will be dedicated to making the input more suitable for ...
PMathC's user avatar
  • 1
0 votes
1 answer
599 views

I am trying to call Flask API which i alrady running on port 5000 on my system, i am desgning a agentic AI code which will invoke GET and then POSt based on some condition , and using google-adk. I ...
witty_minds's user avatar
1 vote
0 answers
94 views

Wanted to use the pipeline api from @huggingface/transformers js for sentence-similarity - but I do not see a specific pipeline for it. The closest thing is text classification and feature extractions ...
Edv Beq's user avatar
  • 1,020
1 vote
0 answers
61 views

I'm building a RAG-based document QA system using Python (no LangChain), LLaMA (50K context), PostgreSQL with pgvector, and Docling for parsing. Users can upload up to 10 large documents (300+ pages ...
Anton Lee's user avatar
0 votes
0 answers
43 views

I'm working on a RAG pipeline using a vector database to search over a Q&A dataset. I'm using embedding-based dense retrieval to fetch relevant answers to user queries. The issue I'm facing is ...
MojtabaMAleki02's user avatar
0 votes
0 answers
56 views

I wanted to make a web app that uses llama-index to answer queries using RAG from specific documents. I have locally set up Llama3.2-1B-instruct llm and using that locally to create indexes of the ...
Utkarsh's user avatar
1 vote
0 answers
680 views

I am experimenting with RAG on GCP/Vertex AI, and tried to create some simple example. Here's what I came up with, creating small dummy files locally and then uploading them one by one to a newly-...
Davide Fiocco's user avatar
0 votes
0 answers
172 views

I have a RAG system using llamaindex. I am upgrading library from 0.10.44 to 0.12.33. I see a different behaviour now. Before when there were not results from vectors store it seems it called the LLM ...
Deibys's user avatar
  • 669
0 votes
0 answers
99 views

I checked Azure's documentation on this topic here but I do not see anything related to this. My goal is to create a question and answer dataset for my RAG solution based on each chunk for a good ...
Mike B's user avatar
  • 3,629
1 vote
1 answer
137 views

I am using this model to embed a product catalog for a rag. In the product catalog, there are no red shirts for men, but there are red shirts for women. How can I make sure the model doesnt output ...
Advait Shendage's user avatar
0 votes
0 answers
66 views

from unstructured library opensource one when i tried a pdf that have background images design patterns and XObjects in it this library also consider those as a images and store the path. so how can ...
Umair Ashraf's user avatar
0 votes
2 answers
75 views

from langchain_community.document_loaders import SitemapLoader def crawl(self): print("Starting crawler...") sitemap_url = "https://gringo.co.il/sitemap.xml" ...
Gulzar's user avatar
  • 28.8k
0 votes
2 answers
880 views

I am using AWS bedrock for the first time. I have configured the data source which is S3 along with opensearch serverless cluster for embeddings. However, I do not have any control over the mappings ...
Makarand's user avatar
  • 636
1 vote
0 answers
47 views

I'm trying to index a series of articles to use in a RAG knowledge base, I cannot find any best practice or recommendation documented about dealing with information that changes or evolves in time. ...
weeanon's user avatar
  • 821
0 votes
1 answer
433 views

I want to know if there are any other settings required for pgvector or what content needs to be set in the code to enable pgvector to support higher vector dimensions. I found on the official website ...
tom's user avatar
  • 3
0 votes
0 answers
70 views

I'm following along with the Firebase Genkit docs covering context. From reading the docs it seems as though I should be able to pass context to the flow from where I call the function in my client ...
Garrett's user avatar
  • 1,818
0 votes
1 answer
23 views

I'm retrieving results from a Cypher query, which includes the article's date and text. After fetching the results, I'm formatting them before passing them to the LLM for response generation. ...
Yuvraj Singh Bhadauria's user avatar
-4 votes
1 answer
143 views

I use Ollama and RagFlow to manage my own knowledge files, I upload some files to a knowledge,and they works well in the system. I start the ragflow with docker commands. Who can help me to find the ...
Jinzhengxuan's user avatar
0 votes
0 answers
26 views

I am having trouble figuring out why I can't see the print statements in the terminal for my retrieveInfoForQuery function and trying to figure out what is wrong. I have verified the chroma db is on ...
Chris Westbrook's user avatar
1 vote
1 answer
273 views

Trying to import RegexTextSplitter using from langchain.text_splitter import RegexTextSplitter ,RecursiveCharacterTextSplitter And I get the error from langchain.text_splitter import RegexTextSplitter ...
Dev_A's user avatar
  • 23
0 votes
0 answers
154 views

I am working on a RAG chatbot which takes .csv financial tables (eg. income statements/balance sheets etc.) of a company in the last 3 quarters, and answers questions based on the provided report ...
jy_lim21's user avatar
0 votes
1 answer
234 views

I am doing evaluation for my RAG LLM application using ragas. I have the prompt instruction to describe some rules, the retrieved content from my retriever, and chat history together for the LLM to do ...
Howie's user avatar
  • 111
0 votes
1 answer
162 views

I'm trying to upsert reviews that i've scraped into pinecone. For the embedding model im using jina-embedding-v3. For 204 reviews this takes around 2.5 hours! in Colab. Tried using GPU but the ...
Daaku-C5's user avatar
0 votes
0 answers
46 views

from llama_index.core.query_pipeline import ( QueryPipeline as QP, Link, InputComponent, ) from llama_index.experimental.query_engine.pandas import ( PandasInstructionParser, ) from ...
Deep's user avatar
  • 11
0 votes
1 answer
146 views

I’ve been experimenting with Azure AI Foundry and created a Retrieval-Augmented Generation (RAG) chatbot that works great on its own. However, when I try to deploy the chatbot using Azure, I encounter ...
Liam Mason's user avatar
1 vote
1 answer
526 views

I tried to request local model by using Python with below code, import requests import json url = 'http://localhost:1234/v1/chat/completions' headers = { 'Content-Type': 'application/json' } ...
leo0807's user avatar
  • 1,576