Skip to main content
0 votes
0 answers
40 views

I am stuck at this error - Error Loading the Langachain pipeline 'field set'. How can i resolve this error? Below i have mention my code , import os from langchain_community.vectorstores import FAISS ...
Dilip Verma's user avatar
Advice
4 votes
3 replies
62 views

I am building a local RAG chatbot using LangChain and ChromaDB (PersistentClient). I’m encountering 'hallucinations' when the similarity search returns documents with a low relevance score. How can I ...
grace h's user avatar
Advice
0 votes
3 replies
119 views

How to build an AI-powered chatbot for a company that can answer users' questions based on company-specific data, interact with users through natural question-and-answer conversations similar to other ...
Sornab's user avatar
  • 1
Advice
1 vote
1 replies
37 views

I am planning to deploy llama 3.3 70b(FP8) Model in my EC2 instance, and I am wondering which would be good for better performance, GPU memory utilization, and operational complexity? I will be just ...
SawDeC's user avatar
  • 1
Best practices
1 vote
1 replies
33 views

I have a set of established Web Services (W3C) on IIS servers and I want to expose their functionalities to LLMs using the Model Context Protocol (MCP). What is the recommended architectural pattern ...
YuSiang's user avatar
3 votes
0 answers
52 views

I am attempting to load a large local Hugging Face model ( ./ltx-2-19b-dev.safetensors", .safetensors format) inside JupyterLab using transformers, but the kernel crashes immediately during ...
Yilmaz's user avatar
  • 51.3k
0 votes
2 answers
39 views

I'm planning to evaluate a fine-tuned LLM in the same RAG system as the base model. Therefore, I set up a PromptFoo evaluation. In the process, I came across an error that I just can't wrap my head ...
hurgel's user avatar
  • 31
Tooling
1 vote
0 replies
48 views

I want to build an npm package which any business can install into their project which then parses it and create client tools which will mimic a real user on the website. The main idea is to help ...
Amartya's user avatar
  • 35
Best practices
5 votes
8 replies
267 views

This has probably been talked about to death on this site but I still wanted to hear thoughts from other devs regarding best practices in using LLMs as guiding tools in coding I specifically avoid ...
hele's user avatar
  • 1
Best practices
2 votes
1 replies
58 views

I am trying to learn how to fine-tune models with the Huggingface suite of libraries (Transformers, PEFT, and TRL). On the Huggingface Hub there are many models that have base and instruction-tuned ...
QAH's user avatar
  • 4,269
Tooling
2 votes
1 replies
111 views

VSCode lets you access models served from Ollama. But in the Manage Models selection screen, only the models hosted by services like OpenAI, Claude, etc. show "Capabilities" that are ...
NewEndian's user avatar
  • 579
0 votes
1 answer
74 views

I'm writing a custom lm-evaluation-harness model backend that runs inference through Apple's MLX framework via mlx_lm, so I can benchmark quantized LLMs natively on Apple Silicon without routing ...
NullPointerDepressiveDisorder's user avatar
0 votes
0 answers
56 views

I have several ETL jobs from DataStage in .dsx format. I use a PowerShell script to automatically run the migration for a larger number of files. And one job migrates successfully, while another ...
danio900409's user avatar
Best practices
4 votes
0 replies
92 views

I'm building a RAG-based chatbot (using OpenAI + FastAPI + Weaviate) and I've hit a point where I need to let users upload their own documents (PDFs, DOCX, images, TXT) directly inside a chat thread ...
umair mehmood's user avatar
-1 votes
0 answers
41 views

import pdfplumber import json from ctransformers import AutoModelForCausalLM class Extractor: def __init__(self, model_name="Mistral-7B-Instruct-v0.3-Q4_K_M.gguf"): self.text = &...
Ioan's user avatar
  • 54

15 30 50 per page
1
2 3 4 5
109