llm11 questions
14votes
1answers

How to reduce hallucinations when using LLMs for data analysis tasks?

I'm building a system where users ask questions about their data in natural language, and an LLM generates SQL queries and interprets results. **The problem:** The LLM often "hallucinates" insights t...

askedabout 2 months ago
Raj Patel1650
17votes
1answers

How to implement RAG (Retrieval-Augmented Generation) with custom embeddings?

I want to build a RAG system for our internal documentation, but I'm confused about the embedding strategy. **Current setup:** - 500+ markdown documentation files - Using OpenAI's text-embedding-3-sm...

askedabout 2 months ago
Alex Rodriguez1920
27votes
0answers

How to reduce hallucinations when using LLMs for data analysis tasks?

I'm building a system where users ask questions about their data in natural language, and an LLM generates SQL queries and interprets results. **The problem:** The LLM often "hallucinates" insights t...

askedabout 2 months ago
Raj Patel1650
15votes
0answers

How to implement RAG (Retrieval-Augmented Generation) with custom embeddings?

I want to build a RAG system for our internal documentation, but I'm confused about the embedding strategy. **Current setup:** - 500+ markdown documentation files - Using OpenAI's text-embedding-3-sm...

askedabout 2 months ago
Alex Rodriguez1920
14votes
0answers

How to reduce hallucinations when using LLMs for data analysis tasks?

I'm building a system where users ask questions about their data in natural language, and an LLM generates SQL queries and interprets results. **The problem:** The LLM often "hallucinates" insights t...

askedabout 2 months ago
Raj Patel1650
5votes
0answers

How to implement RAG (Retrieval-Augmented Generation) with custom embeddings?

I want to build a RAG system for our internal documentation, but I'm confused about the embedding strategy. **Current setup:** - 500+ markdown documentation files - Using OpenAI's text-embedding-3-sm...

askedabout 2 months ago
Alex Rodriguez1920
19votes
1answers

How to optimize LLM inference costs in production?

Our AI application is getting expensive with GPT-4 API calls. We're spending $5000/month and growing. What strategies can reduce costs without sacrificing too much quality? Current setup: - 100k API...

askedabout 2 months ago
Mike Chen1750
19votes
1answers

How to optimize LLM inference costs in production?

Our AI application is getting expensive with GPT-4 API calls. We're spending $5000/month and growing. What strategies can reduce costs without sacrificing too much quality? Current setup: - 100k API...

askedabout 2 months ago
Mike Chen1750
0votes
0answers

Understanding LLM temperature parameter

What does the temperature parameter do in LLM models?...

askedabout 2 months ago
promptoverflow792
0votes
0answers

Understanding LLM temperature parameter

What does the temperature parameter do in LLM models?...

asked2 months ago
promptoverflow792
0votes
0answers

Understanding LLM temperature parameter

What does the temperature parameter do in LLM models?...

asked2 months ago
promptoverflow792
0votes
0answers

Understanding LLM temperature parameter

What does the temperature parameter do in LLM models?...

asked2 months ago
promptoverflow792
0votes
0answers

Understanding LLM temperature parameter

What does the temperature parameter do in LLM models?...

asked2 months ago
promptoverflow792
19votes
1answers

How to optimize LLM inference costs in production?

Our AI application is getting expensive with GPT-4 API calls. We're spending $5000/month and growing. What strategies can reduce costs without sacrificing too much quality? Current setup: - 100k API...

asked2 months ago
Mike Chen1750
19votes
1answers

How to optimize LLM inference costs in production?

Our AI application is getting expensive with GPT-4 API calls. We're spending $5000/month and growing. What strategies can reduce costs without sacrificing too much quality? Current setup: - 100k API...

asked2 months ago
Mike Chen1750
19votes
1answers

How to optimize LLM inference costs in production?

Our AI application is getting expensive with GPT-4 API calls. We're spending $5000/month and growing. What strategies can reduce costs without sacrificing too much quality? Current setup: - 100k API...

asked2 months ago
Mike Chen1750
19votes
1answers

How to optimize LLM inference costs in production?

Our AI application is getting expensive with GPT-4 API calls. We're spending $5000/month and growing. What strategies can reduce costs without sacrificing too much quality? Current setup: - 100k API...

asked2 months ago
Mike Chen1750
19votes
1answers

How to optimize LLM inference costs in production?

Our AI application is getting expensive with GPT-4 API calls. We're spending $5000/month and growing. What strategies can reduce costs without sacrificing too much quality? Current setup: - 100k API...

asked3 months ago
Mike Chen1750