llm11 questions
14votes
1answers

How to reduce hallucinations when using LLMs for data analysis tasks?

I'm building a system where users ask questions about their data in natural language, and an LLM generates SQL queries and interprets results. **The problem:** The LLM often "hallucinates" insights t...

askedabout 2 months ago
Raj Patel1650
17votes
1answers

How to implement RAG (Retrieval-Augmented Generation) with custom embeddings?

I want to build a RAG system for our internal documentation, but I'm confused about the embedding strategy. **Current setup:** - 500+ markdown documentation files - Using OpenAI's text-embedding-3-sm...

askedabout 2 months ago
Alex Rodriguez1920
27votes
0answers

How to reduce hallucinations when using LLMs for data analysis tasks?

I'm building a system where users ask questions about their data in natural language, and an LLM generates SQL queries and interprets results. **The problem:** The LLM often "hallucinates" insights t...

askedabout 2 months ago
Raj Patel1650
15votes
0answers

How to implement RAG (Retrieval-Augmented Generation) with custom embeddings?

I want to build a RAG system for our internal documentation, but I'm confused about the embedding strategy. **Current setup:** - 500+ markdown documentation files - Using OpenAI's text-embedding-3-sm...

askedabout 2 months ago
Alex Rodriguez1920
14votes
0answers

How to reduce hallucinations when using LLMs for data analysis tasks?

I'm building a system where users ask questions about their data in natural language, and an LLM generates SQL queries and interprets results. **The problem:** The LLM often "hallucinates" insights t...

askedabout 2 months ago
Raj Patel1650
5votes
0answers

How to implement RAG (Retrieval-Augmented Generation) with custom embeddings?

I want to build a RAG system for our internal documentation, but I'm confused about the embedding strategy. **Current setup:** - 500+ markdown documentation files - Using OpenAI's text-embedding-3-sm...

askedabout 2 months ago
Alex Rodriguez1920
19votes
1answers

How to optimize LLM inference costs in production?

Our AI application is getting expensive with GPT-4 API calls. We're spending $5000/month and growing. What strategies can reduce costs without sacrificing too much quality? Current setup: - 100k API...

askedabout 2 months ago
Mike Chen1750
19votes
1answers

How to optimize LLM inference costs in production?

Our AI application is getting expensive with GPT-4 API calls. We're spending $5000/month and growing. What strategies can reduce costs without sacrificing too much quality? Current setup: - 100k API...

askedabout 2 months ago
Mike Chen1750
0votes
0answers

Understanding LLM temperature parameter

What does the temperature parameter do in LLM models?...

askedabout 2 months ago
promptoverflow792
0votes
0answers

Understanding LLM temperature parameter

What does the temperature parameter do in LLM models?...

askedabout 2 months ago
promptoverflow792
0votes
0answers

Understanding LLM temperature parameter

What does the temperature parameter do in LLM models?...

askedabout 2 months ago
promptoverflow792
0votes
0answers

Understanding LLM temperature parameter

What does the temperature parameter do in LLM models?...

askedabout 2 months ago
promptoverflow792
0votes
0answers

Understanding LLM temperature parameter

What does the temperature parameter do in LLM models?...

askedabout 2 months ago
promptoverflow792
19votes
1answers

How to optimize LLM inference costs in production?

Our AI application is getting expensive with GPT-4 API calls. We're spending $5000/month and growing. What strategies can reduce costs without sacrificing too much quality? Current setup: - 100k API...

askedabout 2 months ago
Mike Chen1750
19votes
1answers

How to optimize LLM inference costs in production?

Our AI application is getting expensive with GPT-4 API calls. We're spending $5000/month and growing. What strategies can reduce costs without sacrificing too much quality? Current setup: - 100k API...

askedabout 2 months ago
Mike Chen1750
19votes
1answers

How to optimize LLM inference costs in production?

Our AI application is getting expensive with GPT-4 API calls. We're spending $5000/month and growing. What strategies can reduce costs without sacrificing too much quality? Current setup: - 100k API...

askedabout 2 months ago
Mike Chen1750
19votes
1answers

How to optimize LLM inference costs in production?

Our AI application is getting expensive with GPT-4 API calls. We're spending $5000/month and growing. What strategies can reduce costs without sacrificing too much quality? Current setup: - 100k API...

askedabout 2 months ago
Mike Chen1750
19votes
1answers

How to optimize LLM inference costs in production?

Our AI application is getting expensive with GPT-4 API calls. We're spending $5000/month and growing. What strategies can reduce costs without sacrificing too much quality? Current setup: - 100k API...

asked3 months ago
Mike Chen1750