Files
sure/docs/hosting/langfuse.md
Juan José Mata f6dde1a098 Add Langfuse-based LLM observability (#86)
* Add Langfuse-based LLM observability

* Document Langfuse configuration

* Don't hardcode model in use
2025-08-06 23:23:07 +02:00

38 lines
1.2 KiB
Markdown
Raw Blame History

This file contains ambiguous Unicode characters
This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.
# Langfuse
This app can send traces of all LLM interactions to [Langfuse](https://langfuse.com) for debugging and usage analytics. Find them here [on
GitHub](https://github.com/langfuse/langfuse) and look at their [Open
Source statement](https://langfuse.com/open-source).
## Prerequisites
1. Create a Langfuse project (selfhosted or using their cloud offering).
2. Copy the **public key** and **secret key** from the project's settings.
## Configuration
Set the following environment variables for the Rails app:
```bash
LANGFUSE_PUBLIC_KEY=your_public_key
LANGFUSE_SECRET_KEY=your_secret_key
# Optional if selfhosting or using a nondefault domain
LANGFUSE_HOST=https://your-langfuse-domain.com
```
In Docker setups, add the variables to `compose.yml` and the accompanying `.env` file.
The initializer reads these values on boot and automatically enables tracing. If the keys are absent, the app runs normally without Langfuse.
## What Gets Tracked
* `chat_response`
* `auto_categorize`
* `auto_detect_merchants`
Each call records the prompt, model, response, and token usage when available.
## Viewing Traces
After starting the app with the variables set, visit your Langfuse dashboard to see traces and generations grouped under the `openai.*` traces.