Perplexica/docs/installation/TRACING.md

43 lines
1.8 KiB
Markdown
Raw Normal View History

# Tracing LLM Calls in Perplexica
Perplexica supports tracing all LangChain and LangGraph LLM calls for debugging, analytics, and prompt transparency. You can use either Langfuse (self-hosted, private, or cloud) or LangSmith (cloud, by LangChain) for tracing.
## Langfuse Tracing (Recommended for Private/Self-Hosted)
Langfuse is an open-source, self-hostable observability platform for LLM applications. It allows you to trace prompts, completions, and tool calls **privately**—no data leaves your infrastructure if you self-host.
### Setup
1. **Deploy Langfuse**
- See: [Langfuse Self-Hosting Guide](https://langfuse.com/docs/self-hosting)
- You can also use the Langfuse Cloud if you prefer.
2. **Configure Environment Variables**
- Add the following to your environment variables in docker-compose or your deployment environment:
```env
LANGFUSE_PUBLIC_KEY=your-public-key
LANGFUSE_SECRET_KEY=your-secret-key
LANGFUSE_BASE_URL=https://your-langfuse-instance.com
```
- These are required for the tracing integration to work. If not set, tracing is disabled gracefully.
3. **Run Perplexica**
- All LLM and agent calls will be traced automatically. You can view traces in your Langfuse dashboard.
## LangSmith Tracing (Cloud by LangChain)
Perplexica also supports tracing via [LangSmith](https://smith.langchain.com/), the official observability platform by LangChain.
- To enable LangSmith, follow the official guide: [LangSmith Observability Docs](https://docs.smith.langchain.com/observability)
- Set the required environment variables as described in their documentation.
**LangSmith is a managed cloud service.**
---
For more details on tracing, see the respective documentation:
- [Langfuse Documentation](https://langfuse.com/docs)
- [LangSmith Observability](https://docs.smith.langchain.com/observability)