mirror of
https://github.com/we-promise/sure.git
synced 2026-05-10 22:25:00 +00:00
* Wire conversation history through OpenAI responses API * Fix RuboCop hash brace spacing in assistant tests * Pipelock ignores * Batch fixes --------- Co-authored-by: sokiee <sokysrm@gmail.com>
This commit is contained in:
23
.env.example
23
.env.example
@@ -25,19 +25,34 @@ OPENAI_ACCESS_TOKEN=
|
||||
OPENAI_MODEL=
|
||||
OPENAI_URI_BASE=
|
||||
|
||||
# Optional: LLM token budget (applies to chat, auto-categorize, merchant detection, PDF processing).
|
||||
# Lower these for small-context local models (Ollama, LM Studio, LocalAI).
|
||||
# Defaults work for modern cloud OpenAI models without configuration.
|
||||
# LLM_CONTEXT_WINDOW=2048
|
||||
# LLM_MAX_RESPONSE_TOKENS=512
|
||||
# LLM_MAX_HISTORY_TOKENS=
|
||||
# LLM_SYSTEM_PROMPT_RESERVE=256
|
||||
# LLM_MAX_ITEMS_PER_CALL=25
|
||||
|
||||
# Optional: OpenAI-compatible capability flags
|
||||
# OPENAI_REQUEST_TIMEOUT=60 # HTTP timeout in seconds; raise for slow local models
|
||||
# OPENAI_SUPPORTS_PDF_PROCESSING=true # Set to false for endpoints without vision support
|
||||
# OPENAI_SUPPORTS_RESPONSES_ENDPOINT= # Override Responses-API vs chat.completions routing
|
||||
# LLM_JSON_MODE= # auto | strict | json_object | none
|
||||
|
||||
# Optional: External AI Assistant — delegates chat to a remote AI agent
|
||||
# instead of calling LLMs directly. The agent calls back to Sure's /mcp endpoint.
|
||||
# See docs/hosting/ai.md for full details.
|
||||
# ASSISTANT_TYPE=external
|
||||
# EXTERNAL_ASSISTANT_URL=https://your-agent-host/v1/chat/completions
|
||||
# EXTERNAL_ASSISTANT_TOKEN=your-api-token
|
||||
# EXTERNAL_ASSISTANT_TOKEN=your-api-token # pipelock:ignore
|
||||
# EXTERNAL_ASSISTANT_AGENT_ID=main
|
||||
# EXTERNAL_ASSISTANT_SESSION_KEY=agent:main:main
|
||||
# EXTERNAL_ASSISTANT_ALLOWED_EMAILS=user@example.com
|
||||
|
||||
# Optional: MCP server endpoint — enables /mcp for external AI assistants.
|
||||
# Both values are required. MCP_USER_EMAIL must match an existing user's email.
|
||||
# MCP_API_TOKEN=your-random-bearer-token
|
||||
# MCP_API_TOKEN=your-random-bearer-token # pipelock:ignore
|
||||
# MCP_USER_EMAIL=user@example.com
|
||||
|
||||
# Optional: Langfuse config
|
||||
@@ -82,7 +97,7 @@ EMAIL_SENDER=
|
||||
# Database Configuration
|
||||
DB_HOST=localhost # May need to be changed to `DB_HOST=db` if using devcontainer
|
||||
DB_PORT=5432
|
||||
POSTGRES_PASSWORD=postgres
|
||||
POSTGRES_PASSWORD=postgres # pipelock:ignore
|
||||
POSTGRES_USER=postgres
|
||||
|
||||
# Redis configuration
|
||||
@@ -94,7 +109,7 @@ REDIS_URL=redis://localhost:6379/1
|
||||
# REDIS_SENTINEL_HOSTS=sentinel1:26379,sentinel2:26379,sentinel3:26379
|
||||
# REDIS_SENTINEL_MASTER=mymaster
|
||||
# REDIS_SENTINEL_USERNAME=default
|
||||
# REDIS_PASSWORD=your-redis-password
|
||||
# REDIS_PASSWORD=your-redis-password # pipelock:ignore
|
||||
|
||||
# App Domain
|
||||
# This is the domain that your Sure instance will be hosted at. It is used to generate links in emails and other places.
|
||||
|
||||
Reference in New Issue
Block a user