File size: 4,337 Bytes
a7d2416
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
# ============================================
# AllyCAT GraphRAG - Cloud Configuration
# ============================================
# This configuration uses cloud services for all components
# Recommended for production and free-tier deployments
# Docker image size: ~800 MB

# ============================================
# Pipeline Automation (Docker Only)
# ============================================
# Set to 'true' to automatically run the complete pipeline on container startup
# This will: crawl → process → save to vector DB → process graph → save to graph DB
# Recommended for cloud deployments (Heroku, AWS, Google Cloud Run)
AUTO_RUN_PIPELINE=true

# Website to crawl (required if AUTO_RUN_PIPELINE=true)
WEBSITE_URL=https://your-website.com

# Memory Optimization: Remove pipeline dependencies after completion
# Saves ~350-500 MB RAM - Highly recommended for 1GB containers
# Enables deployment on cheaper plans: DigitalOcean $12/mo (1GB) vs $25/mo (2GB)
CLEANUP_PIPELINE_DEPS=true

# ============================================
# LLM Configuration - Cloud Mode
# ============================================
LLM_RUN_ENV=cloud
# Choose your preferred cloud LLM (via LiteLLM)
LLM_MODEL=cerebras/llama3.1-8b
# Alternative models:
# LLM_MODEL=gemini/gemini-1.5-flash

# ============================================
# LLM API Keys (Set at least one)
# ============================================
# Cerebras (Fast, free tier available)
CEREBRAS_API_KEY=your_cerebras_api_key_here

# Google Gemini (Good for graph extraction)
GEMINI_API_KEY=your_gemini_api_key_here

# Nebius (Alternative provider)
NEBIUS_API_KEY=your_nebius_api_key_here

# ============================================
# Vector Database - Zilliz Cloud
# ============================================
VECTOR_DB_TYPE=cloud_zilliz
ZILLIZ_CLUSTER_ENDPOINT=https://your-cluster.zilliz.cloud
ZILLIZ_TOKEN=your_zilliz_token_here

# ============================================
# Graph Database - Neo4j Aura Cloud
# ============================================
NEO4J_URI=neo4j+s://your-instance.databases.neo4j.io
NEO4J_USERNAME=neo4j
NEO4J_PASSWORD=your_neo4j_password_here
NEO4J_DATABASE=neo4j

# ============================================
# Application Settings
# ============================================
# Choose app type: flask_graph, chainlit_graph, flask
APP_TYPE=flask_graph

# ============================================
# Port Configuration
# ============================================
# Flask Applications
FLASK_VECTOR_PORT=8081      # app_flask.py (vector-only RAG)
FLASK_GRAPH_PORT=8080       # app_flask_graph.py (GraphRAG - default)

# Chainlit Applications
CHAINLIT_VECTOR_PORT=8082   # app_chainlit.py
CHAINLIT_GRAPH_PORT=8083    # app_chainlit_graph.py

# Docker & External Services
DOCKER_PORT=8080            # External Docker exposed port (host side)
DOCKER_APP_PORT=8080        # Internal container port (container side, set to match your APP_TYPE)
OLLAMA_PORT=11434           # Ollama server port (not used in cloud mode)

# Workspace directory
# For native execution: use relative path 'workspace'
# For Docker: use absolute path '/allycat/workspace'
WORKSPACE_DIR=/allycat/workspace

# ============================================
# Website Crawling Configuration
# ============================================
WEBSITE_URL=https://example.com
CRAWL_MAX_DOWNLOADS=100
CRAWL_MAX_DEPTH=3
WAITTIME_BETWEEN_REQUESTS=0.1

# ============================================
# Embedding Model Configuration
# ============================================
EMBEDDING_MODEL=ibm-granite/granite-embedding-30m-english
EMBEDDING_LENGTH=384
HF_ENDPOINT=https://hf-mirror.com

# ============================================
# Chunking Configuration
# ============================================
CHUNK_SIZE=512
CHUNK_OVERLAP=20

# ============================================
# Graph Extraction Configuration
# ============================================
GRAPH_MIN_ENTITIES=5
GRAPH_MAX_ENTITIES=15
GRAPH_MIN_RELATIONSHIPS=3
GRAPH_MAX_RELATIONSHIPS=8
GRAPH_MIN_CONFIDENCE=0.8
GRAPH_MAX_CONTENT_CHARS=12000
GRAPH_SENTENCE_BOUNDARY_RATIO=0.7

# ============================================
# UI Settings
# ============================================
UI_STARTER_PROMPTS=What is this website? | What are upcoming events? | Who are the partners?