-
Notifications
You must be signed in to change notification settings - Fork 13
/
config.yaml.example
64 lines (51 loc) · 2.12 KB
/
config.yaml.example
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
# CodeGate Example Configuration
# Network settings
port: 8989 # Port to listen on (1-65535)
proxy_port: 8990 # Proxy port to listen on (1-65535)
host: "localhost" # Host to bind to (use localhost for all interfaces)
# Logging configuration
log_level: "INFO" # One of: ERROR, WARNING, INFO, DEBUG
# Note: This configuration can be overridden by:
# 1. CLI arguments (--port, --proxy-port, --host, --log-level)
# 2. Environment variables (CODEGATE_APP_PORT, CODEGATE_APP_PROXY_PORT, CODEGATE_APP_HOST, CODEGATE_APP_LOG_LEVEL)
# Provider URLs
provider_urls:
openai: "https://api.openai.com/v1"
anthropic: "https://api.anthropic.com/v1"
vllm: "http://localhost:8000" # Base URL without /v1 path, it will be added automatically
# Note: Provider URLs can be overridden by environment variables:
# CODEGATE_PROVIDER_OPENAI_URL
# CODEGATE_PROVIDER_ANTHROPIC_URL
# CODEGATE_PROVIDER_VLLM_URL
# Or by CLI flags:
# --vllm-url
# --openai-url
# --anthropic-url
# Certificate configuration
certs_dir: "./certs" # Directory for certificate files
ca_cert: "ca.crt" # CA certificate file name
ca_key: "ca.key" # CA key file name
server_cert: "server.crt" # Server certificate file name
server_key: "server.key" # Server key file name
# Note: Certificate configuration can be overridden by:
# 1. CLI arguments (--certs-dir, --ca-cert, --ca-key, --server-cert, --server-key)
# 2. Environment variables:
# CODEGATE_CERTS_DIR
# CODEGATE_CA_CERT
# CODEGATE_CA_KEY
# CODEGATE_SERVER_CERT
# CODEGATE_SERVER_KEY
# Storage configuration
vec_db_path: "./sqlite_data/vectordb.db" # Path to SQLite vector database file for similarity search
# Embedding model configuration
model_base_path: "./codegate_volume/models" # Base path for model files
embedding_model: "all-minilm-L6-v2-q5_k_m.gguf" # Model to use for embeddings
####
# Inference model configuration
##
# Model to use for chatting
chat_model_path: "./codegate_volume/models/qwen2.5-coder-1.5b-instruct-q5_k_m.gguf"
# Context length of the model
chat_model_n_ctx: 32768
# Number of layers to offload to GPU. If -1, all layers are offloaded.
chat_model_n_gpu_layers: -1