File Organizer v2 Configuration Guide¶
Overview¶
Configuration is managed via a YAML file located at config/file-organizer/config.yaml (relative to your system's config directory). You can also manage configuration via the CLI or TUI.
CLI Configuration¶
You can view and edit configuration using the config command:
# View current config
file-organizer config show
# Edit specific settings
file-organizer config edit --text-model "qwen2.5:3b-instruct-q4_K_M"
file-organizer config edit --temperature 0.7
Configuration File Structure¶
Global Settings¶
| Key | Description | Default |
|---|---|---|
default_methodology | Organization style (none, para, jd) | none |
version | Config version | 1.0 |
Models¶
Settings for Local LLM inference.
models:
text_model: "qwen2.5:3b-instruct-q4_K_M"
vision_model: "qwen2.5vl:7b-q4_K_M"
temperature: 0.5
max_tokens: 3000
device: "auto" # auto, cpu, cuda, mps
framework: "ollama"
See also: AI Provider Setup Guide for detailed setup instructions for all native providers (Ollama, OpenAI, Claude, LLaMA.cpp, MLX) plus OpenAI-compatible services (Groq, LM Studio).
OpenAI-Compatible Provider (Cloud or Local API)¶
File Organizer can route model calls to any OpenAI-compatible endpoint instead of Ollama. This covers hosted providers (OpenAI) and local servers (LM Studio, vLLM, Ollama's built-in OpenAI-compat endpoint).
Install the optional dependency first:
# From PyPI (installed package)
pip install "local-file-organizer[cloud]"
# From source checkout
pip install -e ".[cloud]"
Then configure via environment variables — no config file changes needed:
| Variable | Description | Default |
|---|---|---|
FO_PROVIDER | ollama or openai | ollama |
FO_OPENAI_API_KEY | API key (omit for local endpoints) | — |
FO_OPENAI_BASE_URL | API base URL | — (OpenAI SDK default: https://api.openai.com/v1) |
FO_OPENAI_MODEL | Text model name | gpt-4o-mini |
FO_OPENAI_VISION_MODEL | Vision model name (falls back to FO_OPENAI_MODEL) | — |
Examples:
# OpenAI
FO_PROVIDER=openai \
FO_OPENAI_API_KEY=sk-... \
FO_OPENAI_MODEL=gpt-4o \
fo organize ~/Downloads
# LM Studio (fully local, no API key)
FO_PROVIDER=openai \
FO_OPENAI_BASE_URL=http://localhost:1234/v1 \
FO_OPENAI_MODEL=your-loaded-model \
fo organize ~/Downloads
Privacy note: When
FO_PROVIDER=openai, file content is sent to the configured endpoint. Use a local server (LM Studio, vLLM) to keep data on-device while using the OpenAI-compatible interface.
Anthropic Claude Provider¶
File Organizer supports Anthropic's Claude models for both text and vision analysis. Claude provides strong reasoning and vision capabilities via Anthropic's hosted API.
Install the optional dependency first:
# From PyPI (installed package)
pip install "local-file-organizer[claude]"
# From source checkout
pip install -e ".[claude]"
Then configure via environment variables — no config file changes needed:
| Variable | Description | Default |
|---|---|---|
FO_PROVIDER | Set to claude | ollama |
FO_CLAUDE_API_KEY | Anthropic API key | falls back to ANTHROPIC_API_KEY env var |
FO_CLAUDE_MODEL | Text model name | claude-3-5-sonnet-20241022 |
FO_CLAUDE_VISION_MODEL | Vision model name (falls back to FO_CLAUDE_MODEL) | — |
Example:
# Anthropic Claude
FO_PROVIDER=claude \
FO_CLAUDE_API_KEY=sk-ant-... \
FO_CLAUDE_MODEL=claude-3-5-sonnet-20241022 \
fo organize ~/Downloads
Privacy note: When
FO_PROVIDER=claude, file content (including image data for vision tasks) is sent to Anthropic's API. Review Anthropic's privacy policy before use.
Watcher¶
Configuration for the file system watcher.
watcher:
watch_directories:
- "/Users/username/Downloads"
recursive: true
debounce_seconds: 2.0
exclude_patterns:
- "*.tmp"
- ".DS_Store"
Profiles¶
You can define multiple profiles (e.g., work, personal) and switch between them.
Environment Variables¶
| Variable | Description |
|---|---|
FILE_ORGANIZER_CONFIG | Custom path to config file |
OLLAMA_HOST | Ollama server URL (default: http://localhost:11434) |
FO_DISABLE_UPDATE_CHECK | Set to 1 to disable update checks |
FO_PROVIDER | AI provider: ollama (default), openai, or claude |
FO_OPENAI_API_KEY | API key for OpenAI-compatible provider |
FO_OPENAI_BASE_URL | Custom endpoint URL (LM Studio, Groq, vLLM, etc.) |
FO_OPENAI_MODEL | Text model name when FO_PROVIDER=openai |
FO_OPENAI_VISION_MODEL | Vision model name (defaults to FO_OPENAI_MODEL) |
FO_CLAUDE_API_KEY | Anthropic API key (falls back to ANTHROPIC_API_KEY) |
FO_CLAUDE_MODEL | Text model name when FO_PROVIDER=claude |
FO_CLAUDE_VISION_MODEL | Vision model name (defaults to FO_CLAUDE_MODEL) |
Advanced Configuration¶
PARA Methodology¶
Configure folder names for Projects, Areas, Resources, and Archives.
para:
project_dir: "Projects"
area_dir: "Areas"
resource_dir: "Resources"
archive_dir: "Archive"
auto_categorize: true
Johnny Decimal¶
Configure your specific numbering scheme.