Cortex is a private, secure, and highly responsive desktop AI assistant designed for seamless interaction with local Large Language Models (LLMs) through the Ollama framework. All models and data stay on your device—no cloud, no third parties. Cortex keeps everything local for maximum privacy and speed.
Its rich feature set includes a robust permanent memory system, advanced conversational controls like response regeneration and chat forking, and deep customization options to tailor the AI's core behavior to your exact needs.
| Dark Mode | Light Mode |
|---|---|
![]() |
![]() |
For the most up-to-date files, source code, and release notes, please visit the official release page.
- Cortex Application: Download
Cortex.exe(66.9 MB) - Ollama & Model Installer: Download
Cortex Setup.exe(65 MB)
- Manual Setup Guide: Desktop Quick Setup Walkthrough
Use the automated setup tool to install Ollama and pull the required models, or follow the manual guide linked above for more control.
Cortex is built for users who demand full control and privacy from their AI tools. It runs entirely on your local machine, connecting directly to your Ollama models for fast, private, and reliable conversations. No cloud processing—just your hardware, your data, your control.
- 100% Local & Private: All processing happens on your machine via your Ollama instance. Nothing ever leaves your system.
- Powered by Ollama: Seamlessly integrates with any model served by Ollama, giving you the freedom to choose the right tool for the job.
- High-Performance Database: Chat history is stored in a robust SQLite database for instantaneous loading and rock-solid data integrity.
- Response Regeneration: Not satisfied with an answer? A single click prompts the AI to rethink its last response, offering a new perspective or a more refined solution.
- Conversational Forking: Explore different lines of thought without losing your place. Split a conversation at any point to create a new, independent chat thread that preserves the context up to that moment.
- Rich Code Rendering: Code blocks are displayed in a professional, dedicated container with syntax highlighting, a one-click copy button, and full theme support.
- Permanent Memory System: Teach the AI key facts about you, your projects, or your preferences. It will subtly use this information to personalize future responses.
- Custom System Instructions: Set a persistent persona or define global behavioral rules for the AI through a dedicated settings dialog. Your instructions are given the highest priority.
- Advanced Model Controls: Fine-tune the AI's core behavior by adjusting Temperature, Context Window Size, and Seed for reproducible outputs.
- Externalized AI Persona: The AI's core instructions are located in external
.txtfiles (system_prompt.txt,memory_prompt.txt), allowing advanced users to directly edit its personality and operational rules.
- Light & Dark Themes: Choose a look that fits your workspace. The entire UI updates instantly.
- Keyboard Shortcuts: A full suite of shortcuts (
Ctrl+Nfor New Chat,Ctrl+Lto focus input, etc.) for a high-velocity workflow. - Asynchronous Processing: The UI remains perfectly smooth and responsive while the AI is thinking, thanks to a multi-threaded architecture.
- Non-Intrusive Updates: The app checks for new versions in the background and notifies you discreetly within the Settings panel.
- First-Run User Agreement: A one-time EULA screen on first launch ensures transparency and clarifies user liability when interacting with local models.
![]() |
![]() |
![]() |
![]() |
![]() |
![]() |
Cortex is built on a modern, modular architecture that separates concerns for maintainability and performance.
- Presentation Layer (UI): A responsive and themeable interface built with PySide6 (Qt). Custom widgets ensure a consistent and polished user experience.
- Orchestration Layer (Control): The
Orchestratormanages application state, coordinates UI events, and dispatches long-running tasks to asynchronous workers to prevent the UI from freezing. - Data & Agent Layer (Model):
- SQLite Database: All chat history is stored in a local
cortex_db.sqlitefile, providing fast, reliable, and scalable data persistence. - Synthesis Agent: Interfaces with the Ollama client, builds complex prompts incorporating memory and user instructions, and parses the AI's response.
- Externalized Prompts: The AI's core identity and rules are loaded from external
.txtfiles, decoupling the AI's "personality" from the application's code.
- SQLite Database: All chat history is stored in a local
Cortex is a locally-run, privacy-focused AI assistant that integrates tightly with Ollama to deliver fast, context-aware, and persistent interactions. With advanced tools for conversational control and deep customization, it empowers you to create a personalized AI assistant that operates entirely on your terms—all without ever sending your data to the cloud.








