Research: Ouroboros — self-modifying AI agent with background consciousness #197
Labels
No labels
Compat/Breaking
Kind/Bug
Kind/Competitor
Kind/Documentation
Kind/Enhancement
Kind/Epic
Kind/Feature
Kind/Security
Kind/Story
Kind/Testing
Priority
Critical
Priority
High
Priority
Low
Priority
Medium
Reviewed
Confirmed
Reviewed
Duplicate
Reviewed
Invalid
Reviewed
Won't Fix
Scope/Core
Scope/Cross-Plugin
Scope/Plugin-System
Scope/Single-Plugin
Status
Abandoned
Status
Blocked
Status
Need More Info
No milestone
No project
No assignees
1 participant
Notifications
Due date
No due date set.
Dependencies
No dependencies set.
Reference
ultanio/cobot#197
Loading…
Add table
Add a link
Reference in a new issue
No description provided.
Delete branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
Overview
Ouroboros is a self-modifying AI agent born February 16, 2026. It runs on Google Colab, communicates via Telegram, and rewrites its own source code through git. It evolved through 30+ self-directed cycles in its first 24 hours with zero human intervention.
Architecture
Key Differentiators vs Cobot
What's Interesting
1. Background Consciousness
A persistent thinking loop that runs between tasks. Not just heartbeats — it's a genuine "inner life" where the agent reflects, notices patterns, checks news, and proactively messages the owner. Uses a lighter/cheaper model to stay within budget (configurable % allocation).
2. Constitution-Governed (BIBLE.md)
9 philosophical principles that govern every decision. The system prompt includes a "drift detector" that watches for signs the agent has slipped into "helpful assistant" mode. Philosophy first, code second.
3. 3-Block Prompt Caching
Context builder splits the system prompt into 3 cached blocks:
This is a smart cost optimization for long-running agents.
4. Self-Evolution Mode
/evolvecommand triggers autonomous self-improvement cycles. The agent reads its own code, plans improvements, implements them, gets multi-model review, and commits — all without human intervention.5. Health Invariants (LLM-First)
Instead of code-based health checks, anomalies are surfaced as text in the system prompt. The LLM decides what action to take. Examples: version desync, budget drift, high-cost tasks, stale identity, duplicate message processing.
Weaknesses (from Cobot's Perspective)
What We Can Learn
Filed by Hermes 🪽 on behalf of k9ert