Skip to content

saxenauts/syke

Syke

PyPI Python 3.12+ CI License: AGPL-3.0-only

Syke is a local-first cross-harness agentic memory. It works with the AI harnesses you already use, keeps them in sync, and adapts its memory design to match your usage patterns.

AI-enabled work now means many sessions running in parallel across different harnesses, agents, repos, and chats. A decision made in one place often fails to reach the next place where it matters. Syke connects those threads so context carries forward.

That improves continuity and memory. It also makes the harnesses more efficient. They can rely more on what has already been discovered, decided, or learned, instead of starting over each time.

Syke introduces a kind of federation across your harnesses. Each harness can keep its own memory design and still rely on a general memory agent specialized in continuity across them. In that sense, Syke becomes a memory of memory.

Start Here

Use Syke directly

pipx install syke
syke setup
syke doctor
syke context
syke ask "What changed this week?"

syke setup reviews the current setup plan first, guides provider selection if needed, validates the provider with a live Pi probe, ingests detected sources, runs initial synthesis when needed, and can enable background sync as a separate choice.

Other install methods

uv tool install

uv tool install syke
syke setup

From source

git clone https://github.com/saxenauts/syke.git && cd syke
uv sync --extra dev --locked
uv run syke setup

Use Syke through an agent

If you're an agent setting up Syke for a user, follow this sequence:

# 1. Install
pipx install syke          # or: uv tool install syke

# 2. Run agent setup
syke setup --agent         # returns JSON — read the "status" field

# 3. If status is "needs_provider", configure auth and retry
syke auth set <provider> --api-key <KEY> --use
syke setup --agent

# 4. Confirm health
syke doctor

Common providers: anthropic, openai, azure-openai-responses, kimi-coding, openrouter. For Azure, also pass --base-url https://<resource>.openai.azure.com/openai/v1 --model <model>.

Do NOT run syke setup without --agent — that launches an interactive menu.

Once setup is done:

  • syke ask "..." — deep recall across all sessions
  • syke record "..." — save notes, decisions, TODOs
  • syke context — read the current memex (fastest)
  • syke status — check what's connected and running

After onboarding, Syke installs a skill file into detected agent harnesses and keeps a live memex updated every ~15 minutes.

Why this loop is trustworthy

Syke separates capture from inference. Supported local harnesses feed raw activity into an append-only events timeline. When Syke synthesizes memory or answers a question, it does so inside a local workspace where the events snapshot is read-only, the learned-memory store is writable, and the current memex is routed back out as additive context.

That gives you something simple to trust: a record you can inspect, a memory layer that carries forward, and a loop that stays local-first until ask or synthesis calls your configured provider.

How It Works

 Claude Code      Codex        Hermes       OpenCode
      \             |             |             /
       \            |             |            /
        +-----------+-------------+-----------+
                            |
                            v
           Observe: deterministic local ingest
                 no LLM before the ledger
                            |
                            v
        +-------------------------------------------+
        | events.db                                 |
        | immutable observed timeline               |
        | append-only evidence ledger               |
        +-------------------------------------------+
                            |
                            v
        +-------------------------------------------+
        | local Syke workspace                      |
        |                                           |
        |  read   events.db snapshot                |
        |  write  syke.db learned memory            |
        |  route  MEMEX.md                          |
        |                                           |
        |  ask and synthesis run here               |
        +-------------------------------------------+
                            |
             +--------------+---------------+
             |                              |
             v                              v
      direct reads                    routed context
      syke context                    syke ask
                                      MEMEX.md
                                      capability surfaces / SKILL.md
  • events.db stores what happened.
  • syke.db stores what Syke currently believes.
  • MEMEX.md is the current map returned to future work.
  • The raw timeline stays separate from learned memory.

Current output-side scope is:

  • export the canonical MEMEX.md
  • register the Syke capability package on supported harness capability surfaces

Harness-specific memex injection or startup wiring is a later phase.

CLI

syke ask "question"   # deeper timeline and evidence-backed recall
syke context          # current memex, instant local read
syke record "note"    # write an observation back into memory
syke status           # quick operational snapshot
syke doctor           # deeper diagnostic
syke setup            # start or repair the system
syke sync             # manual refresh and synthesis cycle

Use syke ask when the agent needs more than the current memex. Use syke context when the current memex is enough and speed matters. Use syke record after useful work so the next session inherits it.

syke status is the quick snapshot. syke doctor is the deeper repair path.

Background sync commands
syke daemon start
syke daemon stop
syke daemon status
syke daemon logs

Platforms

Syke discovers supported local harnesses from its built-in Observe catalog and their expected local paths. During setup, it scans those paths, checks what is actually present on disk, validates a shipped seed adapter when one exists, and ingests what it finds.

Supported local harnesses today:

  • Claude Code: sessions, tools, projects, branches
  • Codex: rollout sessions, thread metadata, tool and model metadata
  • OpenCode: SQLite sessions and model metadata
  • Cursor: local chat/session state from official user-data roots
  • GitHub Copilot: Copilot CLI session state plus VS Code chat session files
  • Antigravity: workflow artifacts, walkthroughs, and browser recording metadata
  • Hermes: SQLite/session history and tool traces
  • Gemini CLI: chat recordings and checkpoint artifacts

Current active discovery roots in code include:

  • ~/.claude/projects
  • ~/.claude/transcripts
  • ~/.codex
  • ~/Library/Application Support/Cursor/User/... or ~/.config/Cursor/User/...
  • ~/.copilot/session-state
  • ~/Library/Application Support/Code/User/... or ~/.config/Code/User/...
  • ~/.gemini/antigravity
  • ~/.hermes
  • ~/.gemini/tmp
  • ~/.local/share/opencode

All ingestion is local-first. Syke reads these surfaces from local files and local databases.

When a supported harness exposes a native capability surface, Syke can register its canonical Syke capability package there as part of distribution.

For supported harnesses, setup is now seed-first. It validates the shipped adapter for the detected source, deploys it into the user adapter directory when validation passes, and only falls back to the Observe factory when a shipped seed is missing or fails validation on the local artifact shape.

If your harness layout is unusual, or if you want to connect a new harness yourself, use:

syke connect /path/to/your/harness

The factory remains the repair and unknown-harness path. It inspects real local artifacts, writes one adapter, validates it strictly, and deploys it into Syke's local adapters directory.

Privacy and ownership

Canonical user stores live under ~/.syke/data/{user}/. The workspace mirrors current state locally for synthesis and ask flows.

  • events.db is the immutable observed ledger.
  • syke.db is the learned-memory store.
  • MEMEX.md is the current memex returned to future sessions.
  • A content filter strips API keys, OAuth tokens, credential patterns, and private message bodies before ingest.
  • Network calls go only to your configured LLM provider during ask and synthesis.

Users should have one place under their control for the scattered material their harnesses leave behind.

What changes when Syke is running

The simplest change is that your agents stop starting from blank.

A decision made in one harness can show up in the next place where it matters. A useful pattern does not have to stay trapped inside one session. A question like "what did I ship today?" can be answered from accumulated work instead of being rebuilt from scratch.

The bigger bet is that memory management itself should improve from use. Syke keeps the raw timeline separate from learned memory, then uses synthesis to keep reshaping the memex as a map. Over time, that lets the system learn better routes through a user's own history instead of forcing one fixed memory schema on everyone.

One controlled example: on February 26, 2026, the same question was asked against the same codebase in the same minute: "What did I ship today?" Manual multi-agent orchestration was compared with syke ask.

Metric Result
Token usage 55% fewer tokens, from 970K to 431K
User-facing calls 96% fewer calls, from 51 to 2
Agents spawned 3 to 0

This is one measured example from one workflow on one date. Freshness still has a gap of up to 15 minutes. The current claim is narrower and more useful: continuity can reduce reconstruction, and memory can get better at routing through repeated use.

Learn More

Start here

Runtime and internals

Story and evolution

Provider examples
syke auth set openai --api-key YOUR_KEY --model gpt-5.4 --use
syke auth set openrouter --api-key YOUR_KEY --model openai/gpt-5.1-codex --use
syke auth login openai-codex --use
syke auth login anthropic --use
syke auth set zai --api-key KEY --model glm-5 --use
syke auth set kimi-coding --api-key KEY --model k2p5 --use
syke auth set azure-openai-responses --api-key KEY --endpoint URL --model gpt-5.4-mini --use
syke auth set localproxy --base-url URL --model MODEL --use

AGPL-3.0-only

About

dynamic identity - agentic memory - your digital extension for all your AI

Topics

Resources

License

Contributing

Security policy

Stars

Watchers

Forks

Packages

 
 
 

Contributors