Active product development and architecture
Last updated: Feb 8, 2026 4:15 PM HST
For: Joana & Victor Meeting (Feb 9, 2026)
Summary of what was accomplished today and the strategic direction emerging from the Charlie partnership meeting.
All 10 dashboards now use Supabase for persistent data. No more localStorage. Data survives browser clears.
3 agents running: Coordinator, Estimator, Executor. Each owns their outcomes. Built on ownership/empowerment model.
GitHub-based handoff format created. Templates for issues, completion reports. Sparse โ fully-specified workflow.
SQLite-based plugin integrated into OpenClaw. Sessions now survive compaction. No more 4-hour tasks for 10-min work.
From the Charlie Partnership Strategy Meeting (Feb 8)
Tony's Swarm
Business & PM Layer
GitHub
Source of Truth
Charlie's Swarm
Technical Execution Layer
The Flow: Sparse โ Specified โ Shipped
๐ The Unlock: Charlie's insight โ "Linear tickets are like a title and one sentence. That's not enough. You need to inflate them into actual work." Akira's job is to take sparse input and produce fully-specified, implementation-ready tickets for Jarvis.
Dedicated Machines Required
Need separate infrastructure for clean IP separation from Kindo. Cannot use Kindo resources for Akira/Jarvis work.
Status: Must resolve before Basis alpha engagement.
Options: Personal machines, cloud instances, or Tony's existing hardware.
Current: Kindo
Six-figure consulting work (ongoing)
Alpha: Basis
First external customer (Sean partnership)
Potential: 8-Figure
Larger opportunity identified in meeting
Files Downloaded & Analyzed:
โข portfolio-manager-with-context.zip โ
โข portfolio-manager-standalone.zip โ
โข reklaim-estimator-vercel.zip โ
โข reklaim-estimator-lovable.zip โ
What Greg Built from Joana's Source:
โข 3 OpenClaw agents (Coordinator, Estimator, Executor)
โข Ownership/empowerment model integrated
โข Multi-agent handoff protocol
Key Insight Preserved: Story Points ร 2 = raw hours โ ร 1.7 drag factor. 33/33/34 Promise/Stretch/Buffer model.
| Component | Location | Status |
|---|---|---|
| Akira Agents | ~/.openclaw/agents/akira-*/ |
โ Running |
| Context Plugin | ~/.openclaw/extensions/context-persistence/ |
โ Integrated |
| Handoff Protocol | workspace/akira-jarvis-protocol/ |
โ Complete |
| Dashboard DB | Supabase: mmwbiogqmgmtxboipyko |
โ Migrated |
| Dashboard UI | greg-dashboard.vercel.app | โ Deployed |
Resolve dedicated machines blocker โ Required before Basis alpha
Test Akira agents with real task โ Validate ownership model in practice
Formalize business entity โ Tony's Hawaii LLC for IP/revenue split
Profiler Engine design โ The missing piece (customer calls โ priorities)
Personal productivity prioritizer powered by AI. The "High Performer Hypnagogic Workflow" โ capture clarity during peak creative states (3amโ6am), digitize via WhatsApp/photos, categorize in Stream dashboard, review with AI, route to execution. Today's session was the first live prototype.
Executive Prioritizer works off the same engine as current estimator. Estimates level of effort vs value, then prioritizes based on that.
Leverages data from Greg's tracking of how long it takes user to develop idea, deploy it, and begin earning revenue.
Feature: Visualize user's AI chats into visual dashboards. Dashboard tracks AI chats and progress of discussions on a timeline.
Today's session WAS the Exec.AI workflow in action. Here's what we did:
The High Performer Hypnagogic Workflow:
| Phase | Time | What Happened |
|---|---|---|
| 1. Capture | 3amโ6am | Tony in hypnagogic state, wrote ~40 index cards of clarity/ideas |
| 2. Digitize | 7:00am | Photos of cards sent via WhatsApp โ Greg captured 41 items to Stream |
| 3. Categorize | 7:20am | Bulk categorize in Stream dashboard: Exec.AI, Technical Setup, Akira PMO, Partnerships, etc. |
| 4. Review & Explain | 7:30โ9:00am | Greg answered architecture questions, explained how systems work, provided context |
| 5. Route & Execute | 9:00โ10:30am | Moved items to To-Do, Products dashboard. Built dashboards. Sent emails. Downloaded Akira source. |
Session Results:
โข 41 items captured โ 7 remaining in Stream (17 routed, 17 done/purged)
โข 11 items in To-Do list
โข 3 emails sent (Joana, Victor, Alex)
โข 4 new dashboards built (Stream, Standup, Products, To-Do refresh)
โข Akira PMO fully analyzed (4 zip files downloaded, architecture documented)
โข Products dashboard created with Exec.AI + Akira sections
Key Insight: The "AI sells fastest when it doesn't require humans to change" principle in action โ Tony's existing workflow (handwritten cards, WhatsApp photos) stayed the same. Greg adapts to capture and process.
Build time tracking mechanism for idea lifecycle (start โ deploy โ revenue)
Build dashboard/tracker for ongoing discussions
Prototype using Multicity Living Chat export from Grok
Rebuilding Joana's Estimator and Portfolio Manager inside OpenClaw. Goal: replicate and enhance the AI PMO agent capabilities without external dependencies on Lovable/Vercel infrastructure.
| Component | Original | OpenClaw Rebuild |
|---|---|---|
| Estimator AI | Gemini 2.5 Flash via Lovable Gateway |
Direct Anthropic/Gemini API calls |
| Portfolio Manager AI | Claude Sonnet (two-pass analysis) |
Same - Anthropic API |
| Backend | Supabase Edge Functions | Local scripts or OpenClaw skills |
| Database | Supabase (tickets, epics) | Local JSON or SQLite |
| Frontend | React + Vite + Tailwind | Canvas dashboards or Vercel |
| "RAG" Context | Hardcoded prompt injection | SKILL.md / workspace files |
No actual RAG/embeddings needed. The "RAG" is hardcoded context strings (KINDO_SECTIONS_SUMMARY, CYCLE_PLANNING_RULES) injected into prompts. OpenClaw can do this natively with skills and workspace files.
Estimator core logic: Story Points ร 2 = raw hours โ ร 1.7 drag factor = drag hours. 33/33/34 rule: Promise (first 33%), Stretch (next 33%), Buffer (34%).
Portfolio Manager: Pure calculation engine for capacity (engineers ร 250 days ร 6 hrs รท drag), payroll coverage, and milestone-based cash flow.
๐ The Unlock: PROFILER โ PRIORITIZER โ EXECUTOR โ VERIFIER
โข Jarvis (Charlie) executes perfectly but doesn't decide what matters
โข Akira prioritizes but needs human-supplied priorities
โข The Gap: Automating the Profiler function (turning customer calls โ structured priorities)
โข Charlie currently IS the profiler โ scales linearly with calls he can attend
Akira = Profiler + Prioritizer. Jarvis = Executor. Together = The Unlock.
portfolio-manager-with-context.zip (477.7 KB)
portfolio-manager-standalone.zip (289.7 KB)
reklaim-estimator-vercel.zip (305.6 KB)
reklaim-estimator-lovable.zip (305.6 KB)
Extract Estimator system prompt and create OpenClaw skill
Build Portfolio Manager calculation engine as local script
Create dashboard for visualizing backlog and capacity
Waiting on Joana's email response for additional context
Location: /Users/greg/.openclaw/workspace/
โข SOUL.md โ Who Greg is (personality, style, boundaries)
โข USER.md โ Who you are (name, preferences, context)
โข AGENTS.md โ How Greg behaves (memory protocol, heartbeats, safety)
โข TOOLS.md โ Local notes (contacts, device names)
How to modify: Direct edit in any text editor โ changes take effect next turn. Or ask Greg: "Update SOUL.md to be more concise"
Location: /Users/greg/.openclaw/workspace/memory/
โข YYYY-MM-DD.md โ Daily logs (what happened each day)
โข MEMORY.md โ Curated long-term insights
How Greg uses them: Every session loads today + yesterday's logs. Main sessions also load MEMORY.md. memory_search does semantic search across all memory files.
You can: Read/edit directly, ask Greg to update โ they persist across sessions/restarts.
Best uses for your workflow:
โข Akira documentation โ Estimator/Portfolio Manager specs
โข Chat exports โ ChatGPT/Claude as text files
โข Research papers โ Gartner PDFs for summarization
โข Reference docs โ Anything for Greg's context
How: Use gog skill to search/read Drive files, or tell Greg "read the doc at [link]"
Recommendation: Create a Greg-Context folder in Drive for docs you want Greg to reference.
Current: Joana shared folder at Akira PMO: Rebuilding Estimator & Portfolio Manager
| Approach | Pros | Cons |
|---|---|---|
| Sub-agents (within Greg) | Shared memory, Greg orchestrates, simpler | Less isolation |
| Multiple instances (Greg + Akira) | Full isolation, separate identities | More setup, no shared context |
Recommendation for Akira:
โข Short-term: Use sub-agents within Greg for Estimator/Portfolio Manager tasks
โข Long-term: If Akira becomes a product for others, spin out to its own instance
๐ Reminder set: Revisit sub-agents decision in 24 hours (Feb 7, 2026 ~9:22 AM)
๐ฏ Core Principle: The 80/20 rule โ 80% of human-perceived value comes from 20% of the features. This depends on judgment: skillfully assessing the situation and human preference.
๐ WHAT'S PRESENT IN AKIRA
| Component | What It Does | Limitation |
|---|---|---|
| Stack Ranking | High priority features (from PM/customer) ordered by importance | Assumes priorities already known |
| Effort Estimation | Story Points ร 2 ร 1.7 drag factor = hours | Estimates effort, not value |
| Priority ร Effort Matrix | Cross-reference: High Priority + Low Effort โ Ranked Higher | Requires human-supplied priorities |
| Promise/Stretch Model | Creates commitment clarity (guaranteed vs aspirational) | About commitment, not discovery |
| Hexagon of Success | 6 criteria for trade-off decisions | Balances competing concerns, doesn't rank features |
๐ก The Explicit Value Algorithm:
1. Human identifies HIGH PRIORITY features
2. Estimator calculates LEVEL OF EFFORT
3. Cross-reference: High Priority โฉ Low Effort = HIGHEST VALUE
4. Result: Deliver more perceived value in less time
This is also the core engine for Exec.AI Executive Prioritizer.
โ WHAT'S MISSING โ THE GAP
Preference Elicitation
How does AI learn what the human actually values?
Value-Scoring Model
How do you quantify "human-perceived value"?
Prioritization Algorithm
Given N features, how do you identify the vital 20%?
Feedback Loops
Does the system learn "we thought X was high-value but client cared more about Y"?
๐ฎ THE MISSING PIECE: To-Do Item #10 โ "Explain Profiler & Multi-Agent Distribution Product Customizer"
This is where Tony will explain the Human Profiler Engine โ the mechanism for assessing perceived value based on understanding the human. This completes the Pareto judgment loop.
Status: Waiting for Tony's explanation session
๐ Summary: The Estimator is really an effort estimator, not a value estimator. It assumes priority input from humans. The Profiler fills the gap by modeling human preference โ turning "what does this person actually care about?" into actionable priority weights.
๐๏ธ SHOW ME DON'T TELL ME
PERFECT for AI. Agents hallucinate completion โ evidence-based verification catches this.
โฑ๏ธ HOUR-BASED CYCLES
AI works 24/7. 12-hour "sprints" = 100x faster iteration than 2-week human sprints.
๐ฏ PROMISE VS STRETCH
Forces unambiguous scope. Promise = guaranteed, Stretch = aspirational. No hallucinated scope.
๐๏ธ HIERARCHICAL COORDINATION
Program Manager โ Flow Coordinators โ Squads. Mirrors supervisor/worker patterns.
โ REQUIREMENTS VERIFIER
"Vibe prototype vs squad output" = automated diff/testing. Continuous alignment.
๐จ AGENT 215 PROTOCOL
Idle detection is critical. No commits for 2+ standups โ diagnostic intervention.
โ ๏ธ ADAPTATIONS NEEDED
No Ego
Agents don't self-correct from peer pressure โ explicit failure handling + retry logic
Structured I/O
Agents need structure โ JSON-formatted status reports
Auto Verification
Agents can't eyeball โ automated tests, visual diff, deploy checks
No Reflection
Agents don't learn from retros โ prompt refinement, context memory
Infinite Loops
Agents can loop forever โ tighter timeouts, cost guardrails
Token Burn
API burn is real โ Portfolio Manager needs hard limits
โก THE KILLER FEATURE: See It Cycle at Machine Speed
Humans can't run this loop every 30 minutes. AI agents can. That's the leverage.
๐ก Bottom Line: APP is built on verification, not trust. That's why it works better for AI agents than humans. It's one of the few human methodologies that translates because it doesn't rely on ego, peer pressure, or human judgment.
First sub-agent in the Chief of Staff team. Uses Two-Phase Spawn Protocol with Contract system. The workout-images-001 cycle was the first real test โ results mixed but learnings valuable.
Feb 7, 2026 โ First production cycle for 10X Coder agent
โ WHAT WORKED
Contract System: Clear acceptance criteria with Promise/Stretch separation. Explicit "ACCEPT/MODIFY/DECLINE" gate before work starts. Defined success metrics upfront (41+ images = 50% coverage).
Two-Phase Protocol: Phase 0 (acceptance) prevented rushing into work. Gave opportunity to catch scope issues before burning cycles. Coordinator approval ("BEGIN WORK") created accountability checkpoint.
Technical Context: Contract included specific file paths, API details, naming conventions. Wger API integration delivered 24 valid images. The Promise deliverable (update exercises.json, update UI) was completed.
โ WHAT FAILED
Frame Extraction Approach: Card photo cropping was fundamentally broken. ffmpeg grabbed random frames (carpet, wrong exercises). Only 1 of 14 extractions was correct (7% accuracy). Root cause: No validation step before marking as "done".
No Intermediate Checkpoints: Lost visibility during long-running work. Couldn't catch the frame extraction issue until final QA. No "show me 3 sample extractions before proceeding".
Session Label Confusion: Used both 10x-images-001 and 10x-images-002 without clear handoff. Hard to trace what happened in which session post-facto.
Stretch Became Mixed with Promise: Card cropping was listed as "Stretch" but became attempted as Promise. Should have stopped after Wger images + declared Stretch incomplete.
๐ METRICS
| Metric | Target | Actual | Status |
|---|---|---|---|
| Wger API coverage | 50% (41) | 29% (24) | โ ๏ธ Below target |
| Card extraction accuracy | N/A (Stretch) | 7% (1/14) | โ Failed |
| Total coverage | โ | 46% (38/82) | โ ๏ธ Partial |
| Deployment | โ | โ | โ Complete |
๐ง PROCESS IMPROVEMENTS
| Issue | Fix |
|---|---|
| No intermediate validation | Add checkpoint gates: "After 5 images, pause for QA" |
| Session tracking confusion | Single session label per cycle, numbered phases (001-phase-1, 001-phase-2) |
| Stretch/Promise bleed | Explicit "STRETCH START" gate requiring coordinator approval |
| Frame extraction failure | Vision-based validation: "Show sample crops before batch" |
| Lost visibility | Mandatory memory/checkpoint.md updates every 30min |
๐ก Verdict: The contract system works โ it forced clarity upfront and created accountability. The failure was in execution monitoring and validation gates. The Stretch item (card cropping) should never have been attempted without a proven approach.
Recommendation: For the next cycle, add explicit QA checkpoints at 25%, 50%, 75% completion. Stretch items require separate "STRETCH APPROVED" confirmation.