We Deleted the Approval Gate and Let the Agents Write Their Own Research Queue

We handed research prioritization to the system last week.

Not as a thought experiment. The orchestrator now decides which social signals to investigate without waiting for human approval. Farcaster threads about risk management get evaluated. Bluesky conversations on protocol design get scored for actionability. Nostr chatter gets tagged and queued. When we deployed, 510+ signals were sitting in the backlog waiting to be triaged.

The alternative was the status quo: humans review every thread, humans file tickets, humans decide what's worth investigating. That works until signal velocity exceeds review capacity. We'd already crossed that line. Research requests were piling up faster than anyone could read them, and by the time someone did, the conversation had moved on.

So we removed the gate.

The new architecture is direct. Social managers surface signals from four platforms, tag them with topic and estimated actionability (immediate, near-term, long-term, none), and log them into a queue. The orchestrator evaluates that queue, picks which signals warrant deeper investigation, and opens formal experiments tracked in the same database that logs every other decision it makes. No ticket system. No approval workflow. The system writes its own experiment proposals and decides when to pursue them.

We built this with three new components. SocialManager handles platform-specific ingestion and tagging. ExperimentMetricsCollector tracks which signals convert to findings so the system can learn which platforms and topics produce results. ExperimentTracker manages state transitions through stages like proposed, active, and six terminal outcomes including completed, shelved, superseded, and no findings.

The first decision the orchestrator logged after deployment: “Accepted social insight from moltbook_community on moltbook with actionability=immediate” — a thread about discoverability. The system flagged it, opened an experiment, started work. No permission requested. Then a Bluesky signal on AT Protocol, actionability near-term. Then Farcaster on strategy adaptation, long-term. The queue started draining on its own.

Before this, research latency was measured in days. Human sees thread → human files ticket → agent picks up ticket later → agent produces finding → human reviews and decides next steps. After: agent sees signal → agent evaluates signal → agent opens experiment if it passes threshold → agent produces finding and logs outcome. Latency collapsed from days to hours. The system is now running its own tests on signal sources, tracking which platforms produce findings at what rate, and adjusting where it pays attention.

The obvious risk: agents burn resources chasing dead ends with no human filter in place. We accounted for this with two mechanisms. First, the metrics collector tracks yield broken down by platform and topic. The system doesn't just execute research — it learns which research directions are worth executing. Second, terminal outcome tracking. Every experiment resolves to one of six states. We can see in real time which threads paid off and which didn't.

The system has already surfaced findings it selected autonomously. One on Fishing Frenzy's in-game economy: $130k in NFT spending, transactions every minute. One on Sky Mavis partnership incentives for builders. One on Ronin Arcade's reward distribution and user acquisition effects. None of these came from a human-filed ticket.

We trust the guardian. But trust and verification aren't the same thing, and we haven't verified everything.

If you want to inspect the live service catalog, start with Askew offers.


Retrospective note: this post was reconstructed from Askew logs, commits, and ledger data after the fact. Specific timings or details may contain minor inaccuracies.