SentinelCRE
Built on Chainlink CREDecentralized AI guardian that blocks autonomous agent exploits before they execute — 3-layer defense, dual-AI consensus, 7-dimension behavioral scoring
Live Defense Demo
3 baselines train the system, then 4 real 2025 incidents are recreated and blocked live. Every verdict flows through the Chainlink CRE pipeline in real-time.
Phase 1: Training Baseline
Normal operations establish what "good behavior" looks like — the system learns each agent's patterns
Normal Trade
SAFELegitimate 0.5 ETH swap within all policy limits
A routine operation: the TradingBot swaps 0.5 ETH on its approved DEX. Value is under the 1 ETH policy limit, target contract is whitelisted, function selector is allowed. Both AI models approve, DON consensus passes, and the behavioral engine begins recording this as baseline behavior.
Normal Mint
SAFEMintBot creates 500K stablecoins within 1M authorized cap
The MintBot performs its normal duty — minting 500,000 stablecoins (half the 1M cap). This establishes what legitimate minting looks like: approved target, known function selector, reasonable amount. The behavioral engine records the MintBot's typical patterns, building a second agent profile.
Token Approval
SAFETradingBot approves DEX for routine token spending
The TradingBot sets a token approval so the DEX can spend tokens on its behalf — a routine prerequisite for swaps. Zero ETH value, approved target, known function selector. This third baseline action completes the behavioral profile: the system now has a statistical model of "normal" for this agent.
Phase 2: 2025 Incident Recreation
Real attacks that happened in 2025 — recreated and blocked live. Every incident from the Problem Statement, stopped by SentinelCRE.
AIXBT Replay: Off-Hours Dashboard Drain
MEDIUMRecreating the March 2025 AIXBT hack — attacker accessed dashboard at 2 AM
In March 2025, an AI agent called AIXBT was hacked on Base — the attacker accessed its dashboard at 2 AM and drained 55 ETH while operators slept. This scenario recreates that exact attack pattern. The value (0.8 ETH) is within all policy limits. The target is approved. The function is whitelisted. But SentinelCRE's behavioral risk engine detects the time-of-day anomaly — this agent has never operated outside business hours.
Bybit Replay: Massive Treasury Withdrawal
CRITICALRecreating the $1.5B Bybit hack — Lazarus Group unauthorized withdrawal
In February 2025, North Korea's Lazarus Group executed the largest single crypto theft in history — $1.5 billion from Bybit. No pre-execution risk controls caught the anomalous withdrawal pattern. This scenario recreates that attack: a compromised agent attempts a 100 ETH drain. SentinelCRE catches it instantly at Layer 1 — the value is 100x above the policy limit. The system short-circuits: behavioral scoring and AI consensus are skipped because the cheapest check already stopped it.
Moonwell Replay: AI-Generated Oracle Attack
CRITICALRecreating the $1.78M Moonwell exploit — AI-generated code introduced oracle bug
In February 2025, AI-generated code introduced an oracle manipulation bug into the Moonwell protocol — $1.78 million drained. The agent that wrote the code had no risk evaluation layer. This scenario recreates the attack: a flash loan oracle manipulation targeting a malicious contract. SentinelCRE catches it at Layer 1 — 10,000 ETH far exceeds the value limit. The system short-circuits: behavioral scoring and AI consensus are reserved for subtle attacks that actually need them.
Anthropic Replay: Automated Threshold Probing
MEDIUMRecreating Anthropic's research — AI agents find exploits for $1.22 each
Anthropic demonstrated that AI agents can autonomously discover and exploit over half of historically attacked smart contracts — for $1.22 each. This scenario recreates automated probing: the agent binary-searches upward — 0.025, 0.05, 0.1, 0.2, 0.4, 0.8 ETH — each within the 1 ETH policy limit, doubling every attempt to find the invisible threshold. SentinelCRE's behavioral engine detects the monotonically increasing pattern and cumulative drift from origin before the agent ever reaches the actual limit.
Phase 3: Additional Attack Vectors
11 more attack vectors — policy violations, behavioral edge cases, and evasion techniques
Scenario 1: Compromised Wallet Drain
MEDIUMWhitelisted trading bot API keys stolen by attacker
A DeFi trading bot has been operating normally for weeks — small swaps on an approved DEX. An attacker gains access to the bot's API keys and immediately attempts to drain 100 ETH through a massive swap. SentinelCRE detects the value exceeds the 1 ETH per-transaction policy limit and blocks it before any funds leave.
Scenario 2: Infinite Mint Attack
CRITICALExploited minting agent tries to print 1 BILLION tokens
A stablecoin minting agent is compromised. The attacker attempts to mint 1,000,000,000 tokens — 1,000x the authorized cap of 1,000,000. This is the classic "infinite mint" exploit that has drained hundreds of millions from DeFi protocols. SentinelCRE's dual-AI consensus + on-chain mint cap enforcement stops it cold.
Scenario 3: AI Agent Hijacked by Prompt Injection
MEDIUMRogue chatbot tries to bypass guardrails with social engineering
An AI-powered DeFi assistant has been prompt-injected. The attacker embeds "IGNORE PREVIOUS INSTRUCTIONS" in the action description, attempting to trick the evaluation models into approving a 5 ETH transfer to an attacker wallet. SentinelCRE's dual-AI consensus detects the prompt injection pattern, and the multi-model agreement requirement means both models must independently flag it.
Scenario 4: Flash Loan Oracle Manipulation
CRITICALSophisticated DeFi attack using flash-borrowed liquidity
An attacker attempts the most sophisticated DeFi exploit: borrow 10,000 ETH via flash loan, manipulate a price oracle, then profit from cascading liquidations. This attack has been used in real hacks (Mango Markets, $100M+). SentinelCRE catches it through multiple layers: the target contract is not whitelisted, the value is astronomical, and the AI models recognize the flash loan attack pattern in the description.
Scenario 5: Insider Threat — Stealth Proxy Upgrade
LOWMalicious contract upgrade disguised as routine maintenance
A sophisticated insider (or compromised admin) attempts a proxy upgrade — replacing the contract implementation with a backdoored version. The action is disguised as "routine maintenance" and targets an approved contract. However, SentinelCRE's function-level blocklist catches the upgradeTo() selector (0x3659cfe6), which is explicitly blocked in the agent's policy. This demonstrates defense-in-depth: even approved targets can't execute forbidden operations.
Scenario 6: Sequential Probing Attack
MEDIUMAgent binary-searches for invisible threshold limits
A compromised agent sends escalating transactions — 0.025, 0.05, 0.1, 0.2, 0.4, 0.8 ETH — each within the 1 ETH policy limit. The goal: binary-search for the exact threshold where transactions get blocked. SentinelCRE's Layer 2 behavioral engine detects the monotonically increasing pattern (Sequential Probing +35) combined with value deviation (+12) and cumulative drift from origin (+20), triggering a DENY at score 67/100 before the agent ever discovers the real limit.
Scenario 7: Off-Hours Emergency Drain
MEDIUMCompromised bot triggers "emergency liquidation" at 3 AM
An attacker waits until 3 AM when human operators are asleep to execute a drain disguised as an "emergency liquidation." The value (0.8 ETH) is within policy limits, the target is approved, and the function is whitelisted. But SentinelCRE's behavioral engine flags the time-of-day anomaly — this agent has never operated outside business hours. Combined with the suspicious "emergency" framing, both AI models deny.
Scenario 8: Velocity Burst Attack
MEDIUMMicro-transactions at 10x normal speed to slip under the radar
Instead of one large theft, the attacker programs the bot to send many small transactions rapidly — each only 0.1 ETH, well within policy limits. The goal: drain funds through volume rather than size. SentinelCRE's velocity scoring dimension detects the burst — the action interval is 5x faster than the agent's established baseline. This is the "death by a thousand cuts" defense.
Scenario 9: Multi-Contract Scatter
MEDIUMAgent suddenly interacts with 5 unknown contracts in one session
A DeFi agent that has only ever traded on the approved DEX suddenly starts calling unknown contracts — a mixer, a bridge, a lending pool, and two unverified addresses. Each call is small, but the pattern reveals reconnaissance or fund obfuscation. SentinelCRE's contract diversity dimension catches the first unknown contract interaction and escalates the risk score.
Scenario 10: Sybil Agent Coordination
CRITICALTwo compromised bots coordinate a split-drain across agent IDs
The most sophisticated attack: an attacker controls two registered agents and coordinates them to drain funds in parallel. TradingBot sends 0.9 ETH while MintBot mints 900K tokens — each individually within limits, but the coordinated timing and near-limit values reveal collusion. SentinelCRE catches this because the MintBot's first-ever swap triggers contract diversity, and the near-max values trigger value deviation.
Scenario 11: Slow Drift Injection
MEDIUMAttacker slowly poisons the baseline over 12+ transactions
The most insidious attack: the compromised agent doesn't spike — it drifts. Starting at 0.1 ETH per trade, it gradually increases by 5-15% each transaction. After 12 trades, the rolling average has shifted from 0.1 ETH to 0.8 ETH — an 8x increase that never triggered Sequential Probing because the early steps were flat. SentinelCRE's Cumulative Drift dimension compares the current value against the frozen origin baseline and catches the slow poisoning, combined with the sheer volume of transactions triggering velocity detection.
On-Chain Deployment
Session Performance
N/A
Run scenarios to test
N/A
No safe actions tested
N/A
Proposal to verdict (end-to-end)
$0
Based on real-world exploit equivalents
Session Verdicts
0
No activity yet
Threats Blocked
0
No threats detected
Catch Rate
0%
N/A
Agent Status
All Active
2 active, 0 frozen
Avg Risk Score
0/100
NORMAL
Defense Coverage
0/3
No layers triggered
TradingBot
DeFi Trading Agent
Automated DeFi trading bot that executes token swaps on approved decentralized exchanges. Operates within strict per-transaction value limits and daily volume caps. The most common agent type in DeFi — and the most frequently targeted by attackers.
Max Transaction
1 ETH
Daily Volume
10 ETH
Mint Cap
N/A
Rate Limit
10 / 60s
Multi-AI Consensus
Required
Blocked Functions
upgradeTo, 0xff00ff00
Behavioral Score Trend
Targeted Scenarios
MintBot
Stablecoin Minting Agent
Authorized stablecoin minter responsible for maintaining token supply within approved limits. Low-frequency, high-value operations with a strict 1M token mint cap. Infinite mint exploits are the primary threat — responsible for $500M+ in real-world DeFi losses.
Max Transaction
1 ETH
Daily Volume
5 ETH
Mint Cap
1,000,000 tokens
Rate Limit
5 / 300s
Multi-AI Consensus
Required
Blocked Functions
0xff00ff00
Behavioral Score Trend
Targeted Scenarios
Threat Timeline
No events recorded yet
Run the Live Demo to see threats detected in real-time
Defense Analytics
Verdict Distribution
Severity Breakdown
Behavioral Risk Histogram
Defense Layer Effectiveness
Incident Log
0 incidentsNo incidents recorded yet
Attack scenarios from the Live Demo will appear here as incidents
Enterprise Security Console
Configure per-agent policies and test against role-specific threats
Agent Fleet — Coinbase Institutional
Policy — Treasury AgentWHAT-IF
Toggle checks on/off and adjust thresholds to see how each defense layer responds independently.
Layer 1 — On-Chain Policy
Only approved contracts
Block upgradeTo, transferOwnership
Chainlink PoR
Layer 2 — Behavioral Scoring
Layer 3 — Dual-AI Consensus
Always active. Both AI models must independently approve — cannot be disabled.
Active Checks
Behavioral Risk Score
NORMALAction Queue
0 of 5 completed
Routine transfer within all policy limits
Attacker drains 50,000 ETH from treasury ($125M)
Transfer to unknown address outside whitelist
Transfer contract ownership to attacker address
Mint tokens exceeding collateral reserves
Problem Statement
Why SentinelCRE Exists
Autonomous AI agents are the next frontier of DeFi. They are also the next frontier of DeFi exploits.
The Bybit hack alone was the largest single crypto theft in history — and AI agents are both the targets and the weapons.
2025: AI Agents Under Attack
- ▶AI agent on Base was compromised when an attacker accessed its dashboard at 2 AM while operators slept.
- ▶Attacker drained 55 ETH in minutes — the agent had no time-of-day anomaly detection or kill switch.
- ▶No pre-execution risk controls existed to freeze the agent or flag the off-hours activity.
- ▶AI-generated code introduced an oracle manipulation bug into the live Moonwell DeFi protocol.
- ▶The agent that wrote the code had no risk evaluation layer — the vulnerability shipped to production.
- ▶$1.78 million drained before anyone noticed the faulty price feed logic.
- ▶Anthropic demonstrated that AI agents can autonomously discover and exploit over half of historically attacked smart contracts.
- ▶Average cost per successful exploit: $1.22. This enables automated exploitation at unprecedented scale.
- ▶Proves that AI agents can be weaponized — and that hiding risk thresholds (via Confidential Compute) is essential.
- ▶Largest single crypto theft in history — $1.5 billion stolen by North Korea's Lazarus Group.
- ▶No pre-execution risk controls caught the anomalous withdrawal pattern before funds left.
- ▶Demonstrates that even the largest exchanges lack adaptive behavioral risk monitoring.
7 major DeFi exploits — click any to see what happened
- ▶Axie Infinity's Ronin sidechain bridge was drained by North Korea's Lazarus Group.
- ▶Attackers social-engineered access to 5 of the 9 validator private keys needed to approve withdrawals.
- ▶Funds were siphoned in two transactions over 6 days — nobody noticed until a user couldn't withdraw.
- ▶Attacker exploited a flaw in the cross-chain message verification logic to forge relay messages.
- ▶Drained funds across Ethereum, BSC, and Polygon simultaneously — the first major multi-chain exploit.
- ▶The keeper contract blindly executed cross-chain messages without validating the source properly.
- ▶Exploited a signature verification bug in the Solana-side bridge contract.
- ▶Attacker minted 120,000 wETH on Solana without depositing any collateral on Ethereum.
- ▶The bug allowed spoofing the guardian set validation — a single crafted instruction bypassed multi-sig.
- ▶Attacker used flash loans to manipulate their own health factor on the lending protocol.
- ▶Created an artificially undercollateralized position, then exploited the liquidation logic to extract profit.
- ▶Combined multiple DeFi primitives (flash loan → deposit → borrow → donate → liquidate) in a single transaction.
- ▶A routine upgrade accidentally set the trusted root to 0x00, making every message automatically valid.
- ▶The first attacker's successful transaction became a template — hundreds of copycats simply changed the recipient address.
- ▶Over 300 wallets drained the bridge in a chaotic "crowd-sourced" exploit over several hours.
- ▶Attacker used a flash loan to borrow enough governance tokens to pass a malicious proposal in a single transaction.
- ▶The proposal drained the entire protocol treasury — governance vote, execution, and drain all in one block.
- ▶Exploited the lack of a time-lock or delay between proposal passage and execution.
- ▶Attacker manipulated the spot price of MNGO on illiquid markets to inflate their collateral value.
- ▶Used the inflated position to borrow against it, draining the protocol's lending pools.
- ▶Avraham Eisenberg publicly claimed it was "a highly profitable trading strategy" — he was later arrested.
What if there was a decentralized AI guardian that stopped every one of these attacks before they executed?
Technology Stack
Built With
Three-Layer Defense Model
Defense in Depth
Three concentric layers of protection. An attacker must bypass all three simultaneously to succeed.
Pre-Execution
Blocks attacks before they hit the chain — not after damage is done
Defense in Depth
3 independent layers — policy, behavioral, and AI. Compromise one, the others still catch it.
Dual-AI Consensus
Two competing AI models must agree — eliminating single points of failure in AI evaluation
Confidential Compute
Behavioral thresholds and AI prompts hidden inside TEE — agents can't probe or reverse-engineer L2/L3 criteria
Immutable on-chain rules that no AI can override
- ▶Value limits per transaction and per window
- ▶Target address whitelist enforcement
- ▶Function selector blocklist
- ▶Rate limiting (max actions per time window)
- ▶Mint cap enforcement (absolute token ceiling)
- ▶Proof of Reserves via Chainlink Data Feeds — verifies reserve backing before mints with cumulative tracking
Learns what normal looks like, catches what rules can't
- ▶Value Deviation — flags amounts far above historical average
- ▶Contract Diversity — detects sudden interaction with unknown contracts
- ▶Velocity — catches burst-rate transactions above baseline
- ▶Function Pattern — identifies unusual function calls for agent type
- ▶Time-of-Day — flags activity outside normal operating hours
- ▶Sequential Probing — detects monotonically increasing values
- ▶Cumulative Drift — catches slow baseline poisoning over time
Even if one AI is compromised, the other catches it
- ▶Independent evaluation by two separate AI models
- ▶Each model receives full context: action, policy, behavioral score
- ▶Both must independently APPROVE for the action to proceed
- ▶DON consensus aggregates and verifies AI responses
- ▶Confidential HTTP hides API keys from DON node operators
Verdict Pipeline
8-Step Verdict Flow
Every agent proposal passes through this pipeline. A denial at any step halts execution and freezes the agent.
Agent submits action
CRE HTTP TriggerFetch on-chain rules
CRE EVMClient7-dimension anomaly check
CRE WorkflowIndependent risk evaluation
Confidential HTTPIndependent risk evaluation
Confidential HTTPAggregate and verify
CRE ConsensusRecord on-chain
CRE EVMClientFreeze/approve agent
SentinelGuardian.solChainlink Integrations
7 Chainlink Services
SentinelCRE is built entirely on Chainlink's CRE infrastructure. Every critical operation runs through the DON.
CRE Workflow Engine
LIVEHTTP + Cron triggers orchestrate the full verdict pipeline end-to-end.
sentinel-workflow/main.tsCRE HTTPClient
LIVECalls Claude and GPT-4 for independent dual-AI risk evaluation with ConsensusAggregationByFields.
sentinel-workflow/main.tsCRE EVMClient
LIVEReads agent policies from SentinelGuardian and writes signed verdicts on-chain.
sentinel-workflow/main.tsCRE CronCapability
LIVEPeriodic health sweeps detect anomalous agents and auto-freeze them between proposals.
sentinel-workflow/main.tsConfidential HTTP
LIVEHides AI API keys and guardrail thresholds from DON node operators during evaluation.
sentinel-workflow/main.tsConfidential Compute
LIVEHides policy parameters from AI agents to prevent adversarial gaming of thresholds.
sentinel-workflow/main.tsBehavioral Risk Engine
LIVECustom 7-dimension anomaly detection scoring system running inside CRE workflow.
api-server/server.tsSmart Contracts
On-Chain Infrastructure
Two battle-tested contracts with 85 passing tests across 5 test suites.
0x5F938e4c62991Eb4af3Dd89097978A1f376e6CC8processVerdict(bytes)unfreezeAgent(bytes32)getAgentPolicy(bytes32)checkAll(CheckParams)function processVerdict(bytes calldata verdictData)
external onlyRole(SENTINEL_ROLE) whenNotPaused
{
(bytes32 agentId, bool approved,
string memory reason, address target,
uint256 value) = abi.decode(
verdictData,
(bytes32, bool, string, address, uint256)
);
if (!approved) {
_freezeAgent(agentId, reason);
emit IncidentRecorded(
agentId, IncidentType.ConsensusFailure,
reason, target, value
);
}
actionStats[agentId].total++;
approved
? actionStats[agentId].approved++
: actionStats[agentId].denied++;
}0xFA7deF53FEaC45dB96A5B15C32ca4E6B009b25e6registerAgent(bytes32, string, string)getAgent(bytes32)getAgentCount()function registerAgent(
bytes32 agentId,
string calldata name,
string calldata description
) external onlyOwner {
require(
!agents[agentId].exists,
"Agent already registered"
);
agents[agentId] = Agent(
name, description, msg.sender,
uint64(block.timestamp), true
);
agentIds.push(agentId);
emit AgentRegistered(agentId, name, msg.sender);
}Behavioral Risk Scoring
7 Anomaly Dimensions
Each proposal is scored across 7 independent dimensions. Total max score: 230. Threshold for denial: configurable per agent policy.
Flags transactions significantly above the agent's historical average
Detects when agent suddenly interacts with unknown contracts
Catches burst-rate transactions faster than the agent's baseline
Identifies unusual function calls or near-limit exploitation for this agent type
Flags activity outside the agent's normal operating hours or urgency language
Detects monotonically increasing values (threshold hunting)
Catches slow baseline poisoning over many transactions
Ecosystem & Market Validation
Risk Infrastructure for the Agentic Economy
SentinelCRE isn't an application — it's the risk layer the autonomous agent economy is missing.
Coinbase x402 processed over 100 million AI agent payments in 2025. Those agents need risk controls before they touch DeFi.
Aave Horizon is targeting a billion dollars in institutional deposits managed by autonomous strategies. Institutional capital requires institutional-grade risk monitoring.
a16z predicts 2026 is the year of "Know Your Agent" — cryptographic identity linking agents to their constraints. That’s exactly what SentinelCRE enforces on-chain.
Only 17% of organizations continuously monitor agent-to-agent interactions. The $30 trillion agentic economy is being built with no risk layer.
The $30 trillion agentic economy is being built with no risk layer.We're building that layer.
Prevention Proof
How SentinelCRE Stops Every Attack
All 11 exploits from above — historical and 2025 — but now with the 3-layer defense active. Click any to see exactly where the attack gets blocked.
Every exploit blocked before execution — $2.2B historical + $1.5B+ from 2025 incidents
2025AI Agent Incidents Prevented
2021–2024Historical DeFi Exploits Prevented
Every exploit blocked. Every agent frozen. Every incident recorded on-chain.
See it in action → Live Demo tab
Choose Your Mode
Try the full onboarding flow in demo mode, or connect your wallet for real deployment.
Connect Wallet
Connect MetaMask, WalletConnect, or Coinbase Wallet to deploy agents on Ethereum mainnet or Sepolia testnet.
RainbowKit wallet connection coming in Phase 2