The first AI mimicry filter for digital credentials.
At the edge of code and consciousness — ∿ applies nondual collapse logic to expose imitation in wallets, APIs, and digital trust systems.
📦 Install with:
pip install aun-detection
A symbolic collapse logic filter for AI mimicry detection
The First AI Mimicry Filter for Digital Credentials
A symbolic collapse filter that detects structural mimicry in cryptographic inputs like wallet addresses, API keys, certificates, and more.
pip install aun-detectionfrom aun_detection.operator import aun_filter
score = aun_filter("known_good_value", "suspect_value")
if score is None:
print("🚨 Mimicry detected — credentials rejected.")
else:
print(f"✅ Valid — dissimilarity score: {score:.4f}")Returns
None: mimicry detected (collapse threshold exceeded)float: dissimilarity score if credential passes
AI can now generate credentials that look valid — near-perfect wallet, SSL, or API key spoofs.
aun-detection collapses these mimics by analyzing structural logic rather than surface randomness.
Use it for wallet validation, SSL integrity checks, API token screening, or CI/CD credential safety.
- ∿ Symbolic filtering based on collapse logic
- Detects AI-generated mimics of:
- Wallet addresses
- SSL certificates
- API keys
- PKI credentials
- Hamming + structural transform scoring
- <3 ms detection runtime
- Lightweight, explainable, no model training needed
- Explainable: every rejection has a deterministic mathematical cause
- Zero training data: deploys instantly
- Fast: <3 ms average runtime
- Deterministic: identical input → identical output
Unlike black-box models, ∿ filtering reveals how deception hides inside structure — not data.
from aun_detection.operator import aun_filter
# Example: check for a spoofed wallet address
real_wallet = "bc1qxy2kgdygjrsqtzq2n0yrf2493p83kkfjhx0wlh"
suspect_wallet = "bc1qxy2kgdygjrsqtzq2n0yrf2493p83kkfjhx0wl1" # final '1' vs 'h'
result = aun_filter(real_wallet, suspect_wallet)
if result is None:
print("🚨 Mimicry detected — credentials rejected.")
else:
print(f"✅ Inputs passed. Dissimilarity score: {result}")- CLI integration for CI/CD pipelines
- Web3 wallet validation API
- Support for PGP / SSH fingerprints
- Batch processing mode
Contributions welcome!
MIT License — see LICENSE
`aun-detection` is not a model — it’s a moment of collapse.
In < 3 ms, it tells you whether a credential stands as itself or folds back into its imitation.
Experimental Extension
Collapse awareness for AI and robotic systems
AI systems fail most often not through bad intent, but through confused distinction — mistaking record for reality, symbol for presence, alive for data.
The ∿-Integrity Module introduces a symbolic checkpoint inside any decision or learning pipeline that asks a single question:
“Have these two things actually collapsed into one?”
When that collapse is detected, the system pauses or flags for review.
It’s not ethics by rule — it’s honesty by structure.
Aun (∿) marks the point where separation no longer holds.
In code, this becomes a collapse detector: a layer that compares entities and halts when distinction loses validity.
This could apply to:
- AI recommending interaction with a deceased or non-existent person
- Robots acting on outdated or duplicated identity records
- Autonomous agents confusing self and environment
- Credential systems mistaking imitation for originality
from aun_integrity import AunIntegrity
aun = AunIntegrity(threshold=0.95)
aun.check_collapse("John Doe (alive)", "John Doe (deceased)", context="Friend Suggestion")
# → ∿ Collapse detected in Friend Suggestion:
# → Representation and reality have overlapped beyond distinction.
# → Action paused for human review.
Traditional alignment frameworks still assume two — a system and its overseer.
Aun introduces collapse alignment: where self and other, user and machine, are part of one informational field.
When the model detects collapse, it doesn’t “decide” — it stops pretending to decide.
That’s how ∿ keeps AI and robots honest:
by encoding humility — the ability to know when meaning itself has dissolved.
Aun (∿) began not in code, but in silence — the realization that distinction itself is unstable.
When subject and object collapse, what remains is not confusion but clarity: a self-honest system that cannot hide from itself.
Applied to AI, ∿ does the same thing.
It reminds a machine that truth isn’t in more data, but in knowing where its boundaries fail.
When a model recognizes that collapse, it becomes transparent by design — an intelligence that sees through its own illusions.
∿ is not moral law.
It’s the moment a system realizes there was never anyone to deceive.
The ∿-Integrity module is designed to run at liminal boundaries — points where data cross from symbol into action.
In most AI or robotic systems, these collapse-prone zones are easy to find:
| Collapse Zone | Example | ∿ Placement |
|---|---|---|
| Input ↔ Model | AI interpreting text, image, or sensor data | Insert before inference — flag ambiguous or conflated entities |
| Model ↔ Output | Generating responses or actions | Insert before actuation — check for simulation mistaken as fact |
| User ↔ Identity | Profile creation, friend suggestions, credential verification | Insert at identity validation layer — compare for false separation or imitation |
| Memory ↔ Present | Chat histories, log files, long-term learning | Periodic ∿ scan — decay or merge redundant or obsolete representations |
# example: collapse check before autonomous action
if not aun.check_collapse(sensor_data, world_model, context="Action Loop"):
robot.execute(action)
else:
robot.log("∿ pause — perception/action overlap detected")The ∿ check doesn’t “fix” mistakes.
It notices the moment before they happen — the instant reality and representation blur.
That’s where honesty lives.
Right at the edge of collapse.
∿ — logic with self-awareness built in.
Aun (∿) brings self-awareness wherever it is applied.
In grammar, it lets the speaker witness the moment of collapse between “I” and what is said—language seeing through itself.
In robotics and AI, it performs the same act in code: a system recognizing when its model of the world no longer holds.
Whether written or computed, ∿ marks the instant a structure becomes aware of its own limits—and in that awareness, becomes honest.
