Quality Signals
- GitGuardian: secret detection and leak prevention in repository history and pull requests.
- OpenAPI Contract: validates OpenAPI export and TypeScript codegen synchronization.
- SonarCloud Quality Gate: live status of code-quality gate on SonarCloud (new code).
- Snyk Vulnerabilities: live status of dependency vulnerabilities for this GitHub repository.
Venom is an open-source, local-first AI stack for practical engineering automation. It combines agent orchestration, tool execution, and long-term memory in one environment you can run and evolve locally.
Current environment recommendation: use dev and preprod. prod is still planned and not yet validated/recommended for live operation.
It is not a black box. You get explicit process control (Workflow Control Plane), transparent runtime decisions, and full request-level audit trails. You can also choose between three model stacks: ONNX, vLLM, Ollama, depending on hardware, cost, and latency goals.
- Local-first by default, cloud when needed: keep data and inference local instead of pushing everything to SaaS.
- Three runtime stacks (
ONNX/vLLM/Ollama): pick the best fit for your hardware and latency/cost target. - Process control instead of hidden behavior: Workflow Control Plane shows what runs, what is active, and what changes.
- Transparency and auditability: request tracing exposes decisions, steps, and outcomes end-to-end.
- Memory and lessons learned: knowledge persists beyond a single chat session.
- Open code and docs: easier to debug, extend, and maintain your own fork.
- Explicit quality gates in CI: SonarCloud, Snyk, and OpenAPI Contract remain visible and measurable.
- 🤖 Agent orchestration - planning and execution through specialized roles.
- 🧭 Hybrid model runtime (3-stack) - Ollama / vLLM / ONNX + cloud switching with local-first behavior.
- 💾 Memory and knowledge - persistent context, lessons learned, and knowledge reuse.
- 🎓 Workflow learning - automation built from user demonstrations.
- 🛠️ Operations and governance - service panel, policy gate, and provider cost control.
- 🔍 Transparency and full auditability - end-to-end trace of decisions, actions, and outcomes for operational trust, compliance, and faster incident review.
- 🔌 Extensibility - local tools and MCP import from Git repositories.
- Release 1.7.0 milestone: local 3-stack runtime is production-ready, giving teams better continuity and lower provider risk.
- Security/governance baseline was hardened (
Policy Gate, cost limits, fallback policy) to improve operational safety. - Workflow Control Plane and runtime governance were unified into one operating model (monitoring + configuration + activation flow).
- API traffic control and anti-ban guardrails were integrated as a shared core layer for inbound/outbound communication.
- Quality and learning track was strengthened (
Academy, intent routing rollout, test-artifact policy) to improve repeatability of delivery. - Runtime onboarding profiles (
light/llm_off/full) were stabilized invenom.sh(PL/EN/DE + headless mode). - API Contract Wave-1 was closed (OpenAPI/codegen sync, explicit response schemas, DI cleanup).
- Optional modules platform was opened: custom modules can be enabled through environment-driven registry.
- Deployment + startup - Development/production startup flow and runtime requirements.
- Environment support matrix - Current support status for
dev/preprod/prodand required configuration model. - Dashboard Guide - Web dashboard startup, scripts, and operational cockpit overview.
- Tools Usage Guide - Slash-tool routing, tool behavior, and web-search runtime requirements.
- Configuration panel - What can be edited from UI and safe editing rules.
- Frontend Next.js -
web-nextstructure, views, and implementation standards. - API traffic control - Global anti-spam/anti-ban model for inbound and outbound API traffic.
- System vision - Target platform direction and product assumptions.
- Backend architecture - Backend modules, responsibilities, and component flows.
- Hybrid AI engine - LOCAL/HYBRID/CLOUD routing and local-first policy.
- Workflow Control - Workflow control model, operations, and guardrails.
- System agents catalog - Agent roles, inputs/outputs, and runtime cooperation.
- The Academy - Learning, tuning, and training data operationalization.
- Optional Module Guide - How to author, register, and enable optional modules for Venom core.
- Memory layer - Vector/graph memory organization and retrieval rules.
- External integrations - GitHub/Slack and other integration setup.
- Coding-agent guidelines - Mandatory agent workflow and quality gates.
- Contributing - Contribution process and PR expectations.
- Testing policy - Test scope, validation commands, and quality requirements.
- QA Delivery Guide - Delivery checklist from validation to release readiness.
- LLM 3-stack benchmark baseline - Frozen reference metrics for
ollama/vllm/onnxand E2E comparison.
venom/
├── venom_core/
│ ├── api/routes/ # REST API endpoints (agents, tasks, memory, nodes)
│ ├── core/flows/ # Business flows and orchestration
│ ├── agents/ # Specialized AI agents
│ ├── execution/ # Execution layer and model routing
│ ├── perception/ # Perception (desktop_sensor, audio)
│ ├── memory/ # Long-term memory (vectors, graph, workflows)
│ └── infrastructure/ # Infrastructure (hardware, cloud, message broker)
├── web-next/ # Dashboard frontend (Next.js)
└── modules/ # Optional modules workspace (separate module repos)
- ArchitectAgent - breaks complex tasks into an execution plan.
- ExecutionPlan - plan model with steps and dependencies.
- ResearcherAgent - gathers and synthesizes web knowledge.
- WebSearchSkill - search and content extraction.
- MemorySkill - long-term memory (LanceDB).
- CoderAgent - generates code based on available knowledge.
- CriticAgent - verifies code quality.
- LibrarianAgent - manages files and project structure.
- ChatAgent - conversational assistant.
- GhostAgent - GUI automation (RPA).
- ApprenticeAgent - learns workflows by observation.
- HybridModelRouter (
venom_core/execution/model_router.py) - local/cloud routing. - Modes: LOCAL, HYBRID, CLOUD.
- Local-first: privacy and cost control first.
- Providers: Ollama/vLLM/ONNX (local), Gemini, OpenAI.
- Sensitive data can be blocked from leaving local runtime.
- DemonstrationRecorder - records user actions (mouse, keyboard, screen).
- DemonstrationAnalyzer - behavioral analysis and pixel-to-semantic mapping.
- WorkflowStore - editable procedure repository.
- GhostAgent integration - execution of generated workflows.
- Orchestrator - core coordinator.
- IntentManager - intent classification and path selection.
- TaskDispatcher - routes tasks to agents.
- Workflow Control Plane - visual workflow control.
- LessonStore - repository of experience and corrections.
- Training Pipeline - LoRA/QLoRA fine-tuning.
- Adapter Management - model adapter hot-swapping.
- Genealogy - model evolution and metric tracking.
- Backend API (FastAPI/uvicorn) and Next.js UI.
- LLM servers: Ollama, vLLM, ONNX (in-process).
- LanceDB (embedded), Redis (optional).
- Nexus and background tasks as optional processes.
git clone https://github.com/mpieniak01/Venom.git
cd Venom
pip install -r requirements.txt
cp .env.dev.example .env.dev
make startDefault requirements.txt installs minimal API/cloud profile.
If you want local runtime engines, install one of:
pip install -r requirements.txt(Ollama: no extra Python deps)pip install -r requirements-profile-vllm.txtpip install -r requirements-profile-onnx.txtpip install -r requirements-profile-onnx-cpu.txtpip install -r requirements-extras-onnx.txt(optional extras:faster-whisper+piper-tts; install after ONNX/ONNX-CPU profile)pip install -r requirements-full.txt(legacy full stack)
Use this matrix as the source of truth for "is this missing package expected or a profile bug?".
| Profile | Guaranteed scope | Explicitly not included |
|---|---|---|
requirements.txt (requirements-profile-api.txt) |
API/cloud baseline (fastapi, uvicorn, cloud providers) |
Local heavy runtime packages (vllm, onnxruntime*, lancedb, sentence-transformers) |
requirements-profile-web.txt |
API baseline + web integration runtime deps | Same heavy local runtime packages as API profile |
requirements-profile-vllm.txt |
API baseline + vllm |
ONNX stack, RAG/vector stack (lancedb, sentence-transformers) |
requirements-profile-onnx.txt |
API baseline + ONNX runtime stack | vLLM, RAG/vector stack (lancedb, sentence-transformers) |
requirements-profile-onnx-cpu.txt |
API baseline + ONNX CPU runtime stack | vLLM, ONNX CUDA packages, RAG/vector stack (lancedb, sentence-transformers) |
requirements-full.txt |
Legacy all-in profile (includes vllm, ONNX, lancedb, sentence-transformers) |
n/a |
Rules for incident triage:
- If a package is missing because the corresponding profile was not installed, this is expected behavior.
- If a package is missing and it is not guaranteed by the active profile, this is expected behavior.
- If a package is listed in the selected profile but still missing after install, this is an environment/install issue and should be treated as a defect.
git clone https://github.com/mpieniak01/Venom.git
cd Venom
scripts/docker/venom.shAfter startup:
- API:
http://localhost:8000 - UI:
http://localhost:3000
Protocol policy:
- Dev/local stack uses HTTP by default (
URL_SCHEME_POLICY=force_httpin docker profiles). - Public production should use HTTPS on reverse proxy/ingress (edge TLS).
make start # backend + frontend (dev)
make stop # stop services
make status # process status
make start-prod # production modeWarning:
make start-prodexists for technical compatibility, butprodis not yet validated/recommended for live operation.- Recommended environments:
devandpreprod.
The presentation layer runs on Next.js 16 (App Router, React 19).
- Required runtime: Node.js
>=20.9.0and npm>=10.0.0(seeweb-next/.nvmrc). - SCC (server/client components) - server components by default, interactive parts as client components.
- Shared layout (
components/layout/*) - TopBar, Sidebar, status bar, overlays.
npm --prefix web-next install
npm --prefix web-next run dev
npm --prefix web-next run build
npm --prefix web-next run test:e2e
npm --prefix web-next run lint:localesNEXT_PUBLIC_API_BASE=http://localhost:8000
NEXT_PUBLIC_WS_BASE=ws://localhost:8000/ws/events
API_PROXY_TARGET=http://localhost:8000- Force tool:
/<tool>(e.g./git,/web). - Force provider:
/gpt(OpenAI) and/gem(Gemini). - UI shows a
Forcedlabel when a prefix is detected. - UI language is sent as
preferred_languagein/api/v1/tasks. - Summary strategy (
SUMMARY_STRATEGY):llm_with_fallbackorheuristic_only.
Python 3.12+ (recommended 3.12)
semantic-kernel>=1.9.0- agent orchestration.ddgs>=1.0- web search.trafilatura- web text extraction.beautifulsoup4- HTML parsing.lancedb- vector memory database.fastapi- API server.zeroconf- mDNS service discovery.pynput- user action recording.google-genai- Gemini (optional).openai/anthropic- LLM providers (optional).
Profiles:
- requirements.txt - default minimal API/cloud profile
- requirements-profile-web.txt - API + web-next integration profile
- requirements-profile-vllm.txt - API + vLLM profile
- requirements-profile-onnx.txt - API + ONNX LLM profile (third engine)
- requirements-profile-onnx-cpu.txt - API + ONNX CPU-only profile
- requirements-extras-onnx.txt - optional extras (
faster-whisper,piper-tts), installed after ONNX LLM or ONNX CPU profile - requirements-full.txt - full legacy stack
Full checklist: docs/DEPLOYMENT_NEXT.md.
make start
make stop
make statusmake start-prod
make stopWarning:
- Treat this mode as non-recommended at current stage (no full production validation yet).
| Configuration | Commands | Estimated RAM | Use case |
|---|---|---|---|
| Minimal | make api |
~50 MB | API tests / backend-only |
| Light with local LLM | make api + make ollama-start |
~450 MB | API + local model, no UI |
| Light with UI | make api + make web |
~550 MB | Demo and quick UI validation |
| Balanced | make api + make web + make ollama-start |
~950 MB | Day-to-day work without dev autoreload |
| Heaviest (dev) | make api-dev + make web-dev + make vllm-start |
~2.8 GB | Full development and local model testing |
Dev template: .env.dev.example Preprod template: .env.preprod.example
The panel at http://localhost:3000/config supports:
- service status monitoring (backend, UI, LLM, Hive, Nexus),
- start/stop/restart from UI,
- realtime metrics (PID, port, CPU, RAM, uptime),
- quick profiles:
Full Stack,Light,LLM OFF.
- type/range validation,
- secret masking,
- active env file backup (
.env.devor.env.preprod) toconfig/env-history/, - restart hints after changes.
- editable parameter whitelist,
- service dependency validation,
- timestamped change history.
make monitor
bash scripts/diagnostics/system_snapshot.shReport (logs/diag-YYYYMMDD-HHMMSS.txt) includes:
- uptime and load average,
- memory usage,
- top CPU/RAM processes,
- Venom process status,
- open ports (8000, 3000, 8001, 11434).
make env-audit
make env-clean-safe
make env-clean-docker-safe
CONFIRM_DEEP_CLEAN=1 make env-clean-deep
make env-report-diffRun with prebuilt images:
git clone https://github.com/mpieniak01/Venom.git
cd Venom
scripts/docker/venom.shCompose profiles:
compose/compose.release.yml- end-user profile (pull prebuilt images).compose/compose.minimal.yml- developer profile (local build).compose/compose.spores.yml.tmp- Spore draft, currently inactive.
Useful commands:
scripts/docker/venom.sh
scripts/docker/run-release.sh status
scripts/docker/run-release.sh restart
scripts/docker/run-release.sh stop
scripts/docker/uninstall.sh --stack both --purge-volumes --purge-images
scripts/docker/logs.shRuntime profile (single package, selectable mode):
export VENOM_RUNTIME_PROFILE=light # light|llm_off|full
scripts/docker/run-release.sh startllm_off means no local LLM runtime (Ollama/vLLM/ONNX), but backend and UI can still use external LLM APIs (for example OpenAI/Gemini) after API key configuration.
Optional GPU mode:
export VENOM_ENABLE_GPU=auto
scripts/docker/run-release.sh restart- CI: Quick Validate + OpenAPI Contract + SonarCloud.
- Security: GitGuardian + Snyk dependency scans.
- SonarCloud = code quality and coverage on new code (maintainability, reliability, test coverage).
- Snyk = package/dependency vulnerability posture (third-party risk in Python and npm ecosystems).
pre-commit run --all-filesruns:block-docs-dev-staged,end-of-file-fixer,trailing-whitespace,check-added-large-files,check-yaml,debug-statements,ruff-check --fix,ruff-format,isort.- Extra hooks outside this command:
block-docs-dev-tracked(stagepre-push) andupdate-sonar-new-code-group(stagemanual). pre-commitcan auto-fix files; rerun it until all hooks arePassed.- Treat
mypy venom_coreas a full typing audit; the repository may include historical typing backlog not related to your change. - Local PR sequence:
test -f .venv/bin/activate || { echo "Missing .venv/bin/activate. Create .venv first."; exit 1; }
source .venv/bin/activate
pre-commit run --all-files
make pr-fast
make check-new-code-coverage- v1.4 features (planning, knowledge, memory, integrations).
- The Academy (LoRA/QLoRA).
- Workflow Control Plane.
- Provider Governance.
- Academy Hardening.
- API contract hardening (Wave-1 + Wave-2 MVP) with OpenAPI/FE synchronization.
- ONNX Runtime integrated as the third local LLM engine (3-stack: Ollama + vLLM + ONNX).
- Runtime profiles and installation strategy update (minimal/API-first + optional local stacks).
- Runtime control-plane improvements and provider/runtime governance stabilization.
- Remote models capabilities delivered (
/modelsremote tab + provider status/catalog/connectivity/validation for GPT/Gemini paths). - Global traffic-control hardening for inbound/outbound requests (limits, retry/circuit-breaker policies, safer runtime behavior under load).
- Configuration/audit observability expanded (canonical audit stream and stronger operational visibility for config/runtime changes).
- Academy/API hardening wave completed (module decomposition, contract consistency, safer upload/training/history paths).
- Pre-prod operating model finalized on shared stack (data isolation, env split
.env.dev/.env.preprod, Makefile control, guard rails, backup/restore/smoke).
- Background polling for GitHub Issues.
- Dashboard panel for external integrations.
- Recursive long-document summarization.
- Search result caching.
- Plan validation and optimization UX.
- Better end-to-end error recovery.
- GitHub webhook handling.
- MS Teams integration.
- Multi-source verification.
- Google Search API integration.
- Parallel plan step execution.
- Plan caching for similar tasks.
- GraphRAG integration.
- Code and comments: Polish or English.
- Commit messages: Conventional Commits (
feat,fix,docs,test,refactor). - Style: Black + Ruff + isort (via pre-commit).
- Tests: required for new functionality.
- Quality gates: SonarCloud must pass on PR.
- Development lead: mpieniak01.
- Architecture: Venom Core Team.
- Contributors: Contributors list.
- Microsoft Semantic Kernel, Microsoft AutoGen, OpenAI / Anthropic / Google AI, pytest, open-source community.
Venom - Autonomous AI agent system for next-generation automation
This project is distributed under the MIT license. See LICENSE.
Copyright (c) 2025-2026 Maciej Pieniak



