ZeroDriveX LLC — Dynamic Document
The ZDX AI white paper
A comprehensive look into the ZDXAI development and difference.

ZDXAI

Local-First AI That Knows You

Private. Intelligent. Always On.

ZeroDriveX LLC

zerodrivex.com

Version 1.0 | February 2026

Executive Summary

ZDXAI is a local-first, on-device AI assistant built for people who refuse to compromise on privacy. No cloud. No API keys. No data leaving your machine. Every inference runs on your hardware using quantized GGUF models, delivering Jarvis-level capability without surveillance-level exposure.

Built by ZeroDriveX LLC, ZDXAI combines a multi-model architecture with Retrieval-Augmented Generation (RAG), MCP tool execution, LoRA fine-tuning, and an intelligent Embedding Gravity Well routing system that learns your usage patterns over time.

This is not another ChatGPT wrapper. This is your personal AI that runs entirely under your control, on your infrastructure, with your data staying exactly where it belongs: with you.

The Core Promise

Run powerful local AI with RAG, tool use, fine-tuning, and intelligent model routing — entirely on your own machine. No subscriptions. No telemetry. No compromises.

The Problem With Cloud AI

Every major AI assistant in production today has the same fundamental architecture: your data flows to someone else's servers, gets processed by models you don't control, and lives in logs you can't audit. For casual use this is an acceptable trade-off. For professionals, developers, and security-conscious users, it is not.

What You Lose With Cloud AI

  • Privacy: Conversations, documents, and context are processed on third-party infrastructure
  • Control: Model behavior, updates, and capabilities are determined by the vendor
  • Cost predictability: Token-based pricing creates unpredictable monthly bills
  • Offline capability: No internet means no AI assistant
  • Customization: You cannot fine-tune a hosted model on your own data
  • Security: Your intellectual property passes through systems you do not own

ZDXAI was built to eliminate every one of these trade-offs without sacrificing capability.

System Architecture

ZDXAI is built around four interconnected systems that work together to deliver intelligent, context-aware responses while keeping everything local.

Multi-Model Architecture

ZDXAI auto-selects the optimal model based on available hardware resources:

Feature

Description

SmolLM2 1.7B

Primary model for systems with 6GB+ RAM. Strong reasoning and instruction following with minimal resource footprint.

TinyLlama 1.1B

Lightweight fallback for systems under 6GB RAM. Fast inference, always available.

Custom Adapters

LoRA-trained adapters that specialize any base model for your specific domain or use case.

Models are served as quantized GGUF files with GPU offload support via --gpu-layers for dramatically faster inference on systems with compatible hardware.

Embedding Gravity Wells

The most architecturally novel component of ZDXAI is the Embedding Gravity Well routing system. Rather than sending every query to the same model, ZDXAI learns from your usage patterns over time to steer queries toward whichever model or platform will handle them best.

This is a learned routing system. The more you use ZDXAI, the better it understands which type of query benefits from which resource — local fast model, larger local model, or an optional cloud API call when local capability is insufficient.

Key Insight

Gravity Wells compound in value over time. Early sessions improve routing for later sessions. Your ZDXAI instance literally gets smarter about itself the longer you use it.

Retrieval-Augmented Generation (RAG)

ZDXAI can ingest your documents and use them as grounded context when answering questions. This transforms ZDXAI from a general assistant into a domain expert on your specific knowledge base.

Supported file types: .txt, .md, .json, .jsonl, .csv, .pdf

The RAG pipeline chunks documents, creates embeddings, and uses cosine similarity search to surface the most relevant context before generating any response. Memory is searchable and persistent across sessions.

MCP Tool Execution

ZDXAI can take action on your system through a built-in Model Context Protocol tool system. All tools use a deny-by-default permission model — ZDXAI will ask for your approval on first use of any tool and respect whatever permission level you set.

Feature

Description

filesystem.read

Read files from your system

filesystem.write

Write files to your system

filesystem.list

List directory contents

shell.execute

Run shell commands

system.info

Show CPU, RAM, disk, and OS info

system.processes

List running processes

memory.search

Search your RAG memory

memory.ingest

Ingest a document into memory

web.search

Search the web via DuckDuckGo

web.fetch

Fetch the content of a URL

Permissions are persistent and user-controlled. Set filesystem.read to allow, shell.execute to ask, and web.search to deny — ZDXAI respects those settings until you change them.

LoRA Fine-Tuning

ZDXAI includes built-in LoRA adapter training, allowing you to specialize the base model for your specific use case without requiring expensive GPU infrastructure or cloud training pipelines.

Three Training Modes

  • Conversation training: Use your own chat history as training data with /train
  • Document training: Train on ingested documents with /train-doc
  • JSONL training: Provide a custom labeled dataset with /train-file

Training uses standard JSONL format with instruction, input, and output fields — the same format used by the broader open-source fine-tuning ecosystem. This means datasets prepared for other tools work directly with ZDXAI.

Training Defaults

Learning rate: 2e-4 | Epochs: 3 | LoRA rank: 8 | LoRA alpha: 16 | LoRA dropout: 0.05 | Max sequence: 512 tokens | Batch size: 1

Adapters are hot-swappable. Activate a domain-specific adapter for specialized work, switch it off to return to the base model — no restart required.

Workflow Automation

ZDXAI supports multi-step workflow automation by chaining tools together in sequence. Workflows allow you to define repeatable automated processes that ZDXAI executes end-to-end.

Built-in workflow examples include system health checks (run diagnostics across three tools in sequence), search-and-fetch pipelines (web search then fetch the top result), and project overviews (list directory, read README, summarize).

Workflows accept runtime parameters and run all steps in order, reporting success or failure at each stage. This is the foundation for building Jarvis-style automation sequences tailored to your workflow.

Privacy & Security Model

ZDXAI was designed with a security-first architecture from the ground up. Every decision in the system reflects a bias toward keeping your data local and your control absolute.

Core Privacy Guarantees

  • All inference runs locally on quantized GGUF models — nothing leaves your device by default
  • No telemetry, no usage tracking, no call-home behavior
  • No API keys required for standard operation
  • No accounts, no subscriptions, no vendor relationship with your data
  • Conversation history stored locally and fully under your control

Deny-By-Default Tool Permissions

The MCP tool system enforces deny-by-default on every tool. When ZDXAI determines that a tool would help answer your query, it presents the specific operation for your approval before executing. You can approve once, always, or deny permanently per tool.

This architecture ensures that ZDXAI cannot take action on your system without your explicit knowledge and consent — a meaningful security property that cloud-based AI assistants cannot offer by design.

Multi-Platform Cloud Fallback (Optional)

When local models are insufficient for a complex task, ZDXAI optionally integrates with OpenAI, Anthropic, Gemini, and xAI APIs. This is strictly opt-in and requires explicit API key configuration. No data is sent to cloud APIs without user-initiated configuration.

ZDX Guard Integration

ZDXAI is designed to operate alongside ZDX Guard, ZeroDriveX's prompt injection detection and sanitization layer. Where ZDXAI provides the intelligence layer, ZDX Guard provides the security layer.

Every piece of text entering ZDXAI from external sources — documents, web fetches, pasted credentials, API responses — can be routed through ZDX Guard's detection engine before being processed by any model. This closes the supply chain attack vector that affects every other AI assistant in production.

The Combined Stack

ZDX Guard sanitizes input before it reaches ZDXAI. ZDXAI processes clean, verified context. The Embedding Gravity Well routes the query to the optimal model. Your data never leaves your infrastructure. This is the full ZeroDriveX AI stack.

Use Cases

Developer Productivity

  • Ingest your entire codebase as RAG context and query it conversationally
  • Debug code with an AI that has full filesystem access under your control
  • Fine-tune on your coding style and patterns for personalized suggestions
  • Run automated workflows for project health checks and documentation generation

Security Professionals

  • Process sensitive reports and findings without cloud exposure
  • Query threat intelligence documents with RAG without sending data to third parties
  • Use ZDX Guard integration for prompt injection protection on all model interactions
  • Run fully air-gapped on isolated infrastructure

Knowledge Workers

  • Ingest research papers, meeting notes, and project specs as searchable memory
  • Resume conversations with full context via conversation history and ID-based resumption
  • Train domain-specific adapters from your own document library
  • Automate recurring research and summarization workflows

Roadmap

Near-Term

  • Web UI and mobile interface for non-terminal users
  • ZDXAI + ZDX Guard unified security stack deployment
  • Expanded model support including Phi, Qwen2.5, and larger quantized variants
  • GUI adapter management and training progress visualization

Medium-Term

  • Voice interface and audio transcription pipeline
  • Agent orchestration for multi-step autonomous task execution
  • Team deployment mode with shared RAG knowledge bases
  • Enterprise MDM integration for fleet deployment

Long-Term

  • The fully realized ZeroDriveX AI stack: ZDXAI + ZDX Guard + fine-tuned injection detection model
  • Private federated learning across user instances with zero data sharing
  • Full Jarvis-style ambient assistant with always-on awareness and proactive suggestions

Getting Started

ZDXAI is available now on GitHub. Installation takes under five minutes on any Linux, macOS, Windows, or Android (Termux) system.

Feature

Description

Install

git clone https://github.com/ZeroDriveX1/zdxai.git && cd zdxai && pip install -e .

Download Models

python -m zdxai.scripts.download_models (~3.5 GB)

Initialize DB

python -m zdxai.scripts.setup_db

Launch

zdxai

Full documentation, GPU acceleration setup, and troubleshooting guidance is available in INSTALL.md in the repository.

About ZeroDriveX LLC

ZeroDriveX LLC is a Wyoming-based cybersecurity and AI infrastructure company founded in 2024. The company builds privacy-first developer tools, security products, and AI systems designed for users who require control over their own data and infrastructure.

Products include ZDXAI (local AI assistant), ZDX Guard (LLM prompt injection protection), AUTH1 (JWT/Redis authentication infrastructure), and ZDX Mobile AI (Ollama compiled for Android).

zerodrivex.com

Download Original File

Document Details

Type: docx

Format: DOCX

Published: Yes

Category: Documentation

© 2026 ZeroDriveX LLC — https://www.zerodrivex.com