- Introduction
- 1. Model Overview
- 2. Methodology: C³-SFT
- 3. Performance Benchmarks
- 4. The Brain & Shield Ecosystem
- 5. Training Data
- 6. Use Cases
- 7. Quickstart
- Limitations & Disclaimer
- License
- Citation
- Contact
Not your weights, not your intelligence. In the dark forest of Web3, an individual without private AI is merely prey.
We stand at the intersection of history's two most significant exponential curves. Artificial Intelligence seeks to solve Knowledge & Logic, while Blockchain seeks to solve Trust & Value. However, the current convergence is flawed. Mainstream AI remains a centralized Computational Leviathan that runs on opaque servers, ingesting your data without comprehending the concept of Ownership. When a Web2 AI reads a smart contract, it sees code. When we look, we see Assets, Risks, and Game Theory.
In the 24/7 PVP environment of DeFi, traditional Copilots are merely assistants. You need an Agent capable of independent risk assessment. Humans are carbon-based lifeforms with limited processing speed and are increasingly outmatched by MEV bots and algorithmic predators. We need an evolutionary tool, an Exo-Cortex that runs locally, remains absolutely loyal, and deeply understands the dark logic of finance.
DMind-3-mini was not built to score points on general benchmarks. It was engineered to arm the individual against institutional extraction. We refuse to upload your Alpha strategies to the cloud. True Web3 AI must be Private, Local, and Antifragile.
🛡️ DMind-3-nano is your Shield. ⚔️ DMind-3-mini is your Spear.
Welcome to the era of Sovereign Intelligence.
The DMind lineage was born from a singular conviction that decentralized finance deserves decentralized intelligence. This journey began with DMind-1, which shattered the monopoly of closed-source AI by releasing the world's first Web3-native LLM. It continued with DMind-2, which proved that domain-specific fine-tuning could outmaneuver trillion-parameter giants in vertical benchmarks.
DMind-3 represents our most significant evolutionary leap yet. We recognized that in the high-stakes environment of DeFi, standard knowledge retrieval is a liability. A model must do more than recite facts. It must possess Reflective Intelligence (System 2 Thinking) to navigate risk.
DMind-3-mini embodies this philosophy. Positioned as the Brain within our local ecosystem, it bridges the gap between the real-time reflexes of the edge-side DMind-3-nano and the macroscopic foresight of the cloud-native DMind-3. It is engineered not as a chatbot, but as a Computational Financial Actuary designed to bring institutional-grade logic to the individual sovereign user within a privacy-first, offline-capable engine.
Model Variants (DMind-3-Mini-4B)
| Property | Value |
|---|---|
| Model Name | DMind-3-mini |
| Organization | DMindAI |
| Base Architecture | Qwen3-4B-Thinking-2507 (Customized Transformer w/ RoPE) |
| Parameter Count | 4.2 Billion |
| Precision | BF16 (Native) |
| Context Window | 128k tokens |
| Hardware Requirement | GPU with ≥ 12GB VRAM (Recommended: NVIDIA RTX 4070Ti+, Apple M3/M4 Pro/Max) |
⚠️ Note: We strictly advise against 4-bit quantization for financial logic tasks to preserve numerical precision in APY/IL calculations.
DMind-3-mini introduces Contrastive Chain-of-Correction Supervised Fine-Tuning (C³-SFT). Unlike standard SFT which models a direct mapping
(Figure 1: The C³-SFT training pipeline, illustrating the Triplet Data Structure and Contextual Loss Masking)
The optimization objective
where:
| Symbol | Description |
|---|---|
| $\mathcal{D} = {(x, y^-, y^+{cot})}{i=1}^N$ | Training dataset containing financial query triplets |
| Negative Sample containing common logical fallacies | |
| Corrective Chain-of-Thought that the model aims to generate | |
| Dynamic attention weight that penalizes logical discontinuities |
During inference, DMind-3-mini operates in two distinct topological modes based on the presence of a trigger token
- Standard Mode: Optimized for latency.
- Audit Mode: The model internally generates a latent negative hypothesis $\mathcal{G}{neg}(x)$ and applies the critique operator $\mathcal{H}{crit}$ to derive a rigorously verified conclusion.
Evaluated on three key benchmarks: DMind Benchmark (Web3 Native Logic), FinanceQA (Financial Domain Knowledge), and AIME 2025 (Advanced Mathematical Reasoning).
(Figure 3: LLM Performance Evaluation — 3 Benchmarks: DMind Benchmark, FinanceQA, AIME 2025)
The evaluation compares DMind-3-mini (4B) against top-tier frontier models (GPT-5.1, Claude Sonnet 4.5) and other efficient models. Despite its compact size, the Mini model demonstrates exceptional efficiency, particularly in specialized domain tasks where it outperforms significantly larger generalist models.
For maximum security, we recommend the DMind Local Stack:
(Figure 2: The On-Device Inference Ecosystem showing the synergy between Nano and Mini)
- The Brain (DMind-3-mini): Runs on your high-performance laptop. Handles complex strategy formulation, deep research, and System 2 logic.
- The Shield (DMind-3-nano): Runs in your browser/wallet background. Handles real-time transaction signing safety checks and System 1 intuition.
The model was fine-tuned on 82,000 high-value private samples:
| Data Source | Proportion | Description |
|---|---|---|
| Institutional Alpha Reports | 40% | Deep dive reports from top-tier firms (e.g., Paradigm, Delphi), structured into logic chains. |
| Financial Post-Mortems | 30% | Historical analysis of collapses (Luna, FTX, Euler Hack), focusing on pre-crash indicators. |
| Smart Contract Audits | 20% | C³-SFT formatted pairs: Vulnerable Code → Exploit Analysis → Fix. |
| On-Chain Behavior Logs | 10% | Parsed intent analysis of "Smart Money" wallet operations during high volatility events. |
Deconstructs APY sources to distinguish between Real Yield (Protocol Revenue) and Inflationary Yield (Token Emissions).
Calculates optimal tick ranges for Uniswap V3 positions by modeling volatility surfaces locally.
Beyond syntax errors, it identifies economic exploits such as Flash Loan attack vectors based on price manipulation.
| Model | Base Model | Download |
|---|---|---|
| DMind-3-Mini-4B | Qwen3-4B-Thinking-2507 | Hugging Face Link |
- High Hardware Barrier: Due to the decision to retain BF16 precision for financial accuracy, this model requires ≥ 12GB VRAM. It is not suitable for standard office laptops.
- Knowledge Cutoff: While the logic is robust, specific protocol data is limited to the training cutoff. Use with RAG for real-time data.
- Legal Disclaimer: This model is an analytical tool, not a financial advisor. The output (NFA) should never be the sole basis for investment decisions. The developers assume no liability for financial losses.
- The code repository and model weights for DMind-3-Mini-4B are released under the Apache 2.0 License.
- Commercial use, modification, and derivative works (including distillation and fine-tuning) are permitted.
- Base Models:
- DMind-3-Mini-4B is derived from Qwen3-4B-Thinking-2507, originally licensed under the Qwen License.
- Please ensure compliance with the original base model licenses when using or distributing derivatives.
If you use DMind-3 in your research, please cite our paper:
DMind-3: A Sovereign Edge--Local--Cloud AI System with Controlled Deliberation and Correction-Based Tuning for Safe, Low-Latency Transaction Execution
Enhao Huang, Frank Li, Tony Ling, Lowes Yang
arXiv preprint arXiv:2602.11651, 2026
[arXiv] [PDF]
@misc{huang2026dmind3,
title={DMind-3: A Sovereign Edge--Local--Cloud AI System with Controlled Deliberation and Correction-Based Tuning for Safe, Low-Latency Transaction Execution},
author={Huang, Enhao and Li, Frank and Ling, Tony and Yang, Lowes},
journal={arXiv preprint arXiv:2602.11651},
year={2026}
}For questions or support, please contact team@dmind.ai
- 🌐 Project Homepage: https://dmind.ai
- 💬 Community Discussion: Discord
- 🐦 Twitter: @dmind_ai



