14 KiB
AI Context - DeBros Network Cluster
Table of Contents
- Project Overview
- Architecture Overview
- Codebase Structure
- Key Components
- Configuration System
- Node vs Client Roles
- Network Protocol & Data Flow
- Gateway Service
- Build & Development
- API Reference
- Troubleshooting
- Example Application: Anchat
Project Overview
DeBros Network Cluster is a decentralized peer-to-peer (P2P) system built in Go, providing distributed database operations, key-value storage, pub/sub messaging, and peer management. It is designed for resilient, distributed data management and communication, with a clear separation between full network nodes and lightweight clients.
Architecture Overview
The architecture is modular and robust, supporting both full nodes (which run core services and participate in discovery) and lightweight clients (which connect to the network via bootstrap peers).
┌─────────────────────────────────────────────────────────────┐
│ DeBros Network Cluster │
├─────────────────────────────────────────────────────────────┤
│ Application Layer │
│ ┌─────────────┐ ┌─────────────┐ ┌───────────────────────┐ │
│ │ Anchat │ │ Custom App │ │ CLI Tools │ │
│ └─────────────┘ └─────────────┘ └───────────────────────┘ │
├─────────────────────────────────────────────────────────────┤
│ Client API │
│ ┌─────────────┐ ┌─────────────┐ ┌───────────────────────┐ │
│ │ Database │ │ Storage │ │ PubSub │ │
│ │ Client │ │ Client │ │ Client │ │
│ └─────────────┘ └─────────────┘ └───────────────────────┘ │
├─────────────────────────────────────────────────────────────┤
│ Network Layer │
│ ┌─────────────┐ ┌─────────────┐ ┌───────────────────────┐ │
│ │ Node │ │ Discovery │ │ PubSub │ │
│ │ (Full P2P) │ │ Manager │ │ Manager │ │
│ └─────────────┘ └─────────────┘ └───────────────────────┘ │
├─────────────────────────────────────────────────────────────┤
│ Database Layer (RQLite) │
│ ┌─────────────┐ │
│ │ RQLite │ │
│ │ Consensus │ │
│ └─────────────┘ │
└─────────────────────────────────────────────────────────────┘
Key Principles:
- Modularity: Each component is independently testable and replaceable.
- Fault Tolerance: Network continues operating with node failures.
- Security: End-to-end encryption, peer authentication, and namespace isolation.
- Performance: Optimized for common operations, with connection pooling and caching.
Codebase Structure
network/
├── cmd/ # Executables
│ ├── node/ # Network node (full participant)
│ │ └── main.go # Node entrypoint
│ └── cli/ # Command-line interface
│ └── main.go # CLI entrypoint
├── pkg/ # Core packages
│ ├── client/ # Lightweight client API
│ ├── node/ # Full node implementation
│ ├── config/ # Centralized configuration management
│ ├── database/ # RQLite integration
│ ├── storage/ # Distributed key-value storage
│ ├── pubsub/ # Pub/Sub messaging
│ ├── discovery/ # Peer discovery (node only)
│ ├── logging/ # Structured and colored logging
│ └── anyoneproxy/ # Optional SOCKS5 proxy support
├── configs/ # YAML configuration files
│ ├── node.yaml # Node config
│ └── bootstrap.yaml # Bootstrap config (legacy, now unified)
├── scripts/ # Install and utility scripts
└── data/ # Runtime data (identity, db, logs)
Key Components
1. Network Client (pkg/client/
)
- Role: Lightweight P2P participant for apps and CLI.
- Features: Connects only to bootstrap peers, no peer discovery, provides Database, Storage, PubSub, and NetworkInfo interfaces.
- Isolation: Namespaced per application.
2. Node (pkg/node/
)
- Role: Full P2P participant, runs core services (RQLite, storage, pubsub), handles peer discovery and network management.
- Features: Peer discovery, service registration, connection monitoring, and data replication.
3. Configuration (pkg/config/
)
- Centralized: All config is managed via YAML files, with CLI flags and environment variables overriding as needed.
- Unified: Node and client configs share structure; bootstrap is just a node with no join address.
4. Database Layer (pkg/database/
)
- RQLite: Distributed SQLite with Raft consensus, automatic leader election, and failover.
- Client API: SQL queries, transactions, schema management.
5. Storage System (pkg/storage/
)
- Distributed KV: Namespace-isolated, CRUD operations, prefix queries, replication.
6. Pub/Sub System (pkg/pubsub/
)
- Messaging: Topic-based, real-time delivery, automatic subscription management, namespace isolation.
7. Discovery (pkg/discovery/
)
- Node Only: Handles peer discovery via peerstore and peer exchange. No DHT/Kademlia in client.
Configuration System
- Primary Source: YAML files (
configs/node.yaml
) - Overrides: CLI flags > Environment variables > YAML > Code defaults
- Examples:
data_dir
,key_file
,listen_addresses
,solana_wallet
rqlite_port
,rqlite_raft_port
,rqlite_join_address
bootstrap_peers
,discovery_interval
- Logging:
level
,file
Client Configuration Precedence:
- Explicit in
ClientConfig
- Environment variables (
RQLITE_NODES
,BOOTSTRAP_PEERS
) - Library defaults (from config package)
Node vs Client Roles
Node (pkg/node/
)
- Runs full network services (RQLite, storage, pubsub)
- Handles peer discovery and network topology
- Participates in consensus and replication
- Manages service lifecycle and monitoring
Client (pkg/client/
)
- Lightweight participant (does not run services)
- Connects only to known bootstrap peers
- No peer discovery or DHT
- Consumes network services via API (Database, Storage, PubSub, NetworkInfo)
- Used by CLI and application integrations
Network Protocol & Data Flow
Connection Establishment
- Node: Connects to bootstrap peers, discovers additional peers, registers services.
- Client: Connects only to bootstrap peers.
Message Types
- Control: Node status, heartbeats, topology updates
- Database: SQL queries, transactions, schema ops
- Storage: KV operations, replication
- PubSub: Topic subscriptions, published messages
Security Model
- Transport: Noise/TLS encryption for all connections
- Authentication: Peer identity verification
- Isolation: Namespace-based access control
Data Flow
- Database: Client → DatabaseClient → RQLite Leader → Raft Consensus → All Nodes
- Storage: Client → StorageClient → Node → Replication
- PubSub: Client → PubSubClient → Node → Topic Router → Subscribers
Gateway Service
The Gateway provides an HTTP(S)/WebSocket surface over the network client with strict namespace enforcement.
- Run:
make run-gateway
# Env overrides: GATEWAY_ADDR, GATEWAY_NAMESPACE, GATEWAY_BOOTSTRAP_PEERS,
# GATEWAY_REQUIRE_AUTH, GATEWAY_API_KEYS
-
Auth: When
RequireAuth
is enabled, endpoints require either:- JWT (issued by this gateway; JWKS:
GET /v1/auth/jwks
or/.well-known/jwks.json
) - API Key (via
Authorization: Bearer <key>
orX-API-Key
), optionally mapped to a namespace - Wallet verification uses Ethereum EIP-191
personal_sign
:POST /v1/auth/challenge
returns{nonce}
. Clients must sign the exact nonce string.POST /v1/auth/verify
expects{wallet, nonce, signature}
with 65-byte r||s||v hex (0x allowed).v
normalized (27/28 or 0/1). Address match is case-insensitive. Nonce is marked used only after successful verification.
- JWT (issued by this gateway; JWKS:
-
Namespace Enforcement: Storage and PubSub are internally prefixed
ns::<namespace>::...
. Ownership of namespace is enforced by middleware for routes under/v1/storage*
,/v1/apps*
, and/v1/pubsub*
.
Endpoints
-
Health/Version
GET /health
,GET /v1/health
GET /v1/status
GET /v1/version
→{version, commit, build_time, started_at, uptime}
-
JWKS
GET /v1/auth/jwks
GET /.well-known/jwks.json
-
Auth
POST /v1/auth/challenge
POST /v1/auth/verify
POST /v1/auth/register
POST /v1/auth/refresh
POST /v1/auth/logout
GET /v1/auth/whoami
-
Storage
POST /v1/storage/get
,POST /v1/storage/put
,POST /v1/storage/delete
GET /v1/storage/list?prefix=...
,GET /v1/storage/exists?key=...
-
Network
GET /v1/network/status
,GET /v1/network/peers
POST /v1/network/connect
,POST /v1/network/disconnect
PubSub
-
WebSocket
GET /v1/pubsub/ws?topic=<topic>
- Server sends messages as binary frames; 30s ping keepalive.
- Client text/binary frames are published to the same namespaced topic.
-
REST
POST /v1/pubsub/publish
→ body{topic, data_base64}
→{status:"ok"}
GET /v1/pubsub/topics
→{topics:["<topic>", ...]}
(names trimmed to caller namespace)
Security note: CORS and WS origin checks are permissive for development; harden for production.
Build & Development
Prerequisites
- Go 1.21+
- RQLite
- Git
- Make
Build Commands
make build # Build all executables
make test # Run tests
make run-node # Start node (auto-detects bootstrap vs regular)
make run-gateway # Start HTTP gateway (env overrides supported)
Development Workflow
- Use
make run-node
for local development. - Edit YAML configs for node settings.
- Use CLI for network operations and testing.
API Reference
Client Creation
import "git.debros.io/DeBros/network/pkg/client"
config := client.DefaultClientConfig("my-app")
config.BootstrapPeers = []string{"/ip4/127.0.0.1/tcp/4001/p2p/{PEER_ID}"}
client, err := client.NewClient(config)
err = client.Connect()
defer client.Disconnect()
Database Operations
result, err := client.Database().Query(ctx, "SELECT * FROM users")
err := client.Database().CreateTable(ctx, "CREATE TABLE ...")
Storage Operations
err := client.Storage().Put(ctx, "key", []byte("value"))
data, err := client.Storage().Get(ctx, "key")
PubSub Operations
err := client.PubSub().Subscribe(ctx, "topic", handler)
err := client.PubSub().Publish(ctx, "topic", []byte("msg"))
Network Information
status, err := client.Network().GetStatus(ctx)
peers, err := client.Network().GetPeers(ctx)
Troubleshooting
Common Issues
- Bootstrap Connection Failed: Check peer ID, port, firewall, and node status.
- Database Timeout: Ensure RQLite ports are open, leader election is complete, and join address is correct.
- Message Delivery Failures: Verify topic names, subscription status, and network connectivity.
- High Memory Usage: Unsubscribe from topics when done, monitor connection pool size.
Debugging
- Enable debug logging:
export LOG_LEVEL=debug
- Check service logs:
sudo journalctl -u debros-node.service -f
- Use CLI for health and peer checks:
./bin/network-cli health
,./bin/network-cli peers
Example Application: Anchat
The anchat/
directory contains a full-featured decentralized chat app built on DeBros Network. Features include:
- Solana wallet integration
- End-to-end encrypted messaging
- Real-time pub/sub chat rooms
- Persistent history
This document provides a modern, accurate context for understanding the DeBros Network Cluster architecture, configuration, and usage patterns. All details reflect the current codebase and best practices.