forked from iarv/meshcore-stats
Compare commits
14 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
64cc352b80 | ||
|
|
e37aef6c5e | ||
|
|
81b7c6897a | ||
|
|
a3015e2209 | ||
|
|
5545ce5b28 | ||
|
|
666ed4215f | ||
|
|
3d0d90304c | ||
|
|
6afc14e007 | ||
|
|
4c5a408604 | ||
|
|
3c5eace220 | ||
|
|
7eee23ec40 | ||
|
|
30de7c20f3 | ||
|
|
19fa04c202 | ||
|
|
6ac52629d3 |
@@ -1,3 +1,3 @@
|
||||
{
|
||||
".": "0.2.3"
|
||||
".": "0.2.8"
|
||||
}
|
||||
|
||||
40
CHANGELOG.md
40
CHANGELOG.md
@@ -4,6 +4,46 @@ All notable changes to this project will be documented in this file.
|
||||
|
||||
This changelog is automatically generated by [release-please](https://github.com/googleapis/release-please) based on [Conventional Commits](https://www.conventionalcommits.org/).
|
||||
|
||||
## [0.2.8](https://github.com/jorijn/meshcore-stats/compare/v0.2.7...v0.2.8) (2026-01-06)
|
||||
|
||||
|
||||
### Bug Fixes
|
||||
|
||||
* normalize reporting outputs and chart tooltips ([e37aef6](https://github.com/jorijn/meshcore-stats/commit/e37aef6c5e55d2077baf4ee35abdff0562983d69))
|
||||
|
||||
## [0.2.7](https://github.com/jorijn/meshcore-stats/compare/v0.2.6...v0.2.7) (2026-01-06)
|
||||
|
||||
|
||||
### Features
|
||||
|
||||
* add telemetry collection for companion and repeater nodes ([#24](https://github.com/jorijn/meshcore-stats/issues/24)) ([a3015e2](https://github.com/jorijn/meshcore-stats/commit/a3015e2209781bdd7c317fa992ced6afa19efe61))
|
||||
|
||||
## [0.2.6](https://github.com/jorijn/meshcore-stats/compare/v0.2.5...v0.2.6) (2026-01-05)
|
||||
|
||||
|
||||
### Bug Fixes
|
||||
|
||||
* add tmpfs mount for fontconfig cache to fix read-only filesystem errors ([3d0d903](https://github.com/jorijn/meshcore-stats/commit/3d0d90304cec5ebcdb34935400de31afd62e258d))
|
||||
|
||||
## [0.2.5](https://github.com/jorijn/meshcore-stats/compare/v0.2.4...v0.2.5) (2026-01-05)
|
||||
|
||||
|
||||
### Features
|
||||
|
||||
* add automatic serial port locking to prevent concurrent access ([3c5eace](https://github.com/jorijn/meshcore-stats/commit/3c5eace2207279c55401dd8fa27294d5a94bb682))
|
||||
|
||||
|
||||
### Documentation
|
||||
|
||||
* fix formatting in architecture diagram ([7eee23e](https://github.com/jorijn/meshcore-stats/commit/7eee23ec40ff9441515b4ac18fbb7cd3f87fa4b5))
|
||||
|
||||
## [0.2.4](https://github.com/jorijn/meshcore-stats/compare/v0.2.3...v0.2.4) (2026-01-05)
|
||||
|
||||
|
||||
### Documentation
|
||||
|
||||
* rewrite README with Docker-first installation guide ([6ac5262](https://github.com/jorijn/meshcore-stats/commit/6ac52629d3025db69f9334d3185b97ce16cd3e4b))
|
||||
|
||||
## [0.2.3](https://github.com/jorijn/meshcore-stats/compare/v0.2.2...v0.2.3) (2026-01-05)
|
||||
|
||||
|
||||
|
||||
24
CLAUDE.md
24
CLAUDE.md
@@ -16,6 +16,8 @@ Always edit the source templates, then regenerate with `python scripts/render_si
|
||||
|
||||
## Running Commands
|
||||
|
||||
**IMPORTANT: Always activate the virtual environment before running any Python commands.**
|
||||
|
||||
```bash
|
||||
cd /path/to/meshcore-stats
|
||||
source .venv/bin/activate
|
||||
@@ -354,11 +356,17 @@ All configuration via `meshcore.conf` or environment variables. The config file
|
||||
|
||||
### Timeouts & Retry
|
||||
- `REMOTE_TIMEOUT_S`: Minimum timeout for LoRa requests (default: 10)
|
||||
- `REMOTE_RETRY_ATTEMPTS`: Number of retry attempts (default: 5)
|
||||
- `REMOTE_RETRY_ATTEMPTS`: Number of retry attempts (default: 2)
|
||||
- `REMOTE_RETRY_BACKOFF_S`: Seconds between retries (default: 4)
|
||||
- `REMOTE_CB_FAILS`: Failures before circuit breaker opens (default: 6)
|
||||
- `REMOTE_CB_COOLDOWN_S`: Circuit breaker cooldown (default: 3600)
|
||||
|
||||
### Telemetry Collection
|
||||
- `TELEMETRY_ENABLED`: Enable environmental telemetry collection from repeater (0/1, default: 0)
|
||||
- `TELEMETRY_TIMEOUT_S`: Timeout for telemetry requests (default: 10)
|
||||
- `TELEMETRY_RETRY_ATTEMPTS`: Retry attempts for telemetry (default: 2)
|
||||
- `TELEMETRY_RETRY_BACKOFF_S`: Backoff between telemetry retries (default: 4)
|
||||
|
||||
### Intervals
|
||||
- `COMPANION_STEP`: Collection interval for companion (default: 60s)
|
||||
- `REPEATER_STEP`: Collection interval for repeater (default: 900s / 15min)
|
||||
@@ -410,6 +418,12 @@ Metrics are classified as either **gauge** or **counter** in `src/meshmon/metric
|
||||
|
||||
Counter metrics are converted to rates during chart rendering by calculating deltas between consecutive readings.
|
||||
|
||||
- **TELEMETRY**: Environmental sensor data (when `TELEMETRY_ENABLED=1`):
|
||||
- Stored with `telemetry.` prefix: `telemetry.temperature.0`, `telemetry.humidity.0`, `telemetry.barometer.0`
|
||||
- Channel number distinguishes multiple sensors of the same type
|
||||
- Compound values (e.g., GPS) stored as: `telemetry.gps.0.latitude`, `telemetry.gps.0.longitude`
|
||||
- Telemetry collection does NOT affect circuit breaker state
|
||||
|
||||
## Database Schema
|
||||
|
||||
Metrics are stored in a SQLite database at `data/state/metrics.db` with WAL mode enabled for concurrent access.
|
||||
@@ -694,16 +708,14 @@ meshcore-cli -s /dev/ttyACM0 reset_path "repeater name"
|
||||
|
||||
## Cron Setup (Example)
|
||||
|
||||
Use `flock` to prevent USB serial conflicts when companion and repeater collection overlap.
|
||||
|
||||
```cron
|
||||
MESHCORE=/path/to/meshcore-stats
|
||||
|
||||
# Companion: every minute
|
||||
* * * * * cd $MESHCORE && flock -w 60 /tmp/meshcore.lock .venv/bin/python scripts/collect_companion.py
|
||||
* * * * * cd $MESHCORE && .venv/bin/python scripts/collect_companion.py
|
||||
|
||||
# Repeater: every 15 minutes (offset by 1 min for staggering)
|
||||
1,16,31,46 * * * * cd $MESHCORE && flock -w 60 /tmp/meshcore.lock .venv/bin/python scripts/collect_repeater.py
|
||||
1,16,31,46 * * * * cd $MESHCORE && .venv/bin/python scripts/collect_repeater.py
|
||||
|
||||
# Charts: every 5 minutes (generates SVG charts from database)
|
||||
*/5 * * * * cd $MESHCORE && .venv/bin/python scripts/render_charts.py
|
||||
@@ -717,7 +729,7 @@ MESHCORE=/path/to/meshcore-stats
|
||||
|
||||
**Notes:**
|
||||
- `cd $MESHCORE` is required because paths in the config are relative to the project root
|
||||
- `flock -w 60` waits up to 60 seconds for the lock, preventing USB serial conflicts
|
||||
- Serial port locking is handled automatically via `fcntl.flock()` in Python (no external `flock` needed)
|
||||
|
||||
## Adding New Metrics
|
||||
|
||||
|
||||
697
README.md
697
README.md
@@ -1,6 +1,6 @@
|
||||
# MeshCore Stats
|
||||
|
||||
A Python-based monitoring system for a MeshCore repeater node and its companion. Collects metrics from both devices, stores them in a SQLite database, and generates a static website with interactive SVG charts and statistics.
|
||||
A monitoring system for MeshCore LoRa mesh networks. Collects metrics from companion and repeater nodes, stores them in SQLite, and generates a static dashboard with interactive charts.
|
||||
|
||||
**Live demo:** [meshcore.jorijn.com](https://meshcore.jorijn.com)
|
||||
|
||||
@@ -9,500 +9,359 @@ A Python-based monitoring system for a MeshCore repeater node and its companion.
|
||||
<img src="docs/screenshot-2.png" width="49%" alt="MeshCore Stats Reports">
|
||||
</p>
|
||||
|
||||
## Features
|
||||
## Quick Start
|
||||
|
||||
- **Data Collection** - Collect metrics from companion (local) and repeater (remote) nodes
|
||||
- **Chart Rendering** - Generate interactive SVG charts from the database using matplotlib
|
||||
- **Static Site** - Generate a static HTML website with day/week/month/year views
|
||||
- **Reports** - Generate monthly and yearly statistics reports
|
||||
|
||||
## Requirements
|
||||
|
||||
### Python Dependencies
|
||||
|
||||
- Python 3.10+
|
||||
- meshcore >= 2.2.3
|
||||
- pyserial >= 3.5
|
||||
- jinja2 >= 3.1.0
|
||||
- matplotlib >= 3.8.0
|
||||
|
||||
### System Dependencies
|
||||
|
||||
- sqlite3 (for database maintenance script)
|
||||
|
||||
## Setup
|
||||
|
||||
### 1. Create Virtual Environment
|
||||
> **Linux only** - macOS and Windows users see [Platform Notes](#platform-notes) first.
|
||||
|
||||
```bash
|
||||
cd /path/to/meshcore-stats
|
||||
python3 -m venv .venv
|
||||
source .venv/bin/activate
|
||||
pip install -r requirements.txt
|
||||
```
|
||||
|
||||
### 2. Configure
|
||||
|
||||
Copy the example configuration file and customize it:
|
||||
|
||||
```bash
|
||||
cp meshcore.conf.example meshcore.conf
|
||||
# Edit meshcore.conf with your settings
|
||||
```
|
||||
|
||||
The configuration file is automatically loaded by the scripts. Key settings to configure:
|
||||
|
||||
- **Connection**: `MESH_SERIAL_PORT`, `MESH_TRANSPORT`
|
||||
- **Repeater Identity**: `REPEATER_NAME`, `REPEATER_PASSWORD`
|
||||
- **Display Names**: `REPEATER_DISPLAY_NAME`, `COMPANION_DISPLAY_NAME`
|
||||
- **Location**: `REPORT_LOCATION_NAME`, `REPORT_LAT`, `REPORT_LON`, `REPORT_ELEV`
|
||||
- **Hardware Info**: `REPEATER_HARDWARE`, `COMPANION_HARDWARE`
|
||||
- **Radio Config**: `RADIO_FREQUENCY`, `RADIO_BANDWIDTH`, etc. (includes presets for different regions)
|
||||
|
||||
See `meshcore.conf.example` for all available options with documentation.
|
||||
|
||||
## Usage
|
||||
|
||||
### Manual Execution
|
||||
|
||||
```bash
|
||||
cd /path/to/meshcore-stats
|
||||
source .venv/bin/activate
|
||||
|
||||
# Collect companion data
|
||||
python scripts/collect_companion.py
|
||||
|
||||
# Collect repeater data
|
||||
python scripts/collect_repeater.py
|
||||
|
||||
# Generate static site (includes chart rendering)
|
||||
python scripts/render_site.py
|
||||
|
||||
# Generate reports
|
||||
python scripts/render_reports.py
|
||||
```
|
||||
|
||||
The configuration is automatically loaded from `meshcore.conf`.
|
||||
|
||||
### Cron Setup
|
||||
|
||||
Add these entries to your crontab (`crontab -e`):
|
||||
|
||||
```cron
|
||||
# MeshCore Stats - adjust path as needed
|
||||
MESHCORE=/home/user/meshcore-stats
|
||||
|
||||
# Every minute: collect companion data
|
||||
* * * * * cd $MESHCORE && flock -w 60 /tmp/meshcore.lock .venv/bin/python scripts/collect_companion.py
|
||||
|
||||
# Every 15 minutes: collect repeater data
|
||||
1,16,31,46 * * * * cd $MESHCORE && flock -w 60 /tmp/meshcore.lock .venv/bin/python scripts/collect_repeater.py
|
||||
|
||||
# Every 5 minutes: render site
|
||||
*/5 * * * * cd $MESHCORE && .venv/bin/python scripts/render_site.py
|
||||
|
||||
# Daily at midnight: generate reports
|
||||
0 0 * * * cd $MESHCORE && .venv/bin/python scripts/render_reports.py
|
||||
|
||||
# Monthly at 3 AM on the 1st: database maintenance
|
||||
0 3 1 * * $MESHCORE/scripts/db_maintenance.sh
|
||||
```
|
||||
|
||||
**Notes:**
|
||||
- `cd $MESHCORE` is required because paths in the config are relative to the project root
|
||||
- `flock` prevents USB serial conflicts when companion and repeater collection overlap
|
||||
|
||||
### Docker Installation
|
||||
|
||||
The recommended way to run MeshCore Stats is with Docker Compose. This provides automatic scheduling of all collection and rendering tasks.
|
||||
|
||||
#### Quick Start
|
||||
|
||||
```bash
|
||||
# Clone the repository
|
||||
# Clone and configure
|
||||
git clone https://github.com/jorijn/meshcore-stats.git
|
||||
cd meshcore-stats
|
||||
|
||||
# Create configuration
|
||||
cp meshcore.conf.example meshcore.conf
|
||||
# Edit meshcore.conf with your settings
|
||||
# Edit meshcore.conf with your repeater name and password
|
||||
|
||||
# Create data directories with correct ownership for container (UID 1000)
|
||||
# Create data directories (container runs as UID 1000)
|
||||
mkdir -p data/state out
|
||||
sudo chown -R 1000:1000 data out
|
||||
# Alternative: chmod -R 777 data out (less secure, use chown if possible)
|
||||
|
||||
# Start the containers
|
||||
docker compose up -d
|
||||
|
||||
# View logs
|
||||
docker compose logs -f
|
||||
```
|
||||
|
||||
The web interface will be available at `http://localhost:8080`.
|
||||
|
||||
#### Architecture
|
||||
|
||||
The Docker setup uses two containers:
|
||||
|
||||
| Container | Purpose |
|
||||
|-----------|---------|
|
||||
| `meshcore-stats` | Runs Ofelia scheduler for data collection and rendering |
|
||||
| `nginx` | Serves the static website |
|
||||
|
||||
#### Configuration
|
||||
|
||||
Configuration is loaded from `meshcore.conf` via the `env_file` directive. Key settings:
|
||||
|
||||
```bash
|
||||
# Required: Serial device for companion node
|
||||
MESH_SERIAL_PORT=/dev/ttyUSB0 # Adjust for your system
|
||||
|
||||
# Required: Repeater identity
|
||||
REPEATER_NAME="Your Repeater Name"
|
||||
REPEATER_PASSWORD="your-password"
|
||||
|
||||
# Display names (shown in UI)
|
||||
REPEATER_DISPLAY_NAME="My Repeater"
|
||||
COMPANION_DISPLAY_NAME="My Companion"
|
||||
```
|
||||
|
||||
See `meshcore.conf.example` for all available options.
|
||||
|
||||
#### Serial Device Access
|
||||
|
||||
For serial transport, the container needs access to your USB serial device. Create a `docker-compose.override.yml` file (gitignored) to specify your device:
|
||||
|
||||
```yaml
|
||||
# docker-compose.override.yml - Local device configuration (not tracked in git)
|
||||
# Add your serial device
|
||||
cat > docker-compose.override.yml << 'EOF'
|
||||
services:
|
||||
meshcore-stats:
|
||||
devices:
|
||||
- /dev/ttyUSB0:/dev/ttyUSB0:rw # Linux example
|
||||
# - /dev/ttyACM0:/dev/ttyACM0:rw # Alternative Linux device
|
||||
- /dev/ttyACM0:/dev/ttyACM0
|
||||
EOF
|
||||
|
||||
# Start
|
||||
docker compose up -d
|
||||
|
||||
# Verify it's working. The various collection and render jobs will trigger after a few minutes.
|
||||
docker compose ps
|
||||
docker compose logs meshcore-stats | head -20
|
||||
|
||||
# View dashboard at http://localhost:8080
|
||||
```
|
||||
|
||||
This file is automatically merged with `docker-compose.yml` when running `docker compose up`.
|
||||
## Features
|
||||
|
||||
> **Note**: TCP transport users (e.g., macOS with socat) don't need a devices section - just configure `MESH_TRANSPORT=tcp` in your `meshcore.conf`.
|
||||
- **Data Collection** - Metrics from local companion and remote repeater nodes
|
||||
- **Interactive Charts** - SVG charts with day/week/month/year views and tooltips
|
||||
- **Statistics Reports** - Monthly and yearly report generation
|
||||
- **Light/Dark Theme** - Automatic theme switching based on system preference
|
||||
|
||||
On the host, ensure the device is accessible:
|
||||
## Prerequisites
|
||||
|
||||
- Docker and Docker Compose V2
|
||||
- MeshCore companion node connected via USB serial
|
||||
- Remote repeater node reachable via LoRa from the companion
|
||||
|
||||
**Resource requirements:** ~100MB memory, ~100MB disk per year of data.
|
||||
|
||||
## Installation
|
||||
|
||||
### Docker (Recommended)
|
||||
|
||||
#### 1. Clone the Repository
|
||||
|
||||
```bash
|
||||
# Add user to dialout group (Linux)
|
||||
sudo usermod -a -G dialout $USER
|
||||
git clone https://github.com/jorijn/meshcore-stats.git
|
||||
cd meshcore-stats
|
||||
```
|
||||
|
||||
#### 2. Configure
|
||||
|
||||
Copy the example configuration and edit it:
|
||||
|
||||
```bash
|
||||
cp meshcore.conf.example meshcore.conf
|
||||
```
|
||||
|
||||
**Minimal required settings:**
|
||||
|
||||
```ini
|
||||
# Repeater identity (required)
|
||||
REPEATER_NAME=Your Repeater Name
|
||||
REPEATER_PASSWORD=your-admin-password
|
||||
|
||||
# Display names
|
||||
REPEATER_DISPLAY_NAME=My Repeater
|
||||
COMPANION_DISPLAY_NAME=My Companion
|
||||
```
|
||||
|
||||
See [meshcore.conf.example](meshcore.conf.example) for all available options.
|
||||
|
||||
#### 3. Create Data Directories
|
||||
|
||||
```bash
|
||||
mkdir -p data/state out
|
||||
sudo chown -R 1000:1000 data out
|
||||
```
|
||||
|
||||
The container runs as UID 1000, so directories must be writable by this user. If `sudo` is not available, you can relaxed the permissions using `chmod 777 data out`, but this is less secure.
|
||||
|
||||
#### 4. Configure Serial Device
|
||||
|
||||
Create `docker-compose.override.yml` to specify your serial device:
|
||||
|
||||
```yaml
|
||||
services:
|
||||
meshcore-stats:
|
||||
devices:
|
||||
- /dev/ttyACM0:/dev/ttyACM0
|
||||
```
|
||||
|
||||
Ensure your user has serial port access:
|
||||
|
||||
```bash
|
||||
sudo usermod -aG dialout $USER
|
||||
# Log out and back in for changes to take effect
|
||||
```
|
||||
|
||||
#### Development Mode
|
||||
|
||||
For local development with live code changes:
|
||||
#### 5. Start the Containers
|
||||
|
||||
```bash
|
||||
docker compose -f docker-compose.yml -f docker-compose.dev.yml up --build
|
||||
docker compose up -d
|
||||
```
|
||||
|
||||
This mounts `src/` and `scripts/` into the container, so changes take effect immediately without rebuilding.
|
||||
After the various collection and render jobs has run, the dashboard will be available at **http://localhost:8080**.
|
||||
|
||||
#### Image Tags
|
||||
#### Verify Installation
|
||||
|
||||
Images are published to `ghcr.io/jorijn/meshcore-stats`:
|
||||
```bash
|
||||
# Check container status
|
||||
docker compose ps
|
||||
|
||||
| Tag | Description |
|
||||
|-----|-------------|
|
||||
| `X.Y.Z` | Specific version (e.g., `0.3.0`) |
|
||||
| `latest` | Latest release |
|
||||
| `nightly` | Latest release rebuilt with OS patches |
|
||||
| `nightly-YYYYMMDD` | Dated nightly build |
|
||||
# View logs
|
||||
docker compose logs -f meshcore-stats
|
||||
```
|
||||
|
||||
Version tags are rebuilt nightly to include OS security patches. For reproducible deployments, pin by SHA digest:
|
||||
### Common Docker Commands
|
||||
|
||||
```bash
|
||||
# View real-time logs
|
||||
docker compose logs -f meshcore-stats
|
||||
|
||||
# Restart after configuration changes
|
||||
docker compose restart meshcore-stats
|
||||
|
||||
# Update to latest version (database migrations are automatic)
|
||||
docker compose pull && docker compose up -d
|
||||
|
||||
# Stop all containers
|
||||
docker compose down
|
||||
|
||||
# Backup database
|
||||
cp data/state/metrics.db data/state/metrics.db.backup
|
||||
```
|
||||
|
||||
> **Note**: `docker compose down` preserves your data. Use `docker compose down -v` only if you want to delete everything.
|
||||
|
||||
### Manual Installation (Alternative)
|
||||
|
||||
For environments where Docker is not available.
|
||||
|
||||
#### Requirements
|
||||
|
||||
- Python 3.10+
|
||||
- SQLite3
|
||||
|
||||
#### Setup
|
||||
|
||||
```bash
|
||||
cd meshcore-stats
|
||||
python3 -m venv .venv
|
||||
source .venv/bin/activate
|
||||
pip install -r requirements.txt
|
||||
cp meshcore.conf.example meshcore.conf
|
||||
# Edit meshcore.conf with your settings
|
||||
```
|
||||
|
||||
#### Cron Setup
|
||||
|
||||
Add to your crontab (`crontab -e`):
|
||||
|
||||
```cron
|
||||
MESHCORE=/path/to/meshcore-stats
|
||||
|
||||
# Companion: every minute
|
||||
* * * * * cd $MESHCORE && .venv/bin/python scripts/collect_companion.py
|
||||
|
||||
# Repeater: every 15 minutes
|
||||
1,16,31,46 * * * * cd $MESHCORE && .venv/bin/python scripts/collect_repeater.py
|
||||
|
||||
# Charts: every 5 minutes
|
||||
*/5 * * * * cd $MESHCORE && .venv/bin/python scripts/render_charts.py
|
||||
|
||||
# Site: every 5 minutes
|
||||
*/5 * * * * cd $MESHCORE && .venv/bin/python scripts/render_site.py
|
||||
|
||||
# Reports: daily at midnight
|
||||
0 0 * * * cd $MESHCORE && .venv/bin/python scripts/render_reports.py
|
||||
```
|
||||
|
||||
Serve the `out/` directory with any web server.
|
||||
|
||||
## Platform Notes
|
||||
|
||||
<details>
|
||||
<summary><strong>Linux</strong></summary>
|
||||
|
||||
Docker can access USB serial devices directly. Add your device to `docker-compose.override.yml`:
|
||||
|
||||
```yaml
|
||||
image: ghcr.io/jorijn/meshcore-stats@sha256:abc123...
|
||||
services:
|
||||
meshcore-stats:
|
||||
devices:
|
||||
- /dev/ttyACM0:/dev/ttyACM0
|
||||
```
|
||||
|
||||
#### Volumes
|
||||
Common device paths:
|
||||
- `/dev/ttyACM0` - Arduino/native USB
|
||||
- `/dev/ttyUSB0` - USB-to-serial adapters
|
||||
|
||||
| Path | Purpose |
|
||||
|------|---------|
|
||||
| `./data/state` | SQLite database and circuit breaker state |
|
||||
| `./out` | Generated static site (served by nginx) |
|
||||
</details>
|
||||
|
||||
Both directories must be writable by UID 1000 (the container user). See Quick Start for setup.
|
||||
<details>
|
||||
<summary><strong>macOS</strong></summary>
|
||||
|
||||
#### Resource Limits
|
||||
Docker Desktop for macOS runs in a Linux VM and **cannot directly access USB serial devices**.
|
||||
|
||||
Default resource limits in `docker-compose.yml`:
|
||||
**Option 1: TCP Bridge (Recommended)**
|
||||
|
||||
| Container | CPU | Memory |
|
||||
|-----------|-----|--------|
|
||||
| meshcore-stats | 1.0 | 512MB |
|
||||
| nginx | 0.5 | 64MB |
|
||||
|
||||
Adjust in `docker-compose.yml` if needed.
|
||||
|
||||
#### Important Notes
|
||||
|
||||
- **Single instance only**: SQLite uses WAL mode which requires exclusive access. Do not run multiple container instances.
|
||||
- **Persistent storage**: Mount `./data/state` to preserve your database across container restarts.
|
||||
- **Health checks**: Both containers have health checks. Use `docker compose ps` to verify status.
|
||||
|
||||
Environment variables always take precedence over `meshcore.conf`.
|
||||
|
||||
### Serving the Site
|
||||
|
||||
The static site is generated in the `out/` directory. You can serve it with any web server:
|
||||
Expose the serial port over TCP using socat:
|
||||
|
||||
```bash
|
||||
# Simple Python server for testing
|
||||
cd out && python3 -m http.server 8080
|
||||
# Install socat
|
||||
brew install socat
|
||||
|
||||
# Or configure nginx/caddy to serve the out/ directory
|
||||
# Bridge serial to TCP (run in background)
|
||||
socat TCP-LISTEN:5000,fork,reuseaddr OPEN:/dev/cu.usbserial-0001,rawer,nonblock,ispeed=115200,ospeed=115200
|
||||
```
|
||||
|
||||
## Project Structure
|
||||
Configure in `meshcore.conf`:
|
||||
|
||||
```
|
||||
meshcore-stats/
|
||||
├── requirements.txt
|
||||
├── README.md
|
||||
├── meshcore.conf.example # Example configuration
|
||||
├── meshcore.conf # Your configuration (create this)
|
||||
├── src/meshmon/
|
||||
│ ├── __init__.py
|
||||
│ ├── env.py # Environment variable parsing
|
||||
│ ├── log.py # Logging helper
|
||||
│ ├── meshcore_client.py # MeshCore connection and commands
|
||||
│ ├── db.py # SQLite database module
|
||||
│ ├── retry.py # Retry logic and circuit breaker
|
||||
│ ├── charts.py # Matplotlib SVG chart generation
|
||||
│ ├── html.py # HTML rendering
|
||||
│ ├── reports.py # Report generation
|
||||
│ ├── metrics.py # Metric type definitions
|
||||
│ ├── battery.py # Battery voltage to percentage conversion
|
||||
│ ├── migrations/ # SQL schema migrations
|
||||
│ │ ├── 001_initial_schema.sql
|
||||
│ │ └── 002_eav_schema.sql
|
||||
│ └── templates/ # Jinja2 HTML templates
|
||||
├── scripts/
|
||||
│ ├── collect_companion.py # Collect metrics from companion node
|
||||
│ ├── collect_repeater.py # Collect metrics from repeater node
|
||||
│ ├── render_charts.py # Generate SVG charts from database
|
||||
│ ├── render_site.py # Generate static HTML site
|
||||
│ ├── render_reports.py # Generate monthly/yearly reports
|
||||
│ └── db_maintenance.sh # Database VACUUM/ANALYZE
|
||||
├── data/
|
||||
│ └── state/
|
||||
│ ├── metrics.db # SQLite database (WAL mode)
|
||||
│ └── repeater_circuit.json
|
||||
└── out/ # Generated site
|
||||
├── .htaccess # Apache config (DirectoryIndex, caching)
|
||||
├── styles.css # Stylesheet
|
||||
├── chart-tooltip.js # Chart tooltip enhancement
|
||||
├── day.html # Repeater pages (entry point)
|
||||
├── week.html
|
||||
├── month.html
|
||||
├── year.html
|
||||
├── companion/
|
||||
│ ├── day.html
|
||||
│ ├── week.html
|
||||
│ ├── month.html
|
||||
│ └── year.html
|
||||
└── reports/
|
||||
├── index.html
|
||||
├── repeater/ # YYYY/MM reports
|
||||
└── companion/
|
||||
```ini
|
||||
MESH_TRANSPORT=tcp
|
||||
MESH_TCP_HOST=host.docker.internal
|
||||
MESH_TCP_PORT=5000
|
||||
```
|
||||
|
||||
## Chart Features
|
||||
**Option 2: Native Installation**
|
||||
|
||||
Charts are rendered as inline SVG using matplotlib with the following features:
|
||||
Use the manual installation method with cron instead of Docker.
|
||||
|
||||
- **Theme Support**: Automatic light/dark mode via CSS `prefers-color-scheme`
|
||||
- **Interactive Tooltips**: Hover to see exact values and timestamps
|
||||
- **Data Point Indicator**: Visual marker shows position on the chart line
|
||||
- **Mobile Support**: Touch-friendly tooltips
|
||||
- **Statistics**: Min/Avg/Max values displayed below each chart
|
||||
- **Period Views**: Day, week, month, and year time ranges
|
||||
</details>
|
||||
|
||||
## Troubleshooting
|
||||
<details>
|
||||
<summary><strong>Windows (WSL2)</strong></summary>
|
||||
|
||||
### Serial Device Not Found
|
||||
WSL2 and Docker Desktop for Windows cannot directly access COM ports.
|
||||
|
||||
If you see "No serial ports found" or connection fails:
|
||||
Use the TCP bridge approach (similar to macOS) or native installation.
|
||||
|
||||
1. Check that your device is connected:
|
||||
```bash
|
||||
ls -la /dev/ttyUSB* /dev/ttyACM*
|
||||
```
|
||||
</details>
|
||||
|
||||
2. Check permissions (add user to dialout group):
|
||||
```bash
|
||||
sudo usermod -a -G dialout $USER
|
||||
# Log out and back in for changes to take effect
|
||||
```
|
||||
|
||||
3. Try specifying the port explicitly:
|
||||
```bash
|
||||
export MESH_SERIAL_PORT=/dev/ttyACM0
|
||||
```
|
||||
|
||||
4. Check dmesg for device detection:
|
||||
```bash
|
||||
dmesg | tail -20
|
||||
```
|
||||
|
||||
### Repeater Not Found
|
||||
|
||||
If the script cannot find the repeater contact:
|
||||
|
||||
1. The script will print all discovered contacts - check for the correct name
|
||||
2. Verify REPEATER_NAME matches exactly (case-sensitive)
|
||||
3. Try using REPEATER_KEY_PREFIX instead with the first 6-12 hex chars of the public key
|
||||
|
||||
### Circuit Breaker
|
||||
|
||||
If repeater collection shows "cooldown active":
|
||||
|
||||
1. This is normal after multiple failed remote requests
|
||||
2. Wait for the cooldown period (default 1 hour) or reset manually:
|
||||
```bash
|
||||
rm data/state/repeater_circuit.json
|
||||
```
|
||||
|
||||
### Docker on macOS: Serial Devices Not Available
|
||||
|
||||
Docker on macOS (including Docker Desktop and OrbStack) runs containers inside a Linux virtual machine. USB and serial devices connected to the Mac host cannot be passed through to this VM, so the `devices:` section in docker-compose.yml will fail with:
|
||||
|
||||
```
|
||||
error gathering device information while adding custom device "/dev/cu.usbserial-0001": no such file or directory
|
||||
```
|
||||
|
||||
**Workarounds:**
|
||||
|
||||
1. **Use TCP transport**: Run a serial-to-TCP bridge on the host and configure the container to connect via TCP:
|
||||
```bash
|
||||
# On macOS host, expose serial port over TCP (install socat via Homebrew)
|
||||
socat TCP-LISTEN:5000,fork,reuseaddr OPEN:/dev/cu.usbserial-0001,rawer,nonblock,ispeed=115200,ospeed=115200
|
||||
```
|
||||
Then configure in meshcore.conf:
|
||||
```bash
|
||||
MESH_TRANSPORT=tcp
|
||||
MESH_TCP_HOST=host.docker.internal
|
||||
MESH_TCP_PORT=5000
|
||||
```
|
||||
|
||||
2. **Run natively on macOS**: Use the cron-based setup instead of Docker (see "Cron Setup" section).
|
||||
|
||||
3. **Use a Linux host**: Docker on Linux can pass through USB devices directly.
|
||||
|
||||
Note: OrbStack has [USB passthrough on their roadmap](https://github.com/orbstack/orbstack/issues/89) but it is not yet available.
|
||||
|
||||
## Environment Variables Reference
|
||||
## Configuration Reference
|
||||
|
||||
| Variable | Default | Description |
|
||||
|----------|---------|-------------|
|
||||
| **Connection** | | |
|
||||
| `MESH_TRANSPORT` | serial | Connection type: serial, tcp, ble |
|
||||
| `MESH_SERIAL_PORT` | (auto) | Serial port path |
|
||||
| `MESH_SERIAL_BAUD` | 115200 | Baud rate |
|
||||
| `MESH_TCP_HOST` | localhost | TCP host |
|
||||
| `MESH_TCP_PORT` | 5000 | TCP port |
|
||||
| `MESH_BLE_ADDR` | - | BLE device address |
|
||||
| `MESH_BLE_PIN` | - | BLE PIN |
|
||||
| `MESH_DEBUG` | 0 | Enable debug output |
|
||||
| **Repeater Identity** | | |
|
||||
| `REPEATER_NAME` | - | Repeater advertised name |
|
||||
| `REPEATER_KEY_PREFIX` | - | Repeater public key prefix |
|
||||
| `REPEATER_PASSWORD` | - | Repeater login password |
|
||||
| **Display Names** | | |
|
||||
| `REPEATER_DISPLAY_NAME` | Repeater Node | Display name for repeater in UI |
|
||||
| `COMPANION_DISPLAY_NAME` | Companion Node | Display name for companion in UI |
|
||||
| `REPEATER_NAME` | *required* | Advertised name to find in contacts |
|
||||
| `REPEATER_PASSWORD` | *required* | Admin password for repeater |
|
||||
| `REPEATER_KEY_PREFIX` | - | Alternative to `REPEATER_NAME`: hex prefix of public key |
|
||||
| **Connection** | | |
|
||||
| `MESH_TRANSPORT` | serial | Transport type: `serial`, `tcp`, or `ble` |
|
||||
| `MESH_SERIAL_PORT` | auto | Serial port path |
|
||||
| `MESH_TCP_HOST` | localhost | TCP host (for TCP transport) |
|
||||
| `MESH_TCP_PORT` | 5000 | TCP port (for TCP transport) |
|
||||
| **Display** | | |
|
||||
| `REPEATER_DISPLAY_NAME` | Repeater Node | Name shown in UI |
|
||||
| `COMPANION_DISPLAY_NAME` | Companion Node | Name shown in UI |
|
||||
| `REPEATER_HARDWARE` | LoRa Repeater | Hardware model for sidebar |
|
||||
| `COMPANION_HARDWARE` | LoRa Node | Hardware model for sidebar |
|
||||
| **Location** | | |
|
||||
| `REPORT_LOCATION_NAME` | Your Location | Full location name for reports |
|
||||
| `REPORT_LOCATION_SHORT` | Your Location | Short location for sidebar/meta |
|
||||
| `REPORT_LAT` | 0.0 | Latitude in decimal degrees |
|
||||
| `REPORT_LON` | 0.0 | Longitude in decimal degrees |
|
||||
| `REPORT_LOCATION_NAME` | Your Location | Full location for reports |
|
||||
| `REPORT_LAT` | 0.0 | Latitude |
|
||||
| `REPORT_LON` | 0.0 | Longitude |
|
||||
| `REPORT_ELEV` | 0.0 | Elevation |
|
||||
| `REPORT_ELEV_UNIT` | m | Elevation unit: "m" or "ft" |
|
||||
| **Hardware Info** | | |
|
||||
| `REPEATER_HARDWARE` | LoRa Repeater | Repeater hardware model for sidebar |
|
||||
| `COMPANION_HARDWARE` | LoRa Node | Companion hardware model for sidebar |
|
||||
| **Radio Config** | | |
|
||||
| `RADIO_FREQUENCY` | 869.618 MHz | Radio frequency for display |
|
||||
| `RADIO_BANDWIDTH` | 62.5 kHz | Radio bandwidth for display |
|
||||
| `RADIO_SPREAD_FACTOR` | SF8 | Spread factor for display |
|
||||
| `RADIO_CODING_RATE` | CR8 | Coding rate for display |
|
||||
| **Intervals** | | |
|
||||
| `COMPANION_STEP` | 60 | Companion data collection interval (seconds) |
|
||||
| `REPEATER_STEP` | 900 | Repeater data collection interval (seconds) |
|
||||
| `REMOTE_TIMEOUT_S` | 10 | Remote request timeout |
|
||||
| `REMOTE_RETRY_ATTEMPTS` | 2 | Max retry attempts |
|
||||
| `REMOTE_RETRY_BACKOFF_S` | 4 | Retry backoff delay |
|
||||
| `REMOTE_CB_FAILS` | 6 | Failures before circuit opens |
|
||||
| `REMOTE_CB_COOLDOWN_S` | 3600 | Circuit breaker cooldown |
|
||||
| **Paths** | | |
|
||||
| `STATE_DIR` | ./data/state | State file path |
|
||||
| `OUT_DIR` | ./out | Output site path |
|
||||
| **Radio** (display only) | | |
|
||||
| `RADIO_FREQUENCY` | 869.618 MHz | Frequency shown in sidebar |
|
||||
| `RADIO_BANDWIDTH` | 62.5 kHz | Bandwidth |
|
||||
| `RADIO_SPREAD_FACTOR` | SF8 | Spread factor |
|
||||
|
||||
## Metrics Reference
|
||||
See [meshcore.conf.example](meshcore.conf.example) for all options with regional radio presets.
|
||||
|
||||
The system uses an EAV (Entity-Attribute-Value) schema where firmware field names are stored directly in the database. This allows new metrics to be captured automatically without schema changes.
|
||||
## Troubleshooting
|
||||
|
||||
### Repeater Metrics
|
||||
| Symptom | Cause | Solution |
|
||||
|---------|-------|----------|
|
||||
| "Permission denied" on serial port | User not in dialout group | `sudo usermod -aG dialout $USER` then re-login |
|
||||
| Repeater shows "offline" status | No data or circuit breaker tripped | Check logs; delete `data/state/repeater_circuit.json` to reset |
|
||||
| Empty charts | Not enough data collected | Wait for 2+ collection cycles |
|
||||
| Container exits immediately | Missing or invalid configuration | Verify `meshcore.conf` exists and has required values |
|
||||
| "No serial ports found" | Device not connected/detected | Check `ls /dev/tty*` and device permissions |
|
||||
| Device path changed after reboot | USB enumeration order changed | Update path in `docker-compose.override.yml` or use udev rules |
|
||||
| "database is locked" errors | Maintenance script running | Wait for completion; check if VACUUM is running |
|
||||
|
||||
| Metric | Type | Display Unit | Description |
|
||||
|--------|------|--------------|-------------|
|
||||
| `bat` | Gauge | Voltage (V) | Battery voltage (stored in mV, displayed as V) |
|
||||
| `bat_pct` | Gauge | Battery (%) | Battery percentage (computed from voltage) |
|
||||
| `last_rssi` | Gauge | RSSI (dBm) | Signal strength of last packet |
|
||||
| `last_snr` | Gauge | SNR (dB) | Signal-to-noise ratio |
|
||||
| `noise_floor` | Gauge | dBm | Background RF noise |
|
||||
| `uptime` | Gauge | Days | Time since reboot (seconds ÷ 86400) |
|
||||
| `tx_queue_len` | Gauge | Queue depth | TX queue length |
|
||||
| `nb_recv` | Counter | Packets/min | Total packets received |
|
||||
| `nb_sent` | Counter | Packets/min | Total packets transmitted |
|
||||
| `airtime` | Counter | Seconds/min | TX airtime rate |
|
||||
| `rx_airtime` | Counter | Seconds/min | RX airtime rate |
|
||||
| `flood_dups` | Counter | Packets/min | Flood duplicate packets |
|
||||
| `direct_dups` | Counter | Packets/min | Direct duplicate packets |
|
||||
| `sent_flood` | Counter | Packets/min | Flood packets transmitted |
|
||||
| `recv_flood` | Counter | Packets/min | Flood packets received |
|
||||
| `sent_direct` | Counter | Packets/min | Direct packets transmitted |
|
||||
| `recv_direct` | Counter | Packets/min | Direct packets received |
|
||||
### Debug Logging
|
||||
|
||||
### Companion Metrics
|
||||
```bash
|
||||
# Enable debug mode in meshcore.conf
|
||||
MESH_DEBUG=1
|
||||
|
||||
| Metric | Type | Display Unit | Description |
|
||||
|--------|------|--------------|-------------|
|
||||
| `battery_mv` | Gauge | Voltage (V) | Battery voltage (stored in mV, displayed as V) |
|
||||
| `bat_pct` | Gauge | Battery (%) | Battery percentage (computed from voltage) |
|
||||
| `contacts` | Gauge | Count | Known mesh nodes |
|
||||
| `uptime_secs` | Gauge | Days | Time since reboot (seconds ÷ 86400) |
|
||||
| `recv` | Counter | Packets/min | Total packets received |
|
||||
| `sent` | Counter | Packets/min | Total packets transmitted |
|
||||
# View detailed logs
|
||||
docker compose logs -f meshcore-stats
|
||||
```
|
||||
|
||||
### Metric Types
|
||||
### Circuit Breaker
|
||||
|
||||
- **Gauge**: Instantaneous values stored as-is (battery voltage, RSSI, queue depth)
|
||||
- **Counter**: Cumulative values where the rate of change is calculated (packets, airtime). Charts display per-minute rates.
|
||||
The repeater collector uses a circuit breaker to avoid spamming LoRa when the repeater is unreachable. After multiple failures, it enters a cooldown period (default: 1 hour).
|
||||
|
||||
## Database
|
||||
To reset manually:
|
||||
|
||||
Metrics are stored in a SQLite database at `data/state/metrics.db` with WAL mode enabled for concurrent read/write access.
|
||||
```bash
|
||||
rm data/state/repeater_circuit.json
|
||||
docker compose restart meshcore-stats
|
||||
```
|
||||
|
||||
### Schema Migrations
|
||||
## Architecture
|
||||
|
||||
Database migrations are stored as SQL files in `src/meshmon/migrations/` and are applied automatically when the database is initialized. Migration files follow the naming convention `NNN_description.sql` (e.g., `001_initial_schema.sql`).
|
||||
```
|
||||
┌─────────────────┐ LoRa ┌─────────────────┐
|
||||
│ Companion │◄─────────────►│ Repeater │
|
||||
│ (USB Serial) │ │ (Remote) │
|
||||
└────────┬────────┘ └─────────────────┘
|
||||
│
|
||||
│ Serial/TCP
|
||||
▼
|
||||
┌─────────────────┐
|
||||
│ Docker Host │
|
||||
│ ┌───────────┐ │
|
||||
│ │ meshcore- │ │ ┌─────────┐
|
||||
│ │ stats │──┼────►│ nginx │──► :8080
|
||||
│ └───────────┘ │ └─────────┘
|
||||
│ │ │
|
||||
│ ▼ │
|
||||
│ SQLite + SVG │
|
||||
└─────────────────┘
|
||||
```
|
||||
|
||||
## Public Instances
|
||||
The system runs two containers:
|
||||
- **meshcore-stats**: Collects data on schedule (Ofelia) and generates charts
|
||||
- **nginx**: Serves the static dashboard
|
||||
|
||||
A list of publicly accessible MeshCore Stats installations. Want to add yours? [Open a pull request](https://github.com/jorijn/meshcore-stats/pulls)!
|
||||
## Documentation
|
||||
|
||||
| URL | Hardware | Location |
|
||||
|-----|----------|----------|
|
||||
| [meshcore.jorijn.com](https://meshcore.jorijn.com) | SenseCAP Solar Node P1 Pro + 6.5dBi Mikrotik antenna | Oosterhout, The Netherlands |
|
||||
- [docs/firmware-responses.md](docs/firmware-responses.md) - MeshCore firmware response formats
|
||||
|
||||
## License
|
||||
|
||||
MIT
|
||||
|
||||
## Public Instances
|
||||
|
||||
Public MeshCore Stats installations. Want to add yours? [Open a pull request](https://github.com/jorijn/meshcore-stats/pulls)!
|
||||
|
||||
| URL | Hardware | Location |
|
||||
|-----|----------|----------|
|
||||
| [meshcore.jorijn.com](https://meshcore.jorijn.com) | SenseCAP Solar Node P1 Pro + 6.5dBi Mikrotik antenna | Oosterhout, The Netherlands |
|
||||
|
||||
@@ -15,7 +15,7 @@ services:
|
||||
# MeshCore Stats - Data collection and rendering
|
||||
# ==========================================================================
|
||||
meshcore-stats:
|
||||
image: ghcr.io/jorijn/meshcore-stats:0.2.3 # x-release-please-version
|
||||
image: ghcr.io/jorijn/meshcore-stats:0.2.8 # x-release-please-version
|
||||
container_name: meshcore-stats
|
||||
restart: unless-stopped
|
||||
|
||||
@@ -47,6 +47,7 @@ services:
|
||||
read_only: true
|
||||
tmpfs:
|
||||
- /tmp:noexec,nosuid,size=64m
|
||||
- /var/cache/fontconfig:noexec,nosuid,size=4m
|
||||
|
||||
# Resource limits
|
||||
deploy:
|
||||
|
||||
@@ -102,6 +102,84 @@ Returns a single dict with all status fields.
|
||||
|
||||
---
|
||||
|
||||
## Telemetry Data
|
||||
|
||||
Environmental telemetry is requested via `req_telemetry_sync(contact)` and returns
|
||||
Cayenne LPP formatted sensor data. This requires `TELEMETRY_ENABLED=1` and a sensor
|
||||
board attached to the repeater.
|
||||
|
||||
### Payload Format
|
||||
|
||||
Both `req_telemetry_sync()` and `get_self_telemetry()` return a dict containing the
|
||||
LPP data list and a public key prefix:
|
||||
|
||||
```python
|
||||
{
|
||||
'pubkey_pre': 'a5c14f5244d6',
|
||||
'lpp': [
|
||||
{'channel': 0, 'type': 'temperature', 'value': 23.5},
|
||||
{'channel': 0, 'type': 'humidity', 'value': 45.2},
|
||||
]
|
||||
}
|
||||
```
|
||||
|
||||
The `extract_lpp_from_payload()` helper in `src/meshmon/telemetry.py` handles
|
||||
extracting the `lpp` list from this wrapper format.
|
||||
|
||||
### `req_telemetry_sync(contact)`
|
||||
|
||||
Returns sensor readings from a remote node in Cayenne LPP format:
|
||||
|
||||
```python
|
||||
[
|
||||
{'channel': 0, 'type': 'temperature', 'value': 23.5},
|
||||
{'channel': 0, 'type': 'humidity', 'value': 45.2},
|
||||
{'channel': 0, 'type': 'barometer', 'value': 1013.25},
|
||||
{'channel': 1, 'type': 'gps', 'value': {'latitude': 51.5, 'longitude': -0.1, 'altitude': 10}},
|
||||
]
|
||||
```
|
||||
|
||||
**Common sensor types:**
|
||||
|
||||
| Type | Unit | Description |
|
||||
|------|------|-------------|
|
||||
| `temperature` | Celsius | Temperature reading |
|
||||
| `humidity` | % | Relative humidity |
|
||||
| `barometer` | hPa/mbar | Barometric pressure |
|
||||
| `voltage` | V | Voltage reading |
|
||||
| `gps` | compound | GPS with `latitude`, `longitude`, `altitude` |
|
||||
|
||||
**Stored as:**
|
||||
- `telemetry.temperature.0` - Temperature on channel 0
|
||||
- `telemetry.humidity.0` - Humidity on channel 0
|
||||
- `telemetry.gps.1.latitude` - GPS latitude on channel 1
|
||||
|
||||
**Notes:**
|
||||
- Requires environmental sensor board (BME280, BME680, etc.) on repeater
|
||||
- Channel number distinguishes multiple sensors of the same type
|
||||
- Not all repeaters have environmental sensors attached
|
||||
- Telemetry collection does not affect circuit breaker state
|
||||
- Telemetry failures are logged as warnings and do not block status collection
|
||||
|
||||
### `get_self_telemetry()`
|
||||
|
||||
Returns self telemetry from the companion node's attached sensors.
|
||||
Same Cayenne LPP format as `req_telemetry_sync()`.
|
||||
|
||||
```python
|
||||
[
|
||||
{'channel': 0, 'type': 'temperature', 'value': 23.5},
|
||||
{'channel': 0, 'type': 'humidity', 'value': 45.2},
|
||||
]
|
||||
```
|
||||
|
||||
**Notes:**
|
||||
- Requires environmental sensor board attached to companion
|
||||
- Returns empty list if no sensors attached
|
||||
- Uses same format as repeater telemetry
|
||||
|
||||
---
|
||||
|
||||
## Derived Metrics
|
||||
|
||||
These are computed at query time, not stored:
|
||||
|
||||
@@ -113,6 +113,23 @@ RADIO_CODING_RATE=CR8
|
||||
# REMOTE_CB_FAILS=6
|
||||
# REMOTE_CB_COOLDOWN_S=3600
|
||||
|
||||
# =============================================================================
|
||||
# Telemetry Collection (Environmental Sensors)
|
||||
# =============================================================================
|
||||
# Enable telemetry collection from repeater's environmental sensors
|
||||
# (temperature, humidity, barometric pressure, etc.)
|
||||
# Requires sensor board attached to repeater (e.g., BME280, BME680)
|
||||
# Default: 0 (disabled)
|
||||
# TELEMETRY_ENABLED=1
|
||||
|
||||
# Telemetry-specific timeout and retry settings
|
||||
# Defaults match status settings. Separate config allows tuning if telemetry
|
||||
# proves problematic (e.g., firmware doesn't support it, sensor board missing).
|
||||
# You can reduce these if telemetry collection is causing issues.
|
||||
# TELEMETRY_TIMEOUT_S=10
|
||||
# TELEMETRY_RETRY_ATTEMPTS=2
|
||||
# TELEMETRY_RETRY_BACKOFF_S=4
|
||||
|
||||
# =============================================================================
|
||||
# Paths (Native installation only)
|
||||
# =============================================================================
|
||||
|
||||
@@ -25,8 +25,9 @@ sys.path.insert(0, str(Path(__file__).parent.parent / "src"))
|
||||
|
||||
from meshmon.env import get_config
|
||||
from meshmon import log
|
||||
from meshmon.meshcore_client import connect_from_env, run_command
|
||||
from meshmon.meshcore_client import connect_with_lock, run_command
|
||||
from meshmon.db import init_db, insert_metrics
|
||||
from meshmon.telemetry import extract_lpp_from_payload, extract_telemetry_metrics
|
||||
|
||||
|
||||
async def collect_companion() -> int:
|
||||
@@ -39,138 +40,132 @@ async def collect_companion() -> int:
|
||||
cfg = get_config()
|
||||
ts = int(time.time())
|
||||
|
||||
log.debug("Connecting to companion node...")
|
||||
mc = await connect_from_env()
|
||||
|
||||
if mc is None:
|
||||
log.error("Failed to connect to companion node")
|
||||
return 1
|
||||
|
||||
# Metrics to insert (firmware field names)
|
||||
metrics: dict[str, float] = {}
|
||||
commands_succeeded = 0
|
||||
|
||||
# Commands are accessed via mc.commands
|
||||
cmd = mc.commands
|
||||
log.debug("Connecting to companion node...")
|
||||
async with connect_with_lock() as mc:
|
||||
if mc is None:
|
||||
log.error("Failed to connect to companion node")
|
||||
return 1
|
||||
|
||||
try:
|
||||
# send_appstart (already called during connect, but call again to get self_info)
|
||||
ok, evt_type, payload, err = await run_command(
|
||||
mc, cmd.send_appstart(), "send_appstart"
|
||||
)
|
||||
if ok:
|
||||
commands_succeeded += 1
|
||||
log.debug(f"appstart: {evt_type}")
|
||||
else:
|
||||
log.error(f"appstart failed: {err}")
|
||||
# Commands are accessed via mc.commands
|
||||
cmd = mc.commands
|
||||
|
||||
# send_device_query
|
||||
ok, evt_type, payload, err = await run_command(
|
||||
mc, cmd.send_device_query(), "send_device_query"
|
||||
)
|
||||
if ok:
|
||||
commands_succeeded += 1
|
||||
log.debug(f"device_query: {payload}")
|
||||
else:
|
||||
log.error(f"device_query failed: {err}")
|
||||
try:
|
||||
# send_appstart (already called during connect, but call again to get self_info)
|
||||
ok, evt_type, payload, err = await run_command(
|
||||
mc, cmd.send_appstart(), "send_appstart"
|
||||
)
|
||||
if ok:
|
||||
commands_succeeded += 1
|
||||
log.debug(f"appstart: {evt_type}")
|
||||
else:
|
||||
log.error(f"appstart failed: {err}")
|
||||
|
||||
# get_bat
|
||||
ok, evt_type, payload, err = await run_command(
|
||||
mc, cmd.get_bat(), "get_bat"
|
||||
)
|
||||
if ok:
|
||||
commands_succeeded += 1
|
||||
log.debug(f"get_bat: {payload}")
|
||||
else:
|
||||
log.error(f"get_bat failed: {err}")
|
||||
# send_device_query
|
||||
ok, evt_type, payload, err = await run_command(
|
||||
mc, cmd.send_device_query(), "send_device_query"
|
||||
)
|
||||
if ok:
|
||||
commands_succeeded += 1
|
||||
log.debug(f"device_query: {payload}")
|
||||
else:
|
||||
log.error(f"device_query failed: {err}")
|
||||
|
||||
# get_time
|
||||
ok, evt_type, payload, err = await run_command(
|
||||
mc, cmd.get_time(), "get_time"
|
||||
)
|
||||
if ok:
|
||||
commands_succeeded += 1
|
||||
log.debug(f"get_time: {payload}")
|
||||
else:
|
||||
log.error(f"get_time failed: {err}")
|
||||
# get_time
|
||||
ok, evt_type, payload, err = await run_command(
|
||||
mc, cmd.get_time(), "get_time"
|
||||
)
|
||||
if ok:
|
||||
commands_succeeded += 1
|
||||
log.debug(f"get_time: {payload}")
|
||||
else:
|
||||
log.error(f"get_time failed: {err}")
|
||||
|
||||
# get_self_telemetry
|
||||
ok, evt_type, payload, err = await run_command(
|
||||
mc, cmd.get_self_telemetry(), "get_self_telemetry"
|
||||
)
|
||||
if ok:
|
||||
commands_succeeded += 1
|
||||
log.debug(f"get_self_telemetry: {payload}")
|
||||
else:
|
||||
log.error(f"get_self_telemetry failed: {err}")
|
||||
# get_self_telemetry - collect environmental sensor data
|
||||
# Note: The call happens regardless of telemetry_enabled for device query completeness,
|
||||
# but we only extract and store metrics if the feature is enabled.
|
||||
ok, evt_type, payload, err = await run_command(
|
||||
mc, cmd.get_self_telemetry(), "get_self_telemetry"
|
||||
)
|
||||
if ok:
|
||||
commands_succeeded += 1
|
||||
log.debug(f"get_self_telemetry: {payload}")
|
||||
# Extract and store telemetry if enabled
|
||||
if cfg.telemetry_enabled:
|
||||
lpp_data = extract_lpp_from_payload(payload)
|
||||
if lpp_data is not None:
|
||||
telemetry_metrics = extract_telemetry_metrics(lpp_data)
|
||||
if telemetry_metrics:
|
||||
metrics.update(telemetry_metrics)
|
||||
log.debug(f"Extracted {len(telemetry_metrics)} telemetry metrics")
|
||||
else:
|
||||
# Debug level because not all devices have sensors attached - this is expected
|
||||
log.debug(f"get_self_telemetry failed: {err}")
|
||||
|
||||
# get_custom_vars
|
||||
ok, evt_type, payload, err = await run_command(
|
||||
mc, cmd.get_custom_vars(), "get_custom_vars"
|
||||
)
|
||||
if ok:
|
||||
commands_succeeded += 1
|
||||
log.debug(f"get_custom_vars: {payload}")
|
||||
else:
|
||||
log.debug(f"get_custom_vars failed: {err}")
|
||||
# get_custom_vars
|
||||
ok, evt_type, payload, err = await run_command(
|
||||
mc, cmd.get_custom_vars(), "get_custom_vars"
|
||||
)
|
||||
if ok:
|
||||
commands_succeeded += 1
|
||||
log.debug(f"get_custom_vars: {payload}")
|
||||
else:
|
||||
log.debug(f"get_custom_vars failed: {err}")
|
||||
|
||||
# get_contacts - count contacts
|
||||
ok, evt_type, payload, err = await run_command(
|
||||
mc, cmd.get_contacts(), "get_contacts"
|
||||
)
|
||||
if ok:
|
||||
commands_succeeded += 1
|
||||
contacts_count = len(payload) if payload else 0
|
||||
metrics["contacts"] = float(contacts_count)
|
||||
log.debug(f"get_contacts: found {contacts_count} contacts")
|
||||
else:
|
||||
log.error(f"get_contacts failed: {err}")
|
||||
# get_contacts - count contacts
|
||||
ok, evt_type, payload, err = await run_command(
|
||||
mc, cmd.get_contacts(), "get_contacts"
|
||||
)
|
||||
if ok:
|
||||
commands_succeeded += 1
|
||||
contacts_count = len(payload) if payload else 0
|
||||
metrics["contacts"] = float(contacts_count)
|
||||
log.debug(f"get_contacts: found {contacts_count} contacts")
|
||||
else:
|
||||
log.error(f"get_contacts failed: {err}")
|
||||
|
||||
# Get statistics - these contain the main metrics
|
||||
# Core stats (battery_mv, uptime_secs, errors, queue_len)
|
||||
ok, evt_type, payload, err = await run_command(
|
||||
mc, cmd.get_stats_core(), "get_stats_core"
|
||||
)
|
||||
if ok and payload and isinstance(payload, dict):
|
||||
commands_succeeded += 1
|
||||
# Insert all numeric fields from stats_core
|
||||
for key, value in payload.items():
|
||||
if isinstance(value, (int, float)):
|
||||
metrics[key] = float(value)
|
||||
log.debug(f"stats_core: {payload}")
|
||||
# Get statistics - these contain the main metrics
|
||||
# Core stats (battery_mv, uptime_secs, errors, queue_len)
|
||||
ok, evt_type, payload, err = await run_command(
|
||||
mc, cmd.get_stats_core(), "get_stats_core"
|
||||
)
|
||||
if ok and payload and isinstance(payload, dict):
|
||||
commands_succeeded += 1
|
||||
# Insert all numeric fields from stats_core
|
||||
for key, value in payload.items():
|
||||
if isinstance(value, (int, float)):
|
||||
metrics[key] = float(value)
|
||||
log.debug(f"stats_core: {payload}")
|
||||
|
||||
# Radio stats (noise_floor, last_rssi, last_snr, tx_air_secs, rx_air_secs)
|
||||
ok, evt_type, payload, err = await run_command(
|
||||
mc, cmd.get_stats_radio(), "get_stats_radio"
|
||||
)
|
||||
if ok and payload and isinstance(payload, dict):
|
||||
commands_succeeded += 1
|
||||
for key, value in payload.items():
|
||||
if isinstance(value, (int, float)):
|
||||
metrics[key] = float(value)
|
||||
log.debug(f"stats_radio: {payload}")
|
||||
# Radio stats (noise_floor, last_rssi, last_snr, tx_air_secs, rx_air_secs)
|
||||
ok, evt_type, payload, err = await run_command(
|
||||
mc, cmd.get_stats_radio(), "get_stats_radio"
|
||||
)
|
||||
if ok and payload and isinstance(payload, dict):
|
||||
commands_succeeded += 1
|
||||
for key, value in payload.items():
|
||||
if isinstance(value, (int, float)):
|
||||
metrics[key] = float(value)
|
||||
log.debug(f"stats_radio: {payload}")
|
||||
|
||||
# Packet stats (recv, sent, flood_tx, direct_tx, flood_rx, direct_rx)
|
||||
ok, evt_type, payload, err = await run_command(
|
||||
mc, cmd.get_stats_packets(), "get_stats_packets"
|
||||
)
|
||||
if ok and payload and isinstance(payload, dict):
|
||||
commands_succeeded += 1
|
||||
for key, value in payload.items():
|
||||
if isinstance(value, (int, float)):
|
||||
metrics[key] = float(value)
|
||||
log.debug(f"stats_packets: {payload}")
|
||||
# Packet stats (recv, sent, flood_tx, direct_tx, flood_rx, direct_rx)
|
||||
ok, evt_type, payload, err = await run_command(
|
||||
mc, cmd.get_stats_packets(), "get_stats_packets"
|
||||
)
|
||||
if ok and payload and isinstance(payload, dict):
|
||||
commands_succeeded += 1
|
||||
for key, value in payload.items():
|
||||
if isinstance(value, (int, float)):
|
||||
metrics[key] = float(value)
|
||||
log.debug(f"stats_packets: {payload}")
|
||||
|
||||
except Exception as e:
|
||||
log.error(f"Error during collection: {e}")
|
||||
except Exception as e:
|
||||
log.error(f"Error during collection: {e}")
|
||||
|
||||
finally:
|
||||
# Close connection
|
||||
if hasattr(mc, "disconnect"):
|
||||
try:
|
||||
await mc.disconnect()
|
||||
except Exception:
|
||||
pass
|
||||
# Connection closed and lock released by context manager
|
||||
|
||||
# Print summary
|
||||
summary_parts = [f"ts={ts}"]
|
||||
@@ -183,6 +178,10 @@ async def collect_companion() -> int:
|
||||
summary_parts.append(f"rx={int(metrics['recv'])}")
|
||||
if "sent" in metrics:
|
||||
summary_parts.append(f"tx={int(metrics['sent'])}")
|
||||
# Add telemetry count to summary if present
|
||||
telemetry_count = sum(1 for k in metrics if k.startswith("telemetry."))
|
||||
if telemetry_count > 0:
|
||||
summary_parts.append(f"telem={telemetry_count}")
|
||||
|
||||
log.info(f"Companion: {', '.join(summary_parts)}")
|
||||
|
||||
|
||||
@@ -27,15 +27,15 @@ sys.path.insert(0, str(Path(__file__).parent.parent / "src"))
|
||||
from meshmon.env import get_config
|
||||
from meshmon import log
|
||||
from meshmon.meshcore_client import (
|
||||
connect_from_env,
|
||||
connect_with_lock,
|
||||
run_command,
|
||||
get_contact_by_name,
|
||||
get_contact_by_key_prefix,
|
||||
extract_contact_info,
|
||||
list_contacts_summary,
|
||||
)
|
||||
from meshmon.db import init_db, insert_metrics
|
||||
from meshmon.retry import get_repeater_circuit_breaker, with_retries
|
||||
from meshmon.telemetry import extract_lpp_from_payload, extract_telemetry_metrics
|
||||
|
||||
|
||||
async def find_repeater_contact(mc: Any) -> Optional[Any]:
|
||||
@@ -143,8 +143,10 @@ async def query_repeater_with_retry(
|
||||
|
||||
|
||||
async def collect_repeater() -> int:
|
||||
"""
|
||||
Collect data from remote repeater node.
|
||||
"""Collect data from remote repeater node.
|
||||
|
||||
Collects status metrics (battery, uptime, packet counters, etc.) and
|
||||
optionally telemetry data (temperature, humidity, pressure) if enabled.
|
||||
|
||||
Returns:
|
||||
Exit code (0 = success, 1 = error)
|
||||
@@ -161,122 +163,154 @@ async def collect_repeater() -> int:
|
||||
# Skip collection - no metrics to write
|
||||
return 0
|
||||
|
||||
# Connect to companion
|
||||
log.debug("Connecting to companion node...")
|
||||
mc = await connect_from_env()
|
||||
|
||||
if mc is None:
|
||||
log.error("Failed to connect to companion node")
|
||||
return 1
|
||||
|
||||
# Metrics to insert (firmware field names from req_status_sync)
|
||||
metrics: dict[str, float] = {}
|
||||
status_metrics: dict[str, float] = {}
|
||||
telemetry_metrics: dict[str, float] = {}
|
||||
node_name = "unknown"
|
||||
status_ok = False
|
||||
|
||||
# Commands are accessed via mc.commands
|
||||
cmd = mc.commands
|
||||
|
||||
try:
|
||||
# Initialize (appstart already called during connect)
|
||||
ok, evt_type, payload, err = await run_command(
|
||||
mc, cmd.send_appstart(), "send_appstart"
|
||||
)
|
||||
if not ok:
|
||||
log.error(f"appstart failed: {err}")
|
||||
|
||||
# Find repeater contact
|
||||
contact = await find_repeater_contact(mc)
|
||||
|
||||
if contact is None:
|
||||
log.error("Cannot find repeater contact")
|
||||
# Connect to companion
|
||||
log.debug("Connecting to companion node...")
|
||||
async with connect_with_lock() as mc:
|
||||
if mc is None:
|
||||
log.error("Failed to connect to companion node")
|
||||
return 1
|
||||
|
||||
# Store contact info
|
||||
contact_info = extract_contact_info(contact)
|
||||
node_name = contact_info.get("adv_name", "unknown")
|
||||
# Commands are accessed via mc.commands
|
||||
cmd = mc.commands
|
||||
|
||||
log.debug(f"Found repeater: {node_name}")
|
||||
try:
|
||||
# Initialize (appstart already called during connect)
|
||||
ok, evt_type, payload, err = await run_command(
|
||||
mc, cmd.send_appstart(), "send_appstart"
|
||||
)
|
||||
if not ok:
|
||||
log.error(f"appstart failed: {err}")
|
||||
|
||||
# Optional login (if command exists)
|
||||
if cfg.repeater_password and hasattr(cmd, "send_login"):
|
||||
log.debug("Attempting login...")
|
||||
try:
|
||||
ok, evt_type, payload, err = await run_command(
|
||||
mc,
|
||||
cmd.send_login(contact, cfg.repeater_password),
|
||||
"send_login",
|
||||
)
|
||||
if ok:
|
||||
log.debug("Login successful")
|
||||
else:
|
||||
log.debug(f"Login failed or not supported: {err}")
|
||||
except Exception as e:
|
||||
log.debug(f"Login not supported: {e}")
|
||||
# Find repeater contact
|
||||
contact = await find_repeater_contact(mc)
|
||||
|
||||
# Query status (using _sync version which returns payload directly)
|
||||
# Use timeout=0 to let the device suggest timeout, with min_timeout as floor
|
||||
log.debug("Querying repeater status...")
|
||||
success, payload, err = await query_repeater_with_retry(
|
||||
mc,
|
||||
contact,
|
||||
"req_status_sync",
|
||||
lambda: cmd.req_status_sync(contact, timeout=0, min_timeout=cfg.remote_timeout_s),
|
||||
)
|
||||
if success and payload and isinstance(payload, dict):
|
||||
status_ok = True
|
||||
# Insert all numeric fields from status response
|
||||
for key, value in payload.items():
|
||||
if isinstance(value, (int, float)):
|
||||
metrics[key] = float(value)
|
||||
log.debug(f"req_status_sync: {payload}")
|
||||
else:
|
||||
log.warn(f"req_status_sync failed: {err}")
|
||||
if contact is None:
|
||||
log.error("Cannot find repeater contact")
|
||||
return 1
|
||||
|
||||
# Update circuit breaker
|
||||
if status_ok:
|
||||
cb.record_success()
|
||||
log.debug("Circuit breaker: recorded success")
|
||||
else:
|
||||
# Store contact info
|
||||
contact_info = extract_contact_info(contact)
|
||||
node_name = contact_info.get("adv_name", "unknown")
|
||||
|
||||
log.debug(f"Found repeater: {node_name}")
|
||||
|
||||
# Optional login (if command exists)
|
||||
if cfg.repeater_password and hasattr(cmd, "send_login"):
|
||||
log.debug("Attempting login...")
|
||||
try:
|
||||
ok, evt_type, payload, err = await run_command(
|
||||
mc,
|
||||
cmd.send_login(contact, cfg.repeater_password),
|
||||
"send_login",
|
||||
)
|
||||
if ok:
|
||||
log.debug("Login successful")
|
||||
else:
|
||||
log.debug(f"Login failed or not supported: {err}")
|
||||
except Exception as e:
|
||||
log.debug(f"Login not supported: {e}")
|
||||
|
||||
# Phase 1: Status collection (affects circuit breaker)
|
||||
# Use timeout=0 to let the device suggest timeout, with min_timeout as floor
|
||||
log.debug("Querying repeater status...")
|
||||
success, payload, err = await query_repeater_with_retry(
|
||||
mc,
|
||||
contact,
|
||||
"req_status_sync",
|
||||
lambda: cmd.req_status_sync(contact, timeout=0, min_timeout=cfg.remote_timeout_s),
|
||||
)
|
||||
if success and payload and isinstance(payload, dict):
|
||||
status_ok = True
|
||||
# Insert all numeric fields from status response
|
||||
for key, value in payload.items():
|
||||
if isinstance(value, (int, float)):
|
||||
status_metrics[key] = float(value)
|
||||
log.debug(f"req_status_sync: {payload}")
|
||||
else:
|
||||
log.warn(f"req_status_sync failed: {err}")
|
||||
|
||||
# Update circuit breaker based on status result
|
||||
if status_ok:
|
||||
cb.record_success()
|
||||
log.debug("Circuit breaker: recorded success")
|
||||
else:
|
||||
cb.record_failure(cfg.remote_cb_fails, cfg.remote_cb_cooldown_s)
|
||||
log.debug(f"Circuit breaker: recorded failure ({cb.consecutive_failures}/{cfg.remote_cb_fails})")
|
||||
|
||||
# CRITICAL: Store status metrics immediately before attempting telemetry
|
||||
# This ensures critical data is saved even if telemetry fails
|
||||
if status_ok and status_metrics:
|
||||
try:
|
||||
inserted = insert_metrics(ts=ts, role="repeater", metrics=status_metrics)
|
||||
log.debug(f"Stored {inserted} status metrics (ts={ts})")
|
||||
except Exception as e:
|
||||
log.error(f"Failed to store status metrics: {e}")
|
||||
return 1
|
||||
|
||||
# Phase 2: Telemetry collection (does NOT affect circuit breaker)
|
||||
if cfg.telemetry_enabled and status_ok:
|
||||
log.debug("Querying repeater telemetry...")
|
||||
try:
|
||||
# Note: Telemetry uses its own retry settings and does NOT
|
||||
# affect circuit breaker. Status success proves the link is up;
|
||||
# telemetry failures are likely firmware/capability issues.
|
||||
telem_success, telem_payload, telem_err = await with_retries(
|
||||
lambda: cmd.req_telemetry_sync(
|
||||
contact, timeout=0, min_timeout=cfg.telemetry_timeout_s
|
||||
),
|
||||
attempts=cfg.telemetry_retry_attempts,
|
||||
backoff_s=cfg.telemetry_retry_backoff_s,
|
||||
name="req_telemetry_sync",
|
||||
)
|
||||
|
||||
if telem_success and telem_payload:
|
||||
log.debug(f"req_telemetry_sync: {telem_payload}")
|
||||
lpp_data = extract_lpp_from_payload(telem_payload)
|
||||
if lpp_data is not None:
|
||||
telemetry_metrics = extract_telemetry_metrics(lpp_data)
|
||||
log.debug(f"Extracted {len(telemetry_metrics)} telemetry metrics")
|
||||
|
||||
# Store telemetry metrics
|
||||
if telemetry_metrics:
|
||||
try:
|
||||
inserted = insert_metrics(ts=ts, role="repeater", metrics=telemetry_metrics)
|
||||
log.debug(f"Stored {inserted} telemetry metrics")
|
||||
except Exception as e:
|
||||
log.warn(f"Failed to store telemetry metrics: {e}")
|
||||
else:
|
||||
log.warn(f"req_telemetry_sync failed: {telem_err}")
|
||||
except Exception as e:
|
||||
log.warn(f"Telemetry collection error (continuing): {e}")
|
||||
|
||||
except Exception as e:
|
||||
log.error(f"Error during collection: {e}")
|
||||
cb.record_failure(cfg.remote_cb_fails, cfg.remote_cb_cooldown_s)
|
||||
log.debug(f"Circuit breaker: recorded failure ({cb.consecutive_failures}/{cfg.remote_cb_fails})")
|
||||
|
||||
except Exception as e:
|
||||
log.error(f"Error during collection: {e}")
|
||||
cb.record_failure(cfg.remote_cb_fails, cfg.remote_cb_cooldown_s)
|
||||
|
||||
finally:
|
||||
# Close connection
|
||||
if hasattr(mc, "disconnect"):
|
||||
try:
|
||||
await mc.disconnect()
|
||||
except Exception:
|
||||
pass
|
||||
# Connection closed and lock released by context manager
|
||||
|
||||
# Print summary
|
||||
summary_parts = [f"ts={ts}"]
|
||||
if "bat" in metrics:
|
||||
bat_v = metrics["bat"] / 1000.0
|
||||
if "bat" in status_metrics:
|
||||
bat_v = status_metrics["bat"] / 1000.0
|
||||
summary_parts.append(f"bat={bat_v:.2f}V")
|
||||
if "uptime" in metrics:
|
||||
uptime_days = metrics["uptime"] // 86400
|
||||
if "uptime" in status_metrics:
|
||||
uptime_days = status_metrics["uptime"] // 86400
|
||||
summary_parts.append(f"uptime={int(uptime_days)}d")
|
||||
if "nb_recv" in metrics:
|
||||
summary_parts.append(f"rx={int(metrics['nb_recv'])}")
|
||||
if "nb_sent" in metrics:
|
||||
summary_parts.append(f"tx={int(metrics['nb_sent'])}")
|
||||
if "nb_recv" in status_metrics:
|
||||
summary_parts.append(f"rx={int(status_metrics['nb_recv'])}")
|
||||
if "nb_sent" in status_metrics:
|
||||
summary_parts.append(f"tx={int(status_metrics['nb_sent'])}")
|
||||
if telemetry_metrics:
|
||||
summary_parts.append(f"telem={len(telemetry_metrics)}")
|
||||
|
||||
log.info(f"Repeater ({node_name}): {', '.join(summary_parts)}")
|
||||
|
||||
# Write metrics to database
|
||||
if status_ok and metrics:
|
||||
try:
|
||||
inserted = insert_metrics(ts=ts, role="repeater", metrics=metrics)
|
||||
log.debug(f"Inserted {inserted} metrics to database (ts={ts})")
|
||||
except Exception as e:
|
||||
log.error(f"Failed to write metrics to database: {e}")
|
||||
return 1
|
||||
|
||||
return 0 if status_ok else 1
|
||||
|
||||
|
||||
|
||||
@@ -1,3 +1,3 @@
|
||||
"""MeshCore network monitoring library."""
|
||||
|
||||
__version__ = "0.2.3" # x-release-please-version
|
||||
__version__ = "0.2.8" # x-release-please-version
|
||||
|
||||
@@ -167,6 +167,7 @@ def load_timeseries_from_db(
|
||||
end_time: datetime,
|
||||
lookback: timedelta,
|
||||
period: str,
|
||||
all_metrics: Optional[dict[str, list[tuple[int, float]]]] = None,
|
||||
) -> TimeSeries:
|
||||
"""Load time series data from SQLite database.
|
||||
|
||||
@@ -179,6 +180,7 @@ def load_timeseries_from_db(
|
||||
end_time: End of the time range (typically now)
|
||||
lookback: How far back to look
|
||||
period: Period name for binning config ("day", "week", etc.)
|
||||
all_metrics: Optional pre-fetched metrics dict for this period
|
||||
|
||||
Returns:
|
||||
TimeSeries with extracted data points
|
||||
@@ -188,7 +190,8 @@ def load_timeseries_from_db(
|
||||
end_ts = int(end_time.timestamp())
|
||||
|
||||
# Fetch all metrics for this role/period (returns pivoted dict)
|
||||
all_metrics = get_metrics_for_period(role, start_ts, end_ts)
|
||||
if all_metrics is None:
|
||||
all_metrics = get_metrics_for_period(role, start_ts, end_ts)
|
||||
|
||||
# Get data for this specific metric
|
||||
metric_data = all_metrics.get(metric, [])
|
||||
@@ -379,10 +382,22 @@ def render_chart_svg(
|
||||
|
||||
# Plot area fill
|
||||
area_color = _hex_to_rgba(theme.area)
|
||||
ax.fill_between(timestamps, values, alpha=area_color[3], color=f"#{theme.line}")
|
||||
area = ax.fill_between(
|
||||
timestamps,
|
||||
values,
|
||||
alpha=area_color[3],
|
||||
color=f"#{theme.line}",
|
||||
)
|
||||
area.set_gid("chart-area")
|
||||
|
||||
# Plot line
|
||||
ax.plot(timestamps, values, color=f"#{theme.line}", linewidth=2)
|
||||
(line,) = ax.plot(
|
||||
timestamps,
|
||||
values,
|
||||
color=f"#{theme.line}",
|
||||
linewidth=2,
|
||||
)
|
||||
line.set_gid("chart-line")
|
||||
|
||||
# Set Y-axis limits and track actual values used
|
||||
if y_min is not None and y_max is not None:
|
||||
@@ -458,7 +473,7 @@ def _inject_data_attributes(
|
||||
|
||||
Adds:
|
||||
- data-metric, data-period, data-theme, data-x-start, data-x-end, data-y-min, data-y-max to root <svg>
|
||||
- data-points JSON array to the chart path element
|
||||
- data-points JSON array to the root <svg> and chart line path
|
||||
|
||||
Args:
|
||||
svg: Raw SVG string
|
||||
@@ -495,22 +510,35 @@ def _inject_data_attributes(
|
||||
r'<svg\b',
|
||||
f'<svg data-metric="{ts.metric}" data-period="{ts.period}" data-theme="{theme_name}" '
|
||||
f'data-x-start="{x_start_ts}" data-x-end="{x_end_ts}" '
|
||||
f'data-y-min="{y_min_val}" data-y-max="{y_max_val}"',
|
||||
f'data-y-min="{y_min_val}" data-y-max="{y_max_val}" '
|
||||
f'data-points="{data_points_attr}"',
|
||||
svg,
|
||||
count=1
|
||||
)
|
||||
|
||||
# Add data-points to the main path element (the line, not the fill)
|
||||
# Look for the second path element (first is usually the fill area)
|
||||
path_count = 0
|
||||
def add_data_to_path(match):
|
||||
nonlocal path_count
|
||||
path_count += 1
|
||||
if path_count == 2: # The line path
|
||||
return f'<path data-points="{data_points_attr}"'
|
||||
return match.group(0)
|
||||
def add_data_to_id(match):
|
||||
return f'<path{match.group(1)} data-points="{data_points_attr}"'
|
||||
|
||||
svg = re.sub(r'<path\b', add_data_to_path, svg)
|
||||
svg, count = re.subn(
|
||||
r'<path([^>]*(?:id|gid)="chart-line"[^>]*)',
|
||||
add_data_to_id,
|
||||
svg,
|
||||
count=1,
|
||||
)
|
||||
|
||||
if count == 0:
|
||||
# Look for the second path element (first is usually the fill area)
|
||||
path_count = 0
|
||||
|
||||
def add_data_to_path(match):
|
||||
nonlocal path_count
|
||||
path_count += 1
|
||||
if path_count == 2: # The line path
|
||||
return f'<path data-points="{data_points_attr}"'
|
||||
return match.group(0)
|
||||
|
||||
svg = re.sub(r'<path\b', add_data_to_path, svg)
|
||||
|
||||
return svg
|
||||
|
||||
@@ -558,9 +586,16 @@ def render_all_charts(
|
||||
for metric in metrics:
|
||||
all_stats[metric] = {}
|
||||
|
||||
for period in periods:
|
||||
period_cfg = PERIOD_CONFIG[period]
|
||||
for period in periods:
|
||||
period_cfg = PERIOD_CONFIG[period]
|
||||
x_end = now
|
||||
x_start = now - period_cfg["lookback"]
|
||||
|
||||
start_ts = int(x_start.timestamp())
|
||||
end_ts = int(x_end.timestamp())
|
||||
all_metrics = get_metrics_for_period(role, start_ts, end_ts)
|
||||
|
||||
for metric in metrics:
|
||||
# Load time series from database
|
||||
ts = load_timeseries_from_db(
|
||||
role=role,
|
||||
@@ -568,6 +603,7 @@ def render_all_charts(
|
||||
end_time=now,
|
||||
lookback=period_cfg["lookback"],
|
||||
period=period,
|
||||
all_metrics=all_metrics,
|
||||
)
|
||||
|
||||
# Calculate and store statistics
|
||||
@@ -579,10 +615,6 @@ def render_all_charts(
|
||||
y_min = y_range[0] if y_range else None
|
||||
y_max = y_range[1] if y_range else None
|
||||
|
||||
# Calculate X-axis range for full period padding
|
||||
x_end = now
|
||||
x_start = now - period_cfg["lookback"]
|
||||
|
||||
# Render chart for each theme
|
||||
for theme_name in themes:
|
||||
theme = CHART_THEMES[theme_name]
|
||||
|
||||
@@ -155,6 +155,14 @@ class Config:
|
||||
self.remote_cb_fails = get_int("REMOTE_CB_FAILS", 6)
|
||||
self.remote_cb_cooldown_s = get_int("REMOTE_CB_COOLDOWN_S", 3600)
|
||||
|
||||
# Telemetry collection (requires sensor board on repeater)
|
||||
self.telemetry_enabled = get_bool("TELEMETRY_ENABLED", False)
|
||||
# Separate settings allow tuning if telemetry proves problematic
|
||||
# Defaults match status settings - tune down if needed
|
||||
self.telemetry_timeout_s = get_int("TELEMETRY_TIMEOUT_S", 10)
|
||||
self.telemetry_retry_attempts = get_int("TELEMETRY_RETRY_ATTEMPTS", 2)
|
||||
self.telemetry_retry_backoff_s = get_int("TELEMETRY_RETRY_BACKOFF_S", 4)
|
||||
|
||||
# Paths (defaults are Docker container paths; native installs override via config)
|
||||
self.state_dir = get_path("STATE_DIR", "/data/state")
|
||||
self.out_dir = get_path("OUT_DIR", "/out")
|
||||
|
||||
@@ -588,8 +588,8 @@ def build_page_context(
|
||||
last_updated = None
|
||||
last_updated_iso = None
|
||||
if ts:
|
||||
dt = datetime.fromtimestamp(ts)
|
||||
last_updated = dt.strftime("%b %d, %Y at %H:%M UTC")
|
||||
dt = datetime.fromtimestamp(ts).astimezone()
|
||||
last_updated = dt.strftime("%b %d, %Y at %H:%M %Z")
|
||||
last_updated_iso = dt.isoformat()
|
||||
|
||||
# Build metrics for sidebar
|
||||
@@ -845,24 +845,24 @@ def build_monthly_table_data(
|
||||
airtime = m.get("airtime", MetricStats())
|
||||
|
||||
# Convert mV to V for display
|
||||
bat_v_mean = bat.mean / 1000.0 if bat.mean else None
|
||||
bat_v_min = bat.min_value / 1000.0 if bat.min_value else None
|
||||
bat_v_max = bat.max_value / 1000.0 if bat.max_value else None
|
||||
bat_v_mean = bat.mean / 1000.0 if bat.mean is not None else None
|
||||
bat_v_min = bat.min_value / 1000.0 if bat.min_value is not None else None
|
||||
bat_v_max = bat.max_value / 1000.0 if bat.max_value is not None else None
|
||||
|
||||
rows.append({
|
||||
"is_summary": False,
|
||||
"cells": [
|
||||
{"value": f"{daily.date.day:02d}", "class": None},
|
||||
{"value": f"{bat_v_mean:.2f}" if bat_v_mean else "-", "class": None},
|
||||
{"value": f"{bat_pct.mean:.0f}" if bat_pct.mean else "-", "class": None},
|
||||
{"value": f"{bat_v_mean:.2f}" if bat_v_mean is not None else "-", "class": None},
|
||||
{"value": f"{bat_pct.mean:.0f}" if bat_pct.mean is not None else "-", "class": None},
|
||||
{"value": _fmt_val_time(bat_v_min, bat.min_time), "class": "muted"},
|
||||
{"value": _fmt_val_time(bat_v_max, bat.max_time), "class": "muted"},
|
||||
{"value": f"{rssi.mean:.0f}" if rssi.mean else "-", "class": None},
|
||||
{"value": f"{snr.mean:.1f}" if snr.mean else "-", "class": None},
|
||||
{"value": f"{noise.mean:.0f}" if noise.mean else "-", "class": None},
|
||||
{"value": f"{rx.total:,}" if rx.total else "-", "class": "highlight"},
|
||||
{"value": f"{tx.total:,}" if tx.total else "-", "class": None},
|
||||
{"value": f"{airtime.total:,}" if airtime.total else "-", "class": None},
|
||||
{"value": f"{rssi.mean:.0f}" if rssi.mean is not None else "-", "class": None},
|
||||
{"value": f"{snr.mean:.1f}" if snr.mean is not None else "-", "class": None},
|
||||
{"value": f"{noise.mean:.0f}" if noise.mean is not None else "-", "class": None},
|
||||
{"value": f"{rx.total:,}" if rx.total is not None else "-", "class": "highlight"},
|
||||
{"value": f"{tx.total:,}" if tx.total is not None else "-", "class": None},
|
||||
{"value": f"{airtime.total:,}" if airtime.total is not None else "-", "class": None},
|
||||
],
|
||||
})
|
||||
|
||||
@@ -877,24 +877,24 @@ def build_monthly_table_data(
|
||||
tx = s.get("nb_sent", MetricStats())
|
||||
airtime = s.get("airtime", MetricStats())
|
||||
|
||||
bat_v_mean = bat.mean / 1000.0 if bat.mean else None
|
||||
bat_v_min = bat.min_value / 1000.0 if bat.min_value else None
|
||||
bat_v_max = bat.max_value / 1000.0 if bat.max_value else None
|
||||
bat_v_mean = bat.mean / 1000.0 if bat.mean is not None else None
|
||||
bat_v_min = bat.min_value / 1000.0 if bat.min_value is not None else None
|
||||
bat_v_max = bat.max_value / 1000.0 if bat.max_value is not None else None
|
||||
|
||||
rows.append({
|
||||
"is_summary": True,
|
||||
"cells": [
|
||||
{"value": "", "class": None},
|
||||
{"value": f"{bat_v_mean:.2f}" if bat_v_mean else "-", "class": None},
|
||||
{"value": f"{bat_pct.mean:.0f}" if bat_pct.mean else "-", "class": None},
|
||||
{"value": f"{bat_v_mean:.2f}" if bat_v_mean is not None else "-", "class": None},
|
||||
{"value": f"{bat_pct.mean:.0f}" if bat_pct.mean is not None else "-", "class": None},
|
||||
{"value": _fmt_val_day(bat_v_min, bat.min_time), "class": "muted"},
|
||||
{"value": _fmt_val_day(bat_v_max, bat.max_time), "class": "muted"},
|
||||
{"value": f"{rssi.mean:.0f}" if rssi.mean else "-", "class": None},
|
||||
{"value": f"{snr.mean:.1f}" if snr.mean else "-", "class": None},
|
||||
{"value": f"{noise.mean:.0f}" if noise.mean else "-", "class": None},
|
||||
{"value": f"{rx.total:,}" if rx.total else "-", "class": "highlight"},
|
||||
{"value": f"{tx.total:,}" if tx.total else "-", "class": None},
|
||||
{"value": f"{airtime.total:,}" if airtime.total else "-", "class": None},
|
||||
{"value": f"{rssi.mean:.0f}" if rssi.mean is not None else "-", "class": None},
|
||||
{"value": f"{snr.mean:.1f}" if snr.mean is not None else "-", "class": None},
|
||||
{"value": f"{noise.mean:.0f}" if noise.mean is not None else "-", "class": None},
|
||||
{"value": f"{rx.total:,}" if rx.total is not None else "-", "class": "highlight"},
|
||||
{"value": f"{tx.total:,}" if tx.total is not None else "-", "class": None},
|
||||
{"value": f"{airtime.total:,}" if airtime.total is not None else "-", "class": None},
|
||||
],
|
||||
})
|
||||
|
||||
@@ -928,21 +928,21 @@ def build_monthly_table_data(
|
||||
tx = m.get("sent", MetricStats())
|
||||
|
||||
# Convert mV to V for display
|
||||
bat_v_mean = bat.mean / 1000.0 if bat.mean else None
|
||||
bat_v_min = bat.min_value / 1000.0 if bat.min_value else None
|
||||
bat_v_max = bat.max_value / 1000.0 if bat.max_value else None
|
||||
bat_v_mean = bat.mean / 1000.0 if bat.mean is not None else None
|
||||
bat_v_min = bat.min_value / 1000.0 if bat.min_value is not None else None
|
||||
bat_v_max = bat.max_value / 1000.0 if bat.max_value is not None else None
|
||||
|
||||
rows.append({
|
||||
"is_summary": False,
|
||||
"cells": [
|
||||
{"value": f"{daily.date.day:02d}", "class": None},
|
||||
{"value": f"{bat_v_mean:.2f}" if bat_v_mean else "-", "class": None},
|
||||
{"value": f"{bat_pct.mean:.0f}" if bat_pct.mean else "-", "class": None},
|
||||
{"value": f"{bat_v_mean:.2f}" if bat_v_mean is not None else "-", "class": None},
|
||||
{"value": f"{bat_pct.mean:.0f}" if bat_pct.mean is not None else "-", "class": None},
|
||||
{"value": _fmt_val_time(bat_v_min, bat.min_time), "class": "muted"},
|
||||
{"value": _fmt_val_time(bat_v_max, bat.max_time), "class": "muted"},
|
||||
{"value": f"{contacts.mean:.0f}" if contacts.mean else "-", "class": None},
|
||||
{"value": f"{rx.total:,}" if rx.total else "-", "class": "highlight"},
|
||||
{"value": f"{tx.total:,}" if tx.total else "-", "class": None},
|
||||
{"value": f"{contacts.mean:.0f}" if contacts.mean is not None else "-", "class": None},
|
||||
{"value": f"{rx.total:,}" if rx.total is not None else "-", "class": "highlight"},
|
||||
{"value": f"{tx.total:,}" if tx.total is not None else "-", "class": None},
|
||||
],
|
||||
})
|
||||
|
||||
@@ -954,21 +954,21 @@ def build_monthly_table_data(
|
||||
rx = s.get("recv", MetricStats())
|
||||
tx = s.get("sent", MetricStats())
|
||||
|
||||
bat_v_mean = bat.mean / 1000.0 if bat.mean else None
|
||||
bat_v_min = bat.min_value / 1000.0 if bat.min_value else None
|
||||
bat_v_max = bat.max_value / 1000.0 if bat.max_value else None
|
||||
bat_v_mean = bat.mean / 1000.0 if bat.mean is not None else None
|
||||
bat_v_min = bat.min_value / 1000.0 if bat.min_value is not None else None
|
||||
bat_v_max = bat.max_value / 1000.0 if bat.max_value is not None else None
|
||||
|
||||
rows.append({
|
||||
"is_summary": True,
|
||||
"cells": [
|
||||
{"value": "", "class": None},
|
||||
{"value": f"{bat_v_mean:.2f}" if bat_v_mean else "-", "class": None},
|
||||
{"value": f"{bat_pct.mean:.0f}" if bat_pct.mean else "-", "class": None},
|
||||
{"value": f"{bat_v_mean:.2f}" if bat_v_mean is not None else "-", "class": None},
|
||||
{"value": f"{bat_pct.mean:.0f}" if bat_pct.mean is not None else "-", "class": None},
|
||||
{"value": _fmt_val_day(bat_v_min, bat.min_time), "class": "muted"},
|
||||
{"value": _fmt_val_day(bat_v_max, bat.max_time), "class": "muted"},
|
||||
{"value": f"{contacts.mean:.0f}" if contacts.mean else "-", "class": None},
|
||||
{"value": f"{rx.total:,}" if rx.total else "-", "class": "highlight"},
|
||||
{"value": f"{tx.total:,}" if tx.total else "-", "class": None},
|
||||
{"value": f"{contacts.mean:.0f}" if contacts.mean is not None else "-", "class": None},
|
||||
{"value": f"{rx.total:,}" if rx.total is not None else "-", "class": "highlight"},
|
||||
{"value": f"{tx.total:,}" if tx.total is not None else "-", "class": None},
|
||||
],
|
||||
})
|
||||
|
||||
@@ -1033,23 +1033,23 @@ def build_yearly_table_data(
|
||||
tx = s.get("nb_sent", MetricStats())
|
||||
|
||||
# Convert mV to V
|
||||
bat_v_mean = bat.mean / 1000.0 if bat.mean else None
|
||||
bat_v_min = bat.min_value / 1000.0 if bat.min_value else None
|
||||
bat_v_max = bat.max_value / 1000.0 if bat.max_value else None
|
||||
bat_v_mean = bat.mean / 1000.0 if bat.mean is not None else None
|
||||
bat_v_min = bat.min_value / 1000.0 if bat.min_value is not None else None
|
||||
bat_v_max = bat.max_value / 1000.0 if bat.max_value is not None else None
|
||||
|
||||
rows.append({
|
||||
"is_summary": False,
|
||||
"cells": [
|
||||
{"value": str(agg.year), "class": None},
|
||||
{"value": f"{monthly.month:02d}", "class": None},
|
||||
{"value": f"{bat_v_mean:.2f}" if bat_v_mean else "-", "class": None},
|
||||
{"value": f"{bat_pct.mean:.0f}" if bat_pct.mean else "-", "class": None},
|
||||
{"value": f"{bat_v_mean:.2f}" if bat_v_mean is not None else "-", "class": None},
|
||||
{"value": f"{bat_pct.mean:.0f}" if bat_pct.mean is not None else "-", "class": None},
|
||||
{"value": _fmt_val_day(bat_v_max, bat.max_time), "class": "muted"},
|
||||
{"value": _fmt_val_day(bat_v_min, bat.min_time), "class": "muted"},
|
||||
{"value": f"{rssi.mean:.0f}" if rssi.mean else "-", "class": None},
|
||||
{"value": f"{snr.mean:.1f}" if snr.mean else "-", "class": None},
|
||||
{"value": f"{rx.total:,}" if rx.total else "-", "class": "highlight"},
|
||||
{"value": f"{tx.total:,}" if tx.total else "-", "class": None},
|
||||
{"value": f"{rssi.mean:.0f}" if rssi.mean is not None else "-", "class": None},
|
||||
{"value": f"{snr.mean:.1f}" if snr.mean is not None else "-", "class": None},
|
||||
{"value": f"{rx.total:,}" if rx.total is not None else "-", "class": "highlight"},
|
||||
{"value": f"{tx.total:,}" if tx.total is not None else "-", "class": None},
|
||||
],
|
||||
})
|
||||
|
||||
@@ -1062,23 +1062,23 @@ def build_yearly_table_data(
|
||||
rx = s.get("nb_recv", MetricStats())
|
||||
tx = s.get("nb_sent", MetricStats())
|
||||
|
||||
bat_v_mean = bat.mean / 1000.0 if bat.mean else None
|
||||
bat_v_min = bat.min_value / 1000.0 if bat.min_value else None
|
||||
bat_v_max = bat.max_value / 1000.0 if bat.max_value else None
|
||||
bat_v_mean = bat.mean / 1000.0 if bat.mean is not None else None
|
||||
bat_v_min = bat.min_value / 1000.0 if bat.min_value is not None else None
|
||||
bat_v_max = bat.max_value / 1000.0 if bat.max_value is not None else None
|
||||
|
||||
rows.append({
|
||||
"is_summary": True,
|
||||
"cells": [
|
||||
{"value": "", "class": None},
|
||||
{"value": "", "class": None},
|
||||
{"value": f"{bat_v_mean:.2f}" if bat_v_mean else "-", "class": None},
|
||||
{"value": f"{bat_pct.mean:.0f}" if bat_pct.mean else "-", "class": None},
|
||||
{"value": f"{bat_v_mean:.2f}" if bat_v_mean is not None else "-", "class": None},
|
||||
{"value": f"{bat_pct.mean:.0f}" if bat_pct.mean is not None else "-", "class": None},
|
||||
{"value": _fmt_val_month(bat_v_max, bat.max_time), "class": "muted"},
|
||||
{"value": _fmt_val_month(bat_v_min, bat.min_time), "class": "muted"},
|
||||
{"value": f"{rssi.mean:.0f}" if rssi.mean else "-", "class": None},
|
||||
{"value": f"{snr.mean:.1f}" if snr.mean else "-", "class": None},
|
||||
{"value": f"{rx.total:,}" if rx.total else "-", "class": "highlight"},
|
||||
{"value": f"{tx.total:,}" if tx.total else "-", "class": None},
|
||||
{"value": f"{rssi.mean:.0f}" if rssi.mean is not None else "-", "class": None},
|
||||
{"value": f"{snr.mean:.1f}" if snr.mean is not None else "-", "class": None},
|
||||
{"value": f"{rx.total:,}" if rx.total is not None else "-", "class": "highlight"},
|
||||
{"value": f"{tx.total:,}" if tx.total is not None else "-", "class": None},
|
||||
],
|
||||
})
|
||||
|
||||
@@ -1113,22 +1113,22 @@ def build_yearly_table_data(
|
||||
tx = s.get("sent", MetricStats())
|
||||
|
||||
# Convert mV to V
|
||||
bat_v_mean = bat.mean / 1000.0 if bat.mean else None
|
||||
bat_v_min = bat.min_value / 1000.0 if bat.min_value else None
|
||||
bat_v_max = bat.max_value / 1000.0 if bat.max_value else None
|
||||
bat_v_mean = bat.mean / 1000.0 if bat.mean is not None else None
|
||||
bat_v_min = bat.min_value / 1000.0 if bat.min_value is not None else None
|
||||
bat_v_max = bat.max_value / 1000.0 if bat.max_value is not None else None
|
||||
|
||||
rows.append({
|
||||
"is_summary": False,
|
||||
"cells": [
|
||||
{"value": str(agg.year), "class": None},
|
||||
{"value": f"{monthly.month:02d}", "class": None},
|
||||
{"value": f"{bat_v_mean:.2f}" if bat_v_mean else "-", "class": None},
|
||||
{"value": f"{bat_pct.mean:.0f}" if bat_pct.mean else "-", "class": None},
|
||||
{"value": f"{bat_v_mean:.2f}" if bat_v_mean is not None else "-", "class": None},
|
||||
{"value": f"{bat_pct.mean:.0f}" if bat_pct.mean is not None else "-", "class": None},
|
||||
{"value": _fmt_val_day(bat_v_max, bat.max_time), "class": "muted"},
|
||||
{"value": _fmt_val_day(bat_v_min, bat.min_time), "class": "muted"},
|
||||
{"value": f"{contacts.mean:.0f}" if contacts.mean else "-", "class": None},
|
||||
{"value": f"{rx.total:,}" if rx.total else "-", "class": "highlight"},
|
||||
{"value": f"{tx.total:,}" if tx.total else "-", "class": None},
|
||||
{"value": f"{contacts.mean:.0f}" if contacts.mean is not None else "-", "class": None},
|
||||
{"value": f"{rx.total:,}" if rx.total is not None else "-", "class": "highlight"},
|
||||
{"value": f"{tx.total:,}" if tx.total is not None else "-", "class": None},
|
||||
],
|
||||
})
|
||||
|
||||
@@ -1140,22 +1140,22 @@ def build_yearly_table_data(
|
||||
rx = s.get("recv", MetricStats())
|
||||
tx = s.get("sent", MetricStats())
|
||||
|
||||
bat_v_mean = bat.mean / 1000.0 if bat.mean else None
|
||||
bat_v_min = bat.min_value / 1000.0 if bat.min_value else None
|
||||
bat_v_max = bat.max_value / 1000.0 if bat.max_value else None
|
||||
bat_v_mean = bat.mean / 1000.0 if bat.mean is not None else None
|
||||
bat_v_min = bat.min_value / 1000.0 if bat.min_value is not None else None
|
||||
bat_v_max = bat.max_value / 1000.0 if bat.max_value is not None else None
|
||||
|
||||
rows.append({
|
||||
"is_summary": True,
|
||||
"cells": [
|
||||
{"value": "", "class": None},
|
||||
{"value": "", "class": None},
|
||||
{"value": f"{bat_v_mean:.2f}" if bat_v_mean else "-", "class": None},
|
||||
{"value": f"{bat_pct.mean:.0f}" if bat_pct.mean else "-", "class": None},
|
||||
{"value": f"{bat_v_mean:.2f}" if bat_v_mean is not None else "-", "class": None},
|
||||
{"value": f"{bat_pct.mean:.0f}" if bat_pct.mean is not None else "-", "class": None},
|
||||
{"value": _fmt_val_month(bat_v_max, bat.max_time), "class": "muted"},
|
||||
{"value": _fmt_val_month(bat_v_min, bat.min_time), "class": "muted"},
|
||||
{"value": f"{contacts.mean:.0f}" if contacts.mean else "-", "class": None},
|
||||
{"value": f"{rx.total:,}" if rx.total else "-", "class": "highlight"},
|
||||
{"value": f"{tx.total:,}" if tx.total else "-", "class": None},
|
||||
{"value": f"{contacts.mean:.0f}" if contacts.mean is not None else "-", "class": None},
|
||||
{"value": f"{rx.total:,}" if rx.total is not None else "-", "class": "highlight"},
|
||||
{"value": f"{tx.total:,}" if tx.total is not None else "-", "class": None},
|
||||
],
|
||||
})
|
||||
|
||||
|
||||
@@ -1,7 +1,10 @@
|
||||
"""MeshCore client wrapper with safe command execution and contact lookup."""
|
||||
|
||||
import asyncio
|
||||
from typing import Any, Optional, Callable, Coroutine
|
||||
import fcntl
|
||||
from contextlib import asynccontextmanager
|
||||
from pathlib import Path
|
||||
from typing import Any, AsyncIterator, Callable, Coroutine, Optional
|
||||
|
||||
from .env import get_config
|
||||
from . import log
|
||||
@@ -100,6 +103,92 @@ async def connect_from_env() -> Optional[Any]:
|
||||
return None
|
||||
|
||||
|
||||
async def _acquire_lock_async(
|
||||
lock_file,
|
||||
timeout: float = 60.0,
|
||||
poll_interval: float = 0.1,
|
||||
) -> None:
|
||||
"""Acquire exclusive file lock without blocking the event loop.
|
||||
|
||||
Uses non-blocking LOCK_NB with async polling to avoid freezing the event loop.
|
||||
|
||||
Args:
|
||||
lock_file: Open file handle to lock
|
||||
timeout: Maximum seconds to wait for lock
|
||||
poll_interval: Seconds between lock attempts
|
||||
|
||||
Raises:
|
||||
TimeoutError: If lock cannot be acquired within timeout
|
||||
"""
|
||||
loop = asyncio.get_running_loop()
|
||||
deadline = loop.time() + timeout
|
||||
|
||||
while True:
|
||||
try:
|
||||
fcntl.flock(lock_file.fileno(), fcntl.LOCK_EX | fcntl.LOCK_NB)
|
||||
return
|
||||
except BlockingIOError:
|
||||
if loop.time() >= deadline:
|
||||
raise TimeoutError(
|
||||
f"Could not acquire serial lock within {timeout}s. "
|
||||
"Another process may be using the serial port."
|
||||
)
|
||||
await asyncio.sleep(poll_interval)
|
||||
|
||||
|
||||
@asynccontextmanager
|
||||
async def connect_with_lock(
|
||||
lock_timeout: float = 60.0,
|
||||
) -> AsyncIterator[Optional[Any]]:
|
||||
"""Connect to MeshCore with serial port locking to prevent concurrent access.
|
||||
|
||||
For serial transport: Acquires exclusive file lock before connecting.
|
||||
For TCP/BLE: No locking needed (protocol handles multiple connections).
|
||||
|
||||
Args:
|
||||
lock_timeout: Maximum seconds to wait for serial lock
|
||||
|
||||
Yields:
|
||||
MeshCore client instance, or None if connection failed
|
||||
"""
|
||||
cfg = get_config()
|
||||
lock_file = None
|
||||
mc = None
|
||||
needs_lock = cfg.mesh_transport.lower() == "serial"
|
||||
|
||||
try:
|
||||
if needs_lock:
|
||||
lock_path: Path = cfg.state_dir / "serial.lock"
|
||||
lock_path.parent.mkdir(parents=True, exist_ok=True)
|
||||
|
||||
# Use 'a' mode: doesn't truncate, creates if missing
|
||||
lock_file = open(lock_path, "a")
|
||||
try:
|
||||
await _acquire_lock_async(lock_file, timeout=lock_timeout)
|
||||
log.debug(f"Acquired serial lock: {lock_path}")
|
||||
except Exception:
|
||||
# If lock acquisition fails, close file before re-raising
|
||||
lock_file.close()
|
||||
lock_file = None
|
||||
raise
|
||||
|
||||
mc = await connect_from_env()
|
||||
yield mc
|
||||
|
||||
finally:
|
||||
# Disconnect first (while we still hold the lock)
|
||||
if mc is not None and hasattr(mc, "disconnect"):
|
||||
try:
|
||||
await mc.disconnect()
|
||||
except Exception as e:
|
||||
log.debug(f"Error during disconnect (ignored): {e}")
|
||||
|
||||
# Release lock by closing the file (close() auto-releases flock)
|
||||
if lock_file is not None:
|
||||
lock_file.close()
|
||||
log.debug("Released serial lock")
|
||||
|
||||
|
||||
async def run_command(
|
||||
mc: Any,
|
||||
cmd_coro: Coroutine,
|
||||
|
||||
@@ -17,17 +17,12 @@ import calendar
|
||||
import json
|
||||
from dataclasses import dataclass, field
|
||||
from datetime import date, datetime, timedelta
|
||||
from pathlib import Path
|
||||
from typing import Any, Optional
|
||||
|
||||
from .db import get_connection, get_metrics_for_period, VALID_ROLES
|
||||
from .env import get_config
|
||||
from .metrics import (
|
||||
is_counter_metric,
|
||||
get_chart_metrics,
|
||||
transform_value,
|
||||
)
|
||||
from . import log
|
||||
|
||||
|
||||
def _validate_role(role: str) -> str:
|
||||
@@ -59,6 +54,32 @@ def get_metrics_for_role(role: str) -> list[str]:
|
||||
raise ValueError(f"Unknown role: {role}")
|
||||
|
||||
|
||||
REPORT_UNITS_RAW = {
|
||||
"battery_mv": "mV",
|
||||
"bat": "mV",
|
||||
"bat_pct": "%",
|
||||
"uptime": "s",
|
||||
"uptime_secs": "s",
|
||||
"last_rssi": "dBm",
|
||||
"last_snr": "dB",
|
||||
"noise_floor": "dBm",
|
||||
"tx_queue_len": "count",
|
||||
"contacts": "count",
|
||||
"recv": "packets",
|
||||
"sent": "packets",
|
||||
"nb_recv": "packets",
|
||||
"nb_sent": "packets",
|
||||
"airtime": "s",
|
||||
"rx_airtime": "s",
|
||||
"flood_dups": "packets",
|
||||
"direct_dups": "packets",
|
||||
"sent_flood": "packets",
|
||||
"recv_flood": "packets",
|
||||
"sent_direct": "packets",
|
||||
"recv_direct": "packets",
|
||||
}
|
||||
|
||||
|
||||
@dataclass
|
||||
class MetricStats:
|
||||
"""Statistics for a single metric over a period.
|
||||
@@ -1116,10 +1137,14 @@ def format_yearly_txt(
|
||||
return format_yearly_txt_companion(agg, node_name, location)
|
||||
|
||||
|
||||
def _metric_stats_to_dict(stats: MetricStats) -> dict[str, Any]:
|
||||
def _metric_stats_to_dict(stats: MetricStats, metric: str) -> dict[str, Any]:
|
||||
"""Convert MetricStats to JSON-serializable dict."""
|
||||
result: dict[str, Any] = {"count": stats.count}
|
||||
|
||||
unit = REPORT_UNITS_RAW.get(metric)
|
||||
if unit:
|
||||
result["unit"] = unit
|
||||
|
||||
if stats.mean is not None:
|
||||
result["mean"] = round(stats.mean, 4)
|
||||
if stats.min_value is not None:
|
||||
@@ -1144,7 +1169,7 @@ def _daily_to_dict(daily: DailyAggregate) -> dict[str, Any]:
|
||||
"date": daily.date.isoformat(),
|
||||
"snapshot_count": daily.snapshot_count,
|
||||
"metrics": {
|
||||
ds: _metric_stats_to_dict(stats)
|
||||
ds: _metric_stats_to_dict(stats, ds)
|
||||
for ds, stats in daily.metrics.items()
|
||||
if stats.has_data
|
||||
},
|
||||
@@ -1167,7 +1192,7 @@ def monthly_to_json(agg: MonthlyAggregate) -> dict[str, Any]:
|
||||
"role": agg.role,
|
||||
"days_with_data": len(agg.daily),
|
||||
"summary": {
|
||||
ds: _metric_stats_to_dict(stats)
|
||||
ds: _metric_stats_to_dict(stats, ds)
|
||||
for ds, stats in agg.summary.items()
|
||||
if stats.has_data
|
||||
},
|
||||
@@ -1190,7 +1215,7 @@ def yearly_to_json(agg: YearlyAggregate) -> dict[str, Any]:
|
||||
"role": agg.role,
|
||||
"months_with_data": len(agg.monthly),
|
||||
"summary": {
|
||||
ds: _metric_stats_to_dict(stats)
|
||||
ds: _metric_stats_to_dict(stats, ds)
|
||||
for ds, stats in agg.summary.items()
|
||||
if stats.has_data
|
||||
},
|
||||
@@ -1200,7 +1225,7 @@ def yearly_to_json(agg: YearlyAggregate) -> dict[str, Any]:
|
||||
"month": m.month,
|
||||
"days_with_data": len(m.daily),
|
||||
"summary": {
|
||||
ds: _metric_stats_to_dict(stats)
|
||||
ds: _metric_stats_to_dict(stats, ds)
|
||||
for ds, stats in m.summary.items()
|
||||
if stats.has_data
|
||||
},
|
||||
|
||||
102
src/meshmon/telemetry.py
Normal file
102
src/meshmon/telemetry.py
Normal file
@@ -0,0 +1,102 @@
|
||||
"""Telemetry data extraction from Cayenne LPP format."""
|
||||
|
||||
from typing import Any
|
||||
from . import log
|
||||
|
||||
__all__ = ["extract_lpp_from_payload", "extract_telemetry_metrics"]
|
||||
|
||||
|
||||
def extract_lpp_from_payload(payload: Any) -> list | None:
|
||||
"""Extract LPP data list from telemetry payload.
|
||||
|
||||
Handles both formats returned by the MeshCore API:
|
||||
- Dict format: {'pubkey_pre': '...', 'lpp': [...]}
|
||||
- Direct list format: [...]
|
||||
|
||||
Args:
|
||||
payload: Raw telemetry payload from get_self_telemetry() or req_telemetry_sync()
|
||||
|
||||
Returns:
|
||||
The LPP data list, or None if not extractable.
|
||||
"""
|
||||
if payload is None:
|
||||
return None
|
||||
|
||||
if isinstance(payload, dict):
|
||||
lpp = payload.get("lpp")
|
||||
if lpp is None:
|
||||
log.debug("No 'lpp' key in telemetry payload dict")
|
||||
return None
|
||||
if not isinstance(lpp, list):
|
||||
log.debug(f"Unexpected LPP data type in payload: {type(lpp).__name__}")
|
||||
return None
|
||||
return lpp
|
||||
|
||||
if isinstance(payload, list):
|
||||
return payload
|
||||
|
||||
log.debug(f"Unexpected telemetry payload type: {type(payload).__name__}")
|
||||
return None
|
||||
|
||||
|
||||
def extract_telemetry_metrics(lpp_data: Any) -> dict[str, float]:
|
||||
"""Extract numeric telemetry values from Cayenne LPP response.
|
||||
|
||||
Expected format:
|
||||
[
|
||||
{"type": "temperature", "channel": 0, "value": 23.5},
|
||||
{"type": "gps", "channel": 1, "value": {"latitude": 51.5, "longitude": -0.1, "altitude": 10}}
|
||||
]
|
||||
|
||||
Keys are formatted as:
|
||||
- telemetry.{type}.{channel} for scalar values
|
||||
- telemetry.{type}.{channel}.{subkey} for compound values (e.g., GPS)
|
||||
|
||||
Returns:
|
||||
Dict mapping metric keys to float values. Invalid readings are skipped.
|
||||
"""
|
||||
if not isinstance(lpp_data, list):
|
||||
log.warn(f"Expected list for LPP data, got {type(lpp_data).__name__}")
|
||||
return {}
|
||||
|
||||
metrics: dict[str, float] = {}
|
||||
|
||||
for i, reading in enumerate(lpp_data):
|
||||
if not isinstance(reading, dict):
|
||||
log.debug(f"Skipping non-dict LPP reading at index {i}")
|
||||
continue
|
||||
|
||||
sensor_type = reading.get("type")
|
||||
if not isinstance(sensor_type, str) or not sensor_type.strip():
|
||||
log.debug(f"Skipping reading with invalid type at index {i}")
|
||||
continue
|
||||
|
||||
# Normalize sensor type for use as metric key component
|
||||
sensor_type = sensor_type.strip().lower().replace(" ", "_")
|
||||
|
||||
channel = reading.get("channel", 0)
|
||||
if not isinstance(channel, int):
|
||||
channel = 0
|
||||
|
||||
value = reading.get("value")
|
||||
base_key = f"telemetry.{sensor_type}.{channel}"
|
||||
|
||||
# Note: Check bool before int because bool is a subclass of int in Python.
|
||||
# Some sensors may report digital on/off values as booleans.
|
||||
if isinstance(value, bool):
|
||||
metrics[base_key] = float(value)
|
||||
elif isinstance(value, (int, float)):
|
||||
metrics[base_key] = float(value)
|
||||
elif isinstance(value, dict):
|
||||
for subkey, subval in value.items():
|
||||
if not isinstance(subkey, str):
|
||||
continue
|
||||
subkey_clean = subkey.strip().lower().replace(" ", "_")
|
||||
if not subkey_clean:
|
||||
continue
|
||||
if isinstance(subval, bool):
|
||||
metrics[f"{base_key}.{subkey_clean}"] = float(subval)
|
||||
elif isinstance(subval, (int, float)):
|
||||
metrics[f"{base_key}.{subkey_clean}"] = float(subval)
|
||||
|
||||
return metrics
|
||||
@@ -58,7 +58,8 @@
|
||||
month: 'short',
|
||||
day: 'numeric',
|
||||
hour: '2-digit',
|
||||
minute: '2-digit'
|
||||
minute: '2-digit',
|
||||
timeZoneName: 'short'
|
||||
};
|
||||
|
||||
// For year view, include year
|
||||
@@ -180,15 +181,23 @@
|
||||
const yMin = parseFloat(svg.dataset.yMin);
|
||||
const yMax = parseFloat(svg.dataset.yMax);
|
||||
|
||||
// Find the path with data-points
|
||||
const path = svg.querySelector('path[data-points]');
|
||||
// Find the primary line path for precise coordinates
|
||||
const path =
|
||||
svg.querySelector('path#chart-line') ||
|
||||
svg.querySelector('path[gid="chart-line"]') ||
|
||||
svg.querySelector('#chart-line path') ||
|
||||
svg.querySelector('[gid="chart-line"] path') ||
|
||||
svg.querySelector('path[data-points]');
|
||||
if (!path) return;
|
||||
|
||||
// Parse and cache data points and path coordinates on first access
|
||||
if (!path._dataPoints) {
|
||||
const pointsSource = path.dataset.points || svg.dataset.points;
|
||||
if (!pointsSource) return;
|
||||
|
||||
// Parse and cache data points on first access
|
||||
if (!svg._dataPoints) {
|
||||
try {
|
||||
const json = path.dataset.points.replace(/"/g, '"');
|
||||
path._dataPoints = JSON.parse(json);
|
||||
const json = pointsSource.replace(/"/g, '"');
|
||||
svg._dataPoints = JSON.parse(json);
|
||||
} catch (e) {
|
||||
console.warn('Failed to parse chart data:', e);
|
||||
return;
|
||||
@@ -220,7 +229,7 @@
|
||||
const targetTs = xStart + clampedRelX * (xEnd - xStart);
|
||||
|
||||
// Find closest data point by timestamp
|
||||
const result = findClosestPoint(path._dataPoints, targetTs);
|
||||
const result = findClosestPoint(svg._dataPoints, targetTs);
|
||||
if (!result) return;
|
||||
|
||||
const { point } = result;
|
||||
|
||||
Reference in New Issue
Block a user