mirror of
https://github.com/jorijn/meshcore-stats.git
synced 2026-05-06 21:42:39 +02:00
* test: add comprehensive pytest test suite with 95% coverage Add full unit and integration test coverage for the meshcore-stats project: - 1020 tests covering all modules (db, charts, html, reports, client, etc.) - 95.95% code coverage with pytest-cov (95% threshold enforced) - GitHub Actions CI workflow for automated testing on push/PR - Proper mocking of external dependencies (meshcore, serial, filesystem) - SVG snapshot infrastructure for chart regression testing - Integration tests for collection and rendering pipelines Test organization: - tests/charts/: Chart rendering and statistics - tests/client/: MeshCore client and connection handling - tests/config/: Environment and configuration parsing - tests/database/: SQLite operations and migrations - tests/html/: HTML generation and Jinja templates - tests/reports/: Report generation and formatting - tests/retry/: Circuit breaker and retry logic - tests/unit/: Pure unit tests for utilities - tests/integration/: End-to-end pipeline tests 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com> * chore: add test-engineer agent configuration Add project-local test-engineer agent for pytest test development, coverage analysis, and test review tasks. 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com> * docs: comprehensive test suite review with 956 tests analyzed Conducted thorough review of all 956 test cases across 47 test files: - Unit Tests: 338 tests (battery, metrics, log, telemetry, env, charts, html, reports, formatters) - Config Tests: 53 tests (env loading, config file parsing) - Database Tests: 115 tests (init, insert, queries, migrations, maintenance, validation) - Retry Tests: 59 tests (circuit breaker, async retries, factory) - Charts Tests: 76 tests (transforms, statistics, timeseries, rendering, I/O) - HTML Tests: 81 tests (site generation, Jinja2, metrics builders, reports index) - Reports Tests: 149 tests (location, JSON/TXT formatting, aggregation, counter totals) - Client Tests: 63 tests (contacts, connection, meshcore availability, commands) - Integration Tests: 22 tests (reports, collection, rendering pipelines) Results: - Overall Pass Rate: 99.7% (953/956) - 3 tests marked for improvement (empty test bodies in client tests) - 0 tests requiring fixes Key findings documented in test_review/tests.md including quality observations, F.I.R.S.T. principle adherence, and recommendations. 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com> * test: implement snapshot testing for charts and reports Add comprehensive snapshot testing infrastructure: SVG Chart Snapshots: - Deterministic fixtures with fixed timestamps (2024-01-15 12:00:00) - Tests for gauge/counter metrics in light/dark themes - Empty chart and single-point edge cases - Extended normalize_svg_for_snapshot_full() for reproducible comparisons TXT Report Snapshots: - Monthly/yearly report snapshots for repeater and companion - Empty report handling tests - Tests in tests/reports/test_snapshots.py Infrastructure: - tests/snapshots/conftest.py with shared fixtures - UPDATE_SNAPSHOTS=1 environment variable for regeneration - scripts/generate_snapshots.py for batch snapshot generation Run `UPDATE_SNAPSHOTS=1 pytest tests/charts/test_chart_render.py::TestSvgSnapshots` to generate initial snapshots. 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com> * test: fix SVG normalization and generate initial snapshots Fix normalize_svg_for_snapshot() to handle: - clipPath IDs like id="p47c77a2a6e" - url(#p...) references - xlink:href="#p..." references - <dc:date> timestamps Generated initial snapshot files: - 7 SVG chart snapshots (gauge, counter, empty, single-point in light/dark) - 6 TXT report snapshots (monthly/yearly for repeater/companion + empty) All 13 snapshot tests now pass. 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com> * test: fix SVG normalization to preserve axis rendering The SVG normalization was replacing all matplotlib-generated IDs with the same value, causing duplicate IDs that broke SVG rendering: - Font glyphs, clipPaths, and tick marks all got id="normalized" - References couldn't resolve to the correct elements - X and Y axes failed to render in normalized snapshots Fix uses type-specific prefixes with sequential numbering: - glyph_N for font glyphs (DejaVuSans-XX patterns) - clip_N for clipPath definitions (p[0-9a-f]{8,} patterns) - tick_N for tick marks (m[0-9a-f]{8,} patterns) This ensures all IDs remain unique while still being deterministic for snapshot comparison. 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com> * chore: add coverage and pytest artifacts to gitignore Add .coverage, .coverage.*, htmlcov/, and .pytest_cache/ to prevent test artifacts from being committed. 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com> * style: fix all ruff lint errors across codebase - Sort and organize imports (I001) - Use modern type annotations (X | Y instead of Union, collections.abc) - Remove unused imports (F401) - Combine nested if statements (SIM102) - Use ternary operators where appropriate (SIM108) - Combine nested with statements (SIM117) - Use contextlib.suppress instead of try-except-pass (SIM105) - Add noqa comments for intentional SIM115 violations (file locks) - Add TYPE_CHECKING import for forward references - Fix exception chaining (B904) All 1033 tests pass. 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com> * docs: add TDD workflow and pre-commit requirements to CLAUDE.md - Add mandatory test-driven development workflow (write tests first) - Add pre-commit requirements (must run lint and tests before committing) - Document test organization and running commands - Document 95% coverage requirement 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com> * fix: resolve mypy type checking errors with proper structural fixes - charts.py: Create PeriodConfig dataclass for type-safe period configuration, use mdates.date2num() for matplotlib datetime handling, fix x-axis limits for single-point charts - db.py: Add explicit int() conversion with None handling for SQLite returns - env.py: Add class-level type annotations to Config class - html.py: Add MetricDisplay TypedDict, fix import order, add proper type annotations for table data functions - meshcore_client.py: Add return type annotation Update tests to use new dataclass attribute access and regenerate SVG snapshots. Add mypy step to CLAUDE.md pre-commit requirements. 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com> * fix: cast Jinja2 template.render() to str for mypy Jinja2's type stubs declare render() as returning Any, but it actually returns str. Wrap with str() to satisfy mypy's no-any-return check. 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com> * ci: improve workflow security and reliability - test.yml: Pin all actions by SHA, add concurrency control to cancel in-progress runs on rapid pushes - release-please.yml: Pin action by SHA, add 10-minute timeout - conftest.py: Fix snapshot_base_time to use explicit UTC timezone for consistent behavior across CI and local environments Regenerate SVG snapshots with UTC-aware timestamps. 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com> * fix: add mypy command to permissions in settings.local.json * test: add comprehensive script tests with coroutine warning fixes - Add tests/scripts/ with tests for collect_companion, collect_repeater, and render scripts (1135 tests total, 96% coverage) - Fix unawaited coroutine warnings by using AsyncMock properly for async functions and async_context_manager_factory fixture for context managers - Add --cov=scripts to CI workflow and pyproject.toml coverage config - Omit scripts/generate_snapshots.py from coverage (dev utility) 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com> * docs: migrate claude setup to codex skills * feat: migrate dependencies to uv (#31) * fix: run tests through uv * test: fix ruff lint issues in tests Consolidate patch context managers and clean unused imports/variables Use datetime.UTC in snapshot fixtures * test: avoid unawaited async mocks in entrypoint tests * ci: replace codecov with github coverage artifacts Add junit XML output and coverage summary in job output Upload HTML and XML coverage artifacts (3.12 only) on every run --------- Co-authored-by: Claude Opus 4.5 <noreply@anthropic.com>
387 lines
13 KiB
Python
387 lines
13 KiB
Python
"""Tests for HTML builder functions in html.py."""
|
|
|
|
|
|
from meshmon.html import (
|
|
COMPANION_CHART_GROUPS,
|
|
PERIOD_CONFIG,
|
|
REPEATER_CHART_GROUPS,
|
|
_build_traffic_table_rows,
|
|
build_companion_metrics,
|
|
build_node_details,
|
|
build_radio_config,
|
|
build_repeater_metrics,
|
|
get_jinja_env,
|
|
)
|
|
|
|
|
|
class TestBuildTrafficTableRows:
|
|
"""Test _build_traffic_table_rows function."""
|
|
|
|
def test_empty_list(self):
|
|
"""Empty list returns empty list."""
|
|
result = _build_traffic_table_rows([])
|
|
assert result == []
|
|
|
|
def test_rx_tx_packets(self):
|
|
"""RX and TX become Packets row."""
|
|
traffic = [
|
|
{"label": "RX", "value": "1.2k", "raw_value": 1200, "unit": "packets"},
|
|
{"label": "TX", "value": "800", "raw_value": 800, "unit": "packets"},
|
|
]
|
|
result = _build_traffic_table_rows(traffic)
|
|
|
|
assert len(result) == 1
|
|
assert result[0]["label"] == "Packets"
|
|
assert result[0]["rx"] == "1.2k"
|
|
assert result[0]["rx_raw"] == 1200
|
|
assert result[0]["tx"] == "800"
|
|
assert result[0]["tx_raw"] == 800
|
|
|
|
def test_flood_rx_tx(self):
|
|
"""Flood RX/TX become Flood row."""
|
|
traffic = [
|
|
{"label": "Flood RX", "value": "500", "raw_value": 500, "unit": "packets"},
|
|
{"label": "Flood TX", "value": "300", "raw_value": 300, "unit": "packets"},
|
|
]
|
|
result = _build_traffic_table_rows(traffic)
|
|
|
|
assert len(result) == 1
|
|
assert result[0]["label"] == "Flood"
|
|
assert result[0]["rx"] == "500"
|
|
assert result[0]["tx"] == "300"
|
|
|
|
def test_direct_rx_tx(self):
|
|
"""Direct RX/TX become Direct row."""
|
|
traffic = [
|
|
{"label": "Direct RX", "value": "200", "raw_value": 200, "unit": "packets"},
|
|
{"label": "Direct TX", "value": "100", "raw_value": 100, "unit": "packets"},
|
|
]
|
|
result = _build_traffic_table_rows(traffic)
|
|
|
|
assert len(result) == 1
|
|
assert result[0]["label"] == "Direct"
|
|
assert result[0]["rx"] == "200"
|
|
assert result[0]["tx"] == "100"
|
|
|
|
def test_airtime_rx_tx(self):
|
|
"""Airtime TX/RX become Airtime row."""
|
|
traffic = [
|
|
{"label": "Airtime TX", "value": "1h 30m", "raw_value": 5400, "unit": "seconds"},
|
|
{"label": "Airtime RX", "value": "3h 0m", "raw_value": 10800, "unit": "seconds"},
|
|
]
|
|
result = _build_traffic_table_rows(traffic)
|
|
|
|
assert len(result) == 1
|
|
assert result[0]["label"] == "Airtime"
|
|
assert result[0]["tx"] == "1h 30m"
|
|
assert result[0]["rx"] == "3h 0m"
|
|
|
|
def test_output_order(self):
|
|
"""Output follows order: Packets, Flood, Direct, Airtime."""
|
|
traffic = [
|
|
{"label": "Airtime TX", "value": "1h", "raw_value": 3600, "unit": "seconds"},
|
|
{"label": "Direct RX", "value": "100", "raw_value": 100, "unit": "packets"},
|
|
{"label": "Flood RX", "value": "200", "raw_value": 200, "unit": "packets"},
|
|
{"label": "RX", "value": "500", "raw_value": 500, "unit": "packets"},
|
|
]
|
|
result = _build_traffic_table_rows(traffic)
|
|
|
|
labels = [r["label"] for r in result]
|
|
assert labels == ["Packets", "Flood", "Direct", "Airtime"]
|
|
|
|
def test_missing_pair(self):
|
|
"""Missing pair leaves None for that direction."""
|
|
traffic = [
|
|
{"label": "RX", "value": "500", "raw_value": 500, "unit": "packets"},
|
|
]
|
|
result = _build_traffic_table_rows(traffic)
|
|
|
|
assert result[0]["rx"] == "500"
|
|
assert result[0]["tx"] is None
|
|
|
|
def test_unrecognized_label_skipped(self):
|
|
"""Unrecognized labels are skipped."""
|
|
traffic = [
|
|
{"label": "Unknown", "value": "100", "raw_value": 100, "unit": "packets"},
|
|
{"label": "RX", "value": "500", "raw_value": 500, "unit": "packets"},
|
|
]
|
|
result = _build_traffic_table_rows(traffic)
|
|
|
|
assert len(result) == 1
|
|
assert result[0]["label"] == "Packets"
|
|
|
|
|
|
class TestBuildNodeDetails:
|
|
"""Test build_node_details function."""
|
|
|
|
def test_repeater_details(self, configured_env, monkeypatch):
|
|
"""Repeater node details include location info."""
|
|
monkeypatch.setenv("REPORT_LOCATION_SHORT", "Test Location")
|
|
monkeypatch.setenv("REPORT_LAT", "51.5074")
|
|
monkeypatch.setenv("REPORT_LON", "-0.1278")
|
|
monkeypatch.setenv("REPORT_ELEV", "11")
|
|
monkeypatch.setenv("REPORT_ELEV_UNIT", "m")
|
|
monkeypatch.setenv("REPEATER_HARDWARE", "RAK 4631")
|
|
|
|
# Reset config to pick up new values
|
|
import meshmon.env
|
|
meshmon.env._config = None
|
|
|
|
result = build_node_details("repeater")
|
|
|
|
labels = [d["label"] for d in result]
|
|
assert "Location" in labels
|
|
assert "Coordinates" in labels
|
|
assert "Elevation" in labels
|
|
assert "Hardware" in labels
|
|
|
|
# Check specific values
|
|
location = next(d for d in result if d["label"] == "Location")
|
|
assert location["value"] == "Test Location"
|
|
|
|
hardware = next(d for d in result if d["label"] == "Hardware")
|
|
assert hardware["value"] == "RAK 4631"
|
|
|
|
def test_companion_details(self, configured_env, monkeypatch):
|
|
"""Companion node details are simpler."""
|
|
monkeypatch.setenv("COMPANION_HARDWARE", "T-Beam Supreme")
|
|
|
|
# Reset config
|
|
import meshmon.env
|
|
meshmon.env._config = None
|
|
|
|
result = build_node_details("companion")
|
|
|
|
labels = [d["label"] for d in result]
|
|
assert "Hardware" in labels
|
|
assert "Connection" in labels
|
|
|
|
# No location info for companion
|
|
assert "Location" not in labels
|
|
assert "Coordinates" not in labels
|
|
|
|
def test_coordinate_directions(self, configured_env, monkeypatch):
|
|
"""Coordinate directions are correct for positive/negative."""
|
|
monkeypatch.setenv("REPORT_LAT", "-33.8688") # Sydney
|
|
monkeypatch.setenv("REPORT_LON", "151.2093")
|
|
|
|
import meshmon.env
|
|
meshmon.env._config = None
|
|
|
|
result = build_node_details("repeater")
|
|
coords = next(d for d in result if d["label"] == "Coordinates")
|
|
|
|
assert "S" in coords["value"] # South
|
|
assert "E" in coords["value"] # East
|
|
|
|
|
|
class TestBuildRadioConfig:
|
|
"""Test build_radio_config function."""
|
|
|
|
def test_returns_radio_settings(self, configured_env, monkeypatch):
|
|
"""Returns radio configuration from env."""
|
|
monkeypatch.setenv("RADIO_FREQUENCY", "915.0 MHz")
|
|
monkeypatch.setenv("RADIO_BANDWIDTH", "125 kHz")
|
|
monkeypatch.setenv("RADIO_SPREAD_FACTOR", "SF12")
|
|
monkeypatch.setenv("RADIO_CODING_RATE", "CR5")
|
|
|
|
import meshmon.env
|
|
meshmon.env._config = None
|
|
|
|
result = build_radio_config()
|
|
|
|
labels = [d["label"] for d in result]
|
|
assert "Frequency" in labels
|
|
assert "Bandwidth" in labels
|
|
assert "Spread Factor" in labels
|
|
assert "Coding Rate" in labels
|
|
|
|
freq = next(d for d in result if d["label"] == "Frequency")
|
|
assert freq["value"] == "915.0 MHz"
|
|
|
|
|
|
class TestBuildRepeaterMetrics:
|
|
"""Test build_repeater_metrics function."""
|
|
|
|
def test_none_row_returns_empty(self):
|
|
"""None row returns empty metric lists."""
|
|
result = build_repeater_metrics(None)
|
|
assert result["critical_metrics"] == []
|
|
assert result["secondary_metrics"] == []
|
|
assert result["traffic_metrics"] == []
|
|
|
|
def test_empty_row_returns_empty(self):
|
|
"""Empty row returns empty metric lists."""
|
|
result = build_repeater_metrics({})
|
|
assert result["critical_metrics"] == []
|
|
assert result["secondary_metrics"] == []
|
|
assert result["traffic_metrics"] == []
|
|
|
|
def test_full_row_extracts_metrics(self):
|
|
"""Full row extracts all metric categories."""
|
|
row = {
|
|
"bat": 3850.0,
|
|
"bat_pct": 55.0,
|
|
"last_rssi": -85.0,
|
|
"last_snr": 7.5,
|
|
"uptime": 86400,
|
|
"noise_floor": -115.0,
|
|
"tx_queue_len": 0,
|
|
"nb_recv": 1234,
|
|
"nb_sent": 567,
|
|
"recv_flood": 500,
|
|
"sent_flood": 200,
|
|
"recv_direct": 100,
|
|
"sent_direct": 50,
|
|
"airtime": 3600,
|
|
"rx_airtime": 7200,
|
|
}
|
|
result = build_repeater_metrics(row)
|
|
|
|
# Critical metrics
|
|
assert len(result["critical_metrics"]) >= 4
|
|
battery = next(m for m in result["critical_metrics"] if m["label"] == "Battery")
|
|
assert battery["value"] == "3.85"
|
|
assert battery["unit"] == "V"
|
|
|
|
# Secondary metrics
|
|
assert len(result["secondary_metrics"]) >= 2
|
|
|
|
# Traffic metrics
|
|
assert len(result["traffic_metrics"]) >= 8
|
|
|
|
def test_battery_converts_mv_to_v(self):
|
|
"""Battery value is converted from mV to V."""
|
|
row = {"bat": 4200.0, "bat_pct": 100.0}
|
|
result = build_repeater_metrics(row)
|
|
|
|
battery = next(m for m in result["critical_metrics"] if m["label"] == "Battery")
|
|
assert battery["value"] == "4.20"
|
|
|
|
def test_bar_pct_for_battery(self):
|
|
"""Battery has bar_pct for progress display."""
|
|
row = {"bat": 3850.0, "bat_pct": 55.0}
|
|
result = build_repeater_metrics(row)
|
|
|
|
battery = next(m for m in result["critical_metrics"] if m["label"] == "Battery")
|
|
assert battery["bar_pct"] == 55
|
|
|
|
|
|
class TestBuildCompanionMetrics:
|
|
"""Test build_companion_metrics function."""
|
|
|
|
def test_none_row_returns_empty(self):
|
|
"""None row returns empty metric lists."""
|
|
result = build_companion_metrics(None)
|
|
assert result["critical_metrics"] == []
|
|
assert result["secondary_metrics"] == []
|
|
assert result["traffic_metrics"] == []
|
|
|
|
def test_full_row_extracts_metrics(self):
|
|
"""Full row extracts all metric categories."""
|
|
row = {
|
|
"battery_mv": 3850.0,
|
|
"bat_pct": 55.0,
|
|
"contacts": 5,
|
|
"uptime_secs": 86400,
|
|
"recv": 1234,
|
|
"sent": 567,
|
|
}
|
|
result = build_companion_metrics(row)
|
|
|
|
# Critical metrics
|
|
assert len(result["critical_metrics"]) == 4 # Battery, Charge, Contacts, Uptime
|
|
|
|
# Secondary metrics (empty for companion)
|
|
assert result["secondary_metrics"] == []
|
|
|
|
# Traffic metrics
|
|
assert len(result["traffic_metrics"]) == 2 # RX and TX
|
|
|
|
def test_battery_converts_mv_to_v(self):
|
|
"""Battery value is converted from mV to V."""
|
|
row = {"battery_mv": 4000.0, "bat_pct": 80.0}
|
|
result = build_companion_metrics(row)
|
|
|
|
battery = next(m for m in result["critical_metrics"] if m["label"] == "Battery")
|
|
assert battery["value"] == "4.00"
|
|
|
|
def test_contacts_displays_correctly(self):
|
|
"""Contacts are displayed as integer."""
|
|
row = {"contacts": 7}
|
|
result = build_companion_metrics(row)
|
|
|
|
contacts = next(m for m in result["critical_metrics"] if m["label"] == "Contacts")
|
|
assert contacts["value"] == "7"
|
|
assert contacts["unit"] is None
|
|
|
|
|
|
class TestGetJinjaEnv:
|
|
"""Test get_jinja_env function."""
|
|
|
|
def test_returns_environment(self):
|
|
"""Returns a Jinja2 Environment."""
|
|
from jinja2 import Environment
|
|
|
|
# Reset singleton for clean test
|
|
import meshmon.html
|
|
meshmon.html._jinja_env = None
|
|
|
|
env = get_jinja_env()
|
|
assert isinstance(env, Environment)
|
|
|
|
def test_returns_singleton(self):
|
|
"""Returns same instance on subsequent calls."""
|
|
import meshmon.html
|
|
meshmon.html._jinja_env = None
|
|
|
|
env1 = get_jinja_env()
|
|
env2 = get_jinja_env()
|
|
assert env1 is env2
|
|
|
|
def test_registers_custom_filters(self):
|
|
"""Custom format filters are registered."""
|
|
import meshmon.html
|
|
meshmon.html._jinja_env = None
|
|
|
|
env = get_jinja_env()
|
|
|
|
assert "format_time" in env.filters
|
|
assert "format_value" in env.filters
|
|
assert "format_number" in env.filters
|
|
assert "format_duration" in env.filters
|
|
assert "format_uptime" in env.filters
|
|
assert "format_compact_number" in env.filters
|
|
assert "format_duration_compact" in env.filters
|
|
|
|
|
|
class TestChartGroupConstants:
|
|
"""Test chart group configuration constants."""
|
|
|
|
def test_repeater_chart_groups_defined(self):
|
|
"""Repeater chart groups are defined."""
|
|
assert len(REPEATER_CHART_GROUPS) > 0
|
|
for group in REPEATER_CHART_GROUPS:
|
|
assert "title" in group
|
|
assert "metrics" in group
|
|
assert len(group["metrics"]) > 0
|
|
|
|
def test_companion_chart_groups_defined(self):
|
|
"""Companion chart groups are defined."""
|
|
assert len(COMPANION_CHART_GROUPS) > 0
|
|
for group in COMPANION_CHART_GROUPS:
|
|
assert "title" in group
|
|
assert "metrics" in group
|
|
assert len(group["metrics"]) > 0
|
|
|
|
def test_period_config_defined(self):
|
|
"""Period config has all expected periods."""
|
|
assert "day" in PERIOD_CONFIG
|
|
assert "week" in PERIOD_CONFIG
|
|
assert "month" in PERIOD_CONFIG
|
|
assert "year" in PERIOD_CONFIG
|
|
|
|
for _period, (title, subtitle) in PERIOD_CONFIG.items():
|
|
assert isinstance(title, str)
|
|
assert isinstance(subtitle, str)
|