forked from genewildish/Mainline
MAJOR REFACTORING: Consolidate duplicated pipeline code and standardize on capability-based dependency resolution. This is a significant but backwards-compatible restructuring that improves maintainability and extensibility. ## ARCHITECTURE CHANGES ### Data Sources Consolidation - Move engine/sources_v2.py → engine/data_sources/sources.py - Move engine/pipeline_sources/ → engine/data_sources/ - Create unified DataSource ABC with common interface: * fetch() - idempotent data retrieval * get_items() - cached access with automatic refresh * refresh() - force cache invalidation * is_dynamic - indicate streaming vs static sources - Support for SourceItem dataclass (content, source, timestamp, metadata) ### Display Backend Improvements - Update all 7 display backends to use new import paths - Terminal: Improve dimension detection and handling - WebSocket: Better error handling and client lifecycle - Sixel: Refactor graphics rendering - Pygame: Modernize event handling - Kitty: Add protocol support for inline images - Multi: Ensure proper forwarding to all backends - Null: Maintain testing backend functionality ### Pipeline Adapter Consolidation - Refactor adapter stages for clarity and flexibility - RenderStage now handles both item-based and buffer-based rendering - Add SourceItemsToBufferStage for converting data source items - Improve DataSourceStage to work with all source types - Add DisplayStage wrapper for display backends ### Camera & Viewport Refinements - Update Camera class for new architecture - Improve viewport dimension detection - Better handling of resize events across backends ### New Effect Plugins - border.py: Frame rendering effect with configurable style - crop.py: Viewport clipping effect for selective display - tint.py: Color filtering effect for atmosphere ### Tests & Quality - Add test_border_effect.py with comprehensive border tests - Add test_crop_effect.py with viewport clipping tests - Add test_tint_effect.py with color filtering tests - Update test_pipeline.py for new architecture - Update test_pipeline_introspection.py for new data source location - All 463 tests pass with 56% coverage - Linting: All checks pass with ruff ### Removals (Code Cleanup) - Delete engine/benchmark.py (deprecated performance testing) - Delete engine/pipeline_sources/__init__.py (moved to data_sources) - Delete engine/sources_v2.py (replaced by data_sources/sources.py) - Update AGENTS.md to reflect new structure ### Import Path Updates - Update engine/pipeline/controller.py::create_default_pipeline() * Old: from engine.sources_v2 import HeadlinesDataSource * New: from engine.data_sources.sources import HeadlinesDataSource - All display backends import from new locations - All tests import from new locations ## BACKWARDS COMPATIBILITY This refactoring is intended to be backwards compatible: - Pipeline execution unchanged (DAG-based with capability matching) - Effect plugins unchanged (EffectPlugin interface same) - Display protocol unchanged (Display duck-typing works as before) - Config system unchanged (presets.toml format same) ## TESTING - 463 tests pass (0 failures, 19 skipped) - Full linting check passes - Manual testing on demo, poetry, websocket modes - All new effect plugins tested ## FILES CHANGED - 24 files modified/added/deleted - 723 insertions, 1,461 deletions (net -738 LOC - cleanup!) - No breaking changes to public APIs - All transitive imports updated correctly
537 lines
18 KiB
Python
537 lines
18 KiB
Python
"""
|
|
Pipeline controller - DAG-based pipeline execution.
|
|
|
|
The Pipeline class orchestrates stages in dependency order, handling
|
|
the complete render cycle from source to display.
|
|
"""
|
|
|
|
import time
|
|
from dataclasses import dataclass, field
|
|
from typing import Any
|
|
|
|
from engine.pipeline.core import PipelineContext, Stage, StageError, StageResult
|
|
from engine.pipeline.params import PipelineParams
|
|
from engine.pipeline.registry import StageRegistry
|
|
|
|
|
|
@dataclass
|
|
class PipelineConfig:
|
|
"""Configuration for a pipeline instance."""
|
|
|
|
source: str = "headlines"
|
|
display: str = "terminal"
|
|
camera: str = "vertical"
|
|
effects: list[str] = field(default_factory=list)
|
|
enable_metrics: bool = True
|
|
|
|
|
|
@dataclass
|
|
class StageMetrics:
|
|
"""Metrics for a single stage execution."""
|
|
|
|
name: str
|
|
duration_ms: float
|
|
chars_in: int = 0
|
|
chars_out: int = 0
|
|
|
|
|
|
@dataclass
|
|
class FrameMetrics:
|
|
"""Metrics for a single frame through the pipeline."""
|
|
|
|
frame_number: int
|
|
total_ms: float
|
|
stages: list[StageMetrics] = field(default_factory=list)
|
|
|
|
|
|
class Pipeline:
|
|
"""Main pipeline orchestrator.
|
|
|
|
Manages the execution of all stages in dependency order,
|
|
handling initialization, processing, and cleanup.
|
|
"""
|
|
|
|
def __init__(
|
|
self,
|
|
config: PipelineConfig | None = None,
|
|
context: PipelineContext | None = None,
|
|
):
|
|
self.config = config or PipelineConfig()
|
|
self.context = context or PipelineContext()
|
|
self._stages: dict[str, Stage] = {}
|
|
self._execution_order: list[str] = []
|
|
self._initialized = False
|
|
|
|
self._metrics_enabled = self.config.enable_metrics
|
|
self._frame_metrics: list[FrameMetrics] = []
|
|
self._max_metrics_frames = 60
|
|
self._current_frame_number = 0
|
|
|
|
def add_stage(self, name: str, stage: Stage) -> "Pipeline":
|
|
"""Add a stage to the pipeline."""
|
|
self._stages[name] = stage
|
|
return self
|
|
|
|
def remove_stage(self, name: str) -> None:
|
|
"""Remove a stage from the pipeline."""
|
|
if name in self._stages:
|
|
del self._stages[name]
|
|
|
|
def get_stage(self, name: str) -> Stage | None:
|
|
"""Get a stage by name."""
|
|
return self._stages.get(name)
|
|
|
|
def build(self) -> "Pipeline":
|
|
"""Build execution order based on dependencies."""
|
|
self._capability_map = self._build_capability_map()
|
|
self._execution_order = self._resolve_dependencies()
|
|
self._validate_dependencies()
|
|
self._validate_types()
|
|
self._initialized = True
|
|
return self
|
|
|
|
def _build_capability_map(self) -> dict[str, list[str]]:
|
|
"""Build a map of capabilities to stage names.
|
|
|
|
Returns:
|
|
Dict mapping capability -> list of stage names that provide it
|
|
"""
|
|
capability_map: dict[str, list[str]] = {}
|
|
for name, stage in self._stages.items():
|
|
for cap in stage.capabilities:
|
|
if cap not in capability_map:
|
|
capability_map[cap] = []
|
|
capability_map[cap].append(name)
|
|
return capability_map
|
|
|
|
def _find_stage_with_capability(self, capability: str) -> str | None:
|
|
"""Find a stage that provides the given capability.
|
|
|
|
Supports wildcard matching:
|
|
- "source" matches "source.headlines" (prefix match)
|
|
- "source.*" matches "source.headlines"
|
|
- "source.headlines" matches exactly
|
|
|
|
Args:
|
|
capability: The capability to find
|
|
|
|
Returns:
|
|
Stage name that provides the capability, or None if not found
|
|
"""
|
|
# Exact match
|
|
if capability in self._capability_map:
|
|
return self._capability_map[capability][0]
|
|
|
|
# Prefix match (e.g., "source" -> "source.headlines")
|
|
for cap, stages in self._capability_map.items():
|
|
if cap.startswith(capability + "."):
|
|
return stages[0]
|
|
|
|
# Wildcard match (e.g., "source.*" -> "source.headlines")
|
|
if ".*" in capability:
|
|
prefix = capability[:-2] # Remove ".*"
|
|
for cap in self._capability_map:
|
|
if cap.startswith(prefix + "."):
|
|
return self._capability_map[cap][0]
|
|
|
|
return None
|
|
|
|
def _resolve_dependencies(self) -> list[str]:
|
|
"""Resolve stage execution order using topological sort with capability matching."""
|
|
ordered = []
|
|
visited = set()
|
|
temp_mark = set()
|
|
|
|
def visit(name: str) -> None:
|
|
if name in temp_mark:
|
|
raise StageError(name, "Circular dependency detected")
|
|
if name in visited:
|
|
return
|
|
|
|
temp_mark.add(name)
|
|
stage = self._stages.get(name)
|
|
if stage:
|
|
for dep in stage.dependencies:
|
|
# Find a stage that provides this capability
|
|
dep_stage_name = self._find_stage_with_capability(dep)
|
|
if dep_stage_name:
|
|
visit(dep_stage_name)
|
|
|
|
temp_mark.remove(name)
|
|
visited.add(name)
|
|
ordered.append(name)
|
|
|
|
for name in self._stages:
|
|
if name not in visited:
|
|
visit(name)
|
|
|
|
return ordered
|
|
|
|
def _validate_dependencies(self) -> None:
|
|
"""Validate that all dependencies can be satisfied.
|
|
|
|
Raises StageError if any dependency cannot be resolved.
|
|
"""
|
|
missing: list[tuple[str, str]] = [] # (stage_name, capability)
|
|
|
|
for name, stage in self._stages.items():
|
|
for dep in stage.dependencies:
|
|
if not self._find_stage_with_capability(dep):
|
|
missing.append((name, dep))
|
|
|
|
if missing:
|
|
msgs = [f" - {stage} needs {cap}" for stage, cap in missing]
|
|
raise StageError(
|
|
"validation",
|
|
"Missing capabilities:\n" + "\n".join(msgs),
|
|
)
|
|
|
|
def _validate_types(self) -> None:
|
|
"""Validate inlet/outlet types between connected stages.
|
|
|
|
PureData-style type validation. Each stage declares its inlet_types
|
|
(what it accepts) and outlet_types (what it produces). This method
|
|
validates that connected stages have compatible types.
|
|
|
|
Raises StageError if type mismatch is detected.
|
|
"""
|
|
from engine.pipeline.core import DataType
|
|
|
|
errors: list[str] = []
|
|
|
|
for i, name in enumerate(self._execution_order):
|
|
stage = self._stages.get(name)
|
|
if not stage:
|
|
continue
|
|
|
|
inlet_types = stage.inlet_types
|
|
|
|
# Check against previous stage's outlet types
|
|
if i > 0:
|
|
prev_name = self._execution_order[i - 1]
|
|
prev_stage = self._stages.get(prev_name)
|
|
if prev_stage:
|
|
prev_outlets = prev_stage.outlet_types
|
|
|
|
# Check if any outlet type is accepted by this inlet
|
|
compatible = (
|
|
DataType.ANY in inlet_types
|
|
or DataType.ANY in prev_outlets
|
|
or bool(prev_outlets & inlet_types)
|
|
)
|
|
|
|
if not compatible:
|
|
errors.append(
|
|
f" - {name} (inlet: {inlet_types}) "
|
|
f"← {prev_name} (outlet: {prev_outlets})"
|
|
)
|
|
|
|
# Check display/sink stages (should accept TEXT_BUFFER)
|
|
if (
|
|
stage.category == "display"
|
|
and DataType.TEXT_BUFFER not in inlet_types
|
|
and DataType.ANY not in inlet_types
|
|
):
|
|
errors.append(f" - {name} is display but doesn't accept TEXT_BUFFER")
|
|
|
|
if errors:
|
|
raise StageError(
|
|
"type_validation",
|
|
"Type mismatch in pipeline connections:\n" + "\n".join(errors),
|
|
)
|
|
|
|
def initialize(self) -> bool:
|
|
"""Initialize all stages in execution order."""
|
|
for name in self._execution_order:
|
|
stage = self._stages.get(name)
|
|
if stage and not stage.init(self.context) and not stage.optional:
|
|
return False
|
|
return True
|
|
|
|
def execute(self, data: Any | None = None) -> StageResult:
|
|
"""Execute the pipeline with the given input data.
|
|
|
|
Pipeline execution:
|
|
1. Execute all non-overlay stages in dependency order
|
|
2. Apply overlay stages on top (sorted by render_order)
|
|
"""
|
|
if not self._initialized:
|
|
self.build()
|
|
|
|
if not self._initialized:
|
|
return StageResult(
|
|
success=False,
|
|
data=None,
|
|
error="Pipeline not initialized",
|
|
)
|
|
|
|
current_data = data
|
|
frame_start = time.perf_counter() if self._metrics_enabled else 0
|
|
stage_timings: list[StageMetrics] = []
|
|
|
|
# Separate overlay stages from regular stages
|
|
overlay_stages: list[tuple[int, Stage]] = []
|
|
regular_stages: list[str] = []
|
|
|
|
for name in self._execution_order:
|
|
stage = self._stages.get(name)
|
|
if not stage or not stage.is_enabled():
|
|
continue
|
|
|
|
# Safely check is_overlay - handle MagicMock and other non-bool returns
|
|
try:
|
|
is_overlay = bool(getattr(stage, "is_overlay", False))
|
|
except Exception:
|
|
is_overlay = False
|
|
|
|
if is_overlay:
|
|
# Safely get render_order
|
|
try:
|
|
render_order = int(getattr(stage, "render_order", 0))
|
|
except Exception:
|
|
render_order = 0
|
|
overlay_stages.append((render_order, stage))
|
|
else:
|
|
regular_stages.append(name)
|
|
|
|
# Execute regular stages in dependency order
|
|
for name in regular_stages:
|
|
stage = self._stages.get(name)
|
|
if not stage or not stage.is_enabled():
|
|
continue
|
|
|
|
stage_start = time.perf_counter() if self._metrics_enabled else 0
|
|
|
|
try:
|
|
current_data = stage.process(current_data, self.context)
|
|
except Exception as e:
|
|
if not stage.optional:
|
|
return StageResult(
|
|
success=False,
|
|
data=current_data,
|
|
error=str(e),
|
|
stage_name=name,
|
|
)
|
|
continue
|
|
|
|
if self._metrics_enabled:
|
|
stage_duration = (time.perf_counter() - stage_start) * 1000
|
|
chars_in = len(str(data)) if data else 0
|
|
chars_out = len(str(current_data)) if current_data else 0
|
|
stage_timings.append(
|
|
StageMetrics(
|
|
name=name,
|
|
duration_ms=stage_duration,
|
|
chars_in=chars_in,
|
|
chars_out=chars_out,
|
|
)
|
|
)
|
|
|
|
# Apply overlay stages (sorted by render_order)
|
|
overlay_stages.sort(key=lambda x: x[0])
|
|
for render_order, stage in overlay_stages:
|
|
stage_start = time.perf_counter() if self._metrics_enabled else 0
|
|
stage_name = f"[overlay]{stage.name}"
|
|
|
|
try:
|
|
# Overlays receive current_data but don't pass their output to next stage
|
|
# Instead, their output is composited on top
|
|
overlay_output = stage.process(current_data, self.context)
|
|
# For now, we just let the overlay output pass through
|
|
# In a more sophisticated implementation, we'd composite it
|
|
if overlay_output is not None:
|
|
current_data = overlay_output
|
|
except Exception as e:
|
|
if not stage.optional:
|
|
return StageResult(
|
|
success=False,
|
|
data=current_data,
|
|
error=str(e),
|
|
stage_name=stage_name,
|
|
)
|
|
|
|
if self._metrics_enabled:
|
|
stage_duration = (time.perf_counter() - stage_start) * 1000
|
|
chars_in = len(str(data)) if data else 0
|
|
chars_out = len(str(current_data)) if current_data else 0
|
|
stage_timings.append(
|
|
StageMetrics(
|
|
name=stage_name,
|
|
duration_ms=stage_duration,
|
|
chars_in=chars_in,
|
|
chars_out=chars_out,
|
|
)
|
|
)
|
|
|
|
if self._metrics_enabled:
|
|
total_duration = (time.perf_counter() - frame_start) * 1000
|
|
self._frame_metrics.append(
|
|
FrameMetrics(
|
|
frame_number=self._current_frame_number,
|
|
total_ms=total_duration,
|
|
stages=stage_timings,
|
|
)
|
|
)
|
|
|
|
# Store metrics in context for other stages (like HUD)
|
|
# This makes metrics a first-class pipeline citizen
|
|
if self.context:
|
|
self.context.state["metrics"] = self.get_metrics_summary()
|
|
|
|
if len(self._frame_metrics) > self._max_metrics_frames:
|
|
self._frame_metrics.pop(0)
|
|
self._current_frame_number += 1
|
|
|
|
return StageResult(success=True, data=current_data)
|
|
|
|
def cleanup(self) -> None:
|
|
"""Clean up all stages in reverse order."""
|
|
for name in reversed(self._execution_order):
|
|
stage = self._stages.get(name)
|
|
if stage:
|
|
try:
|
|
stage.cleanup()
|
|
except Exception:
|
|
pass
|
|
self._stages.clear()
|
|
self._initialized = False
|
|
|
|
@property
|
|
def stages(self) -> dict[str, Stage]:
|
|
"""Get all stages."""
|
|
return self._stages.copy()
|
|
|
|
@property
|
|
def execution_order(self) -> list[str]:
|
|
"""Get execution order."""
|
|
return self._execution_order.copy()
|
|
|
|
def get_stage_names(self) -> list[str]:
|
|
"""Get list of stage names."""
|
|
return list(self._stages.keys())
|
|
|
|
def get_overlay_stages(self) -> list[Stage]:
|
|
"""Get all overlay stages sorted by render_order."""
|
|
overlays = [stage for stage in self._stages.values() if stage.is_overlay]
|
|
overlays.sort(key=lambda s: s.render_order)
|
|
return overlays
|
|
|
|
def get_stage_type(self, name: str) -> str:
|
|
"""Get the stage_type for a stage."""
|
|
stage = self._stages.get(name)
|
|
return stage.stage_type if stage else ""
|
|
|
|
def get_render_order(self, name: str) -> int:
|
|
"""Get the render_order for a stage."""
|
|
stage = self._stages.get(name)
|
|
return stage.render_order if stage else 0
|
|
|
|
def get_metrics_summary(self) -> dict:
|
|
"""Get summary of collected metrics."""
|
|
if not self._frame_metrics:
|
|
return {"error": "No metrics collected"}
|
|
|
|
total_times = [f.total_ms for f in self._frame_metrics]
|
|
avg_total = sum(total_times) / len(total_times)
|
|
min_total = min(total_times)
|
|
max_total = max(total_times)
|
|
|
|
stage_stats: dict[str, dict] = {}
|
|
for frame in self._frame_metrics:
|
|
for stage in frame.stages:
|
|
if stage.name not in stage_stats:
|
|
stage_stats[stage.name] = {"times": [], "total_chars": 0}
|
|
stage_stats[stage.name]["times"].append(stage.duration_ms)
|
|
stage_stats[stage.name]["total_chars"] += stage.chars_out
|
|
|
|
for name, stats in stage_stats.items():
|
|
times = stats["times"]
|
|
stats["avg_ms"] = sum(times) / len(times)
|
|
stats["min_ms"] = min(times)
|
|
stats["max_ms"] = max(times)
|
|
del stats["times"]
|
|
|
|
return {
|
|
"frame_count": len(self._frame_metrics),
|
|
"pipeline": {
|
|
"avg_ms": avg_total,
|
|
"min_ms": min_total,
|
|
"max_ms": max_total,
|
|
},
|
|
"stages": stage_stats,
|
|
}
|
|
|
|
def reset_metrics(self) -> None:
|
|
"""Reset collected metrics."""
|
|
self._frame_metrics.clear()
|
|
self._current_frame_number = 0
|
|
|
|
def get_frame_times(self) -> list[float]:
|
|
"""Get historical frame times for sparklines/charts."""
|
|
return [f.total_ms for f in self._frame_metrics]
|
|
|
|
|
|
class PipelineRunner:
|
|
"""High-level pipeline runner with animation support."""
|
|
|
|
def __init__(
|
|
self,
|
|
pipeline: Pipeline,
|
|
params: PipelineParams | None = None,
|
|
):
|
|
self.pipeline = pipeline
|
|
self.params = params or PipelineParams()
|
|
self._running = False
|
|
|
|
def start(self) -> bool:
|
|
"""Start the pipeline."""
|
|
self._running = True
|
|
return self.pipeline.initialize()
|
|
|
|
def step(self, input_data: Any | None = None) -> Any:
|
|
"""Execute one pipeline step."""
|
|
self.params.frame_number += 1
|
|
self.pipeline.context.params = self.params
|
|
result = self.pipeline.execute(input_data)
|
|
return result.data if result.success else None
|
|
|
|
def stop(self) -> None:
|
|
"""Stop and clean up the pipeline."""
|
|
self._running = False
|
|
self.pipeline.cleanup()
|
|
|
|
@property
|
|
def is_running(self) -> bool:
|
|
"""Check if runner is active."""
|
|
return self._running
|
|
|
|
|
|
def create_pipeline_from_params(params: PipelineParams) -> Pipeline:
|
|
"""Create a pipeline from PipelineParams."""
|
|
config = PipelineConfig(
|
|
source=params.source,
|
|
display=params.display,
|
|
camera=params.camera_mode,
|
|
effects=params.effect_order,
|
|
)
|
|
return Pipeline(config=config)
|
|
|
|
|
|
def create_default_pipeline() -> Pipeline:
|
|
"""Create a default pipeline with all standard components."""
|
|
from engine.data_sources.sources import HeadlinesDataSource
|
|
from engine.pipeline.adapters import DataSourceStage
|
|
|
|
pipeline = Pipeline()
|
|
|
|
# Add source stage (wrapped as Stage)
|
|
source = HeadlinesDataSource()
|
|
pipeline.add_stage("source", DataSourceStage(source, name="headlines"))
|
|
|
|
# Add display stage
|
|
display = StageRegistry.create("display", "terminal")
|
|
if display:
|
|
pipeline.add_stage("display", display)
|
|
|
|
return pipeline.build()
|