Geofuchs Hydra-8
Enterprise GeoAI Platform
Development Roadmap
Version 3.0
Espoo, Finland
Hetzner Cloud
This comprehensive roadmap charts the transformation of Geofuchs from a Finnish-focused multi-agent GIS platform into an enterprise-grade geospatial AI foundation system. Over the course of three years, we will systematically close critical capability gaps, implement cutting-edge remote sensing technologies, and build autonomous agent systems that rival leading commercial platforms. The journey begins with immediate infrastructure migration to unlock advanced AI capabilities and culminates in a fully-featured platform serving multiple industry verticals across Europe.
The roadmap balances ambitious technical goals with pragmatic execution, allocating resources strategically across remote sensing development and agent intelligence improvements. Each quarter builds upon previous achievements whilst maintaining strict data residency requirements within Finland. Our phased approach ensures continuous value delivery whilst managing technical risk through proven methodologies and comprehensive testing protocols.
3-Year Transformation Journey
From Regional Platform to Enterprise GeoAI Leader
This detailed roadmap outlines Geofuchs' strategic transformation over three years, evolving from its current regional multi-agent GIS platform into a leading enterprise-grade geospatial AI foundation system. Each phase builds upon the previous, focusing on critical capability enhancements and technological advancements.
Current State (2025)
  • 8-agent multi-agent system
  • Finnish data integration
  • Qwen 7B with LoRA
  • Basic GIS capabilities
Year 1 (2026): Foundation & Remote Sensing
  • Infrastructure migration (RTX 6000 Ada)
  • Remote sensing pipeline
  • SAMGeo integration
  • Vision-language models
Investment: €386,400
Year 2 (2027): Intelligence & Autonomy
  • Feedback loops & active learning
  • GraphRAG knowledge graphs
  • Autonomous agents with reflection
  • MCP toolkit integration
Investment: €469,660
Year 3 (2028): Enterprise Scale
  • XAI & compliance frameworks
  • Generative capabilities
  • Industry verticals (Urban, Infrastructure, Environmental)
  • Multi-region deployment
Investment: €623,300
Target State (2028)
  • Full-featured GeoAI platform
  • Feature parity with commercial leaders
  • European data residency
  • 3 industry verticals
Total Investment: €1,479,360
Executive Summary
Current State
Functional 8-agent multi-agent system with Finnish data integration, LangGraph orchestration, and Qwen 7B with LoRA fine-tuning capabilities
Target State
Full-featured GeoAI platform with remote sensing, knowledge graphs, autonomous agents, and MCP toolkit integration
Investment
€1,479,360 total investment over 3 years with high-availability infrastructure revision
This transformation represents a strategic evolution from a capable regional platform to an enterprise-grade solution that can compete with leading commercial offerings. The revised budget reflects enhanced infrastructure requirements for high availability and the addition of three industry-specific verticals: Urban Planning, Infrastructure, and Environmental monitoring. These verticals leverage Finland's exceptional geospatial data resources from organisations like SYKE and GTK.
Our key milestone targets feature parity with leading platforms by Q4 2027, positioning Geofuchs as a competitive alternative whilst maintaining strict European data residency and privacy standards. The roadmap prioritises closing critical gaps in remote sensing capabilities during Year 1, building autonomous agent intelligence in Year 2, and achieving enterprise scale with generative capabilities in Year 3. This phased approach ensures sustainable growth whilst managing technical complexity and resource allocation effectively.
Gap Analysis: Current System vs Target Platform
A comprehensive assessment reveals significant capability gaps between Geofuchs Hydra-8 and leading enterprise geospatial AI platforms. The most critical deficiencies lie in remote sensing image processing, where we currently lack any capability for automated analysis of satellite or aerial imagery. Similarly, advanced AI interpretation using models like G-SAM and LIM is entirely absent, representing a fundamental limitation for modern geospatial applications.
Critical Gaps (P0)
  • Remote sensing image processing - no current capability
  • AI interpretation with G-SAM/LIM equivalents - completely missing
  • These represent immediate blockers for competitive positioning
High Priority Gaps (P1-P2)
  • Spatial knowledge graph - basic Neo4j implementation exists but lacks depth
  • MCP toolkit - no standardised tool integration framework
  • Vision pipeline - planned for Q4 2026 but not yet implemented
Medium Priority Gaps (P2-P3)
  • AgentX autonomous agents - current agents are rule-based, not truly autonomous
  • AI spatial analysis - basic capabilities exist but lack sophistication
  • Vector-to-image generation - no generative capabilities
Lower Priority Gaps (P3)
  • AI indoor mapping - niche capability for future consideration
  • Industry-specific agents - planned for Year 3 vertical development
The analysis reveals that whilst Geofuchs has established a solid foundation with its 8-agent architecture and Finnish data integration, the platform requires substantial investment in computer vision, autonomous reasoning, and enterprise-grade tooling to achieve competitive parity. The roadmap systematically addresses these gaps through a balanced approach that interleaves remote sensing development with agent intelligence improvements, ensuring continuous capability enhancement across all critical dimensions.
Infrastructure Migration
48GB
GPU Memory
RTX 6000 Ada enables unquantized 70B models
256GB
System RAM
DDR5 ECC for full graph topology in memory
24
CPU Cores
Xeon Gold 5412U for accelerated processing
The current GEX44 hardware configuration represents a critical bottleneck that fundamentally limits our ability to implement advanced AI capabilities. With only 20GB of GPU memory, we are forced to use quantized models that introduce artifacts and reduce reasoning quality. The 64GB system RAM constrains our ability to maintain large knowledge graphs in memory, forcing expensive disk operations that slow query response times. Most critically, the limited GPU memory prevents us from running unquantized Llama-3-70B models, which are essential for implementing sophisticated "System 2" thinking and complex multi-step planning.
Migration to the GEX131 configuration addresses these limitations comprehensively. The RTX 6000 Ada GPU with 48GB (or optionally 96GB) of VRAM enables us to run full-precision large language models, supporting larger batch sizes for parallel agent execution and eliminating quantization artifacts that degrade output quality. The Xeon Gold 5412U processor with 24 cores dramatically accelerates Qdrant vector indexing operations and Neo4j graph traversal queries. The 256GB DDR5 ECC memory allows our entire knowledge graph topology to reside in memory, eliminating disk I/O bottlenecks during complex spatial reasoning tasks.
A critical constraint governs this migration: all data must remain within Finland to comply with data residency requirements. The HEL1 data centre in Espoo provides the necessary geographic location, but GEX131 availability in Helsinki is limited. Our contingency plan requires immediate contact with Hetzner Enterprise Sales if standard rental stock is unavailable, exploring options for custom placement or colocation arrangements. This migration represents the foundational investment that unlocks all subsequent roadmap capabilities, making it the highest priority action item for Q1 2026.
Zero-Downtime Migration Strategy
01
Provisioning & Networking
Secure GEX131 instance in HEL1, configure Hetzner vSwitch for 10Gbps private LAN bridging GEX44 and GEX131, provision Debian 12 with Docker and NVIDIA Container Toolkit
02
Data Replication
Mirror qdrant_storage vectors and neo4j_data graph volumes over private vSwitch using ZFS/Rsync, download full FP16 model weights removing quantization artifacts
03
Blue/Green Switchover
Start full stack on GEX131, execute comprehensive benchmark suite, update DNS to GEX131 static IP, secure wipe and decommission GEX44 after 48-hour stability period
The migration strategy employs a blue/green deployment pattern to ensure zero downtime during the transition. Week 1 focuses on procurement and networking, establishing the private 10Gbps VLAN that enables high-speed data transfer between the old and new infrastructure. This private network is essential for efficient replication of our multi-terabyte vector database and knowledge graph without impacting production traffic or incurring data transfer costs.
Week 2 executes the data replication phase, using ZFS snapshots and rsync to create consistent copies of all persistent storage volumes. We deliberately download full FP16 model weights rather than transferring quantized versions, as this represents an opportunity to eliminate the quality degradation introduced by quantization. The replication process runs continuously, maintaining synchronisation until the final cutover moment.
Week 3 implements the actual switchover using a blue/green pattern. We start the complete application stack on GEX131 whilst the GEX44 system continues serving production traffic. The 08-PERFORMANCE.md benchmark suite validates that the new infrastructure meets or exceeds performance targets across all critical metrics: query latency, agent response time, vector search speed, and graph traversal performance. Only after successful benchmark completion do we update DNS records to point api.geofuchs.app to the GEX131 static IP address.
The old GEX44 system remains operational for 48 hours post-cutover, serving as an immediate rollback option if unexpected issues emerge. After confirming stability, we perform a secure wipe of all data volumes and cancel the GEX44 contract, completing the migration with zero data loss and minimal risk exposure.
Year 1: Foundation & Remote Sensing (2026)
Foundation Year
Year 1 establishes the critical foundations for enterprise GeoAI capabilities, with primary focus on closing the remote sensing gap that represents our most significant competitive disadvantage. The year begins with infrastructure migration and builds progressively through raster data processing, image segmentation, pre-trained model integration, and vision-language understanding. By year-end, Geofuchs will possess comprehensive remote sensing capabilities that match or exceed commercial platforms.
1
Q1 2026
Infrastructure migration, GDAL/Rasterio stack, Sentinel Hub integration, tile processing pipeline
2
Q2 2026
SAMGeo segmentation, prompt engineering layer, long-term memory, hierarchical planner
3
Q3 2026
TorchGeo models, inference service, training pipeline, council supervisor
4
Q4 2026
CLIP/SigLIP integration, image queries, object detection, RS visualization UI
The quarterly progression follows a deliberate architectural pattern. Q1 establishes data ingestion capabilities, enabling us to acquire and process imagery from Copernicus Sentinel-2, Sentinel-1, Landsat, and commercial aerial sources. Q2 adds zero-shot segmentation through SAMGeo, allowing automated extraction of buildings, roads, water bodies, and vegetation without training data. Q3 integrates pre-trained models from TorchGeo for land cover classification, building extraction, and change detection. Q4 completes the vision pipeline with multimodal understanding through CLIP/SigLIP, enabling natural language queries about imagery content.
Parallel to remote sensing development, Year 1 advances agent capabilities through webhook integrations, persistent memory systems, hierarchical planning, council-based supervision, and parallel data fetching. This balanced approach ensures that as we add remote sensing capabilities, our agent infrastructure evolves to effectively utilise these new tools. The 54% remote sensing / 46% agent focus reflects the foundational nature of Year 1, where establishing core capabilities takes precedence over advanced autonomy.
Q1 2026: Infrastructure Migration + RS Pipeline Foundation
1
GEX131 Migration
Migrate to RTX 6000 Ada (48GB VRAM) following zero-downtime blue/green deployment strategy
120 hours | €9,000
2
GDAL/Rasterio Stack
Integrate GDAL 3.8+ and Rasterio for comprehensive multi-format raster I/O operations
160 hours | €12,000
3
Sentinel Hub Integration
Connect to Copernicus Sentinel-2/Sentinel-1 APIs for satellite imagery acquisition
120 hours | €9,000
4
Tile Processing Pipeline
Build COG (Cloud Optimised GeoTIFF) ingestion pipeline for efficient raster handling
200 hours | €15,000
5
Original Q1 Tasks
Webhooks, Context Manager, Tool Registry from original roadmap
280 hours | €21,000
Q1 2026 represents the most critical quarter in the entire roadmap, as it establishes the infrastructure and data pipeline foundations upon which all subsequent capabilities depend. The GEX131 migration unlocks the computational capacity required for unquantized large language models and enables the GPU-accelerated image processing that remote sensing demands. Without this migration, the entire roadmap would be constrained by hardware limitations that fundamentally compromise AI quality and performance.
The GDAL/Rasterio stack integration provides the low-level raster manipulation capabilities essential for professional geospatial work. GDAL 3.8+ supports over 200 raster formats and provides the coordinate transformation, resampling, and format conversion operations that form the backbone of any serious remote sensing pipeline. Rasterio adds a Pythonic interface that integrates naturally with our existing codebase whilst maintaining GDAL's performance characteristics.
Sentinel Hub integration connects Geofuchs to the European Union's Copernicus programme, providing access to petabytes of freely available satellite imagery. Sentinel-2 offers 10-metre resolution multispectral imagery ideal for land cover analysis, whilst Sentinel-1 synthetic aperture radar penetrates clouds and darkness. This integration transforms Geofuchs from a vector-focused platform to one capable of comprehensive Earth observation analysis.
The Cloud Optimised GeoTIFF (COG) pipeline represents a critical architectural decision that enables efficient handling of massive raster datasets. COGs support HTTP range requests, allowing clients to fetch only the specific tiles and resolution levels they need without downloading entire images. This dramatically reduces bandwidth requirements and enables responsive web-based visualisation of terabyte-scale imagery collections. The pipeline includes automatic COG conversion for ingested imagery, ensuring consistent performance characteristics across all data sources.
880
Total Hours
Development effort for Q1 2026
€66K
Quarter Cost
Direct development investment
5
Major Deliverables
Infrastructure + pipeline components
Q2 2026: Segment Anything Integration + Memory Systems
Q2 2026 introduces zero-shot image segmentation capabilities through SAMGeo, a geospatially-aware adaptation of Meta's Segment Anything Model. This represents a transformative capability that enables automated extraction of geographic features from imagery without requiring training data or manual annotation. Buildings, roads, water bodies, vegetation, and other features can be identified and vectorised through simple point, box, or polygon prompts, dramatically reducing the manual effort traditionally required for feature extraction.
SAMGeo Integration
Integrate segment-geospatial for zero-shot feature extraction from remote sensing imagery
160 hours | €12,000
Prompt Engineering Layer
Build intuitive UI for point, box, and polygon prompts with real-time preview
120 hours | €9,000
Mask Post-Processing
CRS transformation, vectorisation, and attribute extraction for segmentation outputs
160 hours | €12,000
The prompt engineering layer provides an intuitive interface for users to guide the segmentation process. Rather than requiring extensive training data or complex configuration, users simply indicate what they want to extract through visual prompts. A point prompt identifies a single feature, a box prompt captures all features within a region, and a polygon prompt enables precise boundary definition for complex areas. The system provides real-time preview of segmentation results, allowing iterative refinement until the desired output is achieved.
Mask post-processing transforms raw segmentation outputs into production-ready geospatial data. The pipeline handles coordinate reference system (CRS) transformation to ensure outputs align with existing datasets, vectorises raster masks into polygon geometries suitable for GIS analysis, and extracts relevant attributes such as area, perimeter, and shape metrics. This automated processing eliminates manual cleanup work and ensures consistent data quality across all segmentation operations.
Memory & Planning Capabilities
Parallel to segmentation development, Q2 implements long-term memory through Qdrant vector storage, enabling agents to recall and learn from historical interactions. The hierarchical planner deploys a dedicated Planner Agent that decomposes complex queries into executable sub-tasks, coordinating multiple specialist agents to achieve sophisticated goals that exceed individual agent capabilities.
  • Long-term memory: 200 hours | €15,000
  • Hierarchical planner: 200 hours | €15,000
Q2 Investment Summary
840
Total Hours
€63K
Quarter Cost
Q3 2026: TorchGeo Models + Supervisor Logic
Q3 2026 integrates pre-trained models from the TorchGeo library, providing production-ready capabilities for land cover classification, building extraction, and change detection without requiring extensive training data collection. TorchGeo offers over 10 pre-trained models specifically designed for remote sensing applications, trained on massive datasets that capture global geographic diversity. These models deliver immediate value whilst our custom training pipeline develops Finnish-specific adaptations.
TorchGeo Pre-trained Models
Integrate 10+ pre-trained RS models for land cover, building extraction, and change detection
200 hours | €15,000
Model Inference Service
Dockerised inference endpoint with GPU batching for efficient processing
160 hours | €12,000
Training Data Pipeline
Annotation workflow for Finnish aerial imagery adaptation
240 hours | €18,000
The model inference service architecture employs Docker containerisation to ensure consistent deployment across development and production environments. GPU batching optimises throughput by processing multiple images simultaneously, dramatically reducing per-image inference time compared to sequential processing. The service exposes a RESTful API that integrates seamlessly with our existing agent architecture, allowing agents to request model predictions as part of their reasoning workflows.
The training data pipeline establishes the infrastructure for continuous model improvement through Finnish-specific fine-tuning. Whilst TorchGeo's pre-trained models provide excellent baseline performance, adaptation to Finnish landscapes, building styles, and seasonal variations improves accuracy for our target market. The annotation workflow supports multiple labelling formats and includes quality control mechanisms to ensure training data consistency. This pipeline becomes increasingly valuable in Year 2 as we implement active learning and automated retraining capabilities.
Council Supervisor & Error Handling
Q3 introduces the Council Supervisor pattern, implementing dynamic task routing that selects the most appropriate agent for each sub-task based on capability matching and historical performance. When multiple agents could potentially handle a task, the supervisor evaluates their respective strengths and assigns work to maximise success probability. This architectural pattern scales naturally as we add specialised agents in later quarters.
Error classification employs a lightweight BERT-tiny model to categorise failure modes, enabling systematic improvement of agent reliability. By understanding why tasks fail—whether due to data unavailability, ambiguous queries, capability limitations, or external service errors—we can implement targeted fixes and improve overall system robustness. The classification system feeds into our active learning pipeline, identifying high-value training examples that address common failure patterns.
920
Total Hours
€69K
Quarter Cost
5
Components
Q4 2026: Vision Pipeline + Sub-Agents
Q4 2026 completes Year 1's remote sensing foundation by integrating vision-language models that enable natural language understanding of imagery content. CLIP (Contrastive Language-Image Pre-training) and SigLIP (Sigmoid Loss for Language-Image Pre-training) provide the multimodal reasoning capabilities that allow users to query images using natural language rather than requiring technical knowledge of image processing parameters or feature extraction algorithms.
CLIP/SigLIP Integration
Vision-language model for map and image understanding through natural language queries
200 hours | €15,000
Image Query Understanding
Natural language queries about uploaded imagery with semantic understanding
160 hours | €12,000
Sub-Process Manager
Parallel WFS fetching with child threads for performance optimisation
160 hours | €12,000
The image query understanding system transforms how users interact with geospatial imagery. Instead of manually configuring segmentation parameters or selecting specific analysis algorithms, users can ask questions like "Show me all buildings with solar panels" or "Identify areas of vegetation stress in this agricultural region." The vision-language model interprets these queries, determines appropriate analysis methods, and coordinates the necessary processing steps to generate meaningful responses.
Object detection integration through Detectron2 adds instance-level feature identification capabilities. Whilst SAMGeo excels at segmenting broad categories, Detectron2 provides precise bounding boxes and classifications for specific objects like vehicles, construction equipment, or infrastructure elements. The combination of segmentation and detection creates a comprehensive computer vision pipeline capable of addressing diverse remote sensing analysis requirements.
The sub-process manager implements parallel data fetching for Web Feature Service (WFS) requests, dramatically improving performance when queries require data from multiple sources. Rather than sequentially requesting each dataset, the manager spawns child threads that fetch data concurrently, reducing total query time by up to 80% for multi-source queries. This architectural improvement benefits all agents, not just those working with imagery.
The RS visualisation UI provides frontend components for displaying remote sensing results, including side-by-side comparison views, temporal change visualisation, and interactive segmentation refinement. Users can overlay multiple analysis results, adjust transparency, and export findings in standard GIS formats. The interface balances power-user capabilities with accessibility for non-technical stakeholders, ensuring that sophisticated analysis remains comprehensible to decision-makers.
920
Development Hours
€69K
Quarter Investment
5
Major Components
Year 1 Financial Summary
Year 1's total investment of €386,400 reflects the foundational nature of this phase, with development costs representing 69% of the budget. The 3,560 development hours at €75/hour cover infrastructure migration, remote sensing pipeline construction, agent capability enhancements, and integration work across all four quarters. This rate reflects senior-level expertise required for complex AI and geospatial systems development.
Infrastructure Investment
The €33,600 infrastructure allocation covers Kubernetes cluster operation and the GEX131 dedicated server rental. This represents a 52% increase over standard infrastructure costs, reflecting the high-availability requirements and GPU-accelerated computing demands of enterprise GeoAI. The investment unlocks capabilities that would be impossible on commodity hardware, justifying the premium through dramatically improved AI quality and performance.
External Services & Data
External services (€8,000) include Sentinel Hub API access, geocoding services, and third-party data feeds. Training data acquisition (€25,000) funds the purchase of high-resolution aerial imagery and ground-truth annotations for Finnish-specific model fine-tuning. The cloud burst budget (€2,400) provides RunPod GPU capacity for occasional high-intensity workloads that exceed on-premise capacity.
The 15% contingency (€50,400) provides buffer for unexpected challenges, scope adjustments, and technical risks. Given the ambitious nature of Year 1's objectives—particularly the infrastructure migration and remote sensing pipeline construction—this contingency represents prudent risk management. Historical data suggests that complex AI infrastructure projects typically encounter 10-20% cost overruns, making our 15% allocation appropriate for the risk profile.
3,560
Development Hours
Total effort across all quarters
€386K
Year 1 Total
Complete investment including contingency
54%
RS Focus
Remote sensing development priority
Year 2: Learning, Agents & Knowledge (2027)
Intelligence Year
Year 2 shifts focus from foundational capabilities to autonomous intelligence, implementing self-improving systems that learn from experience and operate with increasing independence. The year introduces feedback loops that capture user corrections, active learning pipelines that automatically improve models, and autonomous agents capable of multi-step reasoning without human intervention. By year-end, Geofuchs will possess sophisticated agent intelligence that rivals leading commercial platforms whilst maintaining explainability and safety.
Q1 2027
Feedback loops, GraphRAG, entity extraction, GeoSPARQL ontology
Q2 2027
LoRA training service, MCP toolkit, tool adapters, discovery service
Q3 2027
Reflection module, autonomous planning, workflow engine, agent memory
Q4 2027
Constitutional AI, spatial analysis suite, clustering, classification
The quarterly progression builds systematically towards full autonomy. Q1 establishes the feedback infrastructure and knowledge graph foundations that enable learning from experience. Q2 implements the MCP toolkit that standardises tool integration and enables automated model improvement through LoRA fine-tuning. Q3 introduces reflection and autonomous planning capabilities that allow agents to critique their own performance and execute complex multi-step tasks. Q4 adds constitutional AI guardrails and completes the spatial analysis suite, ensuring that autonomous agents operate safely within defined boundaries.
Year 2's 34% remote sensing / 66% agent focus reflects the strategic shift towards intelligence and autonomy. With foundational remote sensing capabilities established in Year 1, we can now concentrate on building the reasoning and learning systems that transform Geofuchs from a capable tool into an intelligent partner. The investment in knowledge graphs, autonomous agents, and self-improvement mechanisms positions the platform for exponential capability growth as these systems accumulate experience and refine their performance over time.
Q1 2027: Feedback Loop + GraphRAG Foundation
Q1 2027 implements the feedback infrastructure that enables continuous learning from user interactions. The feedback UI provides intuitive thumbs-up/thumbs-down controls alongside correction interfaces where users can specify exactly what went wrong and what the correct response should have been. This explicit feedback proves far more valuable than implicit signals, as it captures not just that something failed but precisely how it should have succeeded.
Feedback UI
Thumbs up/down controls with correction interface for explicit user feedback on agent responses
120 hours | €9,000
Data Curation Pipeline
Automated JSONL aggregation from high-quality interactions for training data generation
160 hours | €12,000
Microsoft GraphRAG Integration
Deploy graph-based retrieval augmented generation for geospatial knowledge management
240 hours | €18,000
The data curation pipeline automatically aggregates high-quality interactions into JSONL format suitable for model fine-tuning. The system applies quality filters that select interactions with positive feedback, exclude ambiguous cases, and ensure diversity across query types and geographic regions. This automated curation eliminates the manual effort traditionally required for training data preparation whilst maintaining high data quality standards through algorithmic filtering and sampling strategies.
Microsoft GraphRAG integration represents a fundamental architectural enhancement that transforms how the system retrieves and reasons about geospatial knowledge. Traditional RAG systems retrieve documents based on vector similarity, but GraphRAG leverages graph structure to understand relationships between entities, enabling more sophisticated reasoning about spatial relationships, hierarchies, and dependencies. For geospatial applications, this proves particularly valuable as geographic entities naturally form complex relationship networks.
Entity extraction through DeepKE identifies and classifies spatial entities within text, populating the knowledge graph with structured information about places, features, and their relationships. The system recognises not just place names but also spatial relationships ("north of", "adjacent to"), administrative hierarchies, and temporal associations. This structured knowledge enables more precise query understanding and more accurate response generation.
The GeoSPARQL ontology design establishes a formal semantic framework for representing geospatial knowledge across Finland and Europe. GeoSPARQL extends standard SPARQL query language with spatial operators, enabling queries like "find all municipalities within 50km of Helsinki that intersect protected nature reserves." The ontology captures administrative boundaries, topographic features, infrastructure networks, and environmental characteristics in a machine-readable format that supports automated reasoning and inference.
920
Development Hours
€69K
Quarter Investment
50K+
Graph Entities
Q2 2027: Active Learning + MCP Toolkit
Q2 2027 implements active learning through automated LoRA (Low-Rank Adaptation) fine-tuning, enabling the system to continuously improve its performance based on curated feedback data. The LoRA training service operates as a Dockerised microservice that monitors the feedback pipeline, automatically initiating training runs when sufficient high-quality examples accumulate. This automation eliminates manual intervention whilst ensuring models stay current with evolving user needs and geographic contexts.
1
LoRA Training Service
Dockerised auto-training on curated data with automated quality control
200 hours | €15,000
2
Confidence Scoring
vLLM log-probability analysis for uncertainty flagging and quality assessment
120 hours | €9,000
3
MCP Protocol Implementation
Core MCP server/client architecture for standardised tool integration
200 hours | €15,000
Confidence scoring analyses log probabilities from vLLM inference to identify responses where the model exhibits uncertainty. Low confidence scores trigger human review workflows, ensuring that uncertain responses receive expert validation before reaching users. This mechanism prevents the system from confidently presenting incorrect information whilst allowing high-confidence responses to flow through automatically. Over time, as models improve through active learning, the proportion of high-confidence responses increases, reducing human review burden.
The Model Context Protocol (MCP) implementation establishes a standardised framework for tool integration that dramatically simplifies adding new capabilities to the agent ecosystem. Rather than implementing custom integration code for each tool, MCP defines a universal protocol that tools expose and agents consume. This architectural pattern scales naturally as we add tools, with each new integration requiring only MCP adapter implementation rather than modifications to core agent logic.
MCP Tool Adapters
Q2 develops MCP adapters for our core data services: PostGIS for spatial database operations, Neo4j for knowledge graph queries, Qdrant for vector similarity search, and WFS services for standards-based feature access. Each adapter translates MCP protocol messages into service-specific API calls, handling authentication, error recovery, and result formatting. The 280-hour investment (€21,000) reflects the complexity of ensuring robust, production-grade integration across these diverse systems.
MCP Discovery Service
The discovery service maintains tool metadata and capability descriptions, enabling agents to dynamically discover available tools and understand their capabilities without hard-coded knowledge. When an agent encounters a task requiring unfamiliar capabilities, it queries the discovery service to identify appropriate tools, examines their interfaces, and constructs appropriate invocations. This dynamic discovery enables the agent ecosystem to automatically leverage new tools as they become available.
920
Total Hours
€69K
Quarter Cost
95%
Target Success Rate
Q3 2027: Autonomous Agents + Reflection
Q3 2027 introduces true autonomy through reflection capabilities that enable agents to critique their own performance and iteratively improve their approaches. The reflection module analyses execution traces after task completion, identifying inefficiencies, errors, and opportunities for improvement. This self-critique capability transforms agents from reactive tools into learning systems that accumulate expertise through experience, progressively improving their performance on recurring task patterns.
Reflection Module
Post-task self-critique of execution traces with improvement identification
200 hours | €15,000
DSPy Prompt Optimizer
Automated prompt refinement based on feedback and performance metrics
160 hours | €12,000
Autonomous Planning Agent
Multi-step task decomposition with code generation capabilities
280 hours | €21,000
Workflow Agent
Declarative workflow execution engine for complex business processes
200 hours | €15,000
Agent Memory
Individual agent context persistence for learning and adaptation
160 hours | €12,000
DSPy prompt optimisation automates the traditionally manual process of prompt engineering. The system treats prompts as programs that can be optimised through feedback, automatically refining prompt structure, examples, and instructions based on performance metrics. This automated optimisation proves particularly valuable for geospatial applications where optimal prompts vary significantly across different geographic regions, data sources, and analysis types. The system learns these contextual variations and adapts prompts accordingly.
The autonomous planning agent represents the culmination of Year 2's intelligence focus, capable of decomposing complex multi-step tasks into executable plans without human guidance. When presented with a high-level goal like "Identify all flood-risk areas within 10km of critical infrastructure and generate evacuation route recommendations," the planner breaks this into constituent sub-tasks: define flood-risk criteria, query infrastructure databases, perform spatial analysis, calculate optimal routes, and synthesise recommendations. Each sub-task is assigned to appropriate specialist agents, with the planner monitoring execution and adapting the plan based on intermediate results.
The workflow agent provides declarative workflow execution for enterprise scenarios requiring formal process definitions. Unlike the autonomous planner which generates plans dynamically, the workflow agent executes pre-defined workflows specified in a declarative format similar to enterprise workflow systems. This proves essential for regulated industries where process compliance requires auditable, repeatable execution patterns. The workflow engine supports conditional branching, parallel execution, error handling, and human approval gates.
Per-agent memory systems enable individual agents to accumulate specialised knowledge and preferences over time. Each agent maintains its own context store capturing successful strategies, common failure patterns, and domain-specific knowledge relevant to its specialisation. This distributed memory architecture scales naturally as we add agents, with each agent developing expertise in its domain without requiring centralised knowledge management.
1,000
Development Hours
€75K
Quarter Investment
80%
Target Task Completion
Q4 2027: Goal Alignment + AI Spatial Analysis
Q4 2027 completes Year 2 by implementing constitutional AI guardrails that ensure autonomous agents operate safely within defined boundaries, alongside a comprehensive spatial analysis suite that achieves feature parity with leading commercial platforms. The constitution layer verifies that agent plans align with safety constraints, ethical guidelines, and operational policies before execution, preventing autonomous systems from pursuing technically feasible but inappropriate actions.
1
Constitution Layer
Safety and alignment verification for agent plans with policy enforcement
200 hours | €15,000
2
Hot Spot Analysis
Spatial statistics implementation for identifying significant clustering patterns
120 hours | €9,000
3
Geographically Weighted Regression
GWR algorithm integration for spatial relationship modelling
160 hours | €12,000
4
Clustering Algorithms
DBSCAN and K-means for spatial clustering and pattern detection
120 hours | €9,000
5
ML Classification Suite
Decision tree, SVM, and forest-based classifiers for spatial prediction
200 hours | €15,000
6
Analysis API
Unified API for all spatial analysis methods with consistent interfaces
160 hours | €12,000
The constitutional AI implementation draws inspiration from Anthropic's research on constitutional AI, encoding safety constraints as verifiable rules that agents must satisfy before executing plans. The constitution includes data privacy requirements (never expose personally identifiable information), operational boundaries (stay within authorised geographic regions), and ethical guidelines (prioritise human safety in infrastructure recommendations). Agents submit plans for constitutional review, receiving approval only when all constraints are satisfied.
Hot spot analysis implements Getis-Ord Gi* statistics to identify statistically significant spatial clustering. This proves invaluable for applications like crime analysis (identifying high-crime areas), public health (detecting disease clusters), and urban planning (finding areas of concentrated development pressure). The implementation handles both point and polygon data, automatically selecting appropriate spatial weights matrices based on data characteristics and analysis objectives.
Geographically Weighted Regression (GWR) extends traditional regression by allowing relationship parameters to vary across space, capturing geographic heterogeneity that global models miss. For example, the relationship between housing prices and proximity to transit might be strong in urban cores but weak in suburban areas. GWR detects these spatial variations, providing more accurate predictions and revealing geographic patterns in relationships between variables.
The clustering algorithms suite includes DBSCAN for density-based clustering that handles arbitrary shapes and noise, plus K-means for efficient partitioning of large datasets. The ML classification suite adds decision trees, support vector machines, and random forests for spatial prediction tasks like land cover classification, building type prediction, and infrastructure condition assessment. The unified analysis API provides consistent interfaces across all methods, simplifying integration and reducing learning curves for developers.
960
Development Hours
€72K
Quarter Investment
80%
Feature Parity Target
Year 2 Financial Summary
Year 2's total investment of €469,660 reflects the increased complexity of autonomous agent development and knowledge graph construction, with development costs rising to €285,000 (61% of budget) across 3,800 hours. The higher hourly total compared to Year 1 reflects the sophisticated nature of autonomous reasoning, reflection mechanisms, and constitutional AI implementation—all requiring deep expertise in both AI systems and geospatial domain knowledge.
Infrastructure costs nearly double to €62,400, reflecting the expansion from single-node to three-node GEX131 cluster required to support autonomous agent workloads and knowledge graph operations at scale. The Kubernetes cluster orchestrates workload distribution across nodes, ensuring high availability and enabling horizontal scaling as agent complexity increases. This infrastructure investment proves essential for maintaining sub-15-second response times for complex multi-agent queries.
ML Training & Knowledge Graph
ML training compute allocation (€35,000) funds Lambda Labs GPU instances for LoRA fine-tuning and model distillation experiments. Whilst our on-premise GEX131 cluster handles inference workloads efficiently, training benefits from burst access to multiple high-end GPUs that would be cost-prohibitive to maintain continuously. Lambda Labs provides on-demand H100 access at competitive rates, enabling rapid experimentation without capital expenditure.
Knowledge graph data licensing (€20,000) acquires commercial datasets that supplement freely available Finnish government data. This includes detailed building footprints, infrastructure networks, and environmental monitoring data that enhance the knowledge graph's completeness and enable more sophisticated reasoning about spatial relationships and dependencies.
The cloud burst budget increases to €6,000, reflecting higher utilisation as autonomous agents generate more complex workloads that occasionally exceed on-premise capacity. The 15% contingency (€61,260) addresses risks associated with autonomous agent development, where unexpected behaviours and edge cases frequently emerge during testing. This contingency has proven essential in similar projects, where autonomous systems often require additional iteration cycles to achieve production reliability.
3,800
Development Hours
€470K
Year 2 Total
66%
Agent Focus
Year 3: Scale, Trust & Generation (2028)
Enterprise Year
Year 3 transforms Geofuchs into an enterprise-grade platform through explainability, compliance, generative capabilities, and industry-specific verticals. The year implements XAI dashboards that make agent reasoning transparent, GDPR-compliant audit trails that satisfy regulatory requirements, and vector-to-image generation that creates custom cartographic visualisations. By year-end, Geofuchs will serve three industry verticals—Urban Planning, Infrastructure, and Environmental monitoring—with specialised agents and workflows tailored to each domain.
Q1 2028
XAI dashboard, source attribution, GDPR compliance, data lineage, authentication
Q2 2028
NeMo guardrails, conflict resolution, vector-to-image, style transfer, super-resolution
Q3 2028
Model distillation, distilled deployment, simulation sandbox, synthetic data, benchmarking
Q4 2028
Multi-region DR, GPU autoscaling, edge caching, three industry verticals
The quarterly progression addresses enterprise requirements systematically. Q1 implements explainability and compliance infrastructure that satisfies regulatory and audit requirements. Q2 adds production guardrails and generative capabilities that enable safe, creative outputs. Q3 optimises performance through model distillation and establishes simulation environments for agent training. Q4 scales infrastructure to multi-region deployment and launches three industry-specific verticals that demonstrate platform versatility across diverse use cases.
Year 3's 40% remote sensing / 60% agent focus maintains emphasis on intelligence whilst adding generative capabilities that leverage the remote sensing foundation built in Years 1-2. The vector-to-image generation and super-resolution capabilities transform Geofuchs from an analysis platform into a creative tool that generates custom visualisations and enhances imagery quality. The industry verticals demonstrate how the platform's general capabilities adapt to specific domain requirements, establishing patterns for future vertical expansion.
Q1 2028: XAI + Compliance
Q1 2028 implements explainable AI (XAI) infrastructure that makes agent reasoning transparent and auditable, addressing the "black box" criticism often levelled at AI systems. The XAI dashboard visualises agent decision trees, showing the reasoning path from initial query through intermediate steps to final response. Users can inspect which data sources were consulted, what reasoning rules were applied, and why specific conclusions were reached. This transparency proves essential for enterprise adoption, where decision-makers require confidence in AI recommendations before acting upon them.
XAI Dashboard
Agent decision tree visualisation with interactive exploration of reasoning paths and data sources
200 hours | €15,000
Source Attribution UI
Reference document display for all answers with direct links to original data sources
160 hours | €12,000
GDPR Compliance Logger
Automated audit logging for all actions with data retention and deletion capabilities
200 hours | €15,000
Source attribution displays reference documents for every factual claim in agent responses, enabling users to verify information against original sources. Each citation links directly to the relevant document section, whether that's a government dataset, scientific publication, or internal knowledge base entry. This attribution system satisfies academic rigour requirements whilst building user trust through verifiable claims rather than unsupported assertions.
GDPR compliance logging implements automated audit trails that capture all system actions involving personal data, satisfying European data protection regulations. The logger records data access, processing operations, retention periods, and deletion requests with immutable timestamps and user attribution. This comprehensive logging enables organisations to demonstrate compliance during regulatory audits whilst providing users with transparency about how their data is processed.
Data lineage tracking establishes full provenance for all generated outputs, documenting the complete chain from source data through processing steps to final results. When an agent generates a flood risk map, the lineage system records which elevation data, precipitation records, and infrastructure databases were used, what analysis algorithms were applied, and what parameters were configured. This provenance proves essential for scientific reproducibility and regulatory compliance in domains like environmental assessment.
Enterprise Authentication
The user authentication system implements OAuth2/OIDC protocols with role-based access control, enabling integration with enterprise identity providers like Azure AD, Okta, and Google Workspace. Role-based permissions control access to sensitive datasets, restrict certain analysis capabilities to authorised users, and enforce organisational policies about data sharing and export. This enterprise-grade authentication proves essential for organisations with complex security requirements and regulatory obligations.
200 hours | €15,000
920
Development Hours
€69K
Quarter Investment
Q2 2028: Guardrails + Image Generation
Q2 2028 implements production guardrails through NVIDIA's NeMo framework, providing robust filtering for personally identifiable information (PII), toxic content, and other inappropriate outputs. The guardrails operate at multiple levels: input validation prevents malicious queries from reaching agents, output filtering catches inappropriate responses before they reach users, and continuous monitoring detects emerging patterns that require new guardrail rules. This multi-layered approach ensures safe operation even as agents gain autonomy and generative capabilities.
NeMo Guardrails
PII detection and toxic content filtering with real-time validation
200 hours | €15,000
Conflict Resolution
Multi-source data contradiction handling with confidence-weighted synthesis
160 hours | €12,000
Vector-to-Image Generation
Map style generation from vector data with customisable aesthetics
280 hours | €21,000
Conflict resolution handles contradictions when multiple data sources provide inconsistent information about the same geographic feature. The system employs confidence-weighted synthesis that considers source reliability, data recency, and measurement precision when reconciling conflicts. Rather than arbitrarily selecting one source over another, the resolver generates probabilistic estimates that reflect uncertainty, enabling downstream processes to make informed decisions about how to handle ambiguous information.
Vector-to-image generation represents a transformative capability that enables custom cartographic visualisation without manual design work. Users specify desired map styles through natural language descriptions like "minimalist monochrome with emphasis on transportation networks" or "vibrant colours highlighting green spaces and water features." The generation system interprets these descriptions, selects appropriate colour palettes, configures symbol styles, and renders publication-quality maps that match the specified aesthetic whilst maintaining cartographic best practices.
Image-to-Image Transform
Style transfer capabilities enable transformation of existing maps into different visual styles whilst preserving geographic accuracy. A traditional topographic map can be transformed into a watercolour artistic rendering, a satellite image can be stylised to emphasise specific features, or a historical map can be modernised whilst retaining its original character. These transformations prove valuable for public engagement, educational materials, and creative applications where visual appeal enhances communication effectiveness.
200 hours | €15,000
Super-Resolution
Image super-resolution upscales low-resolution imagery using AI-based enhancement that infers missing detail rather than simply interpolating pixels. This proves particularly valuable for historical aerial photography where original high-resolution sources are unavailable, or for satellite imagery where higher resolution would be cost-prohibitive. The super-resolution system preserves geographic accuracy whilst enhancing visual clarity, enabling detailed analysis of imagery that would otherwise be too coarse for practical use.
160 hours | €12,000
1,000
Development Hours
€75K
Quarter Investment
75%
Generation Parity
Q3 2028: Model Distillation + Simulation
Q3 2028 implements model distillation to create efficient 7B parameter student models that capture the knowledge of our 70B parameter teacher models whilst requiring only a fraction of the computational resources. Distillation trains smaller models to mimic larger models' behaviour, achieving 85-90% of the teacher's performance whilst enabling 3x faster inference and dramatically reduced memory requirements. This efficiency gain proves essential for scaling to hundreds of concurrent users without proportional infrastructure costs.
Model Distillation
Train 7B student models from 70B teacher models with knowledge transfer optimisation
280 hours | €21,000
Distilled Model Deployment
Low-latency inference with distilled models achieving 3x speed improvement
120 hours | €9,000
Simulation Sandbox
Agent practice environment for spatial tasks with synthetic scenarios
280 hours | €21,000
The distilled model deployment infrastructure optimises inference pipelines specifically for the smaller models, leveraging their reduced memory footprint to batch more requests per GPU and their faster computation to reduce queue wait times. The deployment system automatically routes queries to appropriate models based on complexity, using distilled models for straightforward queries and reserving full-scale models for complex reasoning tasks that require maximum capability.
The simulation sandbox creates a safe practice environment where agents can attempt complex spatial tasks without risk of corrupting production data or generating incorrect outputs that reach users. The sandbox generates synthetic scenarios that mirror real-world complexity—fictional cities with realistic infrastructure networks, simulated environmental events, and artificial datasets with known ground truth. Agents practice in this environment, receiving immediate feedback on performance and accumulating experience that transfers to production scenarios.
Synthetic data generation creates training scenarios that cover edge cases and rare events that occur infrequently in production data. The generator produces geographically plausible but fictional scenarios: unusual infrastructure configurations, extreme weather events, complex spatial relationships, and ambiguous queries that test agent reasoning limits. This synthetic data supplements real-world examples, ensuring comprehensive training coverage without waiting for rare events to occur naturally.
Performance benchmarking establishes comprehensive regression testing that validates each system update maintains or improves performance across critical metrics. The benchmark suite measures query latency, agent response accuracy, resource utilisation, and user satisfaction across diverse query types and geographic contexts. Automated benchmarking runs after every significant code change, catching performance regressions before they reach production and providing objective evidence of improvement from optimisation efforts.
1,000
Development Hours
€75K
Quarter Investment
3x
Speed Improvement
Q4 2028: Enterprise Scale + Industry Verticals
Q4 2028 completes the three-year roadmap by scaling infrastructure to enterprise requirements and launching three industry-specific verticals that demonstrate platform versatility. The multi-region disaster recovery evaluation assesses requirements for expanding beyond Finland whilst maintaining data residency compliance, examining potential DR sites in Sweden, Estonia, and other EU locations that satisfy regulatory constraints. This evaluation informs future expansion decisions without requiring immediate capital commitment.
Multi-Region DR Evaluation
Assess EU data residency and disaster recovery site requirements for future expansion
160 hours | €12,000
GPU Autoscaling Controller
Custom Kubernetes controller for 2-6 GPU node scaling based on demand
200 hours | €15,000
Edge Caching Layer
Cloudflare CDN integration for tile serving with global distribution
160 hours | €12,000
The GPU autoscaling controller implements custom Kubernetes logic that dynamically adjusts cluster size between 2-6 GEX131 nodes based on workload demand. During peak hours when multiple users submit complex queries simultaneously, the controller provisions additional nodes to maintain response time targets. During quiet periods, it scales down to minimum configuration, reducing infrastructure costs without sacrificing availability. This elastic scaling proves essential for cost-effective operation whilst maintaining enterprise-grade performance guarantees.
Edge caching through Cloudflare CDN dramatically improves tile serving performance for international users by distributing map tiles across global edge locations. Rather than every tile request traversing the internet to Finland, Cloudflare's edge network caches frequently accessed tiles at locations near users, reducing latency from hundreds of milliseconds to tens of milliseconds. This performance improvement proves particularly valuable for interactive map applications where responsive panning and zooming directly impact user experience.
1
Urban Planning Vertical
Zoning analysis, building permits, city development agents leveraging municipal datasets
280 hours | €21,000
2
Infrastructure Vertical
Road networks, utility mapping, construction monitoring with asset management integration
280 hours | €21,000
3
Environmental Vertical
Flood risk, deforestation, biodiversity agents leveraging SYKE/GTK datasets
240 hours | €18,000
The three industry verticals demonstrate how Geofuchs's general capabilities adapt to specific domain requirements. Each vertical includes specialised agents trained on domain-specific data, custom workflows that encode industry best practices, and integrations with sector-specific data sources. The Urban Planning vertical connects to municipal zoning databases and building permit systems, the Infrastructure vertical integrates with asset management platforms and construction monitoring tools, and the Environmental vertical leverages Finland's exceptional environmental datasets from SYKE (Finnish Environment Institute) and GTK (Geological Survey of Finland).
1,320
Development Hours
€99K
Quarter Investment
3
Industry Verticals
Year 3 Financial Summary
Year 3's total investment of €623,300 represents the peak annual expenditure, reflecting enterprise-scale infrastructure and the development of three industry verticals. Development costs reach €318,000 (51% of budget) across 4,240 hours, with the increased effort driven by vertical-specific customisation, generative capabilities implementation, and enterprise compliance requirements. The per-hour rate remains consistent at €75, but total hours increase by 12% over Year 2 to accommodate the expanded scope.
Infrastructure costs reach €114,000, reflecting the 4-6 node GEX131 cluster required for enterprise-scale operation with high availability. The Kubernetes deployment spans 20+ nodes when accounting for control plane, monitoring, and auxiliary services, creating a robust platform capable of handling hundreds of concurrent users with sub-second response times for most queries. This infrastructure investment positions Geofuchs for commercial launch with enterprise-grade reliability and performance.
Cloud GPU & Burst Capacity
Cloud GPU allocation (€50,000) funds model distillation training and generative model fine-tuning on Lambda Labs and RunPod infrastructure. The distillation process requires sustained access to multiple H100 GPUs over several weeks, making cloud burst more cost-effective than expanding on-premise capacity. The cloud burst budget increases to €15,000, reflecting higher utilisation as generative capabilities and industry verticals create more diverse workload patterns that occasionally exceed on-premise capacity.
Industry Data Licensing
Industry data licensing (€45,000) acquires specialised datasets for the three verticals: detailed building information for Urban Planning, infrastructure asset databases for Infrastructure monitoring, and high-resolution environmental monitoring data for the Environmental vertical. These commercial datasets supplement freely available government data, providing the depth and currency required for production applications in each domain.
The 15% contingency (€81,300) addresses risks associated with enterprise deployment, industry vertical development, and generative AI implementation. These areas typically encounter unexpected challenges during production deployment: edge cases in vertical-specific workflows, performance optimisation requirements for generative models, and integration complexities with enterprise systems. The contingency provides buffer for addressing these challenges without compromising delivery timelines or quality standards.
4,240
Development Hours
€623K
Year 3 Total
60%
Agent Focus
Complete 3-Year Investment Summary
The complete three-year investment totals €1,479,360, representing a 22.8% increase over the original roadmap budget of €1,204,625. This increase reflects two primary factors: enhanced high-availability infrastructure requirements (+52% infrastructure costs) and the addition of three industry-specific verticals not included in the original scope. The revised budget maintains the original roadmap's ambitious technical goals whilst adding enterprise-grade reliability and vertical market capabilities that significantly enhance commercial viability.
Development Investment
€870,000 across 11,600 hours represents 59% of total investment, funding the core engineering effort that builds platform capabilities
Infrastructure & Cloud
€233,400 for Kubernetes clusters, dedicated servers, and cloud GPU burst capacity provides the computational foundation
External Services & Data
€183,000 for APIs, datasets, and licensing ensures access to comprehensive geospatial data across Europe
Contingency Reserve
€192,960 (15%) provides buffer for technical risks, scope adjustments, and unexpected challenges
The investment profile shows steady growth across all three years, with Year 1 establishing foundations (€386,400), Year 2 building intelligence (€469,660), and Year 3 achieving enterprise scale (€623,300). This progressive investment pattern aligns with platform maturity, with each year building upon previous achievements whilst managing financial risk through phased capability delivery. The contingency allocation proves particularly important given the ambitious technical scope and the inherent uncertainties in AI system development.
Compared to the original roadmap budget of €1,204,625, the revised figure of €1,479,360 represents prudent adjustment for enterprise requirements whilst maintaining aggressive technical goals. The 6.7% increase over the previous revision (€1,386,900) specifically addresses high-availability infrastructure needs identified through detailed capacity planning. This investment positions Geofuchs for commercial success with enterprise customers who demand reliability, compliance, and performance guarantees that commodity infrastructure cannot provide.
Balanced Development Approach: RS + Agents
The roadmap maintains strategic balance between Remote Sensing (RS) and Agent development throughout the three-year period, ensuring that neither capability area advances at the expense of the other. Year 1 emphasises RS development (54% focus) to close the critical capability gap that represents our most significant competitive disadvantage. Year 2 shifts towards agent intelligence (66% focus) as autonomous reasoning becomes the primary differentiator. Year 3 balances both areas (40% RS / 60% agents) whilst adding generative capabilities and enterprise features.
This balanced approach ensures that as we add sophisticated remote sensing capabilities, our agent infrastructure evolves to effectively utilise these new tools. Conversely, as agents gain autonomy and reasoning capabilities, they have increasingly powerful remote sensing tools at their disposal. The interleaving prevents capability mismatches where one area significantly outpaces the other, ensuring that the platform maintains coherent functionality throughout development.
The quarterly focus percentages guide resource allocation decisions, ensuring that development effort aligns with strategic priorities whilst maintaining progress across both capability areas. Quarters with heavy RS focus (Q4 2026, Q2 2028) coincide with major vision pipeline milestones, whilst agent-heavy quarters (Q2-Q3 2027) implement autonomous reasoning and reflection capabilities. This rhythm creates natural integration points where new capabilities from one area enhance the other, accelerating overall platform evolution.
Team Structure & Operational Roles
The team structure scales progressively from 5 full-time equivalents in Year 1 to 8 FTEs by Year 3, reflecting increasing platform complexity and operational requirements. Year 1 focuses on core development with emphasis on backend engineering and initial ML capabilities. Year 2 expands ML engineering capacity to support autonomous agent development and active learning implementation. Year 3 adds additional agent-focused ML engineering and scales frontend/DevOps resources to support enterprise deployment and industry verticals.
1
Senior Backend Engineer
2 FTEs across all years, focusing on core platform architecture, API development, and system integration
€72,000 average annual cost
2
AI/ML Engineer (RS focus)
Scales from 1 to 2 FTEs, specialising in remote sensing, computer vision, and image processing
€95,000 average annual cost
3
AI/ML Engineer (Agents)
Scales from 1 to 2 FTEs, specialising in LLMs, autonomous agents, and reasoning systems
€95,000 average annual cost
4
Frontend/Fullstack
Scales from 0.5 to 1 FTE, building user interfaces and visualisation components
€65,000 average annual cost
5
DevOps/SRE
Scales from 0.5 to 1 FTE, managing infrastructure, deployment, and reliability
€78,000 average annual cost
The salary figures reflect Finnish market rates for senior-level technical talent in the Helsinki metropolitan area, adjusted for the specialised nature of geospatial AI development. ML engineers command premium compensation due to high demand and limited supply, whilst backend and DevOps roles align with standard senior developer rates. The team structure assumes remote hiring capability, enabling access to talent across Finland and potentially other Nordic countries where necessary.
Year 1 Focus
DevOps resources concentrate on the critical GEX131 migration and establishing CI/CD pipelines that enable rapid iteration. The 0.5 FTE allocation reflects part-time focus during infrastructure establishment, scaling to full-time as Kubernetes complexity increases.
Year 2 Focus
ML Ops capabilities emerge as retraining pipelines and GPU cluster management become operational requirements. The second RS-focused ML engineer joins to handle TorchGeo model integration and training data pipeline development.
Year 3 Focus
Compliance and security expertise becomes essential as autonomous agents require safety auditing and regulatory compliance verification. The second agent-focused ML engineer enables parallel development of industry verticals.
The team structure emphasises senior-level expertise over junior resources, reflecting the complex nature of geospatial AI development where domain knowledge and architectural experience prove more valuable than raw development capacity. This approach minimises coordination overhead whilst ensuring high-quality implementation across all platform components. The progressive scaling pattern manages financial risk whilst ensuring adequate resources for each development phase's requirements.
Feature Parity Milestone Tracker
The feature parity tracker establishes concrete targets for achieving competitive capability levels across all major platform components. Each milestone specifies the target quarter, development hours required, and expected parity level compared to leading commercial platforms. The parity percentages reflect realistic assessments of achievable capability given resource constraints, with most features targeting 80-90% parity rather than attempting perfect replication of mature commercial systems.
The parity levels reflect honest assessments of achievable capability rather than aspirational targets. For example, vector-to-image generation targets 75% parity because generative AI for cartography remains an emerging field where even commercial leaders have limited capabilities. Conversely, MCP toolkit targets 90% parity because the protocol specification is well-defined and our implementation can closely match reference implementations. These realistic targets enable accurate progress tracking and prevent disappointment from unrealistic expectations.
The cumulative hours across all milestones (6,200) represent approximately 53% of total development effort, with remaining hours allocated to integration work, testing, documentation, and operational improvements not captured in specific feature milestones. This distribution reflects the reality that building individual features represents only part of platform development effort, with substantial additional work required to integrate features into a cohesive, production-ready system.
Risk Register & Mitigation Strategies
The risk register identifies key threats to roadmap success alongside probability assessments, impact evaluations, and specific mitigation strategies. Proactive risk management proves essential for ambitious technical projects where unexpected challenges frequently emerge. The register focuses on technical, operational, and resource risks that could derail development timelines or compromise platform quality, enabling early intervention when risk indicators appear.
SAM Accuracy Below Commercial Standards
Probability: Medium | Impact: High
Mitigation: Fine-tune SAMGeo on Finnish aerial imagery to improve accuracy for local building styles and landscape characteristics. Maintain TorchGeo models as fallback option if SAMGeo proves insufficient for production requirements.
GraphRAG Scalability Limitations
Probability: Medium | Impact: Medium
Mitigation: Implement geographic sharding to distribute graph across regions, reducing query complexity. Deploy aggressive caching layer for frequently accessed graph patterns. Monitor query performance and optimise hot paths.
MCP Adoption Uncertainty
Probability: Low | Impact: Low
Mitigation: Build adapter layer that enables fallback to custom APIs if MCP adoption proves limited. Maintain flexibility to pivot to alternative integration patterns without major architectural changes.
GPU Shortage (GEX131 Availability)
Probability: Medium | Impact: High
Mitigation: Pre-negotiate with Hetzner Enterprise Sales for guaranteed capacity. Establish colocation backup plan if dedicated rental proves unavailable. Consider alternative GPU options (A100, H100) if RTX 6000 Ada unavailable.
Talent Acquisition Challenges
Probability: High | Impact: High
Mitigation: Enable remote hiring across Finland and Nordic region. Establish partnership with Aalto University for graduate recruitment. Offer competitive compensation packages that reflect specialised skill requirements.
LLM Hallucination in Agents
Probability: High | Impact: Medium
Mitigation: Implement constitutional AI guardrails and confidence thresholds. Deploy human-in-loop workflows for low-confidence responses. Maintain comprehensive audit trails for debugging hallucination patterns.
Hardware availability represents the highest-priority risk, as GEX131 GPU servers in Helsinki data centres face limited supply. The mitigation strategy emphasises proactive engagement with Hetzner Enterprise Sales to secure capacity commitments before immediate need arises. The colocation backup plan provides alternative path if standard rental proves unavailable, though this increases operational complexity and potentially costs.
Migration data loss, whilst low probability, carries catastrophic impact if it occurs. The mitigation strategy maintains GEX44 as read-only backup for one week post-switchover, enabling rapid rollback if data integrity issues emerge. Checksum verification across all migrated volumes provides mathematical certainty of successful replication before decommissioning old infrastructure. This conservative approach trades slightly higher migration costs for dramatically reduced risk exposure.
Related Documents & Next Steps
This roadmap integrates with comprehensive technical specifications that detail implementation approaches for each major capability area. The related documents provide architectural guidance, data models, infrastructure specifications, and detailed technical requirements that inform development execution. Teams should consult these specifications when planning sprint work and making implementation decisions to ensure alignment with overall platform architecture.
01-ARCHITECTURE.md
System architecture including remote sensing pipeline design, agent orchestration patterns, and integration approaches
02-AGENTS.md
Agent specifications covering autonomous agents, reflection mechanisms, and multi-agent coordination patterns
04-DATA-MODEL.md
Data models including raster types, vector schemas, knowledge graph ontology, and metadata structures
06-INFRASTRUCTURE.md
Infrastructure specifications covering Kubernetes deployment, GPU cluster configuration, and HA requirements
10-REMOTE-SENSING.md
Remote sensing pipeline details including SAMGeo integration, TorchGeo models, and vision-language capabilities
11-MCP-TOOLKIT.md
MCP toolkit specification covering protocol implementation, tool adapters, and discovery service architecture

Immediate Next Steps
01
Secure GEX131 Infrastructure
Contact Hetzner Enterprise Sales immediately to confirm GEX131 availability in HEL1 data centre. Negotiate capacity commitment and establish colocation backup plan if necessary.
02
Initiate Team Recruitment
Begin recruitment for Year 1 team structure (2 backend engineers, 1 RS ML engineer, 1 agent ML engineer, 0.5 frontend, 0.5 DevOps). Prioritise DevOps hire for migration support.
03
Finalise Q1 2026 Sprint Planning
Break down Q1 tasks into two-week sprints with specific deliverables. Establish development environment and CI/CD pipeline before migration begins.
04
Establish Governance Framework
Define decision-making processes, change management procedures, and stakeholder communication cadence for three-year programme execution.
Success requires immediate action on infrastructure procurement and team assembly. The GEX131 migration represents the critical path item that unlocks all subsequent capabilities, making hardware availability the highest-priority concern. Parallel recruitment efforts ensure team capacity aligns with Q1 2026 development requirements, preventing resource constraints from delaying foundational work. With infrastructure secured and team assembled, Geofuchs can execute this ambitious roadmap and achieve its vision of becoming a leading enterprise GeoAI platform.