TerraScope: Pixel-Grounded Visual Reasoning for Earth Observation

CVPR 2026

1University of Trento, 2BIFOLD and TU Berlin, 3Technical University of Munich, 4MBZUAI
TerraScope Teaser

Figure 1: TerraScope performs pixel-grounded visual reasoning for diverse earth observation tasks.

Abstract

Vision-language models (VLMs) have shown promise in earth observation (EO), yet they struggle with tasks that require grounding complex spatial reasoning in precise pixel-level visual representations. To address this problem, we introduce TerraScope, a unified VLM that delivers pixel-grounded geospatial reasoning with two key capabilities: (1) modality-flexible reasoning: it handles single-modality inputs (optical or SAR) and adaptively fuses different modalities into the reasoning process when both are available; (2) multi-temporal reasoning: it integrates temporal sequences for change analysis across multiple time points. In addition, we curate Terra-CoT, a large-scale dataset containing 1 million samples with pixel-level masks embedded in reasoning chains across multiple sources. We also propose TerraScope-Bench, the first benchmark for pixel-grounded geospatial reasoning with six sub-tasks that evaluates both answer accuracy and mask quality to ensure authentic pixel-grounded reasoning. Experiments show that TerraScope significantly outperforms existing VLMs on pixel-grounded geospatial reasoning while providing interpretable visual evidence.

Method

We propose TerraScope, a unified vision-language framework for pixel-grounded visual reasoning in geospatial understanding. Unlike traditional VLMs that rely on language-only reasoning chains, TerraScope explicitly generates segmentation masks and grounds reasoning in the selected masked visual space.

  • Pixel-Grounded Chain-of-Thought: A cooperative mechanism between dual decoders that interleaves segmentation mask generation with text generation. Upon detecting [SEG] tokens, the mask decoder predicts segmentation masks, from which masked visual tokens are selected and injected back into the reasoning sequence.
  • Dynamic Visual Feature Extraction: Generated masks are aligned to the vision encoder's dynamic patch layout and resized to the token grid. Visual tokens covered >50% by the mask are selected, projected, and fed into the LLM to resume autoregressive generation conditioned on the KV cache.
  • Multi-Modal & Temporal Reasoning: For optical-SAR pairs, a text-guided, token-level modality selection mechanism computes cross-attention relevance scores to adaptively choose features from the more informative modality at each spatial position. For temporal sequences, explicit temporal indicators ("Image: ti") before each [SEG] token enable frame-specific mask decoding.
  • Two-Stage Training: Stage 1 trains on 2M referring expression segmentation pairs for basic grounding; Stage 2 fine-tunes on 1M Terra-CoT samples with a combined language modeling loss and segmentation loss (Dice + pixel-wise cross-entropy).
TerraScope Framework Overview

Figure 2: Overview of the TerraScope framework.

Terra-CoT Dataset

Existing EO datasets provide either segmentation labels or VQA pairs, but not both with reasoning traces. We address this with a two-stage automated pipeline to curate large-scale pixel-grounded reasoning data.

  • Grounded Captioning with Chain-of-Thought (Cap-CoT): We prompt a large multimodal model with images where land-cover categories are highlighted using colored masks and labeled accordingly, producing 250K captioning samples with explicit pixel-grounded reasoning traces. An intermediate annotator, TerraScope-Cap, is then trained to scale this to unlabeled imagery.
  • Hierarchical Data Synthesis — Level 1 (Basic Spatial Grounding): Template-based questions for randomly selected categories, covering fundamental spatial tasks such as existence verification, object counting, localization, area quantification, and boundary detection, each paired with pixel-grounded reasoning traces.
  • Hierarchical Data Synthesis — Level 2 (Complex Multi-Step Reasoning): An LLM composes multiple L1 questions into complex reasoning tasks of two types: L2-Spatial (cross-entity relationship inference) and L2-Semantic (domain knowledge beyond visual observation).

This hierarchical process produces 1M Terra-CoT samples with diverse reasoning abilities across optical, SAR, and temporal sources covering global regions.

Terra-CoT Dataset Pipeline

Figure 3: Overview of the Terra-CoT data synthesis pipeline.

TerraScope-Bench

Existing EO benchmarks emphasize coarse-grained tasks like scene classification and image captioning, failing to assess fine-grained spatial reasoning. We introduce TerraScope-Bench, a benchmark of 3,837 carefully curated samples from test sets of existing datasets, designed to evaluate pixel-level spatial understanding.

  • Six Task Categories: Coverage Percentage Analysis (855), Absolute Area Quantification (855), Distance Measurement (129), Comparative Area Ranking (855), Boundary Relationship Detection (855), and Building Change Estimation (288).
  • Automated QA Generation: Spatial properties (coverage ratios, absolute areas, inter-object distances, boundary relationships) are computed from segmentation masks to derive ground-truth answers. Questions are generated via templates, rephrased by an LLM with plausible distractors for multiple-choice format, and filtered by human experts.
  • Dual Evaluation: Unlike existing benchmarks that only assess final answer accuracy, TerraScope-Bench evaluates both response correctness and spatial reasoning quality using IoU-based segmentation metrics, verifying whether models attend to the correct regions during reasoning.
TerraScope-Bench Overview

Figure 4: Overview of TerraScope-Bench task categories and evaluation pipeline.

Experiments

We evaluate TerraScope against general VLMs, EO-specific VLMs, and fine-tuned VLMs across three benchmarks:

  • TerraScope-Bench (optical only): Six fine-grained spatial reasoning tasks — Coverage Analysis, Area Quantification, Comparative Ranking, Boundary Relationship Detection, Distance Measurement, and Building Change Estimation.
  • Landsat30AU: Land-use understanding tasks including Appearance Recognition, Numerical Reasoning, and Spatial Relationship Inference.
  • DisasterM3: Disaster assessment tasks covering Building Damage Classification and Damage Ratio Estimation.

TerraScope (8B) achieves 68.9% average accuracy on TerraScope-Bench, 73.9% on Landsat30AU, and 46.5% on DisasterM3, consistently outperforming all baselines including proprietary models like GPT-4o and reasoning models like GLM-4.1V-Think.

TerraScope Experimental Results

Table 1: Quantitative performance of TerraScope on TerraScope-Bench, Landsat30AU, and DisasterM3.

Citation

@article{shu2026terrascope,
      title={TerraScope: Pixel-Grounded Visual Reasoning for Earth Observation},
      author={Yan Shu and Bin Ren and Zhitong Xiong and Xiaoxiang Zhu and Begüm Demir and Nicu Sebe and Paolo Rota},
      year={2026},
      eprint={},
      archivePrefix={arXiv},
      primaryClass={cs.CV},
      url={},
}