Skip to content

Perception Method Library Overview

This directory is the method-level perception library. Each page should represent one technique, method, benchmark, or dataset-backed evaluation primitive. Broad synthesis pages in 30-autonomy-stack/perception/overview/ remain useful for system design, but this library is where individual methods get enough space for architecture, data, benchmarks, failure modes, deployment fit, Domain Fit, transfer notes for explicitly scoped ODDs, and sources.

Priority Ratings

Priority ratings are editorial reading and deployment triage signals. Learning answers what to read early for general autonomy understanding. Deployment answers what to evaluate early for AV deployment in the tagged context; it is not a certification, product-readiness, or all-domain average claim. If a method's deployment score is driven by a specific domain or stack role, the reason text should name that context.

MethodRatingStageMaturityReason
Availability-Aware Sensor FusionLearning: ★★★★☆
Deployment: ★★★★★
deployment-patternprototypeDirectly targets sensor degradation and availability-aware fusion.
LiDAR-MOSLearning: ★★★★☆
Deployment: ★★★★★
deployment-patternprototypeMoving-object segmentation is central to map hygiene and dynamic-scene handling.
4DSegStreamerLearning: ★★★★☆
Deployment: ★★★★☆
modern-coreprototype4DSegStreamer is rated for motion segmentation, scene flow, or dynamic-object perception workflows.
AutoOccLearning: ★★★★☆
Deployment: ★★★★☆
modern-coreprototypeAutoOcc is rated for BEV, occupancy, or freespace modeling that feeds planning-facing autonomy stacks.
BEVDepthLearning: ★★★★☆
Deployment: ★★★★☆
modern-coreprototypeImportant depth-aware BEV bridge for camera-only 3D perception.
BEVDetLearning: ★★★★☆
Deployment: ★★★★☆
modern-coreprototypeBaseline camera BEV detector that organizes many later BEV methods.
BEVStereoLearning: ★★★★☆
Deployment: ★★★★☆
modern-coreprototypeBEVStereo is rated for BEV, occupancy, or freespace modeling that feeds planning-facing autonomy stacks.
Cam4DOccLearning: ★★★★☆
Deployment: ★★★★☆
modern-coreprototypeCam4DOcc is rated for BEV, occupancy, or freespace modeling that feeds planning-facing autonomy stacks.
Conformal BoxesLearning: ★★★★☆
Deployment: ★★★★☆
deployment-patternprototypePractical uncertainty wrapper for detection risk and release gates.
Cross-Domain LiDAR Scene FlowLearning: ★★★★☆
Deployment: ★★★★☆
modern-coreprototypeCross-Domain LiDAR Scene Flow is rated for motion segmentation, scene flow, or dynamic-object perception workflows.
Dynamic Occupancy FreespaceLearning: ★★★★☆
Deployment: ★★★★☆
modern-coreprototypeDynamic Occupancy Freespace is rated for BEV, occupancy, or freespace modeling that feeds planning-facing autonomy stacks.
FlashOccLearning: ★★★★☆
Deployment: ★★★★☆
modern-coreprototypeFlashOcc is rated for BEV, occupancy, or freespace modeling that feeds planning-facing autonomy stacks.
GaussianOccLearning: ★★★★☆
Deployment: ★★★★☆
modern-coreprototypeGaussianOcc is rated for BEV, occupancy, or freespace modeling that feeds planning-facing autonomy stacks.
GraphBEVLearning: ★★★★☆
Deployment: ★★★★☆
modern-coreprototypeGraphBEV is rated for BEV, occupancy, or freespace modeling that feeds planning-facing autonomy stacks.
InsMOSLearning: ★★★★☆
Deployment: ★★★★☆
modern-coreprototypeInsMOS is rated for motion segmentation, scene flow, or dynamic-object perception workflows.
Instantaneous Motion PerceptionLearning: ★★★★☆
Deployment: ★★★★☆
modern-coreprototypeInstantaneous Motion Perception is rated for motion segmentation, scene flow, or dynamic-object perception workflows.
LiDAR-Camera Occupancy FusionLearning: ★★★★☆
Deployment: ★★★★☆
modern-coreprototypeLiDAR-Camera Occupancy Fusion is rated for BEV, occupancy, or freespace modeling that feeds planning-facing autonomy stacks.
M2-OccLearning: ★★★★☆
Deployment: ★★★★☆
modern-coreprototypeM2-Occ is rated for BEV, occupancy, or freespace modeling that feeds planning-facing autonomy stacks.
MambaMOSLearning: ★★★★☆
Deployment: ★★★★☆
modern-coreprototypeMambaMOS is rated for motion segmentation, scene flow, or dynamic-object perception workflows.
Mask4DLearning: ★★★★☆
Deployment: ★★★★☆
modern-coreprototypeMask4D is rated for motion segmentation, scene flow, or dynamic-object perception workflows.
MotionSeg3DLearning: ★★★★☆
Deployment: ★★★★☆
modern-coreprototypeMotionSeg3D is rated for motion segmentation, scene flow, or dynamic-object perception workflows.
Neural Scene Flow PriorsLearning: ★★★★☆
Deployment: ★★★★☆
modern-coreprototypeNeural Scene Flow Priors is rated for motion segmentation, scene flow, or dynamic-object perception workflows.
Open-Vocabulary Panoptic OccupancyLearning: ★★★★☆
Deployment: ★★★★☆
modern-coreprototypeOpen-Vocabulary Panoptic Occupancy is rated for BEV, occupancy, or freespace modeling that feeds planning-facing autonomy stacks.
OpenADLearning: ★★★★☆
Deployment: ★★★★☆
modern-corefielded-patternOpen-world benchmark for corner cases and unseen categories.
RadarPillarsLearning: ★★★★☆
Deployment: ★★★★☆
classic-baselineprototypeCore radar-native detection baseline for weather-robust perception.
RenderOccLearning: ★★★★☆
Deployment: ★★★★☆
modern-coreprototypeRenderOcc is rated for BEV, occupancy, or freespace modeling that feeds planning-facing autonomy stacks.
SegNet4DLearning: ★★★★☆
Deployment: ★★★★☆
modern-coreprototypeSegNet4D is rated for motion segmentation, scene flow, or dynamic-object perception workflows.
SelfOccLearning: ★★★★☆
Deployment: ★★★★☆
modern-coreprototypeSelfOcc is rated for BEV, occupancy, or freespace modeling that feeds planning-facing autonomy stacks.
SOLOFusionLearning: ★★★★☆
Deployment: ★★★★☆
modern-coreprototypeSOLOFusion is rated for BEV, occupancy, or freespace modeling that feeds planning-facing autonomy stacks.
SparseOccLearning: ★★★★☆
Deployment: ★★★★☆
modern-coreprototypeSparseOcc is rated for BEV, occupancy, or freespace modeling that feeds planning-facing autonomy stacks.
Spatiotemporal Memory Occupancy FlowLearning: ★★★★☆
Deployment: ★★★★☆
modern-coreprototypeSpatiotemporal Memory Occupancy Flow is rated for BEV, occupancy, or freespace modeling that feeds planning-facing autonomy stacks.
Streaming Gaussian OccupancyLearning: ★★★★☆
Deployment: ★★★★☆
modern-coreprototypeStreaming Gaussian Occupancy is rated for BEV, occupancy, or freespace modeling that feeds planning-facing autonomy stacks.
StreamingFlowLearning: ★★★★☆
Deployment: ★★★★☆
modern-coreprototypeStreamingFlow is rated for motion segmentation, scene flow, or dynamic-object perception workflows.
StreamMOSLearning: ★★★★☆
Deployment: ★★★★☆
modern-coreprototypeStreamMOS is rated for motion segmentation, scene flow, or dynamic-object perception workflows.
SurroundOccLearning: ★★★★☆
Deployment: ★★★★☆
modern-coreprototypeFoundational camera occupancy reference for planning-facing perception.
TPVFormerLearning: ★★★★☆
Deployment: ★★★★☆
modern-coreprototypeTPVFormer is rated for BEV, occupancy, or freespace modeling that feeds planning-facing autonomy stacks.
TrackOccLearning: ★★★★☆
Deployment: ★★★★☆
modern-coreprototypeTrackOcc is rated for BEV, occupancy, or freespace modeling that feeds planning-facing autonomy stacks.
3D-KNN Blind-Spot DesnowingLearning: ★★★☆☆
Deployment: ★★★★☆
deployment-patternpilot-proven3D-KNN Blind-Spot Desnowing is rated for cleaning, stress testing, or failure detection in degraded perception conditions.
4D Radar Road Boundaries and FreespaceLearning: ★★★☆☆
Deployment: ★★★★☆
deployment-patternprototype4D Radar Road Boundaries and Freespace is rated for alternative-sensor perception and adverse-weather fallback evaluation.
4D Radar-Camera OccupancyLearning: ★★★☆☆
Deployment: ★★★★☆
deployment-patternprototype4D Radar-Camera Occupancy is rated for alternative-sensor perception and adverse-weather fallback evaluation.
4DMOSLearning: ★★★☆☆
Deployment: ★★★★☆
modern-coreprototypeExtends LiDAR motion segmentation with temporal 4D reasoning.
Adverse-Weather Radar-LiDAR 3D DetectionLearning: ★★★☆☆
Deployment: ★★★★☆
deployment-patternprototypeAdverse-Weather Radar-LiDAR 3D Detection is rated for alternative-sensor perception and adverse-weather fallback evaluation.
AdverseNetLearning: ★★★☆☆
Deployment: ★★★★☆
deployment-patternpilot-provenAdverseNet is rated for cleaning, stress testing, or failure detection in degraded perception conditions.
AevaScenesLearning: ★★★☆☆
Deployment: ★★★★☆
referencefielded-patternAevaScenes is rated as a benchmark or dataset reference for perception robustness and validation coverage.
AIDELearning: ★★★☆☆
Deployment: ★★★★☆
deployment-patternpilot-provenAIDE is rated for operational perception validation, calibration, or safety-screening workflows.
Classical LiDAR Outlier RemovalLearning: ★★★☆☆
Deployment: ★★★★☆
deployment-patternpilot-provenClassical LiDAR Outlier Removal is rated for cleaning, stress testing, or failure detection in degraded perception conditions.
CVFusionLearning: ★★★☆☆
Deployment: ★★★★☆
modern-coreprototypeImportant radar-camera fusion method for degraded visual conditions.
DenoiseCP-NetLearning: ★★★☆☆
Deployment: ★★★★☆
deployment-patternpilot-provenDenoiseCP-Net is rated for cleaning, stress testing, or failure detection in degraded perception conditions.
Ev-3DODLearning: ★★★☆☆
Deployment: ★★★★☆
deployment-patternprototypeEv-3DOD is rated for alternative-sensor perception and adverse-weather fallback evaluation.
EvOccLearning: ★★★☆☆
Deployment: ★★★★☆
deployment-patternprototypeEvOcc is rated for alternative-sensor perception and adverse-weather fallback evaluation.
Fail2DriveLearning: ★★★☆☆
Deployment: ★★★★☆
deployment-patternpilot-provenFail2Drive is rated for operational perception validation, calibration, or safety-screening workflows.
K-RadarLearning: ★★★☆☆
Deployment: ★★★★☆
modern-corefielded-patternKey 4D radar dataset and benchmark for all-weather perception evaluation.
LASPLearning: ★★★☆☆
Deployment: ★★★★☆
deployment-patternpilot-provenLASP is rated for operational perception validation, calibration, or safety-screening workflows.
LiDAR Weather Artifact RemovalLearning: ★★★☆☆
Deployment: ★★★★☆
deployment-patternpilot-provenLiDAR Weather Artifact Removal is rated for cleaning, stress testing, or failure detection in degraded perception conditions.
LIORNetLearning: ★★★☆☆
Deployment: ★★★★☆
deployment-patternpilot-provenLIORNet is rated for cleaning, stress testing, or failure detection in degraded perception conditions.
LiSnowNetLearning: ★★★☆☆
Deployment: ★★★★☆
deployment-patternpilot-provenLiSnowNet is rated for cleaning, stress testing, or failure detection in degraded perception conditions.
M-detector LiDAR Point-Stream MEDLearning: ★★★☆☆
Deployment: ★★★★☆
deployment-patternpilot-provenM-detector LiDAR Point-Stream MED is rated for operational perception validation, calibration, or safety-screening workflows.
MoMELearning: ★★★☆☆
Deployment: ★★★★☆
modern-coreprototypeUseful resilient fusion pattern for adverse sensor failure cases.
MSC-BenchLearning: ★★★☆☆
Deployment: ★★★★☆
referencefielded-patternMSC-Bench is rated as a benchmark or dataset reference for perception robustness and validation coverage.
MultiCorruptLearning: ★★★☆☆
Deployment: ★★★★☆
referencefielded-patternMultiCorrupt is rated as a benchmark or dataset reference for perception robustness and validation coverage.
Occluded nuScenesLearning: ★★★☆☆
Deployment: ★★★★☆
referencefielded-patternOccluded nuScenes is rated as a benchmark or dataset reference for perception robustness and validation coverage.
POD FMCW LiDAR Predictive DetectionLearning: ★★★☆☆
Deployment: ★★★★☆
deployment-patternprototypePOD FMCW LiDAR Predictive Detection is rated for alternative-sensor perception and adverse-weather fallback evaluation.
ProOODLearning: ★★★☆☆
Deployment: ★★★★☆
deployment-patternpilot-provenProOOD is rated for cleaning, stress testing, or failure detection in degraded perception conditions.
RaCFormerLearning: ★★★☆☆
Deployment: ★★★★☆
deployment-patternprototypeRaCFormer is rated for alternative-sensor perception and adverse-weather fallback evaluation.
RC-AutoCalibLearning: ★★★☆☆
Deployment: ★★★★☆
deployment-patternpilot-provenRC-AutoCalib is rated for operational perception validation, calibration, or safety-screening workflows.
RobuRCDetLearning: ★★★☆☆
Deployment: ★★★★☆
deployment-patternprototypeRobuRCDet is rated for alternative-sensor perception and adverse-weather fallback evaluation.
S2R-BenchLearning: ★★★☆☆
Deployment: ★★★★☆
referencefielded-patternS2R-Bench is rated as a benchmark or dataset reference for perception robustness and validation coverage.
SLiDELearning: ★★★☆☆
Deployment: ★★★★☆
deployment-patternpilot-provenSLiDE is rated for cleaning, stress testing, or failure detection in degraded perception conditions.
Sparse4DLearning: ★★★☆☆
Deployment: ★★★★☆
modern-coreprototypePractical sparse-query direction for camera 3D detection and tracking.
TripleMixerLearning: ★★★☆☆
Deployment: ★★★★☆
deployment-patternpilot-provenTripleMixer is rated for cleaning, stress testing, or failure detection in degraded perception conditions.
V2X-RadarLearning: ★★★☆☆
Deployment: ★★★★☆
deployment-patternprototypeV2X-Radar is rated for alternative-sensor perception and adverse-weather fallback evaluation.
3D-AVSLearning: ★★★☆☆
Deployment: ★★★☆☆
frontierresearch3D-AVS is rated for open-world perception, annotation leverage, and long-tail validation workflows.
3D-OutDetLearning: ★★★☆☆
Deployment: ★★★☆☆
modern-coreprototype3D-OutDet is rated as a supporting perception method for autonomy-stack triage and follow-up reading.
ClipomalyLearning: ★★★☆☆
Deployment: ★★★☆☆
frontierresearchUseful anomaly-detection reference for long-tail discovery workflows.
CoHFFLearning: ★★★☆☆
Deployment: ★★★☆☆
frontierprototypeCoHFF is rated for cooperative perception and infrastructure-assisted sensing evaluation.
CoInfraLearning: ★★★☆☆
Deployment: ★★★☆☆
frontierprototypeCoInfra is rated for cooperative perception and infrastructure-assisted sensing evaluation.
CoopTrackLearning: ★★★☆☆
Deployment: ★★★☆☆
frontierprototypeCoopTrack is rated for cooperative perception and infrastructure-assisted sensing evaluation.
CoSDHLearning: ★★★☆☆
Deployment: ★★★☆☆
modern-coreprototypeCoSDH is rated as a supporting perception method for autonomy-stack triage and follow-up reading.
DetAny3DLearning: ★★★☆☆
Deployment: ★★★☆☆
frontierresearchDetAny3D is rated for open-world perception, annotation leverage, and long-tail validation workflows.
DistillNeRFLearning: ★★★☆☆
Deployment: ★★★☆☆
frontierresearchDistillNeRF is rated for neural scene representation learning and simulation-oriented perception research.
DrivingGaussianLearning: ★★★☆☆
Deployment: ★★★☆☆
frontierresearchDrivingGaussian is rated for neural scene representation learning and simulation-oriented perception research.
ForeSightLearning: ★★★☆☆
Deployment: ★★★☆☆
modern-coreprototypeForeSight is rated as a supporting perception method for autonomy-stack triage and follow-up reading.
GaussianFormerLearning: ★★★☆☆
Deployment: ★★★☆☆
frontierresearchGaussianFormer is rated for neural scene representation learning and simulation-oriented perception research.
HoloVICLearning: ★★★☆☆
Deployment: ★★★☆☆
frontierprototypeHoloVIC is rated for cooperative perception and infrastructure-assisted sensing evaluation.
HUGS Urban GaussiansLearning: ★★★☆☆
Deployment: ★★★☆☆
frontierresearchHUGS Urban Gaussians is rated for neural scene representation learning and simulation-oriented perception research.
Mosaic3DLearning: ★★★☆☆
Deployment: ★★★☆☆
frontierresearchMosaic3D is rated for open-vocabulary 3D segmentation, dataset leverage, and long-tail perception validation.
OP3DetLearning: ★★★☆☆
Deployment: ★★★☆☆
frontierresearchOP3Det is rated for open-world perception, annotation leverage, and long-tail validation workflows.
Open3DTrackLearning: ★★★☆☆
Deployment: ★★★☆☆
frontierresearchOpen3DTrack is rated for open-world perception, annotation leverage, and long-tail validation workflows.
OpenVoxLearning: ★★★☆☆
Deployment: ★★★☆☆
frontierresearchOpenVox is rated for open-world perception, annotation leverage, and long-tail validation workflows.
OVAD And OVODA Open-Vocabulary 3D AttributesLearning: ★★★☆☆
Deployment: ★★★☆☆
frontierresearchOVAD And OVODA Open-Vocabulary 3D Attributes is rated for open-world perception, annotation leverage, and long-tail validation workflows.
OW-OVDLearning: ★★★☆☆
Deployment: ★★★☆☆
frontierresearchOW-OVD is rated for open-world perception, annotation leverage, and long-tail validation workflows.
RCooperLearning: ★★★☆☆
Deployment: ★★★☆☆
frontierfielded-patternCooperative-perception dataset relevant to infrastructure-assisted sensing.
S2MLearning: ★★★☆☆
Deployment: ★★★☆☆
frontierresearchS2M is rated for open-world perception, annotation leverage, and long-tail validation workflows.
SAM 3Learning: ★★★☆☆
Deployment: ★★★☆☆
frontierresearchSAM 3 is rated for open-world perception, annotation leverage, and long-tail validation workflows.
SAM4DLearning: ★★★☆☆
Deployment: ★★★☆☆
frontierresearchSAM4D is rated for open-world perception, annotation leverage, and long-tail validation workflows.
SAMFusionLearning: ★★★☆☆
Deployment: ★★★☆☆
frontierresearchSAMFusion is rated for open-world perception, annotation leverage, and long-tail validation workflows.
SOACLearning: ★★★☆☆
Deployment: ★★★☆☆
modern-coreprototypeSOAC is rated as a supporting perception method for autonomy-stack triage and follow-up reading.
SplatADLearning: ★★★☆☆
Deployment: ★★★☆☆
frontierresearchSplatAD is rated for neural scene representation learning and simulation-oriented perception research.
SplatFlowLearning: ★★★☆☆
Deployment: ★★★☆☆
frontierresearchSplatFlow is rated for neural scene representation learning and simulation-oriented perception research.
TacoDepthLearning: ★★★☆☆
Deployment: ★★★☆☆
modern-coreprototypeTacoDepth is rated as a supporting perception method for autonomy-stack triage and follow-up reading.
V2X-ReaLOLearning: ★★★☆☆
Deployment: ★★★☆☆
frontierprototypeV2X-ReaLO is rated for cooperative perception and infrastructure-assisted sensing evaluation.
WildDet3DLearning: ★★★☆☆
Deployment: ★★★☆☆
modern-coreprototypeWildDet3D is rated as a supporting perception method for autonomy-stack triage and follow-up reading.

Domain Fit Guidance

Generic method pages should use Domain Fit, not Airside Fit, as the default deployment lens. Use three to six compact rows or bullets rather than a large matrix.

DomainFitNote
Road AVstrong / conditional / weak / insufficient evidenceState whether the method has road-scale evidence, actor coverage, and runtime maturity.
Airsidestrong / conditional / weak / insufficient evidenceInclude apron, GSE, FOD, aircraft-proximity, and weather relevance only when supported by the method evidence.
Warehouse / logistics yard / port / mining / construction / agriculture / delivery robot / outdoor campusstrong / conditional / weak / insufficient evidenceAdd only the domains where the method assumptions or validation signals materially transfer.

Airside-specific pages may stay airside-first, but generic pages should not make airside the only deployment lens.

How to Use This Library

For loss and residual foundations behind perception methods, use 3D Object Detection Losses and Assignment for detector training objectives and Robust Losses and M-Estimators for outlier-heavy geometric residuals that connect perception outputs to calibration, tracking, and SLAM.

NeedStart here
Camera BEV, occupancy, and freespaceBEVDet, BEVDepth, BEVStereo, SOLOFusion, Sparse4D, TPVFormer, SurroundOcc, SparseOcc, FlashOcc, SelfOcc, RenderOcc, LiDAR-Camera Occupancy Fusion, Dynamic Occupancy and Freespace, Spatiotemporal Memory Occupancy Flow
Gaussian, 3DGS, 4DGS, and 4D occupancySplatAD, GaussianFormer, GaussianOcc, Streaming Gaussian Occupancy, Cam4DOcc, StreamingFlow, DrivingGaussian, HUGS, SplatFlow, DistillNeRF
LiDAR motion, scene flow, and temporal segmentationLiDAR-MOS, 4DMOS, InsMOS, StreamMOS, 4DSegStreamer, SegNet4D, Mask4D, Instantaneous Motion Perception, MotionSeg3D, MambaMOS, Neural Scene Flow Priors, Cross-Domain LiDAR Scene Flow, TrackOcc
LiDAR denoising, removal, and adverse weatherLIORNet, LiSnowNet, SLiDE, TripleMixer, 3D-KNN Blind-Spot Desnowing, 3D-OutDet, AdverseNet, DenoiseCP-Net, Classical LiDAR Outlier Removal, LiDAR Weather Artifact Removal
Radar, 4D radar, event, and FMCW perceptionRadarPillars, K-Radar, V2X-Radar, TacoDepth, RaCFormer, CVFusion, 4D Radar-Camera Occupancy, Adverse-Weather Radar-LiDAR 3D Detection, RobuRCDet, SAMFusion, POD FMCW LiDAR Predictive Detection, Ev-3DOD, AevaScenes
Open-world and open-vocabulary perceptionOpenAD, OP3Det, WildDet3D, DetAny3D, OW-OVD, Clipomaly, S2M, SAM 3, 3D-AVS, Mosaic3D, OpenVox, OVAD/OVODA Open-Vocabulary 3D Attributes, Open-Vocabulary Panoptic Occupancy
Robust fusion and perception validationMoME, GraphBEV, SOAC, RC-AutoCalib, ASF, MSC-Bench, MultiCorrupt, S2R-Bench, Occluded nuScenes, Conformal Boxes
Cooperative, online, and data-engine methodsRCooper, HoloVIC, CoInfra, V2X-ReaLO, CoHFF, CoSDH, CoopTrack, LASP, Fail2Drive, AIDE

File Boundary Rules

RulePractical meaning
One file, one methodA page should not bundle multiple unrelated methods just because they share a modality. If two papers solve the same exact technique lineage, the page can compare versions, but the title must still name the primary method.
Overview pages link outExisting files such as BEV Encoding Architectures, Streaming Temporal Perception, and Infrastructure Cooperative Perception should summarize families and point here for method-level details.
Benchmarks count as methods when they shape evaluationPages such as MSC-Bench, S2R-Bench, LASP, OpenAD, and Fail2Drive deserve first-class treatment because they define what a deployment team measures.
Domain fit is mandatory for generic pagesGeneric method pages should state the domains where the method is a strong, conditional, weak, or insufficient-evidence fit. Airside-specific pages may use a transfer note instead.
Sources stay close to claimsEach method page must include primary paper, project, dataset, or repository links so future refreshes can verify claims quickly.

Standard Page Shape

Each method page should include:

  1. What the method is.
  2. Core technical idea.
  3. Inputs, outputs, and model/data assumptions.
  4. Architecture or pipeline.
  5. Training/evaluation setup and benchmark signals.
  6. Strengths.
  7. Failure modes and deployment risks.
  8. Domain Fit, or an airside transfer note when the page is explicitly airside-specific.
  9. Implementation notes.
  10. Sources.

Relationship to the Perception Stack

Existing synthesis pageMethod-library role
BEV Encoding ArchitecturesExplains the BEV design space, then links to BEVDet/BEVDepth/BEVStereo/SOLOFusion and camera occupancy methods.
Camera-Only Degraded PerceptionUses camera BEV, occupancy, depth, and open-vocabulary method pages to define fallback modes.
LiDAR Semantic SegmentationSummarizes segmentation architecture choices, then links to LiDAR-MOS, 4DMOS, SegNet4D, Mask4D, MotionSeg3D, MambaMOS, neural scene-flow priors, and HeLiMOS-style evaluation.
LiDAR Artifact Removal TechniquesSynthesizes learned denoisers, classical filters, weather artifact handling, ghost/multipath failures, validation, datasets, and map-cleaning links.
Streaming Temporal PerceptionConnects StreamMOS, 4DSegStreamer, MotionSeg3D, MambaMOS, LASP, sparse-query detection, scene flow, and temporal occupancy into a runtime stack.
Open-Vocabulary and Zero-Shot DetectionStays as the broad open-vocabulary primer; OpenAD, OP3Det, WildDet3D, DetAny3D, OW-OVD, Clipomaly, S2M, and SAM 3 get individual pages here.
Infrastructure Cooperative PerceptionSynthesizes V2X deployment tradeoffs; RCooper, HoloVIC, CoInfra, V2X-ReaLO, CoHFF, CoSDH, and CoopTrack live here as atomic references.
Production Perception SystemsUses this library as the evidence base for validation matrices, degradation policies, and sensor-suite decisions.

Expansion Backlog

The first waves focused on methods already identified as P0/P1 in the Perception Coverage Audit. The 2026-05-09 loops promoted SplatAD, GaussianFormer, GaussianOcc, streaming Gaussian occupancy, Cam4DOcc, StreamingFlow, Sparse4D, TacoDepth, RaCFormer, LIORNet, learned LiDAR desnowing/denoising, broad artifact removal, classical outlier filtering, MotionSeg3D, MambaMOS, neural scene-flow priors, CVFusion, 4D radar-camera occupancy, POD/FMCW LiDAR, DrivingGaussian, HUGS, SplatFlow, DistillNeRF, TrackOcc, cross-domain scene flow, LiDAR-camera occupancy fusion, dynamic occupancy/free-space, radar-LiDAR adverse-weather detection, RobuRCDet, SAMFusion, spatiotemporal memory occupancy flow, OVAD/OVODA, and open-vocabulary panoptic occupancy into atomic files. Future waves should split remaining grouped rows into atomic pages, especially:

  • VEON, EvOcc, ProOOD, SA-Occ, DR-REMOVER, and ExelMap.
  • Drive-OccWorld and DFIT-OccWorld where they need separate world-model or planning-facing treatment beyond the dynamic occupancy page.
  • SparseBEV, DETR4D, DySS, and ForeSight.
  • DepthOcc, LinkOcc, missing-view occupancy, Gaussian-rendered occupancy, SAM4D, and related 2026 radar/occupancy follow-ons.
  • SparseCoop, CoDS, JigsawComm, QuantV2X, TruckV2X, and collaborative Gaussian occupancy.
  • DriveBench, Airport-FOD3S data-engine pages, DSERT-RoLL, CMHT, embodied robotics 3D perception, indoor open-vocabulary 3D instance segmentation, and airside-specific dust/de-icing-mist datasets.

Sources

Public research notes collected from public sources.