OccProphet: Pushing Efficiency Frontier of Camera-Only 4D Occupancy Forecasting with Observer-Forecaster-Refiner Framework
Abstract
Predicting variations in complex traffic environments is crucial for the safety of autonomous driving. Recent advancements in occupancy forecasting have enabled forecasting future 3D occupied status in driving environments by observing historical 2D images. However, high computational demands make occupancy forecasting less efficient during training and inference stages, hindering its feasibility for deployment on edge agents. In this paper, we propose a novel framework, i.e., OccProphet, to efficiently and effectively learn occupancy forecasting with significantly lower computational requirements while improving forecasting accuracy. OccProphet comprises three lightweight components: Observer, Forecaster, and Refiner. The Observer extracts spatio-temporal features from 3D multi-frame voxels using the proposed Efficient 4D Aggregation with Tripling-Attention Fusion, while the Forecaster and Refiner conditionally predict and refine future occupancy inferences. Experimental results on nuScenes, Lyft-Level5, and nuScenes-Occupancy datasets demonstrate that OccProphet is both training- and inference-friendly. OccProphet reduces 58%78% of the computational cost with a 2.6 speedup compared with the state-of-the-art Cam4DOcc. Moreover, it achieves 4%18% relatively higher forecasting accuracy. Code and models are publicly available at https://github.com/JLChen-C/OccProphet.
1 Introduction


Autonomous driving holds significant promise for reshaping transportation and urban mobility. Perceiving 3D surroundings is critical for autonomous vehicles. There are typically two 3D perception paradigms: detection-based perception and occupancy-based perception. The detection-based paradigm, such as monocular or multi-view 3D object detection (Chen et al., 2016; Wang et al., 2021; Huang et al., 2021; Wang et al., 2022), equips autonomous vehicles with 3D perception capabilities by detecting traffic participants and assigning them 3D bounding boxes. However, due to limitations in pre-defined object categories and rigid detection boxes, the detection-based paradigm struggles to generalize to unknown objects and irregular structures, which are likely to appear in real-world traffic scenarios. To relieve these constraints, the occupancy-based perception paradigm (Huang et al., 2023a; Wei et al., 2023b; Wang et al., 2023b; Tian et al., 2023; Xu et al., 2025) offers a more general and fine-grained representation of the environment through learning occupied states in 3D space. This paradigm provides a stronger perception ability to autonomous vehicles, enabling them to better comprehend complex traffic environments.
Despite advancements in occupancy perception, most existing methods can only perceive the past and present states of the environment. They lack the ability to capture and understand environmental dynamics and subsequently forecast the future scene. Forecasting is essential for safe planning, which assists autonomous vehicles to avoid potential collisions (Ding et al., 2019; Song et al., 2020; Ding et al., 2021; Huang et al., 2023b; Pan et al., 2024). Although some bird’s-eye view (BEV) approaches have achieved object motion forecasting in the environment(Hu et al., 2021; Zhang et al., 2022; Fang et al., 2023; Ferenczi et al., 2024), their forecasts are restricted on a 2D plane. This 2D forecasting limits the comprehensive understanding of the entire 3D dynamic scene. In light of this limitation, occupancy forecasting methods (Weng et al., 2021; 2022; Mersch et al., 2022; Khurana et al., 2023; Agro et al., 2024) shift to predict future 3D occupancy for the whole environment. However, these methods rely on point cloud inputs from expensive LiDAR kits. To explore more cost-effective solutions, Cam4DOcc (Ma et al., 2024a) introduces a camera-only benchmark and baseline for occupancy forecasting.
While Cam4DOcc achieves remarkable performance in occupancy forecasting compared to its counterparts, the high computational cost makes it less efficient during training and inference. It hampers feasibility of deployment on edge agents, such as autonomous vehicles that operate under restricted computational budgets. In this paper, we propose a novel framework, dubbed as OccProphet (shown in Figure 2), to efficiently and effectively perform camera-only occupancy forecasting. In OccProphet, we design three lightweight components to forecast future states: the Observer, Forecaster, and Refiner. The Observer adopts 4D feature aggregation and a tripling-attention fusion strategy on the reduced-resolution features to extract spatio-temporal information efficiently. The Forecaster then infers future states according to the scene condition and the Observer’s outputs. Finally, the Refiner enhances the quality of the forecast results through spatio-temporal interactions. The main contributions of this paper are summarized as follows:
-
•
We propose OccProphet, a novel camera-only occupancy forecasting framework, which is both efficient and effective during training and inference, towards on-vehicle deployment.
-
•
We design a lightweight Observer-Forecaster-Refiner pipeline for OccProphet. The Observer extracts spatio-temporal features from historical observations; the Forecaster conditionally predicts coarse future states; the Refiner promotes forecasting accuracy.
-
•
Experimental results demonstrate that OccProphet achieves higher forecasting accuracy with less than half the computational cost of Cam4DOcc. These improvements are consistently observed across the nuScenes (Caesar et al., 2020), Lyft-Level5 (Houston et al., 2021), and nuScenes-Occupancy (Wang et al., 2023a) datasets, highlighting the superior efficiency and effectiveness of OccProphet (shown in Figure 2).
2 Related Work
2.1 Occupancy Prediction
Occupancy prediction aims at modeling the current 3D occupancy layout in space, by observing historical and current environments. Occupancy prediction is adept at providing 3D dense descriptions for complex traffic scenarios, thereby garnering increasing attention from academia and industry. SSCNet (Song et al., 2017) was the first semantic occupancy prediction work, which simultaneously predicted occupied voxels and their semantics for an indoor scene using a depth image. MonoScene (Cao & De Charette, 2022) extends SSCNet to outdoor scenarios by using an RGB image and incorporating stronger supervisions. Training and evaluating occupancy prediction networks require benchmarks with ground truth occupancy labels, which are challenging due to the complexity of densely annotating 3D outdoor driving scenes. Wang et al. (Wang et al., 2023b) propose OpenOccupancy, the first large-scale benchmark for semantic occupancy prediction, which covers multiple sensing modalities and provides high-resolution dense occupancy annotations. Occ3d, developed by Tian et al. (Tian et al., 2023), is another widely used benchmark for occupancy prediction, whose high-quality labels are from the proposed technique of image-guided occupancy label refinement.
Considering the inherently dense nature of depicting 3D environments, ideal occupancy perception emphasizes balancing efficiency and accuracy. Although some studies explore sparse queries (Li et al., 2023b) or tri-perspective view (TPV) representation (Huang et al., 2023a) to elevate the efficiency of occupancy prediction, they inevitably sacrifice fine-grained details of 3D space. In contrast, many other methods (Zhang et al., 2023; Wei et al., 2023b; Ma et al., 2024b) utilize 3D feature volumes to preserve 3D details of the scene, leading to higher occupancy prediction accuracy. Recently, COTR (Ma et al., 2024b) proposes a compact occupancy representation that preserves geometric details while reducing computational costs. Despite the significant advancements in occupancy prediction, including comprehensive benchmarks and powerful algorithms, all existing methods focus exclusively on current occupancy and overlook future occupancy, which reflects potential variations in the 3D environment.
2.2 Occupancy Forecasting
Occupancy forecasting targets to predict future occupancy, starting from the current timestamp. Previous dominant works mainly adopt the BEV perspective for occupancy forecasting, reasoning about 2D occupancy changes on a BEV plane. For example, FIERY (Hu et al., 2021) extracts BEV features from multi-view image inputs and utilizes a temporal model with 3D convolution to capture spatio-temporal states, which are then used to recursively forecast future instances states. BEVerse (Zhang et al., 2022) introduces a unified BEV representation framework that jointly achieves object perception and occupancy forecasting using multi-task supervision. To better align spatio-temporal information, TBP-Former (Fang et al., 2023) designs a pose-synchronized BEV encoder to synchronize multi-frame BEV features during occupancy forecasting. While these BEV-based methods deliver impressive performance for forecasting pre-defined semantic categories (e.g., vehicles), they struggle to (1) forecast the motion of out-of-distribution objects and (2) capture height information in the environment. In contrast, our method forecasts class-agnostic occupancy from a 3D perspective, rather than BEV, thus enabling autonomous vehicles to monitor, comprehend, and reason about 3D dynamics in the physical world.
To address the limitations of BEV occupancy forecasting, researchers have recently shifted their focus toward forecasting 3D occupancy without considering semantics. Specifically, Khurana et al. (Khurana et al., 2023) treat LiDAR point cloud forecasting as a proxy task for the occupancy forecasting task, where point cloud rendering is used to bridge the two tasks. UnO (Agro et al., 2024) takes LiDAR point clouds as input and performs occupancy forecasting using the proposed unsupervised learning paradigm, in which the forecasted occupancy should align with pseudo occupancy labels generated from future LiDAR data. However, these methods rely on point clouds from expensive LiDAR kits, leading to increased costs when implemented in autonomous vehicles.
Compared to LiDAR-based approaches, camera-only occupancy forecasting offers a promising alternative with significantly lower costs. Cam4DOcc (Ma et al., 2024a) introduces a comprehensive benchmark and dataset to evaluate camera-only occupancy forecasting algorithms on both movable and static objects beyond pre-defined categories. Additionally, it proposes a strong camera-only baseline for occupancy forecasting. However, this approach is still far from real-world application due to its high computational demands. Drive-OccWorld (Yang et al., 2024) introduces extra action condition inputs and planning supervision to enhance performance. In this paper, we propose a novel end-to-end framework with faster speed, and higher accuracy, enabling efficient and effective occupancy forecasting in a pure camera-only setting.
3 OccProphet
3.1 Overview
The overview of the proposed OccProphet is illustrated in Figure 3. Given multi-frame surround-view RGB images as input, 2D features are extracted using a shared image encoder. These 2D features are subsequently lifted into 3D space and aggregated into multi-frame 3D voxel features through depth estimation and voxel pooling. We design the following unshared pipeline for each of the occupancy and occupancy flow branches. Specifically, the pipeline consists of four components: the Observer, Forecaster, Refiner, and Predictor. The Observer module efficiently and effectively aggregates spatio-temporal information within multi-frame observations (i.e., multi-frame 3D voxel features). The Observer’s output then undergoes a Forecaster, which adaptively predicts future states, ensuring flexibility across diverse traffic conditions. The Refiner module further enhances the quality of these predictions by enabling cross-frame interactions. Finally, the Predictor module decodes the refined future states into either occupancy or occupancy flow.

3.2 Observer
The Observer takes the 4D motion-aware feature as input, and generates a spacetime-aware representation. Let denote the multi-camera RGB images at timestamp , where refers to the total number of surround-view cameras, and is the total number of input frames. A shared image encoder (i.e., ResNet (He et al., 2016)) is applied to to extract 2D features. For each frame, these 2D features are projected to 3D space and then aggregated into voxelized 3D features (Philion & Fidler, 2020). The 3D features from multiple frames are aligned into the current-frame coordinate system using 6 degrees of freedom (6-DoF) ego-vehicle poses. We then concatenate the aligned 3D features into a 4D feature , where is the number of channels, and represents the size of voxelized 3D feature volume. Subsequently, the motion-aware 4D feature is generated by concatenating 6-DoF ego-vehicle poses.
To generate the spacetime-aware representation, directly processing the 4D motion-aware feature using convolutional operations is intuitive. However, the direct processing imposes a substantial computational burden, and ignores the fact that a large portion of the 3D space is unoccupied, which leads to the inherent sparsity of the motion-aware feature. To address this issue, we efficiently and effectively generate the spacetime-aware feature from using the Observer module. The Observer comprises an Efficient 4D Aggregation module and a Tripling-Attention Fusion module.
3.2.1 Efficient 4D Aggregation

Directly aggregating the original 4D feature will incur a high computational cost. For efficiency, we design the Efficient 4D Aggregation (E4A) module to first produce compact features through downsampling, and then exploit spatio-temporal interactions on the compact features to achieve aggregation, followed by the upsampling process to compensate for the information loss. The architecture of the E4A module is shown in Figure 4. We first reduce the channel number of from to through a 3D convolution, thereby forming a feature . We gradually downsample to reduce computation during aggregation. However, such downsampling inevitably results in non-negligible information loss, especially for small objects. To compensate for this loss, we conduct two operations. On one hand, we perform spatio-temporal interactions on the downsampled features, that is, the Tripling-Attention Fusion module (to be illustrated in Section 3.2.2). On the other hand, the post-interaction features are upsampled, and further summed with the features at the same resolution prior to downsampling. These two operations continue until the resolution of the upsampled feature matches the resolution of the initial motion-aware feature.
Style | Np (M) | FLOPs (G) | (%) |
---|---|---|---|
BEV | 102 | 1370 | 26.50 |
TPV | 181 | 1574 | 26.96 |
E4A | 67 | 1737 | 27.50 |
The output 4D representation of the E4A module possesses spatio-temporal awareness. It effectively preserves the 3D geometric details of the environment–details that the BEV and TPV representations are unable to capture. As shown in Table 1, using the E4A representation achieves higher performance than the BEV and TPV representations with fewer parameters and a slight increase in computational costs.
3.2.2 Tripling-Attention Fusion

A 4D feature can be considered as a combination of multiple 3D voxel-wise features along the temporal dimension. The Tripling-Attention Fusion (TAF, shown in Figure 5) module is specifically designed to facilitate spatio-temporal interactions across multiple 3D features, as depicted on the left of Figure 5. Notably, downsampling reduces computational cost by lowering feature resolution, while the Tripling-Attention Fusion module makes a further step through the proposed tripling operation, as illustrated on the right of Figure 5.
The tripling operation is designed to understand the 3D space from three complementary and compact perspectives, which can retain 3D scene information with fewer computational costs. Specifically, a tripling operation decomposes a 3D feature into three distinct branches: scene, height, and BEV. This decomposition compresses 3D features into 1D and 2D features, significantly reducing the subsequent computational overhead. The scene branch can extract the global context of the corresponding frame, providing an overall understanding of the scenario. The height branch retains vertical details, serving as complementary clues to the 2D BEV branch to enhance the representation capability of the 3D geometric information. The three branches can be computed as follows:
(1) | |||
(2) | |||
(3) |
where denotes the -th downsampled feature input to the Tripling-Attention Fusion module. GAP, GAP, and GAP are 3D, 2D, and 1D global average pooling. Linear, Conv, Norm, and Act refer to a single linear layer, 1D convolution, normalization, and activation, respectively. W-MSA is the window-based multi-head self-attention block (Liu et al., 2021). , , and denote the outputs of scene, height, and BEV branches. After the tripling operation, we interact and fuse scene, height, and BEV features. Specifically, we first independently apply the temporal attention along the time axis to different branches, then sum over the three branches using the broadcast technique. The process is formulated as follows:
(4) |
where , , and denote temporal attention for scene, height, and BEV branches; is broadcast-style plus; is the output feature of the TAF module.
3.3 Forecaster

Given a spatio-temporal representation output from the Observer, the Forecaster (shown in Figure 6) is supposed to generate future states. We first reshape by collapsing the time axis into the channel axis, resulting in the reshaped features . A straightforward approach to forecasting is using a single linear layer to predict features for future frames. This approach encourages the network to learn static weights to fit different traffic scenarios. However, traffic situations vary significantly across different traffic scenarios, presenting distinct challenges in predicting future changes using a single linear layer. For instance, forecasting environmental changes is more difficult in a crowded intersection with many moving objects than on an empty highway with few vehicles. The spatio-temporal complexity in the former is far greater than in the latter.
Considering these challenges, we propose forecasting occupancy with flexibility to adapt to various traffic scenarios featuring diverse spatio-temporal complexities. To achieve this, we design a novel Forecaster module that predicts future states based on the overall environmental condition. The Forecaster comprises a Condition Generator and a Conditional Forecaster. We first use a Condition Generator comprising a 3D GAP and a shared linear layer across different frames to extract the condition from the observation :
(5) |
where denotes the overall environmental condition, which is then reshaped into by collapsing the time axis into the channel axis. is subsequently passed to a Conditional Forecaster to predict future states. Specifically, a linear layer is applied to to produce adaptive weights for specific traffic scenarios. Another linear layer is then used to predict future states conditioned by these adaptive weights.
(6) |
where denotes the adaptive weights, is the future states of future frames. is reshaped into to recover the time axis. , as the preliminary feature for future environments, is then refined by the Refiner module.
3.4 Refiner
Since the Forecaster module predicts using linear projection, it inevitably lacks cross-frame interactions. The Refiner is designed to enhance the forecasted results via further interactions between future frames, as well as incorporating historical frames as supplementary information. The E4A module, described in Section Section 3.2.1, is a spatio-temporal interaction module. Furthermore, taking a review of the E4A, for any input feature , the function of the E4A module can be formulated as:
(7) |
where is the output feature of E4A, and denotes the transformation function. Considering residual networks can aid in refinement and network optimization (He et al., 2016), it is reasonable to regard E4A as a refinement transformation for features, which also reduces the learning complexity of earlier modules. Building on this insight, we further introduce a Refiner reusing the E4A architecture to refine the forecasted future states. The Refiner is applied to the forecasted feature from the Forecaster and the feature output from the Observer, producing an enhanced feature for subsequent forecasting of occupancy and flow.
For fair comparisons, the Predictor for occupancy and occupancy flow prediction, and the overall training loss function follow those in Cam4DOcc (Ma et al., 2024a).
4 Experiments
4.1 Experimental Setups
Datasets.
Following Cam4DOcc, 700 out of 850 scenes in nuScenes (Caesar et al., 2020) and nuScenes-Occupancy (Wang et al., 2023a) datasets, and 130 out of 180 scenes in Lyft-Level5 (Houston et al., 2021) with occupancy labels are used for training, while the remaining scenes are used for evaluation. For nuScenes and nuScenes-Occupancy, the numbers of sequences for training and evaluation are 23930 and 5119, respectively. For Lyft-Level5, 15720 and 5880 sequences are used for training and evaluation, respectively. The total length of the sequence is set as 7, including 3 frames as observations (2 past frames and 1 present frame) and 4 future frames for forecasting. The range of the occupancy labels is for and axes, and for axis. The voxel resolution is , and the grid size is for occupancy labels. The forecasting performances are reported with different time intervals due to the different annotated frequencies, i.e., 2 Hz for nuScenes and nuScenes-Occupancy, and 5 Hz for Lyft-Level5.
Evaluation Protocol and Metrics.
To fully evaluate the forecasting performance, we follow Cam4DOcc to adopt the following three-level camera-only occupancy forecasting tasks: (1) Forecasting inflated general movable objects (GMO): the categories for occupancy grids within bounding boxes from nuScenes and Lyft-Level5 datasets are defined as GMO, while the remaining categories are defined as others. (2) Forecasting fine-grained GMO: the same category definition as (1), while the bounding box labels of GMO are replaced with fine-grained voxel-wise annotations from the nuScenes-Occupancy dataset. (3) Forecasting fine-grained GMO and fine-grained general static objects (GSO): the labels of GMO and GSO are from fine-grained annotations. For Lyft-Level5, the evaluation is only conducted on task (1) due to the lack of fine-grained occupancy annotations. For all tasks, intersection over union (IoU (%)) is adopted as the evaluation metric to evaluate occupancy estimation performance for each category of the current frame (), and any of future frames (), and the whole span (). More details can be found in Cam4DOcc.
Implementation Details.
The proposed OccProphet takes 6 images with pixels from different surround views. Following Cam4DOcc, we use ResNet (He et al., 2016) pre-trained on ImageNet (Deng et al., 2009) with FPN (Lin et al., 2017) as the image encoder. For ablation studies, we use ResNet18 for efficiency, while ResNet34 with deformable convolution (Dai et al., 2017) is adopted for main results. All models are trained with a batch size of 4 on 4 RTX 4090 GPUs and tested on a single RTX 4090 with 24G memory. AdamW (Loshchilov & Hutter, 2019) optimizer with an initial learning rate of 3e-4 and a weight decay of 0.01 is adopted to train the models.
4.2 Main Results
Evaluation on forecasting inflated GMO.
The results of forecasting inflated GMO are listed in Table 2, with five comparison methods in total: OpenOccupancy-C, SPC, PowerBEV-3D, BEVDet4D, and OCFNet (Cam4DOcc). OccProphet achieves higher performance than all the compared approaches. Specifically, on the nuScenes dataset, OccProphet surpasses the BEV-based approaches PowerBEV-3D and BEVDet4D by 2.76-11.28%, 2.07-5.69%, 2.28-7.29% in , , respectively. OccProphet also surpasses the voxel-based method Cam4DOcc in all the metrics, especially in by relatively 4.18% and 15.92% on nuScenes and Lyft-Level5 datasets respectively. Qualitative results in Figure 7 demonstrate the superiority of OccProphet. The first group indicates OccProphet’s adaptability in low-light conditions.
nuScenes | Lyft-Level5 | |||||
---|---|---|---|---|---|---|
Method | (2 s) | (0.8 s) | ||||
OpenOccupancy-C (Wang et al., 2023b) | 12.17 | 11.45 | 11.74 | 14.01 | 13.53 | 13.71 |
SPC (Luo et al., 2023; Wei et al., 2023a; Zhu et al., 2021) | 1.27 | - | - | 1.42 | - | - |
PowerBEV-3D (Li et al., 2023a) | 23.08 | 21.25 | 21.86 | 26.19 | 24.47 | 25.06 |
BEVDet4D (Huang & Huang, 2022) | 31.60 | 24.87 | 26.87 | - | - | - |
OCFNet (Cam4DOcc) (Ma et al., 2024a) | 31.30 | 26.82 | 27.98 | 36.41 | 33.56 | 34.60 |
OccProphet (ours) | 34.36 | 26.94 | 29.15 | 43.38 | 37.92 | 40.11 |

Evaluation on forecasting fine-grained GMO.
The second task evaluates the performance of forecasting fine-grained GMO. As shown in Table 3, when changing the GMO labels from inflated format using bounding boxes to fine-grained voxel-based labels, the performances of all approaches decrease significantly except the point cloud prediction method SPC achieving slightly better performance compared to Table 2. As pointed out in Cam4DOcc, the reason is that the fine-grained 3D structures of the moving objects are difficult to estimate using past and current continuous RGB images, while the labels for training SPC are also fine-grained and sparse. However, SPC still performs worse than other counterparts due to the lack of shape consistency. Cam4DOcc keeps the lead over other competitors. Despite the challenging task, OccProphet surprisingly precedes all other approaches by a large margin. Specifically, the performances of OccProphet in , , are relatively 34.3%, 10.43%, 18.61% higher than those of Cam4DOcc. The above results demonstrate the effectiveness of OccProphet in capturing the intricate 3D details of moving objects.
(2 s) | |||||||
---|---|---|---|---|---|---|---|
Method | GMO | GSO | mean | GMO | GSO | mean | GMO |
OpenOccupancy-C (Wang et al., 2023b) | 9.62 | 17.21 | 13.42 | 7.41 | 17.30 | 12.36 | 7.86 |
SPC (Luo et al., 2023; Wei et al., 2023a; Zhu et al., 2021) | 5.85 | 3.29 | 4.57 | 1.08 | 1.40 | 1.24 | 1.12 |
PowerBEV-3D (Li et al., 2023a) | 5.91 | - | - | 5.25 | - | - | 5.49 |
OCFNet (Cam4DOcc) (Ma et al., 2024a) | 11.02 | 17.79 | 14.41 | 9.20 | 17.83 | 13.52 | 9.66 |
OccProphet (ours) | 13.71 | 24.42 | 19.06 | 9.34 | 24.56 | 16.95 | 10.33 |
Evaluation on forecasting fine-grained GMO and fine-grained GSO.
The performances of forecasting fine-grained GMO and fine-grained GSO, are listed in Table 4. OccProphet still surpasses Cam4DOcc in all metrics.
4.3 Ablation Studies
Method | |
---|---|
OccProphet | 28.24 |
w/o Observer | 27.25 |
w/o Forecaster | 27.50 |
w/o Refiner | 27.44 |
w/o All | 26.07 |
Effectiveness of each component.
Table 5 demonstrates the ablation studies on the effectiveness of each component. Removing the Observer (Row 3) leads to around 1% drop in , showing the importance of the Observer extracting spatio-temporal information from observations. If the Forecaster is removed (Row 4), the performance decreases by 0.74%, indicating the advantage of predicting future states adaptively based on the traffic environment compared to direct prediction. Row 5 shows that removing the Refiner brings a 0.8% drop. If all the components are removed (the last row), where only a single linear layer is used to predict future states, the performances are sharply reduced to 26.07%, further revealing the effectiveness of the components. Qualitative results of using the Observer or not are in Figure 8. The integration of Observer promotes spatio-temporal representativeness, enhancing the forecasting consistency with the ground truth. Other qualitative results are shown in the appendix.

Effectiveness of E4A representation.
To further justify the effectiveness of E4A representation, we experiment with different representation styles of the Observer. As shown in Table 1, the UNet like E4A balances computational cost and performance well.
4.4 Comparison of Model Complexity
Method | N (M) | Mem. (G) | FLOPs (G) | FPS | |
---|---|---|---|---|---|
Cam4DOcc | 370 | 57 | 6434 | 1.7 | 27.98 |
OccProphet | 82(78%) | 24(58%) | 1985(69%) | 4.5(165%) | 29.15 (4%) |
In this section, we compare the model complexity between Cam4DOcc and OccProphet. As shown in Table 6, the number of parameters, memory usage, and FLOPs of OccProphet are decreased by 58-78% compared to Cam4DOcc, while OccProphet achieves a 4% relative increase in , and has 2.6 the FPS speed of Cam4DOcc, justifying the efficiency and effectiveness of OccProphet.
5 Conclusion
This paper proposes OccProphet, a novel camera-only framework for occupancy forecasting. The framework employs an Observer-Forecaster-Refiner pipeline, specifically designed for efficient and effective training and inference. Such efficiency and effectiveness are achieved through 4D aggregation and tripling-attention fusion on reduced-resolution features. Experimental results demonstrate OccProphet’s superiority in both forecasting accuracy and efficiency. It outperforms the state-of-the-art Cam4DOcc in occupancy forecasting by relatively 4%18% on three datasets, while operating 2.6× faster and reducing computational costs by 58%-78%. We hope OccProphet can motivate future research in efficient occupancy forecasting and its applications in on-vehicle deployment.
Acknowledgements
The research work was conducted in the JC STEM Lab of Machine Learning and Computer Vision funded by The Hong Kong Jockey Club Charities Trust and was partially supported by the Research Grants Council of the Hong Kong SAR, China (Project No. PolyU 15215824).
References
- Agro et al. (2024) Ben Agro, Quinlan Sykora, Sergio Casas, Thomas Gilles, and Raquel Urtasun. Uno: Unsupervised occupancy fields for perception and forecasting. In CVPR, 2024.
- Caesar et al. (2020) Holger Caesar, Varun Bankiti, Alex H Lang, Sourabh Vora, Venice Erin Liong, Qiang Xu, Anush Krishnan, Yu Pan, Giancarlo Baldan, and Oscar Beijbom. nuscenes: A multimodal dataset for autonomous driving. In CVPR, 2020.
- Cao & De Charette (2022) Anh-Quan Cao and Raoul De Charette. Monoscene: Monocular 3d semantic scene completion. In CVPR, 2022.
- Chen et al. (2016) Xiaozhi Chen, Kaustav Kundu, Ziyu Zhang, Huimin Ma, Sanja Fidler, and Raquel Urtasun. Monocular 3d object detection for autonomous driving. In CVPR, 2016.
- Dai et al. (2017) Jifeng Dai, Haozhi Qi, Yuwen Xiong, Yi Li, Guodong Zhang, Han Hu, and Yichen Wei. Deformable convolutional networks. In Proceedings of the IEEE international conference on computer vision, pp. 764–773, 2017.
- Deng et al. (2009) Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition, pp. 248–255. Ieee, 2009.
- Ding et al. (2019) Wenchao Ding, Lu Zhang, Jing Chen, and Shaojie Shen. Safe trajectory generation for complex urban environments using spatio-temporal semantic corridor. RA-L, 2019.
- Ding et al. (2021) Wenchao Ding, Lu Zhang, Jing Chen, and Shaojie Shen. Epsilon: An efficient planning system for automated vehicles in highly interactive environments. T-RO, 2021.
- Fang et al. (2023) Shaoheng Fang, Zi Wang, Yiqi Zhong, Junhao Ge, and Siheng Chen. Tbp-former: Learning temporal bird’s-eye-view pyramid for joint perception and prediction in vision-centric autonomous driving. In CVPR, 2023.
- Ferenczi et al. (2024) Bryce Ferenczi, Michael Burke, and Tom Drummond. Motionperceiver: Real-time occupancy forecasting for embedded systems. RA-L, 2024.
- He et al. (2016) Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, 2016.
- Houston et al. (2021) John Houston, Guido Zuidhof, Luca Bergamini, Yawei Ye, Long Chen, Ashesh Jain, Sammy Omari, Vladimir Iglovikov, and Peter Ondruska. One thousand and one hours: Self-driving motion prediction dataset. In CoRL, 2021.
- Hu et al. (2021) Anthony Hu, Zak Murez, Nikhil Mohan, Sofía Dudas, Jeffrey Hawke, Vijay Badrinarayanan, Roberto Cipolla, and Alex Kendall. Fiery: Future instance prediction in bird’s-eye view from surround monocular cameras. In ICCV, 2021.
- Huang & Huang (2022) Junjie Huang and Guan Huang. Bevdet4d: Exploit temporal cues in multi-camera 3d object detection. arXiv preprint arXiv:2203.17054, 2022.
- Huang et al. (2021) Junjie Huang, Guan Huang, Zheng Zhu, Yun Ye, and Dalong Du. Bevdet: High-performance multi-camera 3d object detection in bird-eye-view. arXiv preprint arXiv:2112.11790, 2021.
- Huang et al. (2023a) Yuanhui Huang, Wenzhao Zheng, Yunpeng Zhang, Jie Zhou, and Jiwen Lu. Tri-perspective view for vision-based 3d semantic occupancy prediction. In CVPR, 2023a.
- Huang et al. (2023b) Zhiyu Huang, Haochen Liu, and Chen Lv. Gameformer: Game-theoretic modeling and learning of transformer-based interactive prediction and planning for autonomous driving. In ICCV, 2023b.
- Khurana et al. (2023) Tarasha Khurana, Peiyun Hu, David Held, and Deva Ramanan. Point cloud forecasting as a proxy for 4d occupancy forecasting. In CVPR, 2023.
- Li et al. (2023a) Peizheng Li, Shuxiao Ding, Xieyuanli Chen, Niklas Hanselmann, Marius Cordts, and Juergen Gall. Powerbev: a powerful yet lightweight framework for instance prediction in bird’s-eye view. arXiv preprint arXiv:2306.10761, 2023a.
- Li et al. (2023b) Yiming Li, Zhiding Yu, Christopher Choy, Chaowei Xiao, Jose M Alvarez, Sanja Fidler, Chen Feng, and Anima Anandkumar. Voxformer: Sparse voxel transformer for camera-based 3d semantic scene completion. In CVPR, 2023b.
- Lin et al. (2017) Tsung-Yi Lin, Piotr Dollár, Ross Girshick, Kaiming He, Bharath Hariharan, and Serge Belongie. Feature pyramid networks for object detection. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 2117–2125, 2017.
- Liu et al. (2021) Ze Liu, Yutong Lin, Yue Cao, Han Hu, Yixuan Wei, Zheng Zhang, Stephen Lin, and Baining Guo. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF international conference on computer vision, pp. 10012–10022, 2021.
- Loshchilov & Hutter (2019) Ilya Loshchilov and Frank Hutter. Decoupled weight decay regularization. ICLR, 2019.
- Luo et al. (2023) Zhen Luo, Junyi Ma, Zijie Zhou, and Guangming Xiong. Pcpnet: An efficient and semantic-enhanced transformer network for point cloud prediction. IEEE Robotics and Automation Letters, 8(7):4267–4274, 2023.
- Ma et al. (2024a) Junyi Ma, Xieyuanli Chen, Jiawei Huang, Jingyi Xu, Zhen Luo, Jintao Xu, Weihao Gu, Rui Ai, and Hesheng Wang. Cam4docc: Benchmark for camera-only 4d occupancy forecasting in autonomous driving applications. In CVPR, 2024a.
- Ma et al. (2024b) Qihang Ma, Xin Tan, Yanyun Qu, Lizhuang Ma, Zhizhong Zhang, and Yuan Xie. Cotr: Compact occupancy transformer for vision-based 3d occupancy prediction. In CVPR, 2024b.
- Mersch et al. (2022) Benedikt Mersch, Xieyuanli Chen, Jens Behley, and Cyrill Stachniss. Self-supervised point cloud prediction using 3d spatio-temporal convolutional networks. In CoRL, 2022.
- Pan et al. (2024) Chenbin Pan, Burhaneddin Yaman, Tommaso Nesti, Abhirup Mallik, Alessandro G Allievi, Senem Velipasalar, and Liu Ren. Vlp: Vision language planning for autonomous driving. In CVPR, 2024.
- Philion & Fidler (2020) Jonah Philion and Sanja Fidler. Lift, splat, shoot: Encoding images from arbitrary camera rigs by implicitly unprojecting to 3d. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XIV 16, pp. 194–210. Springer, 2020.
- Song et al. (2020) Haoran Song, Wenchao Ding, Yuxuan Chen, Shaojie Shen, Michael Yu Wang, and Qifeng Chen. Pip: Planning-informed trajectory prediction for autonomous driving. In ECCV, 2020.
- Song et al. (2017) Shuran Song, Fisher Yu, Andy Zeng, Angel X Chang, Manolis Savva, and Thomas Funkhouser. Semantic scene completion from a single depth image. In Proceedings of the IEEE conference on computer vision and pattern recognition, 2017.
- Tian et al. (2023) Xiaoyu Tian, Tao Jiang, Longfei Yun, Yucheng Mao, Huitong Yang, Yue Wang, Yilun Wang, and Hang Zhao. Occ3d: A large-scale 3d occupancy prediction benchmark for autonomous driving. NeurIPS, 2023.
- Tong et al. (2023) Wenwen Tong, Chonghao Sima, Tai Wang, Li Chen, Silei Wu, Hanming Deng, Yi Gu, Lewei Lu, Ping Luo, Dahua Lin, et al. Scene as occupancy. In ICCV, 2023.
- Wang et al. (2021) Tai Wang, Xinge Zhu, Jiangmiao Pang, and Dahua Lin. Fcos3d: Fully convolutional one-stage monocular 3d object detection. In ICCV, 2021.
- Wang et al. (2023a) Xiaofeng Wang, Zheng Zhu, Wenbo Xu, Yunpeng Zhang, Yi Wei, Xu Chi, Yun Ye, Dalong Du, Jiwen Lu, and Xingang Wang. Openoccupancy: A large scale benchmark for surrounding semantic occupancy perception. In ICCV, 2023a.
- Wang et al. (2023b) Xiaofeng Wang, Zheng Zhu, Wenbo Xu, Yunpeng Zhang, Yi Wei, Xu Chi, Yun Ye, Dalong Du, Jiwen Lu, and Xingang Wang. Openoccupancy: A large scale benchmark for surrounding semantic occupancy perception. In ICCV, 2023b.
- Wang et al. (2022) Yue Wang, Vitor Campagnolo Guizilini, Tianyuan Zhang, Yilun Wang, Hang Zhao, and Justin Solomon. Detr3d: 3d object detection from multi-view images via 3d-to-2d queries. In CoRL, 2022.
- Wei et al. (2023a) Yi Wei, Linqing Zhao, Wenzhao Zheng, Zheng Zhu, Yongming Rao, Guan Huang, Jiwen Lu, and Jie Zhou. Surrounddepth: Entangling surrounding views for self-supervised multi-camera depth estimation. In Conference on robot learning, pp. 539–549. PMLR, 2023a.
- Wei et al. (2023b) Yi Wei, Linqing Zhao, Wenzhao Zheng, Zheng Zhu, Jie Zhou, and Jiwen Lu. Surroundocc: Multi-camera 3d occupancy prediction for autonomous driving. In ICCV, 2023b.
- Weng et al. (2021) Xinshuo Weng, Jianren Wang, Sergey Levine, Kris Kitani, and Nicholas Rhinehart. Inverting the pose forecasting pipeline with spf2: Sequential pointcloud forecasting for sequential pose forecasting. In CoRL, 2021.
- Weng et al. (2022) Xinshuo Weng, Junyu Nan, Kuan-Hui Lee, Rowan McAllister, Adrien Gaidon, Nicholas Rhinehart, and Kris M Kitani. S2net: Stochastic sequential pointcloud forecasting. In ECCV, 2022.
- Xu et al. (2025) Huaiyuan Xu, Junliang Chen, Shiyu Meng, Yi Wang, and Lap-Pui Chau. A survey on occupancy perception for autonomous driving: The information fusion perspective. Information Fusion, 114:102671, 2025.
- Yang et al. (2024) Yu Yang, Jianbiao Mei, Yukai Ma, Siliang Du, Wenqing Chen, Yijie Qian, Yuxiang Feng, and Yong Liu. Driving in the occupancy world: Vision-centric 4d occupancy forecasting and planning via world models for autonomous driving. arXiv preprint arXiv:2408.14197, 2024.
- Zhang et al. (2022) Yunpeng Zhang, Zheng Zhu, Wenzhao Zheng, Junjie Huang, Guan Huang, Jie Zhou, and Jiwen Lu. Beverse: Unified perception and prediction in birds-eye-view for vision-centric autonomous driving. arXiv preprint arXiv:2205.09743, 2022.
- Zhang et al. (2023) Yunpeng Zhang, Zheng Zhu, and Dalong Du. Occformer: Dual-path transformer for vision-based 3d semantic occupancy prediction. In ICCV, 2023.
- Zhu et al. (2021) Xinge Zhu, Hui Zhou, Tai Wang, Fangzhou Hong, Yuexin Ma, Wei Li, Hongsheng Li, and Dahua Lin. Cylindrical and asymmetrical 3d convolution networks for lidar segmentation. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 9939–9948, 2021.
Appendix A Appendix
A.1 Additional Quantitative Results
Forecasting inflated GMO and fine-grained GSO: the category definitions of GMO and GSO are the same as those in the task of Forecasting fine-grained GMO. The labels of inflated GMO are generated from bounding boxes, while the labels of fine-grained GSO are occupancy annotations. The performances of forecasting inflated GMO and fine-grained GSO are listed in Table 7. SPC is still the worst, where the IoU of inflated GMO remains consistent in Table 2. OccProphet dramatically outperforms other approaches including OpenOccupancy-C and Cam4DOcc by a large margin. Regarding GMO, OccProphet is 3.77% and 0.92% higher than Cam4DOcc in the current and future moments, respectively. For GSO, OccProphet surpasses Cam4DOcc by 6.46% and 6.38% in and , respectively. When taking multiple future frames for evaluation, OccProphet remains the best with an impressive superiority of 2.21% over the second best Cam4DOcc in . These results demonstrate the superiority of our method of extracting spatio-temporal features and predicting future states.
(2 s) | |||||||
---|---|---|---|---|---|---|---|
Method | GMO | GSO | mean | GMO | GSO | mean | GMO |
OpenOccupancy-C (Wang et al., 2023b) | 13.53 | 16.86 | 15.20 | 12.67 | 17.09 | 14.88 | 12.97 |
SPC (Luo et al., 2023; Wei et al., 2023a; Zhu et al., 2021) | 1.27 | 3.29 | 2.28 | - | 1.40 | - | - |
PowerBEV-3D (Li et al., 2023a) | 23.08 | - | - | 21.25 | - | - | 21.86 |
OCFNet (Cam4DOcc) (Ma et al., 2024a) | 29.84 | 17.72 | 23.78 | 25.53 | 17.81 | 21.67 | 26.53 |
OccProphet (ours) | 33.61 | 24.18 | 28.89 | 26.45 | 24.19 | 25.32 | 28.74 |
A.2 Qualitative Results of Ablation Studies
Effectiveness of the Forecaster.
Figure 9 shows the qualitative results with or without using the Forecaster module. The Forecaster is adept at perceiving moving objects, while forecasting with only a single linear layer (w/o Forecaster) tends to miss some moving objects.

Effectiveness of the Refiner.
Figure 10 shows the qualitative results using the Refiner or not. It is evident that using the Refiner yields better forecast results for both moving and static objects, indicating its forecast refinement capability.

A.3 Distinctions from the traditional encoder-decoder architecture
The traditional encoder-decoder architecture comprises an encoder for representation extraction and a decoder for occupancy prediction, as adopted by OccNet (Tong et al., 2023) and Cam4DOcc. However, the traditional architecture either loses 3D geometry details-e.g., OccNet adopts a BEV-based encoder, or introduces high computational cost-e.g., Cam4DOcc utilizes vanilla 3D convolutional networks as the encoder and decoder.
In OccProphet, the Observer works similarly to an encoder, while the combination of Forecaster and Refiner works similarly to a decoder. Unlike the traditional encoder-decoder architecture, OccProphet pushes the efficiency frontier of 4D occupancy forecasting. To achieve this, the Observer and Refiner enable spatio-temporal interaction using the Efficient 4D Aggregation module, and the Forecaster adaptively predicts future states using a lightweight condition mechanism. Overall, our Observer-Forecaster-Refiner framework emphasizes 4D spatio-temporal interaction and conditional forecasting, meanwhile maintaining efficiency.
A.4 Occupancy forecasting over varying time horizons
The performance of occupancy forecasting over varying time horizons is critical in autonomous driving scenarios. We compare OccProphet with OpenOccupancy-C (Wang et al., 2023b), PowerBEV-3D (Li et al., 2023a), and Cam4DOcc (Ma et al., 2024a) in terms of occupancy forecasting accuracy over varying time horizons, as shown in Table 8. We can see that (1) OccProphet consistently outperforms other approaches across all time horizons on three datasets. (2) The longer the forecasting period, the lower the accuracy of all methods, indicating that forecasting becomes increasingly difficult.
nuScenes | Lyft-Level5 | nuScenes-Occupancy | ||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|
Method | 0.5s | 1.0s | 1.5s | 2.0s | 0.2s | 0.4s | 0.6s | 0.8s | 0.5s | 1.0s | 1.5s | 2.0s |
OpenOccupancy-C | 12.07 | 11.80 | 11.63 | 11.45 | 13.87 | 13.77 | 13.65 | 13.53 | 9.17 | 8.64 | 8.29 | 8.02 |
PowerBEV-3D | 22.48 | 22.07 | 21.65 | 21.25 | 25.70 | 25.25 | 24.82 | 24.47 | 5.74 | 5.56 | 5.41 | 5.25 |
OCFNet (Cam4DOcc) | 29.36 | 28.30 | 27.44 | 26.82 | 35.58 | 34.96 | 34.28 | 33.56 | 10.64 | 10.20 | 9.89 | 9.68 |
OccProphet (ours) | 32.17 | 29.60 | 27.95 | 26.94 | 42.34 | 40.87 | 39.38 | 37.92 | 13.64 | 12.10 | 11.27 | 10.69 |
A.5 Failure Cases
A.5.1 Unseen Scenarios
To test forecasting performance in unseen scenarios, we conduct a cross-domain experiment. Specifically, we train the model on the Lyft-Level5 dataset and test it on the nuScenes dataset. The visualization results are shown in Figure 11. We can observe that cross-domain occupancy forecasting performs worse than intra-domain forecasting within a single dataset. We consider that this performance gap may be due to the difference in sensor settings between the two datasets. In the future, research on generalized occupancy forecasting will be a valuable direction worth exploring.

A.5.2 Occluded Scenarios
We investigate the forecasting performance in occluded scenarios, which frequently occur in traffic scenarios. Ground truth occupancy and forecasted future occupancy by OccProphet, over varying time horizons, are qualitatively visualized in Figure 12. It can be observed that the occupancy labels of the occluded object over different time horizons are complete, ensuring the training reliability. The forecasted results of OccProphet demonstrate its capability of handling occluded scenarios to a certain degree, which can still be further improved in future research.


A.5.3 Dense Scenarios
To evaluate model precision at fine granularity, we qualitatively visualize the occupancy forecasting of Cam4DOcc and OccProphet in dense scenarios, as shown in Figure 13. The visualization reveals that both Cam4DOcc and OccProphet encounter challenges in fine-grained forecasting, indicating a huge difficulty. In comparison, OccProphet’s results are closer to the ground truth. We believe that this is attributed to our proposed Observer-Forecaster-Refiner framework. We regard fine-grained occupancy forecasting as a valuable direction for future research.