Poster
Efficient LiDAR Reflectance Compression via Scanning Serialization
Jiahao Zhu · Kang You · Dandan Ding · Zhan Ma
East Exhibition Hall A-B #E-1600
[
Abstract
]
[
Lay Summary
]
[ Project Page ]
[
Poster]
[
OpenReview]
Wed 16 Jul 4:30 p.m. PDT
— 7 p.m. PDT
Abstract:
Reflectance attributes in LiDAR point clouds provide essential information for downstream tasks but remain underexplored in neural compression methods. To address this, we introduce SerLiC, a serialization-based neural compression framework to fully exploit the intrinsic characteristics of LiDAR reflectance. SerLiC first transforms 3D LiDAR point clouds into 1D sequences via scan-order serialization, offering a device-centric perspective for reflectance analysis. Each point is then tokenized into a contextual representation comprising its sensor scanning index, radial distance, and prior reflectance, for effective dependencies exploration. For efficient sequential modeling, Mamba is incorporated with a dual parallelization scheme, enabling simultaneous autoregressive dependency capture and fast processing. Extensive experiments demonstrate that SerLiC attains over 2$\times$ volume reduction against the original reflectance data, outperforming the state-of-the-art method by up to 22\% reduction of compressed bits while using only 2\% of its parameters. Moreover, a lightweight version of SerLiC achieves $\geq 10$ fps (frames per second) with just 111K parameters, which is attractive for real applications.
Lay Summary:
Reflectance attributes in LiDAR point clouds provide essential information for downstream tasks but remain underexplored in neural compression methods. To address this, we introduce SerLiC, a serialization-based neural compression framework to efficiently process reflectance data. Extensive experiments demonstrate that SerLiC attains over 2$\times$ volume reduction against the original reflectance data, which is attractive for real-world applications.
Chat is not available.