{ "id": "2405.14480", "version": "v1", "published": "2024-05-23T12:12:11.000Z", "updated": "2024-05-23T12:12:11.000Z", "title": "Scalable Visual State Space Model with Fractal Scanning", "authors": [ "Lv Tang", "HaoKe Xiao", "Peng-Tao Jiang", "Hao Zhang", "Jinwei Chen", "Bo Li" ], "categories": [ "cs.CV" ], "abstract": "Foundational models have significantly advanced in natural language processing (NLP) and computer vision (CV), with the Transformer architecture becoming a standard backbone. However, the Transformer's quadratic complexity poses challenges for handling longer sequences and higher resolution images. To address this challenge, State Space Models (SSMs) like Mamba have emerged as efficient alternatives, initially matching Transformer performance in NLP tasks and later surpassing Vision Transformers (ViTs) in various CV tasks. To improve the performance of SSMs, one crucial aspect is effective serialization of image patches. Existing methods, relying on linear scanning curves, often fail to capture complex spatial relationships and produce repetitive patterns, leading to biases. To address these limitations, we propose using fractal scanning curves for patch serialization. Fractal curves maintain high spatial proximity and adapt to different image resolutions, avoiding redundancy and enhancing SSMs' ability to model complex patterns accurately. We validate our method in image classification, detection, and segmentation tasks, and the superior performance validates its effectiveness.", "revisions": [ { "version": "v1", "updated": "2024-05-23T12:12:11.000Z" } ], "analyses": { "keywords": [ "scalable visual state space model", "quadratic complexity poses challenges", "curves maintain high spatial", "fractal scanning", "maintain high spatial proximity" ], "note": { "typesetting": "TeX", "pages": 0, "language": "en", "license": "arXiv", "status": "editable" } } }