Convolutional Neural Networks (CNNs) are widely used for object detection tasks, whereas recent studies have shown that they rely more on texture rather than shape for object recognition, a phenomenon known as texture bias. This bias makes them vulnerable to image corruptions, domain shifts, and adversarial perturbations, posing significant challenges for real-world deployment, especially in safety-critical and industrial applications. Despite its significance, texture bias in object detection remains largely underexplored. To address this gap, we first conduct a comprehensive analysis of texture bias across multiple widely-used CNN-based detection architectures, demonstrating the widespread presence and detrimental impact of this issue. Motivated by these findings, we propose a simple yet effective method, TexDrop, to increase shape bias in CNNs and therefore improve their accuracy and robustness. Specifically, TexDrop randomly drops out the texture and color of the training images through straightforward edge detection, forcing models to learn to detect objects based on their shape, thus increasing shape bias. Unlike prior approaches that require architectural modifications, extensive additional training data or complex regularization schemes, TexDrop is model-agnostic, easy to integrate into existing training pipelines, and incurs negligible computational overhead. Intensive experiments on Pascal VOC, COCO, and various corrupted COCO datasets demonstrate that TexDrop not only improves detection performance across multiple architectures but also consistently enhances robustness against various image corruptions and texture variations. Our study provides empirical insights into texture dependence in object detectors and contributes a practical solution for developing more robust and reliable object detection systems in real-world applications.
扫码关注我们
求助内容:
应助结果提醒方式:
