Electroencephalogram (EEG) signal classification is of utmost importance in brain-computer interface (BCI) systems. However, the inherent complex properties of EEG signals pose a challenge in their analysis and modeling. This letter proposes a novel approach of integrating wavelet scattering transform (WST) with convolutional neural network (CNN) for classifying motor imagery (MI) via EEG signals (referred as WST-CNN), capable of extracting distinctive characteristics in signals even when the data is limited. In this architecture, the first layer is nontrainable WST features with fixed initializations in WST-CNN. Furthermore, WSTs are robust to local perturbations in data, especially in the form of translation invariance, and resilient to deformations, thereby enhancing the network's reliability. The performance of the proposed idea is evaluated on the DBCIE dataset for three different scenarios: left-arm (LA) movement, right-arm (RA) movement, and simultaneous movement of both arms (BA). The BCI Competition IV-2a dataset was also employed to validate the proposed concept across four distinct MI tasks, such as movements in: left-hand (LH), right-hand (RH), feet (FT), and tongue (T). The classifications' performance was evaluated in terms of accuracy ( $eta$