首页 > 最新文献

Optical Memory and Neural Networks最新文献

英文 中文
Complex Spatial Structures in the Optically Driven VCSELs 光驱动vcsel中的复杂空间结构
IF 0.8 Q4 OPTICS Pub Date : 2026-01-26 DOI: 10.3103/S1060992X25601927
E. A. Yarunova, D. S. Riashchikov, A. A. Krents, N. E. Molevich

This study investigates the spatiotemporal dynamics of broad-area vertical-cavity surface-emitting lasers (VCSELs) under external optical injection. Using a semiconductor-adapted Maxwell-Bloch model, we demonstrate that weak optical injection effectively suppresses modulation instability—a major constraint to achieving coherent VCSEL emission. Through linear stability analysis and numerical simulations, we found and showed the dependence of spatial patterns on pump current and injection amplitude. Our results reveal that controlled optical injection transforms chaotic emission into ordered structures, including stripes, hexagons, labyrinths, and their hybrid forms, with the pattern scale governed by the wavenumber of maximum growth increment. Notably, reducing the laser aperture size promotes the formation of defect-free patterns. These findings offer key insights for stabilizing VCSEL emission and leveraging self-organized patterns for advanced photonic applications, such as optical computing and on-chip communication systems.

本文研究了外光注入下广域垂直腔面发射激光器(VCSELs)的时空动力学。利用半导体适应的麦克斯韦-布洛赫模型,我们证明了弱光注入有效地抑制了调制不稳定性-这是实现相干VCSEL发射的主要限制。通过线性稳定性分析和数值模拟,我们发现并展示了泵电流和注入幅度对空间模式的依赖性。结果表明,可控光注入将混沌发射转化为有序结构,包括条纹、六边形、迷宫及其混合形式,其模式尺度由最大生长增量的波数决定。值得注意的是,减小激光孔径可以促进无缺陷图案的形成。这些发现为稳定VCSEL发射和利用自组织模式用于先进光子应用(如光计算和片上通信系统)提供了关键见解。
{"title":"Complex Spatial Structures in the Optically Driven VCSELs","authors":"E. A. Yarunova,&nbsp;D. S. Riashchikov,&nbsp;A. A. Krents,&nbsp;N. E. Molevich","doi":"10.3103/S1060992X25601927","DOIUrl":"10.3103/S1060992X25601927","url":null,"abstract":"<p>This study investigates the spatiotemporal dynamics of broad-area vertical-cavity surface-emitting lasers (VCSELs) under external optical injection. Using a semiconductor-adapted Maxwell-Bloch model, we demonstrate that weak optical injection effectively suppresses modulation instability—a major constraint to achieving coherent VCSEL emission. Through linear stability analysis and numerical simulations, we found and showed the dependence of spatial patterns on pump current and injection amplitude. Our results reveal that controlled optical injection transforms chaotic emission into ordered structures, including stripes, hexagons, labyrinths, and their hybrid forms, with the pattern scale governed by the wavenumber of maximum growth increment. Notably, reducing the laser aperture size promotes the formation of defect-free patterns. These findings offer key insights for stabilizing VCSEL emission and leveraging self-organized patterns for advanced photonic applications, such as optical computing and on-chip communication systems.</p>","PeriodicalId":721,"journal":{"name":"Optical Memory and Neural Networks","volume":"34 2","pages":"S346 - S356"},"PeriodicalIF":0.8,"publicationDate":"2026-01-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146043398","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Axial and Off-Axis Focal Diffraction Orders Formation Using of Phase Quantized Non-Paraxial Optical Elements 利用相位量化的非近轴光学元件形成轴向和离轴焦衍射阶数
IF 0.8 Q4 OPTICS Pub Date : 2026-01-26 DOI: 10.3103/S1060992X2560199X
O. A. Dyukareva

In this paper, we consider the beam diffraction on quantized non-paraxial optical elements. We show that the formation of diffraction orders depends on both the numerical aperture and the focal length. In non-paraxial propagation, the main energy falls on the first diffraction orders, and it is also possible to eliminate a larger number of orders compared to the paraxial case. To simultaneously increase the number of longitudinal and transverse orders in the non-paraxial region, we propose using multifocal lenses as focusing elements. We expect that the results obtained can be used in (de-)multiplexing and signal detection problems.

本文研究了量子化非傍轴光学元件上的光束衍射问题。我们证明了衍射阶的形成取决于数值孔径和焦距。在非近轴传播中,主要能量落在第一衍射阶上,与近轴情况相比,也有可能消除更多的衍射阶。为了同时增加非近轴区域的纵向和横向阶数,我们建议使用多焦透镜作为聚焦元件。我们期望得到的结果可以用于(解)复用和信号检测问题。
{"title":"Axial and Off-Axis Focal Diffraction Orders Formation Using of Phase Quantized Non-Paraxial Optical Elements","authors":"O. A. Dyukareva","doi":"10.3103/S1060992X2560199X","DOIUrl":"10.3103/S1060992X2560199X","url":null,"abstract":"<p>In this paper, we consider the beam diffraction on quantized non-paraxial optical elements. We show that the formation of diffraction orders depends on both the numerical aperture and the focal length. In non-paraxial propagation, the main energy falls on the first diffraction orders, and it is also possible to eliminate a larger number of orders compared to the paraxial case. To simultaneously increase the number of longitudinal and transverse orders in the non-paraxial region, we propose using multifocal lenses as focusing elements. We expect that the results obtained can be used in (de-)multiplexing and signal detection problems.</p>","PeriodicalId":721,"journal":{"name":"Optical Memory and Neural Networks","volume":"34 2","pages":"S243 - S251"},"PeriodicalIF":0.8,"publicationDate":"2026-01-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146043479","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Feature Selection for Thick Cloud Classification 厚云分类的特征选择
IF 0.8 Q4 OPTICS Pub Date : 2026-01-26 DOI: 10.3103/S1060992X25602520
A. S. Minkin

A method of feature selection based on hyperspectral data classification is proposed. The features are selected using iterative training of Decision Tree classifiers for further construction of a thick cloud classifier based on spectral features. Classifiers are trained with different hyperparameters for different set of features by recursive elimination. Feature selection is determined by analyzing the correlation between the decrease in Gini impurity and classification accuracy, combined with mean feature importance. Classification model training is performed for three types of surfaces: ocean, vegetation, and urbanized areas. Feature selection improves the accuracy of Random Forest classifier by choosing a limited set of features from the NIR and the lower part of the SWIR spectrum ranges according to their importance.

提出了一种基于高光谱数据分类的特征选择方法。利用决策树分类器的迭代训练选择特征,进一步构建基于谱特征的厚云分类器。通过递归消去,对不同的特征集使用不同的超参数来训练分类器。特征选择是通过分析基尼杂质的减少与分类精度之间的相关性,并结合平均特征重要度来确定的。对海洋、植被和城市化区域三种类型的表面进行分类模型训练。特征选择是根据特征的重要性从近红外光谱和SWIR光谱范围的下半部分中选择一组有限的特征,从而提高随机森林分类器的准确率。
{"title":"Feature Selection for Thick Cloud Classification","authors":"A. S. Minkin","doi":"10.3103/S1060992X25602520","DOIUrl":"10.3103/S1060992X25602520","url":null,"abstract":"<p>A method of feature selection based on hyperspectral data classification is proposed. The features are selected using iterative training of Decision Tree classifiers for further construction of a thick cloud classifier based on spectral features. Classifiers are trained with different hyperparameters for different set of features by recursive elimination. Feature selection is determined by analyzing the correlation between the decrease in Gini impurity and classification accuracy, combined with mean feature importance. Classification model training is performed for three types of surfaces: ocean, vegetation, and urbanized areas. Feature selection improves the accuracy of Random Forest classifier by choosing a limited set of features from the NIR and the lower part of the SWIR spectrum ranges according to their importance.</p>","PeriodicalId":721,"journal":{"name":"Optical Memory and Neural Networks","volume":"34 2","pages":"S327 - S332"},"PeriodicalIF":0.8,"publicationDate":"2026-01-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146043480","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Diffractive Optics in Laser Processing: Digital Approaches to Design and Application 激光加工中的衍射光学:数字化设计与应用
IF 0.8 Q4 OPTICS Pub Date : 2026-01-26 DOI: 10.3103/S1060992X25602647
S. P. Murzin

Effective control of the spatial distribution of laser beam energy plays a key role in material processing technologies. The integration of diffractive optical elements (DOE) with digital tools opens new prospects for the control of laser processes, enabling precise energy distribution, adaptive adjustment of processing parameters, and improved predictability of outcomes. This paper explores digital approaches to the design and application of diffractive optics in laser processing. It describes methods for tuning laser beam parameters, the use of digital twins, and machine learning algorithms to enhance processing accuracy. The paper also thoroughly examines the potential applications of digital technologies for optimizing micro- and nano-processing, laser welding, and improving the quality and stability of materials during laser modification. The future development of these technologies, including integration with adaptive systems and optimization algorithms, is presented, opening new horizons for precision manufacturing processes.

有效控制激光束能量的空间分布在材料加工技术中起着关键作用。衍射光学元件(DOE)与数字工具的集成为激光过程的控制开辟了新的前景,实现了精确的能量分布,加工参数的自适应调整,并提高了结果的可预测性。本文探讨了衍射光学器件在激光加工中的设计和应用的数字化方法。它描述了调整激光束参数的方法,数字双胞胎的使用,以及提高加工精度的机器学习算法。本文还深入探讨了数字技术在优化微纳米加工、激光焊接以及提高激光改性过程中材料质量和稳定性方面的潜在应用。提出了这些技术的未来发展,包括与自适应系统和优化算法的集成,为精密制造工艺开辟了新的视野。
{"title":"Diffractive Optics in Laser Processing: Digital Approaches to Design and Application","authors":"S. P. Murzin","doi":"10.3103/S1060992X25602647","DOIUrl":"10.3103/S1060992X25602647","url":null,"abstract":"<p>Effective control of the spatial distribution of laser beam energy plays a key role in material processing technologies. The integration of diffractive optical elements (DOE) with digital tools opens new prospects for the control of laser processes, enabling precise energy distribution, adaptive adjustment of processing parameters, and improved predictability of outcomes. This paper explores digital approaches to the design and application of diffractive optics in laser processing. It describes methods for tuning laser beam parameters, the use of digital twins, and machine learning algorithms to enhance processing accuracy. The paper also thoroughly examines the potential applications of digital technologies for optimizing micro- and nano-processing, laser welding, and improving the quality and stability of materials during laser modification. The future development of these technologies, including integration with adaptive systems and optimization algorithms, is presented, opening new horizons for precision manufacturing processes.</p>","PeriodicalId":721,"journal":{"name":"Optical Memory and Neural Networks","volume":"34 2","pages":"S357 - S368"},"PeriodicalIF":0.8,"publicationDate":"2026-01-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146043411","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Research of the Influence of Wave Aberrations by Distortioning the Formation of Light Curves 畸变光曲线形成对波像差影响的研究
IF 0.8 Q4 OPTICS Pub Date : 2026-01-26 DOI: 10.3103/S1060992X25602246
L. B. Dubman, P. A. Khorin

The influence of vortex phase singularity and wave aberrations on the distortion of the pattern of formation of light curves is investigated. The Gillis transform was used to form the parametric type of the curve, and the diffraction optical elements were calculated using the Whittaker integral. It was found that the influence of vortex phase singularity mainly affects the change in the peripheral part of the amplitude of the formed curve and leading to the formation of zones with zero intensity in the central part. It is also shown that aberrations in optical systems distort the wavefront and degrading the image quality of formed light curves. Different types of aberrations introduce characteristic changes in the intensity distributions of light curves, which can be further used for detection and recognition of aberrations.

研究了涡旋相位奇点和波像差对光曲线形成模式畸变的影响。利用Gillis变换形成曲线的参数型,利用Whittaker积分计算衍射光学元件。研究发现,涡旋相位奇点的影响主要影响形成曲线外围部分振幅的变化,导致中心部分形成零强度区。光学系统中的像差会引起波前畸变,从而降低形成的光曲线的成像质量。不同类型的像差会引起光曲线强度分布的特征变化,可以进一步用于像差的检测和识别。
{"title":"Research of the Influence of Wave Aberrations by Distortioning the Formation of Light Curves","authors":"L. B. Dubman,&nbsp;P. A. Khorin","doi":"10.3103/S1060992X25602246","DOIUrl":"10.3103/S1060992X25602246","url":null,"abstract":"<p>The influence of vortex phase singularity and wave aberrations on the distortion of the pattern of formation of light curves is investigated. The Gillis transform was used to form the parametric type of the curve, and the diffraction optical elements were calculated using the Whittaker integral. It was found that the influence of vortex phase singularity mainly affects the change in the peripheral part of the amplitude of the formed curve and leading to the formation of zones with zero intensity in the central part. It is also shown that aberrations in optical systems distort the wavefront and degrading the image quality of formed light curves. Different types of aberrations introduce characteristic changes in the intensity distributions of light curves, which can be further used for detection and recognition of aberrations.</p>","PeriodicalId":721,"journal":{"name":"Optical Memory and Neural Networks","volume":"34 2","pages":"S278 - S291"},"PeriodicalIF":0.8,"publicationDate":"2026-01-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146043416","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Hyperspectral Imaging for Soil Type Classification 土壤类型分类的高光谱成像
IF 0.8 Q4 OPTICS Pub Date : 2026-01-26 DOI: 10.3103/S1060992X25700262
A. R. Makarov, A. A. Muzyka, K. E. Savelev, A. A. Rastorguev, V. V. Podlipnov

This paper explores the application of soil type classification algorithms (red clay, loamy soil, chernozem) in agricultural fields based on hyperspectral imaging data (400–1000 nm) acquired from an unmanned aerial vehicle (UAV). As part of the dataset preparation, the data were processed using a set of algorithms, from classical approaches to deep models. The ensemble NM3D-CNN achieved the best performance—0.874 weighted F1. Results of other approaches—PCA segmentation: 0.864, 1D-CNN: 0.842, SSFTT: 0.829, SVM: 0.764. These results indicate that Vis-NIR UAV-borne HIS enables reliable soil mapping, while spectral-spatial deep ensembles provide the strongest gains.

基于无人机(UAV) 400 ~ 1000 nm高光谱成像数据,探讨了红壤、壤土、黑钙土土壤类型分类算法在农田中的应用。作为数据集准备的一部分,数据使用一组算法进行处理,从经典方法到深度模型。集合NM3D-CNN的性能最好,加权F1为0.874。其他方法的结果- pca分割:0.864,1D-CNN: 0.842, SSFTT: 0.829, SVM: 0.764。这些结果表明,Vis-NIR无人机携带的HIS能够实现可靠的土壤制图,而光谱空间深度集成提供了最强的收益。
{"title":"Hyperspectral Imaging for Soil Type Classification","authors":"A. R. Makarov,&nbsp;A. A. Muzyka,&nbsp;K. E. Savelev,&nbsp;A. A. Rastorguev,&nbsp;V. V. Podlipnov","doi":"10.3103/S1060992X25700262","DOIUrl":"10.3103/S1060992X25700262","url":null,"abstract":"<p>This paper explores the application of soil type classification algorithms (red clay, loamy soil, chernozem) in agricultural fields based on hyperspectral imaging data (400–1000 nm) acquired from an unmanned aerial vehicle (UAV). As part of the dataset preparation, the data were processed using a set of algorithms, from classical approaches to deep models. The ensemble NM3D-CNN achieved the best performance—0.874 weighted F1. Results of other approaches—PCA segmentation: 0.864, 1D-CNN: 0.842, SSFTT: 0.829, SVM: 0.764. These results indicate that Vis-NIR UAV-borne HIS enables reliable soil mapping, while spectral-spatial deep ensembles provide the strongest gains.</p>","PeriodicalId":721,"journal":{"name":"Optical Memory and Neural Networks","volume":"34 2","pages":"S312 - S318"},"PeriodicalIF":0.8,"publicationDate":"2026-01-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146043417","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Synthetic Data Generation for Tasks of Recognizing Actions and Interaction Objects of an Agricultural Drone Operator 农业无人机操作人员动作与交互对象识别任务的综合数据生成
IF 0.8 Q4 OPTICS Pub Date : 2026-01-26 DOI: 10.3103/S1060992X25602489
T. D. Kazarkin, L. A. Abakumov, K. S. Gerasimova, R. M. Khabibullin, L. A. Taskina

This paper presents a prototype application for generating synthetic data used to train neural network models for recognizing actions and interaction objects of an agricultural drone operator. This application has been implemented with the ability to customize various generation parameters. The obtained generation results are saved as images and text files in YOLO format.

本文提出了一个原型应用程序,用于生成用于训练神经网络模型的合成数据,以识别农业无人机操作员的动作和交互对象。这个应用程序已经实现了自定义各种生成参数的能力。得到的生成结果保存为YOLO格式的图像和文本文件。
{"title":"Synthetic Data Generation for Tasks of Recognizing Actions and Interaction Objects of an Agricultural Drone Operator","authors":"T. D. Kazarkin,&nbsp;L. A. Abakumov,&nbsp;K. S. Gerasimova,&nbsp;R. M. Khabibullin,&nbsp;L. A. Taskina","doi":"10.3103/S1060992X25602489","DOIUrl":"10.3103/S1060992X25602489","url":null,"abstract":"<p>This paper presents a prototype application for generating synthetic data used to train neural network models for recognizing actions and interaction objects of an agricultural drone operator. This application has been implemented with the ability to customize various generation parameters. The obtained generation results are saved as images and text files in YOLO format.</p>","PeriodicalId":721,"journal":{"name":"Optical Memory and Neural Networks","volume":"34 2","pages":"S319 - S326"},"PeriodicalIF":0.8,"publicationDate":"2026-01-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146043477","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Reverse Flow during Propagation of Half a Plane Wave 半平面波传播过程中的反向流动
IF 0.8 Q4 OPTICS Pub Date : 2025-12-24 DOI: 10.3103/S1060992X25600909
V. V. Kotlyar, A. A. Kovalev, A. G. Nalimov, A. M. Telegin

We explore the diffraction of a plane wave by an opaque rectangular screen and show that at any on-axis distance from the obstacle plane, a multitude of diffraction fringes occur in a screen-parallel plane, characterized by a (negative) canonical backflow. Remarkably, the localized fringe recurrence interval is found to decrease with larger distance from the screen edge, with the first fringe of the canonical energy backflow departing from the edge with increasing distance from the optical axis. The energy backflow is shown to occur in the diffraction pattern areas characterized by subwavelength values of phase and amplitude modulation, i.e. where the local wave-vector expressed through the phase gradient is larger than the incident wave wave-vector.

我们探索了平面波在不透明矩形屏幕上的衍射,并表明在距离障碍物平面的任何轴上距离处,在屏幕平行的平面上出现了大量的衍射条纹,其特征是(负)正则回流。值得注意的是,局部条纹的重现间隔随着距离屏幕边缘的增大而减小,正则能回流的第一条带随着距离光轴的增加而离开屏幕边缘。能量回流发生在以相位和幅度调制的亚波长值为特征的衍射图案区域,即通过相位梯度表示的局部波矢量大于入射波波矢量的地方。
{"title":"Reverse Flow during Propagation of Half a Plane Wave","authors":"V. V. Kotlyar,&nbsp;A. A. Kovalev,&nbsp;A. G. Nalimov,&nbsp;A. M. Telegin","doi":"10.3103/S1060992X25600909","DOIUrl":"10.3103/S1060992X25600909","url":null,"abstract":"<p>We explore the diffraction of a plane wave by an opaque rectangular screen and show that at any on-axis distance from the obstacle plane, a multitude of diffraction fringes occur in a screen-parallel plane, characterized by a (negative) canonical backflow. Remarkably, the localized fringe recurrence interval is found to decrease with larger distance from the screen edge, with the first fringe of the canonical energy backflow departing from the edge with increasing distance from the optical axis. The energy backflow is shown to occur in the diffraction pattern areas characterized by subwavelength values of phase and amplitude modulation, i.e. where the local wave-vector expressed through the phase gradient is larger than the incident wave wave-vector.</p>","PeriodicalId":721,"journal":{"name":"Optical Memory and Neural Networks","volume":"34 4","pages":"476 - 484"},"PeriodicalIF":0.8,"publicationDate":"2025-12-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145808676","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep Attention-Multistage GAN with Sparse Dense Fusion R-CNN for High-Resolution and Object Detection in Surveillance System 深度注意-多阶段GAN与稀疏密集融合R-CNN用于监控系统的高分辨率和目标检测
IF 0.8 Q4 OPTICS Pub Date : 2025-12-24 DOI: 10.3103/S1060992X2560003X
Anu Yadav,  Ela Kumar

Video surveillance utilization has developed significantly in sectors such as traffic monitoring, private institution protection, and landmark protection. Identifying an object in a captured surveillance image is difficult because of the poor excellence of the images. The quality of Low-Resolution (LR) images can be enhanced using an image Super-Resolution (SR) reconstruction method. Sophisticated Deep learning methods have been utilized in order to attain state-of-the-art performance in SR. Nevertheless, these techniques are typically prone to losing essential information and perform poorly on complex computations. To overcome these challenges, this research develop a deep Attention-Multistage Generative Adversarial Network (DA-MGAN) for image super-resolution and integrate SDR (Sparse Dense fusion R-CNN) with Non-Maximum Suppression (NMS) to enhance object detection accuracy in surveillance images. DA-MGAN is used to generate high-resolution surveillance images by utilizing attention mechanisms for improved feature extraction and integrating a multistage GAN that progressively enhances the image quality at each stage. After image resolution, Sparse Dense Fusion R-CNN (SDR) is used for object detection in super-resolved images to improve feature extraction through Sparse Dense Fusion. The R-CNN leverages these enhanced features to accurately detect and segment objects at the pixel level. Subsequently, Non-Maximum Suppression (NMS) was applied to improve localization by eliminating overlapping bounding boxes and minimizing false positives. This integrated method boosts overall detection precision and reliability in real time surveillance scenarios. The proposed model achieves a Super Resolution Error Rate (SRER) of 0.19%, a Bit Error Rate (BER) of 0.125%, a Packet Error Rate (PER) of 0.0990%, and a Deep End-to-End Image Metric (DEEIM) of 0.04963%, showcasing its superior performance. In contrast with existing methodologies, these results highlight the effectiveness of the suggested approach in reducing error rates and enhancing image quality metrics. As a result, these methods are ideally suited for real-time applications, particularly in high-resolution scenarios and object detection within surveillance systems.

视频监控在交通监控、私人机构保护和地标保护等领域的应用得到了显著发展。由于捕获的监控图像质量不佳,很难识别目标。采用图像超分辨率(SR)重建方法可以提高低分辨率(LR)图像的质量。为了在sr中获得最先进的性能,已经使用了复杂的深度学习方法。然而,这些技术通常容易丢失基本信息,并且在复杂的计算中表现不佳。为了克服这些挑战,本研究开发了一种用于图像超分辨率的深度注意-多阶段生成对抗网络(DA-MGAN),并将SDR(稀疏密集融合R-CNN)与非最大抑制(NMS)相结合,以提高监视图像中的目标检测精度。DA-MGAN通过利用注意力机制来改进特征提取,并集成多阶段GAN来逐步提高每阶段的图像质量,从而生成高分辨率的监控图像。图像分辨率完成后,利用SDR (Sparse Dense Fusion R-CNN)对超分辨率图像进行目标检测,通过Sparse Dense Fusion改进特征提取。R-CNN利用这些增强的功能来准确地检测和分割像素级的对象。随后,采用非最大抑制(NMS)方法,通过消除重叠的边界框和减少误报来改进定位。这种集成方法在实时监控场景中提高了整体检测精度和可靠性。该模型的超分辨错误率(SRER)为0.19%,误码率(BER)为0.125%,包错误率(PER)为0.0990%,深度端到端图像度量(DEEIM)为0.04963%,显示了其优越的性能。与现有的方法相比,这些结果突出了所建议的方法在降低错误率和提高图像质量指标方面的有效性。因此,这些方法非常适合实时应用,特别是在监控系统中的高分辨率场景和目标检测中。
{"title":"Deep Attention-Multistage GAN with Sparse Dense Fusion R-CNN for High-Resolution and Object Detection in Surveillance System","authors":"Anu Yadav,&nbsp; Ela Kumar","doi":"10.3103/S1060992X2560003X","DOIUrl":"10.3103/S1060992X2560003X","url":null,"abstract":"<p>Video surveillance utilization has developed significantly in sectors such as traffic monitoring, private institution protection, and landmark protection. Identifying an object in a captured surveillance image is difficult because of the poor excellence of the images. The quality of Low-Resolution (LR) images can be enhanced using an image Super-Resolution (SR) reconstruction method. Sophisticated Deep learning methods have been utilized in order to attain state-of-the-art performance in SR. Nevertheless, these techniques are typically prone to losing essential information and perform poorly on complex computations. To overcome these challenges, this research develop a deep Attention-Multistage Generative Adversarial Network (DA-MGAN) for image super-resolution and integrate SDR (Sparse Dense fusion R-CNN) with Non-Maximum Suppression (NMS) to enhance object detection accuracy in surveillance images. DA-MGAN is used to generate high-resolution surveillance images by utilizing attention mechanisms for improved feature extraction and integrating a multistage GAN that progressively enhances the image quality at each stage. After image resolution, Sparse Dense Fusion R-CNN (SDR) is used for object detection in super-resolved images to improve feature extraction through Sparse Dense Fusion. The R-CNN leverages these enhanced features to accurately detect and segment objects at the pixel level. Subsequently, Non-Maximum Suppression (NMS) was applied to improve localization by eliminating overlapping bounding boxes and minimizing false positives. This integrated method boosts overall detection precision and reliability in real time surveillance scenarios. The proposed model achieves a Super Resolution Error Rate (SRER) of 0.19%, a Bit Error Rate (BER) of 0.125%, a Packet Error Rate (PER) of 0.0990%, and a Deep End-to-End Image Metric (DEEIM) of 0.04963%, showcasing its superior performance. In contrast with existing methodologies, these results highlight the effectiveness of the suggested approach in reducing error rates and enhancing image quality metrics. As a result, these methods are ideally suited for real-time applications, particularly in high-resolution scenarios and object detection within surveillance systems.</p>","PeriodicalId":721,"journal":{"name":"Optical Memory and Neural Networks","volume":"34 4","pages":"566 - 580"},"PeriodicalIF":0.8,"publicationDate":"2025-12-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145808680","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Unsupervised Background Estimation Using a Neural Integrator 基于神经积分器的无监督背景估计
IF 0.8 Q4 OPTICS Pub Date : 2025-12-24 DOI: 10.3103/S1060992X25600156
Shiva Kamkar, Hamid Abrishami Moghaddam, Wolfram Erlhagen

Background estimation is an important part of many computer vision applications. However, it is a challenging task due to illumination changes, camouflage, occlusion, dynamic background, rain or snow fall, and shadows. In this paper, we propose a method to predict the background of videos recorded by fixed cameras. The proposed algorithm is unsupervised and online. It takes inspiration from the processing mechanisms of neural integrator circuits in recurrently connected networks. The neural activities of three distinct integrators, each responsible for processing a color channel in L * a * b color space, are updated according to the recent changes of the scene covering both spatial and temporal aspects. The maxima of the evolving activity distributions in color space are used to predict the background color value of each pixel. Evaluation results demonstrate that the proposed method outperforms several recent competitors on the Scene Background Initialization (SBI) and LASIESTA datasets, based on mean squared error (MSE) metrics.

背景估计是许多计算机视觉应用的重要组成部分。然而,由于光照变化、伪装、遮挡、动态背景、雨雪降落和阴影,这是一项具有挑战性的任务。本文提出了一种预测固定摄像机拍摄的视频背景的方法。该算法是无监督的在线算法。它的灵感来自于递归连接网络中神经积分器电路的处理机制。三个不同的积分器的神经活动,每个负责处理L * a * b颜色空间中的颜色通道,根据场景的最近变化进行更新,包括空间和时间方面。利用颜色空间中不断变化的活动分布的最大值来预测每个像素的背景颜色值。评估结果表明,基于均方误差(MSE)指标,该方法在场景背景初始化(SBI)和LASIESTA数据集上优于最近的几个竞争对手。
{"title":"Unsupervised Background Estimation Using a Neural Integrator","authors":"Shiva Kamkar,&nbsp;Hamid Abrishami Moghaddam,&nbsp;Wolfram Erlhagen","doi":"10.3103/S1060992X25600156","DOIUrl":"10.3103/S1060992X25600156","url":null,"abstract":"<p>Background estimation is an important part of many computer vision applications. However, it is a challenging task due to illumination changes, camouflage, occlusion, dynamic background, rain or snow fall, and shadows. In this paper, we propose a method to predict the background of videos recorded by fixed cameras. The proposed algorithm is unsupervised and online. It takes inspiration from the processing mechanisms of neural integrator circuits in recurrently connected networks. The neural activities of three distinct integrators, each responsible for processing a color channel in <i>L</i> * <i>a</i> * <i>b</i> color space, are updated according to the recent changes of the scene covering both spatial and temporal aspects. The maxima of the evolving activity distributions in color space are used to predict the background color value of each pixel. Evaluation results demonstrate that the proposed method outperforms several recent competitors on the Scene Background Initialization (SBI) and LASIESTA datasets, based on mean squared error (MSE) metrics.</p>","PeriodicalId":721,"journal":{"name":"Optical Memory and Neural Networks","volume":"34 4","pages":"581 - 591"},"PeriodicalIF":0.8,"publicationDate":"2025-12-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145808715","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Optical Memory and Neural Networks
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1