Pub Date : 2024-09-12DOI: 10.1016/j.optlaseng.2024.108553
In response to the challenges of acquiring spatial target position information and achieving high precision in existing methods, this paper proposes a multi-dimensional high-precision positioning method for spatial targets through multi-sensor fusion. Utilizing optical detection technology, the method extracts two-dimensional positional information of spatial targets on the observation plane. By deriving a fusion positioning formula for visible light and infrared based on the Gaussian mixture TPHD, the proposed method enhances positioning accuracy by 0.2 m compared to using visible light or infrared alone. Additionally, by integrating laser ranging for distance dimension information, precise target positioning in the world coordinate system is achieved. Outdoor experiments for spatial target positioning validate the method's effectiveness, utilizing visible light and infrared cameras along with laser ranging. Comparative analysis with a binary star angular measurement-only method demonstrates 17.9 % improvement in positioning accuracy, with the proposed method achieving 0.12 m accuracy for 5 cm spatial targets at 5 km distance.
{"title":"Research on high precision localization of space target with multi-sensor association","authors":"","doi":"10.1016/j.optlaseng.2024.108553","DOIUrl":"10.1016/j.optlaseng.2024.108553","url":null,"abstract":"<div><p>In response to the challenges of acquiring spatial target position information and achieving high precision in existing methods, this paper proposes a multi-dimensional high-precision positioning method for spatial targets through multi-sensor fusion. Utilizing optical detection technology, the method extracts two-dimensional positional information of spatial targets on the observation plane. By deriving a fusion positioning formula for visible light and infrared based on the Gaussian mixture TPHD, the proposed method enhances positioning accuracy by 0.2 m compared to using visible light or infrared alone. Additionally, by integrating laser ranging for distance dimension information, precise target positioning in the world coordinate system is achieved. Outdoor experiments for spatial target positioning validate the method's effectiveness, utilizing visible light and infrared cameras along with laser ranging. Comparative analysis with a binary star angular measurement-only method demonstrates 17.9 % improvement in positioning accuracy, with the proposed method achieving 0.12 m accuracy for 5 cm spatial targets at 5 km distance.</p></div>","PeriodicalId":49719,"journal":{"name":"Optics and Lasers in Engineering","volume":null,"pages":null},"PeriodicalIF":3.5,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142172733","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-12DOI: 10.1016/j.optlaseng.2024.108582
A multi-line laser scanning system for 3D topography measurement is proposed. This method not only has the advantages of high precision of laser scanning technology, but also has high reconstruction efficiency. In this paper, speckle reconstruction technique, multi-line laser technique and Biocular reconstruction technique are used to construct a 3D reconstruction system, and test equipment is built, and the problems existing in the system establishment process are actually studied. In order to solve the problem of mismatching in binocular multi-line laser matching, a method to sort out the correspondence of multiple laser lines in binocular images based on speckle matching results is proposed. In order to optimize the multi-line laser matching effect, a speckle matching network based on deep learning is proposed, which integrates the grayscale images of the left and right cameras as supplementary information, and takes the speckle image and grayscale image as the input of the network model to obtain more accurate and edge-complete matching results. Finally, the matching results of the multi-line laser and the camera calibration parameters were used to reconstruct the object point cloud. Experimental results show that the proposed speckle matching method can make binocular multiline laser point cloud reconstruction more robust and stable than the traditional method, and through the accuracy analysis of the system, it is proved that the average measurement accuracy of the proposed method can reach 0.05 mm.
{"title":"Multi-line laser scanning reconstruction with binocularly speckle matching and trained deep neural networks","authors":"","doi":"10.1016/j.optlaseng.2024.108582","DOIUrl":"10.1016/j.optlaseng.2024.108582","url":null,"abstract":"<div><p>A multi-line laser scanning system for 3D topography measurement is proposed. This method not only has the advantages of high precision of laser scanning technology, but also has high reconstruction efficiency. In this paper, speckle reconstruction technique, multi-line laser technique and Biocular reconstruction technique are used to construct a 3D reconstruction system, and test equipment is built, and the problems existing in the system establishment process are actually studied. In order to solve the problem of mismatching in binocular multi-line laser matching, a method to sort out the correspondence of multiple laser lines in binocular images based on speckle matching results is proposed. In order to optimize the multi-line laser matching effect, a speckle matching network based on deep learning is proposed, which integrates the grayscale images of the left and right cameras as supplementary information, and takes the speckle image and grayscale image as the input of the network model to obtain more accurate and edge-complete matching results. Finally, the matching results of the multi-line laser and the camera calibration parameters were used to reconstruct the object point cloud. Experimental results show that the proposed speckle matching method can make binocular multiline laser point cloud reconstruction more robust and stable than the traditional method, and through the accuracy analysis of the system, it is proved that the average measurement accuracy of the proposed method can reach 0.05 mm.</p></div>","PeriodicalId":49719,"journal":{"name":"Optics and Lasers in Engineering","volume":null,"pages":null},"PeriodicalIF":3.5,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142172729","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-12DOI: 10.1016/j.optlaseng.2024.108554
This paper proposes a neural network and least squares method to retrieve phase from three-frame random phase-shifting interferograms. The phase retrieval method involves two processes. Firstly, a neural network is utilized to predict phase shifts of the three-frame random phase-shifting interferograms. After the phase shifts are determined, the phase is retrieved using the least squares method. The method is simple, and does not require iterative calculation. The accuracy of the proposed method is verified by comparing the advanced iterative algorithm. Through the analysis of the simulated interferograms, the root mean square (RMS) of phase error can approach 0.1 rad. The interferograms recorded in the interferometer verifies the feasibility.
{"title":"Phase retrieval from random phase-shifting interferograms using neural network and least squares method","authors":"","doi":"10.1016/j.optlaseng.2024.108554","DOIUrl":"10.1016/j.optlaseng.2024.108554","url":null,"abstract":"<div><p>This paper proposes a neural network and least squares method to retrieve phase from three-frame random phase-shifting interferograms. The phase retrieval method involves two processes. Firstly, a neural network is utilized to predict phase shifts of the three-frame random phase-shifting interferograms. After the phase shifts are determined, the phase is retrieved using the least squares method. The method is simple, and does not require iterative calculation. The accuracy of the proposed method is verified by comparing the advanced iterative algorithm. Through the analysis of the simulated interferograms, the root mean square (RMS) of phase error can approach 0.1 rad. The interferograms recorded in the interferometer verifies the feasibility.</p></div>","PeriodicalId":49719,"journal":{"name":"Optics and Lasers in Engineering","volume":null,"pages":null},"PeriodicalIF":3.5,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142172832","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-12DOI: 10.1016/j.optlaseng.2024.108585
Phase unwrapping is a crucial step in laser interferometry for obtaining accurate physical measurement of object. To reduce the impact of speckle noise on wrapped phase during actual measurement and improve the subsequent measurement accuracy, a multi-feature fusion phase unwrapping method for different speckle noises named MFR-Net is proposed in this paper. The network is composed of a front-end multi-module filter processing layer and a back-end network with dilated convolution and coordinate attention mechanism. By reducing random phase differences introduced by different levels of noise, the network enhances its capability to extract spatial features such as gradient information between pixels under speckle noise, so that it successfully unwraps the wrapped phase with different speckle noises and accurately recovers the real phase information. Taking the wrapped phases with multiplicative speckle noise and additive random noise as dataset, the results of ablation and comparison experiments show that the MFR-Net has superior unwrapped results. Under three different levels of speckle noise, the average values of MSE, SSIM, PSNR and AU for MFR-Net are at least improved by 84.80 %, 10.99 %, 29.00 % and 7.72 %, respectively, compared to PDVQG, TIE, DLPU and VURNet algorithms. When the standard deviation of speckle noise varies continuously in the range [1.0, 2.0], the average values of four indexes reaches 0.12 rad, 0.91, 31.80 dB and 99.96 %, respectively, indicating the stronger robustness of MFR-Net. In addition, the phase step unwrapping is performed by MFR-Net. Compared to DLPU and VURNet, MFR-Net method reduced MSE by 80 % and 87.35 %, respectively, demonstrating the outstanding generalization capability. The proposed MFR-Net can realize the correct phase unwrapping under different speckle noises. It may be applied in laser interferometry applications such as digital holography and interferometric synthetic aperture radar.
{"title":"MFR-Net: A multi-feature fusion phase unwrapping method for different speckle noises","authors":"","doi":"10.1016/j.optlaseng.2024.108585","DOIUrl":"10.1016/j.optlaseng.2024.108585","url":null,"abstract":"<div><p>Phase unwrapping is a crucial step in laser interferometry for obtaining accurate physical measurement of object. To reduce the impact of speckle noise on wrapped phase during actual measurement and improve the subsequent measurement accuracy, a multi-feature fusion phase unwrapping method for different speckle noises named MFR-Net is proposed in this paper. The network is composed of a front-end multi-module filter processing layer and a back-end network with dilated convolution and coordinate attention mechanism. By reducing random phase differences introduced by different levels of noise, the network enhances its capability to extract spatial features such as gradient information between pixels under speckle noise, so that it successfully unwraps the wrapped phase with different speckle noises and accurately recovers the real phase information. Taking the wrapped phases with multiplicative speckle noise and additive random noise as dataset, the results of ablation and comparison experiments show that the MFR-Net has superior unwrapped results. Under three different levels of speckle noise, the average values of MSE, SSIM, PSNR and AU for MFR-Net are at least improved by 84.80 %, 10.99 %, 29.00 % and 7.72 %, respectively, compared to PDVQG, TIE, DLPU and VURNet algorithms. When the standard deviation of speckle noise varies continuously in the range [1.0, 2.0], the average values of four indexes reaches 0.12 rad, 0.91, 31.80 dB and 99.96 %, respectively, indicating the stronger robustness of MFR-Net. In addition, the phase step unwrapping is performed by MFR-Net. Compared to DLPU and VURNet, MFR-Net method reduced MSE by 80 % and 87.35 %, respectively, demonstrating the outstanding generalization capability. The proposed MFR-Net can realize the correct phase unwrapping under different speckle noises. It may be applied in laser interferometry applications such as digital holography and interferometric synthetic aperture radar.</p></div>","PeriodicalId":49719,"journal":{"name":"Optics and Lasers in Engineering","volume":null,"pages":null},"PeriodicalIF":3.5,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142172833","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-11DOI: 10.1016/j.optlaseng.2024.108581
Camera imaging through refractive interfaces is a crucial issue in photogrammetric measurements. Most past studies adopted numerical optimization algorithms based on refractive ray tracing procedures. In these studies, the camera and interface parameters are usually calculated iteratively with numerical optimization algorithms. Inappropriate initial values can cause iterations to diverge. Meanwhile, these iterations cannot efficiently reveal the accurate nature of refractive imaging. Therefore, obtaining camera calibration results that are both flexible and physically interpretable continues to be challenging. Consequently, in this study, we modeled refractive imaging by employing ray transfer matrix analysis. Subsequently, we deduced an analytical refractive imaging (ARI) equation that explicitly describes the refractive geometry in a matrix form. Although this equation is built upon the paraxial approximation, we executed a numerical experiment that shows that the developed analytical equation can accurately illustrate refractive imaging with a considerable object distance and a slightly tilted angle of the flat interface. This ARI equation can be used to define the expansion center and the normal vector of the flat interface. Finally, we also propose a flexible measurement method to determine the orientation of the flat interface, wherein the orientation can be measured rather than calculated by iterative procedures.
相机通过折射界面成像是摄影测量中的一个关键问题。以往的研究大多采用基于折射光线跟踪程序的数值优化算法。在这些研究中,相机和界面参数通常通过数值优化算法进行迭代计算。不恰当的初始值会导致迭代发散。同时,这些迭代无法有效揭示折射成像的精确本质。因此,获得既灵活又能从物理角度解释的相机校准结果仍然是一项挑战。因此,在本研究中,我们采用射线传递矩阵分析法对折射成像进行建模。随后,我们推导出一个分析折射成像(ARI)方程,该方程以矩阵形式明确描述了折射几何。虽然该方程是建立在准轴向近似基础上的,但我们进行的数值实验表明,所建立的分析方程可以准确地说明在物体距离较大、平面界面角度略微倾斜的情况下的折射成像。该 ARI 方程可用于定义平面界面的膨胀中心和法向量。最后,我们还提出了一种灵活的测量方法来确定平面界面的方向,其中方向可以通过测量而不是迭代程序计算得出。
{"title":"Analytical equation for camera imaging with refractive interfaces","authors":"","doi":"10.1016/j.optlaseng.2024.108581","DOIUrl":"10.1016/j.optlaseng.2024.108581","url":null,"abstract":"<div><p>Camera imaging through refractive interfaces is a crucial issue in photogrammetric measurements. Most past studies adopted numerical optimization algorithms based on refractive ray tracing procedures. In these studies, the camera and interface parameters are usually calculated iteratively with numerical optimization algorithms. Inappropriate initial values can cause iterations to diverge. Meanwhile, these iterations cannot efficiently reveal the accurate nature of refractive imaging. Therefore, obtaining camera calibration results that are both flexible and physically interpretable continues to be challenging. Consequently, in this study, we modeled refractive imaging by employing ray transfer matrix analysis. Subsequently, we deduced an analytical refractive imaging (ARI) equation that explicitly describes the refractive geometry in a matrix form. Although this equation is built upon the paraxial approximation, we executed a numerical experiment that shows that the developed analytical equation can accurately illustrate refractive imaging with a considerable object distance and a slightly tilted angle of the flat interface. This ARI equation can be used to define the expansion center and the normal vector of the flat interface. Finally, we also propose a flexible measurement method to determine the orientation of the flat interface, wherein the orientation can be measured rather than calculated by iterative procedures.</p></div>","PeriodicalId":49719,"journal":{"name":"Optics and Lasers in Engineering","volume":null,"pages":null},"PeriodicalIF":3.5,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142169551","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-11DOI: 10.1016/j.optlaseng.2024.108552
Fluorescence imaging necessitates precise matching of excitation source, dichroic mirror, emission filter, detector and dyes, which is complex and time-consuming, especially for applications of probe multiplexing. We propose a novel method for multicolor imaging based on a brightness coded set. Each brightness code consists of 12 bits (), denoting probe type, cube, emission filter, imaging result and priority, respectively. The brightness of a probe in an imaging system is defined as the product of extinction coefficient, quantum yield and the filter transmittance. When the brightness exceeds the threshold, indicates a clear image, otherwise . The higher the brightness value the higher the priority (TT). To validate the efficacy and efficiency of the coding method, we conducted two separate experiments involving four-color imaging. The proposed method offers a substantial simplification of the conventional approach to device matching in multicolor imaging by leveraging spectrograms, and presents a promising avenue for the advancement of intelligent multicolor imaging systems.
{"title":"Multicolor imaging based on brightness coded set","authors":"","doi":"10.1016/j.optlaseng.2024.108552","DOIUrl":"10.1016/j.optlaseng.2024.108552","url":null,"abstract":"<div><p>Fluorescence imaging necessitates precise matching of excitation source, dichroic mirror, emission filter, detector and dyes, which is complex and time-consuming, especially for applications of probe multiplexing. We propose a novel method for multicolor imaging based on a brightness coded set. Each brightness code consists of 12 bits (<span><math><mi>O</mi><mi>O</mi><mi>O</mi><mi>X</mi><mi>X</mi><mi>X</mi><mi>Y</mi><mi>Y</mi><mi>Y</mi><mi>Z</mi><mi>T</mi><mi>T</mi></math></span>), denoting probe type, cube, emission filter, imaging result and priority, respectively. The brightness of a probe in an imaging system is defined as the product of extinction coefficient, quantum yield and the filter transmittance. When the brightness exceeds the threshold, <span><math><mi>Z</mi><mo>=</mo><mn>1</mn></math></span> indicates a clear image, otherwise <span><math><mi>Z</mi><mo>=</mo><mn>0</mn></math></span>. The higher the brightness value the higher the priority (<em>TT</em>). To validate the efficacy and efficiency of the coding method, we conducted two separate experiments involving four-color imaging. The proposed method offers a substantial simplification of the conventional approach to device matching in multicolor imaging by leveraging spectrograms, and presents a promising avenue for the advancement of intelligent multicolor imaging systems.</p></div>","PeriodicalId":49719,"journal":{"name":"Optics and Lasers in Engineering","volume":null,"pages":null},"PeriodicalIF":3.5,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142169550","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-11DOI: 10.1016/j.optlaseng.2024.108575
Due to the absorption and scattering of light and the influence of suspended particles, underwater images commonly exhibit color distortions, reduced contrast, and diminished details. This paper proposes an attenuated color channel adaptive correction and bilateral weight fusion approach called WLAB to address the aforementioned degradation issues. Specifically, a novel white balance method is first applied to balance the color channel of the input image. Moreover, a local-block-based fast non-local means method is proposed to obtain a denoised version of the color-corrected image. Then, an adaptive stretching method that considers the histogram's local features to get a contrast-enhanced version of the color-corrected image. Finally, a bilateral weight fusion method is proposed to fuse the above two image versions to obtain an output image with complementary advantages. Experimental studies are conducted on three benchmark underwater image datasets and compared with ten state-of-the-art methods. The results show that WLAB has a significant advantage over the comparative methods. Notably, WLAB exhibits a degree of independence from camera settings and enhances the precision of various image processing applications, including key points and saliency detection. Additionally, it demonstrates commendable adaptability in improving low-light and foggy images.
{"title":"Attenuated color channel adaptive correction and bilateral weight fusion for underwater image enhancement","authors":"","doi":"10.1016/j.optlaseng.2024.108575","DOIUrl":"10.1016/j.optlaseng.2024.108575","url":null,"abstract":"<div><p>Due to the absorption and scattering of light and the influence of suspended particles, underwater images commonly exhibit color distortions, reduced contrast, and diminished details. This paper proposes an attenuated color channel adaptive correction and bilateral weight fusion approach called WLAB to address the aforementioned degradation issues. Specifically, a novel white balance method is first applied to balance the color channel of the input image. Moreover, a local-block-based fast non-local means method is proposed to obtain a denoised version of the color-corrected image. Then, an adaptive stretching method that considers the histogram's local features to get a contrast-enhanced version of the color-corrected image. Finally, a bilateral weight fusion method is proposed to fuse the above two image versions to obtain an output image with complementary advantages. Experimental studies are conducted on three benchmark underwater image datasets and compared with ten state-of-the-art methods. The results show that WLAB has a significant advantage over the comparative methods. Notably, WLAB exhibits a degree of independence from camera settings and enhances the precision of various image processing applications, including key points and saliency detection. Additionally, it demonstrates commendable adaptability in improving low-light and foggy images.</p></div>","PeriodicalId":49719,"journal":{"name":"Optics and Lasers in Engineering","volume":null,"pages":null},"PeriodicalIF":3.5,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142169400","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-11DOI: 10.1016/j.optlaseng.2024.108556
Object recognition poses a critical challenge for firefighters' search and reconnaissance equipment in smoky environments. Indeed, the absorption and scattering of smoke particles are the primary obstacles hindering recognition. This paper proposes an infrared laser line synchronous imaging system (ILLS) to improve the degradation of images. Firstly, the classical Mie scattering theory background is introduced. Then, the principles and components of ILLS are introduced. Finally, the contrast experiment is conducted on the imaging performance of LED floodlighting, Infrared (IR), and ILLS under different smoke visibility levels, illustrating each experiment step. The results demonstrate that the ILLS system performs excellently compared to other imaging methods under different smoke visibility levels. In the analysis of object image contrast (C), ILLS achieves contrast enhancements of 3.0 and 7.0 times compared to IR and LED at the visibility of 0.8 m, respectively. In evaluating the Mean Squared Error (MSE) for the object image, ILLS exhibits a deviation compared to IR of greater than or equal to 20%, regardless of visibility. ILLS effectively improve the challenge of image degradation due to severe backscattering of particles and provide sufficiently accurate solutions for engineering applications.
{"title":"Active imaging system based on infrared laser line synchronous scanning","authors":"","doi":"10.1016/j.optlaseng.2024.108556","DOIUrl":"10.1016/j.optlaseng.2024.108556","url":null,"abstract":"<div><p>Object recognition poses a critical challenge for firefighters' search and reconnaissance equipment in smoky environments. Indeed, the absorption and scattering of smoke particles are the primary obstacles hindering recognition. This paper proposes an infrared laser line synchronous imaging system (ILLS) to improve the degradation of images. Firstly, the classical Mie scattering theory background is introduced. Then, the principles and components of ILLS are introduced. Finally, the contrast experiment is conducted on the imaging performance of LED floodlighting, Infrared (IR), and ILLS under different smoke visibility levels, illustrating each experiment step. The results demonstrate that the ILLS system performs excellently compared to other imaging methods under different smoke visibility levels. In the analysis of object image contrast (C), ILLS achieves contrast enhancements of 3.0 and 7.0 times compared to IR and LED at the visibility of 0.8 m, respectively. In evaluating the Mean Squared Error (MSE) for the object image, ILLS exhibits a deviation compared to IR of greater than or equal to 20%, regardless of visibility. ILLS effectively improve the challenge of image degradation due to severe backscattering of particles and provide sufficiently accurate solutions for engineering applications.</p></div>","PeriodicalId":49719,"journal":{"name":"Optics and Lasers in Engineering","volume":null,"pages":null},"PeriodicalIF":3.5,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142169553","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-11DOI: 10.1016/j.optlaseng.2024.108555
The impact of elemental composition on the engineering properties of high-density tungsten alloy is crucial, particularly in relation to grain size, hardness, elastic modulus, and surface degradation. This study introduces a new multimodal laser opto-ultrasonic spectroscopy (LOUS) technique, which simultaneously integrates the benefits of laser-induced breakdown spectroscopy (LIBS) and laser ultrasonic testing to evaluate the engineering properties of tungsten alloys. The findings indicate that the increasing tungsten concentration significantly enhances the samples' hardness, grain size and elastic modulus. The composition of elements and hardness were assessed using a calibration curve derived from the emission intensity ratio (W-II/W-I) and plasma electron temperature (Te) in optical emission. The correlation results of (W-II/W-I) and Te showed significant enhancement with the increase of hardness with a regression coefficient (R2 ≥ 0.996), validating the Saha-Eggert relation and underscoring model effectiveness. Additionally, the correlation of the laser ultrasonic testing parameters (attenuation coefficient and velocities) in assessing grain size and elastic modulus showed good reliability (R2≥0.993) when compared to the results obtained from conventional optical microscopy and tensile testing. The results underscore the accuracy and predictive ability of the LOUS method for in-situ characterization.
元素组成对高密度钨合金工程特性的影响至关重要,尤其是在晶粒尺寸、硬度、弹性模量和表面降解方面。本研究介绍了一种新的多模式激光光学超声波谱(LOUS)技术,该技术同时集成了激光诱导击穿光谱(LIBS)和激光超声波测试的优点,用于评估钨合金的工程特性。研究结果表明,钨浓度的增加会显著提高样品的硬度、晶粒度和弹性模量。根据光学发射中的发射强度比(W-II/W-I)和等离子体电子温度(Te)得出的校准曲线,对元素组成和硬度进行了评估。(W-II/W-I) 和 Te 的相关结果表明,随着硬度的增加,相关系数(R2 ≥ 0.996)显著提高,验证了 Saha-Eggert 关系并强调了模型的有效性。此外,在评估晶粒尺寸和弹性模量时,激光超声波测试参数(衰减系数和速度)与传统光学显微镜和拉伸测试结果的相关性显示出良好的可靠性(R2≥0.993)。这些结果凸显了 LOUS 方法在原位表征方面的准确性和预测能力。
{"title":"Multimodal laser opto-ultrasonic spectroscopy (LOUS) Approach to investigate the impact of elemental composition on the engineering properties of high-density tungsten alloys'","authors":"","doi":"10.1016/j.optlaseng.2024.108555","DOIUrl":"10.1016/j.optlaseng.2024.108555","url":null,"abstract":"<div><p>The impact of elemental composition on the engineering properties of high-density tungsten alloy is crucial, particularly in relation to grain size, hardness, elastic modulus, and surface degradation. This study introduces a new multimodal laser opto-ultrasonic spectroscopy (LOUS) technique, which simultaneously integrates the benefits of laser-induced breakdown spectroscopy (LIBS) and laser ultrasonic testing to evaluate the engineering properties of tungsten alloys. The findings indicate that the increasing tungsten concentration significantly enhances the samples' hardness, grain size and elastic modulus. The composition of elements and hardness were assessed using a calibration curve derived from the emission intensity ratio (W-II/W-I) and plasma electron temperature (<em>T<sub>e</sub></em>) in optical emission. The correlation results of (W-II/W-I) and <em>T<sub>e</sub></em> showed significant enhancement with the increase of hardness with a regression coefficient (R<sup>2</sup> ≥ 0.996), validating the Saha-Eggert relation and underscoring model effectiveness. Additionally, the correlation of the laser ultrasonic testing parameters (attenuation coefficient and velocities) in assessing grain size and elastic modulus showed good reliability (R<sup>2</sup>≥0.993) when compared to the results obtained from conventional optical microscopy and tensile testing. The results underscore the accuracy and predictive ability of the LOUS method for <em>in-situ</em> characterization.</p></div>","PeriodicalId":49719,"journal":{"name":"Optics and Lasers in Engineering","volume":null,"pages":null},"PeriodicalIF":3.5,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142169554","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-09-11DOI: 10.1016/j.optlaseng.2024.108577
Metalenses have been widely used in various optical systems due to their compact size, lightweight nature and high efficiency. This paper presents a novel metalense for generating multiple optical trap arrays. The proposed metalense is composed of many unit cells of half-wave plate. By changing the polarization state of the incident light, the optical trap array generated by the metalense can be controlled. To evaluate its performance, we simulate the metalense with total trap numbers of 4 and 6 respectively. The results show that the produced optical trap arrays exhibit precise positioning, uniform size, and intensity distribution. Furthermore, the arrays maintain excellent uniformity and shape integrity even when the number of optical traps is increased. Compared with the previous metalenses only produce a specific optical trap array, the designed metalense offers higher flexibility and better meets the experimental requirements of multi-trap optical tweezers. Our metalense exhibits significant potential for the integration and miniaturization of optical trap arrays, as well as other focusing optical systems.
{"title":"Generating controllable optical trap arrays with a polarization-based metalense","authors":"","doi":"10.1016/j.optlaseng.2024.108577","DOIUrl":"10.1016/j.optlaseng.2024.108577","url":null,"abstract":"<div><p>Metalenses have been widely used in various optical systems due to their compact size, lightweight nature and high efficiency. This paper presents a novel metalense for generating multiple optical trap arrays. The proposed metalense is composed of many unit cells of half-wave plate. By changing the polarization state of the incident light, the optical trap array generated by the metalense can be controlled. To evaluate its performance, we simulate the metalense with total trap numbers of 4 and 6 respectively. The results show that the produced optical trap arrays exhibit precise positioning, uniform size, and intensity distribution. Furthermore, the arrays maintain excellent uniformity and shape integrity even when the number of optical traps is increased. Compared with the previous metalenses only produce a specific optical trap array, the designed metalense offers higher flexibility and better meets the experimental requirements of multi-trap optical tweezers. Our metalense exhibits significant potential for the integration and miniaturization of optical trap arrays, as well as other focusing optical systems.</p></div>","PeriodicalId":49719,"journal":{"name":"Optics and Lasers in Engineering","volume":null,"pages":null},"PeriodicalIF":3.5,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142169552","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}