Based on the existing plant layout and process flow, a simulation analysis was conducted using the Plant Simulation platform with the utilization efficiency of each station and production capacity of the dismantling system as indicators. A problem with long-term suspension in the disassembly process was determined. Based on the two optimization directions of increasing material transportation equipment and expanding the buffer capacity, a cost-oriented optimization model is established. A genetic algorithm and model simulation were used to solve the model. An optimization scheme that satisfies the production needs and has the lowest cost is proposed. The results show that the optimized dismantling system solves the suspended work problem at the dismantling station and a significant improvement in productivity and station utilization efficiency compared with the previous system.
{"title":"Simulation and optimization of scrap wagon dismantling system based on Plant Simulation.","authors":"Hai-Qing Chen, Yu-De Dong, Fei Hu, Ming-Ming Liu, Shi-Bao Zhang","doi":"10.1186/s42492-023-00134-7","DOIUrl":"https://doi.org/10.1186/s42492-023-00134-7","url":null,"abstract":"<p><p>Based on the existing plant layout and process flow, a simulation analysis was conducted using the Plant Simulation platform with the utilization efficiency of each station and production capacity of the dismantling system as indicators. A problem with long-term suspension in the disassembly process was determined. Based on the two optimization directions of increasing material transportation equipment and expanding the buffer capacity, a cost-oriented optimization model is established. A genetic algorithm and model simulation were used to solve the model. An optimization scheme that satisfies the production needs and has the lowest cost is proposed. The results show that the optimized dismantling system solves the suspended work problem at the dismantling station and a significant improvement in productivity and station utilization efficiency compared with the previous system.</p>","PeriodicalId":52384,"journal":{"name":"Visual Computing for Industry, Biomedicine, and Art","volume":"6 1","pages":"7"},"PeriodicalIF":0.0,"publicationDate":"2023-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10126191/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9389329","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-03-29DOI: 10.1186/s42492-023-00133-8
Zhenxing Xu, Aizeng Wang, Fei Hou, Gang Zhao
Gears play an important role in virtual manufacturing systems for digital twins; however, the image of gear tooth defects is difficult to acquire owing to its non-convex shape. In this study, a deep learning network is proposed to detect gear defects based on their point cloud representation. This approach mainly consists of three steps: (1) Various types of gear defects are classified into four cases (fracture, pitting, glue, and wear); A 3D gear dataset was constructed with 10000 instances following the aforementioned classification. (2) Gear-PCNet+ + introduces a novel Combinational Convolution Block, proposed based on the gear dataset for gear defect detection to effectively extract the local gear information and identify its complex topology; (3) Compared with other methods, experiments show that this method can achieve better recognition results for gear defects with higher efficiency and practicability.
{"title":"Defect detection of gear parts in virtual manufacturing.","authors":"Zhenxing Xu, Aizeng Wang, Fei Hou, Gang Zhao","doi":"10.1186/s42492-023-00133-8","DOIUrl":"https://doi.org/10.1186/s42492-023-00133-8","url":null,"abstract":"<p><p>Gears play an important role in virtual manufacturing systems for digital twins; however, the image of gear tooth defects is difficult to acquire owing to its non-convex shape. In this study, a deep learning network is proposed to detect gear defects based on their point cloud representation. This approach mainly consists of three steps: (1) Various types of gear defects are classified into four cases (fracture, pitting, glue, and wear); A 3D gear dataset was constructed with 10000 instances following the aforementioned classification. (2) Gear-PCNet+ + introduces a novel Combinational Convolution Block, proposed based on the gear dataset for gear defect detection to effectively extract the local gear information and identify its complex topology; (3) Compared with other methods, experiments show that this method can achieve better recognition results for gear defects with higher efficiency and practicability.</p>","PeriodicalId":52384,"journal":{"name":"Visual Computing for Industry, Biomedicine, and Art","volume":"6 1","pages":"6"},"PeriodicalIF":0.0,"publicationDate":"2023-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10060497/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9576684","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-03-17DOI: 10.1186/s42492-023-00132-9
István Csoba, Roland Kunkli
Vision-simulated imagery-the process of generating images that mimic the human visual system-is a valuable tool with a wide spectrum of possible applications, including visual acuity measurements, personalized planning of corrective lenses and surgeries, vision-correcting displays, vision-related hardware development, and extended reality discomfort reduction. A critical property of human vision is that it is imperfect because of the highly influential wavefront aberrations that vary from person to person. This study provides an overview of the existing computational image generation techniques that properly simulate human vision in the presence of wavefront aberrations. These algorithms typically apply ray tracing with a detailed description of the simulated eye or utilize the point-spread function of the eye to perform convolution on the input image. Based on the description of the vision simulation techniques, several of their characteristic features have been evaluated and some potential application areas and research directions have been outlined.
{"title":"Rendering algorithms for aberrated human vision simulation.","authors":"István Csoba, Roland Kunkli","doi":"10.1186/s42492-023-00132-9","DOIUrl":"https://doi.org/10.1186/s42492-023-00132-9","url":null,"abstract":"<p><p>Vision-simulated imagery-the process of generating images that mimic the human visual system-is a valuable tool with a wide spectrum of possible applications, including visual acuity measurements, personalized planning of corrective lenses and surgeries, vision-correcting displays, vision-related hardware development, and extended reality discomfort reduction. A critical property of human vision is that it is imperfect because of the highly influential wavefront aberrations that vary from person to person. This study provides an overview of the existing computational image generation techniques that properly simulate human vision in the presence of wavefront aberrations. These algorithms typically apply ray tracing with a detailed description of the simulated eye or utilize the point-spread function of the eye to perform convolution on the input image. Based on the description of the vision simulation techniques, several of their characteristic features have been evaluated and some potential application areas and research directions have been outlined.</p>","PeriodicalId":52384,"journal":{"name":"Visual Computing for Industry, Biomedicine, and Art","volume":"6 1","pages":"5"},"PeriodicalIF":0.0,"publicationDate":"2023-03-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10023823/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9145801","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
This study presents a robustness optimization method for rapid prototyping (RP) of functional artifacts based on visualized computing digital twins (VCDT). A generalized multiobjective robustness optimization model for RP of scheme design prototype was first built, where thermal, structural, and multidisciplinary knowledge could be integrated for visualization. To implement visualized computing, the membership function of fuzzy decision-making was optimized using a genetic algorithm. Transient thermodynamic, structural statics, and flow field analyses were conducted, especially for glass fiber composite materials, which have the characteristics of high strength, corrosion resistance, temperature resistance, dimensional stability, and electrical insulation. An electrothermal experiment was performed by measuring the temperature and changes in temperature during RP. Infrared thermographs were obtained using thermal field measurements to determine the temperature distribution. A numerical analysis of a lightweight ribbed ergonomic artifact is presented to illustrate the VCDT. Moreover, manufacturability was verified based on a thermal-solid coupled finite element analysis. The physical experiment and practice proved that the proposed VCDT provided a robust design paradigm for a layered RP between the steady balance of electrothermal regulation and manufacturing efficacy under hybrid uncertainties.
{"title":"Robustness optimization for rapid prototyping of functional artifacts based on visualized computing digital twins.","authors":"Jinghua Xu, Kunqian Liu, Linxuan Wang, Hongshuai Guo, Jiangtao Zhan, Xiaojian Liu, Shuyou Zhang, Jianrong Tan","doi":"10.1186/s42492-023-00131-w","DOIUrl":"https://doi.org/10.1186/s42492-023-00131-w","url":null,"abstract":"<p><p>This study presents a robustness optimization method for rapid prototyping (RP) of functional artifacts based on visualized computing digital twins (VCDT). A generalized multiobjective robustness optimization model for RP of scheme design prototype was first built, where thermal, structural, and multidisciplinary knowledge could be integrated for visualization. To implement visualized computing, the membership function of fuzzy decision-making was optimized using a genetic algorithm. Transient thermodynamic, structural statics, and flow field analyses were conducted, especially for glass fiber composite materials, which have the characteristics of high strength, corrosion resistance, temperature resistance, dimensional stability, and electrical insulation. An electrothermal experiment was performed by measuring the temperature and changes in temperature during RP. Infrared thermographs were obtained using thermal field measurements to determine the temperature distribution. A numerical analysis of a lightweight ribbed ergonomic artifact is presented to illustrate the VCDT. Moreover, manufacturability was verified based on a thermal-solid coupled finite element analysis. The physical experiment and practice proved that the proposed VCDT provided a robust design paradigm for a layered RP between the steady balance of electrothermal regulation and manufacturing efficacy under hybrid uncertainties.</p>","PeriodicalId":52384,"journal":{"name":"Visual Computing for Industry, Biomedicine, and Art","volume":"6 1","pages":"4"},"PeriodicalIF":0.0,"publicationDate":"2023-02-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9971427/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9379049","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-01-23DOI: 10.1186/s42492-023-00130-x
Qingsong Yang, Donna L Lizotte, Wenxiang Cong, Ge Wang
Over recent years, the importance of the patent literature has become increasingly more recognized in the academic setting. In the context of artificial intelligence, deep learning, and data sciences, patents are relevant to not only industry but also academe and other communities. In this article, we focus on deep tomographic imaging and perform a preliminary landscape analysis of the related patent literature. Our search tool is PatSeer. Our patent bibliometric data is summarized in various figures and tables. In particular, we qualitatively analyze key deep tomographic patent literature.
{"title":"Preliminary landscape analysis of deep tomographic imaging patents.","authors":"Qingsong Yang, Donna L Lizotte, Wenxiang Cong, Ge Wang","doi":"10.1186/s42492-023-00130-x","DOIUrl":"https://doi.org/10.1186/s42492-023-00130-x","url":null,"abstract":"<p><p>Over recent years, the importance of the patent literature has become increasingly more recognized in the academic setting. In the context of artificial intelligence, deep learning, and data sciences, patents are relevant to not only industry but also academe and other communities. In this article, we focus on deep tomographic imaging and perform a preliminary landscape analysis of the related patent literature. Our search tool is PatSeer. Our patent bibliometric data is summarized in various figures and tables. In particular, we qualitatively analyze key deep tomographic patent literature.</p>","PeriodicalId":52384,"journal":{"name":"Visual Computing for Industry, Biomedicine, and Art","volume":"6 1","pages":"3"},"PeriodicalIF":0.0,"publicationDate":"2023-01-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9868030/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10614260","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-01-14DOI: 10.1186/s42492-022-00129-w
Nathan Wang, Mengzhou Li, Petteri Haverinen
Thermal ablation procedures, such as high intensity focused ultrasound and radiofrequency ablation, are often used to eliminate tumors by minimally invasively heating a focal region. For this task, real-time 3D temperature visualization is key to target the diseased tissues while minimizing damage to the surroundings. Current computed tomography (CT) thermometry is based on energy-integrated CT, tissue-specific experimental data, and linear relationships between attenuation and temperature. In this paper, we develop a novel approach using photon-counting CT for material decomposition and a neural network to predict temperature based on thermal characteristics of base materials and spectral tomographic measurements of a volume of interest. In our feasibility study, distilled water, 50 mmol/L CaCl2, and 600 mmol/L CaCl2 are chosen as the base materials. Their attenuations are measured in four discrete energy bins at various temperatures. The neural network trained on the experimental data achieves a mean absolute error of 3.97 °C and 1.80 °C on 300 mmol/L CaCl2 and a milk-based protein shake respectively. These experimental results indicate that our approach is promising for handling non-linear thermal properties for materials that are similar or dissimilar to our base materials.
{"title":"Photon-counting computed tomography thermometry via material decomposition and machine learning.","authors":"Nathan Wang, Mengzhou Li, Petteri Haverinen","doi":"10.1186/s42492-022-00129-w","DOIUrl":"https://doi.org/10.1186/s42492-022-00129-w","url":null,"abstract":"<p><p>Thermal ablation procedures, such as high intensity focused ultrasound and radiofrequency ablation, are often used to eliminate tumors by minimally invasively heating a focal region. For this task, real-time 3D temperature visualization is key to target the diseased tissues while minimizing damage to the surroundings. Current computed tomography (CT) thermometry is based on energy-integrated CT, tissue-specific experimental data, and linear relationships between attenuation and temperature. In this paper, we develop a novel approach using photon-counting CT for material decomposition and a neural network to predict temperature based on thermal characteristics of base materials and spectral tomographic measurements of a volume of interest. In our feasibility study, distilled water, 50 mmol/L CaCl<sub>2</sub>, and 600 mmol/L CaCl<sub>2</sub> are chosen as the base materials. Their attenuations are measured in four discrete energy bins at various temperatures. The neural network trained on the experimental data achieves a mean absolute error of 3.97 °C and 1.80 °C on 300 mmol/L CaCl<sub>2</sub> and a milk-based protein shake respectively. These experimental results indicate that our approach is promising for handling non-linear thermal properties for materials that are similar or dissimilar to our base materials.</p>","PeriodicalId":52384,"journal":{"name":"Visual Computing for Industry, Biomedicine, and Art","volume":"6 1","pages":"2"},"PeriodicalIF":0.0,"publicationDate":"2023-01-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9840722/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9098823","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2023-01-03DOI: 10.1186/s42492-022-00128-x
Min Li, Hongshu Li, Weiliang Meng, Jian Zhu, Gary Zhang
In classical smoothed particle hydrodynamics (SPH) fluid simulation approaches, the smoothing length of Lagrangian particles is typically constant. One major disadvantage is the lack of adaptiveness, which may compromise accuracy in fluid regions such as splashes and surfaces. Attempts to address this problem used variable smoothing lengths. Yet the existing methods are computationally complex and non-efficient, because the smoothing length is typically calculated using iterative optimization. Here, we propose an efficient non-iterative SPH fluid simulation method with variable smoothing length (VSLSPH). VSLSPH correlates the smoothing length to the density change, and adaptively adjusts the smoothing length of particles with high accuracy and low computational cost, enabling large time steps. Our experimental results demonstrate the advantages of the VSLSPH approach in terms of its simulation accuracy and efficiency.
{"title":"An efficient non-iterative smoothed particle hydrodynamics fluid simulation method with variable smoothing length.","authors":"Min Li, Hongshu Li, Weiliang Meng, Jian Zhu, Gary Zhang","doi":"10.1186/s42492-022-00128-x","DOIUrl":"https://doi.org/10.1186/s42492-022-00128-x","url":null,"abstract":"<p><p>In classical smoothed particle hydrodynamics (SPH) fluid simulation approaches, the smoothing length of Lagrangian particles is typically constant. One major disadvantage is the lack of adaptiveness, which may compromise accuracy in fluid regions such as splashes and surfaces. Attempts to address this problem used variable smoothing lengths. Yet the existing methods are computationally complex and non-efficient, because the smoothing length is typically calculated using iterative optimization. Here, we propose an efficient non-iterative SPH fluid simulation method with variable smoothing length (VSLSPH). VSLSPH correlates the smoothing length to the density change, and adaptively adjusts the smoothing length of particles with high accuracy and low computational cost, enabling large time steps. Our experimental results demonstrate the advantages of the VSLSPH approach in terms of its simulation accuracy and efficiency.</p>","PeriodicalId":52384,"journal":{"name":"Visual Computing for Industry, Biomedicine, and Art","volume":"6 1","pages":"1"},"PeriodicalIF":0.0,"publicationDate":"2023-01-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9810768/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10489490","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-12-19DOI: 10.1186/s42492-022-00126-z
Kang Wu, Xiao-Ming Fu, Renjie Chen, Ligang Liu
Visual arts refer to art experienced primarily through vision. 3D visual optical art is one of them. Artists use their rich imagination and experience to combine light and objects to give viewers an unforgettable visual experience. However, the design process involves much trial and error; therefore, it is often very time-consuming. This has prompted many researchers to focus on proposing various algorithms to simplify the complicated design processes and help artists quickly realize the arts in their minds. To help computer graphics researchers interested in creating 3D visual optical art, we first classify and review relevant studies, then extract a general framework for solving 3D visual optical art design problems, and finally propose possible directions for future research.
{"title":"Survey on computational 3D visual optical art design.","authors":"Kang Wu, Xiao-Ming Fu, Renjie Chen, Ligang Liu","doi":"10.1186/s42492-022-00126-z","DOIUrl":"https://doi.org/10.1186/s42492-022-00126-z","url":null,"abstract":"<p><p>Visual arts refer to art experienced primarily through vision. 3D visual optical art is one of them. Artists use their rich imagination and experience to combine light and objects to give viewers an unforgettable visual experience. However, the design process involves much trial and error; therefore, it is often very time-consuming. This has prompted many researchers to focus on proposing various algorithms to simplify the complicated design processes and help artists quickly realize the arts in their minds. To help computer graphics researchers interested in creating 3D visual optical art, we first classify and review relevant studies, then extract a general framework for solving 3D visual optical art design problems, and finally propose possible directions for future research.</p>","PeriodicalId":52384,"journal":{"name":"Visual Computing for Industry, Biomedicine, and Art","volume":"5 1","pages":"31"},"PeriodicalIF":0.0,"publicationDate":"2022-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9760587/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10399702","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-12-09DOI: 10.1186/s42492-022-00125-0
Jed D Pack, Mufeng Xu, Ge Wang, Lohendran Baskaran, James Min, Bruno De Man
This review paper aims to summarize cardiac CT blooming artifacts, how they present clinically and what their root causes and potential solutions are. A literature survey was performed covering any publications with a specific interest in calcium blooming and stent blooming in cardiac CT. The claims from literature are compared and interpreted, aiming at narrowing down the root causes and most promising solutions for blooming artifacts. More than 30 journal publications were identified with specific relevance to blooming artifacts. The main reported causes of blooming artifacts are the partial volume effect, motion artifacts and beam hardening. The proposed solutions are classified as high-resolution CT hardware, high-resolution CT reconstruction, subtraction techniques and post-processing techniques, with a special emphasis on deep learning (DL) techniques. The partial volume effect is the leading cause of blooming artifacts. The partial volume effect can be minimized by increasing the CT spatial resolution through higher-resolution CT hardware or advanced high-resolution CT reconstruction. In addition, DL techniques have shown great promise to correct for blooming artifacts. A combination of these techniques could avoid repeat scans for subtraction techniques.
{"title":"Cardiac CT blooming artifacts: clinical significance, root causes and potential solutions.","authors":"Jed D Pack, Mufeng Xu, Ge Wang, Lohendran Baskaran, James Min, Bruno De Man","doi":"10.1186/s42492-022-00125-0","DOIUrl":"https://doi.org/10.1186/s42492-022-00125-0","url":null,"abstract":"<p><p>This review paper aims to summarize cardiac CT blooming artifacts, how they present clinically and what their root causes and potential solutions are. A literature survey was performed covering any publications with a specific interest in calcium blooming and stent blooming in cardiac CT. The claims from literature are compared and interpreted, aiming at narrowing down the root causes and most promising solutions for blooming artifacts. More than 30 journal publications were identified with specific relevance to blooming artifacts. The main reported causes of blooming artifacts are the partial volume effect, motion artifacts and beam hardening. The proposed solutions are classified as high-resolution CT hardware, high-resolution CT reconstruction, subtraction techniques and post-processing techniques, with a special emphasis on deep learning (DL) techniques. The partial volume effect is the leading cause of blooming artifacts. The partial volume effect can be minimized by increasing the CT spatial resolution through higher-resolution CT hardware or advanced high-resolution CT reconstruction. In addition, DL techniques have shown great promise to correct for blooming artifacts. A combination of these techniques could avoid repeat scans for subtraction techniques.</p>","PeriodicalId":52384,"journal":{"name":"Visual Computing for Industry, Biomedicine, and Art","volume":"5 1","pages":"29"},"PeriodicalIF":0.0,"publicationDate":"2022-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9733770/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9334832","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-12-09DOI: 10.1186/s42492-022-00127-y
Lin Fu, Bruno De Man
Deep learning (DL) has shown unprecedented performance for many image analysis and image enhancement tasks. Yet, solving large-scale inverse problems like tomographic reconstruction remains challenging for DL. These problems involve non-local and space-variant integral transforms between the input and output domains, for which no efficient neural network models are readily available. A prior attempt to solve tomographic reconstruction problems with supervised learning relied on a brute-force fully connected network and only allowed reconstruction with a 1284 system matrix size. This cannot practically scale to realistic data sizes such as 5124 and 5126 for three-dimensional datasets. Here we present a novel framework to solve such problems with DL by casting the original problem as a continuum of intermediate representations between the input and output domains. The original problem is broken down into a sequence of simpler transformations that can be well mapped onto an efficient hierarchical network architecture, with exponentially fewer parameters than a fully connected network would need. We applied the approach to computed tomography (CT) image reconstruction for a 5124 system matrix size. This work introduces a new kind of data-driven DL solver for full-size CT reconstruction without relying on the structure of direct (analytical) or iterative (numerical) inversion techniques. This work presents a feasibility demonstration of full-scale learnt reconstruction, whereas more developments will be needed to demonstrate superiority relative to traditional reconstruction approaches. The proposed approach is also extendable to other imaging problems such as emission and magnetic resonance reconstruction. More broadly, hierarchical DL opens the door to a new class of solvers for general inverse problems, which could potentially lead to improved signal-to-noise ratio, spatial resolution and computational efficiency in various areas.
{"title":"Deep learning tomographic reconstruction through hierarchical decomposition of domain transforms.","authors":"Lin Fu, Bruno De Man","doi":"10.1186/s42492-022-00127-y","DOIUrl":"https://doi.org/10.1186/s42492-022-00127-y","url":null,"abstract":"<p><p>Deep learning (DL) has shown unprecedented performance for many image analysis and image enhancement tasks. Yet, solving large-scale inverse problems like tomographic reconstruction remains challenging for DL. These problems involve non-local and space-variant integral transforms between the input and output domains, for which no efficient neural network models are readily available. A prior attempt to solve tomographic reconstruction problems with supervised learning relied on a brute-force fully connected network and only allowed reconstruction with a 128<sup>4</sup> system matrix size. This cannot practically scale to realistic data sizes such as 512<sup>4</sup> and 512<sup>6</sup> for three-dimensional datasets. Here we present a novel framework to solve such problems with DL by casting the original problem as a continuum of intermediate representations between the input and output domains. The original problem is broken down into a sequence of simpler transformations that can be well mapped onto an efficient hierarchical network architecture, with exponentially fewer parameters than a fully connected network would need. We applied the approach to computed tomography (CT) image reconstruction for a 512<sup>4</sup> system matrix size. This work introduces a new kind of data-driven DL solver for full-size CT reconstruction without relying on the structure of direct (analytical) or iterative (numerical) inversion techniques. This work presents a feasibility demonstration of full-scale learnt reconstruction, whereas more developments will be needed to demonstrate superiority relative to traditional reconstruction approaches. The proposed approach is also extendable to other imaging problems such as emission and magnetic resonance reconstruction. More broadly, hierarchical DL opens the door to a new class of solvers for general inverse problems, which could potentially lead to improved signal-to-noise ratio, spatial resolution and computational efficiency in various areas.</p>","PeriodicalId":52384,"journal":{"name":"Visual Computing for Industry, Biomedicine, and Art","volume":"5 1","pages":"30"},"PeriodicalIF":0.0,"publicationDate":"2022-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9733764/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9334833","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}