{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,9]],"date-time":"2025-12-09T16:57:16Z","timestamp":1765299436038,"version":"3.46.0"},"reference-count":38,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2025,12,1]],"date-time":"2025-12-01T00:00:00Z","timestamp":1764547200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2025,12,1]],"date-time":"2025-12-01T00:00:00Z","timestamp":1764547200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2025,12,1]],"date-time":"2025-12-01T00:00:00Z","timestamp":1764547200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2025,12,1]],"date-time":"2025-12-01T00:00:00Z","timestamp":1764547200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2025,12,1]],"date-time":"2025-12-01T00:00:00Z","timestamp":1764547200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2025,12,1]],"date-time":"2025-12-01T00:00:00Z","timestamp":1764547200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,12,1]],"date-time":"2025-12-01T00:00:00Z","timestamp":1764547200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100005356","name":"Zhejiang Shuren University","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100005356","id-type":"DOI","asserted-by":"publisher"}]},{"name":"2024 General Research Program Funded by Zhejiang Provincial Department of Education","award":["Y202456162"],"award-info":[{"award-number":["Y202456162"]}]},{"name":"Key Laboratory of Artificial Organs and Computational Medicine ofZhejiang Province","award":["2025A\/B10XX"],"award-info":[{"award-number":["2025A\/B10XX"]}]},{"DOI":"10.13039\/501100014786","name":"Northern Border University","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100014786","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100008867","name":"Department of Education of Zhejiang Province","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100008867","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Special funds for the basic scientific research in Provincial Universities from Zhejiang Shuren University","award":["2022XZ014"],"award-info":[{"award-number":["2022XZ014"]}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Biomedical Signal Processing and Control"],"published-print":{"date-parts":[[2025,12]]},"DOI":"10.1016\/j.bspc.2025.108127","type":"journal-article","created":{"date-parts":[[2025,6,4]],"date-time":"2025-06-04T20:11:49Z","timestamp":1749067909000},"page":"108127","update-policy":"https:\/\/linproxy.fan.workers.dev:443\/https\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"special_numbering":"PA","title":["A hybrid fusion approach combining salience perception and generative adversarial networks for multimodal medical image integration"],"prefix":"10.1016","volume":"110","author":[{"given":"Fangfang","family":"Ye","sequence":"first","affiliation":[]},{"given":"Jinming","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Ting","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Congcong","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Salwa","family":"Othmen","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"issue":"4","key":"10.1016\/j.bspc.2025.108127_b0005","doi-asserted-by":"crossref","first-page":"1540","DOI":"10.1109\/TCSS.2022.3165559","article-title":"Hierarchical progressive network for multimodal medical image fusion in healthcare systems","volume":"10","author":"Yang","year":"2022","journal-title":"IEEE Trans. Comput. Social Syst."},{"issue":"15","key":"10.1016\/j.bspc.2025.108127_b0010","doi-asserted-by":"crossref","first-page":"22577","DOI":"10.1007\/s11042-022-14266-0","article-title":"Multimodal medical volumetric image fusion using 3-D shearlet transform and TS fuzzy reasoning","volume":"82","author":"Luo","year":"2023","journal-title":"Multimed. Tools Appl."},{"key":"10.1016\/j.bspc.2025.108127_b0015","doi-asserted-by":"crossref","DOI":"10.1016\/j.bspc.2024.106561","article-title":"AMMNet: a multimodal medical image fusion method based on an attention mechanism and MobileNetV3","volume":"96","author":"Di","year":"2024","journal-title":"Biomed. Signal Process. Control"},{"key":"10.1016\/j.bspc.2025.108127_b0020","doi-asserted-by":"crossref","DOI":"10.1016\/j.bspc.2023.105301","article-title":"IE-CFRN: Information exchange-based collaborative feature representation network for multimodal medical image fusion","volume":"86","author":"Han","year":"2023","journal-title":"Biomed. Signal Process. Control"},{"key":"10.1016\/j.bspc.2025.108127_b0025","doi-asserted-by":"crossref","DOI":"10.1016\/j.bspc.2024.106358","article-title":"Multimodal medical image fusion via three-scale decomposition and structure similarity balanced optimization","volume":"95","author":"Zhao","year":"2024","journal-title":"Biomed. Signal Process. Control"},{"issue":"1","key":"10.1016\/j.bspc.2025.108127_b0030","doi-asserted-by":"crossref","first-page":"232","DOI":"10.1186\/s12880-024-01418-x","article-title":"Multimodal medical image fusion based on interval gradients and convolutional neural networks","volume":"24","author":"Gu","year":"2024","journal-title":"BMC Med. Imaging"},{"key":"10.1016\/j.bspc.2025.108127_b0035","article-title":"Medical image fusion based on multi-level bidirectional feature interaction network","author":"Shi","year":"2024","journal-title":"IEEE Sens. J."},{"issue":"4","key":"10.1016\/j.bspc.2025.108127_b0040","doi-asserted-by":"crossref","first-page":"38","DOI":"10.70023\/sahd\/241104","article-title":"Federated and multimodal learning algorithms for healthcare and cross-domain analytics","volume":"1","author":"Begum","year":"2024","journal-title":"PatternIQ Mining.,"},{"issue":"7","key":"10.1016\/j.bspc.2025.108127_b0045","doi-asserted-by":"crossref","first-page":"5907","DOI":"10.1007\/s10489-024-05431-z","article-title":"Feature extraction of multimodal medical image fusion using novel deep learning and contrast enhancement method","volume":"54","author":"Bhutto","year":"2024","journal-title":"Appl. Intell."},{"key":"10.1016\/j.bspc.2025.108127_b0050","doi-asserted-by":"crossref","DOI":"10.1016\/j.cmpb.2024.108384","article-title":"A diffusion model multiscale feature fusion network for imbalanced medical image classification research","author":"Zhu","year":"2024","journal-title":"Comp. Methods Progr. Biomed."},{"issue":"3","key":"10.1016\/j.bspc.2025.108127_b0055","article-title":"Classification of recurrent depression using brain CT images through feature fusion","volume":"17","author":"Yang","year":"2024","journal-title":"J. Radiat. Res. Appl. Sci."},{"issue":"3","key":"10.1016\/j.bspc.2025.108127_b0060","doi-asserted-by":"crossref","DOI":"10.32604\/cmc.2024.052060","article-title":"Research on multiscale feature fusion network algorithm based on brain tumor medical image classification","volume":"79","author":"Zhou","year":"2024","journal-title":"Computers, Materials & Continua"},{"key":"10.1016\/j.bspc.2025.108127_b0065","doi-asserted-by":"crossref","DOI":"10.1109\/TIM.2023.3317470","article-title":"MRSCFusion: Joint residual Swin transformer and multiscale CNN for unsupervised multimodal medical image fusion","author":"Xie","year":"2023","journal-title":"IEEE Trans. Instrum. Measure."},{"key":"10.1016\/j.bspc.2025.108127_b0070","doi-asserted-by":"crossref","DOI":"10.1016\/j.bspc.2022.104402","article-title":"DFENet: A dual-branch feature enhanced network integrating transformers and convolutional feature learning for multimodal medical image fusion","volume":"80","author":"Li","year":"2023","journal-title":"Biomed. Signal Process. Control"},{"issue":"1","key":"10.1016\/j.bspc.2025.108127_b0075","doi-asserted-by":"crossref","first-page":"203","DOI":"10.1186\/s12880-023-01160-w","article-title":"MedFusionGAN: multimodal medical image fusion using an unsupervised deep generative adversarial network","volume":"23","author":"Safari","year":"2023","journal-title":"BMC Med. Imaging"},{"key":"10.1016\/j.bspc.2025.108127_b0080","first-page":"1","article-title":"End-to-end dynamic residual focal transformer network for multimodal medical image fusion","author":"Zhang","year":"2024","journal-title":"Neural Comput. & Applic."},{"key":"10.1016\/j.bspc.2025.108127_b0085","article-title":"MSAIF-Net: A Multi-stage spatial attention based invertible fusion network for MR images","author":"Zhang","year":"2023","journal-title":"IEEE Trans. Instrum. Measure."},{"key":"10.1016\/j.bspc.2025.108127_b0090","doi-asserted-by":"crossref","DOI":"10.1016\/j.compbiomed.2023.107531","article-title":"ADDNS: An asymmetric dual deep network with sharing mechanism for medical image fusion of CT and MR-T2","volume":"166","author":"Huang","year":"2023","journal-title":"Comput. Biol. Med."},{"issue":"7","key":"10.1016\/j.bspc.2025.108127_b0095","doi-asserted-by":"crossref","first-page":"3489","DOI":"10.1109\/JBHI.2023.3264819","article-title":"An improved hybrid network with a transformer module for medical image fusion","volume":"27","author":"Liu","year":"2023","journal-title":"IEEE J. Biomed. Health Inform."},{"issue":"11","key":"10.1016\/j.bspc.2025.108127_b0100","doi-asserted-by":"crossref","first-page":"3348","DOI":"10.1109\/TMI.2023.3283517","article-title":"F-DARTS: foveated differentiable architecture search based multimodal medical image fusion","volume":"42","author":"Ye","year":"2023","journal-title":"IEEE Trans. Med. Imaging"},{"key":"10.1016\/j.bspc.2025.108127_b0105","doi-asserted-by":"crossref","DOI":"10.1109\/ACCESS.2023.3310245","article-title":"Fusion of textural and visual information for medical image modality retrieval using deep learning-based feature engineering","author":"Iqbal","year":"2023","journal-title":"IEEE Access"},{"issue":"5","key":"10.1016\/j.bspc.2025.108127_b0110","doi-asserted-by":"crossref","first-page":"494","DOI":"10.1109\/TRPMS.2023.3239520","article-title":"Medical image fusion using a new entropy measure between intuitionistic fuzzy sets joint Gaussian curvature filter","volume":"7","author":"Jiang","year":"2023","journal-title":"IEEE Trans. Radiat. Plasma Med. Sci."},{"key":"10.1016\/j.bspc.2025.108127_b0115","doi-asserted-by":"crossref","first-page":"944","DOI":"10.1109\/TMM.2023.3273924","article-title":"Msgfusion: Medical semantic guided two-branch network for multimodal brain image fusion","volume":"26","author":"Wen","year":"2023","journal-title":"IEEE Trans. Multimedia"},{"key":"10.1016\/j.bspc.2025.108127_b0120","doi-asserted-by":"crossref","first-page":"70851","DOI":"10.1109\/ACCESS.2024.3402965","article-title":"Multimodal medical image fusion network based on target information enhancement","volume":"12","author":"Zhou","year":"2024","journal-title":"IEEE Access"},{"issue":"1","key":"10.1016\/j.bspc.2025.108127_b0125","doi-asserted-by":"crossref","first-page":"17609","DOI":"10.1038\/s41598-024-68183-3","article-title":"A multibranch and multiscale neural network based on semantic perception for multimodal medical image fusion","volume":"14","author":"Lin","year":"2024","journal-title":"Sci. Rep."},{"issue":"5","key":"10.1016\/j.bspc.2025.108127_b0130","doi-asserted-by":"crossref","first-page":"264","DOI":"10.1007\/s00530-024-01473-y","article-title":"C2IENet: multibranch medical image fusion based on contrastive constraint features and information exchange","volume":"30","author":"Di","year":"2024","journal-title":"Multimedia Syst."},{"issue":"19","key":"10.1016\/j.bspc.2025.108127_b0135","doi-asserted-by":"crossref","first-page":"14379","DOI":"10.1007\/s00521-023-08493-1","article-title":"DoubleU-NetPlus: a novel attention and context-guided dual U-Net with multiscale residual feature fusion network for semantic segmentation of medical images","volume":"35","author":"Ahmed","year":"2023","journal-title":"Neural Comput. & Applic."},{"issue":"1","key":"10.1016\/j.bspc.2025.108127_b0140","doi-asserted-by":"crossref","first-page":"19261","DOI":"10.1038\/s41598-024-69997-x","article-title":"Enhanced multimodal medical image fusion via modified DWT with arithmetic optimization algorithm","volume":"14","author":"Alzahrani","year":"2024","journal-title":"Sci. Rep."},{"issue":"11","key":"10.1016\/j.bspc.2025.108127_b0145","doi-asserted-by":"crossref","first-page":"3067","DOI":"10.1007\/s11517-023-02902-2","article-title":"MBRARN: multibranch residual attention reconstruction network for medical image fusion","volume":"61","author":"Zhang","year":"2023","journal-title":"Med. Biol. Eng. Compu."},{"issue":"6","key":"10.1016\/j.bspc.2025.108127_b0150","doi-asserted-by":"crossref","first-page":"2019","DOI":"10.1007\/s00779-019-01317-x","article-title":"Medical image fusion based on saliency and adaptive similarity judgment","volume":"27","author":"Li","year":"2023","journal-title":"Pers. Ubiquit. Comput."},{"issue":"1","key":"10.1016\/j.bspc.2025.108127_b0155","doi-asserted-by":"crossref","first-page":"317","DOI":"10.1007\/s40747-022-00792-9","article-title":"Multimodal medical image fusion with convolution sparse representation and mutual information correlation in NSST domain","volume":"9","author":"Guo","year":"2023","journal-title":"Complex Intell. Syst."},{"issue":"Suppl 1","key":"10.1016\/j.bspc.2025.108127_b0160","doi-asserted-by":"crossref","first-page":"421","DOI":"10.1007\/s10462-023-10529-w","article-title":"SS-SSAN: a self-supervised subspace attentional network for multimodal medical image fusion","volume":"56","author":"Zhang","year":"2023","journal-title":"Artif. Intell. Rev."},{"key":"10.1016\/j.bspc.2025.108127_b0165","doi-asserted-by":"crossref","DOI":"10.1016\/j.eswa.2024.124467","article-title":"DBEF-Net: diffusion-based boundary-enhanced fusion network for medical image segmentation","author":"Huang","year":"2024","journal-title":"Expert Syst. Appl."},{"key":"10.1016\/j.bspc.2025.108127_b0170","doi-asserted-by":"crossref","DOI":"10.1016\/j.inffus.2024.102364","article-title":"Federated fusion learning with attention mechanism for multi-client medical image analysis","volume":"108","author":"Irfan","year":"2024","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.bspc.2025.108127_b0175","doi-asserted-by":"crossref","DOI":"10.1016\/j.jocs.2024.102324","article-title":"LitefusionNet: boosting the performance for medical image classification with an intelligent and lightweight feature fusion network","author":"Asif","year":"2024","journal-title":"J. Comput. Sci."},{"key":"10.1016\/j.bspc.2025.108127_b0180","doi-asserted-by":"crossref","DOI":"10.1016\/j.bspc.2023.105671","article-title":"Multimodal medical image fusion via multi-dictionary and truncated Huber filtering","volume":"88","author":"Jie","year":"2024","journal-title":"Biomed. Signal Process. Control"},{"issue":"8","key":"10.1016\/j.bspc.2025.108127_b0185","doi-asserted-by":"crossref","first-page":"4045","DOI":"10.3390\/s23084045","article-title":"Automation and optimization of rat heart decellularization using a vibrating fluid column","volume":"23","author":"Bonciog","year":"2023","journal-title":"Sensors"},{"key":"10.1016\/j.bspc.2025.108127_b0190","series-title":"Automation of Decellularization Process Using Artificial Neural Networks","first-page":"1","author":"Bonciog","year":"2023"}],"container-title":["Biomedical Signal Processing and Control"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/api.elsevier.com\/content\/article\/PII:S174680942500638X?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/api.elsevier.com\/content\/article\/PII:S174680942500638X?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2025,12,9]],"date-time":"2025-12-09T16:51:35Z","timestamp":1765299095000},"score":1,"resource":{"primary":{"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/linkinghub.elsevier.com\/retrieve\/pii\/S174680942500638X"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,12]]},"references-count":38,"alternative-id":["S174680942500638X"],"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/doi.org\/10.1016\/j.bspc.2025.108127","relation":{},"ISSN":["1746-8094"],"issn-type":[{"type":"print","value":"1746-8094"}],"subject":[],"published":{"date-parts":[[2025,12]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"A hybrid fusion approach combining salience perception and generative adversarial networks for multimodal medical image integration","name":"articletitle","label":"Article Title"},{"value":"Biomedical Signal Processing and Control","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/linproxy.fan.workers.dev:443\/https\/doi.org\/10.1016\/j.bspc.2025.108127","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2025 Elsevier Ltd. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"108127"}}