{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,18]],"date-time":"2026-03-18T06:07:52Z","timestamp":1773814072132,"version":"3.50.1"},"reference-count":58,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T00:00:00Z","timestamp":1775001600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T00:00:00Z","timestamp":1775001600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T00:00:00Z","timestamp":1775001600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T00:00:00Z","timestamp":1775001600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T00:00:00Z","timestamp":1775001600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T00:00:00Z","timestamp":1775001600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T00:00:00Z","timestamp":1775001600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100012166","name":"National Key Research and Development Program of China","doi-asserted-by":"publisher","award":["2022YFF0800601"],"award-info":[{"award-number":["2022YFF0800601"]}],"id":[{"id":"10.13039\/501100012166","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62276008"],"award-info":[{"award-number":["62276008"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62250037"],"award-info":[{"award-number":["62250037"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62076010"],"award-info":[{"award-number":["62076010"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Neural Networks"],"published-print":{"date-parts":[[2026,4]]},"DOI":"10.1016\/j.neunet.2025.108290","type":"journal-article","created":{"date-parts":[[2025,11,5]],"date-time":"2025-11-05T08:15:38Z","timestamp":1762330538000},"page":"108290","update-policy":"https:\/\/linproxy.fan.workers.dev:443\/https\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"special_numbering":"C","title":["Repetitive contrastive learning enhances Mamba\u2019s selectivity in time series prediction"],"prefix":"10.1016","volume":"196","author":[{"ORCID":"https:\/\/linproxy.fan.workers.dev:443\/https\/orcid.org\/0009-0003-2618-9526","authenticated-orcid":false,"given":"Wenbo","family":"Yan","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/linproxy.fan.workers.dev:443\/https\/orcid.org\/0009-0009-5973-2648","authenticated-orcid":false,"given":"Hanzhong","family":"Cao","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/linproxy.fan.workers.dev:443\/https\/orcid.org\/0000-0001-8243-4731","authenticated-orcid":false,"given":"Ying","family":"Tan","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.neunet.2025.108290_bib0001","doi-asserted-by":"crossref","unstructured":"Ahamed, M. A., & Cheng, Q. (2024). Timemachine: A time series is worth 4 mambas for long-term forecasting. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2403.09898.","DOI":"10.3233\/FAIA240677"},{"key":"10.1016\/j.neunet.2025.108290_bib0002","unstructured":"Cao, H., Yan, W., Tan, Y., 2025. Numerion: A multi-hypercomplex model for time series forecasting. arXiv preprint arXiv: 2510.03251."},{"key":"10.1016\/j.neunet.2025.108290_bib0003","unstructured":"Challu, C., Olivares, K. G., Oreshkin, B. N., Garza, F., Mergenthaler-Canseco, M., & Dubrawski, A. (2022). N-hiTS: Neural hierarchical interpolation for time series forecasting. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2201.12886."},{"key":"10.1016\/j.neunet.2025.108290_bib0004","unstructured":"Cheng, X., Zhang, R., Zhou, J., & Xu, W. (2023). Deeptransport: Learning spatial-temporal dependency for traffic condition forecasting. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/1709.09585."},{"key":"10.1016\/j.neunet.2025.108290_bib0005","article-title":"Hybrid PDES simulation of HPC networks using zombie packets","author":"Cruz-Camacho","year":"2024","journal-title":"ACM Transactions on Modeling and Computer Simulation"},{"key":"10.1016\/j.neunet.2025.108290_bib0006","unstructured":"Gu, A., & Dao, T. (2024). Mamba: Linear-time sequence modeling with selective state spaces. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2312.00752."},{"key":"10.1016\/j.neunet.2025.108290_bib0007","unstructured":"Gu, A., Goel, K., & Christopher, R. (2022). Efficiently modeling long sequences with structured state spaces. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2111.00396."},{"key":"10.1016\/j.neunet.2025.108290_bib0008","unstructured":"Haviv, D., Rivkind, A., & Barak, O. (2019). Understanding and controlling memory in recurrent neural networks. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/1902.07275."},{"issue":"8","key":"10.1016\/j.neunet.2025.108290_bib0009","doi-asserted-by":"crossref","first-page":"1735","DOI":"10.1162\/neco.1997.9.8.1735","article-title":"Long short-term memory","volume":"9","author":"Hochreiter","year":"1997","journal-title":"Neural Computation"},{"key":"10.1016\/j.neunet.2025.108290_bib0010","unstructured":"Huang, W., Pan, J., Tang, J., Ding, Y., Xing, Y., Wang, Y., Wang, Z., & Hu, J. (2024). Ml-mamba: Efficient multi-modal large language model utilizing mamba-2. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2407.19832."},{"key":"10.1016\/j.neunet.2025.108290_bib0011","unstructured":"Jaiswal, A., Babu, A. R., Zadeh, M. Z., Banerjee, D., & Makedon, F. (2021). A survey on contrastive self-supervised learning. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2011.00362."},{"key":"10.1016\/j.neunet.2025.108290_bib0012","doi-asserted-by":"crossref","unstructured":"Kuang, Z., Bi, H., Xu, C., & Sun, J. (2025). Ecp-mamba: An efficient multi-scale self-supervised contrastive learning method with state space model for polsar image classification. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2506.01040.","DOI":"10.1109\/TGRS.2025.3601583"},{"key":"10.1016\/j.neunet.2025.108290_bib0013","unstructured":"Lee, S., Park, T., & Lee, K. (2024). Soft contrastive learning for time series. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2312.16424."},{"key":"10.1016\/j.neunet.2025.108290_bib0014","unstructured":"Li, S., Jin, X., Xuan, Y., Zhou, X., Chen, W., Wang, Y.-X., & Yan, X. (2020). Enhancing the locality and breaking the memory bottleneck of transformer on time series forecasting. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/1907.00235."},{"key":"10.1016\/j.neunet.2025.108290_bib0015","doi-asserted-by":"crossref","unstructured":"Li, S., Singh, H., & Grover, A. (2024). Mamba-ND: Selective state space modeling for multi-dimensional data. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2402.05892.","DOI":"10.1007\/978-3-031-73414-4_5"},{"key":"10.1016\/j.neunet.2025.108290_bib0016","unstructured":"Li, T., Liu, Z., Shen, Y., Wang, X., Chen, H., & Huang, S. (2023). Master: Market-guided stock transformer for stock price forecasting. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2312.15235."},{"key":"10.1016\/j.neunet.2025.108290_bib0017","doi-asserted-by":"crossref","unstructured":"Li, X., Gong, Y., Shen, Y., Qiu, X., Zhang, H., Yao, B., Qi, W., Jiang, D., Chen, W., & Duan, N. (2022). Coderetriever: Unimodal and bimodal contrastive learning for code search. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2201.10866.","DOI":"10.18653\/v1\/2022.emnlp-main.187"},{"key":"10.1016\/j.neunet.2025.108290_bib0018","unstructured":"Liang, A., Jiang, X., Sun, Y., Shi, X., & Li, K. (2024). Bi-mamba+: Bidirectional mamba for time series forecasting. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2404.15772."},{"key":"10.1016\/j.neunet.2025.108290_bib0019","unstructured":"Liu, Y., Hu, T., Zhang, H., Wu, H., Wang, S., Ma, L., & Long, M. (2024). itransformer: Inverted transformers are effective for time series forecasting. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2310.06625."},{"key":"10.1016\/j.neunet.2025.108290_bib0020","unstructured":"Luo, D., Cheng, W., Wang, Y., Xu, D., Ni, J., Yu, W., Zhang, X., Liu, Y., Chen, Y., Chen, H., & Zhang, X. (2023). Time series contrastive learning with information-aware augmentations. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2303.11911."},{"key":"10.1016\/j.neunet.2025.108290_bib0021","unstructured":"Manzoor, M. A., Albarri, S., Xian, Z., Meng, Z., Nakov, P., & Liang, S. (2024). Multimodality representation learning: A survey on evolution, pretraining and its applications. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2302.00389."},{"key":"10.1016\/j.neunet.2025.108290_bib0022","unstructured":"McInnes, L., Healy, J., & Melville, J. (2020). Umap: Uniform manifold approximation and projection for dimension reduction. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/1802.03426."},{"key":"10.1016\/j.neunet.2025.108290_bib0023","doi-asserted-by":"crossref","unstructured":"Mehrabian, A., Hoseinzade, E., Mazloum, M., & Chen, X. (2025). Mamba meets financial markets: A graph-mamba approach for stock price prediction. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2410.03707.","DOI":"10.1109\/ICASSP49660.2025.10888749"},{"key":"10.1016\/j.neunet.2025.108290_bib0024","series-title":"2017\u202fIEEE Conference on visual analytics science and technology (VAST)","first-page":"13","article-title":"Understanding hidden memories of recurrent neural networks","author":"Ming","year":"2017"},{"key":"10.1016\/j.neunet.2025.108290_bib0025","unstructured":"Nam, H., Kim, J., & Yeom, J. (2024). An adversarial learning approach to irregular time-series forecasting. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2411.19341."},{"key":"10.1016\/j.neunet.2025.108290_bib0026","unstructured":"Nie, Y., Nguyen, N. H., Sinthong, P., & Kalagnanam, J. (2023). A time series is worth 64 words: Long-term forecasting with transformers. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2211.14730."},{"key":"10.1016\/j.neunet.2025.108290_bib0027","unstructured":"Oreshkin, B. N., Carpov, D., Chapados, N., & Bengio, Y. (2020). N-BEATS: Neural basis expansion analysis for interpretable time series forecasting. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/1905.10437."},{"key":"10.1016\/j.neunet.2025.108290_bib0028","unstructured":"Patro, B. N., & Agneeswaran, V. S. (2024). SiMBA: Simplified mamba-based architecture for vision and multivariate time series. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2403.15360."},{"key":"10.1016\/j.neunet.2025.108290_bib0029","series-title":"Semantic cosine similarity","author":"Rahutomo","year":"2012"},{"key":"10.1016\/j.neunet.2025.108290_bib0030","unstructured":"Ramponi, G., Protopapas, P., Brambilla, M., & Janssen, R. (2019). T-CGAN: Conditional generative adversarial network for data augmentation in noisy time series with irregular sampling. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/1811.08295."},{"issue":"3","key":"10.1016\/j.neunet.2025.108290_bib0031","first-page":"1","article-title":"Ldheatmap: An r function for graphical display of pairwise linkage disequilibria between single nucleotide polymorphisms","volume":"16","author":"Shin","year":"2006","journal-title":"Journal of Statistical Software, Code Snippets"},{"issue":"42","key":"10.1016\/j.neunet.2025.108290_bib0032","doi-asserted-by":"crossref","first-page":"95410","DOI":"10.1007\/s11356-023-29044-0","article-title":"Quarterly electricity consumption prediction based on time series decomposition method and gray model","volume":"30","author":"Sun","year":"2023","journal-title":"Environmental Science and Pollution Research"},{"key":"10.1016\/j.neunet.2025.108290_bib0033","unstructured":"van den, O. A., Li, Y., & Vinyals, O. (2018). Representation learning with contrastive predictive coding. arXiv preprint arXiv: 1807.03748."},{"key":"10.1016\/j.neunet.2025.108290_bib0034","unstructured":"van den Oord, A., Li, Y., & Vinyals, O. (2019). Representation learning with contrastive predictive coding. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/1807.03748."},{"key":"10.1016\/j.neunet.2025.108290_bib0035","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., Kaiser, L., & Polosukhin, I. (2023). Attention is all you need. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/1706.03762."},{"issue":"6","key":"10.1016\/j.neunet.2025.108290_bib0036","doi-asserted-by":"crossref","first-page":"8682","DOI":"10.1109\/TITS.2025.3552010","article-title":"A lightweight spatio-temporal neural network with sampling-based time series decomposition for traffic forecasting","volume":"26","author":"Wang","year":"2025","journal-title":"IEEE Transactions on Intelligent Transportation Systems"},{"key":"10.1016\/j.neunet.2025.108290_bib0037","unstructured":"Wang, S., Wu, H., Shi, X., Hu, T., Luo, H., Ma, L., Zhang, J. Y., & Zhou, J. (2024a). Timemixer: Decomposable multiscale mixing for time series forecasting. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2405.14616."},{"key":"10.1016\/j.neunet.2025.108290_bib0038","unstructured":"Wang, W., Yao, L., Chen, L., Lin, B., Cai, D., He, X., & Liu, W. (2021). Crossformer: A versatile vision transformer hinging on cross-scale attention. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2108.00154."},{"key":"10.1016\/j.neunet.2025.108290_bib0039","doi-asserted-by":"crossref","unstructured":"Wang, Z., Kong, F., Feng, S., Wang, M., Yang, X., Zhao, H., Wang, D., & Zhang, Y. (2024b). Is mamba effective for time series forecasting?https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2403.11144.","DOI":"10.2139\/ssrn.4832898"},{"key":"10.1016\/j.neunet.2025.108290_bib0040","doi-asserted-by":"crossref","unstructured":"Wen, Q., Zhou, T., Zhang, C., Chen, W., Ma, Z., Yan, J., & Sun, L. (2023). Transformers in time series: A survey. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2202.07125.","DOI":"10.24963\/ijcai.2023\/759"},{"key":"10.1016\/j.neunet.2025.108290_bib0041","unstructured":"Wu, H., Hu, T., Liu, Y., Zhou, H., Wang, J., & Long, M. (2023). Timesnet: Temporal 2d-variation modeling for general time series analysis. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2210.02186."},{"key":"10.1016\/j.neunet.2025.108290_bib0042","unstructured":"Wu, H., Xu, J., Wang, J., & Long, M. (2022). Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2106.13008."},{"key":"10.1016\/j.neunet.2025.108290_bib0043","unstructured":"Wu, M., Zhuang, C., Mosse, M., Yamins, D., & Goodman, N. (2020). On mutual information in contrastive learning for visual representations. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2005.13149."},{"key":"10.1016\/j.neunet.2025.108290_bib0044","doi-asserted-by":"crossref","DOI":"10.1016\/j.knosys.2024.112257","article-title":"Simple debiased contrastive learning for sequential recommendation","volume":"300","author":"Xie","year":"2024","journal-title":"Knowledge-based systems"},{"key":"10.1016\/j.neunet.2025.108290_bib0045","unstructured":"Xu, S., Zhang, X., Wu, Y., & Wei, F. (2022). Sequence level contrastive learning for text summarization. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2109.03481."},{"key":"10.1016\/j.neunet.2025.108290_bib0046","doi-asserted-by":"crossref","unstructured":"Yan, W., Tan, Y., 2024a. Double-path adaptive-correlation spatial-temporal inverted transformer for stock time series forecasting. arXiv preprint arXiv: 2409.15662.","DOI":"10.3233\/FAIA251117"},{"key":"10.1016\/j.neunet.2025.108290_bib0047","unstructured":"Yan, W., Tan, Y., 2024b. Tcgpn: Temporal-correlation graph pre-trained network for stock forecasting. arXiv preprint arXiv: 2407.18519."},{"key":"10.1016\/j.neunet.2025.108290_bib0048","series-title":"Joint European Conference on Machine Learning and Knowledge Discovery in Databases","first-page":"22","article-title":"Hierarchical information-guided spatio-temporal mamba for stock time series forecasting","author":"Yan","year":"2025"},{"issue":"4","key":"10.1016\/j.neunet.2025.108290_bib0049","doi-asserted-by":"crossref","first-page":"5453","DOI":"10.1109\/TVT.2024.3508021","article-title":"Dynamic spatial-temporal graph convolution network for e-bike traffic flow forecasting","volume":"74","author":"Yu","year":"2025","journal-title":"IEEE Transactions on Vehicular Technology"},{"key":"10.1016\/j.neunet.2025.108290_bib0050","doi-asserted-by":"crossref","unstructured":"Yue, Z., Wang, Y., Duan, J., Yang, T., Huang, C., Tong, Y., & Xu, B. (2022). Ts2vec: Towards universal representation of time series. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2106.10466.","DOI":"10.1609\/aaai.v36i8.20881"},{"key":"10.1016\/j.neunet.2025.108290_bib0051","unstructured":"Zeng, A., Chen, M., Zhang, L., & Xu, Q. (2022). Are transformers effective for time series forecasting?https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2205.13504."},{"key":"10.1016\/j.neunet.2025.108290_bib0052","unstructured":"Zhang, C., Li, Q., Hua, L., & Song, D. (2020). Assessing the memory ability of recurrent neural networks. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2002.07422."},{"key":"10.1016\/j.neunet.2025.108290_bib0053","doi-asserted-by":"crossref","DOI":"10.1016\/j.rser.2021.111768","article-title":"Solar forecasting with hourly updated numerical weather prediction","volume":"154","author":"Zhang","year":"2022","journal-title":"Renewable and Sustainable Energy Reviews"},{"key":"10.1016\/j.neunet.2025.108290_bib0054","doi-asserted-by":"crossref","unstructured":"Zhang, W., Huang, J., Wang, R., Wei, C., Huang, W., & Qiao, Y. (2024). Integration of mamba and transformer \u2013 MAT for long-short range time series forecasting with application to weather dynamics. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2409.08530.","DOI":"10.1109\/ICECCE63537.2024.10823516"},{"key":"10.1016\/j.neunet.2025.108290_bib0055","doi-asserted-by":"crossref","unstructured":"Zhang, X., Zhao, Z., Tsiligkaridis, T., & Zitnik, M. (2022b). Self-supervised contrastive pre-training for time series via time-frequency consistency. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2206.08496.","DOI":"10.52202\/068431-0288"},{"key":"10.1016\/j.neunet.2025.108290_bib0056","unstructured":"Zhou, H., Zhang, S., Peng, J., Zhang, S., Li, J., Xiong, H., & Zhang, W. (2021). Informer: Beyond efficient transformer for long sequence time-series forecasting. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2012.07436."},{"key":"10.1016\/j.neunet.2025.108290_bib0057","unstructured":"Zhou, T., Ma, Z., Wen, Q., Wang, X., Sun, L., & Jin, R. (2022). Fedformer: Frequency enhanced decomposed transformer for long-term series forecasting. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2201.12740."},{"key":"10.1016\/j.neunet.2025.108290_bib0058","unstructured":"Zhu, Y., Jiang, B., Jin, H., Zhang, M., Gao, F., Huang, J., Lin, T., & Wang, X. (2023). Networked time series prediction with incomplete data via generative adversarial network. https:\/\/linproxy.fan.workers.dev:443\/https\/arxiv.org\/abs\/2110.02271."}],"container-title":["Neural Networks"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/api.elsevier.com\/content\/article\/PII:S0893608025011712?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/api.elsevier.com\/content\/article\/PII:S0893608025011712?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,3,18]],"date-time":"2026-03-18T05:13:29Z","timestamp":1773810809000},"score":1,"resource":{"primary":{"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/linkinghub.elsevier.com\/retrieve\/pii\/S0893608025011712"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,4]]},"references-count":58,"alternative-id":["S0893608025011712"],"URL":"https:\/\/linproxy.fan.workers.dev:443\/https\/doi.org\/10.1016\/j.neunet.2025.108290","relation":{},"ISSN":["0893-6080"],"issn-type":[{"value":"0893-6080","type":"print"}],"subject":[],"published":{"date-parts":[[2026,4]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"Repetitive contrastive learning enhances Mamba\u2019s selectivity in time series prediction","name":"articletitle","label":"Article Title"},{"value":"Neural Networks","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/linproxy.fan.workers.dev:443\/https\/doi.org\/10.1016\/j.neunet.2025.108290","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2025 Published by Elsevier Ltd.","name":"copyright","label":"Copyright"}],"article-number":"108290"}}