菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-05-13
📄 Abstract - Compact Latent Manifold Translation: A Parameter-Efficient Foundation Model for Cross-Modal and Cross-Frequency Physiological Signal Synthesis

The analysis of physiological time series, such as electrocardiograms (ECG) and photoplethysmograms (PPG), is persistently hindered by modality and frequency gaps stemming from heterogeneous recording devices. Existing foundation models typically rely on continuous latent spaces, which frequently suffer from severe modality entanglement, lack high-fidelity cross-frequency generative capacity, and impose high computational costs that prohibit edge-device deployment. In this paper, we propose Compact Latent Manifold Translation (CLMT), a highly parameter-efficient (0.09B) unified framework that bridges these gaps through a novel two-stage discrete translation paradigm. First, we introduce a Universal Tokenizer utilizing Hierarchical Residual Vector Quantization (RVQ) to decouple heterogeneous signals into isolated, well-structured discrete latent manifolds, effectively preventing inter-modality interference. Second, a Context-Prompted Latent Translator maps these discrete tokens across modalities by integrating static physiological priors, reframing complex signal synthesis as a pure latent sequence translation task. Extensive evaluations demonstrate that our 0.09B model significantly outperforms massive baselines. In cross-modal PPG-to-ECG synthesis, it resolves temporal phase drift and dramatically improves the clinical R-peak detection F1-score from 0.37 (baseline) to 0.83. Furthermore, in extreme cross-frequency super-resolution (25Hz to 100Hz), it successfully recovers high-frequency diagnostic landmarks, achieving an unprecedented Pearson correlation of 0.9956. By learning a universal discrete language for biological signals with a fraction of the computational footprint, our approach sets a new trajectory for edge-deployable, multi-modal medical foundation models.

顶级标签: medical machine learning model training
详细标签: physiological signals discrete latent space vector quantization cross-modal synthesis biomedical foundation model 或 搜索:

紧凑潜在流形翻译:一种用于跨模态与跨频率生理信号合成的参数高效基础模型 / Compact Latent Manifold Translation: A Parameter-Efficient Foundation Model for Cross-Modal and Cross-Frequency Physiological Signal Synthesis


1️⃣ 一句话总结

本文提出了一种名为CLMT的超轻量级AI模型(仅0.09B参数),通过先将心电图与光电容积描记图等不同生理信号分解为独立且结构化的离散代码,再像翻译句子一样跨模态和跨频率地转换这些代码,从而在极低计算成本下大幅提升了信号合成的精度,例如能将脉搏波合成心电图的心跳检测准确率从0.37提升至0.83。

源自 arXiv: 2605.13248