This project aims to determine the scaling laws for the neural networks when applied to ECG (Electrocardiogram) data. Scaling laws help in understanding how the performance of a model changes with respect to its size, the amount of data, and available compute.
This work has been accepted in the European Heart Journal – Digital Health (2025): Learning to scale: deriving data-driven scaling laws for ECG-optimized CNNs
- 👉 Link to article (to be added)
- Extend the design space to include Transformer based models. Do Transformers scale better than CNNs?
- Use the learned scaling laws to scale models to (much) larger sizes. How big can we go?
- Optimizer has a significant impact on performance and scaling capabilities
- Just using more compute does not improve performance. A careful choice of model architecture, optimizer, and hyperparamters is required.
Figure 1: Pareto front of model performance (Physionet Accuracy is a balanced accuracy described here) vs. total compute (FLOPs) for different optimizers. AdamW clearly outperforms SGD and SGD with Momentum.
Figure 2: Random Forest predicts perforamance based on hyperparameters. Shown are the regression plots for a 3-fold cross validation and the average pearson correlaction coeafficient.
- Number of stages is consistently the most important hyperparameter
- AdamW is required to optimally use model large network width (
$w_m$ ) - SGD(-Momentum) does not work well with deep networks (
$depth$ ). Use AdamW instead.
Figure 3: Gini importance of scaling hyperparameters and their spearman rank correlation with Physionet Accuracy performance metric.
Figure 3: Inference time benchmark on various hardware platforms and OpenVINO optimization levels. DML=Data Manipulation Language, GPU=NVIDIA A100, M4=Apple M4, AUTO=OpenVINO+Intel COREi7, CPU=Intel COREi7, NPU=Neural Processing Unit .
We use the PhysioNet 2021 dataset. The dataset include >88.000 annotated twelve-lead ECG recordings from six sources in four countries across three continents.
- CPSC Database and CPSC-Extra Database
- INCART Database
- PTB and PTB-XL Database
- The Georgia 12-lead ECG Challenge (G12EC) Database
- Chapman-Shaoxing and Ningbo Database
- The University of Michigan (UMich) Database
We utilize the ConvNeXt model, a convolutional neural network architecture, known for its efficiency and performance in image and signal processing tasks. Key features are:
- Stochastic depth
- LayerScale
- LayerNorm
- Inverted bottleneck
- GELU
- Depth wise convolution
- ... and more


