Demajh logo Demajh, Inc.

ReDeLEx: A Framework for Relational Deep Learning Exploration: what it means for business leaders

ReDeLEx converts everyday SQL databases into graph structures and benchmarks cutting-edge graph-neural models, letting leaders trial advanced AI on existing tables before funding costly data-re-engineering projects.

1. What the method is

ReDeLEx is an open-source framework that introspects relational schemas, builds heterogeneous graphs, and systematically evaluates multiple graph-neural-network architectures alongside classical baselines. It standardises data splits, hyper-parameter sweeps, and reporting, giving teams a reproducible sandbox for testing relational deep-learning ideas across dozens of real-world databases.

2. Why the method was developed

Enterprises store mission-critical data in normalized SQL tables, yet typical ML pipelines flatten those schemas, losing relational signal and demanding heavy feature-engineering. Existing GNN studies use hand-picked datasets and incomparable setups. The authors built ReDeLEx to provide a rigorous benchmark, reveal when GNNs actually beat tabular models, and guide data leaders toward architectures that generalise across diverse relational workloads.

3. Who should care

4. How the method works

The toolkit connects to a relational database, maps each row to a node and each foreign-key link to a typed edge, producing a heterogeneous graph. GraphSAGE, GAT, R-GCN and other models then learn over this structure with task-specific heads. A YAML pipeline automates preprocessing, Bayesian hyper-parameter tuning, seed sweeps, and leaderboard generation. Classic tabular algorithms—gradient boosting, logistic regression—are run on the same train-validation-test splits, enabling apples-to-apples comparisons without manual munging.

5. How it was evaluated

ReDeLEx benchmarks 70+ datasets from the CTU Relational Learning Repository, covering finance, health, and retail. Each dataset is scored on node- and graph-level prediction tasks using accuracy, F1, and AUC. Five random seeds quantify variance, and results are reported with 95 % confidence intervals to ensure statistical robustness.

6. How it performed

The best GNN beat classical baselines on 82 % of datasets, delivering a median 6.3-point accuracy boost while halving feature-engineering effort. Performance scaled smoothly with schema size and relational depth, suggesting strong generalisation potential. (Source: arXiv 2506.22199, 2025)

← Back to dossier index