TAG-DS TDL Challenge 2026#
Welcome to the Topological Deep Learning Challenge 2026: Bridging the Gap, sponsored by New Theory and hosted at the second Topology, Algebra, and Geometry in Data Science (TAG-DS) Conference.
Organizers: Guillermo Bernardez, Lev Telyatnikov, Mathilde Papillon, Marco Montagna, Miquel Ferriol, Raffael Theiler, Louis Van Langendonck, Olga Fink, Nina Miolane
See also
Link to the challenge repository: geometric-intelligence/TopoBench.
Important
👉 Now live: Join us in Bridging the Gap between Graph Neural Networks and Topological Deep Learning!
Motivation#
Despite rapid infrastructure advances and deep theoretical connections, the Graph Neural Network (GNN) and Topological Deep Learning (TDL) communities have largely operated in parallel. While these fields are often treated as distinct paradigms, they are profoundly intertwined: a growing body of work studies the interplay between standard message-passing and higher-order representations. Yet, the broader geometric deep learning field still lacks a unified, side-by-side comparison.
The 2026 TDL Challenge: Bridging the Gap sets out to unite these worlds. Rather than treating topological tools as an isolated ecosystem, this year’s challenge invites researchers from both communities to collaboratively map the frontier of relational learning. We invite participants to contribute and implement recent State-of-the-Art (SOTA) models across two dedicated tracks: Track 1 for GNNs, and Track 2 for TNNs.
For the first time, the TDL Challenge will go beyond implementation to feature a rigorous performance analysis of the submitted models. To achieve a truly objective comparison, both tracks will be evaluated through a shared pipeline powered by TopoBench and GraphUniverse. By leveraging GraphUniverse’s framework for generating controlled synthetic graphs, models will be tested against specific structural properties. This will allow both communities to gain insight on how different architectures from both domains handle varying degrees of homophily, heterophily, and complex degree distributions.
Through this shared benchmarking ecosystem of GNNs and TNNs, we aim to formulate data-driven answers to long-standing scientific questions:
Structural Sensitivity: How do specific graph properties (e.g., severe heterophily) impact the performance of classical GNNs versus their higher-order topological counterparts?
The Topological Component: Under what specific data regimes and controlled environments do TDL models consistently provide unique capabilities over standard SOTA GNN approaches (if any)?
Description of the Challenge#
We propose that participants implement recent, SOTA message-passing models from either the GNN or TDL literature. The core objective is to move beyond standard leaderboard metrics and rigorously evaluate how these different architectures behave under specific, controlled topological conditions.
To achieve this, participants will integrate their models into the TopoBench ecosystem and evaluate them using synthetic datasets generated by GraphUniverse. By leveraging this framework, participants will conduct a performance analysis that tests their implemented models against strict, predefined graph properties—such as varying homophily/heterophily ratios and complex degree distributions.
Embracing Modularity. Beyond just the core message-passing backbone, we strongly encourage participants to take full advantage of TopoBench’s modular architecture. For example, if your chosen SOTA model relies on a novel feature encoder, a specialized readout mechanism, or a custom loss function, you can seamlessly integrate these components into the pipeline. Beyond ensuring faithful implementations of complex architectures, this will help enrich the TopoBench ecosystem with highly reusable modules for future research across both communities.
To foster fair and structured comparison, the challenge is divided into two distinct tracks:
Track 1 — Graph Neural Networks (GNNs)#
Focuses on classical, pairwise message-passing architectures operating strictly on graph structures (e.g., modern GCNs, GATs, or deep MPNNs).
Track 2 — Topological Neural Networks (TNNs)#
Focuses on higher-order message-passing models that leverage rich topological domains (e.g., Simplicial, Cellular, or Hypergraph Neural Networks).
While the tracks are separate, both will share the exact same evaluation pipeline and controlled datasets. Participants are tasked with building the model, integrating it into the benchmarking suite, and reporting how its performance scales across different structural regimes.
Reward Outcomes#
⭐ White paper: Every submission that meets the requirements will be included in a white paper summarizing the challenge’s findings (planned via PMLR through Topology, Algebra, and Geometry in Machine Learning/Data Science 2026). Authors of qualifying submissions will be offered co-authorship. [1]
🏆 Cash prizes: Two winning teams (one per track) will be announced at TAG-DS 2026 during the Awards Ceremony.
💰 Track 1 (GNNs) winner: $1,000 USD (sponsored by New Theory).
💰 Track 2 (TNNs) winner: $1,000 USD (sponsored by New Theory).
🌴 Research internship — Geometric Intelligence Lab, UCSB (USA): A team, pending evaluation results and interest, will be invited for a visit of up to two months at the Geometric Intelligence Lab, University of California, Santa Barbara. During the visit, winners will work on cutting-edge methods bridging GNNs and TDL. Travel costs will be reimbursed and financial assistance for lodging will be provided. Remote participation is also available. [2]
🏔️ Research internship — IMOS Lab, EPFL (Switzerland): A team, pending evaluation results and interest, will be invited for a research internship at the Intelligent Maintenance and Operations Systems (IMOS) Lab at EPFL in Lausanne, Switzerland. Winners will develop advanced relational learning and TDL methods in a world-class academic environment. MSc enrollment at the time of the internship is required. Financial assistance for lodging will be provided; winners will likely need to secure a visa and work authorization.
Note
Organizers reserve the right to reallocate prize money between tracks in the event of a significant disparity in the number or quality of submissions.
Deadline#
The final submission deadline is August 12th, 2026 (AoE). Participants may continue modifying their PRs until this time.
Guidelines#
Eligibility: Participation is free and open to all. However, for legal reasons, individuals affiliated with institutions that appear in the sections of the Restricted Foreign Research Institutions list are not eligible for the reward outcomes of the challenge — including the cash prizes, internships, and co-authorship on the white paper summarizing the challenge findings.
Registration: To participate in the challenge, participants must (1) open a Pull Request (PR) on TopoBench and (2) fill out the Registration Google Form with their PR and team information. Each submission (i.e., each PR) must be accompanied by a Registration Form to be valid.
Picking a Model: Please refer to the open Pull Requests in TopoBench (see open PRs) to see which architectures are already being implemented in each track. We encourage a diverse representation of recent SOTA models, so please check open PRs to avoid duplicating efforts.
Submission: For a submission to be valid, teams must:
Submit a valid PR before the deadline.
Fill out the registration form before the deadline.
Ensure the PR passes all integration tests for the TopoBench and GraphUniverse evaluation pipeline.
Tag the PR with the appropriate category (one of:
track-1-gnn,track-2-tnn).Respect all code, documentation, and submission requirements.
Run the official evaluation pipeline on the provided GraphUniverse settings and report the results in the PR. [Exact evaluation procedure TBD — see TODO below.]
TODO
Replace the Run Evaluation bullet above with the finalized evaluation procedure (exact GraphUniverse settings, reporting format, evaluation notebook reference).
Model Implementations:
Each PR may contain at most one core model architecture.
If a model supports multiple significant variants or parameter configurations, the PR should include separate configuration files for each variant to ensure proper integration with the evaluation pipeline.
Teams:
Teams are allowed, with a maximum of 2 members. (If you wish to form a larger team, please contact the organizers — see the Questions section — for discussion and approval.)
The same team can submit multiple models through different PRs. Make sure to register on the Google Form (link) for each PR.
The same team can participate in both challenge tracks.
Early submissions:
We strongly encourage participants to submit PRs early. This allows ample time to resolve potential integration issues with the synthetic evaluation datasets.
In cases where multiple high-quality submissions cover the exact same model architecture, earlier submissions will be given priority consideration.
Submission Requirements#
A submission consists of a Pull Request (PR) to TopoBench. The PR title must follow this format:
Track: [Track1|Track2]; Team name: <team name>; Model: <Model Name>
Submissions may implement models already proposed in the literature or novel architectures. For novel models, neither the challenge nor its related publications will prevent you from publishing your own work — you retain full credit for your contribution.
Core Requirements (Both Tracks)#
Backbone — Required
Implement your model as a
torch.nn.Moduleand store it in:topobench/nn/backbones/{domain}/{model_name}.pywhere
{domain}is one ofgraph,simplicial,cell,hypergraph,combinatorial, ornon_relational(Track 1 usesgraph; Track 2 uses the appropriate topological domain).Note
The backbone is automatically discovered by TopoBench’s
ModelExportsManager— no manual registration is needed.Hydra Configuration — Required
Provide a YAML configuration file for your model at:
configs/model/{domain}/{model_name}.yamlThis file must specify the full
TBModelcomposition:feature_encoder,backbone,backbone_wrapper, andreadout. Use an existing config (e.g.,configs/model/graph/gcn.yamlfor Track 1,configs/model/simplicial/sccnn.yamlfor Track 2) as a template.
Optional Components#
Custom Feature Encoder (only if needed)
If your model requires non-standard input preprocessing, implement a custom encoder in:
topobench/nn/encoders/{encoder_name}.pyYour encoder class must inherit from
topobench.nn.encoders.base.AbstractFeatureEncoderand implement aforward(data)method returning atorch_geometric.data.Dataobject. It is automatically discovered via theLoadManager.Custom Readout (only if needed)
If your model requires a non-standard output aggregation strategy, implement a custom readout in:
topobench/nn/readouts/{readout_name}.pyYour readout class must inherit from
topobench.nn.readouts.base.AbstractZeroCellReadOut. It is automatically discovered viaReadoutExportsManager.Custom Backbone Wrapper (only if needed)
If your model has non-standard output handling (i.e., the standard
GNNWrapperor domain wrappers intopobench/nn/wrappers/are insufficient), implement a custom wrapper in:topobench/nn/wrappers/{domain}/{model_name}_wrapper.pyYour wrapper must inherit from
topobench.nn.wrappers.base.AbstractWrapper.Custom Loss (only if needed)
If your model requires an auxiliary or custom training loss (beyond the standard task loss), implement it in:
topobench/loss/model/{ModelName}Loss.pyYour loss class must inherit from
topobench.loss.base.AbstractLossand implementforward(model_out, batch). It is automatically discovered via theLoadManager. Reference it in your model config underbackbone.loss._target_(seeconfigs/model/graph/graph_mlp.yamlfor an example).
Testing — Required#
Unit Tests. All contributed files must pass the pre-existing unit tests. Any method or class not currently covered must be accompanied by new test files placed in the appropriate subdirectory mirroring the source structure:
Contributed File |
Test Location |
|---|---|
|
|
|
|
|
|
|
(test alongside backbone or in |
Each test file should include functions that correspond one-to-one with
contributed classes and methods (see test/nn/backbones/graph/test_graphmlp.py
as an example).
TopoBench uses Codecov to measure coverage. Your PR must match or exceed 93% coverage. The Codecov report will appear as a bot comment on your PR after CI runs.
Pipeline Test — Required. Fill in test/pipeline/test_pipeline.py with
your model and a compatible dataset to demonstrate that the full training
pipeline runs successfully end-to-end:
DATASET = "graph/MUTAG" # or another compatible dataset
MODELS = ["graph/your_model"] # your config path under configs/model/
The evaluation notebook provided with the challenge must also run successfully with your submitted model. No minimum training performance is required — the goal is to evaluate architectural correctness, not accuracy.
Tip
For Track 1 models that map directly onto PyG’s standard GCN/GAT/GIN
API, the existing GNNWrapper and AllCellFeatureEncoder can be
reused without modification. For Track 2 models, inspect the existing
simplicial/cell/hypergraph wrappers in topobench/nn/wrappers/ for the
closest analog.
Evaluation#
Award Categories#
One winning submission will be crowned per track:
Track 1: Best GNN model implementation
Track 2: Best TNN model implementation
Additionally, two teams will be selected for invited visits across both tracks based on overall quality, level of difficulty, and impact of contribution. Honorable mentions may also be awarded.
Evaluation Procedure#
A panel of TopoBench maintainers and collaborators will vote using the Condorcet method to select the best submission in each track.
Evaluation criteria include:
Correctness: Does the submission implement the model correctly? Is it reasonable and well-defined?
Code quality: How readable and clean is the implementation? How well does the submission respect the requirements?
Documentation & tests: Are docstrings clear? Are unit tests robust?
Important
These criteria do not reward final model performance on the dataset. The goal is to deliver well-written, usable model implementations and infrastructure that enable further experimentation and insight.
A panel of TopoBench developers and TDL experts will decide on the two teams to be invited for visits, pending interest as indicated in their Registration Forms. Internship opportunities and cash prizes are not mutually exclusive.
Questions#
Feel free to contact the organizers at topological.intelligence@gmail.com.



