The George B. Moody PhysioNet Challenges are annual competitions that invite participants to develop automated approaches for addressing important physiological and clinical problems. Chagas disease is a parasitic disease in Central and South America. It affects an estimated 6.5 million people and causes nearly 10,000 deaths annually. Serological testing can detect Chagas disease and allow for treatment that prevents or slows damage to the cardiovascular system, but testing capacity is limited. Chagas disease symptoms may also appear in electrocardiograms (ECG), so ECGs can help to prioritize individuals for the limited numbers of serological tests and inform the impacts of and treatments for Chagas disease. The George B. Moody PhysioNet Challenge 2025 invites teams to develop open-source algorithmic approaches for identifying potential cases of Chagas disease from standard 12-lead ECGs.
Chagas disease is a tropical parasitic disease caused by Trypanosoma cruzi and is transmitted mainly by triatomine bugs, also known as “kissing bugs”. It affects an estimated 6.5 million people in endemic countries, and causes nearly 10,000 deaths annually1. Chagas disease can also be found in immigrants living in high-income countries2. There is no human vaccine for Chagas disease.
After an acute phase, which generally occurs in childhood, Chagas disease enters a life-long chronic phase3,4. In the early stages of infection, Chagas disease has no or mild symptoms, and can be treated by specific drugs that can prevent the progression of the disease. In the later stages of infection, Chagas disease can cause cardiomyopathy, leading to heart failure, cardiac arrhythmias, and thromboembolism, and is associated with a higher risk of death. Serological testing has shown the widespread prevalence of Chagas disease in some areas, and such tests can be used for diagnosis in individual patients, but serological testing capacities are limited. However, Chagas cardiomyopathy often manifests in electrocardiograms (ECG), providing a signal for Chagas disease and informing the treatment of the resulting heart conditions.
The George B. Moody PhysioNet Challenge 2025 provides an opportunity to advance the Chagas disease screening by inviting teams to develop algorithmic approaches for using ECGs to prioritize patients for confirmatory testing for Chagas disease.
For the 2025 Challenge, we ask participants to develop and implement open-source algorithms for identifying potential cases of Chagas disease from standard 12-lead ECGs. The winners of the Challenge will be the team with the highest score on the hidden test set.
The Challenge data include standard 12-lead ECG records from several sources from Central and South America and binary labels for Chagas disease.
We are gradually introducing data for the 2025 Challenge, beginning with the CODE-15% Database5.
The CODE-15% Database5 contains over 300,000 12-lead ECG records collected in Brazil between 2010 and 2016. Most recordings have a duration of either 7.3 s or 10.2 s and a sampling frequency of 400 Hz. The binary Chagas labels are self-reported and therefore may or may not have been validated. The database may or may not include answers to a questionnaire. These data are publicly available and included in the Challenge training set.
Like all data, these data have different data collection, processing, and documentation procedures, but they are realistic examples of real-world data.
If you use these or other data as part of your method, then please cite them appropriately to clarify your method and attribute the data contributors.
The data will be split into training, validation, and test sets. The training data are publicly shared, and the validation and test data will be privately sequestered to allow for the principled assessment of the submitted algorithms. The validation and test data include sources that are not represented in the training set.
The data were minimally processed from the machines to de-identify the data as necessary and provide the ECG recordings in a consistent, standardized format for the Challenge.
The data are stored in WFDB files. We provide example code for preparing, loading, and processing the data in both MATLAB and Python. Each ECG recording includes a signal file in a WFDB-compatible binary format for the ECG signal data and a plain text file in WFDB header format for metadata about the recording as well as patient attributes, including the Chagas labels.
The WFDB header file includes the sampling frequency, signal length, signal resolution, and signal names of the channels in the ECG waveform; initial and checksum values for the channels; any available demographic information; and the Chagas labels. For example, the CODE-15% Database includes a record 113
. This record has 12 channels, a sampling frequency of 400Hz, and 4096 samples per channel:
113 12 400 4096
113.dat 16 1000(0)/mV 16 0 7767 28904 0 I
113.dat 16 1000(0)/mV 16 0 2820 -4132 0 II
113.dat 16 1000(0)/mV 16 0 -4947 32499 0 III
113.dat 16 1000(0)/mV 16 0 -5290 18601 0 AVR
113.dat 16 1000(0)/mV 16 0 6357 -9558 0 AVL
113.dat 16 1000(0)/mV 16 0 -1059 -23504 0 AVF
113.dat 16 1000(0)/mV 16 0 4451 27243 0 V1
113.dat 16 1000(0)/mV 16 0 3186 15474 0 V2
113.dat 16 1000(0)/mV 16 0 4451 28365 0 V3
113.dat 16 1000(0)/mV 16 0 3826 22578 0 V4
113.dat 16 1000(0)/mV 16 0 3948 -12925 0 V5
113.dat 16 1000(0)/mV 16 0 4367 -14130 0 V6
# Exam ID: 113
# Patient ID: 16860
# Age: 68
# Sex: Female
# Chagas label: False
You can access the data by following the instructions in this repository.
To participate in the Challenge, register your team by providing the full names, affiliations, and official email addresses of your entire team before you submit your algorithm. The details of all authors must be exactly the same as the details in your abstract submission to Computing in Cardiology. You may update your author list by completing this form again (read the form for details), but changes to your authors must not contravene the rules of the Challenge.
Each team must develop and implement an algorithm that, given a 12-lead ECG recording, provides a binary Chagas label and a probability of a positive Chagas case.
We implemented example algorithms in MATLAB and Python. Other implementation languages will be considered upon request. The code repositories contain details for the examples and other helpful scripts and functions. These examples were not designed to perform well but to provide minimal working examples of how to work with the data for the Challenge task. We will also provide a version of this code, which uses a deep neural network to detect Chagas disease.
Teams submit their code, including working training code, in a GitHub or Gitlab repository by submitting a Google Form. During the Challenge, we will train each entry on the training set and evaluate it on the validation set. After the Challenge, we will train entries on the training set and evaluate them on the test set. We try to provide feedback about the entry (a score on the validation set for successful entries and an error message for unsuccessful entries) within 72 hours of submission.
Please see the submission instructions for detailed information about how to submit a successful Challenge entry, double check your code (we cannot debug your code for you), and submit your algorithm after we begin accepting code submissions. We will provide feedback on your entry as soon as possible, so please wait at least 72 hours before contacting us about the status of your entry.
Please note that you remain the owners of any code that you submit, and we encourage you to use an open-source license.
Each region has a limited capacity for serological testing to validate potential cases of Chagas disease. We will evaluate each algorithm by prioritizing patients according to the algorithm’s probability of a positive Chagas case and reporting the true positive rate (TPR) for Chagas cases in the patients with the 5% highest probabilities. The team with the highest TPR on the hidden test set wins the Challenge.
This metric and other traditional metrics, including AUC, are implemented in the evaluate_model
script. We invite feedback about these metrics.
There are two phases for the Challenge: an unofficial phase and an official phase. The unofficial phase of the Challenge allows us to introduce and “beta test” the data, scores, and submission system before the official phase of the Challenge. Participation in the unofficial phase is mandatory for participating in the official phase of the Challenge because it helps us to improve the official phase.
Entrants may have an overall total of up to 15 scored entries over both the unofficial and official phases of the competition (see the below table). We will evaluate these entries on the validation set during the unofficial and official phases, and we will evaluate at most on successful official phase entry from each team on the test set after the official phase. All deadlines occur at 11:59pm GMT on the dates mentioned below, and all dates are during 2025 unless indicated otherwise. If you do not know the difference between GMT and your local time, then find it out before the deadline!
Please submit your entries early to ensure that you have the most chances for success. If you wait until the last few days to submit your entries, then you may not receive feedback before the submission deadline, and you may be unable to resubmit your entries if there are unexpected errors or issues with your submissions. Every year, several teams wait until the last few days to submit their first entry and are unable to debug their work before the deadline.
Although we score on a first-come-first-serve basis, please note that if you submit more than one entry in a 24-hour period, your second entry may be deprioritized compared to other teams’ first entries. If you submit more than one entry in the final 24 hours before the Challenge deadline, then we may be unable to provide feedback or a score for more than one of your entries. It is unlikely that we will be able to debug any code in the final days of the Challenge.
For these reasons, we strongly suggest that you start submitting entries at least 5 days before the unofficial deadline and 10 days before the official deadline. We have found that the earlier teams enter the Challenge, the better they do because they have time to digest feedback and performance. We therefore suggest entering your submissions many weeks before the deadline to give yourself the best chance for success.
Start | End | Submissions | |
---|---|---|---|
Unofficial phase | 09 January 2025 | 09 April 2025 | 1-5 scored entries (*) |
Hiatus | 10 April 2025 | 09 May 2025 | N/A |
Abstract deadline | 15 April 2025 | 15 April 2025 | 1 abstract |
Official phase | 10 May 2025 | 20 August 2025 | 1-10 scored entries (*) |
Abstract decisions released | Mid-June 2025 | Mid-June 2025 | N/A |
Wild card entry date | 31 July 2025 | 31 July 2025 | N/A |
Hiatus | 21 August 2025 | 13 September 2025 | N/A |
Deadline to choose algorithm for test data | 27 August 2025 | 27 August 2025 | N/A |
Preprint deadline | 27 August 2025 | 27 August 2025 | One 4-page paper (**) |
Conference | 14 September 2025 | 17 September 2025 | 1 presentation (**) |
Final scores released | Mid-September 2025 | Mid-September 2025 | N/A |
Final paper deadline | 1 October 2025 | 1 October 2025 | One 4-page paper (***) |
(* Entries that fail to score do not count against limits.)
(** Must include preliminary scores.)
(*** Must include final scores, your ranking in the Challenge, and any updates to your work as a result of feedback after presenting at CinC. This final paper deadline is earlier than the deadline given by CinC so that we can check these details.)
To be eligible for the open-source award, you must do all the following:
You must not submit an analysis of this year’s Challenge data to other conferences or journals until after CinC 2025 so that we can discuss the Challenge in a single forum. If we discover evidence that you have submitted elsewhere before the end of CinC 2025, then you will be disqualified and de-ranked on the website, banned from future Challenges, and the journal/conference will be contacted to request your article be withdrawn for contravention of the terms of use.
There are many reasons for this policy: (1) we do not release results on the test data before the end of CinC, and only reporting results on the training data increases the likelihood of overfitting and is not comparable to the official results on the test data, and (2) attempting to publish on the Challenge data before the Challengers present their results is unprofessional and comes across as a territorial grab. This requirement stands even if your abstract is rejected, but you may continue to enter the competition and receive scores. (However, unless you are accepted into the conference at a later date as a “wild card” entry, you will not be eligible to win a prize.) Of course, any publicly available data that was available before the Challenge is exempted from this condition, but any of the novelty of the Challenge (the Challenge design, the Challenge data that you downloaded from this page because it was processed for the Challenge, the scoring function, etc.) is not exempted.
After the Challenge is over and the final scores have been posted (in late September), everyone may then submit their work to a journal or another conference.
If your abstract is rejected or if you otherwise failed to qualify during the unofficial period, then there is still a chance to present as CinC and win the Challenge. A “wild card” entry has been reserved for a high-scoring entry from a team that was unable to submit an accepted abstract to CinC by the original abstract submission deadline. A successful entry must be submitted by the wild card entry deadline. We will contact eligible teams and ask them to submit an abstract. The abstract will still be reviewed as thoroughly as any other abstract accepted for the conference. See Advice on Writing an Abstract.
To improve your chances of having your abstract accepted, we offer the following advice:
You will be notified if your abstract has been accepted by email from CinC in June. You may not enter more than one abstract describing your work in the Challenge. We know you may have multiple ideas, and the actual abstract will evolve over the course of the Challenge. More information, particularly on discounts and scholarships, can be found here. We are sorry, but the Challenge Organizers do not have extra funds to enable discounts or funding to attend the conference.
Again, we cannot guarantee that your code will be run in time for the CinC abstract deadline, especially if you submit your code immediately before the deadline. It is much more important to focus on writing a high-quality abstract describing your work and submit this to the conference by abstract deadline. Please follow these instructions here carefully.
Please make sure that all of your team members are authors on your abstract. If you need to add or subtract authors, do this at least a week before the abstract deadline. Asking us to alter your team membership near or after the deadline is going to lead to confusion that could affect your score during review. It is better to be more inclusive on the abstract in terms of authorship, though, and if we find authors have moved between abstracts/teams without permission, then this is likely to lead to disqualification. As noted above, you may change the authors/team members later in the Challenge.
Please make sure that you include your team name, your official score as it appears on the leaderboard, and cross validation results in your abstract using the scoring metrics for this year’s Challenge (especially if you are unable to receive a score or are scoring poorly). The novelty of your approach and the rigor of your research is much more important during the unofficial phase. Please make sure you describe your technique and any novelty very specifically. General statements such as “a 1D CNN was used” are uninformative and will score poorly in review.
The Challenge Organizers have no ability to help with any problems with the abstract submission system. We do not operate it. Please do not email us with issues related to the abstract submission system.
Please note that each team remains the owners of their code and retain the intellectual property (IP) of their code. We encourage the use of open-source licenses for your entries.
Entries with non open-source licenses will be scored but not ranked in the official competition. All scores will be made public. At the end of the competition, all entries will be posted publicly, and therefore automatically mirrored on several sites around the world. We have no control over these sites, so we cannot remove your code even on request. Code which the organizers deem to be functional will be made publicly available after the end of the Challenge. You can request to withdraw from the Challenge, so that your entry’s performance is no longer listed in the official leaderboard, up until a week before the end of the official phase. However, the Organizers reserve the right to publish any submitted open-source code after the official phase is over. The Organizers also retain the right to use a copy of submitted code for non-commercial use. This allows us to re-score if definitions change and validate any claims made by competitors.
If no license is specified in your submission, then the license given in the example code will be added to your entry, i.e., we will assume that you have released your code under the BSD 3-Clause license.
To maintain the scientific impact of the Challenges, it is important that all Challengers contribute truly independent ideas. For this reason, we impose the following rules on team composition/collaboration:
If we discover evidence of the contravention of these rules, then you will be ineligible for a prize and your entry publicly marked as possibly associated with another entry. Although we will contact the team(s) in question, time and resources are limited and the Organizers must use their best judgment on the matter in a short period of time. The Organizers’ decision on rule violations will be final.
Similarly, no individual who is affiliated with the same research group, department, or similar organization unit (academic or industry) as one or more of the Organizers of that year’s Challenge researchers may officially participate in the Challenge for that year, even if they do not collaborate with Organizers. If you are uncertain if your shared affiliation disallows you from officially participating, then please contact the Challenge Organizers for clarification. This rule is to prevent concerns about independence or favoritism.
Please note, as we do every year, we will perform some simple tests on your code to ensure it is more usable and reusable. We suggest you also try these similar approaches, including:
Again, this is a simplified process, and we may change how we stress test your code in future tests (such as randomizing the labels), so please think about how you can ensure that your code isn’t dependent on a single set of data and labels or a single test for robustness. Of course, you should also try similar steps to check the rest of your code as well.
All of this work is in service of protecting your scientific contributions over the course of the Challenge, and we appreciate, as always, your feedback and help.
CinC 2025 will take place from 14-17 September 2025 in São Paulo, Brazil. You must attend the whole conference to be eligible for prizes. If you send someone in your place who is not a team member or co-author, then you will be disqualified and your abstract will be removed from the proceedings. In particular, it is vital that the presenter (oral or poster) can defend your work and have in-depth knowledge of all decisions made during the development of your algorithm. Due to this year’s challenges, both in-person and remote attendance are allowed, but only in-person attendees are eligible for prize money. If you require a visa to attend the conference, we strongly suggest that you apply as soon as possible. Please contact the local conference organizing committee (not the Challenge Organizers) for any visa sponsorship letters and answer any questions concerning the conference.
This year’s Challenge is generously sponsored by MathWorks.
MathWorks has generously decided to sponsor this Challenge by providing complimentary licenses to all teams that wish to use MATLAB. Users can apply for a license and learn more about MATLAB support by visiting the PhysioNet Challenge page from MathWorks. If you have questions or need technical support, then please contact MathWorks at studentcompetitions@mathworks.com.
Supported by the National Institute of Biomedical Imaging and Bioengineering (NIBIB) under NIH grant R01EB030362.
Cucunubá, Z. M., Gutiérrez-Romero, S. A., Ramírez, J., Velásquez-Ortiz, N., Ceccarelli, S., Parra-Henao, G., Henao-Martínez, A. F., Rabinovich, J., Basáñez, M., Nouvellet, P., & Abad-Franch, F. (2024). The epidemiology of Chagas disease in the Americas. The Lancet Regional Health - Americas, 37, 100881. doi: 10.1016/j.lana.2024.100881
Nepomuceno de Andrade G, Bosch-Nicolau P, Nascimento BR, Martins-Melo FR, Perel P, Geissbühler Y, Demacq C, Quijano M, Mosser JF, Cousin E, Machado ÍE, Rodrigues MLAC, Ribeiro ALP, Molina I. Prevalence of Chagas disease among Latin American immigrants in non-endemic countries: an updated systematic review and meta-analysis. Lancet Reg Health Eur. 2024 Sep 6;46:101040. doi: 10.1016/j.lanepe.2024.101040. PMID: 39290806; PMCID: PMC11407232.
Sabino EC, Nunes MCP, Blum J, Molina I, Ribeiro ALP. Cardiac involvement in Chagas disease and African trypanosomiasis. Nat Rev Cardiol. 2024 Jul 15. doi: 10.1038/s41569-024-01057-3. Epub ahead of print. PMID: 39009679.
Nunes MCP, Beaton A, Acquatella H, Bern C, Bolger AF, Echeverría LE, Dutra WO, Gascon J, Morillo CA, Oliveira-Filho J, Ribeiro ALP, Marin-Neto JA; American Heart Association Rheumatic Fever, Endocarditis and Kawasaki Disease Committee of the Council on Cardiovascular Disease in the Young; Council on Cardiovascular and Stroke Nursing; and Stroke Council. Chagas Cardiomyopathy: An Update of Current Clinical Knowledge and Management: A Scientific Statement From the American Heart Association. Circulation. 2018 Sep 18;138(12):e169-e209. doi: 10.1161/CIR.0000000000000599. PMID: 30354432.
Ribeiro AH, Ribeiro MH, Paixão GMM, Oliveira DM, Gomes PR, Canazart JA, Ferreira MPS, Andersson CR, Macfarlane PW, Meira W Jr, Schön TB, Ribeiro ALP. Automatic diagnosis of the 12-lead ECG using a deep neural network. Nat Commun. 2020 Apr 9;11(1):1760. doi: 10.1038/s41467-020-15432-4. Erratum in: Nat Commun. 2020 May 1;11(1):2227. doi: https://doi.org/10.1038/s41467-020-16172-1. PMID: 32273514; PMCID: PMC7145824.
This year’s Challenge is generously sponsored by MathWorks.
Supported by the National Institute of Biomedical Imaging and Bioengineering (NIBIB) under NIH grant number R01EB030362.
© PhysioNet Challenges. Website content licensed under the Creative Commons Attribution 4.0 International Public License.