Prediction of Pulmonary Graft Dysfunction After Double-lung Transplantation (PGD3-AI Study)
NCT ID: NCT04643665
Last Updated: 2020-11-25
Study Results
The study team has not published outcome measurements, participant flow, or safety data for this trial yet. Check back later for updates.
Basic Information
Get a concise snapshot of the trial, including recruitment status, study phase, enrollment targets, and key timeline milestones.
COMPLETED
478 participants
OBSERVATIONAL
2012-01-01
2020-10-05
Brief Summary
Review the sponsor-provided synopsis that highlights what the study is about and why it is being conducted.
We included all double lung transplantation from 2012 to 2019 and excluded multi-organ transplant, cardiopulmonary bypass, or repeated transplantation during the study period for the same patient. Recipient, donor and intraoperative data were added in a gradient boosting algorithm step-by-step according to standard transplantation stages. Dataset will be split randomly as 80% training set and 20% testing set. Relationship between predictive factors and PGD3 will be represented as ShHapley Additive exPlanation (SHAP) values.
Detailed Description
Dive into the extended narrative that explains the scientific background, objectives, and procedures in greater depth.
Continuous variables are presented as median + interquartile range (IQR) or mean and 95%CI, and were compared using independent T-test or Mann-Whitney test. Categorical variables are presented as n (%) and were compared using Chi-squared test or Fisher's exact test. We applied machine learning algorithm to predict 3-day ahead primary graft dysfunction after lung transplant surgery among patients. Machine learning is a branch of artificial intelligence where computer systems can learn from available data and identify patterns with minimal human intervention. Machine learning algorithm tests on data and performance metrics were used to obtain the higher performing algorithm. In this study, we performed a XGBoost (Gradient Boosting) algorithm which was a combination of decisions trees. Each decision tree typically learned from its precursor and passed on the improved function to the following. The weighted combination of these trees provided the prediction.
No particular data transformation has been performed on numerical variables. Categorical variables have been encoded as integer, without any further pre-processing steps. In particular, no specific processing has been performed to deal with missing data. The default behavior of XGBoost has been used. It consists in treating missing data as a specific modality. During the training step of XGBoost models, missing values are treated as other values, and left or right decisions at any branch of a tree are learned by optimizing the outcome.
In order to reflect the sequential nature of this predictive medicine problem, nine steps have been defined to take into account incrementally observed variables acquired at various stages of the surgery.
Step 1: recipient variables Step 2: donor variables Step 3: arrival in the OR Step 4: after anesthetic induction Step 5: during first pneumonectomy Step 6: after first graft implantation Step 7: second pneumonectomy Step 8: second graft implantation Step 9: end surgery status At each of the nine steps, a cross-validation procedure is employed to assess the predictive performance of a machine learning model (XGBoost). One repetition of the cross-validation procedure is designed as follows: the dataset of subjects is randomly split into eight disjoint parts. Successively, the performance of the XGBoost model on each of the eight subset,while training the machine learning model using the remaining seven subsets. For such a repetition, the predictive probability of 3-day ahead primary graft dysfunction for each subject is retained to finally compute the area under ROC (receiving operator curve). To evaluate the variability of the predictive performance of the machine learning model, this cross-validation procedure is repeated fifty times, with randomly chosen subjects partitions. For each of the fifty times eight times nine (repetitions, partitions, surgical steps), hence 3600 models training, a conservative approach has been adopted for XGBoost training, consisting in a unique set of training parameters. These parameters have been chosen to prevent overfitting due to a relatively small number of subjects compared to the number of variables, especially categorical variables, which yield a high degree of freedom. Specifically XGBoost has been trained for 400 rounds (no early stopping), a maximum depth of 5 for each tree, a minimum child weight of 3, and a learning parameter eta equals to 0.0002. Besides those conservative parameters chosen to prevent overfitting, only 40 percents of available columns are selected for tree construction at each round, and 95 % of subjects. These parameters have been kept fixed and chosen to ensure stability of results. Small perturbations around these values could result in local performance improvements, but would not be practically chosen given the size of the dataset.
In order to gain some insights into the most useful variables in terms of predictive power, we then conducted a post-hoc analysis based on the following methodology: at each surgical step, 400 models have been trained for the repeated cross-validation procedure. For each model, we retain the rank of each variable as given by the variable importance procedure of XGBoost. The average rank of each variable for each step is then computed by averaging the ranks obtained by variables for each of the 400 models. At step 9, variables are ordered based on their average rank (increasing average ranks). They are then incrementally used as input of a new cross-validation procedure (repeated 20 times).
Conditions
See the medical conditions and disease areas that this research is targeting or investigating.
Study Design
Understand how the trial is structured, including allocation methods, masking strategies, primary purpose, and other design elements.
COHORT
RETROSPECTIVE
Study Groups
Review each arm or cohort in the study, along with the interventions and objectives associated with them.
No grade 3 Pulmonary graft dysfunction at postoperative day 3
patients having not a grade 3 Pulmonary graft dysfunction at postoperative day 3
Double-lung transplantation
Grade 3 Pulmonary graft dysfunction at postoperative day 3
patients having a grade 3 Pulmonary graft dysfunction at postoperative day 3
Double-lung transplantation
Interventions
Learn about the drugs, procedures, or behavioral strategies being tested and how they are applied within this trial.
Double-lung transplantation
Eligibility Criteria
Check the participation requirements, including inclusion and exclusion rules, age limits, and whether healthy volunteers are accepted.
Inclusion Criteria
Exclusion Criteria
* use of a cardiopulmonary bypass
* repeated transplantation during the study period for the same patient.
12 Years
ALL
No
Sponsors
Meet the organizations funding or collaborating on the study and learn about their roles.
Hopital Foch
OTHER
Responsible Party
Identify the individual or organization who holds primary responsibility for the study information submitted to regulators.
Principal Investigators
Learn about the lead researchers overseeing the trial and their institutional affiliations.
Elisabeth Hulier Ammar, PhD
Role: STUDY_DIRECTOR
Hopital Foch
Other Identifiers
Review additional registry numbers or institutional identifiers associated with this trial.
1111111111111
Identifier Type: -
Identifier Source: org_study_id