Development and Validation of Delirium Recognition Using Computer Vision in Neuro-critical Patients

NCT ID: NCT07136207

Last Updated: 2025-08-22

Study Results

Results pending

The study team has not published outcome measurements, participant flow, or safety data for this trial yet. Check back later for updates.

Basic Information

Get a concise snapshot of the trial, including recruitment status, study phase, enrollment targets, and key timeline milestones.

Recruitment Status

RECRUITING

Total Enrollment

1000 participants

Study Classification

OBSERVATIONAL

Study Start Date

2025-08-30

Study Completion Date

2026-01-30

Brief Summary

Review the sponsor-provided synopsis that highlights what the study is about and why it is being conducted.

This research project employs machine learning algorithms integrated with computer vision, image processing, and pattern recognition technologies to perform digital analysis of facial expression behaviors in neurocritical care patients with delirium. By constructing multidimensional high-level features of delirium, the investigators have established a classification model based on behavioral. The primary objective of this study is to address the critical challenge of achieving precise and efficient delirium diagnosis in neurologically critically ill patients through automated facial expression behavior recognition.

Detailed Description

Dive into the extended narrative that explains the scientific background, objectives, and procedures in greater depth.

This study is a prospective cohort study approved by the Ethics Committee of Beijing Tiantan Hospital. It aims to support the accurate and efficient diagnosis of delirium in neurocritical patients through a facial expression recognition system. A mobile application was developed for this study, collaboratively designed by senior clinicians and engineers from the Institute of Computing Technology, Chinese Academy of Sciences. The application is based on a stimulus paradigm designed using CAM-ICU (Confusion Assessment Method for the Intensive Care Unit) questions to record dynamic facial videos of neurocritical patients following delirium evaluation based on the DSM-V criteria.

Patients were assessed for delirium and facial expression behavior data were collected twice daily during ICU admission, in two time slots: 8:00-10:00 AM and 8:00-10:00 PM, following the study's inclusion and exclusion criteria. A trained and experienced specialist used the gold standard DSM-V to diagnose delirium. Within five minutes after completing the assessment, dynamic facial behavior video data were collected to prepare images for subsequent model development.

Various image preprocessing and data augmentation techniques were employed to prepare the images for the VGG16 model. These techniques are standard for running convolutional neural network (CNN) models. Using the "preprocess\_input"function from the Keras VGGFace module, the investigators standardized image color and size to ensure that each image met the expected input requirements for model training. For data augmentation, the investigators applied TensorFlow's "ImageDataGenerator" function to perform horizontal flipping, rotation, scaling, width and height shifting, and shearing. These augmentation techniques created a more diverse dataset, helping to prevent overfitting and improving the model's generalizability to new faces.

The investigators developed a binary classification model to identify delirium using a CNN with a pretrained backbone. The VGG16 model, based on deep learning, was adopted, leveraging transfer learning from VGGFace2, which possesses pre-existing facial feature recognition capabilities. Transfer learning allowed us to utilize prior knowledge to detect features more quickly, accurately, and with lower computational cost. The VGGFace2 model was employed for training.

Model performance was evaluated through internal validation at Beijing Tiantan Hospital and external validation at Guiyang Second People's Hospital, with metrics including accuracy, sensitivity, specificity, and F1 score. Additionally, to address the "black box" issue of machine learning, occlusion heatmap techniques were used to identify the most critical facial regions for delirium assessment, with the results visualized on a virtual face.

This model aims to support precise and efficient identification of delirium in neurocritical care units.

Conditions

See the medical conditions and disease areas that this research is targeting or investigating.

Delirium Artificial Intelligence (AI)

Study Design

Understand how the trial is structured, including allocation methods, masking strategies, primary purpose, and other design elements.

Observational Model Type

COHORT

Study Time Perspective

PROSPECTIVE

Study Groups

Review each arm or cohort in the study, along with the interventions and objectives associated with them.

Neurocritical non-delirium patients

For neurocritical non-delirium patients, the investigators record facial expression videos, which are used during model development to compare with the facial expressions of delirium patients.

No interventions assigned to this group

Neurocritical delirium patients

The investigators record facial expression videos of neurocritical delirium patients and perform frame sampling on the videos to analyze and extract the facial expression features specific to delirium. Based on this analysis, the investigators develop a model for delirium recognition in neurocritical patients.

No interventions assigned to this group

Eligibility Criteria

Check the participation requirements, including inclusion and exclusion rules, age limits, and whether healthy volunteers are accepted.

Inclusion Criteria

1. Neurocritical patients admitted to the ICU, including postoperative neurosurgical patients, stroke patients, and those receiving ICU care due to other neurological conditions.
2. Age over 18 years.
3. Signed informed consent.

Exclusion Criteria

1. Age under 18 years.
2. Persistent coma (GCS ≤ 8) within 7 days pre- and post-surgery, making delirium assessment impossible.
3. Did not survive more than 24 hours in the ICU.
4. Patients with facial paralysis, post-traumatic facial disfigurement, or other conditions that could significantly affect facial recognition.
5. Exclusion of patients with severe dementia, Parkinson's disease, depression, or other conditions that might impact facial emotional expressions.
Minimum Eligible Age

18 Years

Maximum Eligible Age

80 Years

Eligible Sex

ALL

Accepts Healthy Volunteers

No

Sponsors

Meet the organizations funding or collaborating on the study and learn about their roles.

Affiliated Jinyang Hospital of Guizhou Medical University

UNKNOWN

Sponsor Role collaborator

Affiliated Hospital of Guizhou Medical University

UNKNOWN

Sponsor Role collaborator

Beijing Tiantan Hospital

OTHER

Sponsor Role lead

Responsible Party

Identify the individual or organization who holds primary responsibility for the study information submitted to regulators.

Responsibility Role SPONSOR

Locations

Explore where the study is taking place and check the recruitment status at each participating site.

Beijing Tiantan Hospital

Beijing, Beijing Municipality, China

Site Status RECRUITING

Countries

Review the countries where the study has at least one active or historical site.

China

Central Contacts

Reach out to these primary contacts for questions about participation or study logistics.

Huang Huawei, Doctoral degree

Role: CONTACT

+8613599058877 ext. 59978000

Facility Contacts

Find local site contact details for specific facilities participating in the trial.

Shi Guangzhi Department Director, Doctoral degree

Role: primary

+8613599058877

References

Explore related publications, articles, or registry entries linked to this study.

Heintz TA, Badathala A, Wooten A, Cu CW, Wallace A, Pham B, Wallace AW, Cobert J. Preliminary Development and Validation of Automated Nociception Recognition Using Computer Vision in Perioperative Patients. Anesthesiology. 2025 Apr 1;142(4):726-737. doi: 10.1097/ALN.0000000000005370. Epub 2025 Jan 13.

Reference Type BACKGROUND
PMID: 39804295 (View on PubMed)

Atee M, Hoti K, Parsons R, Hughes JD. A novel pain assessment tool incorporating automated facial analysis: interrater reliability in advanced dementia. Clin Interv Aging. 2018 Jul 16;13:1245-1258. doi: 10.2147/CIA.S168024. eCollection 2018.

Reference Type BACKGROUND
PMID: 30038491 (View on PubMed)

Goldberg TE, Chen C, Wang Y, Jung E, Swanson A, Ing C, Garcia PS, Whittington RA, Moitra V. Association of Delirium With Long-term Cognitive Decline: A Meta-analysis. JAMA Neurol. 2020 Nov 1;77(11):1373-1381. doi: 10.1001/jamaneurol.2020.2273.

Reference Type BACKGROUND
PMID: 32658246 (View on PubMed)

Aldecoa C, Bettelli G, Bilotta F, Sanders RD, Audisio R, Borozdina A, Cherubini A, Jones C, Kehlet H, MacLullich A, Radtke F, Riese F, Slooter AJ, Veyckemans F, Kramer S, Neuner B, Weiss B, Spies CD. European Society of Anaesthesiology evidence-based and consensus-based guideline on postoperative delirium. Eur J Anaesthesiol. 2017 Apr;34(4):192-214. doi: 10.1097/EJA.0000000000000594.

Reference Type BACKGROUND
PMID: 28187050 (View on PubMed)

Ely EW, Margolin R, Francis J, May L, Truman B, Dittus R, Speroff T, Gautam S, Bernard GR, Inouye SK. Evaluation of delirium in critically ill patients: validation of the Confusion Assessment Method for the Intensive Care Unit (CAM-ICU). Crit Care Med. 2001 Jul;29(7):1370-9. doi: 10.1097/00003246-200107000-00012.

Reference Type BACKGROUND
PMID: 11445689 (View on PubMed)

Ahmed A, Garcia-Agundez A, Petrovic I, Radaei F, Fife J, Zhou J, Karas H, Moody S, Drake J, Jones RN, Eickhoff C, Reznik ME. Delirium detection using wearable sensors and machine learning in patients with intracerebral hemorrhage. Front Neurol. 2023 Jun 9;14:1135472. doi: 10.3389/fneur.2023.1135472. eCollection 2023.

Reference Type BACKGROUND
PMID: 37360342 (View on PubMed)

Al-Hindawi A, Vizcaychipi M, Demiris Y. A Dual-Camera Eye-Tracking Platform for Rapid Real-Time Diagnosis of Acute Delirium: A Pilot Study. IEEE J Transl Eng Health Med. 2024 May 7;12:488-498. doi: 10.1109/JTEHM.2024.3397737. eCollection 2024.

Reference Type BACKGROUND
PMID: 39050621 (View on PubMed)

Oh J, Cho D, Park J, Na SH, Kim J, Heo J, Shin CS, Kim JJ, Park JY, Lee B. Prediction and early detection of delirium in the intensive care unit by using heart rate variability and machine learning. Physiol Meas. 2018 Mar 27;39(3):035004. doi: 10.1088/1361-6579/aaab07.

Reference Type BACKGROUND
PMID: 29376502 (View on PubMed)

Eeles E, Tronstad O, Teodorczuk A, Flaws D, Fraser JF, Dissanayaka N. Face and content validity of a mobile delirium screening tool adapted for use in the medical setting (eDIS-MED): Welcome to the machine. Australas J Ageing. 2024 Jun;43(2):415-419. doi: 10.1111/ajag.13288. Epub 2024 Feb 28.

Reference Type BACKGROUND
PMID: 38415380 (View on PubMed)

Nejati V, Khorrami AS, Fonoudi M. Neuromodulation of facial emotion recognition in health and disease: A systematic review. Neurophysiol Clin. 2022 Jun;52(3):183-201. doi: 10.1016/j.neucli.2022.03.005. Epub 2022 Apr 12.

Reference Type BACKGROUND
PMID: 35428551 (View on PubMed)

Provided Documents

Download supplemental materials such as informed consent forms, study protocols, or participant manuals.

Document Type: Study Protocol and Statistical Analysis Plan

View Document

Document Type: Informed Consent Form

View Document

Other Identifiers

Review additional registry numbers or institutional identifiers associated with this trial.

PX2023021

Identifier Type: -

Identifier Source: org_study_id

More Related Trials

Additional clinical trials that may be relevant based on similarity analysis.