BSKI: Advancing Vision-Language Models for Understanding Activities of Daily Living – Accepted at AAAI 2025

Categories: AI4Health News

We are excited to share that our paper, “BSKI Models: Skeleton-Induced Vision-Language Embeddings for Understanding Activities of Daily Living,” has been accepted at AAAI 2025. This research introduces BSKI, a Skeleton-Induced Vision-Language model designed for recognizing Activities of Daily Living (ADLs), with critical applications in healthcare, elderly care, and behavioral monitoring.

Enhancing Health Monitoring with BSKI

Monitoring ADLs is essential for early detection of health anomalies, fall risks, and behavioral changes that may indicate underlying medical conditions. While existing vision-based models analyze daily activities, they often lack robustness in recognizing fine-grained actions and biomechanical cues. BSKI addresses these challenges by leveraging 3D skeleton-based representations alongside vision-language embeddings to enhance ADL understanding.

Key Contributions of BSKI:

  • Skeleton-Induced Vision-Language Modeling: Integrates skeletal pose estimation with vision-language embeddings to improve activity recognition.
  • Enhanced ADL Detection for Senior Care: Facilitates early intervention in fall risks and behavioral changes, promoting independent living.
  • Real-Time Insights for Caregivers & Healthcare Professionals: Provides continuous monitoring to ensure timely medical assistance.

Impact and Applications

BSKI has the potential to revolutionize elderly care by enabling continuous, real-time ADL tracking. This can enhance personalized healthcare, improve patient safety, and support independent living while assisting caregivers in providing proactive care.

Acknowledgments

This work was developed by Arkaprava Sinha, Dominick Reilly, Francois Bremond, Pu Wang, and Srijan Das, and will be presented at AAAI 2025.

📄 Read more about BSKI: Preprint

Stay tuned for further updates on AI-driven advancements in healthcare and ADL monitoring.