Home

Awesome

<img src='https://s3.amazonaws.com/drivendata-public-assets/logo-white-blue.png' width='600'> <br><br>

Banner Image

Senior Data Science: Safe Aging with SPHERE

Goal of the Competition

This challenge is part of a large research project which centers around using sensors and algorithms to help older people live safely at home while maintaining their privacy and independence. Using passive, automated monitoring, the ultimate goal is to look out for a person's well-being without being burdensome or intrusive.

To gather data, researchers in the SPHERE Inter-disciplinary Research Collaboration (IRC) equipped volunteers with accelerometers similar to those found in cell phones or fitness wearables, and then had the subjects go about normal activities of daily living in a home-like environment that was also equipped with motion detectors. After gathering a robust set of sensor data, they had multiple annotators use camera footage to establish the ground truth, labeling chunks of sensor data as one of twenty specifically chosen activities (e.g. walk, sit, stand-to-bend, ascend stairs, descend stairs, etc).

What's in this Repository

This repository contains code volunteered from leading competitors in the Senior Data Science: Safe Aging with SPHERE on DrivenData. Code for all winning solutions are open source under the MIT License. Code for all winning solutions are open source under the MIT License.

Winning code for other DrivenData competitions is available in the competition-winners repository.

Winning Submissions

PlaceTeam or UserPublic ScorePrivate ScoreSummary of Model
1Daniel_FG0.12930.1346I selected 7 subsets of features by random/optimization and I trained a bunch of Layer 1 models with them (ExtraTrees, RandomForest, XGBoost, Neural Networks and Linear Regression). In Layer 2, I selected again 4 subset of meta-features (L1 models) by random/optimization and I trained 7 L2 models. The final submission is a weighted average of theses predictions.
2dataiku0.12920.1381We decided to give the stacking technique a try. This ensemble learning technique combines the predictions of diverse individual models and leverages their strengths. Stacking 10 individual models – including linear regressions, Naive Bayes classifiers, random forests, extra-trees and XGBoost models – turned out to be very performant.
3n20.13390.1397We then trained dozens of models using various algorithms such as logistic regression, random forest, and extreme gradient boosting tree. Our final submission is an ensemble model of these base models.

Winner's Interview: "Meet the winners of the safe aging challenge"

Benchmark Blog Post: "There's no place like $HOME"