Skip to content

Columbia-ICSL/SoundTrackDB

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

10 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

πŸƒβ€β™‚οΈ SoundTrackDB

SoundTrackDB is a comprehensive dataset of treadmill running sounds, designed to support research and development in contactless running metric estimation.

Dataset Overview

πŸ“Š Highlights:

  • 40 hours of audio data
  • 61 subjects, 363 sessions
  • Recorded in 13 public gyms
  • Developed in collaboration with a licensed running coach

This dataset captures real-world variability and is ideal for training and evaluating ML models that estimate running metrics such as:

  • Cadence
  • Ground Contact Time (GCT)
  • And many other possible metrics with our labels

Contents & Data Alignment

There are 5 folders in the "data" directory, each containing 363 files with the same prefixes, corresponding to the 363 running sessions. The filenames are in the format "date-code_subject-id_speed_trial.filetype". For example, "d01_P01_6_0.m4a" corresponds to the audio recording of the running session on date "d01" of subject "P01" running at "6" mph, and is the "0"th trial running 6 mph for that day.

Here is an overview of what a file in each folder corresponds to.

  • audio_m4a: the audio recording of the phone placed on the treadmill control panel. Consider this to be the "main" timeline such that the timestamps in all other files refer to the time of the audio recording.
  • gct_csv: the ground truth labels of cadence and ground contact time (GCT). Each row represents a single step. The important columns are (1) "YOLO_Start_Time" that corresponds to when the specifc step makes contact with the treadmill belt in the audio recording and (2) "YOLO_Contact_Time" that corresponds to the contact time of the specific step. Cadence is provided as a column but can also be derived in a custom way depending on the implementation. This file also includes and the metrics from the baseline devices which we aligned and derived. The raw output file of the baseline devices are also provide in the other folders.
  • imu_csv: the 3-axis (x, y, z) imu data of the running session collected by the phone placed on the treadmill control panel, also aligned with the audio recording. We experimented with this modality in our Microbenchmark but found no added value to the audio.
  • runscribe_csv: the output csv from the RunScribe app. We manually aligned the steps based on the method describe in Data Syncing Section of the manuscript such that the "time_sec" column is aligned with the audio.
  • watch_csv: the output csv from the sport watches we used. We pre-processed to have consistent format and aligned it to the audio.

Meta Data

We have organized data collected from our Pre-Run and Post-Run Questionnaires.

  • SoundTrack_session_info.csv contains the meta data of the 363 running session, including information such as exertion level, muscle sore, etc.
  • SoundTrack_subject_info.csv contains the meta data of the 61 participants, including information such as age, weight, running expertise, etc.

πŸ“± SoundTrack

We also introduce SoundTrack, a contactless mobile solution that estimates cadence and GCT by analyzing treadmill running sounds using on-device machine learning β€” no wearables required.

Watch the demo

Citation

If you use this dataset in your research, please cite the following paper:

Jingping Nie, Yuang Fan, Ziyi Xuan, Minghui Zhao, Runxi Wan, Matthias Preindl, and Xiaofan Jiang.
"SoundTrack: A Contactless Mobile Solution for Real-time Running Metric Estimation for Treadmill Running in the Wild."
Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies (IMWUT),
Vol. 9, No. 2, Article 42, June 2025, 30 pages.
DOI: https://doi.org/10.1145/3729486


About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 2

  •  
  •