Detailed Information

Cited 13 time in webofscience Cited 17 time in scopus
Metadata Downloads

End-to-End Sleep Staging Using Nocturnal Sounds from Microphone Chips for Mobile Devicesopen access

Authors
Hong, JoonkiTran, Hai HongJung, JinhwanJang, HyeryungLee, DongheonYoon, In-YoungHong, Jung KyungKim, Jeong-Whun
Issue Date
Jun-2022
Publisher
Dove Medical Press Ltd.
Keywords
respiratory sounds; sleep stages; deep learning; smartphone; polysomnography
Citation
Nature and Science of Sleep, v.14, pp 1187 - 1201
Pages
15
Indexed
SCIE
SCOPUS
Journal Title
Nature and Science of Sleep
Volume
14
Start Page
1187
End Page
1201
URI
https://scholarworks.dongguk.edu/handle/sw.dongguk/3884
DOI
10.2147/NSS.S361270
ISSN
1179-1608
Abstract
Purpose: Nocturnal sounds contain numerous information and are easily obtainable by a non-contact manner. Sleep staging using nocturnal sounds recorded from common mobile devices may allow daily at-home sleep tracking. The objective of this study is to introduce an end-to-end (sound-to-sleep stages) deep learning model for sound-based sleep staging designed to work with audio from microphone chips, which are essential in mobile devices such as modern smartphones. Patients and Methods: Two different audio datasets were used: audio data routinely recorded by a solitary microphone chip during polysomnography (PSG dataset, N=1154) and audio data recorded by a smartphone (smartphone dataset, N=327). The audio was converted into Mel spectrogram to detect latent temporal frequency patterns of breathing and body movement from ambient noise. The proposed neural network model learns to first extract features from each 30-second epoch and then analyze inter-epoch relationships of extracted features to finally classify the epochs into sleep stages. Results: Our model achieved 70% epoch-by-epoch agreement for 4-class (wake, light, deep, REM) sleep stage classification and robust performance across various signal-to-noise conditions. The model performance was not considerably affected by sleep apnea or periodic limb movement. External validation with smartphone dataset also showed 68% epoch-by-epoch agreement. Conclusion: The proposed end-to-end deep learning model shows potential of low-quality sounds recorded from microphone chips to be utilized for sleep staging. Future study using nocturnal sounds recorded from mobile devices at home environment may further confirm the use of mobile device recording as an at-home sleep tracker.
Files in This Item
There are no files associated with this item.
Appears in
Collections
College of Advanced Convergence Engineering > Department of Computer Science and Artificial Intelligence > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Researcher Jang, Hye Ryung photo

Jang, Hye Ryung
College of Advanced Convergence Engineering (Department of Computer Science and Artificial Intelligence)
Read more

Altmetrics

Total Views & Downloads

BROWSE