Detailed Information

Cited 0 time in webofscience Cited 0 time in scopus
Metadata Downloads

An Interpersonal Dynamics Analysis Procedure With Accurate Voice Activity Detection Using Low-Cost Recording Sensorsopen access

Authors
Lim, DongcheolKang, HyewonChoi, BeomseokHong, WoonkiLee, Junghye
Issue Date
2024
Publisher
IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
Keywords
Feature extraction; Sensors; Recording; Smoothing methods; Voice activity detection; Wearable sensors; Machine learning; Behavioral sciences; Verbal activity detection; organizational behavior; conversational feature; machine learning; noise smoothing filter
Citation
IEEE ACCESS, v.12, pp 68427 - 68440
Pages
14
Indexed
SCIE
SCOPUS
Journal Title
IEEE ACCESS
Volume
12
Start Page
68427
End Page
68440
URI
https://scholarx.skku.edu/handle/2021.sw.skku/111226
DOI
10.1109/ACCESS.2024.3387279
ISSN
2169-3536
Abstract
Voice data is of special interest to organizational behavior researchers since it can be easily collected and provides a wealth of information for understanding interpersonal dynamics. Due to these benefits, voice activity detection (VAD) using wearable sensor badges has received considerable attention as one of the more objective and effective data-driven analysis methods rather than self-report methods for analyzing interpersonal dynamics. Moreover, with the VAD results, several prior works extracted conversational features, such as speaking times, turn-taking, and overlapped speaking, for attaining high-level organizational insights. However, there is still room for improvement in the accuracy of the VAD models, and there is a lack of research to develop reliable conversational feature extracting algorithms in the previous works. Most of these prior studies relied on sociometric badges, which are costly electronic devices, including voice recording machines and redundant sensors. In this paper, we propose an interpersonal dynamics analysis procedure based on low-cost commercial recording sensors, consisting of data-driven VAD modeling and conversational feature extracting. To accurately identify voice presence, the VAD modeling incorporates three steps: signal preprocessing, derived-variable generation, and VAD with machine learning-based classification and smoothing technique. We conducted experiments, from collecting datasets comprising the human verbal interaction of 15 three-person groups by commercial recording sensors to implementing our procedure on the datasets. Our results show that the proposed procedure excels in voice activity detection, achieving superior accuracy compared to prior studies. This remarkable accuracy subsequently ensures greater reliability in our conversational feature extraction. Thus, the proposed procedure can encourage organizational behavior researchers to acquire objective information about interpersonal dynamics and efficiently obtain high-level organizational insights with cost-effectiveness.
Files in This Item
There are no files associated with this item.
Appears in
Collections
Computing and Informatics > Convergence > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Altmetrics

Total Views & Downloads

BROWSE