Literature DB >> 34007870

Indoor and outdoor environmental data: A dataset with acoustic data acquired by the microphone embedded on mobile devices.

Ivan Miguel Pires1,2,3, Nuno M Garcia1, Eftim Zdravevski4, Petre Lameski4.   

Abstract

All mobile devices include a microphone that can be used for acoustic data acquisition. This article presents a dataset of acoustic signals related to nine environments, captured with a microphone embedded on off-the-shelf mobile devices. The mobile phone can be placed in the pants pockets, in a wristband, over the bedside table, on a table, or on other furniture. Data collection environments are bar, classroom, gym, kitchen, library, street, hall, living room, and bedroom. The data was collected by 25 individuals (15 men and 10 women) in different environments around Covilhã and Fundão municipalities (Portugal). The microphone data was sampled with 44,100 Hz into an array with 16-bit unsigned integer values in the range [0, 255] with a 128 offset for zero. The dataset presented in this paper presents at least 2000 samples of 5 s of data for each environment, corresponding to around 2.8 h for each environment into text files. In total, it includes at least 25.2 h of acoustic data for the implementation of data processing techniques, e.g., Fast Fourier Transform (FFT), and other machine learning methods for the different analysis.
© 2021 The Author(s). Published by Elsevier Inc.

Entities:  

Keywords:  Acoustic data; Environment; Microphone; Mobile devices

Year:  2021        PMID: 34007870      PMCID: PMC8111260          DOI: 10.1016/j.dib.2021.107051

Source DB:  PubMed          Journal:  Data Brief        ISSN: 2352-3409


Specifications Table

Value of the Data

The importance of this dataset is related to the creation of methods and/or patterns for the automatic and easy identification of environments for the monitoring systems for different types of people. The presented dataset includes general environments that can be frequented by diverse people; The data will allow the development of automatic methods for the identification of the environments that the persons are frequenting, and the promotion of healthy habits at the different locations [3]; The use of mobile devices for the data acquisition allows the identification of uncontrolled environments in different circumstances, allowing the identification of the environments for the improvement of the results of a Personal Digital Life Coach; Some people need assistance during the daily activities with the monitoring of the environment with acoustic data that is largely available around the world, allowing to prevent some possible problems that may occur with different people; Big data and machine learning techniques may be combined to support the different daily activities [4]. These data may be the start of developing a solution for the identification of environments that may be combined with other data around the world.

Data Description

Recently, various studies have been performed for the identification of the different acoustic environments using the data acquired from mobile devices, since complex algorithm or simple methods implemented locally in a mobile device [5], [6], [7]. The data intends to present acoustic data related to different environments for the future development of a Personal Digital Life Coach [8], [9], [10]. The acquisition of data with people with various lifestyles, and different locations allows the future generalization of the acquired data for the including in a reliable instrument for the identification of different environments [11]. This paper presents a dataset with the microphone data byte values related to nine environments, including bar, classroom, gym, kitchen, library, street, hall, living room, and bedroom. A BQ Aquaris 5.7 smartphone [1], placed in the several positions, including pants' front pocket, on a wristband, over the bedside table, on a table, or on other furniture, was used for the data acquisition. The dataset is distributed by five repositories composed, in total, of eleven main folders, i.e., one folder for each environment, and each folder contains more than 2000 numbered folders with the files related to the data acquired from the raw data of the microphone available in the off-the-shelf mobile device. Each subfolder contains one file named as "sound.txt". In total, the dataset contains around 18,000 files related to the acoustic data. The different files contain the values of the bit arrays acquired for 5 s, where each value of the byte arrays acquired during the defined time are presented in only one column with integer values. As an example, Fig. 1 presents an excerpt of the data acquired in the gym environment. The original file contains 348,160 rows, where Fig. 1 presents 10,000 rows.
Fig. 1

Example of microphone data related to gym environment are byte array.

Example of microphone data related to gym environment are byte array. The analysis of the raw data acquired from the microphone available in the off-the-shelf mobile device. For the analysis, 26 Mel-frequency cepstral coefficients (MFCC) [12] and statistical features are extracted from the data recorded in the different files. For the calculation of MFCC, the Fast Fourier Transform (FFT) was applied. Thus, the MFCC coefficients retrieves data related to the sound as a short-term power spectrum composed by a linear cosine transform of a log power spectrum on a nonlinear Mel-scale of frequency, which, as previously used in other studies, e.g., [7,13], it consists in the codification of audio data for the identification of the different frequencies. It allows the categorization of the different data collected, because it discretizes the data for the correct identification. Finally, the statistical features are: Standard deviation: The standard deviation of values was calculated; Average: The mean of values was calculated; Variance: The variance of values was calculated; Median: The median of values was calculated. Table 1 presents the average of the features of all samples of the microphone raw data related to each environment.
Table 1

Average of the features calculated for each environment with the acoustic data.

ParametersBarClassroomGymHallKitchenLibraryBedroomStreetLiving room
MFCC 11.4253.2021.3001.2103.9411.9400.7161.1391.631
MFCC 21.9044.3241.7071.6785.2552.5951.0101.5782.238
MFCC 31.6093.7681.3661.5844.3942.1951.0001.4862.047
MFCC 41.2273.0140.9371.4473.2341.6620.9851.3541.771
MFCC 50.8632.2380.5511.2882.0661.1310.9681.2021.459
MFCC 60.5861.5800.2851.1301.1200.7040.9511.0551.154
MFCC 70.4101.1010.1450.9840.5050.4190.9330.9290.885
MFCC 80.3090.7920.0880.8580.2090.2580.9150.8340.664
MFCC 90.2490.6060.0690.7520.1500.1770.8980.7680.488
MFCC 100.2060.4900.0590.6640.2290.1350.8790.7260.350
MFCC 110.1740.4090.0550.5900.3640.1070.8590.6990.242
MFCC 120.1560.3430.0570.5290.5010.0860.8380.6770.158
MFCC 130.1570.2870.0630.4790.6010.0710.8160.6540.090
MFCC 140.1720.2370.0630.4400.6320.0610.7930.6240.034
MFCC 150.1920.1930.0520.4090.5780.0550.7680.585−0.012
MFCC 160.2080.1530.0360.3840.4410.0490.7420.537−0.049
MFCC 170.2140.1170.0230.3630.2530.0450.7150.483−0.078
MFCC 180.2080.0870.0220.3440.0650.0410.6860.427−0.100
MFCC 190.1900.0660.0310.326−0.0630.0390.6570.376−0.115
MFCC 200.1610.0530.0410.307−0.0930.0380.6280.333−0.127
MFCC 210.1270.0490.0420.287−0.0250.0380.5990.304−0.138
MFCC 220.0900.0500.0340.2650.1080.0360.5710.290−0.150
MFCC 230.0530.0490.0250.2410.2500.0330.5430.286−0.161
MFCC 240.0190.0450.0190.2170.3490.0280.5150.289−0.170
MFCC 25−0.0110.0350.0210.1950.3800.0230.4850.292−0.173
MFCC 26−0.0340.0210.0260.1750.3520.0190.4550.291−0.171
Standard Deviation0.0370.0340.0380.0400.0170.0470.0050.0410.031
Average0.6320.5170.7010.5700.5050.4390.7720.5270.573
Variance0.0010.0010.0020.0020.0010.0030.0000.0020.001
Median0.6320.5170.7010.5700.5050.4390.7720.5270.573
Average of the features calculated for each environment with the acoustic data.

Experimental Design, Materials and Methods

The smartphone acquired samples with five seconds of acoustic data collected from the microphone of the smartphone. The data were collected with a smartphone in different places according to the different environments, as presented in Table 2.
Table 2

Position of the smartphone in different environments.

EnvironmentsPlacement
BarOver a table; Front pocket of the pants; Over other furniture
ClassroomOver a table; Over other furniture
GymOver a table; Front pocket of the pants; On a wristband; Over other furniture
HallFront pocket of the pants; On a wristband
KitchenOver a table; Over other furniture
LibraryOver a table; Over other furniture
BedroomOver the bedside table; Over a table; Over other furniture
StreetFront pocket of the pants; On a wristband
Living roomOver a table; Over other furniture
Position of the smartphone in different environments.

Participants

The experiments were performed by twenty-five randomly selected individuals aged between 16 and 60 years old. The participants' lifestyle is significant for the Bar, Gym, Hall, and Street environments, where the mobile device is positioned related to the user's body. Thus, from the different individuals, 10 participants are mainly active, and the remaining 15 participants are mainly sedentary. Age = 33.5200 ± 13.5250 years old

Procedure

After the smartphone placement on the furniture or in the user's garments, the microphone data was recorded with an Android application. Initially, the person selects the environment in the mobile application. Then, the user presses the button to start the data acquisition. After that, the user positions the mobile device adequately as presented in Fig. 2.
Fig. 2

Positioning of the smartphone during the data acquisition. (a) The smartphone is positioned in a waistband. (b) The smartphone is positioned over a table. (c) The smartphone is positioned in the pocket of the pants. Image partilly adapted from https://www.needpix.com/photo/1811318/man-business-man-business-person-people-tie-professional-grown-up-businessman.

Positioning of the smartphone during the data acquisition. (a) The smartphone is positioned in a waistband. (b) The smartphone is positioned over a table. (c) The smartphone is positioned in the pocket of the pants. Image partilly adapted from https://www.needpix.com/photo/1811318/man-business-man-business-person-people-tie-professional-grown-up-businessman. This dataset was acquired with an easy positioning of the mobile device for the data acquisition related to the different environments, such as bar, classroom, gym, kitchen, library, street, hall, living room, and bedroom. All people with different conditions can to understand the information for the data acquisition. These are: Install the mobile application on the mobile device; Open the mobile application designed for the acquisition of the sensors’ data; The user selects the environment where the mobile device is in; Press the button to start the data acquisition; The data acquisition starts after 10 s; The user positions the mobile device adequately; The data acquisition is performed during slots of 5 s; The data acquisition stops for 5 min; The flow returns to point 7, and it repeats continuously until the user press the button to stop data acquisition.

Ethics Statement

The participants signed an ethical agreement to allow us to share the results of the tests in an anonymous form. The agreement also provided the participants’ informed consent considering the risks and the objective of the study. Only the data related to the individuals that sign the consent to participate in the study were recorded. The participants were also informed that about the inclusion of the data anonymously in Mendeley Data. Ethics Committee from Universidade da Beira Interior approved the study with the number CE-UBI-Pj-2020–035.

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships which have, or could be perceived to have, influenced the work reported in this article.
SubjectAcoustics and UltrasonicsElectrical and Electronic EngineeringBiomedical EngineeringHealth
Specific subject areaEnvironmentsMicrophoneMobile Devices
Type of dataTable
How data were acquiredThe data was acquired from the microphone available in a BQ Aquaris 5.7 smartphone [1] with a mobile application. The mobile device has a Quad Core CPU and 16 GB of internal memory. Different places were used to position the mobile device during the data acquisition, including the front pocket of the pants, a wristband, a bedside table, a table, or other furniture. The mobile device automatically acquires the acoustic data related to the different environments, and the user selects the environment where the mobile device is placed. There are no specific specifications needed for the acquisition of the data. Still, it was acquired at the same time as the inertial sensors data for the activities presented in [2].
Data formatRaw text files
Parameters for data collectionDepending on the environments, the mobile device was placed in different places according to the environments' restrictions. Initially, the individual selects the environment in the mobile application to label the various records. The protocol of using the mobile application and its actions were explained to the participants before starting the data acquisition.
Description of data collectionAfter selecting the environment in the user interface of the mobile application, the user places the mobile device in a position that she/he chooses, including the front pocket of the pants, a wristband, a bedside table, a table, or other furniture. The microphone data is collected as a byte array and stored in text files for further analysis during the data acquisition. The microphone acquires the data with a sample rate of 44,100 Hz in a mono channel as an array of 16-bit unsigned integer values in the range [0, 255] with a 128 offset for zero.
Data source locationPrimary data sources:City/Town/Region: CovilhãCountry: PortugalLatitude and longitude (and GPS coordinates, if possible) for collected samples/data: 40° 16′ 50.037′′ N 7° 30′ 15.555′′ WCity/Town/Region: FundãoCountry: PortugalLatitude and longitude (and GPS coordinates, if possible) for collected samples/data: 40° 7′ 30.129′′ N 7° 30′ 39.966′′ W
Data accessibilityRepository name: Raw dataset with acoustic data for environments - Part 1Data identification number: 10.17632/yjp2cdyrnj.2Direct URL to data: http://dx.doi.org/10.17632/yjp2cdyrnj.2Repository name: Raw dataset with acoustic data for environments - Part 2Data identification number: 10.17632/yczyfrx2rp.2Direct URL to data: http://dx.doi.org/10.17632/yczyfrx2rp.2Repository name: Raw dataset with acoustic data for environments - Part 3Data identification number: 10.17632/bmgn76p7b6.2
Direct URL to data: http://dx.doi.org/10.17632/bmgn76p7b6.2Repository name: Raw dataset with acoustic data for environments - Part 4Data identification number: 10.17632/vtx583krxt.2Direct URL to data: http://dx.doi.org/10.17632/vtx583krxt.2Repository name: Raw dataset with acoustic data for environments - Part 5Data identification number: 10.17632/rk7zg7v7s5.2Direct URL to data: http://dx.doi.org/10.17632/rk7zg7v7s5.2
Related research articleI.M. Pires, G. Marques, N.M. Garcia, N. Pombo, F. Flórez-Revuelta, S. Spinsante, M.C. Teixeira, E. Zdravevski, Recognition of Activities of Daily Living and Environments Using Acoustic Sensors Embedded on Mobile Devices. Electronics 2019, 8, 1499https://doi.org/10.3390/electronics8121499
  2 in total

1.  Mobile Computing Technologies for Health and Mobility Assessment: Research Design and Results of the Timed Up and Go Test in Older Adults.

Authors:  Vasco Ponciano; Ivan Miguel Pires; Fernando Reinaldo Ribeiro; María Vanessa Villasana; Rute Crisóstomo; Maria Canavarro Teixeira; Eftim Zdravevski
Journal:  Sensors (Basel)       Date:  2020-06-19       Impact factor: 3.576

2.  Activities of daily living with motion: A dataset with accelerometer, magnetometer and gyroscope data from mobile devices.

Authors:  Ivan Miguel Pires; Nuno M Garcia; Eftim Zdravevski; Petre Lameski
Journal:  Data Brief       Date:  2020-12-08
  2 in total
  1 in total

1.  Daily motionless activities: A dataset with accelerometer, magnetometer, gyroscope, environment, and GPS data.

Authors:  Ivan Miguel Pires; Nuno M Garcia; Eftim Zdravevski; Petre Lameski
Journal:  Sci Data       Date:  2022-03-25       Impact factor: 6.444

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.