Skip to content

dasari-mohana-zz/Human_Activity_Recognition_project

Repository files navigation

Human_Activity_Recognition_project

Abstract:

Human Activity Recognition database built from the recordings of 30 subjects performing activities of daily living (ADL) while carrying a waist-mounted smartphone with embedded inertial sensors.

Data Set Information:

The experiments have been carried out with a group of 30 volunteers within an age bracket of 19-48 years. Each person performed six activities (WALKING, WALKING_UPSTAIRS, WALKING_DOWNSTAIRS, SITTING, STANDING, LAYING) wearing a smartphone (Samsung Galaxy S II) on the waist. Using its embedded accelerometer and gyroscope, we captured 3-axial linear acceleration and 3-axial angular velocity at a constant rate of 50Hz. The experiments have been video-recorded to label the data manually. The obtained dataset has been randomly partitioned into two sets, where 70% of the volunteers was selected for generating the training data and 30% the test data.

The sensor signals (accelerometer and gyroscope) were pre-processed by applying noise filters and then sampled in fixed-width sliding windows of 2.56 sec and 50% overlap (128 readings/window). The sensor acceleration signal, which has gravitational and body motion components, was separated using a Butterworth low-pass filter into body acceleration and gravity. The gravitational force is assumed to have only low frequency components, therefore a filter with 0.3 Hz cutoff frequency was used. From each window, a vector of features was obtained by calculating variables from the time and frequency domain.

Attribute Information:

For each record in the dataset it is provided:

  • Triaxial acceleration from the accelerometer (total acceleration) and the estimated body acceleration.
  • Triaxial Angular velocity from the gyroscope.
  • A 561-feature vector with time and frequency domain variables.
  • Its activity label.
  • An identifier of the subject who carried out the experiment.

Feature names

  1. These sensor signals are preprocessed by applying noise filters and then sampled in fixed-width windows(sliding windows) of 2.56 seconds each with 50% overlap. ie., each window has 128 readings.

  2. From Each window, a feature vector was obtianed by calculating variables from the time and frequency domain. In our dataset, each datapoint represents a window with different readings.

  3. The accelertion signal was saperated into Body and Gravity acceleration signals(tBodyAcc-XYZ and tGravityAcc-XYZ) using some low pass filter with corner frequecy of 0.3Hz.

  4. After that, the body linear acceleration and angular velocity were derived in time to obtian jerk signals (tBodyAccJerk-XYZ and tBodyGyroJerk-XYZ).

  5. The magnitude of these 3-dimensional signals were calculated using the Euclidian norm. This magnitudes are represented as features with names like tBodyAccMag, tGravityAccMag, tBodyAccJerkMag, tBodyGyroMag and tBodyGyroJerkMag.

  6. Finally, We've got frequency domain signals from some of the available signals by applying a FFT (Fast Fourier Transform). These signals obtained were labeled with prefix 'f' just like original signals with prefix 't'. These signals are labeled as fBodyAcc-XYZ, fBodyGyroMag etc.,.

  7. These are the signals that we got so far.

    • tBodyAcc-XYZ
    • tGravityAcc-XYZ
    • tBodyAccJerk-XYZ
    • tBodyGyro-XYZ
    • tBodyGyroJerk-XYZ
    • tBodyAccMag
    • tGravityAccMag
    • tBodyAccJerkMag
    • tBodyGyroMag
    • tBodyGyroJerkMag
    • fBodyAcc-XYZ
    • fBodyAccJerk-XYZ
    • fBodyGyro-XYZ
    • fBodyAccMag
    • fBodyAccJerkMag
    • fBodyGyroMag
    • fBodyGyroJerkMag
  8. We can esitmate some set of variables from the above signals. ie., We will estimate the following properties on each and every signal that we recoreded so far.

    • mean(): Mean value
    • std(): Standard deviation
    • mad(): Median absolute deviation
    • max(): Largest value in array
    • min(): Smallest value in array
    • sma(): Signal magnitude area
    • energy(): Energy measure. Sum of the squares divided by the number of values.
    • iqr(): Interquartile range
    • entropy(): Signal entropy
    • arCoeff(): Autorregresion coefficients with Burg order equal to 4
    • correlation(): correlation coefficient between two signals
    • maxInds(): index of the frequency component with largest magnitude
    • meanFreq(): Weighted average of the frequency components to obtain a mean frequency
    • skewness(): skewness of the frequency domain signal
    • kurtosis(): kurtosis of the frequency domain signal
    • bandsEnergy(): Energy of a frequency interval within the 64 bins of the FFT of each window.
    • angle(): Angle between to vectors.
  9. We can obtain some other vectors by taking the average of signals in a single window sample. These are used on the angle() variable.

    • gravityMean
    • tBodyAccMean
    • tBodyAccJerkMean
    • tBodyGyroMean
    • tBodyGyroJerkMean

Y_Labels(Encoded)

  • In the dataset, Y_labels are represented as numbers from 1 to 6 as their identifiers.

    • WALKING as 1
    • WALKING_UPSTAIRS as 2
    • WALKING_DOWNSTAIRS as 3
    • SITTING as 4
    • STANDING as 5
    • LAYING as 6

Train and test data

  • The readings from 70% of the volunteers were taken as trianing data and remaining 30% subjects recordings were taken for test data

Data

  • All the data is present in 'UCI_HAR_dataset/' folder in present working directory.
    • Feature names are present in 'UCI_HAR_dataset/features.txt'

    • Train Data

      • 'UCI_HAR_dataset/train/X_train.txt'
      • 'UCI_HAR_dataset/train/subject_train.txt'
      • 'UCI_HAR_dataset/train/y_train.txt'
    • Test Data

      • 'UCI_HAR_dataset/test/X_test.txt'
      • 'UCI_HAR_dataset/test/subject_test.txt'
      • 'UCI_HAR_dataset/test/y_test.txt'

Files

  1. HumanActivity_EDA.ipynb --> Exploratory data Analysis of the data and preprocessing the data
  2. Machine_Learning_Predictions_Model.ipynb --> Built a Classic Machine Learning Model with different type of classification algorithms on feature enginnered data
  3. DeepNN_LSTM_model.ipynb --> Built a Deep Learning Sequential Model with single layer of LSTM and with dropouts on raw data.

Performance

  1. Machine Learning Model --> Of all classifiers Logistic Regression, Linear SVC and Rbf Kernal SVM perform well compare to tree type classifiers with above 95% accuracy (with featured data)
  2. Deep learing model --> With a simple 2 layer architecture we got 90.09% accuracy and a loss of 0.30 (without featured data)

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published