Skip to main content

Posts

Showing posts from June, 2019

Summer - Week 10

This week, I am working to get the dynamic time warping functionality into my program. The process of doing so includes re-processing the features to include the time series, putting each series back together when we construct sequences, and then performing the DTW to generate a number that will be used to compute the kNN of each sequence which can then be used for predictions with the models. The processing time of these activities has gone up significantly since we have been using five different metrics with each of the F phase datasets. I am returning to school next week, and once I've completed the DTW processing all that will remain before we put together our second paper (The date for the reach journal we would like to submit it to is October 1), I am hoping I will have time to look again into the Agglomerative Hierarchical Clustering concept, which I did not successfully complete when we explored it earlier in the summer and then changed focus to the paper. We heard back

Summer - Week 3

This week, I performed experiments with the first and second phase data together and varied different factors. These were various model types from sklearn, length of the sequence (How many periods precede the one that we are attempting to predict), number of nearby sequences to use and subgroupings (gender, phase, injury type, experiment group). The primary aspect of this that we didn't get to experiment with when we were finishing our final report was testing out how different models worked. My model for the report was a from-scratch k-Nearest-Neighbor Regressor, and I took an equal-weighted approach for calculating my final predictions, using all of the similar active/inactive minute values and taking the average. Using the sklearn's kNN model takes into account the distance of each series from the series that we are attempting to predict and weight closer sequences higher than farther ones. Other models that we intend to test out include a linear SVM, decision trees (Alexa

Summer - Week 2

I spent most of this week cleaning up my code, organizing our files and reviewing Alexa's code. I am mostly done with this and I am looking forward to working on our first objective, wrapping up the topics that we were working on at the end of the year.

Summer - Week 1

Alexa is not going to continue working on the project throughout the summer, but Gina and I intend to branch out based on what we've done so far. In order to do so, we have set a few objectives for ourselves: 1. Tie up some loose ends from Alexa and I's initial project 2. Complete 1-2 papers, since we were not able to get our research during the year to that point         3. Push the "Nearest Sequences" concept farther by associating data from farther back with a given night or day period we are predicting - as far back as we can go without having to use baseline days in the experiment while not losing any data. 3. Explore the built-in sleep detection methods that are automatically classified by the Actigraph's software, and how they compare to Dr. Skornyakov's sleep/wake algorithm and the inactive minutes method we've used previously. One of the variables we addressed when deciding to proceed with the "k-nearest sequences" approach was the