Difference between revisions of "User:Cnapun"

From REU@MU
Jump to: navigation, search
(Week 4 (26 Jun - 30 Jun))
(Forecasting with Deep Learning)
Line 46: Line 46:
 
=== Forecasting with Deep Learning ===
 
=== Forecasting with Deep Learning ===
 
* [https://arxiv.org/abs/1610.09460 "Building Energy Load Forecasting using Deep Neural Networks"]
 
* [https://arxiv.org/abs/1610.09460 "Building Energy Load Forecasting using Deep Neural Networks"]
** Discusses the direct application of LSTMs to load forecasting
+
** Discusses the direct application of seq2seq LSTMs to load forecasting
 
* [http://ieeexplore.ieee.org/document/6796853/ "Training Recurrent Networks by Evolino"]
 
* [http://ieeexplore.ieee.org/document/6796853/ "Training Recurrent Networks by Evolino"]
 
** Describes the use of genetic algorithms to train RNNs
 
** Describes the use of genetic algorithms to train RNNs
 
** Should offer better performance than Echo State Networks (ESNs)
 
** Should offer better performance than Echo State Networks (ESNs)
 
** No description of how crossover and mutate operations work; I referred to [http://davidmontana.net/papers/hybrid.pdf "Neural Network Weight Selection Using Genetic Algorithms"] for a more complete explanation
 
** No description of how crossover and mutate operations work; I referred to [http://davidmontana.net/papers/hybrid.pdf "Neural Network Weight Selection Using Genetic Algorithms"] for a more complete explanation
 +
 
=== Hybrid Methods ===
 
=== Hybrid Methods ===
 
* [https://openreview.net/pdf?id=ByD6xlrFe "Hybrid Neural Networks Over Time Series For Trend Forecasting"]
 
* [https://openreview.net/pdf?id=ByD6xlrFe "Hybrid Neural Networks Over Time Series For Trend Forecasting"]

Revision as of 15:11, 1 July 2017

Personal Information

  • Case Western Reserve University Class of 2019
  • Applied Mathematics and Computer Science Major

Weekly Log

Week 0 (30 May - 2 Jun)

  • Met Dr. Povinelli 4 times: to discuss possible project topics, to get oriented with the lab, to decide on a project topic, and to establish weekly milestones
  • Obtained MU and MSCS account logins and ID card
  • Read most of "Learning Deep Architectures for AI"
  • Read section on Sequence Modeling (Ch 10) of The Deep Learning book
  • Read various other papers

Week 1 (5 Jun - 9 Jun)

  • Attended GasDay camp and learned about what GasDay does
  • Attended responsible conduct of research training
  • Continued to read papers
  • Decided to use TensorFlow and Keras for now
  • Got Anaconda, TensorFlow, and Keras installed on a couple lab computers
  • Learned how to access customer data

Week 2 (12 Jun - 16 Jun)

  • Started playing with TensorFlow, mostly using the GEFCom2014-E dataset to allow me to continue working when not at the lab
  • Things I got working in Tensorflow:
    • n-layer Sequence to sequence (seq2seq) model (encoder-deocder architecture)
    • Autoregressive seq2seq model (using slower ops)
    • Multilayer Perceptron
    • 1D ConvNet on weather inputs, in parallel with MLP to process load and time inputs
  • Will try ConvNet on all 3 inputs, feeding into RNN, with the seed state being the month, day, and day of week at the initial timestamp
  • Will try Clockwork RNN

Week 3 (19 Jun - 23 Jun)

  • Reinstalled CUDA
  • Started training seq2seq models on GasDay (electric) data
  • Got TensorBoard to work
  • Trained AR LSTM on GasDay gas data

Week 4 (26 Jun - 30 Jun)

  • Continued to experiment with hyperparameters
  • Achieved suspiciously good performance on both gas and electricity datasets
  • Prepared a presentation of my research for other REU students (slides)
  • Attempted to train a model with objective that was not at all convex; it failed miserably
  • Still can't figure out why Dropout makes everything worse

Reading

Here are some of the papers I have read, skimmed, and partially read:

Forecasting with Deep Learning

Hybrid Methods

Review Papers

LSTMs, Training, and Possible Improvements