Skip to main content

Rei - Blog Post 4 & 5

Due to external factors, I accidentally missed the last Blog Post, however I shall make up for it by writing two posts in one! The last couple of weeks have been at least a little productive.

Blog Post 4

Over these weeks, Matthew and I were both working on solidifying and exploring the context portion of our project. Of course, we have been exploring the context since nearly the beginning of the year, but because of both capstone classes have asked us to write something about the context of our paper.

The context of our project can be found throughout our previous blog post, but a short and summarized version of out context section boils down to a couple main points.

  1.  Neural Networks are not too present in video games.
  2. When neural nets are present, they are used for research or marketing.

As for the rest of our Design Document, unfortunately we rushed it a bit. We got caught up in some upsetting events these weeks and allowed our work to suffer instead of staying on task. However, we did eventually get ourselves together and write something. We realize that something may have not been our best work, but it was work none the less.

Matthew was kind enough to setup a repository for our project, this is where all of our public facing code and documentation will live on the internet. This will not include any of our training data, except for the training data that me and Matthew Supply. That repository can be found here.

Blog Post 5

Over these weeks, we were reeling from the busyness of Thanksgiving and Black Friday, but we did get some important work done. Matthew made a reddit post soliciting replay data from the Rivals of Aether Community. We very quickly got many user submissions from the community, these replays are considered private data so we will not be sharing them with the public.

Now that we have data though, we need to label it! Luckily the first few lines of the replay file contain how we want to label it. The labels we want will most likely be:

  1. The character we are playing as.
  2. The map we are playing on.
  3. If the game is vs Human or vs CPU
  4. Who provided the data. (If we want to train it off of one person.)
All of this is, again, incredibly easy to extract from the replay file. After we label this data, we should also figure out how to parse the data into data we can use, there are community tools in the works that we can use as reference data, but creating our own will most likely be the course we need to take.

During this blog posts frame, we also received feedback for our design document. We did make some dumb mistakes, like not compiling out LaTeX files into a pdf, which we will be sure not to do again. We also know that our literature review itself was not as good as it should have been. The goal for our final few weeks of this semester is to reevaluate our literature review and move on to creating a better problem statement. The reevaluation will mostly consist of really solidifying where our sources fit into our project. In all honesty, most of our research was just us building a body of knowledge for A.I., but we should have stated this more clearly.

As for the problem statement, Matthew and I will need to really work on nailing down a better wording. We understand what we wish our project to be, we just failed to explain it properly.

Comments

Popular posts from this blog

Rei - Blog Post 10

So,  I missed blog post 9. This is me acknowledging that for consistency. Anyway, the past couple of weeks have been incredibly productive for ContentsMayBeHot. Matthew has finished collecting all the replay data, we have refactored our project to reduce complexity, we have improved the runtime of our code, and finally we have started seriously training our model. The Changes Matthew implemented multi-threading for the model loading. Which reduced our load time between files from about 3-5 Seconds to 1 Second or less. Which allows us to fully train a model in much less time! While Matthew did this I reduced the code duplication in our project. This way, if we needed to change how we loaded our training data, we didn't have to change it in multiple places. This just allows us to make hot-fixes much more efficiently. We also started working on some unittests for our project using pytest. These tests were written because of a requirement for another class, but we thought it

Matthew - Blog Post 10

At the time of my last blog post, we were managing quite a few problems. Our model was essentially vaporware, our training and testing was hindered by slow, blocking function calls from our loader, and our VRAM was continually getting exhausted during training sessions. But there is nothing to worry about. We have made major strides since then. Major strides. Model improvements First, we have completely overhauled our model's architecture. We are now using a model composed of special layers that combine the functionality of a 2D convolutional neural network with that of an LSTM. Here is a summary of  our model as printed by Keras: This model was made with the help of the wonderful community over on Stack Overflow . I would also like to mention that Professor Auerbach made invaluable contributions. In general, his tutelage made this project possible. We dropped our Sequence subclass, and replaced it with training and testing loops. In these loops, we iterate over the whol

Matthew - Blog Post 8

replaymanager.py is only 339 lines long but it feels at least three times that when I'm working on it. The module is definitely due for a refactor. For one, the term subdataset should be renamed to version_set and extracted into a class. version_set more accurately and describes what it is, and the class extraction would clean up the namespace in ReplayManager. There is probably some kind of class extraction possible for replays, so that their names, paths, file streams, and Twitter profiles can all be neatly encapsulated, thereby cleaning up the namespace even more. However, I do not want to worry about having two kinds of replays: the one used by ReplayManager and the one used by ReplayParser. Even though ReplayManager does not use ReplayParser, the prospect of making things more muddier deters me. There's probably a right way to refactor this code, but, to put it simply, now is not the time. Speaking of time, I came up with a great way to get work done, even when I am slee