Skip to main content

Rei - Blog Post 4 & 5

Due to external factors, I accidentally missed the last Blog Post, however I shall make up for it by writing two posts in one! The last couple of weeks have been at least a little productive.

Blog Post 4

Over these weeks, Matthew and I were both working on solidifying and exploring the context portion of our project. Of course, we have been exploring the context since nearly the beginning of the year, but because of both capstone classes have asked us to write something about the context of our paper.

The context of our project can be found throughout our previous blog post, but a short and summarized version of out context section boils down to a couple main points.

  1.  Neural Networks are not too present in video games.
  2. When neural nets are present, they are used for research or marketing.

As for the rest of our Design Document, unfortunately we rushed it a bit. We got caught up in some upsetting events these weeks and allowed our work to suffer instead of staying on task. However, we did eventually get ourselves together and write something. We realize that something may have not been our best work, but it was work none the less.

Matthew was kind enough to setup a repository for our project, this is where all of our public facing code and documentation will live on the internet. This will not include any of our training data, except for the training data that me and Matthew Supply. That repository can be found here.

Blog Post 5

Over these weeks, we were reeling from the busyness of Thanksgiving and Black Friday, but we did get some important work done. Matthew made a reddit post soliciting replay data from the Rivals of Aether Community. We very quickly got many user submissions from the community, these replays are considered private data so we will not be sharing them with the public.

Now that we have data though, we need to label it! Luckily the first few lines of the replay file contain how we want to label it. The labels we want will most likely be:

  1. The character we are playing as.
  2. The map we are playing on.
  3. If the game is vs Human or vs CPU
  4. Who provided the data. (If we want to train it off of one person.)
All of this is, again, incredibly easy to extract from the replay file. After we label this data, we should also figure out how to parse the data into data we can use, there are community tools in the works that we can use as reference data, but creating our own will most likely be the course we need to take.

During this blog posts frame, we also received feedback for our design document. We did make some dumb mistakes, like not compiling out LaTeX files into a pdf, which we will be sure not to do again. We also know that our literature review itself was not as good as it should have been. The goal for our final few weeks of this semester is to reevaluate our literature review and move on to creating a better problem statement. The reevaluation will mostly consist of really solidifying where our sources fit into our project. In all honesty, most of our research was just us building a body of knowledge for A.I., but we should have stated this more clearly.

As for the problem statement, Matthew and I will need to really work on nailing down a better wording. We understand what we wish our project to be, we just failed to explain it properly.

Comments

Popular posts from this blog

Rei - Blog Post 10

So,  I missed blog post 9. This is me acknowledging that for consistency. Anyway, the past couple of weeks have been incredibly productive for ContentsMayBeHot. Matthew has finished collecting all the replay data, we have refactored our project to reduce complexity, we have improved the runtime of our code, and finally we have started seriously training our model. The Changes Matthew implemented multi-threading for the model loading. Which reduced our load time between files from about 3-5 Seconds to 1 Second or less. Which allows us to fully train a model in much less time! While Matthew did this I reduced the code duplication in our project. This way, if we needed to change how we loaded our training data, we didn't have to change it in multiple places. This just allows us to make hot-fixes much more efficiently. We also started working on some unittests for our project using pytest. These tests were written because of a requirement for another class, but we thought it...

Matthew - Blog Post 7

Since January, we've been working hard to not only finish writing the Replay Parser and Frame Collector but also totally synchronize them. I'm pleased to report our success. This is an amazing milestone for us because it means that we've surmounted one of our most troubling obstacles. I have also made sure to keep our documentation up to date. So, if you like, you can follow along with this blog post by replicating its results. The Frame Collector uses timed input sequences to start each replay associated with the currently running game version. Then, after waiting a set amount of time for playback to begin, it starts grabbing 1/4-scale frames at a rate of 10 frames per second. The Frame Collector takes these down-scaled frames, which are NumPy arrays, and rapidly pickles and dumps them into the file system. Here's a screenshot of the Frame Collector in action: If you look at the image above, you'll see that each pickle (the .np files) ...

Matthew - Blog Post 10

At the time of my last blog post, we were managing quite a few problems. Our model was essentially vaporware, our training and testing was hindered by slow, blocking function calls from our loader, and our VRAM was continually getting exhausted during training sessions. But there is nothing to worry about. We have made major strides since then. Major strides. Model improvements First, we have completely overhauled our model's architecture. We are now using a model composed of special layers that combine the functionality of a 2D convolutional neural network with that of an LSTM. Here is a summary of  our model as printed by Keras: This model was made with the help of the wonderful community over on Stack Overflow . I would also like to mention that Professor Auerbach made invaluable contributions. In general, his tutelage made this project possible. We dropped our Sequence subclass, and replaced it with training and testing loops. In these loops, we iterate over the whol...