If you are not familiar with Microsoft CoCos, you should be. Its a treasure trove of data for your learning pleasure! There just happens to be one pesky problem with it, and that is the fact that when attempting to find the files for training/testing; the Annotation file that ships with MS CoCo does not include the actual file name, but rather the image id. This sounds fine, except the data when you download it has a bunch of trailing stuff! In this article we will go through how to get it ready.
In this article I’m going to go through how to set up CNTK with Visual Studio Code and take advantage of those PASCAL GPUs I know everybody has these days. I will also do a breif overview of what CNTK and Visual Studio Code are and why they are so incredible for machine learning scientists.
So today we are going to do something really awesome. Operationalize Keras with Azure Machine Learning. Why in the world would we want to do this? Well we can configure Deep Neural Nets and train them on GPU. In fact, in this article, we will train a 2 depth neural network which outputs a linear prediction of energy efficiency of buildings and then operationalize that GPU trained network on Azure Machine Learning for production API usage.
So today I have decided to actually begin a write up on where I’m headed. Its taken a long while to come up with what my next adventure after food trucks was going to be, especially with all of the things you can do with embedded technologies and machine learning. So here it is, a self driving race car.
In this article I aim to lay out the high level plan of attack for what my team and I are building and some of the direction on where we are headed.
So I just completed an incredible project with Brain Thermal Tunnel Genix, where I learned so much about pattern recognition, machine learning and taking research and algorithms and pushing those into a production environment where it can be integrated into a real product. Today’s article takes those lessons and provides a sample on how to perform complex modelling and operationalize it in the cloud. The accompanying Gallery Example can be found here.
So there are a ton of articles out there on the theory of Reinforcement Learning, but very few with an actual application. I watched a few lectures from Berkley, and read a few articles by NVidia and thought, “Well, lets just give this a shot”. 8 hours later, this is what I had.
Herby V1 simply learns to go forward as much as possible while avoiding obstacles.
So here is a pretty raw blog article; not unlike most of my articles. The cognitive revolution. I’m going to coin this term today. What the heck is this thing? What does it mean for you? What does it mean for me? Where did it come from? What is it? These are questions I aim to answer in this blog article.
So this is an interesting problem. You are collecting data from somewhere and you want to feed it into a neural network for classification. There is one main problem with this. The shape of the data! Neural networks and really just anything require specifically shaped data, you can’t just like give it something of ambiguous size. There are tons of papers out there on dimensionality reduction, but nothing on dimensionality reduction to a specified size. This article explains my approach.
So today, I was asked to put some thought into what we should focus our entry level data scientists on in terms of tech skills. After I put a bunch of thought into it, I ended up coming up with this. I decided that the most important aspect of this was a few items fold
Don’t overload them
Can deliver to production where the target can be anything, including IoT.
They will not be concerned with building front ends.