DeepMind Builds AI That Codes as Well as the Average Human Programmer

This site may earn affiliate commissions from the links on this page. Terms of use.

(Credit: Google)
Google’s DeepMind AI division has already built machines that can wreck you in StarCraft II and predict millions of protein structures, but now it’s taking on an even harder task: writing coherent code. In an apparent effort to put themselves out of a job, DeepMind researchers have created an AI that can write code, and it performs surprisingly well in programming challenges. It’s neither the best nor worst when compared with humans participating in the same challenges, but the StarCraft AI was only a middling player at first.

While machine learning has advanced by leaps and bounds, it’s hard to create an AI that’s good at more than one thing. So, a machine could be trained with data to handle one class of programming challenges, but it would fail when given a different problem to tackle. So, the team decided to skip all the training on algorithms and code structure, instead treating it more like a translation problem.

Programming challenges usually include a description of the task, and the resulting code submitted by a human participant is technically just an expression of the description. The AI works in two phases: It takes the description and converts it to an internal representation. Then, it uses that representation to generate functional code based on the data it was shown in training. And there was a lot of data.

DeepMind fed the AI 700GB of code on GitHub, complete with the comments that explain the code. As Ars Technica points out, that’s a huge amount of text data. With the essence of programming internalized, DeepMind set up its own programming contests and fed the results into the AI to fine-tune the model’s performance. The team says this is an order of magnitude more training data than past coding machines have gotten, and that made all the difference.

The researchers found that AlphaCode was able to generate a huge number of potential answers to a coding problem, but about 40 percent of them would run through all the available system memory or fail to reach the answer in a reasonable amount of time. The data needs to be filtered to find the one percent of solutions that are actually good code. DeepMind found that clusters of similar code indicated better answers, whereas the wrong ones were randomly distributed. By focusing on those answers, AlphaCode was able to correctly answer about one-third of coding challenges. It turns out a lot of human programmers are little better, so AlphaCode placed in the top 54 percent. It’s not about to take jobs from DeepMind engineers, but give it time.

Now read:

Leave a Reply

Next Post

The first photo of Samsung's smart wireless charger spotted

Robert Triggs / Android Authority TL;DR Samsung has filed an FCC application for its new smart wireless charger. The application includes a photo that gives us our first look at the device. The documents say it has Zigbee and WPT capabilities with Bluetooth LE and Wi-Fi connectivity. Samsung has been […]
The first photo of Samsung’s smart wireless charger spotted

You May Like