This is according to Rachel Greenstadt, an associate professor of computer science at Drexel University, who together with Aylin Caliskan who was her former PhD student and is now an assistant professor at George Washington University, who presented their findings at the DefCon hacking experience.
How the AI works is by being fed examples of a programmer’s work where it studies the coding structure. From there, it will then be able to train itself to be capable of spotting that programmer’s work in the future. Based on the testing that they did using Google’s Code Jam, their AI appeared to be relatively adept as it was capable of identifying the programmers 83% of the time.
So why is this useful? It could be useful in investigating instances of hacks, or identifying who created certain pieces of malware, which generally tend to be anonymous. It could also be used in legal situations where a developer might accuse another of copying their code. However the downside is privacy, where there might be some instances in which programmers might choose to remain anonymous for certain reasons, and being able to be identified is not necessarily a good thing.