In June, the developers of a Russian chatbot posing as a 13-year-old boy from Ukraine claimed it had passed the Turing test. While a lot of people doubt the result's validity because the testers used a sketchy methodology and the event was organized by a man fond of making wild claims, it's clear we need a better way to determine if an AI possesses human levels of intelligence. Enter Lovelace 2.0, a test proposed by Georgia Tech associate professor Mark Riedl.
Here's how Lovelace 2.0 works:
For the test, the artificial agent passes if it develops a creative artifact from a subset of artistic genres deemed to require human-level intelligence and the artifact meets certain creative constraints given by a human evaluator. Further, the human evaluator must determine that the object is a valid representative of the creative subset and that it meets the criteria. The created artifact needs only meet these criteria but does not need to have any aesthetic value. Finally, a human referee must determine that the combination of the subset and criteria is not an impossible standard.