DIGITAL LIFE
The AI that taught itself: How AI can learn what it never knew
For years, the guiding assumption of artificial intelligence has been simple: an AI is only as good as the data it has seen. Feed it more, train it longer, and it performs better. Feed it less, and it stumbles. A new study from the USC Viterbi School of Engineering was accepted at the IEEE SoutheastCon 2026, taking place March 12–15. It suggests something far more surprising: with the right method in place, an AI model can dramatically improve its performance in territory it was barely trained on, pushing well past what its training data alone would ever allow.
The method was developed by Minda Li, a USC Viterbi undergraduate who has been pursuing research since her freshman year, working alongside her advisor Bhaskar Krishnamachari, a Faculty Fellow and Systems Professor in the Ming Hsieh Department of Electrical and Computer Engineering, with a joint appointment in the Thomas Lord Department of Computer Science at the USC Viterbi School of Engineering and the USC School of Advanced Computing. Their paper is available on the arXiv preprint server.
Together, they tested GPT-5's ability to write code in Idris, an extraordinarily obscure programming language with a fraction of the online presence of mainstream languages like Python. The results were striking: by giving the AI feedback on its errors and letting it try again, Li pushed the model's success rate from a dismal 39% all the way to 96%.
"Our AI tools are now able to transcend their initial training. Used to be, maybe a year or two ago, you would say an AI model is only as good as the data it has seen. This paper is saying something different." said Prof. Bhaskar Krishnamachari
A language so obscure, even the researchers didn't know it...Python, the world's most popular programming language, has over 24 million code repositories publicly available online, a vast library that AI models like GPT-5 learn from during training. Idris, the language Li and Krishnamachari chose to test, has approximately 2,000. That is roughly 10,000 times less data.
The choice of Idris was deliberate, and, as Krishnamachari describes it, a little playful. "We were hunting for a language so obscure that we hadn't heard of it," he said. "I think we were just in my office together, googling around, trying to find some crazy language that no one's ever heard of." They found Idris, a dependently typed functional programming language used by a small community of specialists, and decided it was the perfect test case.
Crucially, neither researcher could write a line of it themselves. "Neither Minda nor I had ever coded in it, and frankly, we could not tell you if the code was correct or wrong," Krishnamachari admitted. That is part of what makes the findings so striking: Li was guiding an AI to master a language that its own guides could not speak.
The breakthrough: A feedback loop that changes everything...Li started by simply asking GPT-5 to solve 56 Idris coding exercises on Exercism, a popular coding practice platform. Out of the box, the model solved only 22 of them, a 39% success rate, far below its 90% success rate in Python and 74% in Erlang.
She then tried several approaches to improve performance: providing documentation, error manuals, and reference guides. These helped somewhat, pushing the success rate to the low 60s, but never dramatically.
The breakthrough came when she implemented what they call a compiler feedback loop. A compiler is the software that translates human-written code into instructions a computer can execute. When code is wrong, the compiler says so, precisely and in technical detail. She began capturing those error messages and feeding them directly back to GPT-5, asking it to fix the specific problems identified and try again. Up to 20 times per problem.
"I thought we'd probably get a 10% jump," said Li, who designed and ran the experiments. "I was surprised that just that alone, seemingly one simple thing, just keep recompiling, keep trying, was able to get to 96%."
Beyond code: Why this changes everything...What Li and Krishnamachari built is essentially a method for unlocking capability that was always there but inaccessible. By engineering the right kind of structured feedback, they found a way to get far more out of an AI model than its training data alone would ever produce.
Krishnamachari envisions this approach being applied far outside the world of software and niche programming languages. "Imagine you're trying to get an AI to build 3D models of buildings," he said. "You have something that gives feedback: this model is structurally unsafe, it doesn't have the right distribution of materials, it's too expensive to build. Whatever it is, it just gives feedback on everything the AI generates, every iteration.
No comments:
Post a Comment