Knowledge Science - Alles über KI, ML und NLP

AI Generated (E): KS Pulse - Scaling Smarter The Art of Deliberate Practice in AI

Sigurd Schacht, Carsten Lanquillon

Send us a text

English Version - The German Version also exists, but the content differs minimally:
AI-generated News of the Day. The Pulse is an experiment to see if it is interesting to get the latest news in 5 minutes small packages generated by an AI every day.

It is completely AI-generated. Only the content is curated. Carsten and I select suitable news items. After that, the manuscript and the audio file are automatically created.

Accordingly, we cannot always guarantee accuracy.

Taxonomy, Opportunities, and Challenges of Representation Engineering for Large Language Models  -  https://arxiv.org/pdf/2502.19649

Improving the Scaling Laws of Synthetic Data with Deliberate Practice - https://arxiv.org/pdf/2502.15588

Support the show

Welcome to the Knowledge Science Pulse podcast. I'm Sigurd, and I’m here with my co-host Carsten.
Today, we’re diving into two intresting AI research topics: improving synthetic data scaling and a new way to control large language models.
Carsten, have you ever wondered if AI can get better at learning the way humans do?
####That’s an interesting question!
AI is great at processing massive amounts of data, but does it actually learn efficiently like humans?
####That's precisely what the paper Improving the Scaling Laws of Synthetic Data with Deliberate Practice explores
It introduces a method called Deliberate Practice which enables AI models to train smarter, not harder
####Deliberate Practice sounds like something humans do when learning a new skill
####Thats right!
Instead of just giving AI more data, Deliberate Practice dynamically selects the most challenging and informative examples for training
Think of it like a student practicing the hardest problems instead of repeating easy ones.
This way models learn more efficiently!
####So instead of dumping a huge dataset on the AI, Deliberate Practice picks the toughest cases to improve learning. Does it actually work?
####Absolutely!
The paper shows that Deliberate Practice achieves better performance while using up to 10 times less data
On ImageNet-1k, models trained with DP needed 30% fewer training iterations while outperforming standard training methods.
That’s a massive efficiency boost!
####Impressive!
So by focusing on what is challenging the model most, we get better AI with less data. What is the catch here?
####The main challenge is designing good selection criteria.
So how do we know which data is most useful?
The researchers address this by using entropy-based sampling, ensuring AI focuses on the hardest, most uncertain cases.
####Makes sense! Now, you mentioned another breakthrough in AI control.
What’s that about?
####The second paper called: Taxonomy, Opportunities, and Challenges of Representation Engineering for Large Language Models, introduces Representation Engineering, or in short RepE.
This is a new way to control AI behavior by directly modifying its internal representations.
####But how are Internal representations that different from prompting or fine-tuning?
####Good question!
Instead of changing the input (like prompting) or adjusting model weights (like fine-tuning), RepE tweaks how the model processes information internally.
This allows more precise, efficient, and interpretable control.
####So, it’s like adjusting the gears inside an AI instead of just changing what we feed into it! What can this be used for?
####Tons of applications.
RepE can steer AI behavior, align outputs with human preferences and even reduce biases!
For example, if an AI is generating responses that are too aggressive, RepE can shift it toward more neutral or polite responses without retraining!
####That sounds very powerful.
This should raise the question of whether there a limitations?
####Yes! The biggest challenges are figuring out how different concepts are represented inside the AI and ensuring that changing one thing doesn’t cause unintended side effects.
The paper highlights the need for better tools to analyze and control these representations!
####That’s fascinating! Between smarter training and deeper AI control, it sounds like we’re moving toward more efficient and adaptable AI systems.
####Exactly! Both papers show exciting ways to make AI smarter and more controllable. We’re just scratching the surface of what’s possible!
####I can’t wait to see what’s next. Thanks for the breakdown Sigurd
####My pleasure! That’s all for today’s Pulse episode. Thanks for listening, and Join us again next time on the Knowledge Science Pulse podcast