There is a consensus growing in the AI research community that the traditional ‘scaling law’ of deep learning might be reaching its limits. This law has been a driving force behind recent AI advancements, suggesting that simply increasing the amount of data, training time, and model parameters would lead to better performance. However, recent reports indicate that scaling has plateaued, with researchers struggling to create models that outperform OpenAI’s GPT-4, which is nearly two years old. One of the reasons behind this stagnation could be the exhaustion of high-quality training data. Supplementation with synthetic data is causing the new models to resemble older models, leading to problematic performance. Despite this, some researchers believe that alternative approaches like OpenAI’s ‘o1’ model, which focuses on extended ‘thinking’ time for tasks, might offer a way forward even if scaling plateaus.