The era of scaling large language models (LLMs) by simply increasing parameters, data, and compute may be ending, as improvements plateau. However, a new scaling paradigm focuses on increasing "test-time compute," allowing models to think longer and solve more complex problems, as demonstrated by OpenAI's 03 model. This approach shows promise for achieving artificial general intelligence and may extend to other AI modalities.