share_log

阶跃星辰姜大昕:大模型已同时具备人类大脑快思考与慢思考的能力

Jieyue Xingchen Jiang Daxin: The large-scale model now has the ability to simultaneously perform fast thinking and slow thinking like the human brain.

cls.cn ·  Sep 19 13:12

①OpenAI's large model o1 enables AI to have the ability of human slow thinking (actively reflecting, correcting complex thinking) through reinforcement learning, and the next key breakthrough will be to improve the generalization ability of the reinforcement learning model and accelerate the integration of multimodal understanding and generation, which is a further breakthrough. ②GPT-4o signifies the progress of multimodal integration, and it is the foundation for better modeling the physical world and thoroughly realizing the simulated world.

On September 19th, in the 2024 Yunqi Conference, Dr. Jiang Daxin, the founder and CEO of Star Leap, stated that the development of AI is undergoing a critical technological paradigm shift: OpenAI's large model o1 has explored a way to enable AI to have the ability of human slow thinking (actively reflecting, correcting complex thinking) through reinforcement learning. The next key steps for AI technology are to improve the generalization ability of the reinforcement learning model and to accelerate the integration of multimodal understanding and generation.

1Cn4qVx9cs.jpg

The person second from the left is Dr. Jiang Daxin, CEO of Star Leap.

He disclosed that Star Leap is actively exploring new technological paradigms and has already implemented the methodology of reinforcement learning training on trillion-parameter models. At the same time, they are continuously improving the performance of the underlying large models and enhancing the end-user product experience. Recently, Star Leap's self-developed Step-2 trillion-parameter MoE language model has been integrated into the intelligent assistant 'Leap Question,' demonstrating stronger command following, creation, and reasoning capabilities.

At the Yunqi Conference, the latest release of the o1 model by OpenAI has sparked discussions. The o1 model is considered to have explored a way to enable AI to have the ability of human slow thinking (actively reflecting, correcting, trying different strategies, and conducting complex reasoning) through reinforcement learning technology. This is the first time a large model has simultaneously possessed the ability of human brain System 1 (fast thinking, giving direct answers) and System 2 (slow thinking). Jiang Daxin believes that this is a critical step for large models to begin to possess the ability to infer the world.

"We divide the path to AGI into three parallel stages: simulating the world, exploring the world, and inferring the world. In the past year, there have been breakthrough technological advancements in all three areas, and the pace of development can be described as 'a day in AI, a year on Earth'". He stated that in addition to the progress made by o1 in inferring the world, GPT-4o signifies the progress of multimodal integration, which is the foundation for better modeling the physical world and completely realizing the simulated world. Regarding exploring the world, Tesla's release of the fully autonomous driving system FSDV12 indicates the technological direction for how embodied intelligent devices can be combined with large models, transitioning from the digital world to exploring the physical world.

In climbing the road to AGI, strong foundational models are needed for reinforcement learning, multimodal understanding, and industry models. This year, Star Leap has continued to iterate the trillion-parameter language model Step-2, and in March, it was the first in China to release a preview version of the trillion-parameter language model. In July, it was officially released. After testing, Step-2's comprehensive capabilities have increased by nearly 50% compared to the hundred-billion-parameter language model Step-1, with significant improvements in logical reasoning, mathematics, programming, and knowledge.

Meanwhile, YUE Xingchen announced the comprehensive upgrade of its smart assistant "YUEWEN", offering users the Step-2 trillion parameter MoE language model capability for free for a limited time, and launching the multi-modal search and question-answering feature "PAI ZHAO WEN". Users can achieve "capture and ask" through image interaction, solving the scenario demands that are difficult to efficiently describe in text and voice interaction modes.

The translation is provided by third-party software.


The above content is for informational or educational purposes only and does not constitute any investment advice related to Futu. Although we strive to ensure the truthfulness, accuracy, and originality of all such content, we cannot guarantee it.
    Write a comment