This Trillion-Parameter AI Just DESTROYED GPT-5 & Gemini — Live Demo + Results!
10:11

This Trillion-Parameter AI Just DESTROYED GPT-5 & Gemini — Live Demo + Results!

Universe of AI 18.10.2025 12 733 просмотров 350 лайков обн. 18.02.2026
Поделиться Telegram VK Бот
Транскрипт Скачать .md
Анализ с AI
Описание видео
🚨 Open-source just struck back. InclusionAI’s brand-new Ling-1T model is here — a 1-trillion-parameter AI that’s outperforming GPT-5, Gemini 2.5, and DeepSeek V3 in complex reasoning, code generation, and even visual understanding. Built on the Ling 2.0 architecture, Ling-1T is the largest FP8-trained foundation model ever released, activating only 50 billion parameters per token for insane efficiency. It uses Evo-CoT (Evolutionary Chain-of-Thought) to improve reasoning over time, making it one of the smartest — and most efficient — AIs ever built. In this video, we break down: How Ling-1T destroys closed-source models in reasoning benchmarks The new Mixture-of-Experts architecture with only 1/32 activation FP8 training breakthroughs and trillion-scale efficiency Its aesthetic intelligence in front-end code generation and design And why this model might change the open-source landscape forever This isn’t just another AI drop — this is open-source evolution. 0:00 – Intro: Open-Source Strikes Back 0:57 - The Architecture 2:13 - How they did it 3:21 - Results 4:31 - Demo 9:16 - Conclusion 🔗 Links 🔹 Model Page → https://huggingface.co/inclusionAI/Ling-1T 🔗 My Links: 📩 Sponsor a Video or Feature Your Product: intheuniverseofaiz@gmail.com 🔥 Become a Patron (Private Discord): /worldofai 🧠 Follow me on Twitter: /intheworldofai 🌐 Website: https://www.worldzofai.com #Ling1T #InclusionAI #AInews #OpenSourceAI #GPT5 #Gemini2 #DeepSeek #MachineLearning #LLMs #ArtificialIntelligence #TechNews #AITools #UniverseofAI Ling-1T, InclusionAI, GPT-5, Gemini 2.5, DeepSeek V3, open-source AI, trillion-parameter model, FP8 training, Evo-CoT, Mixture-of-Experts, AI reasoning, machine learning news, AI updates, reasoning model, AI architecture, universe of ai, ai breakthroughs, ai revolution, ai comparison, gpt competitor

Оглавление (6 сегментов)

  1. 0:00 Intro: Open-Source Strikes Back 156 сл.
  2. 0:57 The Architecture 179 сл.
  3. 2:13 How they did it 169 сл.
  4. 3:21 Results 164 сл.
  5. 4:31 Demo 784 сл.
  6. 9:16 Conclusion 148 сл.
0:00

Intro: Open-Source Strikes Back

The AI race just hit another milestone. We've talked about OpenAI's GPT5, Google's Gemini 3 coming out soon, and even Deep Seek's insane speed improvements. But today, something huge has quietly dropped in the open-source world. Meet Ling 1T, a 1 trillion parameter model built by Inclusion AI. It's the first flagship non-thinking model in their new Ling 2. 0 O series and is designed to push the limits of efficient reasoning and scalable cognition. Now that phrase might sound complicated, but what it really means is Ling 1T isn't trying to act like a human brain. It's trying to reason better than one. Trained on more than 20 trillion highquality reasoning dense tokens and running on one trillion parameters with only 50 billion active per token. Ling 1T is aiming for something no open model has done before. flagship level reasoning power at a fraction of the compute. At its core, Ling 1T is powered
0:57

The Architecture

by the Ling 2. 0 architecture, a system built entirely around trillion scale efficiency. Instead of activating all 1 trillion parameters at once, it uses a mixture of experts approach. Only about 132nd of its parameters are active per token. That's roughly 50 billion at a time. This makes Ling OneT behave more like a network of specialized brains that light up only when needed, giving it immense scale without crushing compute or memory limits. It also introduces several new tricks under the hood. MTP layers to improve compositional reasoning, sigmoid scoring, expert routing with zero mean updates for better stability, and QK normalization to keep training smooth and balanced. And here's the wild part. Ling 1T is the largest FP8 train foundation model ever built. FP8 mix precision training gives it over 15% speed improvement, 40% better GPU utilization, and barely any accuracy loss compared to BF-116. All of this means trillion scale models are no longer just for massive tech companies. They're becoming accessible and efficient in open research. To make Ling OneT this capable, Inclusion AI
2:13

How they did it

built one of the most carefully crafted training pipelines we've seen. They started with over 20 trillion tokens, gradually increasing the difficulty of the data as training progressed. More than 40% of those tokens were reasoning dense math, logic, programming, symbolic language, the kind of material that forces a model to think in steps. In the middle of training, they introduce a stage called reasoning pre-activation. It uses a cured chain of thought corpra to warm up the model's logic pathways before the final training happens. Then in post-raining they implemented something called EVO coot or evolutionary chain of thought. Think of it as a natural selection for reasoning. The model generates multiple reasoning paths, learns which one works best, and evolves over time to get more precise without exploding its compute costs. They even redesigned the learning rateuler using a warm-up stable merge or WSM schedule to mimic how learning slows down naturally. Every detail of this pipeline was built for one goal. Maximize reasoning power. Okay, numbers
3:21

Results

time. How did it actually perform? Linget was tested against both open and closed source heavyweights models like Deepseek version 3. 1 Terminus, Kimmy K2 Instruct, GPT5 Main, and Gemini 2. 5 Pro. Across code generation, software development, competition level math and logical reasoning. Ling 1T consistently led in complex reasoning ability and overall efficiency. In the AME25 benchmark, it even extend the parental frontier, basically setting a new balance point between reasoning accuracy and reasoning length. That's a fancy way of saying it doesn't just give more accurate answers, it gives smarter answers with less wasted computation. Then there's the BFCL version 3 tool use test, a benchmark that measures how well a model can use external tools or APIs. Ling 1T scored around 70% tool call accuracy with only light instruction tuning. Despite never being trained on massive trajectory data, it achieved that. So even though it's labeled as a non-thinking model, it starts to behave like something that thinks. Now, here's
4:31

Demo

where things get really interesting. Ling1T doesn't just reason, it designs. It excels at front-end code generation and visual reasoning thanks to a new syntax function aesthetics reward mechanism. That means it doesn't only produce correct and functional code. It produces code that looks good that aligns with user intent and that understands layout, color, and design principles. On Artifacts Bench, a benchmark for front-end and design intelligence, Lingi actually ranked first among all open- source models. And get this, the benchmark visualization shown on his hugging face card were generated by Linguenti itself. We're seeing the early signs of models that not only solve logical problems, but understand aesthetic structure and human visual preferences. On the screen, you can see me ask Ling to generate me a website for a clothing store specifically for gymwear. It's generating a HTML file for me that I can use. Let's see what it creates. All right, let me just paste my code in here. Um, okay. Looks okay. Uh, as you can see, the code is in here. It called itself the flexware. Um, there's buttons on there. Moisture wicking, four-way stretch, recycle materials. Oh, it even offered a discount. The collection button doesn't really work obviously. And then the other pop-up buttons don't work, but not too bad for what it has created. Okay. It even validates if my email works or not, which is pretty good. And there's some returns section. Okay, I think that's where it stops. All right, let's try another prompt uh with the new model. We're going to ask it a reasoning question, which is that Anna takes a ball, puts it in a red box, then leaves the room. Bob takes the ball out of the red box, and then puts it into the yellow box, then leaves the room. Anna returns to the room. Where will she look for the ball? So, the answer should be red box because Anna put it in the red box. And I'm going to ask the model to not only solve this, but show me what it's thinking and how it got to its answer. So, I'm going to ask, show me how you got the answer. All right, let's see. Okay, so I did get the answer right and I will look in the red box because that's where she last place the ball and believes it still is. And obviously she didn't see Bob moving into the yellow box. Okay, makes sense. Let's see how it reasons it. Number one, it says Anna's initial action places the ball in the red box. Her belief the ball is in the red box. Okay, that's pretty nice. Bob's action while Anna is absent. So it understands that Anna is not in the room. Whatever Bob does should not be clear to Anna. And then when Anna returns, she has no information about the ball's location. Her mental state still affects her last observation. Therefore, she searches where she thinks it is. The red box. This test theory of mind. The Okay, this is kind of cool. It's giving us a philosophy or like the reasoning it used to help solve it, which is the theory of mind. The ability to attribute false beliefs to others. Children under four years typically fail this task, searching the yellow box where it actually is. Success indicates understanding that others can hold incorrect beliefs about the world. That's pretty sick. What's fascinating is how this simple experiment reveals whether someone grasps that beliefs don't always match reality. Do you work with development psychology or is this curiositydriven? That I think is very cool and kind of scary why it's asking me that question. As we can see, this model is clearly smart and it definitely has a reasoning step-by-step situation in mind. So, it obviously is ready for reasoning tasks and harder level questions. Just for testing purposes, we're going to ask chat GPT the same question and let's see what it does. Okay, it's thinking. So obviously you did get the answer right. Basic question which is look in the red box. Uh how I got it. No spoilersy mind readading just facts and beliefs timeline and it puts the ball in the red box leaves and the returns. It does obviously walk us through how it did it, but obviously the other model did it much better in more detail and also had a follow-up question which I think was pretty cool. So, you know, obviously the answer is pretty straightforward, but we can see how both of these models think differently and reason differently and provide different answers at the end. In conclusion
9:16

Conclusion

Linguent represents more than another giant model. It's a philosophy shift. Instead of chasing thinking models that mimic human minds, inclusion AI built a system that focuses purely on efficient reasoning. Less emotion, more logic, less imitation, more precision. And the results speak for themselves. Near flagship reasoning power, strong visual and code synthesis, and all of it openly released to the public. So, what does this mean for AI's future? Maybe the next leap won't come from making models larger, but from teaching them to think smarter. Lingui proves that open source can still innovate at the trillion parameter level, and that efficient design might just be the next arms race. What do you think? Can open models like Ling1T really challenge GPT5 and Gemini headon? Let me know in the comments. If you enjoyed this breakdown, hit that like button, subscribe, and I'll see you next

Ещё от Universe of AI

Ctrl+V

Экстракт Знаний в Telegram

Транскрипты, идеи, методички — всё самое полезное из лучших YouTube-каналов.

Подписаться