Game OVER! Chinas New AI Video Tool BEATS SORA! (KLING AI Text-To-Video)
Вставка
- Опубліковано 9 лют 2025
- Join My Private Community - / theaigrid
🐤 Follow Me on Twitter / theaigrid
🌐 Checkout My website - theaigrid.com/
Links From Todays Video:
kling.kuaishou...
Welcome to my channel where i bring you the latest breakthroughs in AI. From deep learning to robotics, i cover it all. My videos offer valuable insights and perspectives that will expand your knowledge and understanding of this rapidly evolving field. Be sure to subscribe and stay updated on my latest videos.
Was there anything i missed?
(For Business Enquiries) contact@theaigrid.com
#LLM #Largelanguagemodel #chatgpt
#AI
#ArtificialIntelligence
#MachineLearning
#DeepLearning
#NeuralNetworks
#Robotics
#DataScience
My five year old is sitting next to me watching the panda play the guitar and his eyes got big then ran off. I'm obliged to explain AI video tools to him now. What a strange world to grow up in.
😂😂😂😂😂
😂😂😂😂😂
There is something really powerful about being the first to put it out there and even if it’s second-best, it might get a lot of adoption and people will build workflows around it, and it will win in the long run
Not with AI tools ...
Anything humans build Ai will take over.
Post-labor world dude 1 to post-labor world dude 2: Remember when workflows were a thing?
@TeamLorie Prompting into an AI system is not the same as mastering an application like Autocad, Premiere, Protools, Maya, Excel , etc etc ... If you get a more powerful data node that can be feed with your history ( for those attached to their past) , just migration and prompt in the new platform will increment the productivity.... that thing of staying with the first developer is a thing of times where/when technologies develop slow and require a set of specific skills to master the resources... slow humans are animals of habits, so changing their habits is difficult for most of them.. but Prompting is an habit that doesn't make a single platform an exclusive necessity ... Too, when you got systems that can make from the Prompting up to the final outcome... what people will take will be the better and cheaper... If the cheaper is the best, eventually, everyone will tend to migrate into those.... the issue is that greedy Westerners would try to set their platforms for preventing users migration of data... but that will make their platforms obsolete compared with fully open and compatible open source platforms....
With time that thing of "branding " will disappear because Prompting is not about who makes the outcomes, Prompting is about who enjoys those outcomes.... If you enjoy AI outcomes, it doesn't matter where they come from... If you like more the AI outcomes from new platforms , you don't need to get attached to obsolete platforms.
if it was VHS vs Betamax, sure. But things are moving at such pace now the competitor will be realasing bluray in a couple months and not betamax to compete with your VHS.
Sora isn’t even available
@timer4times2 "Safety" reasons. Nah it's probably compute resources. "Open"AI is too busy making GPT-5.
@timer4times2you obviously saw all the important people just walk out of OpenAI
@@youdontneedmyrealnamethis is a very good point
Ashton Kutcher has a copy.😮
CloseAI lmao
The remarkable consistency in the use of the word remarkable is remarkably remarkable in terms of consistency.
That’s literally a remarkably remarkable remark 😂
What a capable and incredibly convincing depiction of realistic reasoning, showing a very consistent ability to output remarkable accurate impressions of this impressively consistent video.
IMO people aren't talking about the most important factor about AI video generator that is how much customizable the videos are and the ability of following a prompt, it can create perfect videos but if it's as limited as image generators, then it's useless
remarkable: 26 times, consistency: 16 times from transcript lol
This definitely beats Sora in most aspects.
Anyone who says otherwise is just anti-chinese.
But the fact that Sora got dethroned before it was even released is quite impressive.
I think that 2024 is going to be remembered as the year when China started to surpass the US in AI development and robotics.
The quality does not look definitively better than Sora’s to me, although it’s in the same ballpark. The larger point is that it would be extremely naive to blindly trust any results coming out of China as being fully authentic. We can barely even trust our own companies not to doctor their results (Google was just caught doing it last year), so the idea that we blindly trust results coming out of China is pretty foolish IMO.
Its not anti-chinese, its you that are pro-chinese....
I agree. But it's no surprise. Several years ago, they announced that this would be the decade that they surpass the US on several key measures.
Researchers are training text-video models at least 3 years ago. There is a website to download those models, I couldn't remember the name, but it was a professional website for researchers. Sora doesn't look special, but they have the money and equipment to train high resolution videos.
This model absolutely does NOT “beat Sora in most respects,” and it is not remotely “anti-Chinese” to say that. It’s an impressive model for sure. It’s not quite at Sora’s level, based on what we’ve seen. That’s just a fact.
I was at the WAIC scene, and the head of Kling said that they were iterating 3 times a month, indicating that the learning ability of the video large model is still very strong, and the emergence rate seems to have exceeded the large language model and the picture large model. In addition to the launch of the HD version and the first and last frame control, there are also camera control, a single generation of more than 10s, online web side. The Kling official at the site also said that some other functions have been developed, but have not been released yet. These functions include character ID retention, voice face matching, screen structure control, etc
Even if Sora is better China is doing the only thing that matters. Releasing a usable product.
@timer4times2 or if their claims are only that since we haven’t seen it perform.
Documentary Lifestyle 60p is an untapped market atm. Most of the current models appeal for that cinematic movie feel. So this is timely, lots of potential for long form storytelling. Exciting times! Let the games begin 🏹
Documentary lifestyle 60p? What's that
@@a.nobodys.nobody 60 frames per second-which is what live tv, most documentaries, sports, vlogs etc. use. Though it looks like it can only output 30p atm.
Movies use 24p for the cinematic look. So this model opens up an entirely new market of lifestyle storytelling.
The main eating Ramen is AI generated.
Look at his left hand. His finger morphs around the bowl and you can see the wrinkles on his right hand fingers change when he moves the chopsticks. Pretty amazing stuff China.
Good to see some real competition in the field. I’m sick of OpenAI not releasing more useful models at a consistent clip. And let’s not even talk about flops like Gemini. Claude has also been shit lately, at least for my use case.
The beauty about AI is that you don’t have to be married to a model. If a better one comes along, it should be a plug and play replacement.
Claude banned me before i could even try it for the first time.. They are shit company.
But, kind Sir, Claude is not shit. I used it to create an Autonomous Trading Bot and it's working wonderfully.
@@charlesromelus3 Thats good but i like to be in control of my trades.
@@charlesromelus3 Claude is not shit, it seems nerfed for some use cases. I’m talking about Opus specifically in the health field.
GPT 4-0 is better than Opus at certain tasks, not even close. Opus is better at conversation, feels more natural. At least that’s been my experience.
Yes, kling shocked me so much. In addition to text input commands, it also supports image input commands. The generated effects are very shocking and can’t be seen to be fake at all.
How does that gutter oil taste
Most impressive clip was the Lego clip. Looked like an actual movie scene. Rediculous 😂
It’s nice to know that our future AI overlords think that we are messy eaters.
I'm sick of all these fake releases. If I can't use it now, It sucks, really.
Yeah, I call at least partial BS if it isn’t publicly available to confirm their claims.
They want us to think that things are happening really fast so we will stay engaged. Turns out things are progressing at the normal pace.
The website doesn’t even work
Bro, this is not text tool this is fake video machine lol. They still dont have good idea how to make sure its not rly good fake video
Can't trust anything from China, sorry but that's fact.
That's very impressive. It looks diffusiony from some of the morphing but the composition is outstanding.
It makes me wonder if they're constructing a 3d model and from that, creating/combining edge & depth maps with ray tracing to create light maps, maybe with a fine contrast map for intricate details, then plumbing that through a diffusion model for finishing. If there's no 3d model involved, this AI clearly understands the world better than we give it credit for.
I especially like the train, the night sky keeping maintaining all the stars and the guy eating noodles, wow!
It's because models like Sora and those from China are all Video Diffusion Transformers, a new kind of model, they will basically output the same Diffusion images style but in videos.
What I liked about the Sora demonstration is that along with their best ones, they showed its flaws too. That way we could see it's limitations, potential, and ways they can improve it. Kinda wish the Chinese Sora did this too.
Chop sticks was on point he was holding them perfect hands looks good to me.
I wonder if allowing everyone to make Hollywood quality movies will result in better movies... Saturation yes but some gems too?
Terrifier 5 , yes!!! I WILL MAKE THIS HAPPEN!
@@charlesromelus3 OK, I need to watch that!
Considering how well indie creators are doing on UA-cam recently... Most definitely
I thought the chinese guy eating the noodles was a real video. How can you tell if they didn't just record a guy eating noodles and called it AI?
look at his fingers there is no light bouncing off of the skin. Looks like plastic
@@unityman3133And infinite pasta.
The plate behaves like it is on a table, completely static, while it is being hold by his left hand. The level of quality is remarkable enough to deceive us at first glance, but a huge part of it is due to the fact that we are not used to question whether a video is real or not. When you start to observe every detail you spot many inconsistencies, but they're completely ignored by our brains when we're not paying close attention to them, which is really shocking!
@@ggmanerareality is what the mind believes it to be
Also there shouldn't be any yellowish liquid on his mouth after he has eaten the noodles..beacuse when they are being pulled off the plate..there is no liquid stuck to themm..the are just white..so thats another hint
the ai does better shadows than nasa lol.
Read a book not a Alex Jones website
@@Seriouslydave keep the faith, stay strong.
My favorite video was the one where we got to use it. LOL
I do t think the game is over. Just getting started.
How can we be sure that Sora is really capable of what we were showed? Its been months since presentation and still none can use it. The same is about 4o
Not efficient enough, Video Diffusion Transformers have the highest cost of compute.
So they are not ready, that is telling us all 😂
4o is out already! Free desktop access
Have they caught up?
It’s not as good as Sora - not far way but it’s less stable. Check the paving under the bike in the 2m video. Or the typers and sides of the car. It has flaws yet that Sora doesn’t have.
Anyone who thinks this is better than sora should see an eye doctor. It is not a matter of opinion
Lol, training Sora is like training GPT-2, it's just a new kind of models (Diffusion Transformer), no secret sauce behind, of course there will be plenty of competitors catching up.
open ai is hiding something somewhere
This caught up majority of things that come out of China anyhow are usually fake but even if they did end up generating really good video what does that supposed to mean a video generator? Yeah it’s impressive but the funny thing is is you’re also gotta look at who is producing the product it’s like saying wow the devil can really play some music
Are these video's being written by AI? The tend to be getting more pretty, pretty, pretty gibberish, hyperbolic and filled TRULY silly hallucinations.... Sora stomps this on every level, from fidelity to consistency... not one did I need to double take on this video as to whether it might be real... the was not TRULY SHOCKING about this. I guess Wes is the last man standing of the first wave of successful AI channels that hasn’t gone total cringe.
I like AIExplained too.
If you can't notice that hallucinations are 10 times less than SORA, you must've pretty shitty eyes. Sora was unable to respect dimensions and the consistency was highly lacking, visual glitches and impossibilities were omnipresent and the respect to the prompt only worked for very short videos. Those problems have been greatly reduced.
You claim that it's laughably bad? Alright buddy. Ler's take one random video: The astronaut. What's clearly visible during your first watch which gives away the fact that it's AI?
@@user-cg7gd5pw5b nope... obviously your drinking the koolaid.
@@Vitaphone Then go ahead and tell me what I asked for. If you can't, your point is BS...
Bro you're attacking their videos but can't give examples of why. All I do is point it out. Who's really wrong here?
@@Vitaphone It feels like comparing Stable Diffusion 3 with Dall-E 3, just they have different styles, they're the same kind of model (Video Diffusion Transformer).
I found the train most impressive, driving past the realistic waves and beaches and just the length of the video... hope they release it soon
So it beats SORA in the AI category of "things we show people, but you will apparently never have access to". Honestly, I'm beginning to really HATE these demos, because we cannot use them. What is the point? Give me cool new AI tech I can actually use.
In due time. You will have your God. And you will have made It - with your own Hands.
I think released by Chinese companies not CHINA
Let's keep doing game overs everyday.
Leon Ding helped them ?
"The FBI estimated in its report that the annual cost to the U.S. economy of counterfeit goods, pirated software, and theft of trade secrets is between $225 billion and $600 billion."
its not a cost its a Loss, but there's novway to tell if the losses would have occurred, apart from commercial licenses.
Sora, and Stable Diffision entered the chat
Sora is still in development before they will open, don't know when it will be available for everyone to use.
Hollywood is cooked fr
The noodles guy left and you can see his thumb on the bowl doesn't look right. The bowl is a static image
Except that the big slurp of noodles he ate had no residual sauce on them. The sauce from the bowl just teleported to his lips.
"Game over" for the 10 millionth time 🤪
That happened to my coffee this morning.
The Rabbit R1 looked cool on stage, too.
Any software?
The buildings and cat's ears reflected off the car in the driving one...
I have to be skeptical because these companies usually put their best forward, or hide any tailoring they had to do.
But even still that's impressive.
Is it possible to create multiple videos with same context or story?
We are in an era similar to the one that invented steam engines.
Maybe kuaishou can take tiktoks market share now with this video generator that content creators can use
The clips look like a cell phone video. Very realistic.
sorry but we are looking to different videos then, are you sure this is better than Sora? you still want to extract "extraordinary" from all your videos
where is the platform or app for this?
快影
快影
"China just went ahead and released"
Wrong. If it's not available to use, then It has not been released. You don't consider a video game to be "released" when it's in closed alpha or beta testing, do you? It's released when it's ready to be used by everyone.
When will this be available for public use?
It already is unfortunately you need a Chinese phone number its only available on china but a couple of twitter accounts have been accepting prompts and showing the results
Where do i download this?
Consistently consistent in a consistential consistentiality. Truly. Jk. Very impressive. But Sora isn’t even out, was a simple capability demo in an extremely limited release. I guess this isn’t out either, but the fact that it’s up to 2 mins long is really impressive. I can’t wait for the time when we’re able to create our own home movies with ourselves as the main characters. Who needs actors when you can be the actor?
Yeah even sora from months ago is better. Compare to the music videos/ transportation video/ dog in Italy sora is better. Not to mention sora has progressed and is far more capable than we are aware.
Sounds like an LLM hallucination, plot twist: Sora is just a new kind of model (Diffusion Transformer), no secret sauce, everyone can catch up.
Congratulatiins for de vide and for the fratures.
Impressive, but still too much warping. I am betting that will be solved in the next 1-2 years.
most likely scenario would be pre-training using video that closely matches, or even worse, is filmed specifically to match, the specific prompts. Without a release of any of these ai tools (chinese or not) there is no way to trust these announcements. In depth analysis if the marketing results is pointless.
Maby thanks for the update 😊
Google just left the chat 😂😂 , no doubt china is not falling behind
It doesn't sounds weird to you to say "China" as if the entire country were behind this Chinese startup ?
@@TheRealUsername do you know how the communist party of china works?? They are behind every company that is in china.
@@TheRealUsername
China : "Ours" (communist meme)
Wow, I can't wait to not try it!
IMAO SORA has nothing to sweat about. This one has train tracks appearing out of nowhere. (5:62) and lots of other eye jarring artifacts. The desert between the rider's body, the horse, and the reins was much grayer than the rest of the desert. This is good. It's not up to great yet.
{^_^}
Well, that doesn't really matter, Sora is just the same kind of model (Video Diffusion Transformer) as those presented anyway.
My Ai-Asian isekai desire is the real thing. Hmmm case in point! That rabbit isn't myopic. I am enjoying the ultra effects. I am so happy. I don't have to "toy with fantasy" for the sake of your humanity. Strong Concept combination Ability is my pleasure now.
Alright.. A little bit of geography lessons. The red sunset looks really cool, but it's impossible to see it like that BOTH in front of you AND in the car mirror, because we only have one sun in our solar system. So yeah, the colors are indeed impressive, but the video can be seen as generated right in the first second when you watch it. And don't ChatGPT me now; there are some rare exceptions, but this is not it :)
That's true the clouds would be lit up too rather than shadows.
Bro when will they make it available for thr rest if thr world. Can't log in without chinese number
I take it they have to answer any copywriting concerns.
I think it just better that more countries join, so it kills that America 1st. trend which lately has been taken over, companies first realising US and then only later to the rest of the world.
It doesn't beat SORA. It looks way worse. Sora is very smooth, doesn't have legs blurring with themselves.
Game over until the next time the game is over.
Given how things can be faked in demonstrations, I wouldn't be surprised if there were real videos in here and told they were AI generated. An example being 9:35.
That is AI generated. Look at his left hand. His finger morphs around the bowl and you can see the wrinkles on his right hand fingers change when he moves the chopsticks.
4- The emergence of AI and its necessities
X Mohammad Rahim Jamshidi
Honestly, I hope we have access to some legacy software. Like the stuff that made the whacky will Smith eating spaghetti. That kind of mental stuff. Though I suppose you could just include that in the prompt, huh?
Not at all, you just look at that bicycle boy video, that's the reality(I think their mothed is basically cobble images together), however this generator proves scaling law works again.
Why are you saying "genuamly @ 21.40? are you AI generated? :) Nice video
Leon Ding sent China a lot of AI code.
Kuaishou is a short-video platform , rival of TikTok. But it mainly operates in China. So they can use abundant videos as reference to generate short clips
Can we use Kling?
They're not AI generated videos, they're REALLLL :))
I like your vids, but you seriously need an auto loop instead of constantly manually clicking play, kinda annoying
These are going to be interesting US elections…
How would you know it wasn't Sora with a wrapper?
Sora is not publicly available, that's how.
It's like stating the Stable Diffusion 3 demos where made using Midjourney
Open AI, vs china, not Google!!!
We don't get the best from Google as that goes to the military n Govs watching us.
Well done.
What's next? 😂😂
This uses PINNs with 3D spatiotemporal attention?
Is Hollywood ruined?
Remarkable!
Don’t get me wrong it’s good, but idk if it beats sora tho.
What makes you say that?
You haven't even tested Sora, besides Sora is like the LLM of video generation, you can train your own if you have the compute, its training process is freely available in some research papers
a sauce god talking about sauce
Too much cherry picking I would like to see an independent evaluation side by side with Sora. The cat vid was not that good. And by side by side I mean the same input running them side by side
Cope better
Truly truly Remarkably remarkable. Actually, Genuinely pretty pretty crazy. Trust me.
This is not about catching up. They just steal models and adapt them for themselves. This makes it so they are consistently just 1 year or so behind. We are not seeing the latest from Sora etc
Fake probably! And he makes this long video like it’s real with bunch of speculations without testing it himself
lol @8:47 world model - milk is pouring above milk level in jug and milk level not diminishing as cup level raises
Yeah, you should probably show them how to do it.
Where is the European company?!😂
But in reality, this is all speculation because no one on the public side has access to do a true comaparsions lets be real here.
8:00
They create a company with my surname
That's cool but don't care until I can use it. Sick of all this hype. I want results
Cherrypicked videos and post edited after.
Naah, not good as Sora. However, this shows that yes, we are heading towards a terminal race condition. We're not there yet, but real soon we shall be. Frightening.
meh, still a long road ahead for perfection
And nobody cares where the training data comes from..
Why can't runwayml or pika labs do anything lol they have a decent amount of money behind them
They should but the compute power to train a Video Diffusion Transformer seems higher than training a SOTA LLM
it looks shit an incoherent in comparison to Sora. Obviously its running at a lower level of complexity.
Neither could do reflections correctly.