Game OVER! Chinas New AI Video Tool BEATS SORA! (KLING AI Text-To-Video)
Вставка
- Опубліковано 16 вер 2024
- Join My Private Community - / theaigrid
🐤 Follow Me on Twitter / theaigrid
🌐 Checkout My website - theaigrid.com/
Links From Todays Video:
kling.kuaishou...
Welcome to my channel where i bring you the latest breakthroughs in AI. From deep learning to robotics, i cover it all. My videos offer valuable insights and perspectives that will expand your knowledge and understanding of this rapidly evolving field. Be sure to subscribe and stay updated on my latest videos.
Was there anything i missed?
(For Business Enquiries) contact@theaigrid.com
#LLM #Largelanguagemodel #chatgpt
#AI
#ArtificialIntelligence
#MachineLearning
#DeepLearning
#NeuralNetworks
#Robotics
#DataScience
My five year old is sitting next to me watching the panda play the guitar and his eyes got big then ran off. I'm obliged to explain AI video tools to him now. What a strange world to grow up in.
😂😂😂😂😂
😂😂😂😂😂
Sora isn’t even available
Not even their newest voice feature. Seriously, what the hell is going on with Closed, sorry OpenAI?
@@timer4times2 "Safety" reasons. Nah it's probably compute resources. "Open"AI is too busy making GPT-5.
@@timer4times2you obviously saw all the important people just walk out of OpenAI
@@youdontneedmyrealnamethis is a very good point
Ashton Kutcher has a copy.😮
There is something really powerful about being the first to put it out there and even if it’s second-best, it might get a lot of adoption and people will build workflows around it, and it will win in the long run
Not with AI tools ...
Anything humans build Ai will take over.
Post-labor world dude 1 to post-labor world dude 2: Remember when workflows were a thing?
@TeamLorie Prompting into an AI system is not the same as mastering an application like Autocad, Premiere, Protools, Maya, Excel , etc etc ... If you get a more powerful data node that can be feed with your history ( for those attached to their past) , just migration and prompt in the new platform will increment the productivity.... that thing of staying with the first developer is a thing of times where/when technologies develop slow and require a set of specific skills to master the resources... slow humans are animals of habits, so changing their habits is difficult for most of them.. but Prompting is an habit that doesn't make a single platform an exclusive necessity ... Too, when you got systems that can make from the Prompting up to the final outcome... what people will take will be the better and cheaper... If the cheaper is the best, eventually, everyone will tend to migrate into those.... the issue is that greedy Westerners would try to set their platforms for preventing users migration of data... but that will make their platforms obsolete compared with fully open and compatible open source platforms....
With time that thing of "branding " will disappear because Prompting is not about who makes the outcomes, Prompting is about who enjoys those outcomes.... If you enjoy AI outcomes, it doesn't matter where they come from... If you like more the AI outcomes from new platforms , you don't need to get attached to obsolete platforms.
if it was VHS vs Betamax, sure. But things are moving at such pace now the competitor will be realasing bluray in a couple months and not betamax to compete with your VHS.
The remarkable consistency in the use of the word remarkable is remarkably remarkable in terms of consistency.
That’s literally a remarkably remarkable remark 😂
What a capable and incredibly convincing depiction of realistic reasoning, showing a very consistent ability to output remarkable accurate impressions of this impressively consistent video.
IMO people aren't talking about the most important factor about AI video generator that is how much customizable the videos are and the ability of following a prompt, it can create perfect videos but if it's as limited as image generators, then it's useless
remarkable: 26 times, consistency: 16 times from transcript lol
This definitely beats Sora in most aspects.
Anyone who says otherwise is just anti-chinese.
But the fact that Sora got dethroned before it was even released is quite impressive.
I think that 2024 is going to be remembered as the year when China started to surpass the US in AI development and robotics.
The quality does not look definitively better than Sora’s to me, although it’s in the same ballpark. The larger point is that it would be extremely naive to blindly trust any results coming out of China as being fully authentic. We can barely even trust our own companies not to doctor their results (Google was just caught doing it last year), so the idea that we blindly trust results coming out of China is pretty foolish IMO.
Its not anti-chinese, its you that are pro-chinese....
I agree. But it's no surprise. Several years ago, they announced that this would be the decade that they surpass the US on several key measures.
Researchers are training text-video models at least 3 years ago. There is a website to download those models, I couldn't remember the name, but it was a professional website for researchers. Sora doesn't look special, but they have the money and equipment to train high resolution videos.
This model absolutely does NOT “beat Sora in most respects,” and it is not remotely “anti-Chinese” to say that. It’s an impressive model for sure. It’s not quite at Sora’s level, based on what we’ve seen. That’s just a fact.
I was at the WAIC scene, and the head of Kling said that they were iterating 3 times a month, indicating that the learning ability of the video large model is still very strong, and the emergence rate seems to have exceeded the large language model and the picture large model. In addition to the launch of the HD version and the first and last frame control, there are also camera control, a single generation of more than 10s, online web side. The Kling official at the site also said that some other functions have been developed, but have not been released yet. These functions include character ID retention, voice face matching, screen structure control, etc
Even if Sora is better China is doing the only thing that matters. Releasing a usable product.
Agreed. What's the point of Sora being better if only rich people from Hollywood can afford it in 2025?
@@timer4times2 or if their claims are only that since we haven’t seen it perform.
What I liked about the Sora demonstration is that along with their best ones, they showed its flaws too. That way we could see it's limitations, potential, and ways they can improve it. Kinda wish the Chinese Sora did this too.
Yes, kling shocked me so much. In addition to text input commands, it also supports image input commands. The generated effects are very shocking and can’t be seen to be fake at all.
How does that gutter oil taste
Good to see some real competition in the field. I’m sick of OpenAI not releasing more useful models at a consistent clip. And let’s not even talk about flops like Gemini. Claude has also been shit lately, at least for my use case.
The beauty about AI is that you don’t have to be married to a model. If a better one comes along, it should be a plug and play replacement.
Claude banned me before i could even try it for the first time.. They are shit company.
But, kind Sir, Claude is not shit. I used it to create an Autonomous Trading Bot and it's working wonderfully.
@@charles2353 Thats good but i like to be in control of my trades.
@@charles2353 Claude is not shit, it seems nerfed for some use cases. I’m talking about Opus specifically in the health field.
GPT 4-0 is better than Opus at certain tasks, not even close. Opus is better at conversation, feels more natural. At least that’s been my experience.
Documentary Lifestyle 60p is an untapped market atm. Most of the current models appeal for that cinematic movie feel. So this is timely, lots of potential for long form storytelling. Exciting times! Let the games begin 🏹
Documentary lifestyle 60p? What's that
@@a.nobodys.nobody 60 frames per second-which is what live tv, most documentaries, sports, vlogs etc. use. Though it looks like it can only output 30p atm.
Movies use 24p for the cinematic look. So this model opens up an entirely new market of lifestyle storytelling.
Most impressive clip was the Lego clip. Looked like an actual movie scene. Rediculous 😂
*Least Impressive.
The main eating Ramen is AI generated.
Look at his left hand. His finger morphs around the bowl and you can see the wrinkles on his right hand fingers change when he moves the chopsticks. Pretty amazing stuff China.
He said that in video Captain Obvious.
I'm sick of all these fake releases. If I can't use it now, It sucks, really.
Yeah, I call at least partial BS if it isn’t publicly available to confirm their claims.
They want us to think that things are happening really fast so we will stay engaged. Turns out things are progressing at the normal pace.
The website doesn’t even work
Bro, this is not text tool this is fake video machine lol. They still dont have good idea how to make sure its not rly good fake video
Can't trust anything from China, sorry but that's fact.
I wonder if allowing everyone to make Hollywood quality movies will result in better movies... Saturation yes but some gems too?
Terrifier 5 , yes!!! I WILL MAKE THIS HAPPEN!
@@charles2353 OK, I need to watch that!
Considering how well indie creators are doing on UA-cam recently... Most definitely
That's very impressive. It looks diffusiony from some of the morphing but the composition is outstanding.
It makes me wonder if they're constructing a 3d model and from that, creating/combining edge & depth maps with ray tracing to create light maps, maybe with a fine contrast map for intricate details, then plumbing that through a diffusion model for finishing. If there's no 3d model involved, this AI clearly understands the world better than we give it credit for.
I especially like the train, the night sky keeping maintaining all the stars and the guy eating noodles, wow!
It's because models like Sora and those from China are all Video Diffusion Transformers, a new kind of model, they will basically output the same Diffusion images style but in videos.
I thought the chinese guy eating the noodles was a real video. How can you tell if they didn't just record a guy eating noodles and called it AI?
look at his fingers there is no light bouncing off of the skin. Looks like plastic
@@unityman3133And infinite pasta.
The plate behaves like it is on a table, completely static, while it is being hold by his left hand. The level of quality is remarkable enough to deceive us at first glance, but a huge part of it is due to the fact that we are not used to question whether a video is real or not. When you start to observe every detail you spot many inconsistencies, but they're completely ignored by our brains when we're not paying close attention to them, which is really shocking!
@@ggmanerareality is what the mind believes it to be
Also there shouldn't be any yellowish liquid on his mouth after he has eaten the noodles..beacuse when they are being pulled off the plate..there is no liquid stuck to themm..the are just white..so thats another hint
Chop sticks was on point he was holding them perfect hands looks good to me.
It’s nice to know that our future AI overlords think that we are messy eaters.
the ai does better shadows than nasa lol.
Read a book not a Alex Jones website
@@Seriouslydave keep the faith, stay strong.
How can we be sure that Sora is really capable of what we were showed? Its been months since presentation and still none can use it. The same is about 4o
Not efficient enough, Video Diffusion Transformers have the highest cost of compute.
So they are not ready, that is telling us all 😂
4o is out already! Free desktop access
I do t think the game is over. Just getting started.
My favorite video was the one where we got to use it. LOL
I found the train most impressive, driving past the realistic waves and beaches and just the length of the video... hope they release it soon
It's not as good as SORA but. .at least will be available to the public soon. .
Have they caught up?
It’s not as good as Sora - not far way but it’s less stable. Check the paving under the bike in the 2m video. Or the typers and sides of the car. It has flaws yet that Sora doesn’t have.
Anyone who thinks this is better than sora should see an eye doctor. It is not a matter of opinion
Lol, training Sora is like training GPT-2, it's just a new kind of models (Diffusion Transformer), no secret sauce behind, of course there will be plenty of competitors catching up.
open ai is hiding something somewhere
This caught up majority of things that come out of China anyhow are usually fake but even if they did end up generating really good video what does that supposed to mean a video generator? Yeah it’s impressive but the funny thing is is you’re also gotta look at who is producing the product it’s like saying wow the devil can really play some music
Maybe kuaishou can take tiktoks market share now with this video generator that content creators can use
So it beats SORA in the AI category of "things we show people, but you will apparently never have access to". Honestly, I'm beginning to really HATE these demos, because we cannot use them. What is the point? Give me cool new AI tech I can actually use.
In due time. You will have your God. And you will have made It - with your own Hands.
I think released by Chinese companies not CHINA
I'm sure he didn't mean CCP.
Sora, and Stable Diffision entered the chat
The buildings and cat's ears reflected off the car in the driving one...
I have to be skeptical because these companies usually put their best forward, or hide any tailoring they had to do.
But even still that's impressive.
The noodles guy left and you can see his thumb on the bowl doesn't look right. The bowl is a static image
The Rabbit R1 looked cool on stage, too.
Impressive, but still too much warping. I am betting that will be solved in the next 1-2 years.
Sora is still in development before they will open, don't know when it will be available for everyone to use.
Hollywood is cooked fr
Consistently consistent in a consistential consistentiality. Truly. Jk. Very impressive. But Sora isn’t even out, was a simple capability demo in an extremely limited release. I guess this isn’t out either, but the fact that it’s up to 2 mins long is really impressive. I can’t wait for the time when we’re able to create our own home movies with ourselves as the main characters. Who needs actors when you can be the actor?
IMAO SORA has nothing to sweat about. This one has train tracks appearing out of nowhere. (5:62) and lots of other eye jarring artifacts. The desert between the rider's body, the horse, and the reins was much grayer than the rest of the desert. This is good. It's not up to great yet.
{^_^}
Well, that doesn't really matter, Sora is just the same kind of model (Video Diffusion Transformer) as those presented anyway.
Yeah even sora from months ago is better. Compare to the music videos/ transportation video/ dog in Italy sora is better. Not to mention sora has progressed and is far more capable than we are aware.
Sounds like an LLM hallucination, plot twist: Sora is just a new kind of model (Diffusion Transformer), no secret sauce, everyone can catch up.
Bro when will they make it available for thr rest if thr world. Can't log in without chinese number
That happened to my coffee this morning.
Are these video's being written by AI? The tend to be getting more pretty, pretty, pretty gibberish, hyperbolic and filled TRULY silly hallucinations.... Sora stomps this on every level, from fidelity to consistency... not one did I need to double take on this video as to whether it might be real... the was not TRULY SHOCKING about this. I guess Wes is the last man standing of the first wave of successful AI channels that hasn’t gone total cringe.
I like AIExplained too.
If you can't notice that hallucinations are 10 times less than SORA, you must've pretty shitty eyes. Sora was unable to respect dimensions and the consistency was highly lacking, visual glitches and impossibilities were omnipresent and the respect to the prompt only worked for very short videos. Those problems have been greatly reduced.
You claim that it's laughably bad? Alright buddy. Ler's take one random video: The astronaut. What's clearly visible during your first watch which gives away the fact that it's AI?
@@user-cg7gd5pw5b nope... obviously your drinking the koolaid.
@@Vitaphone Then go ahead and tell me what I asked for. If you can't, your point is BS...
Bro you're attacking their videos but can't give examples of why. All I do is point it out. Who's really wrong here?
@@Vitaphone It feels like comparing Stable Diffusion 3 with Dall-E 3, just they have different styles, they're the same kind of model (Video Diffusion Transformer).
Except that the big slurp of noodles he ate had no residual sauce on them. The sauce from the bowl just teleported to his lips.
most likely scenario would be pre-training using video that closely matches, or even worse, is filmed specifically to match, the specific prompts. Without a release of any of these ai tools (chinese or not) there is no way to trust these announcements. In depth analysis if the marketing results is pointless.
The Chinese guys are missing a finger. And the cat one, the people on the sidewalk are walking backwards. It didn’t translate relational movement properly.
Not at all, you just look at that bicycle boy video, that's the reality(I think their mothed is basically cobble images together), however this generator proves scaling law works again.
Congratulatiins for de vide and for the fratures.
Why are you saying "genuamly @ 21.40? are you AI generated? :) Nice video
Maby thanks for the update 😊
My Ai-Asian isekai desire is the real thing. Hmmm case in point! That rabbit isn't myopic. I am enjoying the ultra effects. I am so happy. I don't have to "toy with fantasy" for the sake of your humanity. Strong Concept combination Ability is my pleasure now.
We are in an era similar to the one that invented steam engines.
Honestly, I hope we have access to some legacy software. Like the stuff that made the whacky will Smith eating spaghetti. That kind of mental stuff. Though I suppose you could just include that in the prompt, huh?
The clips look like a cell phone video. Very realistic.
I think it just better that more countries join, so it kills that America 1st. trend which lately has been taken over, companies first realising US and then only later to the rest of the world.
"China just went ahead and released"
Wrong. If it's not available to use, then It has not been released. You don't consider a video game to be "released" when it's in closed alpha or beta testing, do you? It's released when it's ready to be used by everyone.
Leon Ding helped them ?
"The FBI estimated in its report that the annual cost to the U.S. economy of counterfeit goods, pirated software, and theft of trade secrets is between $225 billion and $600 billion."
its not a cost its a Loss, but there's novway to tell if the losses would have occurred, apart from commercial licenses.
Alright.. A little bit of geography lessons. The red sunset looks really cool, but it's impossible to see it like that BOTH in front of you AND in the car mirror, because we only have one sun in our solar system. So yeah, the colors are indeed impressive, but the video can be seen as generated right in the first second when you watch it. And don't ChatGPT me now; there are some rare exceptions, but this is not it :)
That's true the clouds would be lit up too rather than shadows.
Let's keep doing game overs everyday.
Any software?
4- The emergence of AI and its necessities
X Mohammad Rahim Jamshidi
This is not about catching up. They just steal models and adapt them for themselves. This makes it so they are consistently just 1 year or so behind. We are not seeing the latest from Sora etc
I take it they have to answer any copywriting concerns.
Kuaishou is a short-video platform , rival of TikTok. But it mainly operates in China. So they can use abundant videos as reference to generate short clips
sorry but we are looking to different videos then, are you sure this is better than Sora? you still want to extract "extraordinary" from all your videos
I'm ELECTROCUTED. Singed and aromatic.
"Game over" for the 10 millionth time 🤪
Wow, I can't wait to not try it!
When will this be available for public use?
It already is unfortunately you need a Chinese phone number its only available on china but a couple of twitter accounts have been accepting prompts and showing the results
This uses PINNs with 3D spatiotemporal attention?
Remarkable!
Leon Ding sent China a lot of AI code.
I like your vids, but you seriously need an auto loop instead of constantly manually clicking play, kinda annoying
Where do i download this?
Is it possible to create multiple videos with same context or story?
That would be the real strength right there. Right now it's stupid memes or gimmicks like food turning into penguins.
Can we use Kling?
It doesn't beat SORA. It looks way worse. Sora is very smooth, doesn't have legs blurring with themselves.
Fake probably! And he makes this long video like it’s real with bunch of speculations without testing it himself
These are going to be interesting US elections…
It would be great if the Chinese trolled and showed real videos mildly processed to appear simulated and said they are AI 😄
BTW I play your videos at 1,75x speed and these demos look pretty pretty real.
Why can't runwayml or pika labs do anything lol they have a decent amount of money behind them
They should but the compute power to train a Video Diffusion Transformer seems higher than training a SOTA LLM
Well done.
What's next? 😂😂
Given how things can be faked in demonstrations, I wouldn't be surprised if there were real videos in here and told they were AI generated. An example being 9:35.
That is AI generated. Look at his left hand. His finger morphs around the bowl and you can see the wrinkles on his right hand fingers change when he moves the chopsticks.
Google just left the chat 😂😂 , no doubt china is not falling behind
It doesn't sounds weird to you to say "China" as if the entire country were behind this Chinese startup ?
@@TheRealUsername do you know how the communist party of china works?? They are behind every company that is in china.
@@TheRealUsername
China : "Ours" (communist meme)
Game over until the next time the game is over.
Truly truly Remarkably remarkable. Actually, Genuinely pretty pretty crazy. Trust me.
Is this even real? Can you use it now? UA-camrs will believe anything for a good post
where is the platform or app for this?
快影
快影
How would you know it wasn't Sora with a wrapper?
Sora is not publicly available, that's how.
It's like stating the Stable Diffusion 3 demos where made using Midjourney
Where is the European company?!😂
a sauce god talking about sauce
They create a company with my surname
Open AI, vs china, not Google!!!
We don't get the best from Google as that goes to the military n Govs watching us.
They're not AI generated videos, they're REALLLL :))
Is Hollywood ruined?
Don’t get me wrong it’s good, but idk if it beats sora tho.
What makes you say that?
You haven't even tested Sora, besides Sora is like the LLM of video generation, you can train your own if you have the compute, its training process is freely available in some research papers
But in reality, this is all speculation because no one on the public side has access to do a true comaparsions lets be real here.
Too much cherry picking I would like to see an independent evaluation side by side with Sora. The cat vid was not that good. And by side by side I mean the same input running them side by side
Cope better
No offense, but this is China we’re talking about. They have zero qualms about faking it all.
Nah its real this guy on twitter was asking people to comment what they wanted it to generate. And he got them all though sometimes the ai made mistakes and appearently it can't copy famous people either
meh, still a long road ahead for perfection
That's cool but don't care until I can use it. Sick of all this hype. I want results
it looks shit an incoherent in comparison to Sora. Obviously its running at a lower level of complexity.