Sam Sartor
Sam Sartor
  • 2
  • 99 900
AI Generated Art (Journal Club Presentation)
This is a presentation I gave on October 6th 2022 at the William & Mary Graduate Student Association's weekly Journal Club.
I talk about the past, present, and future of computers spitting out pretty pictures. Specifically how computers generate art and the extent to which this is an extension of human creativity. This includes recent AI projects like DALL-E and Stable Diffusion, but also more traditionally generated art: fractals, minecraft worlds, and the "wave function collapse" algorithm.
Demo Links:
- There are countless ways to run Stable Diffusion for free. For this presentation I set up github.com/AUTOMATIC1111/stable-diffusion-webui.
- lexica.art is an archive of previously made AI art and the corresponding prompts.
- dreamfusion3d.github.io uses Stable Diffusion to make 3D art.
- oskarstalberg.com/Townscaper is an online demo of the Townscaper game.
Переглядів: 626

Відео

Why neural networks aren't neural networks
Переглядів 99 тис.3 роки тому
There is a better way to understand how AIs sort data, process images, and make decisions! Made for the 2021 Summer of Math Exposition: www.3blue1brown.com/blog/some1 Source code available here: gitlab.com/samsartor/nn_vis The background music is an excerpt of the endless ambient generative music system "At Sunrise," available at generative.fm/music/alex-bainter-at-sunrise

КОМЕНТАРІ

  • @skye-sys
    @skye-sys 4 місяці тому

    The youtube algorithm blessed me today what a video!

  • @ikartikthakur
    @ikartikthakur 4 місяці тому

    I am learning new things each day ..thanks man , for sharing this perspective . I never thought of transforming the data , I was used to rotating the lines..! 👍

  • @avadhsavsani1148
    @avadhsavsani1148 5 місяців тому

    Hi there, this idea of looking at what the deep learning model is doing internally is really intuitive. I wanted to know, if there is any underlying connection between Transformations and Fine tuning the weights and biases with Backpropogation.

    • @samsartorial
      @samsartorial 5 місяців тому

      In the sense that fine-tuning/training is how you work out what the transformations (linear weights + biases) should be, yes! But I pretty much ignored training with backprop in this video because there are already really good explanations of it on UA-cam, and I just wanted to explain what the model itself actually consists of.

  • @DoYouHaveAName1
    @DoYouHaveAName1 6 місяців тому

    AMAZING! The animations and explanation are perfect! Thank you for your hard work

  • @ShantanuSingh-wc4ou
    @ShantanuSingh-wc4ou 10 місяців тому

    Please keep sharing your knowledge with us, you are just awesome, and we love your content.🙏

  • @pacukluka
    @pacukluka Рік тому

    Amazing ! Please keep it up, we have so much yet to learn.

  • @sebastianp4023
    @sebastianp4023 Рік тому

    brilliant!!!

  • @barbaricmelons
    @barbaricmelons Рік тому

    Even in the face of the almost insurmountable and fascinating amount of "ummmmmmm" during the speech...this is an incredible breakdown. Fantastic job! Despite the internet's data being hijacked for commercial purposes (images/personal/location) it seems like similar to old high school typing classes the ramification to what you're becoming a part of from just posting something on Reddit, Facebook, or getting your first smartphone (despite how cool it is) needs to be explained to users in almost a classroom setting because it's becoming so overwhelming to your average user how their data is being extrapolated. I always think about my mom saying she took a typing class in college and her family asking why she would need to learn how to type, the same question may arise from asking why would you need to know how AI works. Granted it's a different time, but relatable. I feel like 98% of people that would randomly watch this video would be completely blown away or at least ignorant to how this technology is evolving. As an artist I'm still adding new plugins or going to certain apps for certain functionality (for my real job, not just making dumbass cartoons on UA-cam) but I still feel like I'm lagging behind. Additionally, I hope there's a way to explain that the technology wielded at 39:00 is comfortably used by certain groups of people based on a morally loose justification, and that morally loose justification is leading to more and more questionable places. And generally like your explanation of denoising a photograph or a render engine, it's almost impossible to escape, and I say that using voice recognition because I'm not physically able to type this comment.

  • @incredulouschordate
    @incredulouschordate Рік тому

    This is seriously the best NN video I've seen, and yes that's after watching 3B1B's series

  • @BlueBirdgg
    @BlueBirdgg Рік тому

    Great video (and I am on ~3min onto the video)! Ty

  • @geraldkenneth119
    @geraldkenneth119 Рік тому

    From this perspective an ANN resembles an SVM with a learned kernel

  • @ZubairKhan-sp8vb
    @ZubairKhan-sp8vb Рік тому

    Just awsome!!! I have been thinking about neural networks this way, and how the key word has stuck with this statistical process. You illustrated beautifully just amazing!!

  • @hannyaespinoza5111
    @hannyaespinoza5111 Рік тому

    wow, you are so talented, man!

  • @Axl_K
    @Axl_K Рік тому

    It's all just Math!

  • @ming3706
    @ming3706 Рік тому

    I can't believe my highschool math is this important

    • @dabidmydarling5398
      @dabidmydarling5398 Рік тому

      I thought the same thing as well! I love ML and statistics!! :)

  • @tiagoverissimo1404
    @tiagoverissimo1404 Рік тому

    clean

  • @y.8901
    @y.8901 Рік тому

    Hello, what is at 6:55 in the 3d plot, is it what's in 1 node of the hidden layer or the 3 nodes ? Because 1 node does not make sense for me, we activates these 3 nodes independently, and we do the sum of these 3 nodes after that, which is shown at 7:03 What does it means at 6:54 "The three new numbers are nonlinearly altered" ?

    • @samsartorial
      @samsartorial Рік тому

      Because the middle layer has 3 nodes, the plot is 3 dimensional (to visualize the 3 x,y,z values for every sample). When the network has 2 nodes, all the points are on a plane. And when it has one node, all the points are on a line. I did mess with the basis vectors a bit to improve clarity. The nonlinearity in this example is the tanh function, which is applied to each value in-between the linear transforms.

  • @aren6
    @aren6 Рік тому

    brilliant video

  • @essendossev362
    @essendossev362 Рік тому

    absolutely beautiful visualizations

  • @TUMENG-TSUNGF
    @TUMENG-TSUNGF Рік тому

    Does one transformation map to one layer in the NN? I can see that if the input is in 2D and to transform it into a representation in 3D it corresponds to a layer that takes two input and outputs 3. But if several transformations are from 2D to 2D can they be encapsulated into one layer or do you have to do them one after another (aka in diff layers)?

    • @samsartorial
      @samsartorial Рік тому

      That's a really good question! Sadly, it doesn't have a very clear answer. Different people draw the borders between layers in different places. Traditionally, the nonlinearities inside a neural network are basic functions, and are exactly the same for every layer. For example, the linear transformations in this video are all completely different, but each is punctuated by a tanh function. If you have 10 2D linear transformations, each followed by tanh, that would count as 10 layers. Sadly it isn't usually that simple. Although a matrix multiply followed by tanh makes perfect sense when transforming data through 2, 3, or even 100 dimensions, a modern image model might transform the data across a hundred million dimensions at a time. That requires more specialized kinds of linear transformations, sometimes grouped together in odd ways. For example, I'm currently working with a network that repeatedly transforms the data through a depthwise 7x7 linear layer, a normalization layer, a 1x1 linear layer, a GLEU nonlinear function, another 1x1 linear layer, and a final skip connection. You could argue that those all amount to just one very convoluted linear transform punctuated with the single GLEU nonlinearity. But since the individual parts are important for performance, reliability, and accuracy, they are also called "layers" and the whole combination is more often called a "block". A ConvNeXt block to be specific. 6 layers per ConvNeXt block, 12 ConvNeXt blocks per ConvNeXt backbone, 2 backbones in the actual U-Net, and some miscellaneous stuff on the side.

  • @MathForLife
    @MathForLife Рік тому

    Amazing video! Thank you.

  • @weirdsciencetv4999
    @weirdsciencetv4999 Рік тому

    This is top notch! Hope you make more educational videos

  •  Рік тому

    But the process is similar and we do get something similar to what our brains do, not coincidentally.

  • @iwanttwoscoops
    @iwanttwoscoops Рік тому

    came from your other video. Thanks for posting this

  • @lucasgivens3750
    @lucasgivens3750 Рік тому

    Why isn't it equivalent? What do neurons do differently?

  • @mms7146
    @mms7146 Рік тому

    I loved the video! Just one question though: why not simply use a GLM instead of doing all that complicated data transformation?

  • @tooljack4439
    @tooljack4439 Рік тому

    learned a lot. great presentation Sam.

  • @kamalemelkemi6066
    @kamalemelkemi6066 Рік тому

    help to explain the success of the deep Neural Nets.

  • @robkim55
    @robkim55 Рік тому

    A different look at the problem that is refreshing.

  • @davidg8497
    @davidg8497 2 роки тому

    Thanks for the video. What I would add, you "reduce" the neural network to "simple statistics". And you explain very trivial examples of decision making (very nice) that should show us that AI is not real intelligence. But: Neural networks are a "complete basis" of any function between two vector spaces, so any information processing that can be done can be done with a neural network. At 8:23 you show what artificial networks ignore, but all these things don't give us anything anymore. In the sense that a new way of processing information is added by these things. The reason for this is that NN are complete, we cannot "add" anything. These points at 8:23 could make a NN simulating the brain's information processing more complex, but not impossible in itself! To me, the blue beer-strawberry example shows how a very low level of intelligence works, say an IQ of 0.1, but it's not "proof" that AI isn't a form of intelligence. And you can see it different: Take away from a human being all its ability to make "simple" decisions and process information that a function of any kind could. What would remain would be pure "human intelligences", but what is that? hp even a quantum computer adds nothing to what can be computed with a universal Turing machine, it is only faster!

  • @Sydra.
    @Sydra. 2 роки тому

    You made the world a better place with this video!

  • @nembobuldrini
    @nembobuldrini 2 роки тому

    Like n. 7001! ;) Brilliant! The best visual explanation of NN I have encountered so far. The visuals are extremely helpful in getting the gist of what a feed-forward neural network does. It's important to point out - and this would have spiced up the ending of the video too ;) - that there are other types of neural networks that are more similar to how the brain works. Hebbian learning and recursion are involved in these other types of neural networks, for which a simplification in the terms used in the video would be not so quite straightforward. It would actually be great to see a follow-up video on these kind of NN!

    • @diadetediotedio6918
      @diadetediotedio6918 Рік тому

      They are also, extreme simplifications of the processes that occur in a brain and, fundamentally, they end up being used as if they were the current "artificial neural networks" when they depend on statistical methods such as backpropagation.

  • @stevengill1736
    @stevengill1736 2 роки тому

    Thank you! I didn't have the vocabulary to say this when trying to explain this to people...

  • @eriksaari4430
    @eriksaari4430 2 роки тому

    nobody has claimed they would be. just brain inspired. which is quite amazing

  • @johnathan8174
    @johnathan8174 2 роки тому

    ƤRO𝓂O𝕤ᗰ

  • @alivecoding4995
    @alivecoding4995 2 роки тому

    Wonderful! Thanks so much!

  • @mrnobody.4069
    @mrnobody.4069 2 роки тому

    We've been lied to! But how is AI even considered an intelligence!?

    • @samsartorial
      @samsartorial 2 роки тому

      Currently AIs are "narrow" intelligences, meaning they make specific sorts of predetermined decisions. Humans are "general" intelligences. I think we'll have general AIs in the next decades, but that is hotly debated. Edit: also "lied to" is going a bit far lol. I just think there are better metaphors.

    • @mrnobody.4069
      @mrnobody.4069 2 роки тому

      @@samsartorial Humm agreed but I think we should also research AI"s without current networks with what works best and is more efficient.

  • @igormorgado
    @igormorgado 2 роки тому

    no. translation isn't a linear transfomation. We call linear transformation a function that preserves T(xa+yb) = xT(a) + yT(b).

    • @samsartorial
      @samsartorial 2 роки тому

      That's true! The technical term is in an affine transformation. As I've mentioned before, I wanted to avoid overloading the viewer with terminology given neural networks can always replicate affine transformations given an extra linear dimension. We tend to include it anyway just because it makes optimization easier.

  • @zeten___9908
    @zeten___9908 2 роки тому

    4:20 why can't you just use a parabolic function?

    • @samsartorial
      @samsartorial 2 роки тому

      Thank you for asking this question! I learned a lot while trying to answer it. You could absolutely fit a single parabola to that particular example. But the goal is to find a process that can draw a decision boundary through any data set, no matter how complicated. We call any such process a "universal approximator". I thought that almost any sufficiently long sequence of linear and nonlinear transformations was capable of universal approximation, and that parabolas weren't used as nonlinear transformations for some practical reason. They extrapolate badly? But looking into it further, Allan Inkus proved in 1999 that polynomial transformations aren't sufficient. Our nonlinear function has to exponential or trigonometric, or piecewise, or something else similarly weird. Sigmoid functions like tanh were the first to be proved as universal, and I used them in this video because they looked nice, but these days everyone uses a piecewise function called RELU.

  • @sedevacantist1
    @sedevacantist1 2 роки тому

    I find it intellectually dishonest to draw two circles, call them neurons, then draw a line between them and call it a neural link network, and artificial intelligence. So far the majority use of this so-called artificial intelligence is to steal our humanity, to sort us like berries to take advantage, and control us. To call this computer process intelligence is an insult to our real intelligence. Am I to ooh-and-ahh over this emerging evil?

  • @NhanTrann
    @NhanTrann 2 роки тому

    hope your PhD is going well!!

  • @iestynne
    @iestynne 2 роки тому

    What a *fantastic* video! I've watched a large number of videos on artificial neural networks over the past few years... yet I learned such a lot from this one! Such a (shockingly) clean perspective on how these systems work. The choice of the examples and the clarity of the writing and animation are just superb. If you didn't win, it's a travesty.

  • @PedroTricking
    @PedroTricking 2 роки тому

    There's a chance that after I watch this video I will think that neural networks are more like neural networks than I initially did.

  • @chadx8269
    @chadx8269 2 роки тому

    Thanks for debunking the "Neural Network".

  • @Marc-tm4xh
    @Marc-tm4xh 2 роки тому

    This video blew my goddamn mind. Please make more 😃

  • @ZuvielDrama
    @ZuvielDrama 2 роки тому

    Awesome, thank u for that masterpiece.

  • @benjaminpierce8835
    @benjaminpierce8835 2 роки тому

    Also the people in these comments are definitely lying, this would literally be a small part of the first day of a ML course and just at par level at best, they never took a ML course.

  • @benjaminpierce8835
    @benjaminpierce8835 2 роки тому

    You make a big statement and then very little to back it up. You gave a great explanation and argument for why current neural networks could model anything that a brain could produce but then your takeaway is that because it doesn't use the chemical interactions that you just argued where unnecessary, that it can't learn. Looks like you understand the concept but not fully. What about these chemical and dendritic processes do you believe would cause behavior that is impossible to model with neural networks?

    • @samsartorial
      @samsartorial 2 роки тому

      I never said that behavior in the brain is impossible to model with neural networks. The basic theory of NNs tells us that any dendritic process is modelable by a deep enough network. The thesis of this video was "the biological metaphor confuses the idea of neural networks as iterated linear and nonlinear transformations derived through a statistical process". If neural networks are fundamentally emulations of nature, then what is the point of multi-head attention? What about the SPADE block used in img2img tasks? What the heck is it emulating? Neural networks absolutely _could_ emulate nature, but they don't _have to_ emulate nature. When we finally create a model that can truly learn from its experiences, it doesn't have to do it in the same way nature does.

    • @diadetediotedio6918
      @diadetediotedio6918 Рік тому

      I can do a even bigger claim, it is not possible to simulate even a single neuron on a digital computer. Neurons are physical, real, beings, they change their shapes and they create magnetic fields, and there are good theories out there that even state tha they interact directly with the quantum nature of our universe. So good luck modeling that.

  • @randomNunber
    @randomNunber 2 роки тому

    Very interesting, viewing the content in another directions

  • @classiccj2569
    @classiccj2569 2 роки тому

    I don't know why the Avocado Armchair made me laugh so much, I just did not expect it I guess