Evolutionary Intelligence
Evolutionary Intelligence
  • 35
  • 67 574
Prolog Programming Basics
#prolog #firstorderlogic #logic #predicate #propositionallogic
Переглядів: 304

Відео

Fuzzy Logic
Переглядів 57Рік тому
Fuzzy Logic
First Order Logic
Переглядів 105Рік тому
Material for slides mostly taken from the book "Artificial Intelligence" by Stuart Russell and Peter Norvig
Big Data Technologies : MapReduce, Apache Hadoop, Spark and Kafka
Переглядів 9603 роки тому
For further reading: www.edureka.co/blog/top-big-data-technologies/ www.cs.amherst.edu/~ccmcgeoch/cs34/papers/p107-dean.pdf www.educba.com/hadoop-vs-spark/ phoenixnap.com/kb/hadoop-vs-spark www.educba.com/kafka-vs-spark/ www.educba.com/apache-kafka-vs-flume/ #DataScience #BigData #MapReduce #Spark #Apache #Hadoop #Kafka #ParallelProcessing #MachineLearning #DeepLearning #Petabyte #Exabyte #Zett...
Reinforcement Learning : Tic-Tac-Toe
Переглядів 24 тис.3 роки тому
#DataScience #ReinforcementLearning #TicTacToe
Principal Component Analysis (PCA) : Mathematical Derivation
Переглядів 3,1 тис.3 роки тому
PCA from intuitive perspective: ua-cam.com/video/cERNIfg9TLM/v-deo.html Python Code for PCA: scikit-learn.org/stable/modules/generated/sklearn.decomposition.PCA.html www.analyticsvidhya.com/blog/2016/03/pca-practical-guide-principal-component-analysis-python/ shankarmsy.github.io/posts/pca-sklearn.html PCA vs Linear Regression: shankarmsy.github.io/posts/pca-vs-lr.html #MachineLearning #PCA​​​ ...
Principal Component Analysis (PCA) : Intuitive Perspective
Переглядів 6263 роки тому
PCA Mathematical Details: ua-cam.com/video/ml1d1zC0jsg/v-deo.html Python Code for PCA: scikit-learn.org/stable/modules/generated/sklearn.decomposition.PCA.html www.analyticsvidhya.com/blog/2016/03/pca-practical-guide-principal-component-analysis-python/ shankarmsy.github.io/posts/pca-sklearn.html PCA vs Linear Regression: shankarmsy.github.io/posts/pca-vs-lr.html #MachineLearning #PCA​​​ #Unsup...
Gaussian Mixture Models
Переглядів 4693 роки тому
Python Code for GMM: www.analyticsvidhya.com/blog/2019/10/gaussian-mixture-models-clustering/ K-Means Clustering: ua-cam.com/video/92-h3_1Yfz8/v-deo.html Expectation Maximisation Algorithm: ua-cam.com/video/PNrghbfK5r0/v-deo.html Image Source: www.amazon.in/Pattern-Recognition-Learning-Information-Statistics/dp/0387310738 commons.wikimedia.org/wiki/File:ClusterAnalysis_Mouse.svg #MachineLearnin...
K-Means Clustering
Переглядів 4173 роки тому
Python Code Example: scikit-learn.org/stable/modules/generated/sklearn.cluster.KMeans.html Supervised Learning: ua-cam.com/video/DwqbxP5BDFo/v-deo.html Image Source: www.amazon.in/Pattern-Recognition-Learning-Information-Statistics/dp/0387310738 commons.wikimedia.org/wiki/File:ClusterAnalysis_Mouse.svg #MachineLearning​​​ #K-MeansClustering #UnsupervisedLearning​​​ #DataScience​​​ #GaussianMixt...
Expectation Maximisation Algorithm for Machine Learning
Переглядів 5343 роки тому
Bayesian Network: ua-cam.com/video/xJtyVQMV1A8/v-deo.html Code for python implementation: people.duke.edu/~ccc14/sta-663/EMAlgorithm.html #MachineLearning​​​ #ExpectationMaximisation #BayesianNetworks #ExpectationMaximization #GraphicalModels #ArtificialNeuralNetworks #DeepLearning #ANN #UnsupervisedLearning​​​ #DataScience​​​ #GenerativeModels
Probabilistic Graphical Models : Bayesian Networks
Переглядів 7 тис.3 роки тому
#MachineLearning​​​ #GraphicalModels #BayesianNetworks #ArtificialNeuralNetworks #DeepLearning #ANN #SupervisedLearning​​​ #DataScience​​​ #ExplainableAI #GenerativeModels #MarkovChain
Generative vs. Discriminative Models
Переглядів 1,2 тис.3 роки тому
Python code example for GAN: machinelearningmastery.com/how-to-develop-a-generative-adversarial-network-for-an-mnist-handwritten-digits-from-scratch-in-keras/ Bayesian Parameter Estimation: ua-cam.com/video/8P7tdwFF0is/v-deo.html Maximum Likelihood Estimation: ua-cam.com/video/WIPUh9yWM4c/v-deo.html Image source: www.flickr.com/photos/crazyeddie/4118740532/ #MachineLearning​​​ #GenerativeModels...
Artificial Neural Networks : Activation Functions and Optimization Algorithms
Переглядів 4463 роки тому
Activation functions: towardsdatascience.com/comparison-of-activation-functions-for-deep-neural-networks-706ac4284c8a Optimization Algorithms: arxiv.org/pdf/1609.04747 Loss Functions: towardsdatascience.com/understanding-different-loss-functions-for-neural-networks-dd1ed0274718 Vanishing Gradient Problem: towardsdatascience.com/the-vanishing-gradient-problem-69bf08b15484 #MachineLearning ​​​ #A...
Deep Learning : A Broad Overview
Переглядів 2963 роки тому
MNIST Dataset Analysis: yann.lecun.com/exdb/mnist/ #MachineLearning​​​ #ArtificialNeuralNetwork #DeepLearning #ANN #SupervisedLearning​​​ #DataScience​​​ #RepresentationLearning #Convolution #ConvolutionalNeuralNetwork #RecurentNeuralNetwork #ExplainableAI
Artificial Neural Networks : Solving ODEs and PDEs
Переглядів 3,7 тис.3 роки тому
Python Code for solving ODEs using ANNs: towardsdatascience.com/how-to-solve-an-ode-with-a-neural-network-917d11918932 Physics Inspired Neural Networks: www.sciencedirect.com/science/article/pii/S0021999118307125 Artificial Neural Networks : A Simple Introduction ua-cam.com/video/6nkylSKqaAc/v-deo.html Backpropagation algorithm: ua-cam.com/video/ntnwjWEpnkk/v-deo.html​ #DataScience​​​​ #Machine...
Artificial Neural Networks : Backpropagation Algorithm
Переглядів 8053 роки тому
Artificial Neural Networks : Backpropagation Algorithm
Artificial Neural Networks : A Simple Introduction
Переглядів 6473 роки тому
Artificial Neural Networks : A Simple Introduction
Random Forest : Comparison with Logistic Regression and Support Vector Machine (SVM)
Переглядів 3 тис.3 роки тому
Random Forest : Comparison with Logistic Regression and Support Vector Machine (SVM)
Decision Trees for Supervised Learning
Переглядів 5023 роки тому
Decision Trees for Supervised Learning
Support Vector Machine : Slack Variables and Nonlinear Kernels
Переглядів 1,1 тис.3 роки тому
Support Vector Machine : Slack Variables and Nonlinear Kernels
Support Vector Machine with Linear Kernel
Переглядів 1,3 тис.3 роки тому
Support Vector Machine with Linear Kernel
Logistic Regression for Multi-Class Classification
Переглядів 2 тис.3 роки тому
Logistic Regression for Multi-Class Classification
Logistic Regression for Binary Classification
Переглядів 1 тис.3 роки тому
Logistic Regression for Binary Classification
Linear Regression : Normal Equation & Regularisation
Переглядів 1,6 тис.3 роки тому
Linear Regression : Normal Equation & Regularisation
Supervised Learning : Gradient Descent & Regularisation
Переглядів 5283 роки тому
Supervised Learning : Gradient Descent & Regularisation
Maximum A Posteriori Estimation (MAP)
Переглядів 1,5 тис.3 роки тому
Maximum A Posteriori Estimation (MAP)
Maximum Likelihood Estimation (MLE)
Переглядів 5193 роки тому
Maximum Likelihood Estimation (MLE)
Bayesian Parameter Estimation
Переглядів 1,1 тис.3 роки тому
Bayesian Parameter Estimation
Supervised Learning : Spectrum of Algorithms, KNN Algorithm and the Curse of Dimensionality
Переглядів 5743 роки тому
Supervised Learning : Spectrum of Algorithms, KNN Algorithm and the Curse of Dimensionality
Naive Bayes : A simple algorithm for spam detection
Переглядів 1,4 тис.3 роки тому
Naive Bayes : A simple algorithm for spam detection

КОМЕНТАРІ

  • @JJ-fq3dh
    @JJ-fq3dh Місяць тому

    total waste of time

  • @androidterminal3924
    @androidterminal3924 2 місяці тому

    this is too general, please don't expect much from this. Don't watch if you have exam tomorrow :)

    • @Fandrii
      @Fandrii 2 місяці тому

      I have exam in next 30 minutes 😂😂

  • @aryanshrajsaxena6961
    @aryanshrajsaxena6961 2 місяці тому

    This channel is drastically underrated

  • @aryanshrajsaxena6961
    @aryanshrajsaxena6961 2 місяці тому

    Sir aap kamaal ho sir!!!

  • @ifgfqageneration6939
    @ifgfqageneration6939 3 місяці тому

    Its pronounced as Gowsen as in cow.

  • @ofelipegimenes
    @ofelipegimenes 4 місяці тому

    Excelent video! Thankyou so much!!!!

  • @ReversedFootage
    @ReversedFootage 5 місяців тому

    Excellent video, I learned a lot!

  • @user-ti8gx3on6b
    @user-ti8gx3on6b 6 місяців тому

    00:01 Unsupervised learning involves learning from unlabeled data 01:27 The video discusses reinforcement learning in Tic-Tac-Toe. 04:06 Understanding the state details and policy in reinforcement learning. 06:47 Importance of subscribing and value of association 08:57 Reinforcement learning involves updating the value of the current state based on the difference between the next and current values. 11:15 Computer learning to play Tic-Tac-Toe 13:19 The video discusses reinforcement learning in Tic-Tac-Toe 15:20 Reinforcement learning has various applications and benefits.

  • @RaviShankar-jm1qw
    @RaviShankar-jm1qw 9 місяців тому

    Very well explained. Thanks much :)

  • @kingdomofknowledge5960
    @kingdomofknowledge5960 Рік тому

    Waste of time and data

  • @sohamgosavi6797
    @sohamgosavi6797 Рік тому

    Does this formula derive from the Bellman's equation in some way? (Q(s,a)=R(s)+lambda×Q(s',a')) Where s and a are state and action corresponding to current state, s' and a' corresponding to the next state and R is the reward

  • @varshanhs7654
    @varshanhs7654 Рік тому

    What is the difference between Bayesian Network and Bayesian Belief Network?

  • @user-bu8mg7uq3s
    @user-bu8mg7uq3s Рік тому

    thanks

  • @prosenjitbanerjee6662
    @prosenjitbanerjee6662 Рік тому

    tough concept....yet to catch up completely

  • @furqanhasifi7014
    @furqanhasifi7014 2 роки тому

    bullshit Edited: Im joking

  • @willti123
    @willti123 2 роки тому

    I think you always make the wrong exmple for how an outlier screws up lr. The outlier has to be on the wrong side of the decision boundary. As you showed it, it poses no Problem for lr at All.

  • @Kishan31468
    @Kishan31468 2 роки тому

    Neural network for solving ODE is at very nascent stage. There is a lot to do. Numerical methods serve very well as of now. We don't need neural network for that. Can I have your linkedin ID.

  • @vascoabreu412
    @vascoabreu412 2 роки тому

    Does anyone have the code for this tic tac toe example?

  • @whoosshhaa
    @whoosshhaa 2 роки тому

    Very well explained !! Thank you 👍

  • @sanaullahsaqib2168
    @sanaullahsaqib2168 2 роки тому

    Why we use ANNs on ODEs. What is difference to get Solution from numerical methode and also get the solution from the ANNs. Why we use ANNs ?

  • @Rangarang
    @Rangarang 2 роки тому

    it was effictive

  • @beizhou4025
    @beizhou4025 2 роки тому

    How do we know if the RL algorithm converges on the TicTacToe? Which criteria should we look into?

    • @rotviepe
      @rotviepe Рік тому

      Assuming the opponent is a perfect player, then we could say our RL agent has converged to the optimal policy if all of the games end in a tie

  • @VijayChakravarty-pz4qv
    @VijayChakravarty-pz4qv 3 роки тому

    In the example used, we had the structure of the Bayesian network given and then we could estimate the parameters using BPE. I have read that we can do three things in Bayesian Networks...parameter learning, structure learning and latent structure learning. How do we go about predicting the structure of the Bayesian Network if the parameters are given instead?

  • @VijayChakravarty-pz4qv
    @VijayChakravarty-pz4qv 3 роки тому

    Is there a threshold that the generator aspires to achieve in defeating the discriminator? is it possible that such a threshold might be unachievable in certain cases?

    • @EvolutionaryIntelligence
      @EvolutionaryIntelligence 3 роки тому

      The threshold is determined by the human programmer and of course, it can be unachievable depending on how high its set.

  • @AayushMishra-pt4yx
    @AayushMishra-pt4yx 3 роки тому

    Can PCA be used for data compression like K-Means clustering? If it can be used, then which of these two will lead to less loss of data?

  • @aishaummer5817
    @aishaummer5817 3 роки тому

    simply what does it mean by training a network???plss answer...

    • @EvolutionaryIntelligence
      @EvolutionaryIntelligence 3 роки тому

      Watch this video on the Backpropagation algorithm : ua-cam.com/video/ntnwjWEpnkk/v-deo.html

  • @RahulSharma-tf6zk
    @RahulSharma-tf6zk 3 роки тому

    Sir, do we always get real eigenvectors of sx? if we get imaginary, a possible workaround exists?

    • @EvolutionaryIntelligence
      @EvolutionaryIntelligence 3 роки тому

      If your input features are real, Sx is a real symmetric matrix and so has real eigenvalues and eigenvectors.

  • @NamanJoshi-pi4bn
    @NamanJoshi-pi4bn 3 роки тому

    17:32 Temperature is not playing any role in decision making here, does this mean that temperature will also not play any role if the decision is being made in real life by a human? or is this limited to this algorithm.

    • @EvolutionaryIntelligence
      @EvolutionaryIntelligence 3 роки тому

      It is of course limited to this algorithm! Human beings do not always follow strict logic, thankfully!

  • @VatsalyaSharan
    @VatsalyaSharan 3 роки тому

    Is it possible to use K-Means Clustering for classification? Given the number of Clusters, can we assign a subset of clusters into a given class and then maybe use this way for classification, even though this is not the best form of classification we have available.

  • @NamanJoshi-pi4bn
    @NamanJoshi-pi4bn 3 роки тому

    Can robot be an example of "nearly" strong AI, humans take input via their sense organs, eye, ear, nose, skin and tongue and the data is in form of images, sound, smell, touch/texture and taste. If a robot is able to take data in these format and learn from it and can be rewarded for doing right things and penalised for doing wrong things (for eg: a baby feels pain when he touches fire, therefore learns that he shouldn't touch it), then we can say that robot is an example of strong AI? Do the researchers know what should be the "reward" or "penalty" to a robot?

    • @EvolutionaryIntelligence
      @EvolutionaryIntelligence 3 роки тому

      Not at all!! No matter how complex a robot gets, it still has zero self-awareness. Making Strong AI is going to require a fundamental shift in our approach to AI and currently we have absolutely no clue how to get there.

  • @sidchaini
    @sidchaini 3 роки тому

    What is the reason for calling this the "Expectation Maximisation" Algorithm? Which expectation is being maximised?

    • @EvolutionaryIntelligence
      @EvolutionaryIntelligence 3 роки тому

      Check the video on Gaussian Mixture Models : ua-cam.com/video/he4lI0w1G-g/v-deo.html

  • @siddharthsethi7773
    @siddharthsethi7773 3 роки тому

    Sir can you explain a more about Space through model space? I didn't understand.

    • @EvolutionaryIntelligence
      @EvolutionaryIntelligence 3 роки тому

      Thats a very complicated procedure! Check this out: link.springer.com/article/10.1007/s13748-019-00194-y

  • @varundhankhar30
    @varundhankhar30 3 роки тому

    If all the eigen values are same then out of them, which one would be the principal component?

    • @EvolutionaryIntelligence
      @EvolutionaryIntelligence 3 роки тому

      Thats a very unlikely situation, but if it occurs, PCA is clearly not going to be of any use.

  • @siddharthsethi7773
    @siddharthsethi7773 3 роки тому

    Does Probabilistic Graphical Model have some connection with Decision Trees as if we can use Probabilistic Graphical Model to find which data set plays an important role in predicting the output same as in Decision Trees where we find Entropy and Information gain. Are these two somewhere related?

    • @EvolutionaryIntelligence
      @EvolutionaryIntelligence 3 роки тому

      The only connection is that both use boxes and arrows! In Decision Trees, there is no notional of conditional dependence. You just start from the root node and follow the appropriate leaves to get to your conclusion. But Probabilistic Graphical Models are used to model conditional probabilities and are generative in nature, whereas Decision Trees are purely discriminative.

  • @siddharthsethi7773
    @siddharthsethi7773 3 роки тому

    Is GANs only used in binary classification or it can also be used in multi class classification? Also won't the computation will increase if we do multi class Classification using GANs.

    • @EvolutionaryIntelligence
      @EvolutionaryIntelligence 3 роки тому

      I don't see any reason why GAN cannot handle multi class classification. Of course, computational requirements will increase.

  • @TanishqKumarBaswal
    @TanishqKumarBaswal 3 роки тому

    As stated in post pruning method, the removal of the nodes is done randomly or there is some method to select which node has to be removed. Is there a loss of information.(suppose we remove the 'wind' node from the optimal decision tree stated in the video, does the new decision tree will halt on encountering wind feature)

    • @EvolutionaryIntelligence
      @EvolutionaryIntelligence 3 роки тому

      Pruning the nodes randomly is one option, but then you have to keep track of how it impacts your validation accuracy. So if you choose to remove one node randomly, and if it reduces your validation accuracy, then you put that node back. Another method is to do pruning using information gain. If a particular node has a very low information gain, then you prune that node.

  • @ShirshakkPurkayastha
    @ShirshakkPurkayastha 3 роки тому

    Can Reinforcement Learning be considered as the ML analogy of Closed-loop control systems (because both take into account the feedback of previous state)?

    • @EvolutionaryIntelligence
      @EvolutionaryIntelligence 3 роки тому

      In conventional control systems, there is no learning involved. But a lot of modern day control systems use Reinforcement Learning for automatic control.

  • @ShirshakkPurkayastha
    @ShirshakkPurkayastha 3 роки тому

    Is a linear relationship between all variables required for PCA?

    • @EvolutionaryIntelligence
      @EvolutionaryIntelligence 3 роки тому

      The variables don't need to have a linear relationship with each other. What PCA does is to find a linear combination of the variables which gives you the new coordinate system.

  • @ShuhulHandoo-md9vm
    @ShuhulHandoo-md9vm 3 роки тому

    Can reinforcement learning be used for feature extraction? And to how much extent is it better than any other approaches?

  • @ShuhulHandoo-md9vm
    @ShuhulHandoo-md9vm 3 роки тому

    Does there exist any trade-off between model complexity and dimensionality of the data while using PCA?

    • @EvolutionaryIntelligence
      @EvolutionaryIntelligence 3 роки тому

      Not much, except that if the dimensions are very high and dataset is relatively small, then you find the eigenvalues and eigenvectors for a slightly different matrix as explained in this video : ua-cam.com/video/ml1d1zC0jsg/v-deo.html

  • @VatsalyaSharan
    @VatsalyaSharan 3 роки тому

    The example given in the video contained 3^9 possible states. In real world applications like these, these numbers could go very very high. So instead of going through each and every possibilities which might be very time and energy consuming, is it possible to model our approach in a way which takes into consideration only those samples/states which are much more likely to happen and then proceed to do Re-enforced learning from there?

    • @EvolutionaryIntelligence
      @EvolutionaryIntelligence 3 роки тому

      Yes, for problems with large number of states, sampling the entire state space is one option and is done using Monte Carlo methods which are also very popular in other areas of science and engineering. Another option is to use Deep Learning which learns intelligent ways to represent the states and thereby reduce the effective size of the state space.

  • @ShivanBhatt-sx5ib
    @ShivanBhatt-sx5ib 3 роки тому

    during training, does the model learn values for every possible state?

    • @EvolutionaryIntelligence
      @EvolutionaryIntelligence 3 роки тому

      For low number of states, that is surely a possibility, but fails to work when the state space is very large, in which case it is generally advised to use Deep Learning which learns intelligent ways to represent the states, thereby reducing the effective size of the state space.

    • @EvolutionaryIntelligence
      @EvolutionaryIntelligence 3 роки тому

      @Prashant Shukla 16139 Tic-Tac-Toe is too small a problem to appreciate this. Think of a game like Go, where the number of possible states is too large and regular reinforcement learning would just take forever to learn! Check this for a detailed explanation : nikcheerla.github.io/deeplearningschool/2018/01/01/AlphaZero-Explained/

  • @VaibhavSingh-iz4bf
    @VaibhavSingh-iz4bf 3 роки тому

    Why did we stop at a linear combination of the inputs and not explore combination of higher powers?

    • @EvolutionaryIntelligence
      @EvolutionaryIntelligence 3 роки тому

      That is just variable transformation and does not impact the PCA technique itself.

  • @lucaswehmuth
    @lucaswehmuth 3 роки тому

    How did you get the value of P(W = T) ?

    • @EvolutionaryIntelligence
      @EvolutionaryIntelligence 3 роки тому

      By summing over the relevant conditional probabilities.

    • @lucaswehmuth
      @lucaswehmuth 3 роки тому

      @@EvolutionaryIntelligence Which are?

    • @EvolutionaryIntelligence
      @EvolutionaryIntelligence 3 роки тому

      @@lucaswehmuth By using the chain rule of probability theory, we can write: P(C, S, R, W) = P(C) * P(S|C) * P(R|C,S) * P(W|C,S,R) This can be simplified by using the conditional independence relationships described in the video through the graph: P(C, S, R, W) = P(C) * P(S|C) * P(R|C) * P(W|S,R) Now the numerator in the expression at time 17:14 is obtained by summing over various possibilities of R and C. For the denominator, you need to additionally sum over the possibilities of S as well.

  • @TanishqKumarBaswal
    @TanishqKumarBaswal 3 роки тому

    Should we use the primal or the dual form of the svm problem to train a model on a training set with millions of instances and hundreds of features

    • @EvolutionaryIntelligence
      @EvolutionaryIntelligence 3 роки тому

      If your data is not linearly separable, then you have to use the kernel trick, in which case you have to use the dual form. No choice! For data that is linearly separable, there are some quadratic programming algorithms that can solve the dual problem faster than the primal formulation. So in general, I think you should use the dual formulation, unless there is a very specific reason to use the primal form.

  • @ShirshakkPurkayastha
    @ShirshakkPurkayastha 3 роки тому

    Is it predetermined to the Discriminator that the Generator always produces fake data (is the Discriminator biased into believing that Generator produces fake data)?

    • @EvolutionaryIntelligence
      @EvolutionaryIntelligence 3 роки тому

      The Discriminator has no clue where the images are coming from. It only gets an image as input and has to figure out whether its real or fake.

  • @GopikaSR
    @GopikaSR 3 роки тому

    How to know how many times we need to do exploitation and exploration to gain an optimal solution for a particular dataset? Rather than randomly deciding how many times we need to do exploitation and exploration, is there a fixed way to find the which combination or which selection works the best to find the optimal solution.

    • @EvolutionaryIntelligence
      @EvolutionaryIntelligence 3 роки тому

      There is no such fixed way and one has to do trial and error for each given problem.

    • @GopikaSR
      @GopikaSR 3 роки тому

      @@EvolutionaryIntelligence Alright. Thank you sir.

  • @koushiksrinivasula3584
    @koushiksrinivasula3584 3 роки тому

    Sir, If I train an RNN for a task and in it if I add this concept of reward from one state to another, is that considered to be reinforcement learning.

    • @EvolutionaryIntelligence
      @EvolutionaryIntelligence 3 роки тому

      Whether it is considered reinforcement learning or supervised learning depends on the actual task. In supervised learning, the algorithm learns to map inputs to outputs. In reinforcement learning, the algorithm learns to perform a certain task (eg. navigating an environment or playing a game). Neural Networks are surely used for Reinforcement Learning, and the combination is called Deep Reinforcement Learning: www.baeldung.com/cs/reinforcement-learning-neural-network

    • @koushiksrinivasula3584
      @koushiksrinivasula3584 3 роки тому

      @@EvolutionaryIntelligence oh! Thank you sir

  • @koushiksrinivasula3584
    @koushiksrinivasula3584 3 роки тому

    So, are the values of the states determined recursively, after reaching the end state? like if that is the case isn't that similar to training our weights and biases for each node in ANN. But yes, I do understand that here we directly deal with different states.

    • @EvolutionaryIntelligence
      @EvolutionaryIntelligence 3 роки тому

      At the fundamental level, it surely boils to function approximation, but the application, approach and models are different. It's like we use the same language for all our communication. The basic words and grammar remains the same and what changes is the outside form.

  • @VatsalyaSharan
    @VatsalyaSharan 3 роки тому

    Is there any way to decrease the correlation by certain design choices of our model Or are these correlations completely determined by the datasets we use?

    • @EvolutionaryIntelligence
      @EvolutionaryIntelligence 3 роки тому

      These correlations are determined by the datasets, but you can compute them incorrectly if you are using the wrong model. And thats why model selection becomes very important in these problems.

    • @EvolutionaryIntelligence
      @EvolutionaryIntelligence 3 роки тому

      @Prashant Shukla 16139 Bayesian Networks have no relation with Naive Bayes classifier. If the graph structure of a Bayesian Network is not known apriori, estimating it from data is an NP-hard problem and you could look at this paper for some of the useful algorithms: link.springer.com/article/10.1007/s13748-019-00194-y