2. Introduction to Statistics (cont.)

Поділитися
Вставка
  • Опубліковано 1 лют 2025

КОМЕНТАРІ • 134

  • @aungkyaw9353
    @aungkyaw9353 4 роки тому +21

    After reviewing probability courses and this one itself many times, I found this lecture to be a great one. Some vague knowledge in real analysis will help. But I still need to come back later after digging deep into other sound footings for foundation. Thank MIT and professor.

  • @mitocw
    @mitocw  7 років тому +40

    The first lecture was not recorded. We are hoping to record it when the course is taught again in the Fall. In the meantime, please check out the course site here ocw.mit.edu/18-650F16 where you will find the slides that were used in all of the lectures.

    • @MuhammadGhufran5
      @MuhammadGhufran5 7 років тому +3

      Hi! thank u very much for the recordings !
      First part of this lecture (lecture 2) is missing and thats why its really difficult to catch up with the rest of the lecture ! It would be really helpful if u guys could upload the newest version of this lecture

    • @aldousmehan8186
      @aldousmehan8186 4 роки тому +8

      Even though there is now a lecture 1 recording from 2017, it doesn't cover the same material as the missing lecture 1 in 2016 had covered (the Central Limit Theorem, this quantity q=alpha/2), so anyone trying to learn from this will have to do a lot of their own legwork.

  • @hxxzxtf
    @hxxzxtf Рік тому +19

    🎯 Key Takeaways for quick navigation:
    00:00 📊 *The speaker discusses the central limit theorem and its application to estimating proportions using averages.*
    02:19 🔄 *The central limit theorem implies that the distribution of square root of n times the average converges to a standard normal random variable.*
    03:16 🔄 *The probability that a Gaussian random variable exceeds q alpha over 2 is equal to alpha.*
    05:08 📉 *Confidence intervals are built by transforming the estimate in a way that results in a pivotal distribution, not dependent on unknown parameters.*
    07:28 🧠 *The concept of pivotal distributions is introduced to create estimates that are asymptotically independent of unknown parameters.*
    13:33 📉 *Hoeffding's inequality provides a useful tool for bounding the probability that the sample average deviates from its expectation for any n.*
    16:47 📉 *Solving Hoeffding's inequality for a specific case shows a method for constructing confidence intervals without the need for large sample sizes.*
    19:43 📈 *The Hoeffding inequality's worst-case scenario is discussed.*
    20:12 🔄 *Combining intervals to determine the probability of a variable being outside a specified range.*
    21:34 🤔 *Comparing the margin square root of log 2 over alpha divided by 2n to q alpha over 2/3n.*
    22:55 🎲 *The role of assumptions in statistics, and the importance of balancing assumptions for confident statements.*
    23:25 🔄 *Different types of convergence in statistics, emphasizing convergence in distribution as crucial.*
    24:51 🔄 *Convergence in distribution explained through the concept of probability computations on random variables.*
    26:17 🧐 *Stronger conditions needed for convergence in distribution to allow combining random variables effectively.*
    27:42 🔄 *Almost sure convergence and convergence in probability explained, highlighting their differences.*
    29:06 📉 *Convergence in Lp and its relation to the weakening of convergence conditions.*
    31:27 🔄 *The importance of the characteristic function in proving convergence in distribution, especially in the central limit theorem.*
    33:19 🔄 *Equivalence between convergence in almost sure, convergence in probability, and convergence in distribution.*
    38:35 🔄 *The continuous mapping theorem: if Tn goes to T, then f(Tn) goes to f(T) for continuous functions f.*
    39:32 📊 *Convergence in probability involves decreasing index values implying convergence, eventually leading to convergence in distribution.*
    40:29 🔄 *Operations and limits differ between convergence almost surely and convergence in probability, allowing various manipulations for the former.*
    41:55 📈 *Convergence in distribution criteria includes the convergence of characteristic functions or bounded continuous functions of the random variable.*
    42:24 🔄 *Slutsky's theorem states that if one random variable converges in probability and another in distribution, certain operations are still valid.*
    43:23 🚇 *Inter-arrival times in queuing theory, modeled by exponential distribution, are useful in systems with memoryless properties.*
    45:16 📉 *Exponential distribution is often employed in modeling positive random variables, such as inter-arrival times.*
    48:39 🔄 *The average of inter-arrival times in queuing theory, denoted by Tn bar, serves as an estimator for 1 over lambda, the rate parameter.*
    50:32 🔄 *Strong and weak laws of large numbers support the convergence of Tn bar to 1 over lambda.*
    52:49 📚 *The variance of the exponential distribution with parameter lambda is 1 over lambda squared, impacting the central limit theorem.*
    56:53 📊 *Construction of a confidence interval for 1 over lambda involves manipulation of inequalities and dependency on Tn bar.*
    [01:00:11 URL](ua-cam.com/video/C_W1adH-NVE/v-deo.html) *📈 First-order Taylor expansion involves finding a theta bar between two values at which the expansion is performed, making them equal.*
    [01:00:38 URL](ua-cam.com/video/C_W1adH-NVE/v-deo.html) *📊 Multiplying by root n in Taylor expansion leads to root n Zn minus theta times g prime of theta bar.*
    [01:01:33 URL](ua-cam.com/video/C_W1adH-NVE/v-deo.html) *📉 Law of large numbers implies that theta bar converges to theta as Zn approaches theta.*
    [01:02:29 URL](ua-cam.com/video/C_W1adH-NVE/v-deo.html) *🍔 The "sandwich theorem" visualizes the convergence of theta bar to theta, considering the movement of Zn and theta.*
    [01:03:55 URL](ua-cam.com/video/C_W1adH-NVE/v-deo.html) *📜 Slutsky's theorem allows combining the convergence in distribution of Xn and convergence in probability of Yn if the limit of Yn is a constant.*
    [01:05:48 URL](ua-cam.com/video/C_W1adH-NVE/v-deo.html) *🔄 Slutsky's theorem facilitates combining sequences of random variables converging in distribution and probability under specific conditions.*
    [01:07:16 URL](ua-cam.com/video/C_W1adH-NVE/v-deo.html) *🔍 Delta method extends the central limit theorem to functions of averages, considering the derivative of the function.*
    [01:08:11 URL](ua-cam.com/video/C_W1adH-NVE/v-deo.html) *📊 Application of the Delta method to the function g(x) = 1/x results in a confidence interval for lambda.*
    [01:10:59 URL](ua-cam.com/video/C_W1adH-NVE/v-deo.html) *🔧 Replacing lambda by lambda hat in the confidence interval is justified by Slutsky's theorem, allowing for easier computation.*
    [01:12:22 URL](ua-cam.com/video/C_W1adH-NVE/v-deo.html) *🧮 Slutsky's theorem is crucial for replacing parameters with their estimates, simplifying computations in statistical applications.*
    [01:15:14 URL](ua-cam.com/video/C_W1adH-NVE/v-deo.html) *🎓 Understanding Slutsky's theorem enables the legitimate replacement of parameters with their estimates, maintaining convergence properties.*
    Made with HARPA AI

  • @arisweedler4703
    @arisweedler4703 3 роки тому +6

    29:51 that’s why electron clouds are drawn as balloons with really solid surface areas. We can draw the surface of the solid that the electron is in 95% of the time (or whatever P is if it is not 95%) because of “convergence in probability”.

  • @OttoFazzl
    @OttoFazzl 7 років тому +72

    This is very cryptic, especially without the first part.

    • @maxivy
      @maxivy 7 місяців тому

      so then watch the first part you spoiled bozo.

  • @nadekang8198
    @nadekang8198 6 років тому +11

    Am I the only one who felt hard to follow the course? 6:31 means basically, using properties of variance, we have X bar as X_bar, the n observations of x are i.i.d. Bernoulli, Bernoulli's variance = E[x^2] - E[x]^2 = 1^2*p - 0^2*(1-p) - (1^2*p - 0^2*(1-p))^2 = p - p^2 = p(1-p). Having this variance, we can derive the variance of X_bar, which is var(X_bar) = var([X1+...+Xn]/n), know properties of variance, we take n out, as var(X_bar) = 1/n^2 * var(X1+...+Xn), since i.i.d., these xi all have the same variance, therefore we have n*var(xi) divided by n^2, so the variance(X_bar) = var(Xi)/n, so to standardize it (in CTL), we divide it by the square root of var(Xi)/n, which is what he wrote on the board. So, it all roots back to Prof. Tsitsiklis' Probability course.

  • @crunchycho
    @crunchycho 8 місяців тому +1

    i'm a fan, professor! the breakdowns. the pacing. the reading of the room. noice.

  • @nearlymagicman3809
    @nearlymagicman3809 3 роки тому +14

    This lecture is basically broken by the fact that there is a continuation mismatch to 1. - references to concepts and theories are made that were never discussed in the previous lecture.

  • @Gisariasecas
    @Gisariasecas 6 років тому +11

    You can get the slides if you go to the link in the description (and read the missing part in this video).
    Excelent lecture,it took me 3 days to understand all the concepts exposed.
    At 55:23 is not equal to alpha, is equal to (1- alpha) since that is the way he defined the q (you can see that in the slides).

    • @prithvip6360
      @prithvip6360 4 роки тому +1

      Notice the inequality sign being reversed

    • @phillustrator
      @phillustrator Рік тому +1

      Is it really excellent if it took you 3 days to understand?

    • @freeeagle6074
      @freeeagle6074 Рік тому

      If one has not learned all the prerequisites, using 3 days to understand all concepts is very good. Learning the five or more prerequisites of linear algebra, calculus I and II, real analysis, and 18.600 Probability and random variable takes months or one year.@@phillustrator

  • @dimitargueorguiev9088
    @dimitargueorguiev9088 2 роки тому +16

    Nothing is broken - go to the URL link get the lecture slides , get the HW assignments and everything you need to understand what is going on is there. Phillipe Rigollet is excellent and discusses with ease fairly complicated material. Excellent job. My two complaints are that the sound volume is a way too low - whoever recorded the first lectures did lousy job. Also the slides are not properly exported to pdf - there are some missing math symbols in some of the equations

  • @crypto_quant_pro
    @crypto_quant_pro 7 років тому +98

    Why is sound volume so low?

    • @moeinhasani8718
      @moeinhasani8718 3 роки тому +6

      for anyone watching the course now, you can install the volume boost extension for chrome to increase the sound volume

    • @karanvashisth2933
      @karanvashisth2933 Рік тому

      My dear friend he was lying in first lecture. Number can incorporate in my many different combinations. Which is definitely human consciousness.

  • @Alma-im4qk
    @Alma-im4qk 23 дні тому

    It feels like playlist has skipped a video before this. Lots of unexplained concepts are thrown in the beginning.

  • @kingtriclee9529
    @kingtriclee9529 6 років тому +9

    The first video was recorded in Fall 2017. The rest of the lectures were recorded in Fall 2016, but video of Lecture 1 was not available. Only me notice this?

    • @bnglr
      @bnglr 5 років тому +3

      I guess you are right, I watched video 1 but I feel difficult to follow this (cont.) video

  • @laodrofotic7713
    @laodrofotic7713 2 роки тому +1

    Ahh dont you love when the teacher says "I am going to pick up this guy.." and the camera just zooms in to the point you CANT SEE WHAT THE "GUY" IS!!! ahhhhhhh so good bravo bravo.

  • @ebrahimfeghhi1777
    @ebrahimfeghhi1777 Рік тому +2

    When he talks about Hoeffding's inequality, shouldn't the tails be larger compared to a standard normal? I think smaller tails would result in a tighter upper bound (i.e. more confidence that the sample mean is close to mu), which wouldn't make sense when n is small.

  • @jonaseschle1104
    @jonaseschle1104 3 роки тому +4

    The sound level is too low, maybe re-upload with increased volume?

  • @pablock0
    @pablock0 Рік тому +1

    I really liked the lecture. Thanks MIT!

  • @richardmerckling592
    @richardmerckling592 4 роки тому +22

    was not expecting that scooter after watching part 1

    • @BenW.-lx9eq
      @BenW.-lx9eq 29 днів тому

      It's the temporary comfort animal.

  • @kopoiosnikodhmos5102
    @kopoiosnikodhmos5102 6 років тому +39

    how did it go from intro 1 to this? When did he tal about central klimit theorem etc? is he solving exercises? wtf

    • @theplayfulbot8447
      @theplayfulbot8447 5 років тому +14

      When taking this class you are supossed to have taken a probability course first

    • @adarshtiwari6374
      @adarshtiwari6374 5 років тому +4

      @Irving Ceron He meant the probability course provided by MIT (6.041). ua-cam.com/video/j9WZyLZCBzs/v-deo.html

    • @hmd1986
      @hmd1986 4 роки тому +4

      this video and the rest of the series is from Fall 2016, but the first video was from Fall 2017 (mentioned in OCW @ ocw.mit.edu/courses/mathematics/18-650-statistics-for-applications-fall-2016/lecture-videos/lecture-1-introduction-to-statistics/) - hence the inconsistency in the content. As for CLT, you can refer to MIT 6.041 video lectures 19 and 20.

  • @Guesswhoshaillina
    @Guesswhoshaillina 5 місяців тому +4

    Is it supposed to be undergraduate level stat? I absolutely understood nothing.

    • @placeholder6811
      @placeholder6811 4 місяці тому +1

      This is undergrad yes :) It assumes you know some real analysis which many undergrads know as well!

    • @YyNRCyY
      @YyNRCyY 3 місяці тому

      @@placeholder6811 eh, more like you know basic probability. Like Bertsekas Intro to probability, Probability and statistical inference by hogg tanis etc, or blitzstein intro to probability.

  • @brandomiranda6703
    @brandomiranda6703 7 років тому +14

    Oh no! Part 1 is missing! :(

  • @vincentkaruri2393
    @vincentkaruri2393 4 роки тому +3

    The discontinuity from the first video adds quite some confusion.

  • @nischalsubedi9432
    @nischalsubedi9432 3 роки тому +4

    @7:00 ish.. If I am understanding correctly, limiting distribution is not always Gaussian. There might be cases when CLT isn't valid such as when n is small. Are there any other cases? And how were the bounds defined for kiss example? If someone can explain, I would appreciate it.

    • @freeeagle6074
      @freeeagle6074 Рік тому

      Another case is when variance of distribution is unbounded such as the Cauchy distribution. When n -> infinity, Cauchy is still Cauchy (not Gaussian). 18.600 discusses this.

  • @FineFlu
    @FineFlu 4 роки тому +5

    A new intro stats series should be recorded either online (COVID baby!) or in person

    • @normalperson1130
      @normalperson1130 4 роки тому +3

      I wish they upload their covid stuff online.. But I think it is restricted by some US laws that state that educational online content released for the mass public should be accessible by all people and hence they have to take care of subtitling and also blurring faces of class students because of privacy laws.
      I wish it was simpler because all the best universities are teaching online and I hope the content becomes accessible to common folks like us.

  • @jamesgohigh26
    @jamesgohigh26 4 роки тому +5

    why did he ride a scooter during lecture?

  • @meetpatel-lq7xb
    @meetpatel-lq7xb 5 місяців тому

    Slide on 07:43 seems not correct. If we are targeting confidence interval of 1-alpha, shouldn't it be q_(alpha/2) and not q_alpha?

  • @DalpMaths
    @DalpMaths 2 роки тому

    I find these videos quite interesting, however the disorder at the time of writing that many teachers deal with bothers me. Unclear and messy handwriting. It would be nice if they could improve on those aspects.

  • @neilbryanclosa462
    @neilbryanclosa462 6 років тому +9

    I don't understand

  • @hondajacka2
    @hondajacka2 4 роки тому +3

    I can barely hear the audio. Please remake this video next time.

  • @RahulGupta-tf7gd
    @RahulGupta-tf7gd 8 місяців тому

    Pahtetic job of the sound recording ! Can't hear anything even with an external speaker

  • @konstantinoschristopoulos764
    @konstantinoschristopoulos764 3 роки тому +1

    Too bad this course is ruined by discontinuity... There seems to be a whole lecture missing, they should have added also the second lecture from 2017 at least. It's better to have some overlap than missing data.

  • @techodyssey4142
    @techodyssey4142 6 років тому +10

    Of Anyone, Massachusetts "Institute for Technology" should be able to level the volume so we can hear it.. Breaking edge technology, but makes movies with 1990's PBS vibes.. WOW
    Thanks for the videos though!

  • @bnglr
    @bnglr 5 років тому +9

    better to watch the Khan Academy statistics playlist before MIT18.650

  • @srinivasanthirunavukkarasu4499
    @srinivasanthirunavukkarasu4499 6 років тому +3

    Is this basic statics??
    I am worried that i don't understand....
    Can i have the stellar ink wheee the notes and assignment are loaded

    • @mitocw
      @mitocw  6 років тому

      See the course materials on MIT OpenCourseWare at: ocw.mit.edu/18-650F16. Best wishes on your studies!

  • @brainstormingsharing1309
    @brainstormingsharing1309 4 роки тому +1

    Absolutely well done and definitely keep it up!!! 👍👍👍👍👍

  • @asdfafafdasfasdfs
    @asdfafafdasfasdfs Рік тому

    Why can't we get the complete recordings of 2017 so it's continuous? :(

    • @mitocw
      @mitocw  Рік тому +1

      There wasn't any major changes between the two courses. Lecture 1 was missing from the 2016 recording of the course so only lecture 1 was recorded to patch that omission. See the course on MIT OpenCourseWare for the materials at: ocw.mit.edu/18-650F16. Best wishes on your studies!

    • @asdfafafdasfasdfs
      @asdfafafdasfasdfs Рік тому +1

      @@mitocw thanks! understand, the problem is that there was a significant change in the transition from lecture 1 to 2 so it's not possible to understand this lecture without more context. The slides help, but yeah not great in terms of just watching the videos... thus the suggestion to perhaps publish the 2017 course instead (or additionally)? anyway, just what's seems possible as an outsider, assume that there's a reason to having patched it this way instead.

  • @Jeremy-fd2pw
    @Jeremy-fd2pw 6 місяців тому

    Thanks for the video professor

  • @felixmende2693
    @felixmende2693 Рік тому

    amazing lecture. Thank you very much.

  • @itaamelia6715
    @itaamelia6715 Рік тому +1

    in my years of college, i never seen the teacher describe complicated equation so deeply, most of them dont even know the details just put in the power point and explained and the student sleep queitly.

  • @chrisdavies616
    @chrisdavies616 2 роки тому

    Isn't CLT about the distribution of your estimated mean being normal (rather than calculating the mean from a very large sample)? i.e. calculate your mean 1000 times and those 1000 means will follow a normal distribution

    • @julija5949
      @julija5949 2 роки тому

      Well if you look at the statement of the theorem it says that (X1+...+Xn)/n when properly standardized converges to the standard normal, so it is about the sample mean being normally distributed but only in the limit, that's why it's called the central *limit* theorem and that's why we need a large sample size. If you took a small sample a large number of times(like 1000) and calculated the mean each time, there's no reason to think that thus obtained means would be normally distributed(unless the distribution you're sampling from is normal). To take an extreme example assume that your sample size is just 1 and you take the mean 10000 times - the mean will equal the random variable itself each time, so the means will have the same distribution as the r.v you're sampling from. I might have misunderstood your question though.

    • @freeeagle6074
      @freeeagle6074 Рік тому

      Yes. The mean of n samples (A) follow normal distribution, same for the sum of those n samples (B). They are on different scales. Multiply mean of A by n, we get mean of B. Multiple SD of A by sqrt(n), we get SD of B. This follows from linearity of sum of mean/variance of iid distributions (with bounded mean and variance).

  • @Marteenez_
    @Marteenez_ 2 роки тому

    @11:11:38 What does he meant lambda comes from here? Why does it come from the variance of the limit distribution? After this, where does the next expression come from, the one he divides lambda by, why does he do this?

    • @weimingsheng3214
      @weimingsheng3214 8 місяців тому

      that's just standardizing the normal to be N(0,1). He does this to show that slutsky can be applied (together with lln) so that lambda can be replaced by lambda hat

  • @demo5948
    @demo5948 7 років тому

    It's very kind of you

  • @TheAvoca1989
    @TheAvoca1989 5 років тому +1

    At 0:22; you need a better duster

  • @cescob7979
    @cescob7979 4 роки тому +1

    Great lecture! I have enjoyed it; thanks.

  • @KhaledMohamed-is5pl
    @KhaledMohamed-is5pl 4 дні тому

    why wasting my time on a course with 3 lectures un recorded ?
    😡

  • @romchiks
    @romchiks 6 років тому +1

    What's with the sound, guys??!?

  • @profetadosecxxi
    @profetadosecxxi 5 років тому +5

    Here I am trying to learn English, with a french accent teacher who teach statistics. What is the possibility I am be successful?

    • @georgeivanchyk9376
      @georgeivanchyk9376 4 роки тому +1

      Not a lot, considering your grammar. But you can try :)

    • @profetadosecxxi
      @profetadosecxxi 4 роки тому

      @@georgeivanchyk9376 in my native language I learned to speak before to learn grammar

    • @normalperson1130
      @normalperson1130 4 роки тому +1

      @@profetadosecxxi true. English isn't my first language but I easily get my point across and that makes sense. I never even thought about the grammar of my native language. Everything just fits right in and works normally and makes sense ever since I was 4 yr old. But with English it was a long journey till high school

    • @King-Matshobane
      @King-Matshobane 3 роки тому +2

      English have strict grammar rules but it’s not essential when you grasp the language. He is French and that should motivates you. I was taught calculus by a Sri Lankan lecturer and his spoken English was bad but it didn’t deter me from learning the subject.

    • @freeeagle6074
      @freeeagle6074 Рік тому

      America gives one the opportunity to play her or his talent to the full in this great country no matter what accent she or he holds. You'll get used in a few weeks to the accent of any professor who MIT allows to teach.

  • @tomd7841
    @tomd7841 7 років тому +15

    maybe use a microphone next time

    • @manavkashyap1927
      @manavkashyap1927 3 роки тому +1

      😂😂

    • @BiologyIsHot
      @BiologyIsHot 3 роки тому

      Me grabbing a large speaker so I can get a distorted voice that is actually within decibel levels humans can perceive

    • @freeeagle6074
      @freeeagle6074 Рік тому

      Turning volume to maximum on desktop allows one to listen to this lecture without any problem. Using mobile may render this lecture inaudible even with maximum volume.

  • @UnforsakenXII
    @UnforsakenXII 7 років тому +4

    No part uno?

  • @tomaskovarik1215
    @tomaskovarik1215 4 роки тому +1

    the quality is so low

  • @MrPhillips365
    @MrPhillips365 7 років тому +16

    why is the professor on a scooter?

  • @BulbasaurLeaves
    @BulbasaurLeaves 2 роки тому

    Where did q and alpha come from?

  • @calsavestheworld
    @calsavestheworld 4 роки тому +7

    You'd think MIT would figure out the equation for chalkboard erasers. I guess that's more of a Harvard specialization.

  •  Рік тому

    Why is he using a bicycle?

  • @sukursukur3617
    @sukursukur3617 4 роки тому +1

    I spent my years as a student. To teach something to someone, you firstly should motivate them by showing the point they will reach if they grasp taught. This is why firstly you should solve a lot of different real life example. In this course, students cant see the outlet of tunnel.

    • @ryanjackson0x
      @ryanjackson0x 3 роки тому +1

      If everyone thought that way then we wouldn't have the mathematics we do today. Often the real world examples don't exist until dozens or hundreds of years later.

    • @sukursukur3617
      @sukursukur3617 3 роки тому +4

      @@ryanjackson0x why did you write that? That is not true. Maths has developed by physical examples. In the past, scientists were interested in more than one science + maths. They tried to solve their own problems. Their motivation was the problem itself.

  • @prakharsharma7232
    @prakharsharma7232 2 роки тому

    Poor Recording.. can't hear anything

  • @rachanakandi1458
    @rachanakandi1458 8 місяців тому

    Can anyone explain what exactlt is alpha

  • @keepfeatherinitbrothaaaa
    @keepfeatherinitbrothaaaa 4 місяці тому +1

    Confusing

  • @gatorflight74
    @gatorflight74 7 років тому +2

    Someone get that gent a chair and a overhead projector.

    • @fredericmoresmau4303
      @fredericmoresmau4303 5 років тому

      because they lied....
      They lied to me and let me run up then energy exhaustion

  • @JOHNSMITH-ve3rq
    @JOHNSMITH-ve3rq 2 роки тому

    Aw come on, the sound is bad!!

  • @reznoli9656
    @reznoli9656 7 років тому

    why i can't watch it?

    • @mitocw
      @mitocw  7 років тому +5

      If you are having trouble viewing this on UA-cam, you can also get these videos from the Internet Archive and iTunes U: archive.org/details/MIT18.650F16 or itunes.apple.com/us/itunes-u/id1262009852.

  • @Splatpope
    @Splatpope 6 років тому +2

    nice trike bro

  • @Originalimoc
    @Originalimoc 5 років тому +1

    I just remembered 18650 is a battery model 😂

  • @SphereofTime
    @SphereofTime Рік тому

    1:01:08

  • @LC-qm1xs
    @LC-qm1xs 6 років тому

    Thanks Mit

  • @ryanchiang9587
    @ryanchiang9587 Рік тому

    applied mathematics

  • @wayneriley7367
    @wayneriley7367 6 років тому +15

    Ah college hasn’t changed! He started with an interesting example, then went into technical speak. He is now relying on equations he knows but the students don’t. He should be giving real common sense examples then using the equations. I am teaching AP stats and he just seems to jump into complicated equations. If this is a first year stats course I feel sorry for the students. It seems teaching at the university level hasn’t improved in almost 40 years.

    •  5 років тому +1

      True. I felt the same back in 2007 from my first statistic class to the second... I remember I was in doubt if I had skipped a class! And today I felt the same from the video 1 to video 2...

    • @SamMoore19
      @SamMoore19 5 років тому +8

      lol He said in the first video that you should know calculus, linear algebra and some probability. Dont be a puss, if you teach AP stats this should be easy.

    • @MrScattterbrain
      @MrScattterbrain 5 років тому +10

      This is not a first year course. As far as I heard from MIT staff, this is positioned as a graduate level course for non-math students ("non-math" in MIT's understanding still implies quite solid knowlede of math). It assumes the students are hands on calculus, probability theory, and had some exposure to linear algebra (all the way up to eigendecomposition) and multivariable calculus.
      This course requires quite some effort from the learners. It certainly has an "academic', rather than practical, flavor. Personally, i found it very interesting and useful (I completed the 18.6501x version on EDX).

    • @Helpmesubswithoutanyvideos
      @Helpmesubswithoutanyvideos Рік тому

      I agree

  • @mohammadsaifuddin9459
    @mohammadsaifuddin9459 7 років тому +3

    Text book...........Please...!!!!!

    • @mitocw
      @mitocw  7 років тому +5

      There was no required text for this course.

    • @gbrneale2
      @gbrneale2 6 років тому +5

      They had a book called all of statistics by Wasserman that was recommended in the first video

  • @saubaral
    @saubaral 4 роки тому

    How is this an intro?

    • @saubaral
      @saubaral 4 роки тому

      also, why is the teacher looking at students as if they offended his dead grandfather's ashes?

  • @Arkaizen29
    @Arkaizen29 2 роки тому +1

    Introduction to brain damages.

  • @Sharpdus
    @Sharpdus 5 років тому +2

    these statistics do not suit for psychology

  • @laodrofotic7713
    @laodrofotic7713 2 роки тому +3

    He is probably the worst professor I ever saw to be completely honest.

  • @SumoCumLoudly
    @SumoCumLoudly 5 років тому +3

    Teaching is just a paycheck to this guy.

  • @Temp321-r2p
    @Temp321-r2p 2 роки тому

    Test

  • @tomaskovarik1215
    @tomaskovarik1215 4 роки тому

    Totally useless......this shows that not everything that comes from MIT is useful

  • @Infinitesap
    @Infinitesap 6 років тому +6

    Pretentious. MIT you can do better!

  • @hj-core
    @hj-core Рік тому

    Come from 6.041 and this course seems too difficult for me..🥲

    • @mitocw
      @mitocw  Рік тому +1

      Prerequisites
      Probability theory at the level of 18.440 Probability and Random Variables ( ocw.mit.edu/courses/18-440-probability-and-random-variables-spring-2014 ). Some linear algebra (matrices, vectors, eigenvalues). Best wishes on your studies!