Shih-Shinh Huang
Shih-Shinh Huang
  • 32
  • 59 441
quarter DIP: Interactive Graph Cuts
This video introduces a graph-based interactive segmentation algorithm.
The outline of this lecture includes:
(1) Introduction about the motivation and problem description of interactive segmentation, and the central idea of the algorithm using graph cuts.
(2) Graph-based modeling about the construction of a graph, the cost definitions of the edges on the constructed graph, and the process of cost minimization for finding the optimal segmentation.
Any comments are welcome. (email: powwhuang@gmail.com)
All resources are available on the website (quarter.tw)
Переглядів: 35

Відео

quarter CNN: Temporal Convolution Networks (TCN)
Переглядів 62110 місяців тому
This video introduces the temporal convolutional network (TCN), a 1D convolutional neural network designed for sequence modeling tasks. The outline of this lecture includes: (1) Introduction about the motivation, central idea, and architecture overview of the TCN. (2) 1D dilated convolution about the definition and padding rule of the convolution process in TCN. (3) Residual block about the two...
quarter RNN: Simple Recurrent Neural Network (RNN)
Переглядів 168Рік тому
This video introduces the recurrent neural network (RNN) in its simplest form which has been widely used for dealing with sequential data. To clearly explain what the RNN is, we use the problem of price forecasting as an example to explain it in a step-by-step manner. The outline of this lecture includes: (1) Introduction about the RNN model, its main concept, and an example of price forecastin...
quarter RNN: Tokenization in Transformer From Text to Tokens
Переглядів 150Рік тому
This video talks about the tokenization used in Transformer to convert text into tokens for natural language processing. The outline of this lecture includes: (1) Introduction about what the Transformer is and what tokenization is. (2) Tokenization strategy about the main steps and the pros and cons of three tokenization algorithms, character, word, and subword. Any comments are welcome. (email...
quarter CNN: (YOLO v1) You Only Look Once Unified Real-Time Object Detection
Переглядів 7382 роки тому
This video talks about YOLO version 1 short for You Only Look Once. YOLO v1 is a unified real-time object detection algorithm based on convolutional neural networks and was published in CVPR, in 2016. The outline of this lecture includes: (1) Introduction about the background and the idea of YOLO v1. (2) Unified detection about the anchor mechanism used for object detection, network architectur...
quarter: DIP Hough Transform A Case Study of Line Detection
Переглядів 1332 роки тому
This video talks about Hough transform which is a well-known algorithm for detecting the shape of objects from their boundary points. For detailing Hough transform, we take the line detection as a case study. The outline of this lecture includes: (1) Introduction about the history, purpose, and idea of Hough transform. (2) Detection algorithm about the two steps of Hough transform, that are, sh...
quarter CNN Mask R-CNN
Переглядів 3202 роки тому
This video introduces the Mask R-CNN which is a convolution neural network for instance segmentation. The outline of this lecture includes: (1) Introduction about what instance segmentation is and how the Mask R-CNN works. (2) RoI align about the objective, formal statement, and align steps. (3) Mask branch about the design concept, network architecture and the mask loss defined for training. 0...
quarter DIP Introduction to Kalman Filter
Переглядів 2682 роки тому
This video introduces Kalman filter that is a popular algorithm for estimating some unknown variables given the measurements observed over time. Kalman filter was proposed by R. E. Kalman in 1960 and has been widely used in many applications. The outline of this video includes: (1) Introduction about what is Kalman filter and its main idea. (2) Gaussian distribution about the brief introduction...
quarter CNN: Feature Pyramid Networks for Object Detection
Переглядів 1,4 тис.3 роки тому
This video introduces feature pyramid networks that produce a feature pyramid for addressing the scaling problem in the object detection area. The outline of this lecture includes: (1) Introduction about the purpose of object detection, the solution for detecting objects in multiple scales, and the main idea of FPN. (2) Network architecture about the main components of FPN including bottom-up p...
quarter Py: Basic Digital Image Processing Using OpenCV on Google Colab
Переглядів 6943 роки тому
This video introduces basic digital image processing using Opencv on Google Colab which is a free cloud service for machine learning. The outline of this lecture includes: (1) Google Colaboartory about why uses Colab for writing machine learning algorithms, how to start Colab, and how to write the first Colab program. (2) Basic image processing about what is OpenCV and how to write the programs...
quarter CNN: Deep Residual Network
Переглядів 3753 роки тому
This video introduces the deep residual network published in the paper entitled “Deep Residual Learning for Image Classification”, by Kaiming He in CVPR 2016. The outline of this lecture includes: (1) Introduction about the depth evolution in the literature of neural networks and the issue to be addressed when the neural network uses more layers. (2) Deep residual network about the main idea of...
quarter DIP: Determining Optical Flow: Horn and Schunck Method
Переглядів 2 тис.3 роки тому
This video talks about an algorithm proposed by Horn and Schunck in 1981 for determining optical flow between two consecutive images. The outline of this lecture includes: (1) Introduction about the motion and the definition of the optical flow. (2) Brightness constraint about the Taylor expansion for formula derivation of brightness constraint. (3) Horn-Schunck method about the additional smoo...
quarter CNN: RoI Pooling and Align
Переглядів 3,4 тис.3 роки тому
This video introduces two well-known feature extraction operations, called RoI pooling and RoI align, that are widely used in two-stage object detection or object segmentation. The outline of this lecture includes: (1) Introduction about the background of RoI pooling and RoI align and what the RoI feature extraction is. (2) RoI pooling about the overview and pooling steps for RoI pooling (3) Ro...
quarter DIP Two Pass Connected Component Labeling for Binary Image
Переглядів 1,1 тис.4 роки тому
In this video, we introduce a two-pass algorithm for connected component labeling. The outline of this video includes: (1) Introduction about what is connected component labeling and the definitions of connectivity and connected component. (2) Two-pass labeling about the overview of connected component labeling algorithm and the operations performed in pass one and pass two, respectively. 00:00...
quarter Py Beginner's Guide on PyTorch from Linear Regression
Переглядів 2924 роки тому
This lecture introduces how to write a PyTorch program for machine learning. In order to focus on why PyTorch makes machine learning programming easier and more intuitive, we stick with a simple and familiar problem: a linear regression with a single feature. The outline of this lecture includes: (1) Introduction about the description of machine learning, the basic programming paradigm for mach...
quarter CNN: Faster R-CNN
Переглядів 2,5 тис.4 роки тому
quarter CNN: Faster R-CNN
quarter CNN: Region Proposal Network (RPN)
Переглядів 9 тис.4 роки тому
quarter CNN: Region Proposal Network (RPN)
quarter CNN: FaceNet: A Unified Embedding for Face Recognition and Clustering
Переглядів 6 тис.4 роки тому
quarter CNN: FaceNet: A Unified Embedding for Face Recognition and Clustering
quarter DIP: Real time Foreground Background Segmentation Using Codebook Model
Переглядів 7985 років тому
quarter DIP: Real time Foreground Background Segmentation Using Codebook Model
quarter DIP Haar like Feature
Переглядів 5025 років тому
quarter DIP Haar like Feature
quarter DIP: Histogram Matching (Specification)
Переглядів 8635 років тому
quarter DIP: Histogram Matching (Specification)
quarter DIP Support Vector Machine: Two Separable Classes
Переглядів 2905 років тому
quarter DIP Support Vector Machine: Two Separable Classes
quarter DIP Otsu Algorithm Optimal Global Thresholding
Переглядів 3 тис.5 років тому
quarter DIP Otsu Algorithm Optimal Global Thresholding
quarter DIP LBP Local Binary Pattern
Переглядів 8 тис.5 років тому
quarter DIP LBP Local Binary Pattern
quarter DIP HOG Histogram of Oriented Gradients
Переглядів 3,1 тис.6 років тому
quarter DIP HOG Histogram of Oriented Gradients
quarter DIP Gaussian Mixture Models for Background Subtraction
Переглядів 9 тис.6 років тому
quarter DIP Gaussian Mixture Models for Background Subtraction
quarter DIP HLID Histogram of Local Intensity Difference
Переглядів 1216 років тому
quarter DIP HLID Histogram of Local Intensity Difference
quarter DIP Efficient Graph Based Image Segmentation
Переглядів 3,7 тис.6 років тому
quarter DIP Efficient Graph Based Image Segmentation
quarter Java Event Handling
Переглядів 3206 років тому
quarter Java Event Handling
quarter Java while Statement
Переглядів 176 років тому
quarter Java while Statement

КОМЕНТАРІ

  • @maryamjafaripakzad9448
    @maryamjafaripakzad9448 4 місяці тому

    Thank you. You explained very well and easy to understand.

    • @quarter2018
      @quarter2018 4 місяці тому

      Thanks for your recognition. You are so welcome. ^^

  • @longtoZ
    @longtoZ 4 місяці тому

    Very useful, sir!

  • @setarehrezaei9930
    @setarehrezaei9930 6 місяців тому

    How can we have access to the slides?

    • @quarter2018
      @quarter2018 6 місяців тому

      please send an email to quarterhuang2018@gmail.com. We will reply u with the slide in pdf format.

    • @quarter2018
      @quarter2018 5 місяців тому

      Now, you can get the slide from the website "quarter.tw"

  • @改革家-l1y
    @改革家-l1y 7 місяців тому

    最小平方法的概念我懂,但還是不太懂(u,v)怎麼來的。理論上每個點都應該要有一個初步的(u,v)後續才可以求平均值才對

    • @劉勇佑-t3h
      @劉勇佑-t3h 7 місяців тому

      一開始u,v都當零矩陣算,一直迭代下去就能求

    • @quarter2018
      @quarter2018 7 місяців тому

      謝謝你的協助回覆,^^

  • @labibalaskar6965
    @labibalaskar6965 7 місяців тому

    How to connect u...I am interested in doing research under a broad area name graph theory based image segmentation after watching your video on graph based segmentation

    • @quarter2018
      @quarter2018 7 місяців тому

      You can connect me via email "quarterhuang2018@gmail.com"

  • @AmitGupta-lx9wy
    @AmitGupta-lx9wy 7 місяців тому

    I search many paper regarding this topic, but I didn't get it.The way you explained the background subtraction using GMM is fabulous. Thank you so much for making this video.

    • @quarter2018
      @quarter2018 7 місяців тому

      Indeed, I am glad that this video is helpful to u.

  • @amritdahal5177
    @amritdahal5177 8 місяців тому

    Best Explanation Sir

    • @quarter2018
      @quarter2018 8 місяців тому

      Thanks for your recognition. You are so welcome

  • @emperorengineeringsolutions
    @emperorengineeringsolutions 8 місяців тому

    Thank you very much for the informative content sir

    • @quarter2018
      @quarter2018 8 місяців тому

      I am so glad that this video is helpful to you

  • @indukc8960
    @indukc8960 10 місяців тому

    Thanks from India😇

    • @quarter2018
      @quarter2018 9 місяців тому

      Your are so welcome ^^

  • @nhanthien36
    @nhanthien36 Рік тому

    Thank u very much <3

  • @T_rex-te3us
    @T_rex-te3us Рік тому

    Incredible explanation, thank you so much Mr. Huang!

    • @quarter2018
      @quarter2018 Рік тому

      I am glad that you like this video and thanks for your recognition. ^^

  • @tanphanminh3307
    @tanphanminh3307 Рік тому

    Can I ask for the PowerPoint file of this article?

    • @quarter2018
      @quarter2018 Рік тому

      You can download the ppt in pdf format via the link "gg.gg/quarter"

  • @zukofire6424
    @zukofire6424 Рік тому

    Hi Professor Shih-Shinh Huang, thank you for this explanation. Clear and helpful.

  • @CaratacusAD
    @CaratacusAD Рік тому

    Excellent video and explanation, I think this is the clearest explanation on youtube. I have a quick question. Is it possible to swap out the ROI pooling for ROI align in a Faster RCNN network. I believe it is more accurate and i've seen this proposed in a few academic papers. I'm trying to build a custom object detector based on Faster-RCNN but with additional branches to detect additional image properties like occulusion, along with the normal BBOX and classification branch. Many thanks

    • @quarter2018
      @quarter2018 Рік тому

      Since RoI pooling and RoI align are both for normalizing feature maps for further predicting, it is feasible to substitute RoI pooling with RoI align within the Faster R-CNN network. However, this substitution comes at the expense of increased computational complexity.

  • @skarr4616
    @skarr4616 Рік тому

    Thank you very much for the great explanation.

    • @quarter2018
      @quarter2018 Рік тому

      You are so welcome. ^^. We are so glad to have your recognition.

  • @hangngo28092
    @hangngo28092 Рік тому

    Thanks, sir. You're a great teacher !

    • @quarter2018
      @quarter2018 Рік тому

      Thanks for your compliment. I’m really into it! ^^

  • @小草導師國昌
    @小草導師國昌 Рік тому

    老師講的真好,敲碗老師做yolov2~v7的教學影片

    • @quarter2018
      @quarter2018 Рік тому

      很謝謝你的肯定,也很高興你喜歡這個影片,關於 yolov2 ~ yolov7,我們也正考慮要錄製相關的影片,但據我們目前的理解,他們主要是基於 yolov1 進行不同層面的改良,這種類型的影片有點難做 ^^,還在傷腦筋中。

  • @waterspray5743
    @waterspray5743 Рік тому

    You explained it so clearly under 30 minutes. Hands down the best professor.

    • @quarter2018
      @quarter2018 Рік тому

      Thanks a lot for your recognition. ^^

  • @user-jv8vu3sh8q
    @user-jv8vu3sh8q Рік тому

    感謝教授的教學 這個教學真的是核心中的核心

    • @quarter2018
      @quarter2018 Рік тому

      很謝謝你的肯定,也很開心你喜歡這個影片。

  • @mohammadyahya78
    @mohammadyahya78 Рік тому

    Thank you again. Question#1: Why do you upsample by a factor of 2 and what is the purpose of upsampling in a top down fashion please at 13:10? Question#2: Also may you please add a little what is aliasing effect you mentioned in 13:50? Do you mean because we up-sampled the feature maps by 2, the features become very close to each other, which is the aliasing effect if I am not wrong? Why the 3x3 Conv filter you applied solves aliasing effect please? I am new to this area?

    • @quarter2018
      @quarter2018 Рік тому

      Sorry for so late to response. Up-sampling by a factor of 2 in a top-down fashion is to make the feature map with the same dimension to the bottom one. The two features maps with the same dimension can perform element-by-element addition so as to propagate semantics from top to bottom.

  • @mohammadyahya78
    @mohammadyahya78 Рік тому

    How FPN produces feature pyramid please in 6:05?

    • @quarter2018
      @quarter2018 Рік тому

      It propagates semantics from high level down to low level.

  • @mohammadyahya78
    @mohammadyahya78 Рік тому

    May I know please what is dense scale sampling in simple terms at 4:27? Also, why we should use a feature map from the last convolution neural network for prediction please?

    • @quarter2018
      @quarter2018 Рік тому

      The feature map at the last conv. layer is the most semantic and is suitable for detection purposes.

    • @mohammadyahya78
      @mohammadyahya78 Рік тому

      @@quarter2018 Thank you, so what about dense scale sampling at 4:27, may I know its purpose please?

  • @mohammadyahya78
    @mohammadyahya78 Рік тому

    Thank you very much. May I know pleae why you said at 4:16 that The detection of a large object is from the large pyramidal level while the detection of a small object is from the lower pyramidal level?

    • @quarter2018
      @quarter2018 Рік тому

      Since the low pyramidal level has a high resolution, it is feasible to detect small objects. In other words, high pyramidal levels obtained from convolution operations generally have a lower resolution because of sub-sampling. This makes the feature of the small objects in higher layers in-visible.

    • @mohammadyahya78
      @mohammadyahya78 Рік тому

      @@quarter2018 Thank you.

  • @aggelosalexopoulos3039
    @aggelosalexopoulos3039 Рік тому

    Great explanation, Congratulations!

  • @nareshk3885
    @nareshk3885 Рік тому

    Thank you for explaining simplistic and understandable way. Please can you do video on KNN Model for Background Subtraction

    • @quarter2018
      @quarter2018 Рік тому

      Thanks for your recognition. Do you have references for KNN model for background subtraction?

  • @yasminemasmoudi9963
    @yasminemasmoudi9963 Рік тому

    Thank you !

  • @Banefane
    @Banefane 2 роки тому

    Thank you very much :)!

  • @mohammadwahba3077
    @mohammadwahba3077 2 роки тому

    Gooooooooooooooooooood very gooood - Go on

  • @shabbirahammed4596
    @shabbirahammed4596 2 роки тому

    nice...

  • @tejasduseja
    @tejasduseja 2 роки тому

    Thanks for the lecture. It is very concise and clear.

    • @quarter2018
      @quarter2018 2 роки тому

      Thanks for your recognition. You are so welcome ^^

  • @levuquang7483
    @levuquang7483 2 роки тому

    thank Prof!

  • @romanasalim5335
    @romanasalim5335 2 роки тому

    sir..could you plse share the ppt link to download.U have provided the best explanation of RCN.

    • @quarter2018
      @quarter2018 2 роки тому

      You can download our slide in pdf format from our web-site via the link gg.gg/quarter

  • @yacinetorki1740
    @yacinetorki1740 2 роки тому

    Great job thank you 🙏 ❤

    • @quarter2018
      @quarter2018 2 роки тому

      Thanks for your recognition. ^^

  • @yosansu
    @yosansu 2 роки тому

    thank you prof shih-shinh huang

  • @windbug001
    @windbug001 2 роки тому

    Thanks for this online lectures about ML. I can't wait to dive in .

    • @quarter2018
      @quarter2018 2 роки тому

      We are so glad that you like our video

  • @LimonKitchenDesign
    @LimonKitchenDesign 2 роки тому

    This is a very good explanation. Thank you very much!

  • @RR-nq1jb
    @RR-nq1jb 2 роки тому

    Thank you.

  • @spikewong9740
    @spikewong9740 2 роки тому

    The accent of oral English of Taiwanese and Mainland Chinese do have a lot in common which makes your presentation much easier for me to understand, thanks, buddy!

    • @quarter2018
      @quarter2018 2 роки тому

      I am glad that you enjoy this video ^^

  • @ryancusack5909
    @ryancusack5909 2 роки тому

    Im not sure I understand the anchor labeling and anchor sampling step in training. When you first select a random image do you generate random anchors throughout the image at random locations? Why don't we use the ground truth bounding box in training instead of the highest IOU anchor box with the ground truth.

    • @quarter2018
      @quarter2018 2 роки тому

      (1) The anchors are fixed at each point of the conv. feature map but not randomly generated. (2) During training, you have to compute the difference (called loss) between ground-truth bounding boxes and predicted ones. The selected anchors are considered as the predicted bounding boxes for computing the loss.

  • @waleedaiad3411
    @waleedaiad3411 2 роки тому

    Is d in deep cnn mean the number of hidden layer that used?

    • @quarter2018
      @quarter2018 2 роки тому

      $d$ is the number of neurons used in the output layer that makes your output is a $d$-dimensional feature vector.

  • @waleedaiad3411
    @waleedaiad3411 2 роки тому

    How can i khow how many hidden layer used in facenet?

    • @quarter2018
      @quarter2018 2 роки тому

      The number of hidden layers is dependent on your application.

    • @waleedaiad3411
      @waleedaiad3411 2 роки тому

      @@quarter2018 can i talk with you on whatsapp or any way if you can. And thank you for your answer😍🌹🌺

    • @quarter2018
      @quarter2018 2 роки тому

      @@waleedaiad3411 We will open a google meeting (meet.google.com/xya-vuys-vfo) for discussion at 10:00 PM ~ 10:30 PM (UTC+8) every Wednesday.

    • @waleedaiad3411
      @waleedaiad3411 2 роки тому

      @@quarter2018 ok i will open with you Thanks for your answer🌺🌹

  • @valentin5852
    @valentin5852 2 роки тому

    Great explanation! Incredibly clear! Thank you

    • @quarter2018
      @quarter2018 2 роки тому

      Thanks a lot for your comments

  • @danbrown6698
    @danbrown6698 2 роки тому

    Although your oral English is pretty poor, your content is fairly clear

    • @quarter2018
      @quarter2018 2 роки тому

      First, thanks for your comment "your content is fairly clear". For your suggestion, what we currently do is to provide subtitle. In the future, we will spend more time on how to improve our oral English. Thanks for your comments.

  • @cem_kaya
    @cem_kaya 2 роки тому

    thanks for such detailed explanation

  • @lionhuang9209
    @lionhuang9209 2 роки тому

    Very clear.

  • @gradientattack
    @gradientattack 2 роки тому

    this video deserves more visits

    • @quarter2018
      @quarter2018 2 роки тому

      Thanks for your recognition.

  • @farvezfarook3422
    @farvezfarook3422 2 роки тому

    Nice Explanation. Very well articulated. Thank you

  • @sanalcc
    @sanalcc 2 роки тому

    Clear & Complete explanation, Thank you

    • @quarter2018
      @quarter2018 2 роки тому

      Thanks for your recognition.

  • @yagmuraktas2423
    @yagmuraktas2423 2 роки тому

    Best video I have ever seen about RPN!!!! I dont understand this subject for days now all clear thank you a lot!!!! Please keep going with making videos you are the best!!

    • @quarter2018
      @quarter2018 2 роки тому

      Thanks for your recognition. We will keep going with this.

  • @pfl2226
    @pfl2226 2 роки тому

    I heartily request you to make similar videos about Single Object Tracking and Multople Object Tracking. Don't worry if it is too long. I would rather spend 2 hours in a single video rather than wasting time and effort in looking for other non-sense videos.

    • @quarter2018
      @quarter2018 2 роки тому

      Do you have any reference about single or multiple object tracking?