- 32
- 88 878
PegaAnalytics
Приєднався 10 жов 2011
JMP is powerful software for data exploration and statistical analysis.
Pega Analytics Ltd specialises in delivering JMP training courses and creating add-ins and scripts based on the JMP scripting language (JSL).
In this channel, you will learn tips and tricks for working with JMP Software, together with insights into the statistical thinking that underpins a variety of statistical methods.
Pega Analytics Ltd specialises in delivering JMP training courses and creating add-ins and scripts based on the JMP scripting language (JSL).
In this channel, you will learn tips and tricks for working with JMP Software, together with insights into the statistical thinking that underpins a variety of statistical methods.
How to perform a t-test to compare 2 samples using JMP software
"How To" is a new series of videos. They are short, and to the point, to give you quick answers when you need them. They assume you know what you want to do, and why you want to do it ... but you don't know how to do it using JMP Software.
This video addresses the situation where you want to use a t-test to compare 2 samples of data.
This video addresses the situation where you want to use a t-test to compare 2 samples of data.
Переглядів: 42
Відео
How do I test my data for normality using JMP software?
Переглядів 66Місяць тому
"How To" is a new series of videos. They are short, and to the point, to give you quick answers when you need them. They assume you know what you want to do, and why you want to do it ... but you don't know how to do it using JMP Software. This video addresses the situation where you want to determine whether or not it is reasonable to assume that your data are normally distributed.
K-Means Clustering: How Does It Work?
Переглядів 4784 місяці тому
K-Means clustering is a popular technique for partitioning data into distinct groups. In this video I give a summary of how the algorithm works.
Central Composite Designs in JMP Software
Переглядів 6535 місяців тому
One of the most popular design types for a response surface design is the central composite design, consisting of a central factorial design augmented with centre points and axial (star) points. There are 3 variants of this design type: Central Composite Circumscribed (CCC), Central Composite Inscribed (CCI), and Central Composite Face-centred (CCF). These variants are based on how the axial po...
Custom DOE: Comparing a D-Optimal design against an I-Optimal design.
Переглядів 1,3 тис.7 місяців тому
Within JMP Software you can perform design of experiments (DOE) using either classical designs or custom designs. Custom designs are computer-generated based on an algorithm together with an optimisation criterion: the two most common criteria are D-Optimal and I-optimal; so what's the difference? In this video I will compare the two design types.
Multiple ways of performing a simple calculation in JMP Software.
Переглядів 5308 місяців тому
I have results from an experiment. It is an investigation of a material that can reasonably be expected to obey Ohm's Law. If you want to take a look at the data yourself, you can find them here: www.pega-analytics.co.uk/blog/resistance-challenge/ Based on the results of this experiment, what is the resistance? It's a simple calculation - but you will get different answers depending on how you ...
Multiple ways of performing a simple calculation in JMP Software?
Переглядів 518 місяців тому
I have results from an experiment. It is an investigation of a material that can reasonably be expected to obey Ohm's Law. I have a challenge for you … what is the best estimate for the resistance of this material? It seems like a fairly trivial problem to solve, but whenever I ask this question in workshops I get a variety of answers. Please also think about the degree of precision to which yo...
How to use JMP to rapidly identify predictor variables
Переглядів 7758 місяців тому
The Predictor Screening platform in JMP Software is a fast way for identifying important predictor variables. Once they are identified you can use them to propose potential failure modes of a system or use them as variables within a predictive model. This video looks at how to use the platform, and how the results are derived from an ensemble of decision trees.
How to create a dashboard using JMP Software.
Переглядів 1,3 тис.8 місяців тому
The dashboard builder within JMP Software is a convenient tool for combining multiple analysis reports into a single window. But it's not only convenient, it's also smart, and has an awareness of the relative positions of the windows that you are selecting - so it helps to layout your windows correctly prior to creating the dashboard. For a dashboard to be useful you need to combine content tha...
Three layout TIPS for JMP Software
Переглядів 4359 місяців тому
When I am teaching scientists and engineers how to use JMP, often it is the little tips & tricks that help them get the most out of the software. In this video I share 3 tips for changing the layout of results.
I have 2 models. Which one should I choose?
Переглядів 2179 місяців тому
In this video I give an example of how to decide which model to use. I look at a number of statistics including R-square, adjusted R-square, AICc, PRESS, error sum of squares, residuals, p-values, plus a dose of common-sense and experience. I'm using JMP software but it's not about software, its about model building being part art, part science, and about not blindly following rules.
Process Capability: Why you should use the K-Score method and not the Percentile method
Переглядів 16510 місяців тому
Capability indices are defined based on the properties of a Normal distribution. To generalise their definition to non-normal data, there are two methods, the ISO Percentiles method and the K-Score method. If you are working with process capability indices you really need to understand the difference. In this video I discuss both methods, and explain why I think you should consider using the K-...
Exploratory data analysis of (functional) multivariate data using JMP
Переглядів 78710 місяців тому
Some forms of multivariate data are best visualised as a curve. This involves transforming the shape of the data to reveal its functional form. This visualisation can be further enhanced by grouping similar shape curves together - this can be done by performing cluster analysis on the original multivariate structure. Describing this process is much harder than just doing it inside JMP Software!...
How to use JMP's script editor as a calculator
Переглядів 25410 місяців тому
You don't need to be a writing JSL scripts to take advantage of the JSL script editor within JMP Software. You can treat it as a command-line version of JMP, that you can use for performing simple arithmetic operations or more complex statistical calculations that utilise JMP's vast library of mathematical and statistical functions.
Hierarchical Clustering: DENDROGRAMS - what are they, and how are they used?
Переглядів 3,5 тис.11 місяців тому
Hierarchical Clustering: DENDROGRAMS - what are they, and how are they used?
JSL Decoded: How to Write a JSL Script. Part 6.
Переглядів 31611 місяців тому
JSL Decoded: How to Write a JSL Script. Part 6.
JSL Decoded: How to Write a JSL Script. Part 5.
Переглядів 30811 місяців тому
JSL Decoded: How to Write a JSL Script. Part 5.
JSL Decoded: How to Write a JSL Script. Part 4.
Переглядів 31711 місяців тому
JSL Decoded: How to Write a JSL Script. Part 4.
JSL Decoded: How to Write a JSL Script. Part 3.
Переглядів 48511 місяців тому
JSL Decoded: How to Write a JSL Script. Part 3.
JSL Decoded: How to Write a JSL Script. Part 2.
Переглядів 92811 місяців тому
JSL Decoded: How to Write a JSL Script. Part 2.
JSL Decoded: How to Write a JSL Script. Part 1.
Переглядів 2,1 тис.11 місяців тому
JSL Decoded: How to Write a JSL Script. Part 1.
3 Tips for Better Decision Trees in JMP Software
Переглядів 1,1 тис.11 місяців тому
3 Tips for Better Decision Trees in JMP Software
JMP Software: How to Rapidly Discover Relationships Within Your Data
Переглядів 80311 місяців тому
JMP Software: How to Rapidly Discover Relationships Within Your Data
An Introduction to Statistical Design and Analysis of Experiments
Переглядів 2,3 тис.3 роки тому
An Introduction to Statistical Design and Analysis of Experiments
Learn JSL by Example: Oneway Advisor
Переглядів 3,1 тис.4 роки тому
Learn JSL by Example: Oneway Advisor
Model Diagnostics and Transformations
Переглядів 95110 років тому
Model Diagnostics and Transformations
DOE using JMP® Software: Construction and Interpretation of a Response Surface
Переглядів 63 тис.12 років тому
DOE using JMP® Software: Construction and Interpretation of a Response Surface
Thank you so much for the explanation! But I wanted to ask you something: what should I do when I have for instance a and b close, b and c close, but a and c distant? should I still put the three of them together? or should I do them separately, and in this case, which group should I agregate first?
That's an interesting question. We first aggregate the points that are closest - so that would be a and b. You then say a and c are distant. Relative to what? We really need a point d do decide that, I think. Then it's a question of whether c is closer to a&b or closer to d. But there are a couple of further details that we need to think about, and that weren't discussed in this video (I hope I will address these in some new year videos!). First, when we ask about distance of c from a&b what do we mean? We have to define distance more carefully. In the video I used the average location (think of a centre of mass). So we are not comparing the distance of c to a, but the distance of c to the midpoint of a and b. There are alternative distance metrics that we can use - we could take the closest point in a cluster, in which case we would look at the distance of c to b; or we could take the furthest point in a cluster, in which case we would look at the distance of c to a. Different distance metrics allow us to describe different shapes of clusters (think of spherical clusters versus elongated string-like clusters). In all these cases we would compare these distances against the distance from c to d. Any finally, perhaps we want to stop the agglomeration process and keep d and or c separate from a&b because we think that we have reached the optimal number of clusters. To do that we need to define what we mean by "optimal" - but a rule of thumb is to look for the point where we transition from small distances to large distances - and this is something I will discuss in my next video (probably early January because I am decorating the room that I use for recordings!).
Thanks
You're welcome :)
Thanks for practical insights.
You're welcome!
What software you are using for such analysis?
I was using JMP Software.
@PegaAnalytics thank you so much for your reply. Is this software available in online?
It's commercial software, produced originally by SAS: www.jmp.com/
Not clear.
Sorry that it wasn't helpful for you.
@PegaAnalytics it was my short comings not your.
Well I can see that you also produce educational videos. One of the joys of teaching is to try and come up with new ways of explaining concepts. Different explanations work well for different people - ourselves included!
I click the link but error pop up. could you please share jsl. document of lesson 6 to download again? Make Column Switch Handler doesnt work in my code.
Congratulations on getting to the last part of the series. I'm packaged the videos and the associated code into a Udemy course. You'll need to enrol for the course the get access to the files, but it doesn't cost you anything. Plus also, it should give you direct access to me if you have any future JSL questions. Here's the link: www.udemy.com/course/jsl-decoded/
What type of data should we do normality test?
For me personally, I do a lot of process capability analysis, which is very sensitive to the assumption of normality of the data. For many other statistical methods (regression, two-sample t-test, etc) we construct a t-test and derive a p-value; these statistical methods assume that our error terms (residuals) are normally distributed.
Thank you, David. This is the best explanation of dendrogram that I got. I am so grateful to you for not overloading this video with math concepts. Selecting alphabets to represent data points, made it even easier to understand. God bless you sir.
This you so much for taking the time to leave this comment, and for your kind words. This type of feedback gives me huge motivation to make further videos.
Great video!
Thank you! This helped a lot
Great explanation!
Excellent💯 video thank you for sharing!!
Thank you, sensei. This is the most clear explanation I've got about hierarchical clustering. Please I'm interested in learning more about how to know the optimum clusters for a dataset. Thank you.
Thanks for your positive feedback. My next video will discuss the optimal number of clusters
This video is great. Thanks.
Thank you!
Thanks,it`s very useful for me! Look forward to seeing you in live stream. I like 2 pm.
Thank you very insightful 🌹🌹
That's great to hear. Thanks!
Very useful content. Thank you very much
Glad it was helpful!
David: two more use cases: 1) teaching JMP: a journal is a great way to package a training class into one file 2) documenting problem solving: i support manufacturing facilities and am often involved in problem solving. as i work on the problem i save notable reports to a journal. when done, i export the journal to MS Word and then add narration to explain the data, problem solving steps, root causes, and recommendations, etc..
Hi Mark, thanks for your contribution. I totally agree, these are two perfect use-cases for journals. It's interesting that you export a problem-solving journal to MS Word, once you have completed the work. I think that's probably a very helpful step. Something I will look try-out myself!
@@PegaAnalytics Why MS Word? Journal Text Sledgehammer add-in is an improvement, but still nowhere near the features/ease of use of MS Word. Also, customers may not have JMP. MS SharePoint can't index JMP files, but can index Word files.
Thanks a bunch for this simplified and clear explanation, it would be a pleasure if you could share with us how could we make dendrograms from Pulsed field electrophoresis Gel , thank you :)
Funny you should ask that ... the following paper is next on my reading list: "Pulsed-field gel electrophoresis (PFGE) analysis of Listeria monocytogenes isolates from different sources and geographical origins and representative of the twelve serovars" www.academia.edu/111312006/Pulsed_field_gel_electrophoresis_PFGE_analysis_of_Listeria_monocytogenes_isolates_from_different_sources_and_geographical_origins_and_representative_of_the_twelve_serovars
I've looked at this in a bit more detail, and to be honest, handling these type of data is beyond my area of expertise. I did find some general information that I found helpful: A guide to interpreting electrophoresis gels: bento.bio/resources/bento-lab-advice/interpreting-electrophoresis-gels-with-bento-lab/#:~:text=The%20smallest%20bands%20are%20at,is%20up%20the%20ladder%20scale. (pulsed-field addressed larger DNA molecules but I presume the principles on interpretation remain the same). Any analytical technique requires digital data. I found this: Data processing of pulsed-field gel electrophoresis images www.ncbi.nlm.nih.gov/pmc/articles/PMC6940661/ The data processing would seem to me the critical step, which will ultimately result in the generation of tabulated data that would be amenable to cluster analysis. The columns of this tabulation would correspond to metrics that describe the banding, which each sample being represented by a row. I would guess that this data processing is integrated into most laboratory systems that produce pulsed-field electrophoresis gel?
Thank you for helping me understand dendrograms!
Happy to hear the video helped :)
Thank you for the wonderful video! I had a very vague understanding of this concept before watching it. However, after going through the video, everything became crystal clear, and I experienced a profound moment of enlightenment. Your exceptional teaching skills and ability to break down complex ideas into understandable components have truly been an eye-opener for me. I am deeply grateful for your efforts in creating such an informative and insightful resource.
Thank you so much!
how can choose the calculation(simulation)ways will influence the simulation result, so that means must base on the actually status and the experience to choose the perfect mode to simulation, right?
Thanks for the great video! It would be very appreciated if you will discuss how to select the optimum number of clusters in future videos. 🙂
I appreciate your feedback. I'll make a note to make a video about identifying the optimal number of clusters - thanks for the suggestion.
Very helpful video. Congratulation.
Thanks!
This is great. Thanks!
Glad you liked it!
David: Great video! I've been using this platform for about ten years and even I learned a bunch of things from you! For solving manufacturing problems, this platform is outstanding. Thanks for creating this.
Thanks Mark, I always value your feedback.
thank you for video this is amazing content, you are a saint and a scholar cheers
Thank you so much, that is very generous of you!
Great content #PegaAnalytics! Never thought of using arrange in rows in the fit model platform! When would you not use fit separatly?
Hi, thanks for your feedback. I can't think of a good reason to not use the 'fit separately' option. Perhaps I might use it if I am doing a demonstration and want to focus on model visualisation and interpretation without dwelling on the details of model building. But in a real-world scenario, I would always want to spend my time building the best model I can for each response, and for that I would use 'fit separately'. The 'fit separately' option allows you to refine model terms individually for each response, whereas if not selected you refine model terms en-masse across the responses. Prior to version 14 this was the only way of handling multiple responses - apart from running Fit Model separately for each model. And the benefit of specifying multiple responses rather than running multiple Fit Models, is that you can view all the models simultaneously in the prediction profiler and contour profiler, as well as perform numerical optimisations that take account of all the responses.
Can you share the .jsl document of lesson 6 to download?
You should be able to access it here: www.pega-analytics.co.uk/blog/how-to-write-a-jsl-script/
@@PegaAnalytics Thank you!
Thank you so much this is really easy to understand
Glad it was helpful!
Great video! thank you
You are welcome!
I have a JMP question. I am getting only Ppk values, not Cpk values for a non-normal distribution. In your example you shown Cpk. Where is the setting for it? I have search in the JMP preference and I have not found the correct one. Please let me know when you have a moment. Thanks
I had the preference "Ppk Capability Labeling" turned off. This is one of the preferences for the Distribution Platform( File> Preferences> Platforms> Distribution ).
@@PegaAnalytics Thank you
So simple, yet so powerful! I will never look at a fit model the same way again! Thanks for sharing!
Thanks David, How do you get the 84 median? and also is the 5.8 factor dependent of the data set?
Thanks for your questions. The median value of 84 is based on fitting a Lognormal distribution to the data then taking the median of fitted distribution. That median calculation can either be done using a simple script based on the shape and scale parameters for the distribution (see my video on using the script editor as a calculator) or directly from JMP (after fitting a distribution, from the associated red triangle select Profilers> Quantile Profiler, and enter 0.5 for the quantile). Yes, you are correct, the value of 5.8 is specific to these data. Whereas for a Normal distribution we can always say that the width is +/- 3 sigma, that is not true for other distribution types. The value of 5.8 was determined by calculating the interval from the 50% percentile (the median) to the (100-0.135)% percentile. Similar to the median calculation, this can be done with a simple JSL script or by getting the numbers from the Quantile Profiler for the fitted distribution.
Thank you@@PegaAnalytics
Thank you, David! This 30 year JMP user continues to learn from you.
Thanks for your feedback Mark.
Thank you very much great video.
I'm glad you liked it. I appreciate the feedback. Thanks!
Great tips
Thanks!
Thanks for a instructive and simple presentation of the Z-Score method and the Percentile method. I will definitively try the Z-Score method next time I do capability analysis!
Thanks for your positive feedback :)
Very nice to tutorial. Thank you.
Glad you liked it
Hi ,I like all your vedios! Can you make some vedios to introduce the app builder? I want to learn how only use a little script to accomplish a program.
I'm glad you are enjoying the videos. It's really helpful to get requests, so I know what is of interest to people. I'll add the app builder to my list of topics!
@@PegaAnalytics Thanks!
👍
👍
Nice video, David! Going to share it with my team that does global manufacturing support.
That's wonderful. Thanks Mark.
Thank you. I learned a lot.
Glad it was helpful!
Excellent stuff David! Please keep the videos coming!
Thanks Steffen. I appreciate the encouragement!
Great vedio!
Thanks for the feedback. Part 2 tomorrow!
Hello. Thank you for the video. It is interesting. I have downloaded the dataset from JMP Help menu. However, I am interested to know how have you grouped the variables under several sub-groups such as milling, blending, compression, spraying, raw materials
That's a good point. My version of the table was slightly different to the one that you can find in the sample data. On the left of the table, where it lists all the column names, you can select multiple column names and then right-click: you will then see an option to "group columns". JMP will assign a default name to the group e.g. "Mill Time etc.", but you can click on the name and edit it to be something more descriptive. ( You will also find differences in the order of the rows. For illustration purposes I wanted a condition where a control chart had recently gone out of control. To achieve this I think I sorted the data by descending order of mill time. )
awesome stuff please keep posting
Thanks! I really appreciate your feedback.
For 2024 I am planning some new video content relating to DOE. What specific topics would be of interest? Please let me know if the comments :)
RSM CCD design and its analysis to predict the response.
In 2024 I will post videos for the complete build of this oneway advisor ... unless anyone can suggest an alternative use-case ... ? :) The other possibility is to do some of this as a live stream. What are your thoughts, please let me know in the comments.