Descriptive statistics are methods of describing the characteristics of a data set. It includes calculating things such as the average of the data, its spread and the shape it produces.
Descriptive statistics are used to describe and summarize the basic features of data through measures of central tendency like the mean, median, and mode, and measures of variability like range, variance and standard deviation. The mean is the average value and is best for continuous, non-skewed data. The median is less affected by outliers and is best for skewed or ordinal data. The mode is the most frequent value and is used for categorical data. Measures of variability describe how spread out the data is, with higher values indicating more dispersion.
This document provides an overview of inferential statistics. It defines inferential statistics as using samples to draw conclusions about populations and make predictions. It discusses key concepts like hypothesis testing, null and alternative hypotheses, type I and type II errors, significance levels, power, and effect size. Common inferential tests like t-tests, ANOVA, and meta-analyses are also introduced. The document emphasizes that inferential statistics allow researchers to generalize from samples to populations and test hypotheses about relationships between variables.
This document discusses correlation and different types of correlation analysis. It defines correlation as a statistical analysis that measures the relationship between two variables. There are three main types of correlation: (1) simple and multiple correlation based on the number of variables, (2) linear and non-linear correlation based on the relationship between variables, and (3) positive and negative correlation based on the direction of change between variables. The degree of correlation is measured using correlation coefficients that range from -1 to +1. Common methods to study correlation include scatter diagrams and Karl Pearson's coefficient of correlation.
Lecture on Introduction to Descriptive Statistics - Part 1 and Part 2. These slides were presented during a lecture at the Colombo Institute of Research and Psychology.
This document provides an introduction to key statistical concepts and terms. It defines statistics as a branch of mathematics dealing with collecting, organizing, analyzing, and interpreting numerical data. Some key points:
- Data can be quantitative (numerical) or qualitative (descriptive attributes). Population refers to all elements being studied, while a sample is a subset of the population.
- Parameters describe populations and statistics describe samples. Variables differentiate groups within a population or sample.
- Descriptive statistics summarize and present data, while inferential statistics draw conclusions about populations from samples.
- The history of statistics dates back thousands of years to early censuses, though modern statistical theory developed more recently over the 18th-19
This document discusses descriptive statistics used in research. It defines descriptive statistics as procedures used to organize, interpret, and communicate numeric data. Key aspects covered include frequency distributions, measures of central tendency (mode, median, mean), measures of variability, bivariate descriptive statistics using contingency tables and correlation, and describing risk to facilitate evidence-based decision making. The overall purpose of descriptive statistics is to synthesize and summarize quantitative data for analysis in research.
The document discusses t-tests, which are used to compare means between groups. It describes the assumptions of t-tests, the different types of t-tests including independent samples t-tests and dependent samples t-tests, and the steps to conduct t-tests by hand and using SPSS. It provides examples of conducting one-sample t-tests, independent samples t-tests, and dependent samples t-tests, including interpreting the results. It also discusses how to increase statistical power by increasing the difference between means, decreasing variance, increasing sample size, and increasing the alpha level.
Introduction to Statistics - Basic concepts
- How to be a good doctor - A step in Health promotion
- By Ibrahim A. Abdelhaleem - Zagazig Medical Research Society (ZMRS)
This document discusses different types of data including:
- Qualitative data which describes attributes that can be observed but not computed, and quantitative data which can be measured numerically.
- Primary data is collected first-hand for a specific purpose, while secondary data has already been collected in the past.
- Discrete data takes only certain values, while continuous data can take any value in a range.
Ppt for 1.1 introduction to statistical inferencevasu Chemistry
This document provides an introduction to statistical inference. It defines statistics as dealing with collecting, analyzing, and presenting data. The purpose of statistics is to make accurate conclusions or predictions about a population based on a sample. There are two main types of statistics: descriptive statistics, which describes data, and inferential statistics, which helps make predictions and generalizations from data. Statistical inference involves analyzing sample data and making conclusions about the population using statistical techniques, as it is impractical to study entire populations. The key concepts of population, sample, parameters, statistics, and sampling distribution are introduced.
The document defines various statistical measures and types of statistical analysis. It discusses descriptive statistical measures like mean, median, mode, and interquartile range. It also covers inferential statistical tests like the t-test, z-test, ANOVA, chi-square test, Wilcoxon signed rank test, Mann-Whitney U test, and Kruskal-Wallis test. It explains their purposes, assumptions, formulas, and examples of their applications in statistical analysis.
This document discusses inferential statistics, which uses sample data to make inferences about populations. It explains that inferential statistics is based on probability and aims to determine if observed differences between groups are dependable or due to chance. The key purposes of inferential statistics are estimating population parameters from samples and testing hypotheses. It discusses important concepts like sampling distributions, confidence intervals, null hypotheses, levels of significance, type I and type II errors, and choosing appropriate statistical tests.
This document discusses concepts related to data, including collection, organization, presentation, and analysis of data. It defines key terms like qualitative vs quantitative data and primary vs secondary data. It explains methods of collecting primary data through surveys, sampling techniques, and secondary data from published and unpublished sources. The document also covers organizing data through frequency distributions, statistical series, and presenting data in tabular, diagrammatic and graphical forms like pie charts, histograms, bar diagrams and ogives. It concludes with analyzing organized data through measures of central tendency, dispersion, correlation and regression.
This document defines data and different types of data presentation. It discusses quantitative and qualitative data, and different scales for qualitative data. The document also covers different ways to present data scientifically, including through tables, graphs, charts and diagrams. Key types of visual presentation covered are bar charts, histograms, pie charts and line diagrams. Presentation should aim to clearly convey information in a concise and systematic manner.
This document discusses various techniques for analyzing quantitative and qualitative data. It describes editing, coding, classification, and tabulation as methods for processing qualitative data. For quantitative data, it covers univariate analyses like measures of central tendency and dispersion. It also discusses bivariate analyses like correlation and regression, as well as multivariate techniques including multidimensional analysis, factor analysis, and cluster analysis. The goal of data analysis is to discover useful information and support decision making.
The document provides an overview of inferential statistics. It defines inferential statistics as making generalizations about a larger population based on a sample. Key topics covered include hypothesis testing, types of hypotheses, significance tests, critical values, p-values, confidence intervals, z-tests, t-tests, ANOVA, chi-square tests, correlation, and linear regression. The document aims to explain these statistical concepts and techniques at a high level.
This presentation covers statistics, its importance, its applications, branches of statistics, basic concepts used in statistics, data sampling, types of sampling,types of data and collection of data.
This presentation is about Basic Statistics-related to types of Data-Qualitative and Quantitative, and its Examples in everyday life- By: Dr. Farhana Shaheen
This document discusses statistics and their uses in various fields such as business, health, learning, research, social sciences, and natural resources. It provides examples of how statistics are used in starting businesses, manufacturing, marketing, and engineering. Statistics help decision-makers reduce ambiguity and assess risks. They are used to interpret data and make informed decisions. However, statistics also have limitations as they only show averages and may not apply to individuals.
This document summarizes key concepts from an introduction to statistics textbook. It covers types of data (quantitative, qualitative, levels of measurement), sampling (population, sample, randomization), experimental design (observational studies, experiments, controlling variables), and potential misuses of statistics (bad samples, misleading graphs, distorted percentages). The goal is to illustrate how common sense is needed to properly interpret data and statistics.
Statistical inference involves drawing conclusions about a population based on a sample. It has two main areas: estimation and hypothesis testing. Estimation uses sample data to obtain point or interval estimates of unknown population parameters. Hypothesis testing determines whether to accept or reject statements about population parameters. Confidence intervals give a range of values that are likely to contain the true population parameter, with a specified level of confidence such as 90% or 95%.
The document discusses different aspects of research design including what research design is, its key components, and types of research design. It defines research design as the arrangement of conditions for collecting and analyzing data to combine relevance to the research purpose with efficient procedures. The main components of research design discussed are sampling design, observational design, statistical design, and operational design. It also outlines features of a good research design and key concepts like dependent and independent variables, extraneous variables, control, and research hypotheses. Finally, it discusses research design for exploratory, descriptive, diagnostic, and hypothesis-testing research studies.
This document discusses various methods and instruments for collecting data in research studies. It begins by defining data and explaining why data collection is important. It then covers primary and secondary sources of data, as well as internal and external sources. The main methods of collecting primary data discussed are direct personal investigation through interviews, indirect oral investigation, case studies, measurements, and observation. Secondary data sources include published and unpublished sources. The document also discusses self-reported data collection methods like surveys, interviews, and questionnaires. Other methods covered include document review, focus groups, and observation. Mixed methods are also briefly discussed.
This document introduces the concept of data classification and levels of measurement in statistics. It explains that data can be either qualitative or quantitative. Qualitative data consists of attributes and labels while quantitative data involves numerical measurements. The document also outlines the four levels of measurement - nominal, ordinal, interval, and ratio - from lowest to highest. Each level allows for different types of statistical calculations, with the ratio level permitting the most complex calculations like ratios of two values.
1. The document discusses descriptive statistics, which is the study of how to collect, organize, analyze, and interpret numerical data.
2. Descriptive statistics can be used to describe data through measures of central tendency like the mean, median, and mode as well as measures of variability like the range.
3. These statistical techniques help summarize and communicate patterns in data in a concise manner.
This document contains tables and information about quantitative techniques including:
1) An area under the normal curve table that provides the proportion of the normal curve between values of z.
2) A binomial coefficients table that lists coefficients for values up to 20.
3) A table of values of the Poisson probability function for values of m from 0 to 9.
Introduction to Statistics - Basic concepts
- How to be a good doctor - A step in Health promotion
- By Ibrahim A. Abdelhaleem - Zagazig Medical Research Society (ZMRS)
This document discusses different types of data including:
- Qualitative data which describes attributes that can be observed but not computed, and quantitative data which can be measured numerically.
- Primary data is collected first-hand for a specific purpose, while secondary data has already been collected in the past.
- Discrete data takes only certain values, while continuous data can take any value in a range.
Ppt for 1.1 introduction to statistical inferencevasu Chemistry
This document provides an introduction to statistical inference. It defines statistics as dealing with collecting, analyzing, and presenting data. The purpose of statistics is to make accurate conclusions or predictions about a population based on a sample. There are two main types of statistics: descriptive statistics, which describes data, and inferential statistics, which helps make predictions and generalizations from data. Statistical inference involves analyzing sample data and making conclusions about the population using statistical techniques, as it is impractical to study entire populations. The key concepts of population, sample, parameters, statistics, and sampling distribution are introduced.
The document defines various statistical measures and types of statistical analysis. It discusses descriptive statistical measures like mean, median, mode, and interquartile range. It also covers inferential statistical tests like the t-test, z-test, ANOVA, chi-square test, Wilcoxon signed rank test, Mann-Whitney U test, and Kruskal-Wallis test. It explains their purposes, assumptions, formulas, and examples of their applications in statistical analysis.
This document discusses inferential statistics, which uses sample data to make inferences about populations. It explains that inferential statistics is based on probability and aims to determine if observed differences between groups are dependable or due to chance. The key purposes of inferential statistics are estimating population parameters from samples and testing hypotheses. It discusses important concepts like sampling distributions, confidence intervals, null hypotheses, levels of significance, type I and type II errors, and choosing appropriate statistical tests.
This document discusses concepts related to data, including collection, organization, presentation, and analysis of data. It defines key terms like qualitative vs quantitative data and primary vs secondary data. It explains methods of collecting primary data through surveys, sampling techniques, and secondary data from published and unpublished sources. The document also covers organizing data through frequency distributions, statistical series, and presenting data in tabular, diagrammatic and graphical forms like pie charts, histograms, bar diagrams and ogives. It concludes with analyzing organized data through measures of central tendency, dispersion, correlation and regression.
This document defines data and different types of data presentation. It discusses quantitative and qualitative data, and different scales for qualitative data. The document also covers different ways to present data scientifically, including through tables, graphs, charts and diagrams. Key types of visual presentation covered are bar charts, histograms, pie charts and line diagrams. Presentation should aim to clearly convey information in a concise and systematic manner.
This document discusses various techniques for analyzing quantitative and qualitative data. It describes editing, coding, classification, and tabulation as methods for processing qualitative data. For quantitative data, it covers univariate analyses like measures of central tendency and dispersion. It also discusses bivariate analyses like correlation and regression, as well as multivariate techniques including multidimensional analysis, factor analysis, and cluster analysis. The goal of data analysis is to discover useful information and support decision making.
The document provides an overview of inferential statistics. It defines inferential statistics as making generalizations about a larger population based on a sample. Key topics covered include hypothesis testing, types of hypotheses, significance tests, critical values, p-values, confidence intervals, z-tests, t-tests, ANOVA, chi-square tests, correlation, and linear regression. The document aims to explain these statistical concepts and techniques at a high level.
This presentation covers statistics, its importance, its applications, branches of statistics, basic concepts used in statistics, data sampling, types of sampling,types of data and collection of data.
This presentation is about Basic Statistics-related to types of Data-Qualitative and Quantitative, and its Examples in everyday life- By: Dr. Farhana Shaheen
This document discusses statistics and their uses in various fields such as business, health, learning, research, social sciences, and natural resources. It provides examples of how statistics are used in starting businesses, manufacturing, marketing, and engineering. Statistics help decision-makers reduce ambiguity and assess risks. They are used to interpret data and make informed decisions. However, statistics also have limitations as they only show averages and may not apply to individuals.
This document summarizes key concepts from an introduction to statistics textbook. It covers types of data (quantitative, qualitative, levels of measurement), sampling (population, sample, randomization), experimental design (observational studies, experiments, controlling variables), and potential misuses of statistics (bad samples, misleading graphs, distorted percentages). The goal is to illustrate how common sense is needed to properly interpret data and statistics.
Statistical inference involves drawing conclusions about a population based on a sample. It has two main areas: estimation and hypothesis testing. Estimation uses sample data to obtain point or interval estimates of unknown population parameters. Hypothesis testing determines whether to accept or reject statements about population parameters. Confidence intervals give a range of values that are likely to contain the true population parameter, with a specified level of confidence such as 90% or 95%.
The document discusses different aspects of research design including what research design is, its key components, and types of research design. It defines research design as the arrangement of conditions for collecting and analyzing data to combine relevance to the research purpose with efficient procedures. The main components of research design discussed are sampling design, observational design, statistical design, and operational design. It also outlines features of a good research design and key concepts like dependent and independent variables, extraneous variables, control, and research hypotheses. Finally, it discusses research design for exploratory, descriptive, diagnostic, and hypothesis-testing research studies.
This document discusses various methods and instruments for collecting data in research studies. It begins by defining data and explaining why data collection is important. It then covers primary and secondary sources of data, as well as internal and external sources. The main methods of collecting primary data discussed are direct personal investigation through interviews, indirect oral investigation, case studies, measurements, and observation. Secondary data sources include published and unpublished sources. The document also discusses self-reported data collection methods like surveys, interviews, and questionnaires. Other methods covered include document review, focus groups, and observation. Mixed methods are also briefly discussed.
This document introduces the concept of data classification and levels of measurement in statistics. It explains that data can be either qualitative or quantitative. Qualitative data consists of attributes and labels while quantitative data involves numerical measurements. The document also outlines the four levels of measurement - nominal, ordinal, interval, and ratio - from lowest to highest. Each level allows for different types of statistical calculations, with the ratio level permitting the most complex calculations like ratios of two values.
1. The document discusses descriptive statistics, which is the study of how to collect, organize, analyze, and interpret numerical data.
2. Descriptive statistics can be used to describe data through measures of central tendency like the mean, median, and mode as well as measures of variability like the range.
3. These statistical techniques help summarize and communicate patterns in data in a concise manner.
This document contains tables and information about quantitative techniques including:
1) An area under the normal curve table that provides the proportion of the normal curve between values of z.
2) A binomial coefficients table that lists coefficients for values up to 20.
3) A table of values of the Poisson probability function for values of m from 0 to 9.
The document introduces the statistical concepts of mean, median, mode, and range using everyday examples like test scores and family ages. It explains that mean is the average, median is the middle number, mode is the most frequent value, and range is the difference between the highest and lowest numbers. Various examples are provided and explained step-by-step to illustrate how to calculate each statistical measure.
Reporting Statistics in Psychology
This document provides guidelines for reporting statistics in psychology research. It outlines how to round numbers and report means, standard deviations, p-values, effect sizes, and results from t-tests, ANOVAs, and other statistical analyses. Key recommendations include reporting exact p-values to two or three decimal places, using abbreviations like M and SD consistently, and noting any violations of statistical assumptions.
The Content Marketing Metrics That Matter (#CMWorld 2015)PR 20/20
Marketing technology advances have made it easier and more affordable to connect activities to outcomes, but content marketers have largely dropped the ball when it comes to monitoring, reporting and improving performance. The key is to align content marketing KPIs with overall business goals, have a logical and well-documented process for updating and reporting results, and develop systems for turning data into intelligence and intelligence into action.
Attendee takeaways:
* Prioritize content marketing goals.
* Identify Key Performance Indicators (KPIs).
* Select the right analytics technology and tools.
* Optimize your use of Google Analytics.
* Turn data into insights and actions.
This document provides an overview of big data, including definitions, characteristics, and technologies. It defines big data as large datasets that cannot be processed by traditional databases due to size and complexity. It describes the key aspects of big data as volume, variety, velocity, and veracity. The document also discusses how big data differs from traditional transaction systems, the promise and challenges of big data, and Hadoop as a framework for distributed processing of big data.
The document discusses developing an effective enterprise data strategy. It recommends that a data strategy should include identifying and combining multiple data sources, building advanced analytics models, and enabling organizational transformation. An effective strategy also makes data generate business value, identifies critical data assets, defines the data ecosystem, and establishes data governance. The strategy must be flexible, actionable, and provide a clear vision of how data and analytics can improve business results.
The Science behind a Winning Sales CultureBrad Giles
Presentation outlining the different types of salespeople and how a top performing salesperson is impacted by their strengths, skills & severity of weaknesses. Provides a brief summary of the sales assessment we have used on almost 1million salespeople around the world to determine whether they will sell. More details at www.evolutionpartners.com.au
Central tendency refers to measures that characterize the middle or center of a data set. The three most common measures of central tendency are the mean, median, and mode. The mean is the average value found by dividing the sum of all values by the number of values. The median is the middle value when values are arranged from lowest to highest. The mode is the value that occurs most frequently in the data set. These measures help analyze and understand data in a statistical analysis.
This document discusses measures of central tendency (mean, median, mode) and measures of spread (range, variance, standard deviation). It provides formulas and examples to calculate each measure. It also presents two problems, asking to calculate and compare various descriptive statistics for different data sets, such as milk yields from two cow herds and weaning weights of lambs from two breeds. A third problem asks to analyze and compare price data for rice from two markets.
This document discusses key performance indicators (KPIs) for senior staff accountant positions. It provides information on developing KPIs, including defining objectives and key result areas, identifying tasks, and determining how to measure results. The document recommends that KPIs be clearly linked to strategy, answer important questions, and empower employees. It also discusses different types of KPIs and mistakes to avoid when creating KPIs, such as having too many. The document directs the reader to an online source for additional KPI samples and materials.
Measure of dispersion part II ( Standard Deviation, variance, coefficient of ...Shakehand with Life
This tutorial gives the detailed explanation measure of dispersion part II (standard deviation, properties of standard deviation, variance, and coefficient of variation). It also explains why std. deviation is used widely in place of variance. This tutorial also teaches the MS excel commands of calculation in excel.
Training Slides of KPIs, Work flow & evaluating performances discussing the importance of KPI.
For further information regarding the course, please contact:
info@asia-masters.com
www.asia-masters.com
This document discusses key performance indicators (KPIs) for senior accountant positions. It provides steps for creating KPIs for senior accountants, including defining objectives, identifying key result areas and tasks, and determining how to measure results. The document warns against creating too many KPIs and notes that KPIs should be linked to strategy and empower employees. It also lists different types of KPIs and provides a link to additional KPI materials and resources.
Origins of the Marketing Intelligence Engine (INBOUND 2016)PR 20/20
The velocity of change in the marketing industry is accelerating, but what we see today is elementary when we consider the potential of what comes next. This session provides a glimpse into the future of marketing, and the opportunities that exist for those who can harness the power of artificial intelligence and cognitive technology like IBM's Watson. They will be able to do more with less, run personalized campaigns of unprecedented complexity, and analyze massive data sets to predict outcomes. The opportunities are endless for those with the will and vision to transform the industry. Attendees will:
- Learn what the disruption of other industries can teach us about the inevitable impact artificial intelligence will have on the marketing industry.
- Discover existing marketing technologies using artificial intelligence to make marketing more efficient and effective.
- Get inspired to explore what’s possible for the future of marketing, as well as their businesses and careers.
The document provides information on key performance indicators (KPIs). It discusses why KPIs are important for tracking business performance, how to develop a balanced scorecard to measure KPIs across different perspectives like customers, internal processes, learning and growth, and financials. It also provides examples of generic KPI measures and how to implement KPIs through defining strategic goals and drivers, developing new measures, analyzing and reporting on trends, and driving continuous improvement.
One of the best ways to analyze any process is to plot the data. Different graphs can reveal different characteristics of your data such as the central tendency, the dispersion and the general shape for thedistribution.
Data Analyst Interview Questions & AnswersSatyam Jaiswal
Practice Best Data Analyst Interview Questions for the best preparation of the data analyst interview. these interview questions are very popular and asked various times in data analyst interview.
Top 30 Data Analyst Interview Questions.pdfShaikSikindar1
Data Analytics has emerged has one of the central aspects of business operations. Consequently, the quest to grab professional positions within the Data Analytics domain has assumed unimaginable proportions. So if you too happen to be someone who is desirous of making through a Data Analyst .
Unit 2_ Descriptive Analytics for MBA .pptxJANNU VINAY
This document provides an overview of descriptive analytics and data visualization. It discusses descriptive statistics such as measures of central tendency (mean, median, mode) and variability. It also covers data visualization techniques like charts, graphs and dashboards. Key topics include univariate, bivariate and multivariate analysis for data visualization, different types of visualizations, and how to create charts in Microsoft Excel. The document is intended to introduce readers to the fundamental concepts and tools used in descriptive analytics.
Data Processing & Explain each term in details.pptxPratikshaSurve4
Data processing involves converting raw data into useful information through various steps. It includes collecting data through surveys or experiments, cleaning and organizing the data, analyzing it using statistical tools or software, interpreting the results, and presenting findings visually through tables, charts and graphs. The goal is to gain insights and knowledge from the data that can help inform decisions. Common data analysis types are descriptive, inferential, exploratory, diagnostic and predictive analysis. Data analysis is important for businesses as it allows for better customer targeting, more accurate decision making, reduced costs, and improved problem solving.
BA is used to gain insights that inform business decisions and can be used to automate and optimize business processes. Data-driven companies treat their data as a corporate asset and leverage it for a competitive advantage. Successful business analytics depends on data quality, skilled analysts who understand the technologies and the business, and an organizational commitment to data-driven decision-making.
Business analytics examples
Business analytics techniques break down into two main areas. The first is basic business intelligence. This involves examining historical data to get a sense of how a business department, team or staff member performed over a particular time. This is a mature practice that most enterprises are fairly accomplished at using.
This document provides an introduction to data science concepts. It discusses the components of data science including statistics, visualization, data engineering, advanced computing, and machine learning. It also covers the advantages and disadvantages of data science, as well as common applications. Finally, it outlines the six phases of the data science process: framing the problem, collecting and processing data, exploring and analyzing data, communicating results, and measuring effectiveness.
Data Science & AI Road Map by Python & Computer science tutor in MalaysiaAhmed Elmalla
The slides were used in a trial session for a student aiming to learn python to do Data science projects .
The session video can be watched from the link below
https://youtu.be/CwCe1pKOVI8
I have over 20 years of experience in both teaching & in completing computer science projects with certificates from Stanford, Alberta, Pennsylvania, California Irvine universities.
I teach the following subjects:
1) IGCSE A-level 9618 / AS-Level
2) AP Computer Science exam A
3) Python (basics, automating staff, Data Analysis, AI & Flask)
4) Java (using Duke University syllabus)
5) Descriptive statistics using SQL
6) PHP, SQL, MYSQL & Codeigniter framework (using University of Michigan syllabus)
7) Android Apps development using Java
8) C / C++ (using University of Colorado syllabus)
Check Trial Classes:
1) A-Level Trial Class : https://youtu.be/v3k7A0nNb9Q
2) AS level trial Class : https://youtu.be/wj14KpfbaPo
3) 0478 IGCSE class : https://youtu.be/sG7PrqagAes
4) AI & Data Science class: https://youtu.be/CwCe1pKOVI8
https://elmalla.info/blog/68-tutor-profile-slide-share
You can get your trial Class now by booking : https://calendly.com/ahmed-elmalla/30min
And you can contact me on
https://wa.me/0060167074241
by Python & Computer science tutor in Malaysia
DATA ANALYSIS Presentation Computing Fundamentals.pptxAmarAbbasShah1
This document discusses data analysis and provides details on the following:
- It defines data analysis and provides examples of its use.
- It describes the four main types of data analysis: descriptive, diagnostic, predictive, and prescriptive.
- It outlines the six step data analysis process: data requirement gathering, data collection, data cleaning, analyzing data, data interpretation, and data visualization.
- It provides examples to illustrate each type and step of the analysis process.
- It also lists some commonly used data analysis tools.
The document discusses the seven basic quality control tools: (1) flow charts visually illustrate process steps; (2) check sheets collect data at its source; (3) histograms graphically show data distribution; (4) Pareto charts identify the most important causes; (5) cause-and-effect diagrams help determine root causes; (6) control charts distinguish common from special causes of variation; and (7) scatter diagrams study relationships between two variables. Examples are provided for each tool to demonstrate how they are constructed and interpreted for quality improvement.
This document discusses data analytics and related concepts. It defines data and information, explaining that data becomes information when it is organized and analyzed to be useful. It then discusses how data is everywhere and the value of data analysis skills. The rest of the document outlines the methodology of data analytics, including data collection, management, cleaning, exploratory analysis, modeling, mining, and visualization. It provides examples of how data analytics is used in healthcare and travel to optimize processes and customer experiences.
Data Analysis Methods 101 - Turning Raw Data Into Actionable InsightsDataSpace Academy
Data analytics is powerful for organisations. It can help companies improve their overall efficiency and effectiveness. The blog offers a step-by-step narration of the data analysis methods that will help you to comprehend the fundamentals of an analytics project.
Data science involves analyzing data to extract meaningful insights. It uses principles from fields like mathematics, statistics, and computer science. Data scientists analyze large amounts of data to answer questions about what happened, why it happened, and what will happen. This helps generate meaning from data. There are different types of data analysis including descriptive analysis, which looks at past data, diagnostic analysis, which finds causes of past events, and predictive analysis, which forecasts future trends. The data analysis process involves specifying requirements, collecting and cleaning data, analyzing it, interpreting results, and reporting findings. Tools like SAS, Excel, R and Python are used for these tasks.
Data science uses techniques like machine learning and AI to extract meaningful insights from large, complex datasets. It relies on applied mathematics, statistics, and programming to analyze big data. Common data science tools include SAS for statistical analysis, Apache Spark for large-scale processing, BigML for machine learning modeling, Excel for visualization and basic analytics, and programming libraries like TensorFlow, Scikit-learn, and NLTK. These tools help data scientists extract knowledge and make predictions from huge amounts of data.
Focus Group: A Tool for Collecting Valuable Feedback and InsightsCIToolkit
A focus group is a data collection method designed to gather opinions, preferences, insights, and ideas from a group of people on a specific topic. It aims to understand different perspectives by bringing together selected stakeholders and subject matter experts. This method provides an opportunity to obtain qualitative feedback through in-depth discussions, open questions and group dialogue in an interactive group setting. The goal is to capture information and generate ideas through guided discussions and group dynamics.
Simplifying Data Collection and Quality Control with Check SheetsCIToolkit
A check sheet is a manual tool designed to collect and record data in real time, often at the location where the data is generated. Its primary purpose is to simplify and organize data collection while helping analysts identify facts, insights, and patterns that support further analysis. The data recorded on a check sheet can be either quantitative or qualitative, depending on the context.
Designing and Using Questionnaires to Drive Business DecisionsCIToolkit
A questionnaire is a set of questions or statements presented to participants to gather information. It is designed to quickly collect qualitative and quantitative data from a large number of respondents about a particular issue of interest. A questionnaire is a relatively inexpensive way to collect data, and the results are generally easy to analyze. While it is often part of a survey, it can also stand alone as a data collection tool.
Decision Balance Sheet: Evaluating Pros and Cons for Better DecisionsCIToolkit
One of the simplest ways to make decisions is by using a Decision Balance Sheet. It is a simple but effective tool that helps individuals and groups systematically evaluate the advantages and disadvantages of a specific idea or solution. It uses a table to record these advantages and disadvantages for reaching a balanced decision more quickly and confidently, either individually or collaboratively.
Mistake-Proofing: A Lean Approach to Error PreventionCIToolkit
Mistake-proofing is a Lean technique that is used to prevent errors or reduce the chances of their occurrence in processes and products. It involves adding simple design features or modifying existing ones to make it nearly impossible for mistakes to occur. It also involves detecting errors early and allowing for immediate correction before problems escalate to larger issues.
Interviews: A Fundamental Tool for Gathering Valuable InsightsCIToolkit
An interview is a common technique designed to extract information from individuals or groups by engaging them in direct conversations, asking relevant questions, and recording the responses. Typically, interviews are conducted between an interviewer and an interviewee, but they may involve multiple interviewers or participants depending on the context.
Brainstorming: A Key Technique for Idea Generation and Problem SolvingCIToolkit
Brainstorming is a widely used creativity technique designed to generate and collect multiple ideas on a specific topic. It aims to produce as many ideas as possible within a short period of time. It is so popular as a tool for gathering data and ideas and an effective method for exploring problems and generating solutions.
Continuous Improvement Infographics for LearningCIToolkit
The purpose of this section is to provide all the continuous improvement tools in an infographic format. These flashcards are easy to read and understand, and very useful if you are looking for brief, concise, and to-the-point summaries. They are quick refreshers for continuous improvement and can speed up the learning process.
Continuous Improvement Posters for LearningCIToolkit
The intention of this section is to provide all the continuous improvement tools in a poster format that is easy to print and share. These posters are great tools for training, sharing and posting, and can also be distributed as hand-outs during continuous improvement workshops.
Simplifying Complexity: How the Four-Field Matrix Reshapes ThinkingCIToolkit
A Four Field Matrix is an effective model for planning, organizing and making decisions. It is a two-dimensional chart that consists of four equal-sized quadrants, each will describe different aspects of information.
Unlocking Productivity and Personal Growth through the Importance-Urgency MatrixCIToolkit
Importance Urgency Matrix is an effective method of organizing priorities. It is a two-dimensional chart that is used to prioritize work activities as well as personal activities.
Measuring True Process Yield using Robust Yield MetricsCIToolkit
Process yield measures should be able to expose even the smallest inefficiencies within a process, empowering operations to understand their true process yield in order to set realistic targets for improvement. Many organizations employ two primary measures of process yield: First Time Yield (FTY) and Final Yield (FY).
Beyond the Five Whys: Exploring the Hierarchical Causes with the Why-Why DiagramCIToolkit
A why-why diagram is used to identify the root causes of a problem when there are multiple factors to consider. There may be multiple answers at each stage, and each of these answers need to go through a separate process of the why-whys analysis. It is an extension of the 5 Whys approach where they are similar in that they both ask the same Why question multiple times. #WhyWhyDiagram
How-How Diagram: A Practical Approach to Problem ResolutionCIToolkit
How- How Diagram is used when seeking a practical solution to a problem. It works by repeatedly asking: How can this be solved. Multiple answers can be given for a single question, and therefore the result can be represented in a hierarchical tree format.
From Goals to Actions: Uncovering the Key Components of Improvement RoadmapsCIToolkit
An improvement roadmap is an approach used to achieve improvement. It is used to guide through the implementation of a long-term improvement journey. It helps us to understand where we are now as well as where we want to go.
Paired Comparison Analysis: A Practical Tool for Evaluating Options and Prior...CIToolkit
Paired Comparison Analysis is an activity for evaluating a small range of options by comparing them against each other. It is an easy and useful tool for rating and ranking alternatives for decision making where evaluation criteria are subjective.
From Red to Green: Enhancing Decision-Making with Traffic Light AssessmentCIToolkit
Traffic Light Assessment is a rating system for evaluating the performance of a process or variable in relation to a goal. It is a good way to communicate information and have the advantage of being universally recognized by everyone.
Mind Mapping: A Visual Approach to Organize Ideas and ThoughtsCIToolkit
Visually organizing ideas, thoughts and information around a single topic or problem. Mind mapping has many applications in personal, professional and educational situations.
Adapting to Change: Using PEST Analysis for Better Decision-MakingCIToolkit
A strategic and structured planning tool for evaluating the external environment of an organization. PEST stands for Political, Economic, Social, and Technological external factors.
Shane Windmeyer on The Unshakable Importance of DEI in Modern SocietyShane Windmeyer
In every workplace, classroom, boardroom, and public square, the conversation around Diversity, Equity, and Inclusion (DEI) continues to grow in prominence and necessity. DEI is not simply a matter of political correctness or public relations—it is an evolving ethical framework that promotes justice, community, and excellence.
The three pillars of DEI are often misunderstood. Diversity involves acknowledging and embracing differences in identities, perspectives, and backgrounds. Equity seeks to provide individuals with the resources and opportunities they need to succeed, accounting for systemic barriers. Inclusion ensures that all voices are heard, valued, and empowered.
Why Mastering Emotional Intelligence Sets Great Copywriters Apart.pdfSOFTTECHHUB
I've been in the world of words and communication for a long time, and if there's one thing that has remained steadfastly true, it's the power of good copywriting. It’s the bridge between an idea and understanding, a product and a customer, a message and a heart. In recent years, a concept that I believe has become absolutely central to impactful writing is emotional intelligence, or EI. It's that almost magical ability we humans have to connect with one another on a deeper level, to understand and share feelings.
The Bright Minds Top 5 Unstoppable Leaders In 2025 (Final File).pdfaspirenavigators
Meet the trailblazers redefining success, innovation, and leadership in 2025. These visionaries are not just leading—they're transforming the future. read full issue on our website.
https://aspirenavigators.com/
HR & OD professionals navigate inclusion, wellbeing, and change while serving their communities.
This session explores servant leadership’s impact on engagement, retention, and performance
through the evidence-based Engaging Leadership model.
Dissertation Outline: Employee Engagement and Retention in Indian Start-upsbeppamgadu
Here is a concise summary prompt you can give to an AI to create a PowerPoint presentation based on the uploaded report:
---
**Summary Prompt for PPT Creation:**
"Create a comprehensive PowerPoint presentation summarizing a research dissertation on 'The Role of Employee Engagement in Employee Retention in Indian Start-ups.' Include the following key sections:
1. **Introduction:** Context, importance, and research objectives.
2. **Literature Review:** Key concepts of employee engagement (Kahn, Schaufeli), dimensions (vigor, dedication, absorption), and their link to retention.
3. **Research Methodology:** Quantitative approach, sample size (300 employees), data collection tools (UWES scale, surveys), and analysis techniques.
4. **Results & Findings:** Engagement levels, demographic insights, correlation and regression analysis showing the impact of engagement on retention, with emphasis on dedication.
5. **Discussion:** Interpretation of results, practical implications for Indian start-ups, and role of leadership and culture.
6. **Conclusion:** Summary of key findings, importance of engagement, and suggestions for future research.
7. **Additional:** Visuals like data tables, charts, and graphs illustrating engagement scores and retention intentions.
Make the slides visually engaging, use concise bullet points, and highlight major insights with relevant graphics or icons."
---
Would you like me to help craft a more detailed slide outline or content for each slide?
Winning at Work_ Creative Employee Wellness Challenges That Actually Work.pdfEnterprise Wired
Looking to build healthier, happier teams? Learn how thoughtful employee wellness challenges can transform your workplace culture and drive real results.
How does critical thinking relate to problem solving and decision making?Writegenic AI
Explore how critical thinking enhances problem solving and decision making by promoting logical analysis, evaluating solutions, and driving effective outcomes.
SpatzAI - Fairer Teamwork Globally by Addressing Micro-Conflict EarlyDesmond Sherlock
Workplace conflicts don’t always start with blowups—they begin as tiny spats, snide remarks, or ignored frustrations. SpatzAI is a real-time chat app and team review platform that empowers employees to address micro-conflicts before they escalate.
This deck walks you through:
The problem of unreported, unresolved micro-conflict
The 3-step Spatz process: Caution, Objection, Stop
A realistic scenario between two coworkers, Hanna and Pablo
How teams and AI collaborate to resolve issues fairly
Predicted Spatz data showing team impact
Whether you're a founder, team lead, or HR innovator, SpatzAI helps your team communicate with clarity, caution, and courage.
“Scott M. Graffius swings a wrecking ball at the farce of ‘corporate Agile co...Scott M. Graffius
“In ‘Agile Protocol: The Transformation Ultimatum,’ Scott M. Graffius swings a wrecking ball at the farce of ‘corporate Agile cosplay’—where companies slap on the façade of Agile but dodge its soul.”
— excerpt from https://scottgraffius.com/agile-protocol.html
#Agile #AgileProtocol #CorporateAgileCosplay
2. Continuous Improvement Toolkit . www.citoolkit.com
The Continuous Improvement Map
Check Sheets
Data
Collection
Process MappingFlowcharting
Flow Process Charts**
Just in Time
Control Charts
Mistake Proofing
Relations Mapping
Understanding
Performance**
Fishbone Diagram
Design of Experiment
Implementing
Solutions***
Group Creativity
Brainstorming Attribute Analysis
Selecting & Decision Making
Decision Tree
Cost Benefit Analysis
Voting
Planning & Project Management*
Kaizen Events
Quick Changeover
Managing
Risk
FMEA
PDPC
RAID Log*
Observations
Focus Groups
Understanding
Cause & Effect
Pareto Analysis
IDEF0
5 Whys
Kano
KPIs
Lean Measures
Importance-Urgency Mapping
Waste Analysis**
Fault Tree Analysis
Morphological Analysis
Benchmarking***
SCAMPER***
Matrix Diagram
Confidence Intervals
Pugh Matrix
SIPOC*
Prioritization Matrix
Stakeholder Analysis
Critical-to Tree
Paired Comparison
Improvement Roadmaps
Interviews
Quality Function Deployment
Graphical Analysis
Lateral Thinking
Hypothesis Testing
Visual Management
Reliability Analysis
Cross Training
Tree Diagram*
ANOVA
Gap Analysis*
Traffic Light Assessment
TPN Analysis
Decision Balance Sheet
Risk Analysis*
Automation
Simulation
Service Blueprints
DMAIC
Product Family MatrixRun Charts
TPM
Control Planning
Chi-Square
SWOT Analysis
Capability Indices
Policy Deployment
Data collection planner*
Affinity DiagramQuestionnaires
Probability Distributions
Bottleneck Analysis
MSA
Descriptive Statistics
Cost of Quality*
Process Yield
Histograms 5S
Pick Chart
Portfolio Matrix
Four Field Matrix
Root Cause Analysis Data Mining
How-How Diagram***Sampling
Spaghetti **
Mind Mapping*
Project Charter
PDCA
Designing & Analyzing Processes
CorrelationScatter Plots Regression
Gantt Charts
Activity NetworksRACI Matrix
PERT/CPMDaily Planning
MOST
Standard work Document controlA3 Thinking
Multi vari Studies
OEE
Earned Value
Delphi Method
Time Value Map**
Value Stream Mapping**
Force Field Analysis
Payoff Matrix
Suggestion systems Five Ws
Process Redesign
Break-even Analysis
Value Analysis**
FlowPull
Ergonomics
3. Continuous Improvement Toolkit . www.citoolkit.com
Statistics is concerned with the describing, interpretation and
analyzing of data.
It is, therefore, an essential element in any improvement
process.
Statistics is often categorized into descriptive and inferential
statistics.
It uses analytical methods which provide
the math to model and predict variation.
It uses graphical methods to help making
numbers visible for communication
purposes.
- Descriptive Statistics
4. Continuous Improvement Toolkit . www.citoolkit.com
Why do we Need Statistics?
To find why a process behaves the way it does.
To find why it produces defective goods or services.
To center our processes on ‘Target’ or ‘Nominal’.
To check the accuracy and precision of the process.
To prevent problems caused by assignable causes
of variation.
To reduce variability and improve process capability.
To know the truth about the real world.
- Descriptive Statistics
5. Continuous Improvement Toolkit . www.citoolkit.com
Descriptive Statistics:
Methods of describing the characteristics of a data set.
Useful because they allow you to make sense of the data.
Helps exploring and making conclusions about the data in order
to make rational decisions.
Includes calculating things such as the average of the data, its
spread and the shape it produces.
- Descriptive Statistics
6. Continuous Improvement Toolkit . www.citoolkit.com
For example, we may be concerned about describing:
• The weight of a product in a production line.
• The time taken to process an application.
- Descriptive Statistics
7. Continuous Improvement Toolkit . www.citoolkit.com
Descriptive statistics involves describing, summarizing and
organizing the data so it can be easily understood.
Graphical displays are often used along with the quantitative
measures to enable clarity of communication.
- Descriptive Statistics
8. Continuous Improvement Toolkit . www.citoolkit.com
When analyzing a graphical display, you can draw conclusions
based on several characteristics of the graph.
You may ask questions such ask:
• Where is the approximate middle, or center, of the graph?
• How spread out are the data values on the graph?
• What is the overall shape of the graph?
• Does it have any interesting patterns?
- Descriptive Statistics
9. Continuous Improvement Toolkit . www.citoolkit.com
Outlier:
A data point that is significantly greater or smaller than other
data points in a data set.
It is useful when analyzing data to identify outliers
They may affect the calculation of descriptive
statistics.
Outliers can occur in any given data set and in
any distribution.
- Descriptive Statistics
10. Continuous Improvement Toolkit . www.citoolkit.com
Outlier:
The easiest way to detect them is by graphing the data or using
graphical methods such as:
• Histograms.
• Boxplots.
• Normal probability plots.
- Descriptive Statistics
*●
11. Continuous Improvement Toolkit . www.citoolkit.com
Outlier:
Outliers may indicate an experimental error or incorrect
recording of data.
They may also occur by chance.
• It may be normal to have high or low data points.
You need to decide whether to exclude them
before carrying out your analysis.
• An outlier should be excluded if it is due to
measurement or human error.
- Descriptive Statistics
12. Continuous Improvement Toolkit . www.citoolkit.com
Outlier:
This example is about the time taken to process a sample of
applications.
- Descriptive Statistics
Outlier
0 1 2 3 4 5 6 7 8 9
2.8 8.7 0.7 4.9 3.4 2.1 4.0
It is clear that one data point is far distant from the rest of the values.
This point is an ‘outlier’
13. Continuous Improvement Toolkit . www.citoolkit.com
The following measures are used to describe a data set:
Measures of position (also referred to as central tendency or
location measures).
Measures of spread (also referred to as variability or dispersion
measures).
Measures of shape.
- Descriptive Statistics
14. Continuous Improvement Toolkit . www.citoolkit.com
If assignable causes of variation are affecting the process, we
will see changes in:
• Position.
• Spread.
• Shape.
• Any combination of the three.
- Descriptive Statistics
15. Continuous Improvement Toolkit . www.citoolkit.com
Measures of Position:
Position Statistics measure the data central tendency.
Central tendency refers to where the data is centered.
You may have calculated an average of some kind.
Despite the common use of average, there are different
statistics by which we can describe the average of a data set:
• Mean.
• Median.
• Mode.
- Descriptive Statistics
16. Continuous Improvement Toolkit . www.citoolkit.com
Mean:
The total of all the values divided by the size of the data set.
It is the most commonly used statistic of position.
It is easy to understand and calculate.
It works well when the distribution is symmetric and there are
no outliers.
The mean of a sample is denoted by ‘x-bar’.
The mean of a population is denoted by ‘μ’.
- Descriptive Statistics
0 1 2 3 4 5 6 7 8 9
Mean
17. Continuous Improvement Toolkit . www.citoolkit.com
Median:
The middle value where exactly half of the data values are
above it and half are below it.
Less widely used.
A useful statistic due to its robustness.
It can reduce the effect of outliers.
Often used when the data is nonsymmetrical.
Ensure that the values are ordered before calculation.
With an even number of values, the median is the mean of the
two middle values.
- Descriptive Statistics
0 1 2 3 4 5 6 7 8 9
MeanMedian
19. Continuous Improvement Toolkit . www.citoolkit.com
Why can the mean and median be different?
- Descriptive Statistics
0 1 2 3 4 5 6 7 8 9
MeanMedian
20. Continuous Improvement Toolkit . www.citoolkit.com
Mode:
The value that occurs the most often in a data set.
It is rarely used as a central tendency measure
It is more useful to distinguish between unimodal and
multimodal distributions
• When data has more than one peak.
- Descriptive Statistics
21. Continuous Improvement Toolkit . www.citoolkit.com
Measures of Spread:
The Spread refers to how the data deviates from the position
measure.
It gives an indication of the amount of variation in the process.
• An important indicator of quality.
• Used to control process variability and improve quality.
All manufacturing and transactional
processes are variable to some degree.
There are different statistics by which
we can describe the spread of a data set:
• Range.
• Standard deviation.
- Descriptive Statistics
Spread
22. Continuous Improvement Toolkit . www.citoolkit.com
Range:
The difference between the highest and the lowest values.
The simplest measure of variability.
Often denoted by ‘R’.
It is good enough in many practical cases.
It does not make full use of the available data.
It can be misleading when the data is skewed or in the presence
of outliers.
• Just one outlier will increase
the range dramatically.
- Descriptive Statistics
0 1 2 3 4 5 6 7 8 9
Range
23. Continuous Improvement Toolkit . www.citoolkit.com
Standard Deviation:
The average distance of the data points from their own mean.
A low standard deviation indicates that the data points are
clustered around the mean.
A large standard deviation indicates that they are widely
scattered around the mean.
The standard deviation of a sample is
denoted by ‘s’.
The standard deviation of a population
is denoted by “μ”.
- Descriptive Statistics
24. Continuous Improvement Toolkit . www.citoolkit.com
Standard Deviation:
Perceived as difficult to understand because it is not easy to
picture what it is.
It is however a more robust measure of variability.
Standard deviation is computed as follows:
- Descriptive Statistics
Mean (x-bar)
s = standard deviation
x = mean
x = values of the data set
n = size of the data set
s =
∑ ( x – x )2
n - 1
25. Continuous Improvement Toolkit . www.citoolkit.com
Exercise:
This example is about the time taken to process a sample of
applications.
Find the mean, median, range and standard deviation for the
following set of data: 2.8, 8.7, 0.7, 4.9, 3.4, 2.1 & 4.0.
- Descriptive Statistics
Time allowed: 10 minutes
26. Continuous Improvement Toolkit . www.citoolkit.com
If someone hands you a sheet of data and asks you to find the
mean, median, range and standard deviation, what do you do?
- Descriptive Statistics
21 19 20 24 23 21 26 23
25 24 19 19 21 19 25 19
23 23 15 22 23 20 14 20
15 19 20 21 17 15 16 19
13 17 19 17 22 20 18 16
17 18 21 21 17 20 21 21
21 17 17 19 21 22 25 20
19 20 24 28 26 26 25 24
27. Continuous Improvement Toolkit . www.citoolkit.com
Measures of Shape:
Data can be plotted into a histogram to have a general idea of
its shape, or distribution.
The shape can reveal a lot of information about the data.
Data will always follow some know distribution.
- Descriptive Statistics
28. Continuous Improvement Toolkit . www.citoolkit.com
Measures of Shape:
It may be symmetrical or nonsymmetrical.
In a symmetrical distribution, the two sides of the distribution
are a mirror image of each other.
Examples of symmetrical distributions include:
• Uniform.
• Normal.
• Camel-back.
• Bow-tie shaped.
- Descriptive Statistics
29. Continuous Improvement Toolkit . www.citoolkit.com
Measures of Shape:
The shape helps identifying which descriptive statistic is more
appropriate to use in a given situation.
If the data is symmetrical, then we may use the mean or median
to measure the central tendency as they are almost equal.
If the data is skewed, then the median will be a more
appropriate to measure the central tendency.
Two common statistics that measure the shape of the data:
• Skewness.
• Kurtosis.
- Descriptive Statistics
30. Continuous Improvement Toolkit . www.citoolkit.com
Skewness:
Describes whether the data is distributed symmetrically around
the mean.
A skewness value of zero indicates perfect symmetry.
A negative value implies left-skewed data.
A positive value implies right-skewed data.
- Descriptive Statistics
XXX
XXX
XXX
XX
XX
X
X
X
X
XXX
X
X
X
X
XXX
XXX
XXX
XX
XX
X
X
X
X
XXX
X
X
X
X
(+) – SK > 0 (-) – SK < 0
31. Continuous Improvement Toolkit . www.citoolkit.com
Kurtosis:
Measures the degree of flatness (or peakness) of the shape.
When the data values are clustered around the middle, then the
distribution is more peaked.
• A greater kurtosis value.
When the data values are spread around more evenly, then the
distribution is more flatted.
• A smaller kurtosis values.
- Descriptive Statistics
XXXXX
XXX
XXX
X
X
X
XXXXX
XXX
XXXX
XX
XX
X
XXX
XXX
XXX
XX
XX
X
X
X
(-) Platykurtic (0) Mesokurtic (+) Leptokurtic
32. Continuous Improvement Toolkit . www.citoolkit.com
Skewness and kurtosis statistics can be evaluated visually via a
histogram.
They can also be calculated by hand.
This is generally unnecessary with modern statistical software
(such as Minitab).
- Descriptive Statistics
33. Continuous Improvement Toolkit . www.citoolkit.com
Further Information:
Variance is a measure of the variation around the mean.
It measures how far a set of data points are spread out from
their mean.
The units are the square of the units used for the original data.
• For example, a variable measured in meters will have a variance
measured in meters squared.
It is the square of the standard deviation.
- Descriptive Statistics
Variance = s2
34. Continuous Improvement Toolkit . www.citoolkit.com
Further Information:
The Inter Quartile Range is also used to measure
variability.
Quartiles divide an ordered data set into 4 parts.
Each contains 25% of the data.
The inter quartile range contains the middle
50% of the data (i.e. Q3-Q1).
It is often used when the data is not normally
distributed.
- Descriptive Statistics
25%
Interquartile Range
25%
25%
25%
50%
35. Continuous Improvement Toolkit . www.citoolkit.com
Minitab is a statistical software that allows you to enter your
data to perform a wide range of statistical analyses.
It can be used to calculate many types of descriptive statistics.
It tells you a lot about your data in order to make more rational
decisions.
Descriptive statistics summaries in Minitab
can be either quantitative or visual.
- Descriptive Statistics in Minitab
Descriptive Statistics
36. Continuous Improvement Toolkit . www.citoolkit.com
Example:
A hospital is seeking to detect the presence of high glucose
levels in patients at admission.
You may use the glucose_level_fasting worksheet or use data
that you have collected yourself.
Remember to copy the data from
the excel sheet and paste it into
Minitab worksheet.
- Descriptive Statistics in Minitab
79 72 77 85 76 120 78 94
93 70 79 75 68 73 79 85
98 77 77 88 79 79 70 113
75 80 74 83 85 79 87 82
104 106 81 76 68 72 61 95
78 106 84 70 96 70 90 98
69 60 74 67 71 75 105 79
71 75 131 80 75 52 152 106
81 96
37. Continuous Improvement Toolkit . www.citoolkit.com
Example:
To create a quantitative summary of your data:
• Select Stat > Basic Statistics > Display Descriptive Statistics.
• Select the variable to be analyzed, in this case ‘glucose level’.
• Click OK.
Here is a screenshot of the various
descriptive statistics you may
choose when doing your analysis.
- Descriptive Statistics in Minitab
38. Continuous Improvement Toolkit . www.citoolkit.com
Example:
Here is a screenshot of the example result:
- Descriptive Statistics in Minitab
Quantitative Summary
39. Continuous Improvement Toolkit . www.citoolkit.com
Example:
To create a visual summary of your data:
• Select Stat > Basic Statistics > Graphical Summary.
• Select the variable to be analyzed, in this case ‘glucose level’.
• Click OK.
Here is a screenshot
of the example result:
- Descriptive Statistics in Minitab
40. Continuous Improvement Toolkit . www.citoolkit.com
Example:
By default, Minitab fits a normal distribution curve to the
histogram.
A boxplot will also be shown to
display the four quartiles of the
data.
The 95% confidence intervals are
also shown to illustrate where the
mean and median of the population
lie.
- Descriptive Statistics in Minitab
41. Continuous Improvement Toolkit . www.citoolkit.com
Example:
Mean, standard deviation, sample size, and other descriptive
statistic values are shown in the adjacent data table.
The skewed distribution shows the
differences that can occur between
the mean and median.
The mean is pulled to the right by the
high value outliers.
The positive value for skewness indicates
a positive skew of the data set.
- Descriptive Statistics in Minitab