Statistical Analysis: FUNDAMENTAL TO STATISTICS.pptxDr SHAILAJA
"Fundamentals of Statistics" refers to the foundational concepts and principles that underlie statistical analysis. It encompasses a variety of topics essential for understanding how to collect, analyze, interpret, and present data.
Fundamentals of statistics, which is very important to know the basics and significance of statistics.explained with different types of statistics with examples. which can be applied in research purpose and helps in calculating how to calculate the central tendency with mean, median, and mode.
These fundamentals are crucial for applying statistical techniques in various fields such as business, healthcare, social sciences, and research, enabling informed decision-making based on data.
This document discusses descriptive and inferential statistics used in nursing research. It defines key statistical concepts like levels of measurement, measures of central tendency, descriptive versus inferential statistics, and commonly used statistical tests. Nominal, ordinal, interval and ratio are the four levels of measurement, with ratio allowing the most data manipulation. Descriptive statistics describe sample data while inferential statistics allow estimating population parameters and testing hypotheses. Common descriptive statistics include mean, median and mode, while common inferential tests are t-tests, ANOVA, chi-square and correlation. Type I errors incorrectly reject the null hypothesis.
Statistics is the study of collecting, organizing, summarizing, and interpreting data. Medical statistics applies statistical methods to medical data and research. Biostatistics specifically applies statistical methods to biological data. Statistics is essential for medical research, updating medical knowledge, data management, describing research findings, and evaluating health programs. It allows comparison of populations, risks, treatments, and more.
Statistical analysis involves investigating trends, patterns, and relationships using quantitative data. It requires careful planning from the start, including specifying hypotheses and designing the study. After collecting sample data, descriptive statistics summarize and organize the data, while inferential statistics are used to test hypotheses and make estimates about populations. Key steps in statistical analysis include planning hypotheses and research design, collecting a sufficient sample, summarizing data with measures of central tendency and variability, and testing hypotheses or estimating parameters with techniques like regression, comparison tests, and confidence intervals. The results must be interpreted carefully in terms of statistical significance, effect sizes, and potential decision errors.
This document provides an outline for a course on probability and statistics. It begins with an introduction to key concepts like measures of central tendency, dispersion, correlation, and probability distributions. It then lists common probability distributions and hypothesis testing. The document provides examples of how statistics is used in various fields. It also defines key statistical concepts like population and sample, variables, and different scales of measurement. Finally, it discusses data collection methods and ways to represent data through tables and graphs.
This document provides an outline for a course on probability and statistics. It includes an introduction to key statistical concepts like measures of central tendency, dispersion, correlation, probability distributions, and hypothesis testing. Assignments are provided to help students apply these statistical methods to real-world examples from various fields like business, engineering, and the biological sciences. References for further reading on topics in statistics and probability are also listed.
This document provides an outline for a course on probability and statistics. It begins with an introduction to statistics, including definitions and general uses. It then covers various topics that will be taught, such as measures of central tendency, probability, discrete and continuous distributions, and hypothesis testing. References for textbooks are also provided. The document contains sample assignments and examples to illustrate concepts like scales of measurement, data collection methods, and graphical representations of data. It provides instructions for calculating measures of central tendency and examples of frequency distributions and their related graphs.
This document provides an outline for a course on probability and statistics. It begins with an introduction to statistics, including definitions and general uses. It then covers topics like measures of central tendency, probability, discrete and continuous distributions, and hypothesis testing. References for textbooks on the subject are also provided. Assignments include calculating measures of central tendency and constructing frequency distributions from raw data. Various scales of measurement and methods of data collection are defined. Graphical representations like histograms, pie charts, and bar graphs are discussed. Formulas are given for calculating the mean, median, and mode of both grouped and ungrouped data.
This document provides an outline for a course on probability and statistics. It begins with an introduction to statistics, including definitions and general uses. It then covers topics like measures of central tendency, probability, discrete and continuous distributions, and hypothesis testing. References for textbooks on statistics and counterexamples in probability are also provided. Assignments ask students to list contributors to statistics, apply statistics in real life, define independent and dependent variables, and understand scales of measurement. Methods of data collection, tabular and graphical representation of data, and measures of central tendency and location are also discussed.
This document provides an outline for a course on probability and statistics. It begins with an introduction to key concepts like measures of central tendency, dispersion, correlation, and probability distributions. It then lists common probability distributions and the textbook and references used. Later sections define important statistical terms like population, sample, variable types, data collection methods, and ways of presenting data through tables and graphs. It provides examples of how statistics is used and ends with examples of different variable scales.
This document provides an outline for a course on probability and statistics. It begins with an introduction to key concepts like measures of central tendency, dispersion, correlation, and probability distributions. It then lists common probability distributions and the textbook and references used. Later sections define important statistical terms like population, sample, variable types, data collection methods, and ways of presenting data through tables and graphs. It provides examples of each variable scale and ends with assignments for students.
Basics of Educational Statistics (Inferential statistics)HennaAnsari
This document provides information about inferential statistics presented by Dr. Hina Jalal. It defines inferential statistics as using data from a sample to make inferences about the larger population from which the sample was taken. It discusses key areas of inferential statistics like estimating population parameters and testing hypotheses. It also explains the importance of inferential statistics in research for making conclusions from samples, comparing models, and enabling inferences about populations based on sample data. Flow charts are presented for selecting common statistical tests for comparisons, correlations, and regression.
This document provides an overview of key concepts in statistics. It discusses that statistics involves collecting, organizing, analyzing and interpreting data. It also defines important statistical terms like population, sample, parameter, statistic, qualitative and quantitative data, independent and dependent variables, discrete and continuous variables, and different levels of measurement for variables. The different levels of measurement are nominal, ordinal, interval and ratio. Descriptive statistics are used to summarize and describe data, while inferential statistics allow making inferences about populations from samples.
This document provides an overview of basic concepts in inferential statistics. It defines descriptive statistics as describing and summarizing data through measures like mean, median, variance and standard deviation. Inferential statistics is defined as using sample data and statistics to draw conclusions about populations through hypothesis testing and estimates. Key concepts explained include parameters, statistics, sampling distributions, null and alternative hypotheses, and the hypothesis testing process. Examples of descriptive and inferential analyses are also provided.
kelan nyo isubmit yung assignment no. 7 and 8 nyo nasa slides yun ng stats. isubmit nyo sa akin sa lunes during electromagnetism kasi kukulangin yung class participation nyo sa stats.
1. The document discusses quantitative research methods, including comparing groups, examining relationships between variables, different types of data and levels of measurement, sampling techniques, and common statistical tools.
2. Key statistical tools covered include t-tests, ANOVA, correlation analysis, chi-square tests, and non-parametric equivalents for comparing groups and examining relationships.
3. The purpose of quantitative research is to systematically investigate phenomena through collecting and analyzing numerical data.
Statistics is the collection, organization, analysis, and presentation of data. It has become important for professionals, scientists, and citizens to make sense of large amounts of data. Statistics are used across many disciplines from science to business. There are two main types of statistical methods - descriptive statistics which summarize data through measures like the mean and median, and inferential statistics which make inferences about populations based on samples. Descriptive statistics describe data through measures of central tendency and variability, while inferential statistics allow inferences to be made from samples to populations through techniques like hypothesis testing.
The document discusses different types of variables in experimental research:
- Independent variable: Factor manipulated by researcher to determine its effect
- Dependent variable: Factor observed and measured to determine effect of independent variable
- Moderator variable: Factor that modifies relationship between independent and dependent variables
- Control variable: Factors controlled by researcher to neutralize their effects
- Intervening variable: Factor that theoretically affects phenomena but cannot be directly observed
It also discusses data types, central tendency measures, data variability measures, and statistical techniques like correlation analysis, t-tests, ANOVA that are used for quantitative analysis.
When to use, What Statistical Test for data Analysis modified.pptxAsokan R
This document discusses choosing the appropriate statistical test for data analysis. It begins by defining key terminology like independent and dependent variables. It then discusses the different types of variables, including quantitative, categorical, and their subtypes. Hypothesis testing and its key steps are explained. The document outlines assumptions that statistical tests make and categorizes common parametric and non-parametric tests. It provides guidance on choosing a test based on the research question, data structure, variable type, and whether the data meets necessary assumptions. Specific statistical tests are matched to questions about differences between groups, association between variables, and agreement between assessment techniques.
This document provides an introduction and definitions related to key concepts in statistics. It discusses what statistics is as the science of collecting, organizing, and interpreting data. It defines important statistical terms like data, variables, statistics, and parameters. It also outlines the two main branches of statistics as descriptive statistics, which focuses on summarizing and presenting data, and inferential statistics, which analyzes samples to make inferences about populations. Finally, it discusses common sources of data like published sources, experiments, and surveys.
The document discusses various techniques for analyzing different types of data in research. It describes statistical procedures like parametric and non-parametric statistics that have assumptions about the type of data. Qualitative data analysis involves deriving categories from the text or applying existing systems. Descriptive research uses frequencies, central tendencies, and variabilities to analyze data. Correlational research examines relationships between variables using correlations. Multivariate research analyzes multiple dependent and independent variables simultaneously using multiple regression, discriminant analysis, and factor analysis. Experimental research compares groups using t-tests and analyzes more than two groups with one-way ANOVA.
Parametric tests are used to analyze normally distributed numerical data and include Students' t-test and analysis of variance (ANOVA). The Students' t-test compares the means of two groups to determine if they are statistically different. ANOVA compares the means of two or more groups and is more efficient than multiple t-tests. It can be used for one-way, multifactor, or repeated measures designs. Both tests make assumptions about the data distribution and variance between groups.
TOI-421 b: A Hot Sub-Neptune with a Haze-free, Low Mean Molecular Weight Atmo...Sérgio Sacani
Common features of sub-Neptune atmospheres observed to date include signatures of aerosols at moderate equilibrium temperatures (∼500–800 K) and a prevalence of high mean molecular weight atmospheres, perhaps indicating novel classes of planets such as water worlds. Here we present a 0.83–5μm JWST transmission spectrum of the sub-Neptune TOI-421 b. This planet is unique among previously observed counterparts in its high equilibrium temperature (Teq ≈ 920 K) and its Sun-like host star. We find marked differences between the atmosphere of TOI-421 b and those of sub-Neptunes previously characterized with JWST, which all orbit late K and M stars. Specifically, water features in the NIRISS/SOSS bandpass indicate a low mean molecular weight atmosphere consistent with solar metallicity and no appreciable aerosol coverage. Hints of SO2 and CO (but not CO2 or CH4) also exist in our NIRSpec/G395M observations, but not at sufficient signal-to-noise ratio to draw f irm conclusions. Our results support a picture in which sub-Neptunes hotter than ∼850K do not form hydrocarbon hazes owing to a lack of methane to photolyze. TOI-421 b additionally fits the paradigm of the radius valley for planets orbiting FGK stars being sculpted by mass-loss processes, which would leave behind primordial atmospheres overlying rock/iron interiors. Further observations of TOI-421 b and similar hot sub-Neptunes will confirm whether haze-free atmospheres and low mean molecular weights are universal characteristics of such objects.
Astrobiological implications of the stability andreactivity of peptide nuclei...Sérgio Sacani
Recent renewed interest regarding the possibility of life in the Venusian clouds has led to new studies on organicchemistry in concentrated sulfuric acid. However, life requires complex genetic polymers for biological function.Therefore, finding suitable candidates for genetic polymers stable in concentrated sulfuric acid is a necessary firststep to establish that biologically functional macromolecules can exist in this environment. We explore peptidenucleic acid (PNA) as a candidate for a genetic-like polymer in a hypothetical sulfuric acid biochemistry. PNA hex-amers undergo between 0.4 and 28.6% degradation in 98% (w/w) sulfuric acid at ~25°C, over the span of 14 days,depending on the sequence, but undergo complete solvolysis above 80°C. Our work is the first key step towardthe identification of a genetic-like polymer that is stable in this unique solvent and further establishes that con-centrated sulfuric acid can sustain a diverse range of organic chemistry that might be the basis of a form of lifedifferent from Earth’s
Ad
More Related Content
Similar to Basic-Statistics-in-Research-Design.pptx (20)
This document provides an outline for a course on probability and statistics. It begins with an introduction to statistics, including definitions and general uses. It then covers various topics that will be taught, such as measures of central tendency, probability, discrete and continuous distributions, and hypothesis testing. References for textbooks are also provided. The document contains sample assignments and examples to illustrate concepts like scales of measurement, data collection methods, and graphical representations of data. It provides instructions for calculating measures of central tendency and examples of frequency distributions and their related graphs.
This document provides an outline for a course on probability and statistics. It begins with an introduction to statistics, including definitions and general uses. It then covers topics like measures of central tendency, probability, discrete and continuous distributions, and hypothesis testing. References for textbooks on the subject are also provided. Assignments include calculating measures of central tendency and constructing frequency distributions from raw data. Various scales of measurement and methods of data collection are defined. Graphical representations like histograms, pie charts, and bar graphs are discussed. Formulas are given for calculating the mean, median, and mode of both grouped and ungrouped data.
This document provides an outline for a course on probability and statistics. It begins with an introduction to statistics, including definitions and general uses. It then covers topics like measures of central tendency, probability, discrete and continuous distributions, and hypothesis testing. References for textbooks on statistics and counterexamples in probability are also provided. Assignments ask students to list contributors to statistics, apply statistics in real life, define independent and dependent variables, and understand scales of measurement. Methods of data collection, tabular and graphical representation of data, and measures of central tendency and location are also discussed.
This document provides an outline for a course on probability and statistics. It begins with an introduction to key concepts like measures of central tendency, dispersion, correlation, and probability distributions. It then lists common probability distributions and the textbook and references used. Later sections define important statistical terms like population, sample, variable types, data collection methods, and ways of presenting data through tables and graphs. It provides examples of how statistics is used and ends with examples of different variable scales.
This document provides an outline for a course on probability and statistics. It begins with an introduction to key concepts like measures of central tendency, dispersion, correlation, and probability distributions. It then lists common probability distributions and the textbook and references used. Later sections define important statistical terms like population, sample, variable types, data collection methods, and ways of presenting data through tables and graphs. It provides examples of each variable scale and ends with assignments for students.
Basics of Educational Statistics (Inferential statistics)HennaAnsari
This document provides information about inferential statistics presented by Dr. Hina Jalal. It defines inferential statistics as using data from a sample to make inferences about the larger population from which the sample was taken. It discusses key areas of inferential statistics like estimating population parameters and testing hypotheses. It also explains the importance of inferential statistics in research for making conclusions from samples, comparing models, and enabling inferences about populations based on sample data. Flow charts are presented for selecting common statistical tests for comparisons, correlations, and regression.
This document provides an overview of key concepts in statistics. It discusses that statistics involves collecting, organizing, analyzing and interpreting data. It also defines important statistical terms like population, sample, parameter, statistic, qualitative and quantitative data, independent and dependent variables, discrete and continuous variables, and different levels of measurement for variables. The different levels of measurement are nominal, ordinal, interval and ratio. Descriptive statistics are used to summarize and describe data, while inferential statistics allow making inferences about populations from samples.
This document provides an overview of basic concepts in inferential statistics. It defines descriptive statistics as describing and summarizing data through measures like mean, median, variance and standard deviation. Inferential statistics is defined as using sample data and statistics to draw conclusions about populations through hypothesis testing and estimates. Key concepts explained include parameters, statistics, sampling distributions, null and alternative hypotheses, and the hypothesis testing process. Examples of descriptive and inferential analyses are also provided.
kelan nyo isubmit yung assignment no. 7 and 8 nyo nasa slides yun ng stats. isubmit nyo sa akin sa lunes during electromagnetism kasi kukulangin yung class participation nyo sa stats.
1. The document discusses quantitative research methods, including comparing groups, examining relationships between variables, different types of data and levels of measurement, sampling techniques, and common statistical tools.
2. Key statistical tools covered include t-tests, ANOVA, correlation analysis, chi-square tests, and non-parametric equivalents for comparing groups and examining relationships.
3. The purpose of quantitative research is to systematically investigate phenomena through collecting and analyzing numerical data.
Statistics is the collection, organization, analysis, and presentation of data. It has become important for professionals, scientists, and citizens to make sense of large amounts of data. Statistics are used across many disciplines from science to business. There are two main types of statistical methods - descriptive statistics which summarize data through measures like the mean and median, and inferential statistics which make inferences about populations based on samples. Descriptive statistics describe data through measures of central tendency and variability, while inferential statistics allow inferences to be made from samples to populations through techniques like hypothesis testing.
The document discusses different types of variables in experimental research:
- Independent variable: Factor manipulated by researcher to determine its effect
- Dependent variable: Factor observed and measured to determine effect of independent variable
- Moderator variable: Factor that modifies relationship between independent and dependent variables
- Control variable: Factors controlled by researcher to neutralize their effects
- Intervening variable: Factor that theoretically affects phenomena but cannot be directly observed
It also discusses data types, central tendency measures, data variability measures, and statistical techniques like correlation analysis, t-tests, ANOVA that are used for quantitative analysis.
When to use, What Statistical Test for data Analysis modified.pptxAsokan R
This document discusses choosing the appropriate statistical test for data analysis. It begins by defining key terminology like independent and dependent variables. It then discusses the different types of variables, including quantitative, categorical, and their subtypes. Hypothesis testing and its key steps are explained. The document outlines assumptions that statistical tests make and categorizes common parametric and non-parametric tests. It provides guidance on choosing a test based on the research question, data structure, variable type, and whether the data meets necessary assumptions. Specific statistical tests are matched to questions about differences between groups, association between variables, and agreement between assessment techniques.
This document provides an introduction and definitions related to key concepts in statistics. It discusses what statistics is as the science of collecting, organizing, and interpreting data. It defines important statistical terms like data, variables, statistics, and parameters. It also outlines the two main branches of statistics as descriptive statistics, which focuses on summarizing and presenting data, and inferential statistics, which analyzes samples to make inferences about populations. Finally, it discusses common sources of data like published sources, experiments, and surveys.
The document discusses various techniques for analyzing different types of data in research. It describes statistical procedures like parametric and non-parametric statistics that have assumptions about the type of data. Qualitative data analysis involves deriving categories from the text or applying existing systems. Descriptive research uses frequencies, central tendencies, and variabilities to analyze data. Correlational research examines relationships between variables using correlations. Multivariate research analyzes multiple dependent and independent variables simultaneously using multiple regression, discriminant analysis, and factor analysis. Experimental research compares groups using t-tests and analyzes more than two groups with one-way ANOVA.
Parametric tests are used to analyze normally distributed numerical data and include Students' t-test and analysis of variance (ANOVA). The Students' t-test compares the means of two groups to determine if they are statistically different. ANOVA compares the means of two or more groups and is more efficient than multiple t-tests. It can be used for one-way, multifactor, or repeated measures designs. Both tests make assumptions about the data distribution and variance between groups.
TOI-421 b: A Hot Sub-Neptune with a Haze-free, Low Mean Molecular Weight Atmo...Sérgio Sacani
Common features of sub-Neptune atmospheres observed to date include signatures of aerosols at moderate equilibrium temperatures (∼500–800 K) and a prevalence of high mean molecular weight atmospheres, perhaps indicating novel classes of planets such as water worlds. Here we present a 0.83–5μm JWST transmission spectrum of the sub-Neptune TOI-421 b. This planet is unique among previously observed counterparts in its high equilibrium temperature (Teq ≈ 920 K) and its Sun-like host star. We find marked differences between the atmosphere of TOI-421 b and those of sub-Neptunes previously characterized with JWST, which all orbit late K and M stars. Specifically, water features in the NIRISS/SOSS bandpass indicate a low mean molecular weight atmosphere consistent with solar metallicity and no appreciable aerosol coverage. Hints of SO2 and CO (but not CO2 or CH4) also exist in our NIRSpec/G395M observations, but not at sufficient signal-to-noise ratio to draw f irm conclusions. Our results support a picture in which sub-Neptunes hotter than ∼850K do not form hydrocarbon hazes owing to a lack of methane to photolyze. TOI-421 b additionally fits the paradigm of the radius valley for planets orbiting FGK stars being sculpted by mass-loss processes, which would leave behind primordial atmospheres overlying rock/iron interiors. Further observations of TOI-421 b and similar hot sub-Neptunes will confirm whether haze-free atmospheres and low mean molecular weights are universal characteristics of such objects.
Astrobiological implications of the stability andreactivity of peptide nuclei...Sérgio Sacani
Recent renewed interest regarding the possibility of life in the Venusian clouds has led to new studies on organicchemistry in concentrated sulfuric acid. However, life requires complex genetic polymers for biological function.Therefore, finding suitable candidates for genetic polymers stable in concentrated sulfuric acid is a necessary firststep to establish that biologically functional macromolecules can exist in this environment. We explore peptidenucleic acid (PNA) as a candidate for a genetic-like polymer in a hypothetical sulfuric acid biochemistry. PNA hex-amers undergo between 0.4 and 28.6% degradation in 98% (w/w) sulfuric acid at ~25°C, over the span of 14 days,depending on the sequence, but undergo complete solvolysis above 80°C. Our work is the first key step towardthe identification of a genetic-like polymer that is stable in this unique solvent and further establishes that con-centrated sulfuric acid can sustain a diverse range of organic chemistry that might be the basis of a form of lifedifferent from Earth’s
Investigating the central role that theories of the visual arts and creativity played in the development of fascism in France, Mark Antliff examines the aesthetic dimension of fascist myth-making within the history of the avant-garde. Between 1909 and 1939, a surprising array of modernists were implicated in this project, including such well-known figures as the symbolist painter Maurice Denis, the architects Le Corbusier and Auguste Perret, the sculptors Charles Despiau and Aristide Maillol, the “New Vision” photographer Germaine Krull, and the fauve Maurice Vlaminck.
Examining Visual Attention in Gaze-Driven VR Learning: An Eye-Tracking Study ...Yasasi Abeysinghe
This study presents an eye-tracking user study for analyzing visual attention in a gaze-driven VR learning environment using a consumer-grade Meta Quest Pro VR headset. Eye tracking data were captured through the headset's built-in eye tracker. We then generated basic and advanced eye-tracking measures—such as fixation duration, saccade amplitude, and the ambient/focal attention coefficient K—as indicators of visual attention within the VR setting. The generated gaze data are visualized in an advanced gaze analytics dashboard, enabling us to assess users' gaze behaviors and attention during interactive VR learning tasks. This study contributes by proposing a novel approach for integrating advanced eye-tracking technology into VR learning environments, specifically utilizing consumer-grade head-mounted displays.
Eric Schott- Environment, Animal and Human Health (3).pptxttalbert1
Baltimore’s Inner Harbor is getting cleaner. But is it safe to swim? Dr. Eric Schott and his team at IMET are working to answer that question. Their research looks at how sewage and bacteria get into the water — and how to track it.
The Man Who Dared to Challenge Newton: The True Story of Thane Heins, the Canadian Genius
Who Changed the World
By Johnny Poppi – for international press
In a small town in Ontario, among wheat fields and wind-filled silences, a man has worked for decades in
anonymity, armed only with naive curiosity, motors, copper wires, and questions too big to ignore. His
name is Thane C/ Heins, and according to some scientists who have seen him in action, he may have
made—and indeed has made—the most important scientific discovery in the history of humanity.
A discovery which will eventually eliminate the need for oil, coal, and uranium, and at the very least their
harmful effects while eliminating the need to recharge electric vehicles, and even rewrite—as it has already
begun—the very laws of physics as we’ve known them since Aristotle in 300 BC.
Sound like science fiction? Then listen to this story.
Euclid: The Story So far, a Departmental Colloquium at Maynooth UniversityPeter Coles
The European Space Agency's Euclid satellite was launched on 1st July 2023 and, after instrument calibration and performance verification, the main cosmological survey is now well under way. In this talk I will explain the main science goals of Euclid, give a brief summary of progress so far, showcase some of the science results already obtained, and set out the time line for future developments, including the main data releases and cosmological analysis.
Structure formation with primordial black holes: collisional dynamics, binari...Sérgio Sacani
Primordial black holes (PBHs) could compose the dark matter content of the Universe. We present the first simulations of cosmological structure formation with PBH dark matter that consistently include collisional few-body effects, post-Newtonian orbit corrections, orbital decay due to gravitational wave emission, and black-hole mergers. We carefully construct initial conditions by considering the evolution during radiation domination as well as early-forming binary systems. We identify numerous dynamical effects due to the collisional nature of PBH dark matter, including evolution of the internal structures of PBH halos and the formation of a hot component of PBHs. We also study the properties of the emergent population of PBH binary systems, distinguishing those that form at primordial times from those that form during the nonlinear structure formation process. These results will be crucial to sharpen constraints on the PBH scenario derived from observational constraints on the gravitational wave background. Even under conservative assumptions, the gravitational radiation emitted over the course of the simulation appears to exceed current limits from ground-based experiments, but this depends on the evolution of the gravitational wave spectrum and PBH merger rate toward lower redshifts.
Issues in using AI in academic publishing.pdfAngelo Salatino
In this slide deck is a lecture I held at the Open University for PhD students to educated them about the dark side of science: predatory journals, paper mills, misconduct, retractrions and much more.
2. TYPES OF DATA IN
STATISTICS
1. Continuous Data
2. Discrete Data
3. Nominal Data
4. Interval Data
5. Categorical Data
3. CONTINUOUS DATA
are data which come from an interval of possible outcomes.
Examples of continuous data include:
• the amount of rain, in inches, that falls in a randomly selected
storm
• the weight, in pounds, of a randomly selected student
• the square footage of a randomly selected three-bedroom house
4. DISCRETE DATA
data with a finite or countably infinite number of possible outcomes.
Examples of discrete data include
• the number of siblings a randomly selected person has
• the total on the faces of a pair of six-sided dice
• the number of students you need to ask before you find one who
loves
5. NOMINAL DATA
values are grouped into categories that have no meaningful
order. For example, gender and political affiliation are nominal
level variables. Members in the group are assigned a label in that
group and there is no hierarchy. Typical descriptive statistics
associated with nominal data are frequencies and percentages.
6. INTERVAL DATA
is a type of data which is measured along a scale, in which each
point is placed at an equal distance (interval) from one another.
Interval data is one of the two types of discrete data. An example of
interval data is the data collected on a thermometer—its gradation
or markings are equidistant.
7. CATEGORICAL DATA
Categorical variables represent types of data which may be divided
into groups. Examples of categorical variables are race, sex, age
group, and educational level. While the latter two variables may also
be considered in a numerical manner by using exact values for age
and highest grade completed, it is often more informative to
categorize such variables into a relatively small number of groups.
8. STATISTICS
the science concerned with developing and studying methods for
collecting, analyzing, interpreting and presenting empirical data.
Statistics is a highly interdisciplinary field; research in statistics
finds applicability in virtually all scientific fields and research
questions in the various scientific fields motivate the development
of new statistical methods and theory.
9. STATISTICS AND ITS
TYPES
Statistics is a collection of planning experiments
methods, obtaining data, analyzing, interpreting,
and drawing conclusions based on the data
(Alferes & Duro 2010). It is divided into two main
areas: Descriptive and Inferential.
10. DESCRIPTIVE
STATISTICS
• summarizes or describes the essential characteristics of a known
set of data.
• are brief descriptive coefficients that summarize a given data
set, which can be either a representation of the entire or a
sample of a population.
• For example, the Department of Health conducts a tally to
determine the number of CoViD-19 cases per day in the
Philippines.
11. INFERENTIAL
STATISTICS
• uses sample data to make inferences about a population. It
consists of generalizing from samples to populations, performing
hypothesis testing, determining relationships among variables,
and making predictions.
• For example, assuming you want to find out if the Filipinos want
to take a shot on the CoViD-19 vaccine. In such a case, a smaller
sample of the population is considered. The results are drawn,
and the analysis is extended to the larger data set.
12. TOOLS IN DESCRIPTIVE
STATISTICS
Frequency Distribution is a collection of
observations produced by sorting them into
classes and showing their frequency or numbers
of occurrences in each class. For example,
twenty-five students were given a blood test to
determine their blood types.
13. From the given data, here is
how to organize them using
frequency distribution.
Data sets of Blood types of
Twenty-five students.
14. MEASURES OF CENTRAL
TENDENCY OR POSITION OR
AVERAGE
When scores and other measures have been tabulated into a
frequency distribution, the next task is to calculate a measure of
central tendency or central position.
This measure of central tendency is synonymous with the word
“average”. An average is a typical value that tends to describe the
set of data.
15. MEAN
Mean, or simply the average is the most frequently used and
can be described as the arithmetic average of all scores or
groups of scores in a distribution. The process can be done by
adding all the scores or data then divided by the total number
of cases.
16. MEDIAN
Median, or the middle-most value in a list of items arranged in
increasing or decreasing order. If the case is in an odd number or
items, there will be exactly one item in the middle. In case the
number or items is an even number, the midpoint will be
determined by getting the average of the two-middle item.
17. MODE
mode is the score or group of scores that occur
most frequently. Some distributions don’t have
mode at all. Others may have more than one
mode. In cases that the distribution has two
modes, the term used is bimodal.
18. Laboratory tests reveal the
incubation period (measures
in days) of virus among the
30 infected residents of
Brgy. Malinis
In dealing with this, arrange
the given data from highest
to lowest or vice versa
20. MEASURES OF VARIATION/
DISPERSION
The previous section focused on average or
measures of central tendency. The averages
are supposed to be the central scores of a
given set of data, However, not all features of
a given data set may be reflected by the
averages. Suppose, two different groups of 5
Students are given 20-item identical quizzes
in Science. The following data below were the
results.
21. MEASURES OF VARIATION/
DISPERSION
The average of each group
are as follows.
As shown in the second table, the two
sets of averages have no difference. But
both groups show an obvious difference.
Group 2 has more widely scattered data
compared to Group 1. This characteristic
called variability or dispersion is not
reflected by averages. The three basic
measures of dispersion are range,
variance, and standard deviation.
22. RANGE
is the simplest measure of dispersion to calculate. It is
done by getting the difference between the
highest/largest value and lowest/smallest value in
each set of data. A larger range suggests greater
variations or dispersion. On the other hand, a smaller
range suggests lesser variations or dispersion
23. VARIANCE
measures how far a data set is spread out. It is
mathematically defined as the average of the squared
differences from the mean.
24. STANDARD DEVIATION
is the most commonly used measure of dispersion. It indicates
how closely the values of the given data set are clustered
around the mean. It is computed by getting the positive square
root of variance. The lower value of standard deviation means
that the values of the given set of data are spread over a
smaller range around the mean. On the other hand, greater
value means that the values of the given set of data are spread
over a larger range around the mean.
25. USED IN HYPOTHESIS
TESTING
To determine whether a predictor variable has a statistically
significant relationship with an outcome variable and estimate
the difference between two or more groups.
To determine what type of statistical tool is appropriate.
To choose the test that fits the types of predictor or independent
variables and outcome/dependent variables you have collected.
26. TOOLS IN INFERENTIAL
STATISTICS
Statistical tests are used to derive a generalization about the
population from the sample. A statistical test is a formal technique
that relies on the probability distribution for concluding the
reasonableness of the hypothesis. These hypothetical testing related
to differences are classified as parametric and non-parametric tests.
The parametric test is one that has information about the
population parameter. On the other hand, the non-parametric test is
where the researcher has no idea regarding the population
parameter.
27. PARAMETRIC TESTS
usually have stricter requirements than non-parametric tests and
can make more robust inferences from the data. They can only be
conducted with data that adheres to the standard assumptions of
statistical tests.
The most common types of the parametric test include regression
tests, comparison tests, and correlation tests.
29. EXAMPLE
The Effect of the Amount of Chlorine in the Color of Algae. Identify first your
independent and dependent variables, how many are they, and their type,
whether qualitative/ categorical or quantitative/numeric. After identifying
such, look at the diagram above to know the parametric test's right
statistical tool. In the given problem, the amount of chlorine is the
independent variable, it’s numeric or qualitative, and 2 or more amounts of
chlorine may be used in the experiment. The dependent variable is the color
of algae; its categorical and color may vary. So, looking at the above
diagram, logistic regression is the appropriate tool.
30. NON-PARAMETRIC
TEST
They don’t make as many assumptions about the
data and are useful when one or more common
statistical assumptions are violated. However, the
inferences they make aren’t as strong as with
parametric tests.
32. Statistical tools are complex,
especially among beginners.
However, according to Grobman,
2017, the most commonly used
in science investigatory projects
are chi-square, t-tests, and
correlations. In determining
whether there is no statistically
significant relationship between
the independent and dependent
variables, we always consider
the standard rule of thumb. If
the p-value is lower than 0.05,
we reject the null hypothesis
and accept the alternative
hypothesis.
33. Licensed Statisticians
play a vital role in
computing and
interpreting the results
of the data gathered. In
any investigation, it is
important to consult
them to ensure that
your results are
statistically correct.
SPSS and Strata are
some of the most
common software they
are using.