Accendo Reliability

Your Reliability Engineering Professional Development Site

  • Home
  • About
    • Contributors
    • About Us
    • Colophon
    • Survey
  • Reliability.fm
  • Articles
    • CRE Preparation Notes
    • NoMTBF
    • on Leadership & Career
      • Advanced Engineering Culture
      • ASQR&R
      • Engineering Leadership
      • Managing in the 2000s
      • Product Development and Process Improvement
    • on Maintenance Reliability
      • Aasan Asset Management
      • AI & Predictive Maintenance
      • Asset Management in the Mining Industry
      • CMMS and Maintenance Management
      • CMMS and Reliability
      • Conscious Asset
      • EAM & CMMS
      • Everyday RCM
      • History of Maintenance Management
      • Life Cycle Asset Management
      • Maintenance and Reliability
      • Maintenance Management
      • Plant Maintenance
      • Process Plant Reliability Engineering
      • RCM Blitz®
      • ReliabilityXperience
      • Rob’s Reliability Project
      • The Intelligent Transformer Blog
      • The People Side of Maintenance
      • The Reliability Mindset
    • on Product Reliability
      • Accelerated Reliability
      • Achieving the Benefits of Reliability
      • Apex Ridge
      • Field Reliability Data Analysis
      • Metals Engineering and Product Reliability
      • Musings on Reliability and Maintenance Topics
      • Product Validation
      • Reliability by Design
      • Reliability Competence
      • Reliability Engineering Insights
      • Reliability in Emerging Technology
      • Reliability Knowledge
    • on Risk & Safety
      • CERM® Risk Insights
      • Equipment Risk and Reliability in Downhole Applications
      • Operational Risk Process Safety
    • on Systems Thinking
      • Communicating with FINESSE
      • The RCA
    • on Tools & Techniques
      • Big Data & Analytics
      • Experimental Design for NPD
      • Innovative Thinking in Reliability and Durability
      • Inside and Beyond HALT
      • Inside FMEA
      • Institute of Quality & Reliability
      • Integral Concepts
      • Learning from Failures
      • Progress in Field Reliability?
      • R for Engineering
      • Reliability Engineering Using Python
      • Reliability Reflections
      • Statistical Methods for Failure-Time Data
      • Testing 1 2 3
      • The Manufacturing Academy
  • eBooks
  • Resources
    • Accendo Authors
    • FMEA Resources
    • Glossary
    • Feed Forward Publications
    • Openings
    • Books
    • Webinar Sources
    • Podcasts
  • Courses
    • Your Courses
    • Live Courses
      • Introduction to Reliability Engineering & Accelerated Testings Course Landing Page
      • Advanced Accelerated Testing Course Landing Page
    • Integral Concepts Courses
      • Reliability Analysis Methods Course Landing Page
      • Applied Reliability Analysis Course Landing Page
      • Statistics, Hypothesis Testing, & Regression Modeling Course Landing Page
      • Measurement System Assessment Course Landing Page
      • SPC & Process Capability Course Landing Page
      • Design of Experiments Course Landing Page
    • The Manufacturing Academy Courses
      • An Introduction to Reliability Engineering
      • Reliability Engineering Statistics
      • An Introduction to Quality Engineering
      • Quality Engineering Statistics
      • FMEA in Practice
      • Process Capability Analysis course
      • Root Cause Analysis and the 8D Corrective Action Process course
      • Return on Investment online course
    • Industrial Metallurgist Courses
    • FMEA courses Powered by The Luminous Group
    • Foundations of RCM online course
    • Reliability Engineering for Heavy Industry
    • How to be an Online Student
    • Quondam Courses
  • Calendar
    • Call for Papers Listing
    • Upcoming Webinars
    • Webinar Calendar
  • Login
    • Member Home
  • Barringer Process Reliability Introduction Course Landing Page
  • Upcoming Live Events
You are here: Home / Articles / How do I Test my Data for Normality?

by Steven Wachs Leave a Comment

How do I Test my Data for Normality?

How do I Test my Data for Normality?

Many statistical tests and procedures assume that data follows a normal (bell-shaped) distribution.

For example, all of the following statistical tests, statistics, or methods assume that data is normally distributed:

  • Hypothesis tests such as t-tests, Chi-Square tests, F tests
  • Analysis of Variance (ANOVA)
  • Least Squares Regression
  • Control Charts of Individuals with 3-sigma limits
  • Common formulas for process capability indices such as Cp and Cpk

Before applying statistical methods that assume normality, it is necessary to perform a normality test on the data (with some of the above methods we check residuals for normality).  We hypothesize that our data follows a normal distribution, and only reject this hypothesis if we have strong evidence to the contrary.

While it may be tempting to judge the normality of the data by simply creating a histogram of the data, this is not an objective method to test for normality – especially with sample sizes that are not very large.  With small sample sizes, discerning the shape of the histogram is difficult.  Furthermore, the shape of the histogram can change significantly by simply changing the interval width of the histogram bars.

Normal probability plotting may be used to objectively assess whether data comes from a normal distribution, even with small sample sizes.  On a normal probability plot, data that follows a normal distribution will appear linear (a straight line).  For example, a random sample of 30 data points from a normal distribution results in the first normal probability plot (below left).  Here, the data points fall close to the straight line.  The second normal probability plot (below right) illustrates data that does not come from a normal distribution.

The “P-Value” on the plot legend is based on the computed Anderson-Darling statistic and provides an objective assessment of the hypothesis that the data is normally distributed (comes from a normal distribution).  How the p-value is formally defined is an advanced subject and is not discussed in this article.  Here, we focus on the interpretation.

If the p-value is “small” (usually less than 0.05), then we have strong evidence that the data is not normal (does not come from a normal distribution).  If the p-value is “large” (usually more than 0.10), then we assume that the data is normal (comes from a normal distribution).  P-values larger than 0.10 tell us that there is not sufficient evidence to reject the normality assumption (hypothesis).  In this case, we are justified in using statistical methods that assume normality.

Many methods are available to handle non-normal data and these should be utilized when necessary.  Applying methods that assume the normal distribution when this assumption is not valid often results in incorrect conclusions.

[display_form id=419]

Filed Under: Articles, Integral Concepts, on Tools & Techniques

About Steven Wachs

Steven Wachs has 25 years of wide-ranging industry experience in both technical and management positions. Steve has worked as a statistician at Ford Motor Company where he has extensive experience in the development of statistical models, reliability analysis, designed experimentation, and statistical process control.

« Grade Your RCA Effort and Print Out Your Private Report Card…
What is an Asset Management System? »

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Articles by Steven Wachs, Integral Concepts
in the Integral Concepts article series

Join Accendo

Receive information and updates about articles and many other resources offered by Accendo Reliability by becoming a member.

It’s free and only takes a minute.

Join Today

Recent Articles

  • Gremlins today
  • The Power of Vision in Leadership and Organizational Success
  • 3 Types of MTBF Stories
  • ALT: An in Depth Description
  • Project Email Economics

© 2025 FMS Reliability · Privacy Policy · Terms of Service · Cookies Policy