Body of Knowledge - 2008
The last administration of this current Reliability Engineer Body of Knowledge will be June 8, 2009.
Reliability Engineer Body of Knowledge (PDF, 32 KB)
The topics in this Body of Knowledge include additional detail in the form of subtext explanations and the cognitive level at which the questions will be written. This information will provide useful guidance for both the Exam Development Committee and the candidate preparing to take the exam. The subtext is not intended to limit the subject matter or be all-inclusive of what might be covered in an exam. It is meant to clarify the type of content to be included in the exam. The descriptor in parentheses at the end of each entry refers to the maximum cognitive level at which the topic will be tested. A more complete description of cognitive levels is provided at the end of this document.
- RELIABILITY MANAGEMENT (19 Questions)
- Strategic management
- Benefits of reliability engineering
Demonstrate how reliability engineering techniques and methods improve programs, processes, products, and services. (Synthesis) - Interrelationship of quality and reliability
Define and describe quality and reliability and how they relate to each other. (Comprehension) - Role of the reliability function in the organization
Demonstrate how reliability professionals can apply their techniques and interact effectively with marketing, safety and product liability, engineering, manufacturing, logistics, etc. (Analysis) - Reliability in product and process development
Integrate reliability engineering techniques with other development activities (e.g., concurrent engineering). (Synthesis) - Failure consequence and liability management
Use liability and consequence limitation objectives to determine reliability acceptance criteria, and identify development and test methods that verify and validate these criteria. (Application) - Life-cycle cost planning
Determine the impact of failures in terms of service and cost (both tangible and intangible) throughout a product's life-cycle. (Analysis) - Customer needs assessment
Describe how various feedback mechanisms (e.g., QFD, prototyping, beta testing) help determine customer needs and specify product and service requirements. (Comprehension) - Project management
Interpret basic project management tools and techniques, such as Gantt chart, PERT chart, critical path, resource planning, etc. (Comprehension)
- Benefits of reliability engineering
- Reliability program management
- Terminology
Identify and define basic reliability terms such as MTTF, MTBF, MTTR, availability, failure rate, dependability, maintainability, etc. (Analysis) - Elements of a reliability program
Use customer requirements and other inputs to develop a reliability program including elements such as design for reliability, progress assessment, FRACAS, monitoring and tracking components, customer satisfaction and other feedback, etc. (Evaluation) - Product life-cycle and costs
Identify the various life-cycle stages and their relationship to reliability, and analyze various cost-related issues including product maintenance, life expectation, duty cycle, software defect phase containment, etc. (Analysis) - Design evaluation
Plan and implement product and process design evaluations to assess reliability at various life-cycle stages using validation, verification, or other review techniques. (Evaluation) - Requirements management
Describe how requirements management methods are used to help prioritize design and development activities. (Comprehension) - Reliability training programs
Demonstrate the need for training, develop a training plan, and evaluate training effectiveness. (Application)
- Terminology
- Product safety and liability
- Roles and responsibilities
Define and describe the roles and responsibilities of a reliability engineer in terms of safety and product liability. (Application) - Ethical issues
Identify appropriate ethical behaviors for a reliability engineer in various situations. (Evaluation) - System safety program
Identify safety-related issues by analyzing customer feedback, design data, field data, and other information sources. Use risk assessment tools such as hazard analysis, FMEA, FMECA, PRAT, FTA, etc., to identify and prioritize safety concerns, and identify steps to idiot-proofing products and processes to minimize risk exposure. (Analysis)
- Roles and responsibilities
- Strategic management
- PROBABILITY AND STATISTICS FOR RELIABILITY (25 Questions)
- Basic concepts
- Statistical terms
Define and use basic terms such as population, parameter, statistic, random sample, the central limit theorem, etc., and compute expected values. (Application) - Basic probability concepts
Define and use basic probability concepts such as independence, mutually exclusive, complementary and conditional probability, joint occurrence of events, etc., and compute expected values. (Application) - Discrete and continuous probability distributions
Describe, apply, and distinguish between various distributions (binomial, Poisson, exponential, Weibull, normal, log-normal, etc.) and their functions (cumulative distribution functions (CDFs), probability density functions (PDFs), hazard functions, etc.). Apply these distributions and functions to related concepts such as the bathtub curve.(Evaluation) - Statistical process control (SPC)
Define various SPC terms and describe how SPC is related to reliability. (Comprehension)
- Statistical terms
- Statistical inference
- Point and interval estimates of parameters
Define and interpret these estimates. Obtain them using probability plots, maximum likelihood methods, etc. Analyze the efficiency and bias of the estimators. (Evaluation) - Statistical interval estimates
Compute confidence intervals, tolerance intervals, etc., and draw conclusions from the results. (Analysis) - Hypothesis testing (parametric and non-parametric)
Apply hypothesis testing for parameters such as means, variance, and proportions. Apply and interpret significance levels and Type I and Type II errors for accepting/rejecting the null hypothesis. (Analysis) - Bayesian technique
Describe the advantages and limitations of this technique. Define elements including prior, likelihood, and posterior probability distributions, and compute values using the Bayes formula. (Application)
- Point and interval estimates of parameters
- Basic concepts
- RELIABILITY IN DESIGN AND DEVELOPMENT (25 Questions)
- Reliability design techniques
- Use factors
Identify and characterize various use factors (e.g., temperature, humidity, vibration, corrosives, pollutants) and stresses (e.g., severity of service, electrostatic discharge (ESD), radio frequency interference (RFI), throughput) to which a product may be subjected. (Synthesis) - Stress-strength analysis
Apply this technique and interpret the results. (Evaluation). - Failure mode and effects analysis (FMEA) in design
Apply the techniques and concepts and evaluate the results of FMEA during the design phase. (Evaluation) [NOTE: Identifying and using this tool for other aspects of reliability are covered in VII.C.1.] - Failure mode effects and criticality analysis (FMECA) in
design
Apply the techniques and concepts and evaluate the results of FMECA during the design phase. (Evaluation) [NOTE: Identifying and using this tool for other aspects of reliability are covered in VII.C.2.] - Fault tree analysis (FTA) in design
Apply this technique at the design stage to eliminate or minimize undesired events. (Analysis)
[NOTE: Identifying and using the symbols and rules of FTA are covered in VII.C.3.] - Tolerance and worst-case analyses
Use various analysis techniques (e.g., root-sum squared, extreme value, statistical tolerancing) to characterize variation that affects reliability. (Evaluation) - Robust-design approaches
Define terms such as independent and dependent variables, factors, levels, responses, treatment, error, replication, etc. Plan and conduct design of experiments (full-factorial, fractional factorial, etc.) or other methods. Analyze the results and use them to achieve robustness. (Evaluation) - Human factors reliability
Describe how human factors influence the use and performance of products and processes. (Comprehension) - Design for X (DFX)
Apply tools and techniques to enhance a product's producibility and serviceability, including design for assembly, service, manufacturability, testability, etc. (Evaluation)
- Use factors
- Parts and systems management
- Parts selection
Apply techniques such as parts standardization, parts reduction, parallel model, software reuse, etc., to improve reliability in products, systems, and processes. (Application) - Material selection and control
Apply probabilistic methods for proper selection of materials. (Application) - Derating methods and principles
Use methods such as S-N diagram, stress-life relationship, etc., to determine the relationship between applied stress and rated value. (Application) - Establishing specifications
Identify various terms related to reliability, maintainability, and serviceability (e.g., MTBF, MTTF, MTBR, MTBUMA, service interval) as they relate to product specifications.
- Parts selection
- Reliability design techniques
- RELIABILITY MODELING AND PREDICTIONS (23 Questions)
- Reliability modeling
- Sources of reliability data
Identify and describe various types of data (e.g., public, common, On-Site data) and their advantages and limitations, and use data from various sources (prototype, development, test, field, etc.) to measure and enhance product reliability. (Analysis) - Reliability block diagrams and models
Describe, select, and use various types of block diagrams and models (e.g., series, parallel, partial redundancy, time-dependent modeling) and analyze them for reliability. (Evaluation) - Simulation techniques
Identify, select, and apply various simulation methods (e.g., Monte Carlo, Markov) and describe their advantages and limitations. (Analysis)
- Sources of reliability data
- Reliability predictions
- Part count predictions and part stress analysis
Use parts failure rate data to estimate system- and subsystem-level reliability. (Analysis) - Advantages and limitations of reliability predictions
Demonstrate the advantages and limitations of reliability predictions, how they can be used to maintain or improve reliability, and how they relate to and can be used with field reliability data. (Application) - Reliability prediction methods for repairable and non-repairable
devices
Identify and use appropriate prediction methods for these types of devices and systems. (Application) - Reliability apportionment/allocation
Describe the purpose of reliability apportionment/allocation and its relationship to subsystem requirements, and identify when to use equal apportionment or other techniques. (Analysis)
- Part count predictions and part stress analysis
- Reliability modeling
- RELIABILITY TESTING (23 Questions)
- Reliability test planning
- Elements of a reliability test plan
Determine the appropriate elements and reliability test strategies for various development phases. (Analysis) - Types and applications of reliability testing
Identify and evaluate the appropriateness and limitations of various reliability test strategies within available resource constraints. (Evaluation) - Test environment considerations
Evaluate the application environment (including combinations of stresses) to determine the appropriate reliability test environment. (Evaluation)
- Elements of a reliability test plan
- Development testing
Assess the purpose, advantages, and limitations of each of the following types of tests, and use common models to develop test plans, evaluate risks, and interpret test results. (Evaluation)- Accelerated life tests (e.g., single-stress, multiple-stress, sequential stress)
- Step-stress testing (e.g., HALT)
- Reliability growth testing (e.g., Duane, AMSAA, TAAF)
- Software testing (e.g., white-box, fault-injection)
- Product testing
Assess the purpose, advantages, and limitations of each of the following types of tests, and use common models to develop test plans, evaluate risks, and interpret test results. (Evaluation)- Qualification/demonstration testing (e.g., sequential tests, fixed-length tests)
- Product reliability acceptance testing (PRAT)
- Stress screening (e.g., ESS, HASS, burn-in tests)
- Attribute testing (e.g., binomial, hypergeometric)
- Degradation testing (e.g., Arrhenius)
- Software testing (e.g., black-box, operational profile)
- Reliability test planning
- MAINTAINABILITY AND AVAILABILITY (17 Questions)
- Management strategies
- Maintainability and availability planning
Develop maintainability and availability plans that support reliability goals and objectives. (Application) - Maintenance strategies
Identify the advantages and limitations of various maintenance strategies (e.g., reliability-centered maintenance (RCM), predictive maintenance, condition-based maintenance), and determine which strategy to use in specific situations. (Analysis). - Maintainability apportionment/allocation
Describe the purpose of maintainability apportionment/allocation and its relationship to system and subsystem requirements, and determine when to modify the maintainability strategy to achieve maintainability goals. (Synthesis) - Availability tradeoffs
Identify various types of availability (e.g., inherent availability, operational availability), and evaluate the reliability/maintainability tradeoffs associated with achieving availability goals. (Evaluation)
- Maintainability and availability planning
- Analyses
- Maintenance time distributions
Determine the applicable distributions (e.g., log-normal, Weibull) for maintenance times. (Analysis) - Preventive maintenance (PM) analysis
Identify the elements of PM analysis (e.g., types of PM tasks, optimum PM intervals, items for which PM is not applicable) and apply them in specific situations. (Analysis) - Corrective maintenance analysis
Identify the elements of corrective maintenance analysis (e.g., fault-isolation time, repair/replace time, skill level, crew hours) and apply them in specific situations. (Analysis) - Testability
Identify testability requirements and use various methods (e.g., built in tests (BITs), no fault found, retest okay, false-alarm rates, software testability) to achieve reliability goals. (Analysis) - Spare parts strategy
Evaluate the relationship between spare parts requirements and maintainability and availability. (Evaluation)
- Maintenance time distributions
- Management strategies
- DATA COLLECTION AND USE (18 Questions)
- Data collection
- Types of data
Identify, define, classify, and compare various data types (e.g., variables vs. attributes, censored vs. uncensored). (Evaluation) - Data sources
Evaluate the appropriateness of various data sources such as field, On-Site, environment, location, test specification, failure modes, failure mechanisms, time at failure, etc. (Evaluation) - Collection methods
Identify elements of data collection methods such as surveys, automated tests, automated monitoring and reporting, etc. (Application) - Data management
Identify the requirements for an organization-wide product-failure database, including which user groups (e.g., production, research, field service, supplier relations, purchasing, business management/accounting) will use the database and how the information interests and needs of those groups can conflict. Identify and distinguish between the level of detail each user group requires, and explain how reporting formats, coding schemes, and other structural components of the database system can influence the usefulness of the data over time and throughout the organization. (Evaluation)
- Types of data
- Data use
- Data summarization
Analyze, evaluate, and summarize data using techniques such as trend analysis, Weibull, graphic representation, etc., based on data types, sources, and required output. (Evaluation) - Preventive and corrective action
Select and use various root cause and data (failure) analysis tools to determine degradation or failure causes, and identify various preventive or corrective actions to take in specific situations. (Evaluation) - Measures of effectiveness
Select and use various data analysis tools to evaluate the effectiveness of preventive and corrective actions. (Synthesis)
- Data summarization
- Data and failure analysis tools
- Failure mode and effects analysis (FMEA)
Identify the components and steps used to develop a FMEA, and use this tool to analyze problems found in various situations. (Evaluation) - Failure mode, effects, and criticality analysis (FMECA)
Distinguish this analysis tool from FMEA, and use it to evaluate the likelihood of certain effects and their criticality (including identifying and applying various levels of severity) in specific situations. (Evaluation) - Fault tree analysis (FTA) and Success tree analysis (STA)
Identify and use the event and logic symbols and rules of these tools to determine the root cause of product failures or the steps necessary to ensure product success. (Evaluation) - Failure reporting, analysis, and corrective action system
(FRACAS)
Identify the elements necessary for a FRACAS to be effective. (Application)
- Failure mode and effects analysis (FMEA)
- Data collection
NOTE: Approximately 20% of the CRE exam will require candidates to perform mathematical functions.
In addition to content specifics, the subtext detail also indicates the intended complexity level of the test questions for that topic. These levels are based on Levels of Cognition (from Blooms Taxonomy, 1956) and are presented below in rank order, from least complex to most complex.
Knowledge Level
(Also commonly referred to as recognition, recall, or rote knowledge.)
Being able to remember or recognize terminology, definitions, facts, ideas,
materials, patterns, sequences, methodologies, principles, etc.
Comprehension Level
Being able to read and understand descriptions, communications, reports,
tables, diagrams, directions, regulations, etc.
Application Level
Being able to apply ideas, procedures, methods, formulas, principles,
theories, etc., in job-related situations.
Analysis
Being able to break down information into its constituent parts and recognize
the parts relationship to one another and how they are organized;
identify sublevel factors or salient data from a complex scenario.
Synthesis
Being able to put parts or elements together in such a way as to show
a pattern or structure not clearly there before; identify which data or
information from a complex set is appropriate to examine further or from
which supported conclusions can be drawn.
Evaluation
Being able to make judgments regarding the value of proposed ideas, solutions,
methodologies, etc., by using appropriate criteria or standards to estimate
accuracy, effectiveness, economic benefits, etc.