Upload
hoangkhanh
View
293
Download
12
Embed Size (px)
Citation preview
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 1 of 69
aboutpipelines.com
Metal Loss Inline
Inspection Tool
Validation Guidance
Document, 1st Edition January2016
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 2 of 69
Aboutpipelines.com
Notice of Copyright
Copyright ©2014 Canadian Energy Pipeline Association (CEPA). All rights reserved.
Canadian Energy Pipeline Association and the CEPA logo are trademarks and/or registered
trademarks of Canadian Energy Pipeline Association. The trademarks or service marks of all
other products or services mentioned in this document are identified respectively.
Disclaimer of Liability
The Canadian Energy Pipeline Association (CEPA) is a voluntary, non-profit industry
association representing major Canadian transmission pipeline companies. The Metal Loss
Inline Inspection Tool Validation Guidance Document (hereafter referred to as the
“Guidelines”) was prepared to provide common guidelines to enhance industry best practice
and performance.
Use of the Guidelines described herein is wholly voluntary. The Guidelines described are not to
be considered industry standards and no representation as such is made. It is the responsibility
of each pipeline company, or other user of these Guidelines, to implement practices to ensure
the safe operation of assets.
While reasonable efforts have been made by CEPA to assure the accuracy and reliability of the
information contained in these Guidelines, CEPA makes no warranty, representation or
guarantee, express or implied, in conjunction with the publication of these Guidelines as to the
accuracy or reliability of these Guidelines. CEPA expressly disclaims any liability or
responsibility, whether in contract, tort or otherwise and whether based on negligence or
otherwise, for loss or damage of any kind, whether direct or consequential, resulting from the
use of these Guidelines. These Guidelines are set out for informational purposes only.
References to trade names or specific commercial products, commodities, services or
equipment constitutes neither an endorsement nor censure by CEPA of any specific product,
commodity, service or equipment.
The CEPA Metal Loss Inline Inspection Tool Validation Guidelines are intended to be
considered as a whole, and users are cautioned to avoid the use of individual chapters without
regard for the entire Guidelines.
Suite 200, 505-3rd St. SW
Calgary, Alberta T2P 3E6
Tel: 403.221.8777
Fax: 403.221.8760
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 3 of 69
Aboutpipelines.com
Table of Contents
Table of Contents .................................................................................................... 3 List of Tables .......................................................................................................... 5 List of Figures ......................................................................................................... 6 1. Introduction .................................................................................................... 7
1.1. Definition of Terms .................................................................................... 7 1.2. Revisions to this Guidance Document ........................................................... 7 1.3. Background and Philosophy ........................................................................ 8 1.4. Harmonization with Other Industry Documents ............................................. 8
2. Scope............................................................................................................. 8 2.1. Practically Assessing ILI Performance ........................................................... 9
3. ILI Acceptance Overview .................................................................................. 9 3.1. Process Overview ...................................................................................... 9 3.2. Components ............................................................................................10
4. Overall Process ...............................................................................................10 4.1. Process Description ...................................................................................10 4.2. Process Flowchart .....................................................................................11
5. Process Verification .........................................................................................14 5.1. Process Overview .....................................................................................14 5.2. (Pre-Run) Tool Selection ............................................................................15 5.3. Inspection System ....................................................................................16 5.4. (Pre-Run) Planning and Preparation ............................................................18 5.5. (Pre-Run) Function Checks ........................................................................20 5.6. (Pre-Run) Mechanical Checks .....................................................................21 5.7. (In the Pipe) Procedure Execution ...............................................................23 5.8. (Post-Run) Mechanical Check .....................................................................24 5.9. (Post-Run) Function Check .........................................................................26 5.10. (Post-Run) Field Data Quality Check .........................................................28 5.11. (Post-Run) Data Analysis Process Check ...................................................29 5.12. (Post-Run) Cumulative Assessment ..........................................................31
6. Validation ......................................................................................................32 6.1. Known Pipeline Features ............................................................................32 6.2. Comparison with Previous ILI .....................................................................34 6.3. Validation from Excavation Data .................................................................36
A1. Scorecard and Guidance Document .....................................................................39 A1.1. Verification Examples .................................................................................50
A2. NACE Table ......................................................................................................53 A3. Matching ..........................................................................................................55
A3.1. Process overview .......................................................................................55 A3.2. Girth Weld Matching ...................................................................................55 A3.3. Matching of identified Anomalies ..................................................................55 A3.4. Calculating Anomaly Depth Change ..............................................................55
A4. Validation using a Previous ILI ............................................................................56 A4.1. Demonstration of Concept ...........................................................................56 A4.2. ILI Error ....................................................................................................57 A4.3. Comparison with Reference Measurements ....................................................58 A4.4. Acceptance Criteria ....................................................................................59
A5. Opportunities for Future Refinement....................................................................65
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 4 of 69
Aboutpipelines.com
A5.1. Standardization of ILI Reporting...................................................................65 A5.2. Documentation of Procedures ......................................................................67 A5.3. Refinement of Scorecard .............................................................................67 A5.4. Technology Specific Verification ...................................................................67
A6. Scoring – Verification Process Scorecard Summary ...............................................69
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 5 of 69
Aboutpipelines.com
List of Tables Table 1: Verification Procedure Scorecard Parameters as per API 1163 .........................14
Table 2: Pre-Run Tool Selection Scoring ....................................................................15
Table 3: Inspection System Data Check Scoring .........................................................17
Table 4: Pre-Run Planning Scoring ............................................................................18
Table 5: Pre-Run Function Check Scoring ..................................................................20
Table 6: Pre-Run Mechanical Check Scoring ...............................................................22
Table 7: Procedure Execution Scoring .......................................................................23
Table 8: Post-Run Mechanical Check Scoring ..............................................................25
Table 9 Post-Run Function Check Scoring ..................................................................27
Table 10: Post-Run Field Data Check Scoring .............................................................28
Table 11: Post-Run Data Analysis Processes Scoring ...................................................30
Table 12: Post-Run Cumulative Assessment Scoring ...................................................31
Table 13: ILI Validation Parameters ..........................................................................35
Table 14: ILI Validation Parameters ..........................................................................37
Table 16: Guidance for Parameter #1 Pre-Run Tool Selection .......................................39
Table 17: Guidance for Parameter #2 Pre-Run Inspection System Data .........................40
Table 18: Guidance for Parameter #3 Pre-Run Planning ..............................................41
Table 19: Guidance for Parameter #4 Pre-Run Function Checks ...................................42
Table 20: Guidance for Parameter #5 Pre-Run Mechanical Checks ................................43
Table 21 Guidance for Parameter #6 in the pipe Procedure Execution ...........................44
Table 22: Guidance for Parameter #7 Post-Run Mechanical Checks ...............................45
Table 23: Guidance for Parameter #8 Post-Run Function Check ...................................46
Table 24: Guidance for Parameter #9 Post-Run Field Data Check .................................47
Table 25: Guidance for Parameter #10 Post-Run Data Analysis Processes and Quality
Checks ..................................................................................................................48
Table 26: Guidance for Parameter #11 Post-Run Cumulative Assessment ......................49
Table 27: Completed Scorecard Example 1 ................................................................50
Table 28: Completed Scorecard Example 2 ................................................................51
Table 29: Considerations when Dealing with Systematic Bias .......................................60
Table 30: Small Population Case Samples ..................................................................63
Table 31: Key Items to Consider for Future Refinement ...............................................65
Table 32: Summary of Data Provided for Scorecard ....................................................66
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 6 of 69
Aboutpipelines.com
Table 33: Verification Process Scorecard Summary .....................................................70
List of Figures Figure 1: Overall Process .........................................................................................11
Figure 2: Verification Check-Point Flowchart ..............................................................12
Figure 3: Validation Procedure .................................................................................13
............................................................................................................................39
Figure 4: Table 1 from NACE SP 102-2010 giving Guidance on Tool Selection for ILI .......54
Figure 5: Illustration of the Validation Process ............................................................56
Figure6: Random Error and Bias Component Contribute to the Error of any ILI
Measurement .........................................................................................................58
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 7 of 69
Aboutpipelines.com
1. Introduction The intent of this Guidance Document is to supplement key industry standards such as
API 1163 and industry best practices by providing a methodology to assist CEPA
members with a cost-effective method for validating the results from inline
inspections. In particular, this document is meant to enable an operator to establish a
process to identify if validation excavations are required and assess the value of those
excavations versus employing alternative verification or validation processes to accept
an ILI run that has no actionable anomalies.
This Guidance Document outlines the procedure for the acceptance of an ILI run based
on a Verification and Validation of the run. This Guidance Document shifts the
emphasis of ILI validation from field validation to verification that planning,
preparation, execution, and analysis of the ILI run were correctly done using well-
vetted, industry recognized procedures. The shift to process verification is expected to
provide greater confidence in the resulting ILI run.
Note: This Guidance Document outlines a suggested approach for verification and
validating of the run. Other approaches may be taken at the discretion of the pipeline
operator if other processes are more practical, provided the sections of this document are
followed. In essence, there may be internal practices already in place for a member
company, which are consistent with this document and other existing industry documents
(i.e. NACE SP0102-2010 and API 1163).
1.1. Definition of Terms Part of the objective of this Guidance Document is to provide clarity and
consistency regarding terminology. As such, the reader is encouraged to review
the following ASME definitions as their usage was adapted for this document.
1.2. Revisions to this Guidance Document This Guidance Document has been developed by CEPA’s Pipeline Integrity
Working Group (PIWG). It will continue to evolve as new advances and
opportunities for improvement are recognized during its use by CEPA member
Verification: The check of the procedures and
operations to ensure that all aspects of
the inspection have been conducted
according to existing standards and
best practices. A successful ILI should
result as a consequence of proper
procedures and operations.
Validation: The check that results of the
inspection (by comparison to field
measurement, previous ILI or other
independent source of information) are
consistent with stated ILI performance
specifications.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 8 of 69
Aboutpipelines.com
companies and from periodic reviews as deemed necessary by CEPA and/or the
PIWG.
1.3. Background and Philosophy The current (2014) regulatory environment in the US and Canada has lead
CEPA to explore the development of a Guidance Document for ILI Tool
Validation. CEPA members seek an alternative to excavations to validate the
results of an ILI run, as in many situations, excavating a limited number of
shallow anomalies is unlikely to yield any real insight into the performance of
the ILI tool.
This Guidance Document provides an alternative to excavations to validate the
results of an ILI run. The procedure in this Guidance Document will assure the
operator and stakeholders of the high quality of the inspection.
There are a number of key benefits of developing a Guidance Document that is
specific to relatively un-corroded lines. Specifically, operators would have
access to a methodology that would allow them to assess and use the results of
in-line inspection more cost effectively. Also, perhaps most importantly, a
consensus-based Guidance Document released by CEPA would provide a
common foundation for discussions with various jurisdictional authorities as
well as in-line inspection tool vendors.
1.4. Harmonization with Other Industry Documents A number of well-vetted industry documents in the area of ILI acceptance
already exist. To the extent possible, this Guidance Document was designed to
remain consistent with, and leverage to the extent possible, any pre-existing
material. The main documents that were referenced in this way are:
NACE Recommended Practice, SP0102-2010 (formerly RP0102)
NACE 35100, Inline Nondestructive Inspection of Pipelines (December
2000)
API Standard, 1163 (Second Edition – April 2013)
2. Scope The purpose of this Guidance Document is to assist the operator in evaluating the
quality of an ILI run and deciding whether the run should be accepted or rejected.
Previously, the acceptance of an ILI run had been based primarily on a Validation
process where the results of the inspection are compared to the results of NDT
measurements in the field. At best, the comparison can only show that the ILI run is
consistent or inconsistent with results collected in the field. The procedure does not
prove that the ILI meets its performance specification, however, the ILI run is
accepted unless evidence to the contrary is found. If the field results are inconsistent
with the ILI performance specification, then the ILI run would be rejected and a rerun
would be required.
This Guidance Document expands the acceptance procedure to include both
Verification and Validation processes.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 9 of 69
Aboutpipelines.com
The Verification process examines all the aspects of tool selection, run preparation,
running of the tool, and analysis of the results to ensure that the procedures followed
should lead to a successful run. Lacking evidence to the contrary, the ILI run is
accepted.
The Validation procedure is similar to the previous acceptance procedure; however,
this Guidance Document relieves the requirement to excavate in some situations. In
these situations, the Validation procedure is believed to more than compensate for not
excavating the pipeline and should improve the confidence in the quality of the ILI
data.
As such, an overall process has been developed for the Verification and Validation of
ILI runs, consistent with existing industry documents (i.e., NACE SP0102-2010 and
API 1163).
2.1. Practically Assessing ILI Performance At best, excavations provide a limited number of comparisons at a few isolated
locations along the pipeline. Furthermore, the cost of obtaining these
comparisons can be prohibitive.
In this Guidance Document it is assumed that the purpose of an ILI run is to
address one or more threats to the integrity of a specific pipeline. Acceptance
of a run means that the operator accepts that the ILI run can be used to
adequately assess the threat(s). If the inspection is rejected, then the threats
to the pipeline (or portions of it) are not adequately addressed by the
inspection. In some cases, rejection of an inspection may require rerunning the
inspection, but in other cases, the threat can be addressed by other means.
In practice, inspection results usually enable the operator to assess risk on most
of a pipeline, but there are often localized areas where the data has been
compromised in some way and the inspection data is less than optimal for the
assessment of risk. At these locations, if the risk due to the threat is great, then
the threat would need to be addressed by some other method.
The procedure developed depends in large part on documentation of the ILI
inspection process. In this way, the acceptance of the ILI data can be conducted
by persons independent of those who were involved in the inspection.
3. ILI Acceptance Overview
3.1. Process Overview This Guidance Document was developed to supplement API 1163 and industry
best practice, with the overall goal to quantify the value of excavations and
have a rigorous approach to ILI acceptance. The definition of an overall ILI
Verification and Validation process was considered critical in the development
of this Guidance Document. As such, the flowchart in Section 4.2 provides a
systematic and consistent process for the Verification and Validation of an ILI
inspection, based on the available information.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 10 of 69
Aboutpipelines.com
3.2. Components The acceptance of an ILI run depends on its Verification and Validation.
Verification consists of three parts:
1. The ILI tool used in the inspection is appropriately selected to assess the
threat(s) and has a history of successful runs.
2. The actual running of the ILI tool and analysis of the data were
conducted according to existing standards and Guidelines.
3. The results of the ILI data are consistent with expected results
considering the age, condition and history of the pipeline.
Validation may be accomplished by three different processes:
1. If there are no actionable anomalies (i.e., anomalies meeting excavation or
repair criteria, or anomalies that require other mitigative action to be
taken), then the inspection is Validated by ensuring that the inspection
successfully identified and reported the known location of any girth welds,
wall thickness changes, tees, and other features on the pipeline that the
tool can be expected to detect and report.
2. If there are actionable anomalies or previous excavation data, then the
inspection is Validated by the comparison of the ILI report to the results of
the excavation. In the case of a metal-loss inspection, the comparison
might consist of the depth and length of the reported anomalies.
3. If there is a previous inspection of the pipeline, the inspection is Validated
by a comparison of the current ILI results to the previous results.
4. Overall Process
4.1. Process Description This Guidance Document defines a process for the acceptance of an ILI run
without need for excavations when there are no actionable anomalies reported
by the inspection. The process has required the definition of a holistic and
comprehensive approach to the acceptance of the ILI data following the
delivery of the report.
The first step towards the acceptance of an ILI run is the Run Verification.
Verification is an 11-point checklist to ensure that run planning and execution
was conducted according to established standards. Depending on the results of
the Verification phase, the process proceeds to one of the Validation processes.
Validation has three components:
1. Comparison of ILI results to known pipeline features,
2. Comparison of ILI metal-loss anomalies to excavation results, and
3. Comparison of ILI metal-loss anomalies to a previous inspection.
Depending on the available data and the results of the ILI, one, two, or all
three components may be required to validate the run.
The following section shows a flowchart of the overall process.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 11 of 69
Aboutpipelines.com
4.2. Process Flowchart Figure 1 shows the top-level flowchart of the overall process. As discussed
above, the process consists of two steps: Verification and Validation. If the run
fails either the Verification or the Validation, then the run cannot be accepted
to address the threat in question.
Figure 1: Overall Process
4.2.1. Verification Process verification consists of an 11-point check in two parts: part
one is a 10-point check regarding various aspects of the tool then in
part two the final check is a Cumulative Assessment. The checks are:
1. Tool Selection,
2. Historical performance of the inspection system,
3. Planning,
4. Pre-run Function Check,
5. Pre-run Mechanical Check,
6. Procedure execution (e.g., pigging procedure, tool speed, etc.),
7. Post-run Mechanical Check,
8. Post-run Function Check,
9. Field Data Quality Check, and
10. Data analysis processes: quality checks.
The final check is the cumulative assessment.
11. Cumulative Assessment
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 12 of 69
Aboutpipelines.com
For each of the 10 points, the operator would follow the flowchart shown in Figure 2.
Guidance on each decision node is provided in A1. Scorecard and Guidance Document.
Figure 2: Verification Check-Point Flowchart
All checks must have a “Pass” or “Conditional” pass for the run to be
accepted. Once the check of the 10 parameters is concluded, they are
reviewed in the Cumulative Check to decide the final verification
result.
4.2.2. Data Validation The second step in the acceptance of an ILI run is data validation.
Whereas verification examines the inspection process to ensure that
all procedures were followed in the acquisition of the ILI data,
validation examines the results of the inspection to ensure that the
data is accurate and meets performance specifications.
The validation procedure depends on the results of the inspection and
the available data. If there is a previous ILI run, then the current run
can be compared to that previous run for validation. If there are
actionable anomalies or previous excavation data, then the
comparison to the excavation results is used for validation. However,
if there are no previous inspections and no actionable anomalies,
then validation is based on the check that the ILI data successfully
identified and reported known pipeline features such as girth welds,
wall-thickness changes, fittings, valves, tee, etc.
Figure 3 shows the validation procedure flowchart. In the procedure,
the first step is to compare non-metal-loss features to the ILI. Then,
if there are actionable anomalies, those anomalies would need to be
excavated and compared to the ILI data. Once the excavation results
are consistent with performance specification, the inspection can then
be accepted. Previous excavation data can also be used for the
validation. However, if excavation data is unavailable and there is a
previous ILI run, then the metal-loss anomalies, as reported by the
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 13 of 69
Aboutpipelines.com
previous ILI, must be matched and compared to the current ILI to
validate the inspection.
StartValidation
Match known Pipeline feature to
ILI
Pipeline features matching successful
YES
Are there actionable anomalies
YES
Reject ILI run
Is there a previous ILI run on the
line?
No
Excavate anomalies and compare to ILI results
Yes
Are excavation results consistent with performance
Specification?
Compare ILI resultsYes
Are the previous ILI results consistent with performance
specificaiton?
Accept ILI run
YES
Yes NO
No
No
Figure 3: Validation Procedure
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 14 of 69
aboutpipelines.com
5. Process Verification
5.1. Process Overview
Process verification is a systematic and consistent approach to ensure that all
proper procedures were undertaken by the operator and ILI vendor prior to,
during, and after the inspection. The fundamental premise of the methodology is
that high-quality ILI data is a consequence of technology, planning, and
execution. The verification process checks ten parameters. Once these ten
parameters have been assessed, a cumulative assessment of all parameters (and
potential deficiencies) is then reviewed cohesively to ensure that results are still
deemed to be tolerable. The ten parameters are shown in Table 1.
Table 1: Verification Procedure Scorecard Parameters as per API 1163 Item Stage API category API Ref Parameter
1 Pre-run In-Line System Selection
5.4 Tool Selection
2 Pre-run System Results Validation
8.2.2 Inspection System Data (other lines)
3 Pre-run In-Line System Selection
5.3, 7.2 Planning
4 Pre-run System Operational Verification
7.3.2 Function Checks
5 Pre-run System Operational Verification
7.3.3 Mechanical Checks
6 In the pipe System Operational Verification
7.4 Procedure Execution (e.g., pigging procedure, tool speed, etc.)
7 Post-run System Operational Verification
7.5.2 Mechanical Checks
8 Post-run System Operational Verification
7.5.2 Function Check
9 Post-run System Operational Verification
7.5.3 Field Data Quality Check
10 Post-run System Results Validation
8.2.2 & Annex C1
Data Analysis Processes: Quality Checks
For each of these checks, the flowchart in Figure 2 is followed to assign a score to
each item. A score “P” or “Pass” indicates that the proper procedure was followed.
“C” or “Conditional pass” indicates that some irregularities were found in the
procedure, but that the effect on the data is not significant or that additional
actions may be required to ensure the conditional pass is resolved or can be
confirmed to be a pass. For example, in an area where the ILI experienced an
overspeed, the operator may use a different specification to assess the
uncertainty of the features in that particular area. Finally, “F” or “Fail” indicates
that the data is compromised in some way such that the run cannot be accepted.
The final Cumulative Check is a review of all the Conditional Passes to ensure that
the overall effect on the data is acceptable.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 15 of 69
Aboutpipelines.com
The following sections discuss each of these parameters in turn. Specifically,
description of the parameter, motivation of the item, scoring, assessing the
significance of the data impact and potential mitigation options are discussed in
detail. An abridged version of the information is available in the actual scorecard
and guidance document in A1. Scorecard and Guidance Document.
5.2. (Pre-Run) Tool Selection
5.2.1. Parameter Description Given the wide array of tools currently available from a number
of different ILI vendors, this check ensures that an appropriate
tool has been selected in light of the expected defect type(s) on
the pipeline. The primary guidance appears in NACE SP 0102
Table 1; see also A2. NACE Table.
In addition to the NACE guidance, the operator must understand
the capabilities and limitations of the specific tool selected for the
inspection and ensure that the goals of the inspection will be
satisfied.
5.2.2. Motivation The purpose of this check is to ensure that the inspection tool is
capable of assessing the specific threat on the pipeline. An
appropriate inspection tool needs to be selected that can detect
the threats present. During the selection process, the operator
and vendors shall consider the tool’s resolution range (standard
vs. high resolution). The operator shall review the tool’s
performance specifications to verify that the tool is capable of
detecting and sizing the anomalies that are deemed a threat on
the pipeline. The vendor’s tool performance specification shall
contain sizing accuracy standards and confidence levels.
The inclusion of this item ensures that an inspection conducted to
address one threat is not also used to assess threats to which it
is not suited. (For example, an MFL tool may have been run to
assess corrosion, but it should not be used to assess potential
SCC.) Also, the inclusion of this item is required to make the
overall procedure objective. A person not previously involved in
the running of the inspection should verify that the tool is
adequate to the job.
5.2.3. Scoring Table 2: Pre-Run Tool Selection Scoring
Score Scoring Description
F Tool not capable of detection or sizing of expected anomaly type(s).
C Tool capable of detecting anomaly types but limited sizing or detection abilities of expected anomaly type(s).
P Best available technology for detecting and sizing expected anomaly type(s) identified and used.
Note: For a comprehensive Verification Process Scorecard Summary, please
refer to Table 33 in A.6.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 16 of 69
Aboutpipelines.com
The scoring of this parameter is expected to be relatively
straightforward. For example, the use of a transverse MFL tool
will identify general corrosion, but its sizing tolerances are
limited. Thus, if the dimensions of a potentially injurious defect
are beyond the performance specification of the transverse MFL
tool, then this parameter would be deemed a “Fail”.
It should be noted that the “Conditional” pass would be
contingent on the operator confirming that the dimensions of
critical defect size(s), for the pipeline in question, are greater
than the minimum detection and sizing thresholds of the tool.
“Pass” would be reported if high resolution MFL technology was
used to detect general pipeline corrosion. This is intended to
recognize that while MFL technology has some limitations, the
operator is using the best available tool to address corrosion
related pipeline integrity concerns.
5.2.4. Options for Dealing with Compromised Data Quality The impact to data, where the optimal technology is not used,
must be addressed on a case-by-case basis. The guiding principle
remains, as stated above: The operator must ensure that the
dimensions of an injurious defect, for the pipeline in question, are
greater than the minimum detection and sizing thresholds of the
tool.
Should a “Fail” score be appropriate for this parameter, the
options are somewhat limited in that some large-scale program
to prove the integrity of the pipeline must be undertaken. For
example, the operator must re-inspect the line with a more
appropriate tool suited to the specific threat or undertake an
alternative set of activities – such as hydrostatic testing, or direct
assessment.
If a “Conditional” score is given, then a record of the location(s)
(whether the location(s) are limited to specific segment(s) or the
whole pipeline) where the quality of the data may be affected
must be recorded and considered again for the Cumulative
Assessment.
5.3. Inspection System
5.3.1. Parameter Description This check ensures that the inspection tool used in the inspection
has a history of successful runs, and that the inspection system is
likely to perform successfully. An operator may decide to run a
untested technology in a pipeline from time to time, but that run
should not be used to assess the threat on the pipeline without
adequate validation. In addition, if the tool has not previously
conducted an inspection in the operator’s pipeline, but has been
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 17 of 69
Aboutpipelines.com
tested by the vendor, the vendor shall supply relevant
documentation demonstrating successful performance testing.
5.3.2. Motivation Whereas the emphasis of the Tool Selection check is to ensure
that the technology is capable of detecting and sizing the
anomalies, the motivation of this check is to ensure that the
inspection system is able to deliver quality data as demonstrated
by its history of successful runs.
5.3.3. Scoring Table 3: Inspection System Data Check Scoring
Score Scoring Description
F Tool is experimental and there is no established history or it has been demonstrated to have deficiencies in addressing the threat.
C Same model of tool with minor differences (such as diameter) has a history of successful runs to assess the threat, Or the specific model of tool has history of successful runs to assess the threat for other operators, but results of those runs are not available.
P Operator firsthand knowledge of the performance capabilities of the tool and has several successful inspections using the tool.
Note: For a comprehensive Verification Process Scorecard Summary, please
refer to Table 33 in A.6.
The scoring of this parameter is expected to be relatively
straightforward. If the operator has firsthand experience with the
specific ILI vendor’s tool, and the use of that tool has reliably
resulted in successful inspections, then a “Pass” is given.
If, however, the operator does not have firsthand experience with
the specific tool, but has indirect experience or knowledge of the
tool’s performance, a “Conditional” pass is scored. For example, a
“Conditional” is given if the operator has extensive experience
with other tools in the ILI vendor’s fleet, but those tools differ
from the tool used in the current inspection in some way, such as
diameter. Since there are usually a large number of similarities
between a vendor’s 24-inch and 30-inch tool, for example, the
performance of the 30-inch tool is a good indication of the
performance of the 24-inch tool.
This parameter receives a “Fail” if the ILI tool is of an
experimental prototype or if its past runs suggest a high failure
rate.
5.3.4. Options for Dealing with Untested Tools One of the implicit assumptions of this document is that if an
inspection is conducted according to proper procedures, then the
inspection system will perform to its ability. In the case of an
untested inspection system, that ability is not known. Therefore
to accept a tool with no history of successful runs requires a
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 18 of 69
Aboutpipelines.com
more rigorous validation process to ensure that the tool is
accurately reporting the severity of the anomalies.
If a “Conditional” score is given, then a record of the location(s)
(whether the location(s) are limited to specific segment(s) or the
whole pipeline) where the quality of the data may be affected
must be recorded and considered again for the Cumulative
Assessment.
5.4. (Pre-Run) Planning and Preparation
5.4.1. Parameter Description This parameter is a check of the group of activities prior to
executing an inline inspection. The user is referred to NACE
SP0102-2010 (Sections 4, 5, and 6) for details of the types of
activities that are typically undertaken as part of pre-run
planning. As part of the planning procedure, the ILI vendor and
operator should work together to ensure a successful run.
Planning should include, but be not limited to: completion of a
pre-run questionnaire supplied by the ILI vendor, pipeline
cleaning, pipeline geometry assessment, launch and receiver
Trap review, assessment of adequate battery life for inspection
tool (e.g. account for 20 – 30% contingency life longer than
estimated run time), development of an inspection procedure,
inspection scheduling, logistics as well as ensuring appropriate
product type, flows and pressures.
Some of these planning activities may be iterative, such as
logistics, inspection procedures and pipeline operating conditions;
therefore, the operator and vendor shall allow for sufficient time
to complete these activities prior to the launch of the inspection
tool.
5.4.2. Motivation The purpose of this item is to ensure that proper procedures were
followed prior to the running of the tool. In many cases, this
parameter may seem moot after the completion of and
apparently successful inspection. However, the success of an
inspection is dependent on planning and preparations prior to the
running of the tool. Ensuring, for example, that cleaning targets
were met prior to the inspection can avoid degradation of the
data quality. Thus proper planning can sometimes make the
difference between optimal inspection results and simply
adequate results.
5.4.3. Scoring Table 4: Pre-Run Planning Scoring
Score Scoring Description
F Key elements of Pipeline ILI Compatibility Assessment and Inspection planning not conducted.
C Majority of elements of Pipeline ILI Compatibility Assessment and
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 19 of 69
Aboutpipelines.com
Inspection Scheduling completed but undocumented.
P Elements of Pipeline ILI Compatibility and Inspection Scheduling completed and documented.
Note: For a comprehensive Verification Process Scorecard Summary, please
refer to Table 33 in A.6.
The scoring of this parameter is expected to be highly specific to
each situation and somewhat subjective. That is, simple and
straightforward situations (e.g., MFL run in a dry sweet gas line)
for executing an inline inspection will require significantly less
planning compared to more complex scenarios (e.g., ultrasonic
inspection in a liquid slug in a gas line with multiple off-takes and
interconnections). Operators and vendors should document all
planning including the decision whether to conduct or not conduct
specific activities prior to the run.
It is anticipated that at a minimum, operators will follow industry
best practice documentation and conduct planning activities
around the parameters most critical and relevant to the specific
inspection.
It should be noted this is one of a few parameters where a
“Conditional” pass may be assigned even if, in retrospect,
planning activities are deemed insufficient if it is demonstrated
that the data collected by the inline inspection tool was
unaffected. Thus, a “Fail” would only be assigned in situations
where data degradation exists – directly as a result of inadequate
pre-run planning.
5.4.4. Options for Dealing with Compromised Data
Quality The impact to data, where a lack of planning has been identified
as the root cause, must be addressed on a case-by-case basis
since the range of potential outcomes is large. For example, at
one extreme, insufficient planning may lead to a tool lodged in
the line (requiring a cut-out) at a previously unidentified pipe
restriction. At the other end of the spectrum, insufficient planning
of product flows may result in a short speed excursion of the tool
at launch.
The guiding principle remains, as stated -below: The operator
must ensure that the dimensions of an injurious defect, for the
pipeline in question, are greater than the minimum detection and
sizing thresholds of the tool. If the operators cannot be confident
that an injurious defect would be detected, a range of options
exist depending on the length of the area where data has been
compromised. Short sections of data degradation may be
individually assessed and deemed acceptable on the basis of
other integrity related activities such as other ILI runs,
hydrostatic testing, cathodic protection, coating, soil, pipe
properties (i.e., presence of heavy wall pipe), and direct
assessment.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 20 of 69
Aboutpipelines.com
Should a “Fail” score be appropriate for this parameter (i.e.,
significant data degradation), the options are somewhat limited
in that some large-scale program to prove the integrity of the
pipeline must be undertaken. For example, the operator must
re-inspect the line having remedied the planning deficiency or
undertake an alternative set of activities – such as hydrostatic
testing, or direct assessment.
If a “Conditional” score is given, then a record of the location(s)
(whether the location(s) are limited to specific segment(s) or the
whole pipeline) where the quality of the data may be affected
must be recorded and considered again for the Cumulative
Assessment.
5.5. (Pre-Run) Function Checks
5.5.1. Parameter Description This parameter is a check of the group of activities that an ILI
vendor carries out to ensure the functional integrity of the tool
prior to loading the inspection tool into the launcher barrel. As
such, function checks are expected to be specific to each vendor
and technology used. The function checklist should be provided
by the ILI vendor and the items should be standardized and
identified in advance of the inspection. These include, but are not
limited to appropriate initialization of all components, the
adequacy and availability of the power supply, confirming sensors
are operational, and confirming adequacy, and availability of data
storage.
5.5.2. Motivation The purpose of this item is to ensure that the inspection tool is in
good working condition, which requires that the tools’ mechanical
components perform as designed, prior to the inspection. The
documentation of this check is required by API 1163, and its
documentation is indicative of the ILI vendor’s diligence in
following established Standards and Guidelines.
5.5.3. Scoring Table 5: Pre-Run Function Check Scoring
Score Scoring description
F Significant function checks not passed.
C Significant function checks passed but checks are undocumented.
P All function checks passed and documented.
Note: For a comprehensive Verification Process Scorecard Summary,
please refer to Table 33 in A.6.
The scoring of this parameter is expected to be relatively
straightforward. For example, it would be surprising (but
possible) that the tool fails a significant function check(s) and is
still launched. If adequate checks are not performed, the tool
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 21 of 69
Aboutpipelines.com
may experience a failure soon after launch. In such a case, the
oversight would result in some significant data degradation, and
the parameter would be deemed a “Fail”. However, in some cases
documentation confirming the function check may be missing. If
the tool passes a documented post-run function check or if the
tool is received and no data degradation has been identified, then
a “Conditional” pass would be assigned. A “Pass” would be
reported if all function checks were passed and documentation to
this effect was readily available.
It should be noted this is one of a few parameters where a
“Conditional” pass may be assigned, even if the function check
was deemed inadequate, provided that it is demonstrated that
the data collected by the inline inspection tool was unaffected.
5.5.4. Options for Dealing with Compromised Data Quality While not expected to happen often, the situations where
function checks are not passed prior to launch are expected to
have a significant impact on the data collected. While the
mitigation options must be addressed on a case-by-case basis,
the operator must ensure that the dimensions of an injurious
defect, for the pipeline in question, are greater than the
minimum detection and sizing thresholds of the tool.
Should a “Fail” score be appropriate for this parameter, the
options are somewhat limited in that some large-scale program
to prove the integrity of the pipeline must be undertaken. For
example, the operator must re-inspect the line or undertake an
alternative set of activities – such as hydrostatic testing, or direct
assessment.
If a “Conditional” score is given, then a record of the location(s)
(whether the location(s) are limited to specific segment(s) or the
whole pipeline) where the quality of the data may be affected
must be recorded and considered again for the Cumulative
Assessment.
5.6. (Pre-Run) Mechanical Checks
5.6.1. Parameter Description This parameter is a check of the group of activities that an ILI
vendor carries out to ensure the mechanical integrity of the tool
prior to loading it into the launcher barrel. As such, pre-run
mechanical checks are expected to be largely visual and specific
to each vendor and technology used. The pre-run mechanical
checklist should be provided by the ILI vendor, standardized, and
identified in advance of the inspection. These include, but are not
limited to: general visual inspection, confirming good pressure
seals around electronic components, ensuring adequate integrity
of cups, and ensuring all wheels are intact and moving
appropriately.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 22 of 69
Aboutpipelines.com
5.6.2. Motivation The purpose of this item is to ensure that the inspection tool is in
good working condition, which requires that the tools’ mechanical
components perform as designed. The documentation of this
check is required by API 1163 and its documentation is indicative
of the ILI vendor’s diligence in following established Standards
and Guidelines.
5.6.3. Scoring Table 6: Pre-Run Mechanical Check Scoring
Score Scoring Description
F Significant mechanical checks not passed.
C Significant mechanical checks passed but checks are undocumented.
P All mechanical checks passed and documented.
Note: For a comprehensive Verification Process Scorecard Summary, please
refer to Table 33 in A.6.
The scoring of this parameter is expected to be relatively
straightforward. For example, it would be surprising (but
possible) that the tool fails a significant mechanical check but is
still launched. If adequate checks are not performed, the tool
may experience a failure soon after launch. In such a case, the
oversight would result in some significant data degradation, and
the parameter would be deemed a “Fail”. However, in some cases
documentation confirming the mechanical check may be missing.
If the tool passes a documented post-run mechanical check or if
the tool is received and no data degradation has been identified,
then a “Conditional” pass would be assigned. A “Pass” would be
reported if all mechanical checks were passed and documented.
It should be noted this is one of a few parameters where a
“Conditional” pass may be assigned even if, in retrospect, the
mechanical check was deemed inadequate if it is demonstrated
that the data collected by the inline inspection tool was
unaffected.
5.6.4. Options for Dealing with Compromised Data Quality While not expected to happen often, the situations where
mechanical checks are not passed prior to launch are expected to
have a significant impact on the data. While the mitigation
options must be addressed on a case-by-case basis, the operator
must ensure that the dimensions of an injurious defect, for the
pipeline in question, are greater than the minimum detection and
sizing thresholds of the tool.
Should a “Fail” score be appropriate for this parameter, the
options are somewhat limited in that some large-scale program
to prove the integrity of the pipeline must be undertaken. For
example, the operator must re-inspect the line or conduct an
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 23 of 69
Aboutpipelines.com
alternative set of activities – such as hydrostatic testing, or direct
assessment.
If a “Conditional” score is given, then the location(s) (whether
the location(s) are limited to specific segment(s) or the whole
pipeline) where the quality of the data may be affected must be
recorded and considered again for the Cumulative Assessment.
5.7. (In the Pipe) Procedure Execution
5.7.1. Parameter Description A job specific plan that includes technical requirements,
responsibility, emergency contact, procedures, work site
preparation and mobilization is commonly prepared to support
execution of the inspection. This parameter checks the group of
activities required to execute a successful inline inspection based
on the requirements of the plan. These include, but are not
limited to:
Check the tool run was executed as per the planned pigging
procedure.
Check that the line condition parameters (fluid composition,
flow rate, temperature, and pressure) were in accordance
with the planned procedure.
Check that the line conditions for tool launch were as
expected and did launch proceed as planned.
Check that the line conditions for tool receive were as
expected and did receive proceed as planned.
Check that the tool speed was within the planned range for
the length of the run. (If deviations did occur, were they
planned or expected and assessed in advance?)
Check that the tracking of the tool was according to plan.
5.7.2. Motivation This parameter is designed to ensure that the actual inspection
was conducted in such a way as to ensure high-quality inspection
data. The documentation of this check is indicative of the ILI
vendor’s diligence in following established Standards and
Guidelines.
5.7.3. Scoring Table 7: Procedure Execution Scoring
Score Scoring Description
F Inspection not carried out as per inspection procedure with potential material impact to data quality.
C Inspection not carried out as per inspection procedure but deviations are not material to data quality.
P Inspection carried out as per inspection procedure.
Note: For a comprehensive Verification Process Scorecard Summary, please
refer to Table 33 in A.6.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 24 of 69
Aboutpipelines.com
The scoring of this parameter is expected to be relatively
straightforward. For example, if there were deviations from the
planned procedure and significant data impacts resulting (i.e.,
impacts that could not be managed through alternative means),
the parameter would be deemed a “Fail”. However, in some cases
there may have been deviations from the planned procedure but
there was little or no impact to data quality, in these cases, a
“Conditional” pass would be assigned. A “Pass” would be reported
when the inspection procedure was executed as planned and
documentation to this effect was readily available.
5.7.4. Options for Dealing with Compromised Data Quality The impact to data due to deviations from the planned inspection
procedure must be addressed on a case-by-case basis, since the
range of potential outcomes is large. For example, long speed
excursions at wall thickness changes may require restating the
tool performance specification for the entire length of the line.
However, short speed excursions at launch in heavy-wall yard
piping may be deemed tolerable.
The guiding principle remains, as stated above: The operator
must ensure that the dimensions of an injurious defect, for the
pipeline in question, are greater than the minimum detection and
sizing thresholds of the tool. If the operator cannot be confident
that an injurious defect would be detected, a range of options
exist depending on the location and length of the area where
data has been compromised. Short sections of data degradation
may be individually assessed and deemed acceptable on the
basis of other integrity related activities such as other ILI runs,
hydrostatic testing, cathodic protection, coating, soil, pipe
properties (i.e., presence of heavy wall pipe) and direct
assessment.
Should a “Fail” score be appropriate for this parameter (i.e.,
significant data degradation), the options are somewhat limited
in that some large-scale program to prove the integrity of the
pipeline must be undertaken. For example, the operator must
re-inspect the line having remedied the deviation in the
inspection procedure or conduct an alternative set of activities –
such as hydrostatic testing, or direct assessment.
If a “Conditional” score is given, then the location(s) (whether
the location(s) are limited to specific segment(s) or the whole
pipeline) where the quality of the data may be affected must be
recorded and considered again for the Cumulative Assessment.
5.8. (Post-Run) Mechanical Check
5.8.1. Parameter Description This parameter checks the group of activities that an ILI vendor
carries out to ensure the mechanical integrity of the tool upon
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 25 of 69
Aboutpipelines.com
receive at the end of the run. As such, post-run mechanical
checks are expected to be largely visual and specific to each
vendor and technology used. The post-run mechanical checklist
should be provided by the ILI vendor, standardized and identified
in advance of the inspection. These include, but are not limited to
assessing: general state of the tool, pressure seals around
electronic components, integrity of cups, tool cleanliness, location
of debris accumulation, tool wear as well as ensuring all parts are
intact and moving appropriately. It is recommended that the
checklist be appended with photographs of the tool and any
damage to mechanical components
5.8.2. Motivation The purpose of this item is to ensure that the inspection tool was
not damaged during the course of the inspection. The
documentation of this check is indicative of the ILI vendor’s
diligence in following established Standards and Guidelines.
5.8.3. Scoring Table 8: Post-Run Mechanical Check Scoring
Score Scoring Description
F Significant tool wear, damage or debris with material impact to data.
C Tool wear, damage or debris observed with no material impact to data.
P Tool received in good mechanical condition (with no unexpected tool wear, damage or debris).
Note: For a comprehensive Verification Process Scorecard Summary, please
refer to Table 33 in A.6.
The scoring of this parameter is expected to be relatively
straightforward. For example, damage to the tool to the point of
significant data loss or degradation would be deemed a “Fail”.
However, in some cases the tool may have experienced
unexpected damage or wear, but if no data degradation is
identified then a “Conditional” pass would be assigned. A “Pass”
would be reported if all mechanical checks were passed and
documented, upon tool receive.
5.8.4. Options for Dealing with Compromised Data Quality The impact to data, due to a mechanical issue, must be
addressed on a case-by-case basis, since the range of potential
outcomes is large. For example, an unexpected pipeline
restriction may damage the tool and lead to complete data loss.
At the other end of the spectrum, mechanical damage may take
the form of the loss of a single sensor near or at the end of the
run.
If the mechanical check is undocumented, the operator must
demonstrate that the tool was not damaged in the pipeline by
other means. Documentation of this check is important because
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 26 of 69
Aboutpipelines.com
the demonstration that the tool was not damaged may be
difficult.
The guiding principle remains, as stated above: The operator
must ensure that the dimensions of an injurious defect, for the
pipeline in question, are greater than the minimum detection and
sizing thresholds of the tool. If the operators cannot be confident
that an injurious defect would be detected, a range of options
exist depending on the location and length of the area where
data has been compromised.
Short sections of data degradation may be individually assessed
and deemed acceptable on the basis of other integrity related
activities such as other ILI runs, hydrostatic testing, cathodic
protection, coating, soil, pipe properties (i.e., presence of heavy
wall pipe) and direct assessment.
Should a “Fail” score be appropriate for this parameter (i.e.,
significant data degradation), the options are somewhat limited
in that some large-scale program to prove the integrity of the
pipeline must be undertaken. For example, the operator must
re-inspect the line having remedied the cause of the mechanical
problem (e.g., additional cleaning runs or removal of a diameter
restriction) or undertake an alternative set of activities – such as
hydrostatic testing, or direct assessment.
If a “Conditional” score is given, then the location(s) (whether
the location(s) are limited to specific segment(s) or the whole
pipeline) where the quality of the data may be affected must be
recorded and considered again for the Cumulative Assessment.
5.9. (Post-Run) Function Check
5.9.1. Parameter Description This parameter is a check of the group of activities that an ILI
vendor carries out to ensure functional integrity of the tool upon
receive at the end of the run. As such, function checks are
expected to be specific to each vendor and technology used. The
function checklist should be provided by the ILI vendor,
standardized, and identified in advance of the inspection. These
checks include, but are not limited to appropriate operation of all
components, the adequacy and availability of the power supply,
confirming sensors are operational, and confirming adequacy and
availability of data storage.
5.9.2. Motivation The purpose of this item is to ensure that the inspection tool did
not experience an internal failure during the course of the
inspection. The documentation of this check is indicative of the
ILI vendor’s diligence in following established Standards and
Guidelines.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 27 of 69
Aboutpipelines.com
5.9.3. Scoring Table 9 Post-Run Function Check Scoring
Score Scoring Description
F Significant function checks not passed.
C Significant function checks passed were not documented, but that the proper functioning of the tool can be Verified by other means throughout the length of the run.
P Function checks passed and documented
Note: For a comprehensive Verification Process Scorecard Summary, please
refer to Table 33 in A.6.
The scoring of this parameter is expected to be relatively
straightforward. For example, malfunctioning of the tool to the
point where there is significant data loss or degradation would be
deemed a “Fail”. However, if the tool experienced a functional
issue but no significant data degradation, then a “Conditional”
pass would be assigned. A “Pass” would be reported if all function
checks were passed upon tool receive and documented.
5.9.4. Options for Dealing with Compromised Data Quality The impact to data, due to a tool malfunction, must be addressed
on a case-by-case basis since the range of potential outcomes is
large. For example, an electronics failure at launch would require
a re-inspection. At the other end of the spectrum, an electronics
failure at receive would not be expected to be material to the
quality of the data collected.
If the function check is undocumented, the operator must
demonstrate that the tool was operating properly for the entire
length of the inspection. Documentation of this check is
important because the demonstration that the tool was operating
properly may be difficult.
The guiding principle remains, as stated above: The operator
must ensure that the dimensions of an injurious defect, for the
pipeline in question, are greater than the minimum detection and
sizing thresholds of the tool. If the operators cannot be confident
that an injurious defect would be detected, a range of options
exist depending on the location and length of the area where
data has been compromised. However, given the limited ability to
analyze tool data in the field, detailed analysis of data
degradation is unlikely to occur until data analysis in the office
environment is undertaken. As such, this check is intended to
identify major data shortfalls and degradation issues.
Should a “Fail” score be appropriate for this parameter (i.e.,
significant data degradation), the options are somewhat limited
in that some large-scale program to prove the integrity of the
pipeline must be undertaken. For example, the operator must
re-inspect the line having remedied the suspected cause of the
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 28 of 69
Aboutpipelines.com
data degradation or undertake an alternative set of activities –
such as hydrostatic testing, or direct assessment.
If a “Conditional” score is given, then the location(s) (whether
the location(s) are limited to specific segment(s) or the whole
pipeline) where the quality of the data may be affected must be
recorded and considered again for the Cumulative Assessment.
5.10. (Post-Run) Field Data Quality Check
5.10.1. Parameter Description This parameter is the group of checks that an ILI vendor carries
out to ensure integrity of the data collected upon receive at the
end of the run. These checks are expected to be specific to each
vendor and technology used. The Field Data checklist should be
provided by the ILI vendor, standardized, and identified in
advance of the inspection. These checks include, but are not
limited to: amount of data collected, length of line inspected, and
circumferential and linear continuity of data.
5.10.2. Motivation The purpose of this item is to ensure that the inspection tool
collected data for the full length of the line. This check is intended
to identify any major data shortfalls and/or degradation issues,
which would prevent the vendor from meeting the performance
specifications. The documentation of this check is indicative of the
ILI vendor’s diligence in following established Standards and
Guidelines.
5.10.3. Scoring Table 10: Post-Run Field Data Quality Check Scoring
Score Scoring Description
F Tool unable to meet stated specifications due to significant lack of data integrity
C Tool unable to meet stated specifications but manageable through further analysis
P Tool able to meet stated specifications for entire length of run
Note: For a comprehensive Verification Process Scorecard Summary, please
refer to Table 33 in A.6.
The scoring of this parameter is expected to be relatively
straightforward. For example, significant data loss or degradation
would be deemed a “Fail”. However, if the tool experienced data
loss, but based on the level of detail available in the field, the
vendor may feel that the performance specification can largely
still be met then a “Conditional” pass would be assigned. A “Pass”
would be reported if all data checks were passed upon tool
receive and documented.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 29 of 69
Aboutpipelines.com
5.10.4. Options for Dealing with Compromised Data
Quality Data degradation identified through this parameter must be
addressed on a case-by-case basis since the range of potential
outcomes is large. For example, complete loss of data at launch
would require a re-inspection. At the other end of the spectrum,
the loss of a single sensor half way through the run may not be
deemed material to the quality of the data collected.
The guiding principle remains, as stated above: The operator
must ensure that the dimensions of an injurious defect, for the
pipeline in question, are greater than the minimum detection and
sizing thresholds of the tool. If the operators cannot be confident
that an injurious defect would be detected, a range of options
exist depending on the location and length of the area where data
has been compromised. However, given the limited ability to
analyze tool data in the field, detailed analysis of data
degradation is unlikely to occur until data analysis in the office
environment is undertaken. As such, this check is intended to
identify major data shortfalls and degradation issues.
Should a “Fail” score be appropriate for this parameter (i.e.,
significant data degradation), the options are somewhat limited in
that some large-scale program to prove the integrity of the
pipeline must be undertaken. For example, the operator must re-
inspect the line having remedied the suspected cause of the data
degradation or undertake an alternative set of activities – such as
hydrostatic testing, or direct assessment.
If a “Conditional” score is given, then the location(s) (whether the
location(s) are limited to specific segment(s) or the whole
pipeline) where the quality of the data may be affected must be
recorded and considered again for the Cumulative Assessment.
5.11. (Post-Run) Data Analysis Process Check
5.11.1. Parameter Description This parameter is the group of checks that the data was properly
handled and analyzed by the vendor in the production of the final
report. These checks are specific to each vendor and technology
used.
The data analysis should be discussed and decided jointly by the
operator and ILI vendor. The operator and ILI vendor should
agree on items such as sizing algorithms to use, amount of
manual intervention, filtering of reported anomalies, clustering
rules, burst pressure procedure, etc. In addition, the operator
should also discuss analyst’s qualifications (Level 1, 2, or 3) for
who should perform the analysis.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 30 of 69
Aboutpipelines.com
It is recommended that the above listed requirements be
discussed and agreed upon in the Planning and Preparation stage,
to allow the vendor to properly secure the required resources,
including analysts and software programs, for instance, and to
ensure the tool is capable of performing the scope, before the
inspection.
The Data Analysis checklist should be based on the analysis
procedure provided by the ILI vendor, standardized, and
identified in advance of the analysis. These checks include, but
are not limited to: amount of data collected, continuity of data,
appropriate sensor response(s), sizing algorithms, manual
checking, clustering rules, burst pressure procedure, execution of
data analysis procedures as well as use of appropriate input
parameters (such as pipeline diameter, wall thickness, grade,
etc.).
5.11.2. Motivation This final check is to ensure that the raw data from the inspection
has been properly analyzed by the ILI vendor and that the final
report will satisfy the requirements of the inspection. The
documentation of this check is indicative of the ILI vendor’s
diligence in following established Standards and Guidelines.
5.11.3. Scoring Table 11: Post-Run Data Analysis Processes Scoring
Score Scoring Description
F Results of data analysis quality checks are not acceptable.
C Significant data quality checks passed, but quality checks initially undocumented or reanalysis was required.
P Data quality checks passed and documented.
Note: For a comprehensive Verification Process Scorecard Summary, please
refer to Table 33 in A.6.
The scoring of this parameter is expected to be relatively
straightforward. The ILI vendor should provide the operator with
the agreed checklist of all of their procedures for checking and
analyzing the data.
A “Pass” is given if the vendor’s checklist is completed as
previously agreed and the number, distribution, and severity of
anomalies are consistent with the expectations of the vendor
considering the age, coating, previous inspections, and history of
the pipeline.
A “Conditional” pass is given if the vendor has not supplied the
completed checklist or if any deficiency is noted but corrected by
a reanalysis of the data. If following a reanalysis of the data, the
number, distribution, or severity of the anomalies is not
consistent with expectations, then an independent review (audit)
of the analysis procedures may be required.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 31 of 69
Aboutpipelines.com
A “Fail” is given for any situation that cannot be corrected by
reanalysis of the data. Such situations may indicate a problem in
the selection, preparation, or running of the tool.
5.11.4. Options for Data Analysis Issues Depending on the issue, the results for this parameter must be
addressed on a case-by-case basis since the potential causes of
the issue are broad ranging. Issues regarding the analysis of the
data are likely to affect the entire inspection. However some
issues may affect only specific locations on the pipeline. In many
cases, reanalysis of the data or an independent audit of the
analysis may be sufficient to address the concern.
Should a “Fail” score be appropriate for this parameter (i.e.,
significant data degradation), the options are somewhat limited in
that some large-scale program to prove the integrity of the
pipeline must be undertaken. For example, the operator must
re-inspect the line having remedied the suspected cause of the
data degradation or undertake an alternative set of activities –
such as hydrostatic testing, or direct assessment.
If a “Conditional” score is given, then the location(s) (whether the
location(s) are limited to specific segment(s) or the whole
pipeline) where the quality of the data may be affected must be
recorded and considered again for the Cumulative Assessment.
5.12. (Post-Run) Cumulative Assessment
5.12.1. Parameter Description This parameter provides a means of assessing all of the
parameters – taken as a whole – to determine whether the tool
performance was acceptable. That is, this parameter is a
mechanism to ensure that (potential) sub-optimal performance
across all parameters does not result in an unacceptable run even
if all parameters taken individually are deemed acceptable (i.e.,
conditional passes). Relevant considerations include, but are not
limited to, the following:
Can data gaps be mitigated effectively using alternative
methods?
Can any data gaps actually be addressed through re-running
the tool or are line conditions such that similar challenges will
remain?
Are any “Conditional” scores cumulative in nature?
Do “Conditional” scores of different parameters affect the
same locations?
5.12.2. Scoring Table 12: Post-Run Cumulative Assessment Scoring
Score Scoring Description
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 32 of 69
Aboutpipelines.com
F Cumulative impact of “Conditional” passes deemed to materially impact ILI results.
A Cumulative impact of “Conditional” passes is unclear and may have impacted ILI results.
P The cumulative impact of all “Conditional” passes, if any, deemed to be tolerable.
Note: For a comprehensive Verification Process Scorecard Summary, please
refer to Table 33 in A.6.
Scoring for this final parameter is in fact the scoring of the entire
verification Process. The final score may be either “Fail”, “Pass”,
or “Ambiguous”, depending on whether the “Conditional” passes
of the previous checks are cumulative or not. Degradation is
cumulative if one issue magnifies any pre-existing data
degradation (such as a tool over-speed in a location where the
tool performance is already compromised by debris related sensor
lift off). Conversely, issues are not cumulative if their impact on
data quality is largely independent (such as a run where the tool
over-speeds at launch and experiences debris issues at receive).
If two or more “Conditional” passes of the previous checks affect
the same length of pipe and the degradation are cumulative and
materially impact the results of the ILI, a “Fail” would be
assigned. If the operator can clearly confirm that the impact of
any “Conditional” scores are not cumulative and are manageable,
a “Pass” may be assigned. However, should the results be
ambiguous, due to the specifics of the situation, an “Ambiguous”
score is assigned and further analysis is required to manage the
threat on the pipeline.
6. Validation Validation is the process that compares the data collected and reported by the ILI tool to
some independent reference data to ensure the ILI tool meets its performance
specification. Depending on the available data and the results of the ILI inspection, the
validation procedure may consist of different comparisons. In all cases, the validation
must include a comparison of reported pipeline (non-defect) features such as girth welds
and wall thickness changes to as-built records (or similar record). If there are actionable
anomalies, then they must be excavated and compared to the ILI results. In addition,
previous excavation data can also be another reliable source to validate the ILI run.
External features that have been recoated can be used to validate both Magnetic Flux
Leakage (MFL) and Ultrasonic (UT) inspections and external features under a steel repair
sleeve can be used to validate UT inspection only. Finally, if there is a previous ILI run,
then the reported metal-loss anomalies must be compared to the previous inspection.
6.1. Known Pipeline Features
6.1.1. Description The simplest validation process is the comparison of non-defect
features on the pipeline to the ILI report. The features to be
compared should include but are not limited to:
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 33 of 69
Aboutpipelines.com
Girth welds,
Wall thickness changes,
Tees,
Valves,
Above ground markers,
Pipeline asset data that should be validated include but are not
limited to:
Nominal and local wall thickness
Long seam orientation and weld type
Joint length
6.1.2. Procedure The validation procedure using known pipeline features is listed
below. The procedure is meant as a guideline rather than rigorous
set of instructions. Deviations from the procedure may be
required in some circumstances.
1. Obtain the most recent and complete reference list of pipeline
features: The listing may be the as-built, metal-loss inspection
report, non-metal-loss inspection report, or other reliable source.
2. Match the girth weld locations from the reference list to the
reported ILI girth welds.
3. Compile a list of the matched girth welds.
4. Compile a list of the girth welds reported in the reference list but
not reported by the ILI.
5. Compile a list of the girth welds reported by the ILI but not
included in the reference list.
6. Calculate the percentage of matched girth welds:
𝑁𝑢𝑚𝑏𝑒𝑟 𝑜𝑓 𝑚𝑎𝑡𝑐ℎ𝑒𝑑 𝐺𝑖𝑟𝑡ℎ 𝑤𝑒𝑙𝑑𝑠
𝑇𝑜𝑡𝑎𝑙 𝑛𝑢𝑚𝑏𝑒𝑟 𝑜𝑓 𝑔𝑖𝑟𝑡ℎ 𝑤𝑒𝑙𝑑𝑠 𝑖𝑛 𝑡ℎ𝑒 𝑅𝑒𝑓𝑒𝑟𝑒𝑛𝑐𝑒 𝐿𝑖𝑠𝑡𝑖𝑛𝑔 × 100
7. Identify all pipeline features from the reference list and match
them to the corresponding ILI reported feature.
8. Identify all ILI features not listed in the reference list. Thoroughly
investigate the source of all of these discrepancies.
9. Identify all girth weld discrepancies if any. Thoroughly investigate
the source of all the discrepancies.
6.1.3. Acceptance Criteria To satisfy this validation procedure, the ILI report must
successfully meet detection standards and location-accuracy
specifications. The ILI must successfully identify and report all
girth welds and long seam weld types. When required, either the
ILI report or the operator’s reference listing should be updated in
order to get a match of all the girth weld numbers.
In addition, if the pipeline pipe asset data do not match the
reference listing, the cause must be investigated and all
unmatched pipe asset data must be reconciled. The report should
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 34 of 69
Aboutpipelines.com
only be accepted once all girth weld numbers and pipe asset data
completely match the reference listing. The reported location of
all features must meet location-accuracy specifications to enable
the excavation of any reported feature.
6.1.4. Special Considerations If the ILI report initially fails to meet the above requirements, the
operator must investigate the cause. To meet the acceptance
criteria, the operator may choose to ignore the part of the
inspection in the stations at launch or receive where there are
many short pipe joints. Also, when replaced or rerouted segments
have been identified in the ILI report, the operator should
investigate and validate the information and then assign new
Girth Weld (GW) numbers. If the GW numbers assigned are
different from the GW reported on the ILI report, the operator
should request the ILI vendor to update the ILI report with the
new GW numbers. Finally, any unexplained features reported by
the ILI should be investigated to determine their cause.
6.2. Comparison with Previous ILI
6.2.1. Description The comparison with a previous ILI is likely the most
comprehensive method for validating the results of an ILI
inspection. Unlike excavating the pipeline, a previous ILI enables
the operator to systematically compare all anomalies of the
current inspection to the previous reference inspection.
6.2.2. Validation using a previous ILI consists of two parts: detection and accuracy. By using a previous inspection, the operator can confirm both the detection capabilities and the accuracy of the current inspection. Procedure The validation procedure using previous ILI data is listed below.
The procedure is meant as a guideline rather than rigorous set of
instructions. Deviations from the procedure may be required in
some circumstances. Furthermore, a lengthy interval (e.g. more
than 5 years) between ILI inspections or the use of very different
technologies can make matching difficult if not impossible. If
there is insufficient similarity between the inspections to make
adequate matches, then the current inspection cannot be
validated by the comparison. However, it would not necessarily
lead to the rejection of the ILI run, since the cause of the
discrepancy may be the previous ILI run.
6.2.2.1. Validation Parameters
The ILI validation criterion is based on the assumptions and
calculations in A4. Validation using a Previous ILI. The
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 35 of 69
Aboutpipelines.com
validation parameters for a typical run are summarized in
Table 13.
Table 13: ILI Validation Parameters 𝑇 Tolerance to be validated in
the current ILI. This value is 10% if the specified accuracy of the current ILI is ±10% NWT, 80% of the time.
𝑪 Specified confidence level (or certainty) of the current ILI.
This value is 80% if the specified accuracy of the current ILI is ±10% NWT, 80% of the time.
𝑻𝟏 Specified tolerance of the previous ILI inspection.
This value is 10% if the accuracy of the previous ILI is ±10% NWT, 80% of the time.
𝑪𝟏 Specified confidence level (or certainty) of the previous ILI.
This value is 80% if the accuracy of the previous ILI is ±10% NWT, 80% of the time.
𝑻𝒖𝒑𝒑𝒆𝒓 Upper bound of acceptance for the ILI tolerance.
CEPA recommends that this value be 1.1 × 𝑇. Thus in most
cases 𝑇𝑢𝑝𝑝𝑒𝑟 = 11%.
𝟏 − 𝜶 Confidence level for the tolerance validation.
1 − 𝛼 is commonly set at 95%, which makes 𝛼 = 0.05.
𝑵 Minimum matched sample size Based on the above parameters and the calculations in A4.
Validation using a Previous ILI, the minimum sample size is 513 matches.
6.2.2.2. Match ILI Anomalies
Using the procedure outlined in A3. Matching, match the
girth welds and ILI anomalies between the current inspection
and the previous reference inspection.
6.2.2.3. Depth Difference Statistics
For each matched pair of anomalies calculate the apparent
difference in depth:
Δ𝑖 = 𝑑𝑖 − 𝑑𝑟𝑖
Where 𝑑𝑖 is the depth of the 𝑖’th anomaly in the current
inspection;
𝑑𝑟𝑖 is the depth of the 𝑖’th anomaly in the previous
reference inspection; and
Δ𝑖 is the difference in depth of the 𝑖’th anomaly.
Calculate the mean, Δ̅, and standard deviation, 𝑠Δ, of the
difference in depths:
Δ̅̅̅ =1
𝑛∑Δ𝑖
𝑛
𝑖=1
,
And
𝑠Δ = √1
𝑛 − 1∑(Δ𝑖 − Δ̅)2
𝑛
𝑖=1
.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 36 of 69
Aboutpipelines.com
6.2.2.4. Acceptance Criteria
The ILI validation criterion is based on the assumptions and
calculations in A4. Validation using a Previous ILI. The ILI
depth accuracy is validated if:
1. The number of available matches is 513 or more, or if
the conditions in A4.4.4. Minimum Sample Size are
satisfied and 2. The calculated value, 𝑠Δ, of the standard deviation of the
difference in depth is less than 11.6% however, this
value could be different depending on the performance
specifications provided by the vendor.
6.2.3. Detection Validation Due to difference in resolution and reporting criteria of the two
inspections, the number of reported anomalies can differ greatly.
In most cases the difference in reported anomalies between two
inspections is due to small anomalies with depths less than
15%NWT. Differences in the reporting of these small shallow
anomalies are not significant to the performance of the tool.
However, the operator should thoroughly investigate differences
between the inspections for any anomaly with a depth greater
than 40%NWT.
Any failure of the current inspection to detect a bona fide metal-
loss anomaly with a depth of 40% or greater s could invalidate
the run. The remedy of such a situation depends on the cause of
the missed anomaly. In many cases, the deficiency can be
corrected by a reanalysis of the data. If the anomaly was not
detected because of a deficiency in the raw data, then a rerun
should be considered.
6.3. Validation from Excavation Data
6.3.1. Description Comparison with excavation data has been the standard method
for validating the results of an ILI inspection. Unlike the
comparison with a previous ILI, excavation data compares only a
limited number of ILI anomalies. However, in-the-ditch
measurements can be much more accurate than a previous ILI.
Validation using excavation data consists of two parts: detection
and accuracy. It is also useful to include the validation of
identification capabilities.
6.3.2. Procedure The validation procedure using excavation data is listed below.
The procedure is meant as a guideline rather than rigorous set of
instructions. Deviations from the procedure may be required in
some circumstances.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 37 of 69
Aboutpipelines.com
6.3.2.1. Validation Parameters
The ILI validation criterion is based on the assumptions and
calculations in A4. Validation using a Previous ILI. The
validation parameters for a typical run are summarized in
Table 14.
Table 14: ILI Validation Parameters 𝑻 Tolerance to be
Validated in the current ILI.
This value is 10% if the Specified accuracy of the current ILI is ±10% NWT, 80% of the time.
𝑪 Specified confidence level (or certainty) of the current ILI.
This value is 80% if the specified accuracy of the previous ILI is ±10% NWT, 80% of the time.
𝑻𝟏 Specified tolerance of the in-the-ditch measurements.
If the device used in the field is highly accurate, then the field measurements can be assumed to have no error. Or refer PRCI project EC-4-2 for depth error of commonly used NDE devices
𝑪𝟏 Specified confidence level (or certainty) of in-the-ditch measurements.
This value is assumed to be 95%.
𝑻𝒖𝒑𝒑𝒆𝒓 Upper bound of acceptance for the ILI tolerance.
CEPA recommends that this value be 1.1 × 𝑇. Thus in most cases
𝑇𝑢𝑝𝑝𝑒𝑟 = 11%.
𝟏 − 𝜶 Confidence level for the tolerance validation.
1 − 𝛼 is commonly set at 95%, which makes 𝛼 = 0.05.
𝑵 Minimum comparison sample size
Based on the above parameters and the calculations in A4.
Validation using a Previous ILI, the minimum sample size is 134 comparisons. Note that if the NDT measurements in the field have
significant error, then the procedure in in A4. Validation using a
Previous ILI should be used to calculate the minimum sample size.
6.3.2.2. Depth Difference Statistics
For each comparison calculate the apparent difference in
depth:
Δ𝑖 = 𝑑𝑖 − 𝑑𝑟𝑖
Where 𝑑𝑖 is the depth of the 𝑖’th anomaly in the
inspection;
𝑑𝑟𝑖 is the corresponding in-the-ditch depth of
the 𝑖’th anomaly; and
Δ𝑖 is the difference between the in-the-ditch
measurement and the ILI reported depth of the 𝑖’th anomaly.
Calculate the mean, Δ̅, and standard deviation, 𝑠Δ, of the
difference in depths:
Δ̅ =1
𝑛∑ Δ𝑖
𝑛
𝑖=1
,
And
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 38 of 69
Aboutpipelines.com
𝑠Δ = √1
𝑛 − 1∑(Δ𝑖 − Δ̅)2
𝑛
𝑖=1
.
6.3.2.3. Acceptance Criteria
The ILI validation criterion is based on the assumptions and
calculations in A4. Validation using a Previous ILI. The ILI
depth accuracy is validated if:
1. The number of available comparisons is 134 or more. 2. The calculated value, 𝑠Δ, of the standard deviation of the
difference in depth is less than 8.6%, however, this
value could be different depending on the performance
specifications provided by the vendor.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 39 of 69
Aboutpipelines.com
A1. Scorecard and Guidance Document
Tables 15 to 27 provide guidance for the completion of the scorecard. Any supporting
documentation would be put on file with the ILI report.
Table 15: Guidance for Parameter #1 Pre-Run Tool Selection Item # 1
Parameter Tool Selection
Stage Pre-run
API Category In-Line System Selection
API 1163 Reference 5.4
Score (F/C/P) “Fail” - Tool not capable of detection or sizing of expected anomaly type(s). “Conditional” - Tool capable of detecting anomaly types but has limited sizing ability. “Pass” – Tool is best available technology for detecting and sizing expected anomaly types(s).
Flowchart Box 1A (Pass Check?)
Use Guidance Document: SP0102-2010 - specifically Table 1. “Pass” if the tool was the tool is the best available technology relative to the purpose of the inspection.
Flowchart Box 1B (Is the impact on data significant?)
“Conditional” if the operator can establish that the integrity of the pipeline is not jeopardized by the use of the specific tool.
Flowchart Box 1C (Is problem localized?)
Go to Box IE if problem is localized: Localized problems in this check are likely due to changes in the pipeline (diameter, wall thickness, etc.) Go to Box 1D is problem is widespread.
Flowchart Box 1D (Can a re-run fix the problem?)
“Fail” if a rerun of a more suited ILI tool would lead to a successful run. If a rerun is unlikely to be successful, then the threat must be address by other means.
Flowchart Box 1E (Can issue be addressed by other means?)
“Conditional” if alternative integrity management tools such as hydrotesting, direct assessment, etc. can adequately address the localized problems. Go to Box 1D of the areas are too numerous or too long to be addressed economically by other means.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 40 of 69
Aboutpipelines.com
Table 16: Guidance for Parameter #2 Pre-Run Inspection System Data Item # 2
Parameter Inspection system check
Stage Pre-run
API Category System Results Validation
API 1163 Reference 2
Score (F/C/P) “Fail” - Tool is experimental and there is no established history or has been demonstrated to have data gaps. “Conditional” – Tools of the same model with minor differences have a history of successful runs or tool has a history of successful runs, but data is not available to the operator. “Pass” – Tool has a history of successful runs.
Flowchart Box 1A (Pass Check?)
“Pass” if the operator has first-hand knowledge of the performance capabilities of the tool and has several successful inspections using the tool.
Flowchart Box 1B (Is the impact on data significant?)
“Conditional” if the operator has first-hand knowledge of the similar model of tool on other pipelines. The experiences may be with models of tool on different diameters. “Conditional” if the tool has been successfully run on several other pipeline systems, but for other operators and the operator has no access to the data.
Flowchart Box 1C (Is problem localized?)
Go to Box IE if problem is localized: Localized problems in this check are likely due to changes in the pipeline (diameter, wall thickness, etc.) Go to Box 1D is problem is widespread.
Flowchart Box 1D (Can a re-run fix the problem?)
“Fail” if a rerun of a more tested ILI tool would lead to a successful run. If a rerun is unlikely to be successful, then the threat must be address by other means.
Flowchart Box 1E (Can issue be addressed by other means?)
“Conditional” if alternative integrity management tools such as hydrotesting, direct assessment, etc. can adequately address the localized problems. Go to Box 1D of the areas are too numerous or too long to be addressed economically by other means.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 41 of 69
Aboutpipelines.com
Table 17: Guidance for Parameter #3 Pre-Run Planning Item # 3
Parameter Planning
Stage Pre-run
API Category In-Line System Selection
API 1163 Reference 5.3, 7.2
Score (F/C/P) “Fail” - Key elements of Pipeline ILI Compatibility Assessment and Inspection Scheduling not conducted. “Conditional” - Majority of elements of Pipeline ILI Compatibility Assessment and Inspection Scheduling completed but undocumented. “Pass” – All elements of Pipeline ILI Compatibility Assessment and Inspection Scheduling completed and documented.
Flowchart Box 1A (Pass Check?)
Use Guidance Document: SP0102-2010 - specifically sections 4, 5 and 6. “Pass” if the appropriate plan was developed and executed for the expected line conditions.
Flowchart Box 1B (Is the impact on data significant?)
“Conditional” if either the plan was deficient or if the plan was not properly executed, but without affecting the data.
Flowchart Box 1C (Is problem localized?)
Go to Box IE if problem is localized: Localized problems can be caused by event not covered by the plan. Go to Box 1D is problem is widespread.
Flowchart Box 1D (Can a re-run fix the problem?)
“Fail” if the lack of planning in one of the elements described in the RP contributed to compromising data collection. If a rerun is unlikely to be successful, then the threat must be address by other means.
Flowchart Box 1E (Can issue be addressed by other means?)
“Conditional” if alternative integrity management tools such as hydrotesting, direct assessment, etc. can adequately address the localized problems. Go to Box 1D of the areas are too numerous or too long to be addressed economically by other means.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 42 of 69
Aboutpipelines.com
Table 18: Guidance for Parameter #4 Pre-Run Function Checks Item # 4
Parameter Function Checks
Stage Pre-run
API Category System Operational Verification
API 1163 Reference 7.3.2
Score (F/C/P) “Fail” - Significant function checks not passed. “Conditional” - Significant function checks passed completed but undocumented. “Pass” - All function checks passed and documented.
Flowchart Box 1A (Pass Check?)
“Pass” if all relevant function checks passed, including but not limited to
- Adequate power supply available and operational; - Sensors and data storage operating; - Adequate data storage available; - All tool components properly initialized;
Flowchart Box 1B (Is the impact on data significant?)
“Conditional” if the checks were done but not documented or if the functional integrity of the tool can be demonstrated indirectly.
Flowchart Box 1C (Is problem localized?)
Go to Box IE if problem is localized: Localized problems are unlikely in this check. Go to Box 1D is problem is widespread.
Flowchart Box 1D (Can a re-run fix the problem?)
“Fail” if a rerun of the tool with adequate pre-run checks is likely to result in a successful run. If a rerun is unlikely to be successful, then the threat must be address by other means.
Flowchart Box 1E (Can issue be addressed by other means?)
“Conditional” if alternative integrity management tools such as hydrotesting, direct assessment, etc. can adequately address the localized problems. Go to Box 1D of the areas are too numerous or too long to be addressed economically by other means.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 43 of 69
Aboutpipelines.com
Table 19: Guidance for Parameter #5 Pre-Run Mechanical Checks Item # 5
Parameter Mechanical Checks
Stage Pre-run
API Category System Operational Verification
API 1163 Reference 7.3.3
Score (F/C/P) “Fail” - Significant mechanical checks not passed. “Conditional” - Significant mechanical checks completed but undocumented. “Pass” - All mechanical checks passed and documented.
Flowchart Box 1A (Pass Check?)
“Pass” if all relevant mechanical checks passed, including but not limited to: - Visual inspection of tool to ensure it is mechanically sound; - Ensuring electronics are sealed; - Ensuring adequate integrity of cups; - Ensuring all moving parts are functioning as expected;
Flowchart Box 1B (Is the impact on data significant?)
“Conditional” if the checks were done but not documented or if the mechanical integrity of the tool can be demonstrated indirectly.
Flowchart Box 1C (Is problem localized?)
Go to Box IE if problem is localized: Localized problems are unlikely in this check. Go to Box 1D is problem is widespread.
Flowchart Box 1D (Can a re-run fix the problem?)
“Fail” if a rerun of the tool with adequate pre-run checks is likely to result in a successful run. If a rerun is unlikely to be successful, then the threat must be address by other means.
Flowchart Box 1E (Can issue be addressed by other means?)
“Conditional” if alternative integrity management tools such as hydrotesting, direct assessment, etc. can adequately address the localized problems. Go to Box 1D of the areas are too numerous or too long to be addressed economically by other means.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 44 of 69
Aboutpipelines.com
Table 20 Guidance for Parameter #6 in the pipe Procedure Execution Item # 6
Parameter Procedure execution (e.g., pigging procedure/tool speed, etc.)
Stage In the pipe
API Category System Operational Verification
API 1163 Reference 7.4
Score (F/C/P) “Fail” - Inspection not conducted as per inspection procedure with potential material impact to data quality. “Conditional” - Inspection not carried out as per inspection procedure but deviations are not material to data quality. “Pass” - Inspection carried out as per inspection procedure.
Flowchart Box 1A (Pass Check?)
“Pass” – if all relevant checks pass, including but not limited to -Tool run was executed as per the planned pigging procedure; -Line condition (fluid composition, flow rate, temperature, pressure, etc.) was as planned; -Line conditions for tool launch as expected and the launch proceed as planned; -Line conditions for tool receive was as expected and the receive proceed as planned; -Tool speed was within the planned range for the length of the run. -Tool tracking unfold as planned; If deviations did occur, they planned or within expectations.
Flowchart Box 1B (Is the impact on data significant?)
“Conditional” if areas where deviations from the planned procedure are manageable or pose minimal risk to the pipeline.
Flowchart Box 1C (Is problem localized?)
Go to Box IE if problem is localized: Localized problems can be caused by short speed excursions. Go to Box 1D is problem is widespread.
Flowchart Box 1D (Can a re-run fix the problem?)
“Fail” if a rerun of the tool with better planning to address the problem is likely to result in a successful run. If a rerun is unlikely to be successful, then the threat must be address by other means.
Flowchart Box 1E (Can issue be addressed by other means?)
“Conditional” if alternative integrity management tools such as hydrotesting, direct assessment, etc. can adequately address the localized problems. Go to Box 1D of the areas are too numerous or too long to be addressed economically by other means.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 45 of 69
Aboutpipelines.com
Table 21: Guidance for Parameter #7 Post-Run Mechanical Checks Item # 7
Parameter Mechanical Checks
Stage Post run
API Category System Operational Verification
API 1163 Reference 7.5.2
Score (F/C/P) “Fail” - Significant tool wear, damage or debris with material impact to data. “Conditional” - Tool wear, damage or debris observed with no material impact to data. “Pass” - Tool received in good mechanical condition (no unexpected tool wear, damage or debris).
Flowchart Box 1A (Pass Check?)
“Pass” if all relevant mechanical checks passed, including but not limited to -Visual inspection of tool to ensure it is mechanically sound? -Ensuring electronics are sealed; -Ensuring adequate integrity of cups; -Ensuring all moving parts are functioning as expected; -The volume and nature of any debris present was within expectations and not detrimental to data collection.
Flowchart Box 1B (Is the impact on data significant?)
“Conditional” if the checks were done but not documented or if the mechanical integrity of the tool can be demonstrated indirectly. The demonstration of the mechanical integrity of the tool can be difficult unless it has been properly documented.
Flowchart Box 1C (Is problem localized?)
Go to Box IE if problem is localized: Localized problems are unlikely in this check, except when the damage to the tool occurred near the end of the run. Go to Box 1D is problem is widespread.
Flowchart Box 1D (Can a re-run fix the problem?)
“Fail” if a rerun of the tool after addressing the cause of the problem is likely to result in a successful run. If a rerun is unlikely to be successful, then the threat must be address by other means.
Flowchart Box 1E (Can issue be addressed by other means?)
“Conditional” if alternative integrity management tools such as hydrotesting, direct assessment, etc. can adequately address the localized problems. Go to Box 1D of the areas are too numerous or too long to be addressed economically by other means.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 46 of 69
Aboutpipelines.com
Table 22: Guidance for Parameter #8 Post-Run Function Check Item # 8
Parameter Function Check
Stage Post run
API Category System Operational Verification
API 1163 Reference 7.5.2
Score (F/C/P) “Fail” - Significant function checks not passed. “Conditional” - Significant function checks passed but undocumented. “Pass” - Function checks passed and documented.
Flowchart Box 1A (Pass Check?)
“Pass” if all relevant function checks passed, including but not limited to -Adequate power supply available and operational; -Sensors and data storage operating; -Adequate data storage available; -All tool components functioning as expected.
Flowchart Box 1B (Is the impact on data significant?)
“Conditional” if the checks were done but not documented or if the functional integrity of the tool can be demonstrated indirectly. The demonstration of the functional integrity of the tool can be difficult unless it has been properly documented.
Flowchart Box 1C (Is problem localized?)
Go to Box IE if problem is localized: Localized problems are unlikely in this check, except when the functional failure of the tool occurred near the end of the run. Go to Box 1D is problem is widespread.
Flowchart Box 1D (Can a re-run fix the problem?)
“Fail” if a rerun of the tool with better planning to address the problem is likely to result in a successful run. If a rerun is unlikely to be successful, then the threat must be address by other means.
Flowchart Box 1E (Can issue be addressed by other means?)
“Conditional” if alternative integrity management tools such as hydrotesting, direct assessment, etc. can adequately address the localized problems. Go to Box 1D of the areas are too numerous or too long to be addressed economically by other means.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 47 of 69
Aboutpipelines.com
Table 23: Guidance for Parameter #9 Post-Run Field Data Check Item # 9
Parameter Field Data Check
Stage Post run
API Category System Operational Verification
API 1163 Reference 7.5.3
Score (F/C/P) “Fail” - Tool is unable to meet stated specifications due to significant lack of data integrity. “Conditional” - Tool is unable to meet stated specifications but manageable through further analysis. “Pass” - Tool is able to meet stated specification for entire length of run.
Flowchart Box 1A (Pass Check?)
“Pass” if the data collection met all the basic quality and quantity checks, including but not limited to -Confirmation of continuous data stream for full circumference of pipe; -Basic quality requirements have been met; -The “amount” of data captured in line with expectations;
Flowchart Box 1B (Is the impact on data significant?)
“Conditional” if the checks were done but not documented or if the full data collection by the tool can be demonstrated indirectly.
Flowchart Box 1C (Is problem localized?)
Go to Box IE if problem is localized: Localized problems can be due to short losses of data. Go to Box 1D is problem is widespread.
Flowchart Box 1D (Can a re-run fix the problem?)
“Fail” if a rerun of the tool with better planning to address the problem is likely to result in a successful run. If a rerun is unlikely to be successful, then the threat must be address by other means.
Flowchart Box 1E (Can issue be addressed by other means?)
“Conditional” if alternative integrity management tools such as hydrotesting, direct assessment, etc. can adequately address the localized problems. Go to Box 1D of the areas are too numerous or too long to be addressed economically by other means.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 48 of 69
Aboutpipelines.com
Table 24: Guidance for Parameter #10 Post-Run Data Analysis Processes and
Quality Checks Item # 10
Parameter Data analysis processes: quality checks
Stage Post-run
API Category System Results Validation
API 1163 Reference 8.2.2, Annex C
Score (F/C/P) “Fail” - Results of data analysis quality checks are not acceptable. “Conditional” - Significant data quality checks passed but procedure is undocumented. “Pass” - Data analysis procedures followed; data quality checks passed and documented; the number and severity of anomalies meets expectations.
Flowchart Box 1A (Pass Check?)
“Pass” if data analysis checks are met including but not limited to -Continuous recording of data was for the full pipe circumference; -Sensor response was within expected range(s); -Data analysis processes were executed as per pre-defined procedures; -Analysis was conducted by persons with qualification as agreed; -Automated detection and sizing parameters were used as agreed; -Manual intervention by data analysts was conducted as agreed; -Burst pressure calculations where conducted as agreed; -Correct pipeline parameters (pipe diameter, wall thickness, manufacturer, and grade) were documented and used to undertake the analysis; -The number and type of anomalies reported are consistent with expectations.
Flowchart Box 1B (Is the impact on data significant?)
“Conditional” if the analysis of the data deviated from the planned procedure but that the impact on the data is deemed minimal.
Flowchart Box 1C (Is problem localized?)
Go to Box IE if problem is localized: Localized problems are unlikely in this check or can be corrected by reanalysis of the affected areas. Go to Box 1D is problem is widespread.
Flowchart Box 1D (Can a re-run fix the problem?)
“Fail” if a rerun of the tool is likely to result in a successful run. However, rerunning the tool is unlikely to address problems in the analysis of the data. If a rerun is unlikely to be successful, then the threat must be address by other means.
Flowchart Box 1E (Can issue be addressed by other means?)
“Conditional” if alternative integrity management tools such as hydrotesting, direct assessment, etc. can adequately address the localized problems. Go to Box 1D of the areas are too numerous or too long to be addressed economically by other means.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 49 of 69
Aboutpipelines.com
Table 25: Guidance for Parameter #11 Post-Run Cumulative Assessment Item # 11
Parameter Cumulative Assessment
Stage Post-run
Score (F/A/P) “Fail” - Cumulative impact of "Conditional" passes deemed to materially impact ILI results A - Cumulative impact of "Conditional" passes is unclear and may have impacted ILI results “Pass” - The cumulative impact of all "Conditional" passes, if any, deemed to be tolerable
Flowchart Box 1A (Pass Check?)
“Fail” if the number and nature of any Conditional passes provide a cause for concern on a cumulative basis. Relevant considerations include, but are not limited to -Can data gaps be mitigated effectively using alternative methods? -Can any data gaps actually be addressed through re-running the tool or are line conditions such that similar challenges will remain? -Are any “Conditional” pass issues cumulative in nature? Issues would be considered to be cumulative if one issue magnifies any pre-existing data degradation (such as a tool overspeed in a location where the tool performance is already compromised as a result of debris related sensor lift off). Conversely, issues would not be considered cumulative if their impact on data quality is largely independent (such as a run where the tool overspeeds at launch and experiences debris issues at receive).
Additional Comments:
F = Failing assessment of a parameter that cannot be mitigated
C = Conditional passing score that may be investigated, mitigated, documented, and
deemed tolerable
P = Passing score
A = Ambiguous
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 50 of 69
Aboutpipelines.com
A1.1. Verification Examples To illustrate the verification process, this section presents two examples.
A1.1.1. Example 1 The first example is summarized below:
• 100 km NPS 36 run.
• 10 metre over speed at launch.
• Intermittent loss of one sensor (km 1.9 to km 2.0).
• No record of pre-run mechanical and function tests.
• No actionable anomalies; no previous inline inspection.
The completed scorecard is shown in Table 26.
Table 26: Completed Scorecard Example 1 Stage Parameter Score Comment
1 Pre-run Tool selection Pass
2 Pre-run Inspection system Pass
3 Pre-run Planning Pass
4 Pre-run Function check Conditional Pass
-No record -Obtained verbal confirmation -Post run function check passed
5 Pre-run Mechanical check Conditional Pass
-No record -Obtained verbal confirmation -Post run function check passed
6 In pipe Procedure execution
Conditional Pass
-Short over speed (due to line conditions) at launch deemed tolerable since pipe is in station yard and above ground
7 Post- run Mechanical check Pass
8 Post-run Function check Pass
9 Post-run Field data check Pass
10 Post-run Data analysis processes
Conditional Pass
-100 m of data missing (single sensor) -Deepest anomaly non-injurious based on revised tool performance specification
11 Post-run Cumulative assessment
Pass -All “Conditional” scores mitigated -No material cumulative impacts identified
In this example, “Conditional” scores were given for the Pre-run
Function check, Pre-run Mechanical check, Procedure Execution
check and the Post-run Data analysis check.
Although the pre-run checks were undocumented, the post-run
checks were documented, and the tool passed both checks.
Although, the pre-run checks should have been done, the
operator accepted that the condition of the tool as good prior to
the run based on the condition of the tool after the run. The
operator issued a letter to the ILI vendor instructing it to supply
documentation of the pre-run checks on all future runs.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 51 of 69
Aboutpipelines.com
During the running of the tool, a short speed excursion occurred
at the start of the run. The excursion was believed to affect sizing
capabilities of the tool, but would have minimal effect on the
detection capabilities of the tool. Since no metal-loss anomalies
were detected in the area of the speed excursion, the effect on
the data was deemed to be not significant.
In the data analysis process, it was discovered that one sensor
was lost for 100 metres. The loss of one sensor was believed to
affect detection and sizing capabilities of the tool only for metal-
loss anomalies along that sensor’s track. Only very sparse
shallow anomalies were detected in the section of the sensor
loss, and none of the anomalies were in the vicinity of the lost
sensor. The effect on the data was deemed to be not significant.
The final Cumulative Assessment examined the “Conditional”
scores. Since the area of speed excursion and the sensor loss
were along different parts of the pipeline, all aspects of the run
were considered acceptable and none of the issues were
cumulative, the ILI run was accepted.
A1.1.2. Example 2 The second example is:
• 100 km NPS 24 run 1950 asphalt coated line.
• Loss of one sensor bank for last 1 km of inspection.
• All pre-run and post-run tests passed and documented.
• 20 metal loss anomalies reported by ILI tool.
• No actionable anomalies identified; no previous inline
inspection.
The completed scorecard is shown in Table 27.
Table 27: Completed Scorecard Example 2
Stage Parameter Score Comment 1 Pre run Tool selection Pass
2 Pre run Inspection system Conditional -Operator has experience with the 16-inch and 36-inch model of this tool. -The 24-inch model of this tool has been successfully run for several other operating companies, but results of those runs are not available.
3 Pre run Planning Pass
4 Pre run Function check Pass
5 Pre run Mechanical check Pass
6 In pipe Procedure execution Pass
7 Post run Mechanical check Pass
8 Post run Function check Pass
9 Post run Field data check Conditional -Loss of sensor bank for last two km of inspection -Deemed tolerable since section under hydrotest
10 Post run Data analysis processes Conditional -Independent audit identified incorrect threshold set during data analysis process -Data re-analyzed and report re-issued (10,000 metal features identified)
11 Post run Cumulative assessment Pass -All “Conditional” scores mitigated -No material cumulative impacts identified
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 52 of 69
Aboutpipelines.com
In Example 2, “Conditional” scores were given for the Inspection
System check, the Post-run Field data check, and the Data
Analysis process check.
Although the operator had no experience with the 24-inch model
of this particular tool, it did have experience with other diameter
models of the same tool (same vendor but different diameter
tools). The operator’s experience with the tools and the vendor
was very good. In addition, the operator received references
from other operators that the 24-inch model of the tool had also
been successfully run for other pipeline systems. The operator, in
this case, deemed that the ILI tool had a sufficient history to be
accepted.
The Field data check revealed that one sensor bank (24 sensors)
failed for two kilometres of the inspection. This loss was deemed
to have a significant impact on the data (node 1B of the
Verification Check-point Flowchart). The deficiency was localized
(node 1C) and was addressed by a hydrostatic test in the
previous year (node 1E). Therefore, despite that the effect on the
data was significant, the integrity concern in the area of the
issues can be addressed by other means. The item was given a
“Conditional” pass.
In the review of the ILI results, it was found that there were no
reported metal-loss anomalies with a depth less than 10% NWT.
Discussions with the ILI vendor’s Manager of Data Analysis, it
was determined that all metal-loss anomalies with a depth less
than 10% NWT were filtered out in error. The operator requested
all anomalies to be included in the report. The vendor added the
shallow anomalies and reissued the report.
The final Cumulative Assessment examined the “Conditional”
scores. Since none of the issues identified in the verification
process were cumulative, the ILI run was accepted.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 53 of 69
aboutpipelines.com
A2. NACE Table
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 54 of 69
Aboutpipelines.com
Figure 4: Table 1 from NACE SP 0102-2010 giving Guidance on Tool
Selection for ILI
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 55 of 69
aboutpipelines.com
A3. Matching
A3.1. Process overview When validating an ILI inspection using a previous inspection, there are two
ILI run datasets to be compared. Defect matching is done so that defects
from the first run can be compared to the corresponding defect in the
second run. This process involves matching up the girth weld sections,
adjusting the odometer and orientation, and matching the identified
anomalies.
A3.2. Girth Weld Matching Girth welds are very easy to detect with current ILI tools. This means that
the girth weld sections can be matched up from the two ILI runs based on
the length of the sections. Once girth weld sections are matched, the defects
on specific girth weld sections can be matched between runs.
A3.3. Matching of identified Anomalies Once the girth weld sections are matched between the runs, the defects are
matched based on chainage, orientation and depth. Location (as in chainage
and orientation) and size are prioritized over identification (whether the
anomaly is identified as being internal/external on the pipe surface). This
means that external anomalies can be matched to internal anomalies based
on chainage and orientation. This is done because the defect location on the
pipe surface is less certain than the reported chainage and orientation
values.
A3.4. Calculating Anomaly Depth Change With the defects matched between the two runs, it is possible to calculate
the apparent difference in depth of each anomaly. With this information, the
systematic bias and the standard deviation of the difference can be
calculated.
Suppose that we have 𝑛 matched anomalies with depths of 𝑑1, 𝑑2, 𝑑3, … 𝑑𝑛 , as
reported by the current ILI. The depths of the corresponding anomalies in the reference ILI run are 𝑑𝑟1, 𝑑𝑟2, 𝑑𝑟3 … 𝑑𝑟𝑛. The apparent difference in depth
for each anomaly is
Δ𝑖 = 𝑑𝑖 − 𝑑𝑟𝑖
The average difference is
Δ̅ =1
𝑛∑ Δ𝑖
𝑛
𝑖=1
The standard deviation of the difference depth is
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 56 of 69
Aboutpipelines.com
𝑠 = √1
𝑛 − 1 ∑(Δ𝑖 − Δ̅)2
𝑛
𝑖=1
A4. Validation using a Previous ILI The validation Process is a check of the results of the ILI data to ensure that it meets
a performance specification. If there is a previous ILI available, then the accuracy of
the ILI depth measurements can be assessed by comparing the depths with the
previous ILI. This Appendix provides the theory and examples of that process.
Although the examples are based on a comparison of the ILI with a previous ILI run,
the theory for comparing the ILI with the results of in-the-ditch depth measurements
is identical.
A4.1. Demonstration of Concept ILI accuracy is typically given in terms of a tolerance and a confidence level.
ILI accuracy is often stated as being “±10% NWT, 80% of the time.” That
accuracy implies that the ILI reported depth values may differ from the true
depth value, but they are clustered in the vicinity about the true value. To
Validate the ILI run, we compare the reported depth values to a set of
reference depth measurements, which may be another ILI. Like the ILI
measurements, the reference measurements have some measurement
errors, and we assume that the reference measurements are clustered in the
vicinity about the true depth values. If the current ILI measurements are
clustered around the reference measurements, then we can conclude that
the ILI measurements are also clustered about the true depth values.
Figure illustrates the comparison of the validation process.
Figure 5: Illustration of the Validation Process
Assumed tolerance, 𝑇𝑟, of
reference measurements
(e.g., ±𝑇𝑟, 80% of time)
Inferred ILI
accuracy
tolerance, ±𝑇
Calculated
difference, ±𝑇Δ,
between ILI and
reference
measurements,
Reference depth
measurements
ILI depth
measurements
True
defect depth
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 57 of 69
Aboutpipelines.com
The results of an ILI run are compared to the reference measurements. The
difference between the ILI depths and the reference depths puts an upper
bound on the tolerance, 𝑇, of the ILI tool accuracy.
We assume that we have some reference measurements of the depth of the
metal-loss anomalies from some independent source. Ideally, the reference
measurement should have as little error as possible; however, in practice,
they will have some measurement error associated with them. The measurement error of the reference measurements is ±𝑇𝑟, (𝐶 (%) of the
time). 𝑇𝑟 is the tolerance of the reference measurements and 𝐶𝑟 is the
confidence level or certainty for the tolerance.
If the reference measurements are made in the ditch with a highly accurate
device (such as a laser scanning device), the tolerance should be very small: 𝑇𝑟 ≈ 0. However, if the reference measurements are from a previous ILI,
then the tolerance is likely to be 10% of NWT (𝑇𝑟 = 0.10) and the confidence
level is 80% (𝐶𝑟 = 0.80).
The amount of error, as measured by the standard deviation of the errors, in
a set of measurements is conceptually the “distance” between it and the true
value. The validation process sets an upper bound on the distance between
the ILI measurement and the true depths by estimating the distance
between the ILI measurements and the reference measurement.
A4.2. ILI Error ILI depth measurements are a sum of the true depth plus some error.
𝑑𝐼𝐿𝐼 = 𝑑𝑡 ± 𝐸
The variable ±𝐸 is a random variable with some mean and standard
deviation (the “±” symbol is meant to indicate that the variable is random.
We can separate the error term into a random component and a systematic
bias:
𝑑𝐼𝐿𝐼 = 𝑑𝑡 ± 𝐸𝑟 + 𝑠
Where 𝑑𝐼𝐿𝐼 is the depth measured by the ILI tool;
𝑑𝑡 is the true depth of the anomaly;
𝐸𝑟 is the random error of the ILI tool; and
𝑠 is the systematic (constant) bias.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 58 of 69
Aboutpipelines.com
Figure graphically shows the relationships between random and systematic
components of the error. Both the random and systematic components
contribute to the total ILI error.
Figure 6: Random Error and Bias Component
Contribute to the Error of any ILI Measurement
The random error is different for each metal-loss anomaly and causes the
readings to be scattered about the true value of the depth. Random error
has a mean of zero.
ILI accuracy is typically given in terms of a tolerance and a confidence level.
The specified ILI accuracy of “±10% NWT, 80% of the time” implies that the
standard deviation of ILI error is 7.8%: The 80% confidence level translates to the 𝑧-value of 1.28 for a normal distribution. We then calculate the
standard deviation of the error:
𝜎𝑒𝑟𝑟𝑜𝑟 =𝑇
𝑧
𝜎𝑒𝑟𝑟𝑜𝑟 =10%
1.28= 7.8%
A4.3. Comparison with Reference Measurements Since true depth is unavailable, we compare the ILI measured depth with
some reference depth. Assuming corrosion growth is small compared to the
measurement error, a comparison of the ILI to reference depth can be used
to determine the size of the errors. 𝑑1 = 𝑑𝑡 ± 𝐸𝑟1 + 𝑠1
-30% -20% -10% 0% 10% 20% 30%
Freq
uenc
y
Total ILI error
Bias component
Random component
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 59 of 69
Aboutpipelines.com
𝑑2 = 𝑑𝑡 ± 𝐸𝑟2 + 𝑠2
𝑑 = 𝑑2 − 𝑑1 = (±𝐸𝑟1 ± 𝐸𝑟2) + (𝑠2 − 𝑠1)
Note that the plus-minus sign ± indicates a random variable (with mean
zero). If the standard deviation of the random error components, 𝐸𝑟1 and 𝐸𝑟2
are 𝜎𝑟1 and 𝜎𝑟2, then the standard deviation of the sum 𝐸𝑟1 + 𝐸𝑟2 is
±√σ𝑟22 + σ𝑟1
2 . Thus the mean of the Δ𝑑’s is an estimate of the relative
systematic bias as (𝑠2 − 𝑠1), and the standard deviation of the Δ𝑑’s is an
estimate of the random component of the error.
If the reference measurements are from a previous ILI, with an accuracy of ±10% NWT, 80% of the time, then 𝑇𝑟 = 0.10 and 𝐶𝑟 = .80. The standard
deviation, 𝜎𝑟, of the error of the previous ILI is
𝜎𝑟 =𝑇𝑟
𝑧80
=. 10
1.28155= 7.8%
If the specified accuracy of the current ILI is ±10% NWT, 80% of the time, then 𝑇 = 0.10 and 𝐶 = .80. Then we want to demonstrate that the standard
deviation of the current ILI is 𝜎𝐼𝐿𝐼 = 0.078 (7.8%). If the reference
measurements are an ILI, then the standard deviation of the differences, 𝜎Δ,
is given by
𝜎Δ = √𝜎𝑟2 + 𝜎𝐼𝐿𝐼
2 = √(0.078)2 + (0.078)2 = 0.11
Thus we should expect that the standard deviation of the differences
between the current ILI depth measurements and the reference ILI
measurements should be about 0.11 or 11%.
A4.4. Acceptance Criteria The acceptance criterion is the requirements for the acceptance of the ILI
data.
Where 𝐸𝑟1 is the random error component for the previous ILI
measured depth.
𝐸𝑟2 is the random error component for the current ILI
measured depth.
𝑠1 is the systematic bias of the previous ILI measured
depth.
𝑠2 is the systematic bias of the current ILI measured
depth.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 60 of 69
Aboutpipelines.com
A4.4.1. Systematic Bias Criterion This relative bias between two ILI runs does not indicate anything
specifically about either of the ILI runs, but a large relative bias
could indicate that one of the ILI runs does not meet its stated
specification.
To establish how much bias is significant and how one might
adjust for it, consider the following: Assume that the random
component of the error has a normal distribution, then the average of the Δ𝑑’s is the best estimate of the systematic bias, 𝜀𝑠.
Note, however, that the average is only an estimate of the bias:
𝜀𝑠 = Δ𝑑̅̅̅̅ ± 1.96𝜎Δ
√𝑛
Where 1.96𝜎Δ
√𝑛 is the 95% confidence bound of the estimate.
Table 28 outlines the use and significance of the calculated
relative bias.
Table 28: Considerations when Dealing with Systematic Bias Item Description Considerations
1 Usage
Systematic bias can be readily considered when validating an inspection by comparing it to other data sets.
Caution should be used in adjusting for bias when selecting excavations – especially if adjustments result in lower depths.
2 Size of bias Systematic bias may be considered to be significant based on some constant threshold (say 5 - 6% of NWT) and could be associated with tool detection threshold(s).
Alternatively, bias may be viewed in the context of defect depth. That is, if the deepest defect is 10%, a 5% bias may be significant; however, if the deepest defect is 40%, a 5% bias may be less material.
3 Statistical significance
Systematic bias may be considered to be statistically significant based on size of the confidence interval
±1.96σΔd
√n. If Δd̅̅̅̅ > 1.96
σΔd
√n , then the calculated bias is
statistically significant to a 95% confidence level. A calculated bias that is not statistically significant should not be used to adjust the ILI reported depths.
A4.4.2. Random Error Criterion Given an ILI run, we can use the reference measurements to
validate the inspection results. We would like to prove that tolerance, 𝑇, meets the tool specification for the ILI that in
the examples is ±10% NWT, 80% of the time. However, we
cannot statistically prove that the tolerance is exactly 10%;
the best we can do is estimate the tolerance. If the target
tolerance is within the 95% confidence bounds of the
estimate, then we have validated the ILI accuracy.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 61 of 69
Aboutpipelines.com
A4.4.3. Target Tolerance Sample size is the number of reference measurements
needed to statistically validate the accuracy of the ILI. The
minimum sample size depends on the size of the desired
confidence interval. Since the specification is stated as ±10%
and not ±10.0%, we assume that only the first digit is
significant. Thus we want to calculate an estimated tolerance, 𝑇∗, such that the true tolerance is within 10% of the
estimate. That is
𝑇∗ = 𝑇 ± 10% = 𝑇 ± (0.10𝑇) = 10% NWT ± 1% NWT
Since we are only concerned with the one-sided interval, only
the upper bound for the standard deviation of the error is
relevant:
∗ ≤ 11%.
The estimated standard deviation of the ILI error is
𝜎𝐼𝐿𝐼∗ =
𝑇∗
𝑧80
=11%
1.28155= 8.6%
However in practice, we have 𝑇Δ (the comparison of the ILI
with the reference measurements), not 𝑇. So we determine
the upper bounds for the standard deviation of the difference of the reported depths, 𝜎Δ. Assuming that the errors in the
ILI measurements are independent of the errors in the
reference measurements, we calculate the standard deviation
of the difference in the depth values:
𝜎Δ = √𝜎𝐼𝐿𝐼2 + 𝜎𝑟
2
If 𝜎𝑟 = 0.078 and 𝜎𝐼𝐿𝐼∗ = .085, then the upper bound estimated
value for the standard deviation between the reference and
ILI depth values is
𝜎Δ∗ = √(𝜎𝐼𝐿𝐼
∗ )2 + 𝜎𝑟2 = √(0.0858)2 + (0.078)2 = .116 = 11.6%
A4.4.4. Minimum Sample Size The minimum sample size depends on the number of
available matches and the number of anomalies reported in
the ILI inspection. If the number of available matches is
large, then the minimum sample size is calculated by the
Large-Population Case below. If the number of available
matches is less than the minimum sample size for the large
population case, then the Section for the Small-Population
Case determines if the available matches is sufficient.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 62 of 69
Aboutpipelines.com
A4.4.4.1. Large Population Case
In the Section IV.4.3, we assume that the
estimate of the tolerance is within ±1% NWT of
the true tolerance: that is the estimated
tolerance is known within a ±1% NWT
confidence interval with a 95% confidence level.
To satisfy that requirement, we need a sample
size of some minimum size.
We have an upper bound for estimate standard deviation, 𝜎Δ
∗ =0.116.
The best estimate of the standard deviation of
the differences between the ILI depth and the
reference depths is
𝑠Δ = √1
𝑛 − 1∑(Δ𝑖 − Δ̅)2
𝑛
𝑖=1
Thus we want
𝑠Δ ≤ 0.116
The Chi-Squared distribution is used for
calculating the confidence interval of an
estimated standard deviation. For a given calculated value of 𝑠 the confidence interval of
the standard deviation is given by
(𝑁 − 1)𝑠Δ2
𝑈;𝑛−12
≤ 𝜎2 ≤ (𝑁 − 1)𝑠Δ
2
𝐿;𝑛−12
Where 𝑁 is sample size;
𝑠Δ2 is the variance of the sample Δ’s;
𝜎2 is variance of the differences in depth if the ILI meets its performance specification 𝜎Δ
2 = (11.04%)2 =
(√7.8%2 + 7.8%2)2; and
𝑈;𝑁−12 and
𝐿;𝑁−12
are the upper and lower 95% confidence bounds of the 𝜒2 distribution with (𝑁 − 1) degrees of freedom.
By rearranging the above inequality and
considering the upper bound limit, we get
(𝑁 − 1)
𝑈;𝑛−12
≤ 𝜎2
𝑠Δ2
Or
𝑠Δ2 ≤
𝜎2
(𝑁 − 1)
𝑈;𝑛−12
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 63 of 69
Aboutpipelines.com
If we are comfortable accepting a value for the sample standard deviation, 𝑠, that is ± 10% of
the true standard deviation, 𝜎, and a one-sided
confidence level of 95% (where = 0.10) then
we can solve for 𝑁,
√(𝑁 − 1)
𝑈;𝑁−12
= √11.04
11.60= 0.97.
Solving for the above (iteratively), yields
𝑁 = 513.
So in practice, from a random sample of 513
matches, if the calculated standard deviation of the difference in depth, Δ𝑑, is 11.6% or less,
then we can conclude a Specified accuracy of the
ILI tool of ±10% NWT, 80% of the time is within
the 95% confidence interval of the assessed
tolerance.
A4.4.4.2. Small-Population Case
This section provides guidance if the number of
available matches is less than the minimum
sample size, as calculated in the Large-
population case. Suppose the minimum sample size for the Large-population case is 𝑁. There
are four cases to consider, as shown in Table 29.
Table 29: Small Population Case Samples
Case Population of reported anomalies in the previous reference ILI
Population of reported anomalies in the current ILI
Number of available matches
1 Small (< 1.3𝑁) Small (< 1.3𝑁) Small (< 𝑁)
2 Small (< 1.3𝑁) Large (> 1.3𝑁) Small (< 𝑁)
3 Large (> 1.3𝑁) Small (< 1.3𝑁) Small (< 𝑁)
4 Large (> 1.3𝑁) Large (> 1.3𝑁) Small (< 𝑁)
Case 1: If the number of reported anomalies in
both inspections is small, then the number of
available matches is the population of matches
and those matches are sufficient for the
validation process.
Case 2: If the number of reported anomalies in
the first inspection is small but in the second
inspection is larger, then the number of available
matches is the population of matches and those
matches are sufficient for the validation process.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 64 of 69
Aboutpipelines.com
Case 3: If the number of anomalies in the
previous reference inspection is larger than
current inspection, then some care is required.
The operator must ensure that the current
inspection did not miss a large number of
anomalies. (The operator should revisit the data
analysis verification check to ensure that the
proper procedures were followed in the
preparation of the final report.) The number of
matches is sufficient for the validation process if
when the reference inspection is filtered to
include only anomalies deeper than 20% NWT, a
minimum of 75% of the deeper anomalies in the
reference inspection are matched to the current
inspection.
Case 4: If the number of anomalies in both the
previous reference inspection and the current
inspection are large but only a small number of
matches can be made, then the operator must
ensure that the current inspection did not miss a
large number of anomalies. (The operator should
revisit the Data Analysis Verification check to
ensure that the proper procedures were followed
in the preparation of the final report.) The
number of matches is sufficient for the validation
process if when the reference inspection is
filtered to include only anomalies deeper than
20% NWT, a minimum of 75% of the deeper
anomalies in the reference inspection are
matched to the current inspection.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 65 of 69
Aboutpipelines.com
A5. Opportunities for Future Refinement Table 30 contains a summary of the key considerations for future development of this
procedure.
Table 30: Key Items to Consider for Future Refinement Item No.
Key Consideration Recommendation
1 Standardization of ILI Reporting
Reporting of the activities prior, during, and after the inspection should be standardized. Standardization of the reporting of these activities would result in greater consistency of the verification Process.
2 Documentation of Procedures
Documentation of all checks should be required. Proper documentation is also indicative of the ILI vendor’s diligence in following established Standards and Guidelines.
3 Technology Specific Verification
Separate versions of the scorecard should be developed for MFL and UT inspection tools. Also, separate versions of the scorecard should be developed for liquid and gas pipelines.
4 Refinement of Scorecard ILI vendors and operators should provide standardized reporting and the scorecard should be refined to yield a numeric 0-10 score.
A5.1. Standardization of ILI Reporting The purpose of this Guidance Document is to provide a standard procedure
by which operators may verify and validate ILI runs. The procedure is
intended to be independent of any specific operator, ILI vendor, or ILI
technology. Creating a standard procedure has been hampered by the lack
of standardized reporting by the ILI vendors.
Table 31 shows the variability in the data provided to CEPA for various lines.
The documentation of pre-run cleaning, for example was inconsistent. In
some cases it was documented in the report (as indicated by “yes” in the
table). Sometimes pre-run cleaning was not documented (as indicated by
“no” in the table). Sometimes pre-run cleaning was documented by it saying
it may not have been applicable (as indicated by “Not/App?” in the table).
The significance of the differences is not readily obvious.
An area for future consideration is the potential for reporting of the activities
prior, during, and after the inspection to be standardized. Standardization of
the reporting of these activities would result in greater consistency of the
verification process.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 66 of 69
aboutpipelines.com
Table 31: Summary of Data Provided for Scorecard Line Tool/System
Selection Tool
Setup Speed Debris Funct
-ion Check
Data Check
Excavation Data
Sensors Tool Magnet-ization
Pre-Run Cleaning
Tool Temp.
Company
1-100 no yes yes yes yes no yes yes yes ILI Vendor A
1-100 no yes no yes yes yes no no no ILI Vendor B
1-100 Not/App? yes yes yes yes no yes Not/App? yes ILI Vendor A
1-301 Not/App? yes yes yes yes yes yes Not/App? yes ILI Vendor A
1-301 no yes yes yes yes yes yes yes yes ILI Vendor A
1-350 no yes yes yes no no yes yes yes ILI Vendor A
227 Not/App? yes yes yes yes no yes Not/App? yes ILI Vendor A
227 no yes no yes yes no no no no ILI Vendor C
2-301 Not/App? yes yes yes yes no yes Not/App? yes ILI Vendor A
100 Line
no yes no yes no no no no no ILI Vendor B
100 Line
no no no yes yes no no no no ILI Vendor B
100 Line
no yes no yes yes no no no no ILI Vendor B
350 no yes yes yes no no yes yes yes ILI Vendor A
Note: “Yes” indicates the parameter was mentioned in the report whereas “No” means there was no mention of the
parameter.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Nov. 2014 © 2014 Canadian Energy Pipeline Association Page 67 of 69
aboutpipelines.com
A5.2. Documentation of Procedures Throughout the verification process, a score of “Pass” requires the
documentation of certain activities either by the operator or the ILI vendor
or both. API 1163 requires that procedures such as the pre-run function
check must be documented. However, none of the ILI reports supplied to
CEPA for this project included any documentation of the function checks.
To make this document usable with the current practice, documentation is
beneficial, but not required. In most of the verification checks, a “Pass”
requires documentation, but a “Conditional” score is possible if the operator
can demonstrate that the checks were done by indirect methods.
Any lack of documentation weakens the verification and validation process.
Without proper documentation of each step, the process is less likely to
withstand an internal or regulatory audit.
An area for development in future versions of this Guidance Document would
be to, in discussions with operators and vendors, include documentation as
part of the standard ILI report. Once full documentation of all processes is
the norm, then the verification and validation processes could be made more
rigorous by requiring documentation.
A5.3. Refinement of Scorecard The initial intention for the scorecard was that it would have less granular
scores than simply “Pass”, “Fail”, or “Ambiguous”. However, as mentioned
above, this Guidance Document is also intended to be independent of any
specific operator, ILI vendor, or ILI technology. As a result, the scorecard
needed to rely on information that was commonly reported by all vendors
and operators. The requirement that the verification and validation
procedure be applicable to the widest possible situations resulted in the
“Pass” or “Fail” scoring.
The verification and validation procedure would benefit from a finer grained
scoring. Such finer grained scoring would enable the procedure to measure
incremental improvement in the acquisition of ILI data and may enable
comparisons of vendors.
For future consideration: If the ILI vendors and operators can provide
standardized reporting, the scorecard could be refined to yield a numeric
0-10 score.
A5.4. Technology Specific Verification The purpose of this document is to provide a standard procedure by which
operators may verify and validate ILI runs. The procedure is intended to be
independent of any specific operator, ILI vendor, or ILI technology. The
broad scope of this document has forced it to be generic in many areas and
precluded it to contain specifics related to a MFL versus UT technology, for
example.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Sept. 2014 © 2014 Canadian Energy Pipeline Association Page 68 of 69
aboutpipelines.com
In many cases the verification checks need to rely on the ILI vendor to
provide what procedures and checks are required for a specific tool. This
document was unable to include those specifics because they could not be
equally applicable to all vendors and all technologies.
The verification and validation procedure would benefit from a greater
degree of specificity for common technologies. In particular, the process
would benefit from a scorecard that was specific to MFL inspections rather
than generic for all metal-loss inspections. Similarly, a scorecard that is
specific to gas lines and another that is specific to liquid lines would allow a
better verification and validation process.
An area for future consideration would be to develop separate versions of
the scorecard for MFL and UT inspection tools. Also, separate versions of the
scorecard could be developed for liquid and gas pipelines.
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Sept. 2014 © 2014 Canadian Energy Pipeline Association Page 69 of 69
aboutpipelines.com
A6. Scoring – Verification Process Scorecard Summary
Table 33: Verification Process Scorecard Summary
Item Parameter Stage Score
1 Tool Selection Pre-Run F - Tool not capable of detection or sizing of expected anomaly type(s) C - Tool capable of detecting anomaly types but limited sizing or detection abilities of expected anomaly type(s) P - Best available technology for detecting and sizing expected anomaly type(s) identified and used
2 Inspection System
Data
Pre-Run F - Tool is experimental and there is no established history or it has been demonstrated to have deficiencies in addressing the threat C - Same model of tool with minor differences (such as diameter) has a history of successful runs to assess the threat, or the specific model of tool has history of successful runs to assess the threat for other operators, but results of those runs are not available P - Operator firsthand knowledge of the performance capabilities of the tool and has several successful inspections using the tool
3 Planning Pre-Run F - Key elements of Pipeline ILI Compatibility Assessment and Inspection planning not conducted C - Majority of elements of Pipeline ILI Compatibility Assessment and Inspection Scheduling completed but undocumented P - Elements of Pipeline ILI Compatibility and Inspection Scheduling completed and documented
4 Function Checks Pre-Run F - Significant function checks not passed C - Significant function checks passed but checks are undocumented
P - All function checks passed and documented
5 Mechanical Checks Pre-Run F - Significant mechanical checks not passed C - Significant mechanical checks passed but checks are undocumented P - All mechanical checks passed and documented
6 Procedure
Execution (e.g. tool
speed, pigging
procedure, and
etc.)
In the Pipe F - Inspection not carried out as per inspection procedure with potential material impact to data quality C - Inspection not carried out as per inspection procedure but deviations are not material to data quality
P - Inspection carried out as per inspection procedure
7 Mechanical Checks Post-Run F - Significant tool wear, damage or debris with
CEPA Metal Loss Inline Inspection Tool Validation Guidance Document, 1st Edition, Sept. 2014 © 2014 Canadian Energy Pipeline Association Page 70 of 69
aboutpipelines.com
Item Parameter Stage Score material impact to data C - Tool wear, damage or debris observed with no material impact to data P - Tool received in good mechanical condition (with no unexpected tool wear, damage or debris)
8 Function Check Post-Run F - Significant function checks not passed
C - Significant function checks passed were not documented, but that the proper functioning of the tool can be Verified by other means throughout the length of the run P - Function checks passed and documented
9 Field Data Quality
Check
Post-Run F - Tool unable to meet stated specifications due to significant lack of data integrity C - Tool unable to meet stated specifications but manageable through further analysis
P - Tool able to meet stated specifications for entire length of run
10 Data Analysis
Process – Quality
Check
Post-Run F - Results of data analysis quality checks are not acceptable A - Significant data quality checks passed, but quality checks initially undocumented or reanalysis was required P - Data quality checks passed and documented