Skip to main content

A bi-institutional multi-disciplinary failure mode and effects analysis (FMEA) for a Co-60 based total body irradiation technique

Abstract

Background

We aim to assess the risks associated with total body irradiation (TBI) delivered using a commercial dedicated Co-60 irradiator, and to evaluate inter-institutional and inter-professional variations in the estimation of these risks.

Methods

A failure mode and effects analysis (FMEA) was generated using guidance from the AAPM TG-100 report for quantitative estimation of prospective risk metrics. Thirteen radiation oncology professionals from two institutions rated possible failure modes (FMs) for occurrence (O), severity (S), and detectability (D) indices to generate a risk priority number (RPN). The FMs were ranked by descending RPN value. Absolute gross differences (AGD) in resulting RPN values and Jaccard Index (JI; for the top 20 FMs) were calculated. The results were compared between professions and institutions.

Results

A total of 87 potential FMs (57, 15, 10, 3, and 2 for treatment, quality assurance, planning, simulation, and logistics respectively) were identified and ranked, with individual RPN ranging between 1–420 and mean RPN values ranging between 6 and 74. The two institutions shared 6 of their respective top 20 FMs. For various institutional and professional comparison pairs, the number of common FMs in the top 20 FMs ranged from 6 to 13, with JI values of 18–48%. For the top 20 FMs, the trend in inter-professional variability was institution-specific. The mean AGD values ranged between 12.5 and 74.5 for various comparison pairs. AGD values differed the most for medical physicists (MPs) in comparison to other specialties i.e. radiation oncologists (ROs) and radiation therapists (RTs) [MPs-vs-ROs: 36.3 (standard deviation SD = 34.1); MPs-vs-RTs: 41.2 (SD = 37.9); ROs-vs-RTs: 12.5 (SD = 10.8)]. Trends in inter-professional AGD values were similar for both institutions.

Conclusion

This inter-institutional comparison provides prospective risk analysis for a new treatment delivery unit and illustrates the institution-specific nature of FM prioritization, primarily due to operational differences. Despite being subjective in nature, the FMEA is a valuable tool to ensure the identification of the most significant risks, particularly when implementing a novel treatment modality. The creation of a bi-institutional, multidisciplinary FMEA for this unique TBI technique has not only helped identify potential risks but also served as an opportunity to evaluate clinical and safety practices from the perspective of both multiple professional roles and different institutions.

Background

Total body irradiation (TBI) is a specialized radiotherapy technique for cancers involving the entire body, i.e., leukemias, lymphomas, myeloma, or other hematological malignancies [1]. The treatment is typically delivered either with a dedicated facility or as a modified application of radiotherapy equipment used for routine treatments [1, 2]. The uniqueness of TBI dose delivery demands a specialized quality management (QM) program.

Until recently, radiotherapy QM programs have focused primarily on assessing the functional performance of equipment. As suggested by guidelines from various organizations (i.e., AAPM, ACR, ACMP, IAEA, ESTRO, IEC, and ISO), it is always desirable to check and document every measurable parameter. However, given the extent and variety of radiotherapy techniques in clinical practice, this approach is not practical in terms of effective resource utilization. Therefore, the guidelines of the AAPM TG-100 report recommend a new framework for the design of the QM program [3].

This framework is mainly based on introducing prospective QM techniques. The Failure Mode and Effects Analysis (FMEA) is one such technique recommended by TG-100 [3]. Current literature exists on the application of the FMEA for intensity-modulated radiotherapy (IMRT) as well as various special procedures [4,5,6,7,8,9,10,11,12,13,14]. The FMEA technique has also been applied to acceptance and commissioning processes, clinical reference dosimetry, and radiobiological research with small animal irradiators [15,16,17,18,19]. Such an analysis is primarily helpful in the QM of novel processes and treatment techniques as well as for specialized radiotherapy procedures, such as TBI, which are already being performed.

Two investigators have recently published quality improvement analyses specific to TBI. Kim et al. presented a retrospective analysis of TBI treatments by analyzing the data from an incident learning system [20]. Shen et al. recently shared their experience using FMEA for total marrow irradiation (TMI), a more narrowly-targeted modification of TBI [21]. The authors of this study reported that a second FMEA analysis performed for high-risk failure modes (FMs) 1 year following the initial FMEA improved their QM program. While these two analyses have provided insight into FMs in TBI, there is a paucity of data examining whether the identified FMs may be generalizable between institutions. Furthermore, FMs specific to dedicated TBI equipment have not been reported yet.

This study describes the development and evaluation of an FMEA for a novel dedicated Co-60 based TBI delivery unit. It represents the first bi-institutional, multi-disciplinary FMEA for the TBI technique and should also be applicable to other forms of TBI/TMI/TMLI [22]. We will discuss the advantages and difficulties involved in an inter-institutional, inter-professional FMEA.

Materials and methods

Karmanos Cancer Institute (KCI) and Dana Farber/Brigham and Women’s Cancer Center (DFBW) have implemented a commercial dedicated Co-60 based TBI irradiator (GammaBeam 500 by Best Theratronics, Inc., Kanata, ON, Canada). The clinical commissioning of this unit has been previously described [23]. Two multi-disciplinary teams were formed for the FMEA, one at each institution (KCI and DFBW). The KCI team included three medical physicists (MPs), two radiation oncologists (ROs), and two radiation therapists (RTs) whereas the DFBW team included two participants for each of these disciplines. The specific individuals most heavily involved with the respective TBI programs were chosen to participate and all participants were familiar with the TG-100 methodology for FMEA [3].

Some additional information is provided here regarding the processes involved in treating patients with TBI on this unit, both to help the reader better understand the equipment and techniques and to allow a comparison to other treatments and modalities. The two institutions have similar treatment programs with only relatively minor operational differences. The major differences between these programs are in the in-vivo dosimetry and dose calculation procedures. While DFBW performed in-vivo dosimetry during the initiation of the treatment program, these have been discontinued. KCI performs in-vivo dosimetry for all ports for all patients. At both facilities, in-vivo dosimetry is/was performed only for part of the treatment delivery time. This time is manually subtracted from the prescribed treatment time for delivery of the remainder of the treatment. Since commercial treatment planning systems are not designed to handle such large SSDs, dose calculations are performed using in-house developed techniques. KCI uses a Monte Carlo based system to calculate the relative dose distribution [23] and both a spreadsheet system and redundant manual calculations for the absolute dose. DFBW uses a correction based manual calculation and then a different correction-based spreadsheet calculation as a secondary check. The spreadsheet formulas at both institutions are locked and password protected. There are two flattening filters (thick and thin) which are modeled in the MC dose calculation system at KCI. Compensators are created manually from regular geometric shapes of leaded polyethylene. At KCI, in-vivo dosimetry is performed at 12 points (head, suprasternal notch, umbilicus, knees, ankles, and under lung blocks for both AP and PA fields) for multifraction treatments and at the umbilicus for AP and PA fields for single 2 Gy fraction treatments. Doses are compared to both manual and MC calculations at KCI and to both correction-based calculations at DFBW. Image data is managed within a commercial imaging software system provided by Best Theratronics, Inc. Both facilities use the Aria electronic medical record system but not in a mode which interfaces with the treatment delivery system of the GammaBeam 500 system. As such, all data are manually entered into the GammaBeam 500 console and subsequently into Aria.

The process mapping and identification of potential FMs were performed jointly. Participants from all disciplines and both institutions were involved to identify the maximum range of possible FMs, as recommended by the TG-100 guidelines [3]. The list of FMs was then sent to the manufacturer of the TBI unit in an attempt to identify any additional potential FMs not already identified by the staff at the two treatment facilities. The entire treatment procedure was divided into five processes, namely; logistics, simulation, planning, delivery, and quality assurance (QA) with a total of nine sub-processes as depicted by the process map shown in Fig. 1.

Fig. 1
figure1

Process map for the Co-60 based TBI technique

The indices of occurrence (O), severity (S), and detectability (D) were rated separately by the participants in each team on a 1–10 scale as described in Table 2 of the TG-100 report [3]. As recommended, the FMEA was performed with the assumption that there were no specific QA/QC (quality control) measures in place and that O and D should be based entirely on checks that are inherent in routine clinical processes downstream [3]. A risk priority number (RPN) was then obtained for each FM as the product of these three indices and the FMs were ranked with respect to both RPN value and severity. Averaged FMEA index data were calculated, resulting in the following twelve cohorts:

  • One (1) FMEA by all participants from both institutions (named as Aggregate-FMEA)

  • Two (2) FMEA by all participants from one institution (named as KCI-FMEA and DFBW-FMEA respectively)

  • Three (3) FMEA, each with ratings from participants with the same specialty at both institutions (named as MP-FMEA, RO-FMEA, and RT-FMEA).

  • Six (6) FMEA, each with ratings from participants with the same specialty and belong to the same institution (named as KCI-MP-FMEA, KCI-RO-FMEA, KCI-RT-FMEA, DFBW-MP-FMEA, DFBW-RO-FMEA, DFBW-RT-FMEA)

The RPN values for averaged FMEAs were calculated by multiplying the mean O, S, and D values. The standard deviations (SDs) in individual indices are summed in quadrature for getting the SD in RPN.

To provide an overall comparison for the various FMEAs listed above, a single matrix is required that could represent ratings for all the FMs. We chose to report minimum, maximum, median, mean of means (MoM), 1st and 3rd quartiles for O, S, D, and RPN for all the FMs in each of the 12 FMEAs.

A pair-wise comparison was performed for similarities and differences between institutions and specialties as per following comparison pairs (CPs);

  • Inter-institutional (i.e. KCI vs DFBW)—1 CP

  • Inter-professional (i.e. MP vs RO, RO vs RT, RT vs MP)—3 CPs

  • Intra-institutional inter-professional (i.e. KCI-RO vs KCI-RT and so on)—6 CPs with 3 for each institution

  • Inter-institutional intra-professional (i.e. KCI-vs DFBW-MP and so on)—3 CPs, 1 for each specialty

Top 20 ranked FMs (ranked by RPN) were compared using the Jaccard index (JI) defined as below;

$$JI\left( {A,B} \right) = \frac{{\left| {A \cap B} \right|}}{{\left| {A \cup B} \right|}} = \frac{{\left| {A \cap B} \right|}}{{\left| A \right| + \left| B \right| - \left| {A \cap B} \right|}}$$

where A and B are the sets of top 20 FMs in A-FMEA and B-FMEA respectively. \(A \cap B\) denotes the number of common FMs in sets A and B. The higher the number of common FMs in two cohorts, the higher the JI value will be.

Similarly, the 13 CPs were analyzed for the number of FMs with complete data (n). Absolute gross differences (AGD) in RPN values were calculated and evaluated for comparison and contrast. The absolute difference measured between the RPN values (not from the differences in the individual indices) is termed the AGD. The range, mean, SD, and median of AGD in each CP was calculated.

A prioritization value for RPN as well as O, S, and D can be chosen for identifying high-priority FMs for evaluation at the first stage. In subsequent stages, the remaining set of FMs undergoes the same process until all FMs are evaluated for necessary actions. This stage-by-stage appraisal of FMEA based on such a prioritization helps in a smooth translation of changes into the QM program [3]. Here, we introduce an efficient way of prioritizing the FMs based on a plot of the cumulative number of FMs versus mean ratings. As shown in Fig. 4, for a desired number of FMs (on the x-axis) to be prioritized, a prioritization value can be determined on the y-axis.

The most frequently occurring (high O-rating), severe (high S-rating), the most difficult-to-detect (high D-rating), and high RPN FMs in all twelve FMEA spreadsheets are also discussed for inter-institutional and inter-professional comparison. FMEA reproducibility has previously been evaluated in the literature [21, 24]. To provide a sample set of data to evaluate the reproducibility of the risk assessment presented here, the FMEA was repeated by the KCI physicists approximately 2 years after the initial FMEA. These time points represent approximately 2 and 4 years following clinical implementation of the equipment and processes evaluated.

Results

A total of 87 possible FMs were identified for the entire process. The treatment process had the maximum number of identified FMs i.e. 57 (setup: 31, delivery: 26), followed by QA: 15, planning: 10 (preparation of initial plan: 6, ancillary: 4), simulation: 3, and logistics: 2 FMs. A total of 66 FMs were completely evaluated by all participants from DFBW whereas only 51 were completely evaluated by KCI participants. Combined, 46 FMs had complete evaluation data from all participants at both institutions. However, as we note from Table 3, at the minimum, 71 FMs were rated by at least one individual for any CP. MPs rated the maximum FMs (86), whereas some ROs and RTs did not rank some or all indices for a few FMs due to their inability to accurately assess these indices. All data on individual and averaged FMEAs are given in “Supplementary Appendix 1” for reference.

Individual O, S, and D ratings for most participants spanned the full scale (1–10). However, ranges as low as 1–4 are also observed for some participants. Maximum RPN values ranged from 90 to 420 for individual raters.

Figure 2 represents the range and extent of the data obtained for the 12 averaged FMEAs on a box and whisker plot (whiskers show minimum and maximum, whereas the box show 1st, and 3rd quartile along with median, and the crossed datapoint shows the MoM for mean O, S, D, and RPN). We observe that the KCI-MPs generated the highest MoM (i.e. 81.6) and maximum (i.e. 420) for RPN scores compared to the rest of the participants. KCI-RTs generated the lowest MoM (i.e. 9.6) and maximum (i.e. 50) for RPN values, reflecting the inter-professional differences within an institution.

Fig. 2
figure2

Range and extent of the data in a box and whisker plot for various FMEAs

The extent of the aggregate-FMEA data is also illustrated in Fig. 3 which plots mean and SD for O, S, D, and RPN metrics in descending order. In general, participants tended to rate S values highest (8.5, SD:1.6) compared to D (4.6, SD:2.6) and O (3.9, SD:1.7) values. A similar trend exists for the range of mean index ratings in averaged FMEAs, with S having the maximum range (2.2–8.5), followed by D (1.0–4.6 [low ‘detectability’ values show that the FMs were more easily detectable]) and O (1.3–3.9), respectively. It is, however, to be noted that the SD is higher for RPN as compared to the individual index ratings.

Fig. 3
figure3

Statistical representation of mean O, S, D, and RPN data (in A, B, C, and D, respectively) for 87 FMs in descending order for Aggregate-FMEA [vertical bars show SD]

Figure 4 shows the cumulative number of FMs corresponding to a particular rating or RPN value and allows visualization of differences between the two facilities. The top 20 FMs (ranked by RPN) were chosen for the first stage of implementation of our FMEA findings. This corresponds to an RPN prioritization value of 41 and 43 for DFBW and KCI, respectively. This prioritization value was arbitrarily chosen but could be chosen based on a particular RPN cutoff value or potentially by the slope of this curve. This prioritization roughly corresponds to 1/4 of the total number of FMs (i.e. 20 out of total 87 FMs). Table 1 presents the 20 highest ranking FMs for each institution for all groups ranked by total RPN and by Severity index. As observed in Table 1 and Fig. 4, the most pronounced difference between the mean RPN rankings for the two institutions is at the upper end of the RPN range.

Fig. 4
figure4

Cumulative number of FMs versus the overall mean value of O, S, D, and RPN (in A, B, C, and D, respectively)

Table 1 The top 20 FMs ranked by mean RPN and ‘S’ scores for each institution

A number of preventative measures were implemented as a result of this FMEA. Table 2 lists the highest scoring FMs by RPN along with associated preventative measures. Of the 40 measures listed, 11 represent new interventions implemented as a result of this FMEA (5 at KCI and 6 at DFBW).

Table 2 Interventions resulting from the FMEA process

Table 3 presents a variety of parameters that highlight the similarities and differences in various CPs. The number of common FMs (A∩B) in the top 20 ranked FMs ranged between 6 and 13, resulting in JI values of 18–48% respectively. AGD values as high as 407.3 were observed, whereas the mean and median AGD in any CP varied from 12.5 to 74.5, and 9.3 to 38.4 respectively.

Table 3 Similarities and differences for different comparison pairs (CPs)

As shown in Table 1, the two institutions have six common FMs in the top 20 ranked FMs (resulting in a JI of 18%) which is the lowest among all CPs. The inter-institutional intra-professional comparison shows that the number of common FMs in the top 20 list is the highest for MPs (9/20; JI = 29%), followed by ROs (7/20; JI = 21%) and RTs (6/20; JI = 18%). If we compare different specialties without being institution-specific, the number of common FMs in the top 20 is the same for MP vs RO and RO vs RT (10/20; JI = 33% each) and is slightly lower for RT vs MP (9/20; JI = 29%). Further, the JI values for inter-professional comparison differ from institution-to-institution as we evaluate KCI (CP 8, 9, 10 in Table 3) and DFBW (CP 11, 12, 13 in Table 3) data. For KCI, the number of common FMs is the highest for MPs vs ROs (13/20; JI = 48%) followed by ROs vs RTs and RTs vs MPs who both have 10/20 common FMs (JI = 33%). Whereas, for DFBW, the results show 9/20 common FMs for ROs vs RTs (JI = 29%) followed by MPs vs ROs with 7/20 (JI = 21%) and RTs vs MPs with 6/20 (JI = 18%).

The mean inter-institutional AGD value of 14.6 (SD:12.7) is smaller than that observed in either inter-professional comparisons or intra-professional comparisons across institutions. The overall inter-professional comparison (CP 2, 3, 4 in Table 3) suggests that there is a better agreement among ROs and RTs (mean AGD 12.5, SD: 10.8, range: 0.2–48.1) as compared to the other two CPs. This trend is the same even if the data from only one institution (whether KCI or DFBW) are observed.

Evaluation of the reproducibility of the FMEA results by the KCI physicists following a 2 year time interval revealed differences in RPN (new–old) ranging from + 55 to − 276 with a mean value of − 33. Despite this variability, 62 of the 87 FMs had a new RPN within ± 50 of the old RPN, and 16 of the top 20 FMs from the original FMEA were still in the top 20 in the new FMEA. These results demonstrate a small decrease in RPN over time but a relative consistency in the FMEA ranking as a function of time.

Discussion

The results show that a multi-disciplinary, bi-institutional FMEA helps elucidate differences in processes between institutions as well as the prioritization of risks amongst different professional groups. Even when minor operational differences exist, institutions can benefit from mutual experiences, differing experiences, and resultant risk estimation. These operational differences represent the primary reason for the incomplete ratings for some FMs, although these incomplete ratings can still be used for analysis when at least one scorer’s rating is available. The dose calculation and in-vivo dosimetry processes represent an example of the influence of operational differences on FMEA ranking results. While both institutions make use of both a manual and a spreadsheet calculation for treatment time, KCI performs a second check of the manual calculation and performs in-vivo dosimetry for all patients. While corruption of the treatment time calculation spreadsheet is the highest ranking FM for DFBW, it is only the 15th highest for KCI, largely due to the fact that the detectability estimated by KCI is lower by almost a factor of three. While the rankings should not consider QA/QC measures in place, these existing measures may have unintentionally influenced the detectability ranking. Additionally, O and D scores are to be based on checks inherent in routine clinical processes downstream and some raters may have considered redundancies in treatment time calculation and/or in-vivo dosimetry to be routine clinical processes. This highlights one of the complexities involved in the implementation of an FMEA, particularly one which includes multiple institutions, and likely explains the large differences in estimated detectability for this FM.

The MPs tend to rate the greatest number of FMs even in the case of FMs related to bi-institutional operational differences. In contrast, ROs tend to omit rankings for FMs related to QA. Interestingly, some of the RTs also omitted rankings for some of the FMs related to QA as well as treatment setup. This indicates a potential lack of awareness of some safety and QA processes and/or risks by some groups or individuals involved in the treatment process. The fact that these individuals were included in the FMEA process is a valuable first step in recognizing these risks. In addition, the interventions developed to address the highest scoring FMs were shared among all groups participating in the treatment process, thus promoting further awareness of safety and QA processes and potential risks.

As shown for the individual indices, all scorers were inclined to rate ‘S’ more highly as compared to ‘O’ and ‘D’. This trend remains the same for both institutions and all specialties. This suggests either that participants consider the severity of the identified FMs as the most important factor or that the rating scale utilized from the AAPM TG-100 report favors higher ‘S’ scores, followed by difficulty in detectability, and occurrence of the events.

Furthermore, the results shown by Fig. 1 and Table 1 reveal high SD values compared to the mean values for individual indices (O, S, D). This highlights the fact that each FM can be rated quite differently by different scorers. This disparity is even higher for RPN values where SD is sometimes approximately double the mean value. This is because the SD in RPN is based on the propagation of the uncertainties in the factors used in calculating RPN (O, S, D).

It is observed that the highest RPN in the DFBW-FMEA is lower than that of the KCI-FMEA (104.9, SD:72.0 vs 74.9, SD:42.3), owing to inter-interinstitutional differences. The inter-professional comparison shows that MPs have the highest overall mean RPN value (192.5, SD:65.1) followed by ROs (70.1, SD:48.6) and RTs (63.4, SD:35.4). While O, S, and D mean scores were all higher for MPs, the largest differences between the MPs and other groups were in the scoring of occurrence and detectability. If the MP, as the group typically leading safety and QA, is the most knowledgeable about these risks, this result potentially indicates a greater sense of false security in the relative occurrence and detectability of FMs studied here. For the FMs with the most severe consequences, the different specialties tend to agree on ratings. ROs assigned the highest S-rating of 10 (SD:0.0), followed by MPs with 9.3 (SD:0.9) and RTs with 9.00 (SD:1.0). Likewise, for the frequency of occurrence of FMs, the highest mean O-rating was recorded by MPs (5.5, SD:1.5), followed by RTs (5.0, SD:2.0) and ROs (4.3, SD:1.8). Interestingly, participants with different specialties tend to disagree the most on the detectability (D-index). This is potentially due to a lack of familiarity with all mechanisms used to assure safety and quality and the participation of all groups in this FMEA is a positive step toward larger involvement of all groups in the safety process and increased vigilance for the identification of potential errors and unsafe practices. The overall mean value for MPs was 8.3 (SD:1.7) in comparison to 6.5 (SD:2.50) for RTs and 4.3 (SD:3.2) for ROs.

For the similarities in the top 20 FMs, the inter-professional similarity values (9 or 10 common FMs in top 20 i.e. 9–10/20; JI = 29–33%) are better than the inter-institutional similarity (6/20; JI = 18%). The inter-institutional intra-professional comparison shows that the MPs have the most intra-professional similarity (9/20; JI = 29%) compared to other specialties (ROs: 7/20 [JI = 21%], RTs: 6/20 [JI = 18%]). Intra-institutional inter-professional similarity shows that the trend in the similarity between specialties can vary from institution to institution. For KCI, JIMP-RO = 48%, and JIRO-RT = JIRT-MP = 33% based on the number of common FMs in the top 20 i.e. 13/20, 10/20, and 10/20 respectively. Whereas, for DFBW, JIMP-RO = 21%, JIRO-RT = 29%, and JIRT-MP = 18% for the number of common FMs as 7/20, 9/20, and 6/20 respectively. One might expect more commonality among the highest scoring FMs, however, this study serves to illustrate the potential differences, assumed primarily to be in the operational details, between institutions implementing the same procedures on the same equipment.

Evaluating AGD values from the FMEA, there is a good agreement in inter-institutional comparison (Mean AGDKCI-DFBW = 14.6, SD:12.7) relative to the inter-professional comparisons (Mean AGDMP-RO = 36.3, SD:34.1, AGDRO-RT = 12.5, SD:10.8, AGDRT-MP = 41.2, SD: 37.9). Interestingly, the AGD values in inter-institutional intra-profession comparison illustrate that the MPs have a greater difference (AGD = 58.3, SD = 68.8) as compared to ROs (AGD = 16.4, SD = 17.3) and RTs (AGD = 18.0, SD = 18.7). In addition, the inter-professional variability was found to be institution-specific.

A maximum AGD value of 407.3 was observed for FM#49 for KCI-RT ([O, S, D] = [1.5, 8.5, 1.0] and RPN = 12.5) vs KCI-MP ([O, S, D] =  [7, 10, 6] and RPN = 420). This FM was “Programming the incorrect treatment time.” This very large discrepancy highlights the differences in estimation of O, S, and D values across professions. Both RTs and MPs evaluated this FM as relatively severe (S = 8.5 and 10 for RTs and MPs, respectively). However, RTs estimated an occurrence frequency between 0.01 and 0.02%, compared to 0.5–1% for MPs, and RTs estimated a detectability of 99.99%, compared to 95% for MPs. So one may question whose risk estimations are more accurate, the group who oversees departmental quality, safety, and risk assessment, or the group who is actually performing this function. If we had adequate data available to answer this question, we wouldn’t necessarily need to apply prospective risk assessment. Instead, these differences help us understand the relative uncertainties involved in the prospective risk assessment process. Differences between groups in the estimation of occurrence and detectability for this FM result in a relative increase in the overall RPN score by a factor of 4.7 and 6, respectively. Such large changes would substantially change the ranking of a FM. For example, dividing the highest aggregate RPN in our FMEA by a factor of 5 would demote it the 73rd highest ranking FM, thus dramatically reducing the attention it would receive as a potential risk factor. This highlights the importance of participation of as many individuals as possible in performing the prospective risk analysis and that it should not be restricted to individuals within a particular discipline with particular perspectives on the associated processes. This reinforces the TG-100 report recommendation to use “a team-based approach that requires active participation of representatives from all treatment team member categories” [3].

The differences observed between the original and new FMEAs performed by the KCI MP participants highlights the potential variability in the results of an FMEA, as the 2-year time difference was more than enough to assure that there was no recollection of the original FMEA scoring. However, it also likely represents changes in perceived occurrence, severity, and detectability over time. The average overall decrease in RPN over this 2-year period likely represents an increased comfort level that FMs are either less likely to occur (for those that have not yet occurred) or more likely to be detected (due to greater experience with the treatment unit), or both. A similar result was obtained by Mancosu et al. (2021) for repetition of an FMEA after 10 years (25). In those 10 years, there has been a great deal of evolution in the performance of the FMEA in radiation oncology.

One should be very careful in applying AGDs for comparison, whether between institutions or among different specialties. Two FMs with different O, S, and D index values may end up with comparable RPN values thus showing lower AGD values [25]. Similarly, the data on JI is confined to 20 FMs in our case (for the first stage/high priority). Therefore, we recommend that the results on AGD should be used in conjunction with JI for the top 20 FMs to draw any conclusions.

Also, the intra- or inter-institutional and -professional comparisons of the FMs ranked with respect to the individual index values may not be a good approach as such a ranking is highly subjective. As an example, the DFBW-FMEA suggests that the most frequent FM is the therapist hurting themselves while trying to move the patient. In contrast, this FM is ranked 66th most frequent in KCI-FMEA. Such differences could arise purely from the past experiences of one institution or profession.

There is often a discordance between a clinician’s “gut feeling” about the relative risk associated with an event, and a formal FMEA evaluation of that event. As an example, a source stuck in the exposed position accompanied by failure of the emergency shutter to deploy sounds like a catastrophic event. However, there are abundant detection mechanisms for this event inherent in the downstream clinical processes. As such, this failure would be detected relatively quickly even without the consideration of QA/QC measures (as required in the TG-100 FMEA). Estimating the time associated with routine detection and subsequent time for remediation based on our emergency procedures and mock drills, we anticipate that such an event will most likely result only in a minor dosimetric error to the patient. Even assuming a worst case-scenario for severity, the very low predicted occurrence of both the source getting stuck and the emergency shutter failing simultaneously combined with the very high detectability of this event result in an RPN score that is relatively modest.

In other industries, a cut-off RPN value (or of individual indices) is often determined (instead of a prioritization value) [26] to select a handful of FMs for consideration. In those industries, the preventive measures are suggested for that set of FMs and the rest may be discarded. We do not consider any FMs as irrelevant in radiation oncology. On the other hand, prioritizing only a small set of FMs for intervention helps these changes to be translated into the department relatively simply and without major changes [3]. We chose a small representative set for the initial stage of the top 20 ranked by RPN, or 1/4 of the total number of FMs. Of course, a different number can be chosen depending on the human resource availability and readiness of the staff for a change in QM.

Conclusion

We present the first bi-institutional multi-disciplinary FMEA analysis of a Co-60 based dedicated TBI technique. The results highlight the relative variability in ranking FMs when sampled across multiple professions and institutions. Differences in ratings at the two institutions were often linked with operational differences. The inter-professional variability appears to be institution-specific. We find that a bi-institutional and/or multi-disciplinary FMEA is not only feasible but is also helpful in many ways. Such a comparison allows the institutions to learn from one another’s processes, experiences, and risk estimates in seeking areas for improvement in their QM program. Similarly, the inter-professional comparison helps evaluate differences in risk estimation among different specialties. Such variability may be due to either difference in processes or differences in perceptions of risk between institutions and professions. It ultimately helps both in the identification of the most important FMs and in identifying the best practices to address those FMs. The involvement of all groups participating in the treatment process within an FMEA not only improves the quality of data obtainable from the FMEA, but also should have a tangible benefit on departmental QA and safety. Providing all stakeholders with a better understanding of potential risks and their relative occurrence, severity, and detectability, promises to help maximize quality assurance, minimize risks, and improve the overall safety culture of the department.

Availability of data and materials

All data generated or analyzed during this study are included in this published article (and its Additional file 1).

Abbreviations

TBI:

Total body irradiation

FMEA:

Failure mode and effects analysis

FM(s):

Failure mode(s)

O:

Occurrence

S:

Severity

D:

Detectability

RPN:

Risk priority number

SD:

Standard deviation

AGD:

Absolute gross differences

JI:

Jaccard Index

MP(s):

Medical physicist(s)

RO(s):

Radiation oncologist(s)

RT(s):

Radiation therapist(s)

QM:

Quality management

IMRT:

Intensity-modulated radiotherapy

TMI:

Total marrow irradiation

QA:

Quality assurance

QC:

Quality control

Aggregate-FMEA:

An FMEA with ratings by all participants from both institutions

KCI:

Karmanos Cancer Institute

DFBW:

Dana-Farber/Brigham and Women’s Cancer Center

KCI-FMEA:

An FMEA with ratings by all participants from KCI

DFBW-FMEA:

An FMEA with ratings by all participants from DFBW

X-FMEA:

An FMEA with ratings by participants with the same specialty (X, which can either be MP, or RO, or RT) from both institutions.

A-X-FMEA:

The FMEA ratings from participants with the same specialty (X) and belong to the same institution (A, which can either be KCI or DFBW)

MoM:

Mean of means

CP(s):

Comparison pair(s)

References

  1. 1.

    Van Dyk J. American Institute of Physics. American Association of Physicists in Medicine. The physical aspects of total and half body photon irradiation. 1986.

  2. 2.

    Hussain A, Eduardo J, Brown D. Total body irradiation. Qual Saf Radiother. 2010. https://doi.org/10.1201/b10448.

    Article  Google Scholar 

  3. 3.

    Huq MS, Fraass BA, Dunscombe PB, Gibbons JP, Ibbott GS, Mundt AJ, et al. The report of task group 100 of the AAPM: application of risk analysis methods to radiation therapy quality management. Med Phys. 2016;43:4209–62. https://doi.org/10.1118/1.4947547.

    Article  PubMed  PubMed Central  Google Scholar 

  4. 4.

    Xu AY, Bhatnagar J, Bednarz G, Flickinger J, Arai Y, Vacsulka J, et al. Failure modes and effects analysis (FMEA) for Gamma Knife radiosurgery. J Appl Clin Med Phys. 2017;18:152–68. https://doi.org/10.1002/acm2.12205.

    Article  PubMed  PubMed Central  Google Scholar 

  5. 5.

    Faught JT, Balter PA, Johnson JL, Kry SF, Court LE, Stingo FC, et al. An FMEA evaluation of intensity modulated radiation therapy dose delivery failures at tolerance criteria levels. Med Phys. 2017;44:5575–83. https://doi.org/10.1002/mp.12551.

    CAS  Article  PubMed  Google Scholar 

  6. 6.

    Manger RP, Paxton AB, Pawlicki T, Kim GY. Failure mode and effects analysis and fault tree analysis of surface image guided cranial radiosurgery. Med Phys. 2015. https://doi.org/10.1118/1.4918319.

    Article  PubMed  Google Scholar 

  7. 7.

    Ibanez-Rosello B, Bautista-Ballesteros JA, Bonaque J, Celada F, Lliso F, Carmona V, et al. Failure mode and effects analysis of skin electronic brachytherapy using Esteya® unit. J Contemp Brachyther. 2016. https://doi.org/10.5114/jcb.2016.64745.

    Article  Google Scholar 

  8. 8.

    Ibanez-Rosello B, Bautista JA, Bonaque J, Perez-Calatayud J, Gonzalez-Sanchis A, Lopez-Torrecilla J, et al. Failure modes and effects analysis of total skin electron irradiation technique, vol 20. Springer International Publishing; 2018. https://doi.org/10.1007/s12094-017-1721-3.

  9. 9.

    Ciocca M, Cantone MC, Veronese I, Cattani F, Pedroli G, Molinelli S, et al. Application of failure mode and effects analysis to intraoperative radiation therapy using mobile electron linear accelerators. Int J Radiat Oncol Biol Phys. 2012;82:305–11. https://doi.org/10.1016/j.ijrobp.2011.05.010.

    Article  Google Scholar 

  10. 10.

    Veronese I, De Martin E, Martinotti AS, Fumagalli ML, Vite C, Redaelli I, et al. Multi-institutional application of failure mode and effects analysis (FMEA) to CyberKnife stereotactic body radiation therapy (SBRT). Radiat Oncol. 2015;10:1–10. https://doi.org/10.1186/s13014-015-0438-0.

    Article  Google Scholar 

  11. 11.

    Younge KC, Lee C, Moran JM, Feng M, Novelli P, Prisciandaro JI. Failure mode and effects analysis in a dual-product microsphere brachytherapy environment. Pract Radiat Oncol. 2016;6:e299-306. https://doi.org/10.1016/j.prro.2016.03.003.

    Article  PubMed  Google Scholar 

  12. 12.

    Younge KC, Wang Y, Thompson J, Giovinazzo J, Finlay M, Sankreacha R. Practical implementation of failure mode and effects analysis for safety and efficiency in stereotactic radiosurgery. Int J Radiat Oncol Biol Phys. 2015;91:1003–8. https://doi.org/10.1016/j.ijrobp.2014.12.033.

    Article  PubMed  Google Scholar 

  13. 13.

    Ford EC, Gaudette R, Myers L, Vanderver B, Engineer L, Zellars R, et al. Evaluation of safety in a radiation oncology setting using failure mode and effects analysis. Int J Radiat Oncol Biol Phys. 2009;74:852–8.

    Article  Google Scholar 

  14. 14.

    Scorsetti M, Signori C, Lattuada P, Urso G, Bignardi M, Navarria P, et al. Applying failure mode effects and criticality analysis in radiotherapy: lessons learned and perspectives of enhancement. Radiother Oncol. 2010;94:367–74.

    Article  Google Scholar 

  15. 15.

    Poirier Y, Johnstone CD, Anvari A, Brodin NP, Dos SM, Bazalova-Carter M, et al. A failure modes and effects analysis quality management framework for image-guided small animal irradiators: a change in paradigm for radiation biology. Med Phys. 2020. https://doi.org/10.1002/mp.14049.

    Article  PubMed  Google Scholar 

  16. 16.

    Teo PT, Hwang MS, Shields W, Kosterin P, Jang SY, Heron DE, et al. Application of TG-100 risk analysis methods to the acceptance testing and commissioning process of a Halcyon linear accelerator. Med Phys. 2019. https://doi.org/10.1002/mp.13378.

    Article  PubMed  Google Scholar 

  17. 17.

    Ochi Y, Saito A, Kawahara D, Suzuki T, Tsuneda M, Tanaka S, et al. A novel risk analysis of clinical reference dosimetry based on failure modes and effects analysis. Physica Med. 2019. https://doi.org/10.1016/j.ejmp.2019.01.014.

    Article  Google Scholar 

  18. 18.

    Harry T, Yaddanapudi S, Cai B, Stinson K, Murty Goddu S, Noel C, et al. Risk assessment of a new acceptance testing procedure for conventional linear accelerators. Med Phys. 2017. https://doi.org/10.1002/mp.12527.

    Article  PubMed  Google Scholar 

  19. 19.

    Wexler A, Gu B, Goddu S, Mutic M, Yaddanapudi S, Olsen L, et al. FMEA of manual and automated methods for commissioning a radiotherapy treatment planning system. Med Phys. 2017. https://doi.org/10.1002/mp.12278.

    Article  PubMed  Google Scholar 

  20. 20.

    Kim A, Ford E, Spraker M, Zeng J, Ermoian R, Jordan L, et al. Are we making an impact with incident learning systems? Analysis of quality improvement interventions using total body irradiation as a model system. Pract Radiat Oncol. 2017;7:418–24. https://doi.org/10.1016/j.prro.2017.05.010.

    Article  PubMed  Google Scholar 

  21. 21.

    Shen J, Wang X, Deng D, Gong J, Tan K, Zhao H, et al. Evaluation and improvement the safety of total marrow irradiation with helical tomotherapy using repeat failure mode and effects analysis. Radiat Oncol. 2019;14:238. https://doi.org/10.1186/s13014-019-1433-7.

    Article  PubMed  PubMed Central  Google Scholar 

  22. 22.

    Wong JYC, Filippi AR, Scorsetti M, Hui S, Muren LP, Mancosu P. Total marrow and total lymphoid irradiation in bone marrow transplantation for acute leukaemia. Lancet Oncol. 2020;21:e477–87.

    Article  Google Scholar 

  23. 23.

    Burmeister J, Nalichowski A, Snyder M, Halford R, Baran G, Loughery B, et al. Commissioning of a dedicated commercial Co-60 total body irradiation unit. J Appl Clin Med Phys. 2018;19:131–41. https://doi.org/10.1002/acm2.12309.

    Article  PubMed  PubMed Central  Google Scholar 

  24. 24.

    Mancosu P, Signori C, Clerici E, Comito T, D’Agostino GR, Franceschini D, et al. Critical re-evaluation of a failure mode effect analysis in a radiation therapy department after 10 years. Pract Radiat Oncol. 2021;11:e329–38.

    Article  Google Scholar 

  25. 25.

    Thornton E, Brook OR, Mendiratta-Lala M, Hallett DT, Kruskal JB. Quality initiatives: application of failure mode and effect analysis in a radiology department. Radiographics. 2011. https://doi.org/10.1148/rg.311105018.

    Article  PubMed  Google Scholar 

  26. 26.

    Gervais B, D’Arcy DM. Quality risk analysis in a cGMP environment: Multiple models for comprehensive failure mode identification during the computer system lifecycle. Drug Dev Ind Pharm. 2014. https://doi.org/10.3109/03639045.2012.744417.

    Article  PubMed  Google Scholar 

Download references

Acknowledgements

The authors thank Mark Xu from Best Theratronics for valuable discussions regarding the identification of potential equipment failure modes, causes, and effects.

Funding

The Wayne State University/Karmanos Cancer Institute provided the financial support for publishing charges. Shahbaz Ahmed is a recipient of William J. Fulbright PhD Scholarship through the United States Educational Foundation of Pakistan (USEFP).

Author information

Affiliations

Authors

Contributions

All of the authors (SA, TB, AN, JSB, SB, KB, MD1, MD2, MK, GMM, KJM, AN, MT, JB) contributed to acquisition, analysis, and interpretation of the data. SA and JB have drafted the work, revised it and were major contributors in writing. All authors read and approved the final manuscript.

Corresponding author

Correspondence to Shahbaz Ahmed.

Ethics declarations

Ethics approval and consent to participate

Not applicable.

Consent for publication

Not applicable.

Competing interests

The authors declare that they have no competing interests.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Additional file 1.

Aggregate FMEA (the mean FMEA ranked by all participants from both institutions).

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and Permissions

About this article

Verify currency and authenticity via CrossMark

Cite this article

Ahmed, S., Bossenberger, T., Nalichowski, A. et al. A bi-institutional multi-disciplinary failure mode and effects analysis (FMEA) for a Co-60 based total body irradiation technique. Radiat Oncol 16, 224 (2021). https://doi.org/10.1186/s13014-021-01894-3

Download citation

Keywords

  • FMEA
  • TBI
  • Bi-institutional
  • Multidisciplinary
  • Risk assessment
  • Risk estimation
  • TG-100
  • Quality management