Analyzing operating and support costs for Air Force aircraft

Scott C. Hewitson (Department of Systems Engineering and Management, Air Force Institute of Technology, Wright-Patterson AFB, Ohio, USA)
Jonathan D. Ritschel (Department of Systems Engineering and Management, Air Force Institute of Technology, Wright-Patterson AFB, Ohio, USA)
Edward White (Department of Mathematics and Statistics, Air Force Institute of Technology, Wright-Patterson AFB, Ohio, USA)
Gregory Brown (Air Force Life Cycle Management Center, Wright-Patterson AFB, Ohio, USA)

Journal of Defense Analytics and Logistics

ISSN: 2399-6439

Article publication date: 3 August 2018

Issue publication date: 5 September 2018

1774

Abstract

Purpose

Recent legislation resulted in an elevation of operating and support (O&S) costs’ relative importance for decision-making in Department of Defense programs. However, a lack of research in O&S hinders a cost analyst’s abilities to provide accurate sustainment estimates. Thus, the purpose of this paper is to investigate when Air Force aircraft O&S costs stabilize and to what degree. Next, a parametric O&S model is developed to predict median O&S costs for use as a new tool for cost analyst practitioners.

Design/methodology/approach

Utilizing the Air Force total ownership cost database, 44 programs consisting of 765 observations from 1996 to 2016 are analyzed. First, stability is examined in three areas: total O&S costs, the six O&S cost element structures and by aircraft type. Next, stepwise regression is used to predict median O&S costs per total active inventory (CPTAI) and identify influential variables.

Findings

Stability results vary by category but generally are found to occur approximately five years from initial operating capability. The regression model explains 89.01 per cent of the variance in the data set when predicting median O&S CPTAI. Aircraft type, location of lead logistics center and unit cost are the three largest contributing factors.

Originality/value

Results from this research provide insight to cost analysts on when to start using actual O&S costs as a baseline for estimates in lieu of analogous cost program data and also derives a new parametric O&S estimating tool designed as a cross-check to current estimating methodologies.

Keywords

Citation

Hewitson, S.C., Ritschel, J.D., White, E. and Brown, G. (2018), "Analyzing operating and support costs for Air Force aircraft", Journal of Defense Analytics and Logistics, Vol. 2 No. 1, pp. 38-54. https://doi.org/10.1108/JDAL-03-2018-0006

Publisher

:

Emerald Publishing Limited

Copyright © 2018, Scott C. Hewitson, Jonathan D. Ritschel, Edward White and Gregory Brown.


On May 22, 2009, the Weapon Systems Acquisition Reform Act (WSARA) was enacted as Public Law 111-23. For Department of Defense (DoD) cost analysts and acquisition professionals, WSARA represents a watershed moment. Prior to WSARA, the focus of these DoD professionals was overwhelmingly on the acquisition (i.e. research and development or procurement) phases of a program’s life cycle (Ryan et al., 2013). WSARA, however, brought incipient attention and emphasis to the operating and support (O&S) phase of the life cycle (Congress US, 2009). To be clear, it is not the requirement to develop and complete O&S cost estimates that has changed. Rather, it is the elevation of the relative importance of the O&S estimates for decision-making in relation to its acquisition counterparts that has changed. This historically dichotomous relationship is mirrored in the academic literature. In fact, an analysis of the DoD cost literature shows that acquisition cost behavior was the focus of over 130 studies from 1945 to 2009, while published analysis of O&S cost behavior during the same time rarely occurred (Ryan et al., 2012). Since the enactment of WSARA in 2009, however, a burgeoning O&S research stream has appeared in the literature as demonstrated in the works of Boito et al. (2012), Jones et al. (2014) and Ritschel and Ritschel (2016), amongst others.

Additionally, the practical importance of accurate O&S estimates is manifesting itself in programmatic decisions. For example, the decision to divest (or not) the U-2 in favor of the RQ-4 Global Hawk is framed not only by relative capability performance of the platforms but also by their expected O&S costs. The significance of cost in the decision process is clear: “as costs for operating the Global Hawk fell below those of the U-2, the service chose to go with the unmanned platform going forward” (Pomerleau, 2016). While the final decision on retiring the U-2 is in flux, as increased defense funding in the 2017 budget has resulted in both platforms being retained, the importance of O&S costs to decision-makers is evident.

Given the extant need for credible O&S estimates and their known historical inaccuracies (Ryan et al., 2013), we seek to provide improvements in the field of O&S estimating. Cost analysts use a range of estimating techniques, including parametric, analogy, engineering build-up or extrapolation from actuals (actual costs). As data become available and programs mature, analysts shift from techniques such as the analogy method to the extrapolation from actuals method (Mislick and Nussbaum, 2015). Current practice for O&S estimates relies heavily on the analogy method. The development of the Air Force Total Ownership Cost (AFTOC) database in 1996, which collects expenditures on the O&S costs of DoD platforms, provides a relatively recent repository that cost analysts can use to migrate from an analogy technique to the extrapolation from actual method. But when can analysts make the transition? To transition to utilizing actual data, cost analysts strive to understand when a steady-state is achieved. Thus, we seek to answer this question by investigating O&S cost stability properties in US Air Force aircraft platforms.

While determining stability properties helps cost analysts working estimates on mature platforms, there is also a large segment of analysts performing estimates at varied stages in a program’s life cycle. Therefore, the second part of the research develops a regression model (including the results of the stability analysis as an explanatory variable) to predict median O&S costs. The intent of this parametric model is not necessarily to replace current practices. Rather, the model is intended as a secondary technique, or crosscheck, to the analyst’s primary methodology. This fulfills a critical role as cost analysts typically use multiple techniques to garner confidence in the final estimate (Government Accountability Office, 2009). The desired result is convergence around a similar number, thereby providing the decision-maker with the most reliable cost estimate possible.

The literature reveals limited inquiries into stability properties but does provide identification of potentially important independent variables for inclusion in the parametric model developed in this research. Jones et al. (2014) examine the ratio of O&S costs to acquisition costs. A methodological assumption in their analysis is that O&S costs are deemed stable once 10 per cent of the planned procurement quantity are produced. The 10 per cent inclusion criterion is used to simplify their data set; however, there is no data-driven evidence to validate the claim. Dixon (2005) analyzes the effects of aging on commercial aircraft and finds sustainment results in three distinct phases: newness, mature and aging. Dixon discovers that the newness phase ends at approximately five-seven years, which may indicate the initiation of stability. Although commercial aircraft and military aircraft have distinctly different objectives, there are parallels to the analysis, and age is included as a variable in our model. While stability has not been explicitly analyzed in the O&S arena, it has been examined in earned value management (EVM) data. Christensen and Payne use the cost performance index in EVM as a tool to gauge stability in DoD contracts (Christensen and Payne, 1992). Petter et al. (2015) add to this research by investigating earned schedule and its application to stability while analyzing three definitions of stability. While EVM focuses on contract performance and is inherently different than O&S’s perpetual nature, the methodologies provide an insight on how to approach the determination of stability for O&S costs.

In 2001, the Congressional Budget Office (CBO) developed a rudimentary regression model to predict O&S costs that uses variables from a 1990 RAND study (Hildebrandt and Sze, 1990). The 2001 study uses the nascent AFTOC database to create the model. The AFTOC database originated in 1996 resulting in only four years of data in their analysis (CBO, 2001). Key variables identified in the model include age, tempo and unit cost. Incorporating additional years of data and adding programmatic information provides an opportunity for a more robust model to be built.

Database and methodology

The primary data used in this research include annual O&S costs for 44 aircraft programs (765 observations) gathered from the AFTOC database. Research is limited to aircraft in the Air Force inventory and expenditures from 1996 to 2016 because AFTOC does not contain data prior to 1996 or expenditures from 2017 at the time of this analysis. AFTOC provides costs in base-year dollars and then-year dollars. For this research, all costs are converted with Office of the Secretary of Defense (OSD) inflation indices to base-year 2016 to remove the effects of inflation. Initial operational capability (IOC) dates are needed for standardization purposes and are found using selected acquisition reports. To ensure robustness of programs selected for analysis, they must have a total active inventory (TAI) of over 10 in at least one year of cost data, at least five years of cost data for a mission design series (MDS) to be included and an available IOC year. In the data gathering process, some MDS (variants) shared IOC dates which necessitated grouping. Initial screening of the base data set is shown in Table I, while Table II highlights the 44 programs used for analysis grouped by aircraft mission type from AFTOC.

Once data are collected and screened, costs are standardized by calculating cost per flying hour (CPFH) and cost per total active inventory (CPTAI) to control for the variance in hours flown or number of aircraft in the inventory. The three cost types explored are total O&S costs, CPFH and CPTAI. “Time from IOC” is standardized by subtracting the IOC year of a certain MDS from the given year of O&S cost data. For simplicity, the IOC year is looked at, not the month. Data ranges from 1996 to 2016, but the aircraft included are at different points in their sustainment phase. For example, in 2010, the C-17A is 17 years from IOC, while the F-22A is only 5 years from IOC. When all data points are compiled, trends can be seen in terms of years from IOC for a specific program.

The determination of stability for cost analysts relies on a year-to-year per cent cost difference for each year from IOC. In the analysis, the six cost element structures (CES) from the OSD, cost assessment and program evaluation (OSD CAPE) cost estimating guide, shown in Table III, are used to highlight differences in stability properties by CES. Total O&S costs are an aggregate of all six CES, but only the first five CESs are analyzed individually because of the variance in how indirect costs (CES 6) are levied on different programs. Indirect support costs are installation and personnel costs allocated on “a per capita or some other basis,” which can be ambiguous for cost analysts (OSD CAPE, 2014). Once individual differences (percentage) are calculated for each year from IOC for all 44 programs, the mean difference for each year from IOC is determined.

Truncating O&S years beyond 40 occurs because of the reasonable assumption of how long an MDS will be in operation. The 44 selected programs have years of data ranging from 1 to 57 years from IOC. This research aims to determine when stability occurs within the expected service life. Therefore, OSD CAPE’s largest notional service life duration, which between fixed wing and rotary aircraft is 40 years, is the cutoff for truncation (OSD CAPE, 2014). This approach is consistent with the prior literature on O&S costs. Jones et al. (2014) used the OSD CAPE draft estimating guide when finding the ratio of O&S costs to development costs and applied services lives of 20-30 years for fighters and helicopters and 30-40 years for cargo, bomber and tanker aircraft. Given these two determinations of service life, separate analyses are conducted for up to 30 years from IOC and 30-40 years from IOC. This has the added benefit of examining potential aging effect in the data. The base data set has 765 programs and year pairs (e.g. F-15A in 1998 or F-22A in 2008), while the truncated set has 681, which means that only 10.98 per cent of the data points are removed because of truncation.

The degree of stability is determined through testing per cent bounds. The process starts at a 20 per cent bound and finds the first year from IOC in which the mean per cent difference crosses this threshold. That point is then declared the “year from IOC” that stability occurs. The duration of time it stays within the 20 per cent bound is calculated from the stability point to 30 years and the stability point to 40 years. If the mean per cent difference falls within the 20 per cent bound for more than 80 per cent of the time from the stability point to 30 years, then the same process is repeated with a 15 per cent bound. The process is repeated for the 10 and 5 per cent bounds if applicable, with the most restrictive bound being the best point of stability. This process is completed for all 18 cost combinations comprising the three cost types (total costs, CPFH and CPTAI) and six cost categories (Total O&S and CES 1-5). Again, CES 6 is used in the calculation of total O&S but not individually analyzed. The per cent stable to 40 years is not used in the selection of the best bound, but a decrease in per cent of the time stable can reveal potential aging affects.

After the best bound selection, there is a stability point in terms of years from IOC and a per cent bound. Next, descriptive statistics are calculated for the mean per cent differences from the stability point to 30 years from IOC. Using the best bound and descriptive statistics, it can be determined which cost type and which categories exhibit the most or least amount of stability. Understanding these steady-state properties, therefore, provides an insight to cost analysts on the proper time to migrate from the analogy technique to an extrapolation from actual method. The bounds also influence the cost risk modeling of the individual elements when developing a cost estimate. Descriptive statistics for 30-40 years from IOC are also calculated to highlight any aging affects.

This process is then repeated for each aircraft type. Aircraft are divided into the eight operational mission categories given by AFTOC: bomber, fighter/attack, helicopter, reconnaissance, special duty, training, transport/tanker and unmanned aerial vehicle (UAV). It is important to note that splitting the 44 programs into eight categories drastically reduces the number of data points for each aircraft type. Because of the small sample size, inferential tests are not conducted to demonstrate differences in aircraft type, but descriptive statistics can still provide insights. For simplicity, comparisons by aircraft category use the best metric for stability found at the aggregate level (i.e. CPFH, CPTAI or total costs).

The second part of this research creates a top-level regression model as a crosscheck to the cost analyst’s primary estimation techniques for O&S costs. The dependent variable in this regression analysis is the total CPTAI. AFTOC provides additional programmatic information in addition to the cost data that are used to create independent, predictor variables. Examples of programmatic data provided in the database include average age of aircraft, location of lead logistics center, operational mission type and unit-cost, among others, discussed in the next section. Stability analysis results are also incorporated into the model as an additional predictor variable for the point of stability in terms of years from IOC.

Regression analysis requires a re-screen of the initial data set because of the addition of variables. Each program and “years from IOC” pair must have only one value for each potential predictor variable. This screen resulted in five grouped programs that have two different unit-costs and one grouped program with two different logistics centers. C-141B/C, KC-135E/R, MH-53J/M, T-38A/C and UH-1H/N are all programs that had major overhauls at some point in their service life that created a new MDS and an updated unit cost. Because of the programs being previously grouped by IOC year, these five are removed from regression analysis for having two unit-costs. The program with two logistics centers is the EC-130E/H and is included in the regression data set because location of logistics center is a binary variable that may influence the O&S cost. Table IV illustrates the re-screen of the base data set.

Prior to any model building, the re-screened regression database is separated into two groups: the model building data set and validation set. For multiple regression, a random sampling of the data uses approximately 80 per cent for the model building portion and 20 per cent for the validation set. Variables may or may not be included in the model depending on the individual significance and contingent on passing regression diagnostics. The mixed stepwise function in JMP Pro Version 12 is used to determine which predictor variables are initially included in the model. A significance level of 0.05 is the threshold to enter or exit the model. The assumption of normality is assessed using the Shapiro–Wilk test, while the assumption of constant variance is assessed with the Breusch–Pagan test. Both tests are conducted at the 0.05 level of significance. Failing to satisfy the assumption of constant variance can potentially be remedied with natural log transformation of the dependent and/or independent variables. The natural log transformation results in the predicted model value being the median response rather than a mean.

Furthermore, multicollinearity, influential data points and outliers are investigated to prevent additional bias. Variance inflation factors (VIFs) highlight linear relationships between two variables and a VIF higher than 5 indicates multicollinearity. Cook’s distance detects influential data points that could be skewing the model, and any value greater than 0.5 is investigated thoroughly. Any studentized residual that is greater than three standard deviations from the mean is considered an outlier and must also be investigated further. The final assessment is the Bonferroni–Holm correction, which aims to control the family-wise error rate and reduces Type I error (false positives).

Once all diagnostics are passed, the model building set is validated against the validation set using multiple criteria: mean absolute percentage error (MAPE), median absolute percentage error (MdAPE) and adjusted R2. When the model is deemed internally valid, then all data points are combined to update the final model using the variables selected from the model building process.

Results

Stability analysis

The stability analysis begins by examining the mean per cent differences for CPTAI at each year from IOC, up to 40 years (Figure 1). The number of programs used in the calculations of mean per cent difference for a specified “years from IOC” range from 8 to 24. The table displays total CPTAI and the five O&S CESs CPTAI. Mean values are color coordinated from red to green at 5 per cent bins, where dark red is equivalent to greater than 20 per cent difference and dark green is equivalent to under 5 per cent difference. Tables for total O&S cost and CPFH are attached in the Appendix and demonstrate similar characteristics. Note that CES 4 and 5 do not fall under the 20 per cent threshold at any point and are, therefore, removed from further analysis. While it may seem capricious to remove two of the five categories, O’Hanlon (2018) determined that CES 1, 2 and 3 constitute the majority of O&S costs with a mean of 82.39 per cent and median of 80.53 per cent.

The best bound selection process and descriptive statistics for all combinations (minus CES 4 and 5) are summarized in Table V. Looking at the table, several conclusions can be drawn from the data. For instance, CES 1 (Manpower) and 2 (Unit operations) have much lower means and medians than CES 3 (Maintenance). This indicates that manpower and pay are fairly stable. Operation tempo and fuel consumption is usually planned in advance and may contribute to unit operations’ costs having a more consistent nature as well. Maintenance, especially unplanned maintenance, is much harder to predict, and that may explain why less stability is exhibited in CES 3 across all cost combinations. The cost type (total, CPFH or CPTAI) that has the lowest best bounds is CPTAI which displays lower means and medians than the other two cost categories. In addition to the conclusions across cost type and cost categories, there appears to be aging affects in total costs and CPTAI but not for CPFH combinations. Aging can be seen when the per cent stable to 40 years from IOC is less than per cent stable to 30 years from IOC. While it does not show the direction in which costs are going, it does show that there is less stability later in the sustainment phase. In addition, the mean and median values for CPFH are higher, which indicate instability.

The last segment of stability analysis searches for differences across aircraft categories. The sample size to calculate means drops sizably when splitting the 44 programs into eight different categories, which limits the fidelity of inferential statistics. In addition, not all aircraft platform types have data points for all years from IOC in between 1 and 40. For this reason, bounds testing and descriptive statistics provided for the various aircraft platform types are catered to the available data of the aircraft type. CPTAI was the best metric in the aggregate analysis, with stability demonstrated at the 10 per cent bound, a trend that appears the same across all platform types. For this reason, the research hones in on CPTAI as the basis for category comparison. Besides helicopters, which has its first year of data 11 years from IOC, all aircraft types exhibit stability properties till five years from IOC. As shown in Table VI, there are potentially three distinct groupings for the means of per cent differences after stability occurs. Bombers, fighter/attack, training and transport/tanker have total CPTAI means around 7 per cent and medians from 5 to 7 per cent. Reconnaissance and helicopters are at 8.55 and 9.55 per cent, respectively, with medians around 7 per cent. Special duty aircrafts and UAVs appear to be different from the rest in the total CPTAI category with means around 11-12 per cent and medians from 10 to 12 per cent.

Multiple regression analysis

The second portion of analysis creates a regression model by utilizing knowledge from the aforementioned CBO/RAND study and builds upon it with additional programmatic information from AFTOC and results from the stability analysis of this research. The dependent variable in the model is the natural log of total O&S CPTAI. When running diagnostics, the Breusch–Pagan test for constant variance repeatedly failed; however, natural log transformation of CPTAI and two predictor variables provided a more constant variance of residuals. The initial independent variable set to be investigated is shown in Table VII.

Of the independent variables tested in the model building set, two binary variables were removed for insignificance during the step-wise process at the 0.05 level: contract vs organic logistics support and all binary variables for years from IOC. Of the remaining 13 predictor variables, all exhibited individual significance below the comparison-wise error rate of 0.05/13 = 0.00384 for the Holm–Bonferroni correction. For the 80 per cent model building set, the MAPE is 0.0131 (1.31 per cent) and the MdAPE is 0.0095 (0.95 per cent). In the 20 per cent validation set, MAPE is 0.0108 (1.08 per cent) and MdAPE is 0.0071 (0.71 per cent). The adjusted R2 values for the model and validation sets are 0.8878 and 0.902, respectively. The close similarities in absolute percentage error and adjusted R2 values suggest a valid model.

Combining all data points, the updated model has an R2 value of 0.8901, which means that the model explains 89.01 per cent of the variance in the data set when predicting median total CPTAI. The model’s largest VIF is 6.60 for the binary variable representing Warner Robins as the lead logistics center. All remaining values are below 5; however, this indicates there may be some minor multicollinearity in the model. The highest Cook’s D is 0.367, which is below the 0.5 threshold for influential data points. The Shapiro–Wilk test fails the normality test with a p-value of <0.001, and the Breusch–Pagan test fails the test for constant variance with a p-value of <0.001. Though the tests fail statistically, a more in-depth examination of the studentized residuals shows that values centered around the mean and the residual vs predicted plot show no apparent trend. Ordinary least squares regression is also robust against minor or moderate deviations from constant variance and normality (Cohen et al., 2003). Natural log transformation of the response variable as well as two predictor variables, unit-cost and tempo, provided a more constant variance of residuals. Final regression results are shown in Table VIII. The MAPE for the final model is 0.0126 (1.26 per cent) and MdAPE is 0.00913 (0.91 per cent), which indicates the model has a low relative error when predicting the natural log of total CPTAI.

Discussion and conclusion

This research determines to ascertain where stability occurs in O&S costs, and how it is exhibited using a variety of metrics. Overall, we find that the CPTAI metric is the best metric when seeking to determine stability. The cost type that exhibits less stability is the CPHF metric, which could be because of the high variance in flight hours and differences in mission types. Recall that the OSD-CAPE CES (Table III) is the structure used by analysts in constructing their estimates. Analysis of the CES illustrates that manpower (CES 1) and unit operations (CES 2) exhibit the highest degree of stability, while maintenance (CES 3) is the lowest. At the top level, CPTAI reaches stability five years from IOC and coincides with Dixon’s (2005) research on commercial aircraft exiting the newness phase when average aircraft age is five-six years. Dixon’s research is standardized by average age and CPHF but it presents similar results to our analysis. Our categorical analysis (using CPTAI as the metric) finds two-three distinct groupings, with stability typically occurring from 0 to 5 years from IOC. Practical applications of these findings for DoD cost analysts and decision-makers is straightforward. Programmatic decisions, such as the aforementioned U-2 vs Global-Hawk decisions, are based partially on out-year sustainment costs. Understanding the stability properties of these costs provides an increased confidence in the estimates provided and the decisions upon which they are based.

In addition to analyzing stability properties, this research develops a robust regression model to predict median total O&S costs for aircraft. It is important to note that this tool is intended to be used as a crosscheck in conjunction with primary methods when creating estimates. The model has an R2 of 0.8901, which is a significant explanation of the variance in the data. Table IX displays the relative contribution for the variable types as determined by standardized beta estimates. Aircraft type, location of lead logistics center and unit cost are the three largest contributing factors to the model and comprise over 74 per cent of the relative weights in the total model. Standard practice is for analysts to use a multitude of techniques to enhance confidence in the estimates provided. The model developed here serves as another technique for the analyst’s toolkit.

There are several limitations to this research. Perhaps the most significant limitation in determining stability is that the research standardizes by years from IOC. In theory, it is a reasonable way to standardize, but in reality, programs distinguish initial capability differently. Programs may declare IOC earlier than when they should because of schedule or other pressures. In other cases, programs incur O&S costs for multiple years before declaring IOC which skews the data to show stability sooner. This limitation is well documented in other works such as Kozlak et al. (2017) and Jimenez et al. (2016), who also used IOC dates as a standardization point. A second limitation with respect to IOC is that the year is used rather than the month for IOC dates for simplicity reasons. Finally, our investigation is limited to only Air Force aircraft. While the stability results coincide with Dixon’s (2005) research with commercial aircraft, it is impossible to tell if these characteristics are the same as those of other service aircraft or other types of programs.

There are multiple areas where future research is warranted. Stability properties of other military services weapon systems could be investigated using the methodology of this paper. The visibility and management of operating and support costs database contains the necessary O&S costs for other services. Additionally, an investigation into CES 6, indirect costs, to develop an allocation model would allow for future researchers to use this CES in their analysis. Other future research could examine O&S cost property behavior in times of war in comparison to peacetime.

The new emphasis placed on understanding O&S costs from the WSARA legislation has resulted in a nascent research stream that is growing. This research adds to the body of knowledge by providing cost analysts and decision-makers with tools to help determine when Air Force aircraft can be considered to have stabilized O&S costs. A determination of stability allows for better budgeting practices because cost analysts can then switch from using analogous programs to extrapolating from actual data. In addition, we provide a new parametric model that predicts median O&S costs for programs in varying stages in the life cycle. Convergence around an estimate from multiple techniques provides confidence in the decision upon which they are based.

Figures

Mean per cent difference for CPTAI

Figure 1.

Mean per cent difference for CPTAI

Mean per cent difference (total O&S)

Figure A1.

Mean per cent difference (total O&S)

Mean per cent difference CPFH

Figure A2.

Mean per cent difference CPFH

Initial data set screening

Screen Remaining MDS
AFTOC 274
TAI >10 146
MDS years data > 5 110
Available IOC date 57
  Programs
Grouped by IOC Date 44

Programs selected by aircraft type

Fighter/attack Reconnaissance Helicopter
A-10C
A/OA-10
F-117A
F-15A
F-15B
F-15C
F-15D
F-15E
F-16A
F-16B
F-16C
F-16D
F-22A

Bomber
B-1B
B-2A
B-52H
E-3B/C
E-8C
U-2S

Special Duty
AC-130H/U
EC-130E/H
EC-130J
MC-130E/H
MC-130P

Training
T-1A
T-38A/C
T-6A

UAV/Drone
MQ-1B
MQ-9A
RQ-4B
HH-60G
MH-53J/M
UH-1H/N

Transport/Tanker
C-130E/H
C-130J
C-141B/C
C-17A
C-21A
C-5A/B
C-5M
CV-22B
HC-130P/N
KC-10A
KC-135E/R

OSD CAPE CES

CES Name Description
1.0 Unit-level Manpower Cost of operators, maintainers, and other support manpower assigned to operating units. May include military, civilian, and/or contractor manpower
2.0 Unit operations Cost of unit operating material (e.g. fuel and training material), unit support services, and unit travel. Excludes material for maintenance and repair
3.0 Maintenance Cost of all system maintenance other than maintenance manpower assigned to operating units. Consists of organic and contractor maintenance
4.0 Sustaining support Cost of system support activities that are provided by organizations other than the system’s operating units
5.0 Continuing system improvements Cost of system hardware and software modifications
6.0 Indirect support Cost of support activities that provide general services that lack the visibility of actual support to specific force units or systems. Indirect support is generally provided by centrally managed activities that provide a wide range of support to multiple systems and associated manpower

Rescreening of base data set for regression analysis

Screen Programs Observations remaining
Base data set 44 892
Duplicate unit costs 39 708
Has value for each variable 39 609

Summary statistics (excluding CES 4 and 5)

Descriptor O&S Total CES 1 CES 2 CES 3 CPFH total CPFH CES 1 CPFH CES 2 CPFH CES 3 CPTAI Total CPTAI CES 1 CPTAI CES 2 CPTAI CES 3
Stable after (Years) 5 7 9 5 3 3 2 8 5 8 6 5
Bound selected (%) 15 15 15 20 15 20 20 20 10 10 15 20
Stable-30                        
Mean (%) 10.24 9.75 13.27 18.30 11.44 13.39 12.42 19.43 8.44 8.15 12.63 17.68
Std. dev (%) 3.64 3.53 5.41 6.31 3.20 4.38 4.90 6.06 1.89 1.89 4.32 5.80
Median (%) 9.31 9.09 12.49 17.69 10.79 12.16 11.80 18.35 8.71 8.06 12.03 17.06
% stable to 30 92.00 95.65 85.71 72.00 85.19 88.89 85.71 63.64 80.00 86.36 83.33 68.00
30-40  
Mean (%) 13.05 12.21 16.37 18.62 9.55 13.45 8.77 17.36 9.10 7.99 13.27 16.30
SD (%) 3.21 4.09 3.34 3.57 3.39 3.90 2.04 5.21 2.55 3.05 2.81 5.09
Median (%) 13.58 13.68 18.28 20.63 8.98 12.99 8.56 15.49 8.67 8.95 12.76 15.90
% stable to 40 82.86 87.88 70.97 65.71 86.49 89.19 89.47 68.75 74.29 81.25 79.41 37.14

Total CPTAI summary statistics (stable to 30 years from IOC) by aircraft type

Aircraft type Years to stability Best bound (%) Mean (%) SD (%) Median (%)
Training 1 10 7.09 5.44 5.85
Transport/ tanker 5 10 7.17 4.26 6.68
Bombers* 0 15 7.27 5.61 5.97
Fighter/attack 2 10 7.42 2.93 6.94
Reconnaissance 3 15 8.55 6.84 6.94
Helicopters 10 15 9.55 6.34 7.77
Special duty** 0 20 11.26 7.67 10.00
UAV 5 15 12.55 5.08 12.02
Notes:
*

When tested to a 10% bound, bombers had 70.00% stability;

**

when tested to 15% bound, special duty had 76.67% stability

Predictor variables analyzed in regression analysis

Name Type of variable(s) Description Source
Unit cost Continuous Unit cost of the aircraft for the program AFTOC
Tempo Continuous Total flight hours in a year divided by the TAI AFTOC
Average age Continuous Average age of all aircraft in the program at the end of the year AFTOC
Location of lead logistics center Binary Location of the lead air logistics center (ALC). The base case is Ogden Logistics Center (OO-ALC). Binary variables are created for Warner Robins Air Logistics Center (WR-ALC), Aeronautical Systems Center (ASC), and Oklahoma City Logistics Center (OC-ALC) AFTOC
Contract vs organic logistics support Binary Describe if the program is currently being maintained by contract logistic support (CLS) or organic logistics support AFTOC
Mission type Binary There are eight mission types as delineated in AFTOC: Bombers, Fighter/Attack, Helicopters, Reconnaissance, Special Duty, Training, Transport/Tanker, and UAV. The base case in the research is Transport/Tanker AFTOC
Years from IOC Binary These variables include years 1 through 30 from IOC to determine if any specific years may be significant SARs, Deagel.com, Af.mil
Stable vs unstable Binary Derived from the stability portion of the research and indicates whether the program is considered stable or not. Metric used is 5 years from IOC for CPTAI Findings from stability analysis

Final ordinary least squares model

Predictor variable Estimate p-value Standardized estimate V I F
Intercept 3.389 <0.0001 N/A N/A
LN unit cost 0.407 <0.0001 0.634 2.02
LN tempo (Fly hours/TAI) 0.736 <0.0001 0.475 3.39
Average age 0.018 <0.0001 0.232 1.98
Fighter/attack 0.514 <0.0001 0.230 4.71
Bomber 0.802 <0.0001 0.270 3.00
Special duty 0.304 <0.0001 0.131 1.61
Reconnaissance 0.690 <0.0001 0.234 1.61
Helicopter 0.386 <0.0001 0.079 1.34
Trainer −0.572 <0.0001 −0.149 2.29
Stable −0.188 <0.0001 −0.072 1.41
WR-ALC (Warner Robbins) 0.792 <0.0001 0.465 6.60
ASC (Wright-Patt) 0.599 <0.0001 0.255 4.47
OC-ALC (Oklahoma City) 0.503 <0.0001 0.260 5.11
WR-ALC * Fighter/attack interaction −0.611 <0.0001 −0.166 1.84

Predictor variables (grouped) and relative contribution to model

Predictor variables (Grouped) Relative contribution (%)
LN unit cost 17.4
LN tempo 13.0
Average age 6.4
Stable 2.0
Logistics center 26.8
Aircraft type 29.9
Interaction term 4.5

Appendix

Figure A1

Figure A2

References

Boito, M., Light, T., Mills, P. and Baldwin, L. (2012), Managing US Air Force Aircraft Operating and Support Costs (RR-1077), RAND Corporation, Santa Monica, CA.

Christensen, D. and Payne, K. (1992), “Cost performance index stability: fact or fiction?”, Journal of Parametrics, Vol. 12 No. 1, pp. 27-40.

Cohen, J., Cohen, P., West, S.G. and Aiken, L.S. (2003), Applied Multiple Regression/Correlation Analysis for the Behavioral Sciences, 3rd edition. Lawrence Erlbaum Associates, Mahway, NJ.

Congress US (2009), “Weapon systems acquisition reform Act of 2009”, Public Law, 111-23.

Congressional Budget Office (CBO) (2001), The Effects of Aging on the Costs of Operating and Maintaining Military Equipment, Government Printing Office, Washington DC.

Dixon, M. (2005), The Costs of Aging Aircraft; Insights from Commercial Aviation (MG-486), RAND Corporation, Santa Monica, CA.

Government Accountability Office (2009), Gao Cost Estimating and Assessment Guide (09-3SP), United States Government Printing Office, Washington, DC.

Hildebrandt, G.G. and Sze, M. (1990), An Estimation of USAF Aircraft Operating and Support Cost Relations (N3-062), RAND Corporation, Santa Monica, CA.

Jimenez, C.A., White, E.D., Brown, G.E., Ritschel, J.D., Lucas, B.M. and Seibel, M.J. (2016), “Using pre-milestone B data to predict schedule duration for defense acquisition programs”, Journal of Cost Analysis and Parametrics, Vol. 9 No. 2, pp. 112-126.

Jones, G., White, E.D., Ryan, E.T. and Ritschel, J.D. (2014), “Investigation into the ratio of operating and support costs to life-cycle costs for DoD weapon systems”, Defense Acquisition Research Journal, Vol. 21 No. 1, pp. 442-464.

Kozlak, S.J., White, E.D., Ritschel, J.D., Lucas, B.M. and Seibel, J.M. (2017), “Analyzing cost growth at program stages for DoD aircraft”, Defense Acquisition Research Journal, Vol. 24 No. 3, pp. 386-407.

Mislick, G.K. and Nussbaum, D.A. (2015), Cost Estimation: Methods and Tools, John Wiley & Sons, Hoboken, NJ.

Office of the Secretary of Defense, Cost Assessment and Program Evaluation (OSD CAPE) (2014), “Operating and support cost-estimating guide”, Washington, DC, available at: www.cape.osd.mil/files/OS_Guide_v9_March_2014.pdf

O’Hanlon, G. (2018), “An analysis of aircraft operating and support cost element structures and their contribution to total costs (AFIT-ENV-MS-18-M-227)”, Master of science thesis, Air Force Institute of Technology, Department of Systems Engineering and Management, Wright-Patterson AFB, OH.

Petter, J.L., Ritschel, J.D. and White, E.D. (2015), “Stability properties in department of defense contracts: answering the controversy”, Journal of Public Procurement, Vol. 15 No. 3, pp. 341-364.

Pomerleau, M. (2016), “Global Hawk takes another step toward replacing the U2”, Defense Systems, available at: https://defensesystems.com/articles/2016/02/25/global-hawk-sensors-replacing-u2-spy-plane.aspx (accessed 11 January 2018).

Ritschel, J.D. and Ritschel, T.L. (2016), “Organic or contract support? Investigating cost and performance in aircraft sustainment”, Journal of Transportation Management, Vol. 27 No. 2, pp. 47-58.

Ryan, E., Jacques, D., Colombi, J. and Schubert, C. (2012), “A proposed methodology to characterize the accuracy of life cycle cost estimates for DoD programs”, Procedia Computer Science, Vol. 8, pp. 361-369.

Ryan, E., Jacques, D., Ritschel, J. and Schubert, C. (2013), “Characterizing the accuracy of DoD operating and support cost estimates”, Journal of Public Procurement, Vol. 13 No. 1, pp. 103-132.

Acknowledgements

Disclaimer: The views expressed in this article are those of the authors and do not reflect the official policy or position of the United States Air Force, Department of Defense or the United States Government.

Corresponding author

Jonathan D. Ritschel can be contacted at: ritschel1@msn.com

About the authors

First Lt. Scott C. Hewitson is a Cost Analyst at the Air Force Cost Analysis Agency. First Lt. Hewitson graduated with a Bachelor of Science (BS) in Business Management from the Air Force Academy and a Master of Science (MS) in Cost Analysis from the Air Force Institute of Technology (AFIT) in 2018.

Jonathan D. Ritschel, PhD, is an Assistant Professor of Cost Analysis in the Department of Systems Engineering and Management at AFIT. He received his BBA in Accountancy from the University of Notre Dame, MS in Cost Analysis from AFIT and PhD in Economics from George Mason University. Jonathan Ritschel’s research interests include public choice, cost analysis and economic institutional analysis.

Edward White is a Professor of Statistics in the Department of Mathematics and Statistics at AFIT. He received his BS in Mathematics from the University of Tampa, MAS from The Ohio State University and PhD in Statistics from Texas A&M University. His primary research interests include statistical modeling, simulation and data analytics.

Captain Gregory Brown, USAF, is the cost chief for Special Operations Forces and Personnel Recovery Division, Air Force Life Cycle Management Center, Wright-Patterson Air Force Base, Ohio. He received a BA in Economics and a BS in Business-Finance from Colorado State University and an MS in Cost Analysis from the Air Force Institute of Technology.

Related articles