71 Fed. Cl. 571 | Fed. Cl. | 2006
OPINION AND ORDER
This post-award bid protest was before the Court on cross-motions for judgment on the administrative record. Plaintiffs Fort Carson Support Services (“FCSS”) and The Ginn Group, Inc. (“TGG”) separately challenged the award of a contract to intervenor KIRA, Inc. (“KIRA”). For the reasons that follow, the Court DENIES the motions of FCSS and TGG, and GRANTS the cross-motions of the government and KIRA.
I. BACKGROUND
A. The Solicitation
On April 13, 2004, the United States Army issued solicitation number W911RZ-04-R-0002 (“Solicitation”), requesting proposals for a contract to perform base operations and maintenance services for the Directorate of Public Works at Fort Carson, Colorado. See AR Tab 7. This negotiated procurement was for a cost-plus contract, covering a phase-in and one-year base period and nine one-year option periods. See id. § F.2. As was explained in the Solicitation, the award was to be made based on the best value to the government. Offerors were informed:
This decision will not be made by the application of a predefined formula, but by the conduct of a tradeoff process among the factors, subfactors, and elements identified below and by the exercise of the sound business judgment on the part of the Source Selection Authority_ Proposals must conform to solicitation requirements and will be judged by an integrated assessment as being most advantageous to the Government, cost and other factors considered.
Solicitation § M.1.1, AR Tab 7 (emphasis added).
The Army further explained that “cost is not the primary consideration,” and that it “reserve[d] the right to award to other than the lowest proposed cost or to other than the
Offerors are advised that 'primary consideration will be given to the evaluation of the Quality Factor, rather than to the Cost Factor. The Government will determine the best overall value based on an offeror’s proposal, including the Quality and Cost Factors and the risks associated with successful contract performance. The ultimate decision for award will be derived subjectively in the event that the highest quality proposal is not also the lowest cost. In that event, the Source Selection Authority will make a determination, based on the best interests of the Government, whether the differences in quality are sufficient to offset differences in costs.
Id. (emphasis added).
The two evaluation factors for the source selection decision were “Quality and Cost, with Quality being significantly more important than Cost.” Id. § M.2.1 (emphasis added). The Quality factor was composed of two subfactors, Management and Past Performance and Experience (“PPE”), which were “of approximately equal importance.” Solicitation § M.2.1.1, AR Tab 7. The Management subfactor was made up of five elements: the Management and Technical element, which was “substantially more important than any of the other elements,” id.; the Contract Administration, Human Resources Management, and Quality Control elements, which were “of approximately equal importance,” id. and the Phase-In Planning element, which was substantially less important than any of the others. Id. The four elements of the PPE subfaetor — Performance, Managerial, Business Relations, and Cost Control— were “of approximately equal importance.” Id.
The Solicitation further provided that the Quality factor evaluation would be “a composite evaluation using all of the subfactors and elements associated with this factor,” in which “[t]he Management subfaetor and each of its elements will receive an overall adjectival and proposal risk assessment rating,” and “[t]he PPE subfactor and each of its elements will receive a performance risk rating.” Id. § M.2.2. The Solicitation explained, however, that the rating descriptions provided “may not precisely describe the assessment of a specific factor, subfactor, or element. In that case, evaluators will assess the rating with the description most closely describing the actual assessment. ” Id. § M.3 (emphasis added). Thus, the source selection decision process was not geared toward the generation of a particular label to place on the quality of each subfactor and element, but employed labels to assist in the actual assessment of the quality of the proposals.
The adjectival ratings for the Management subfactor and its elements ranged from Outstanding, Good, Satisfactory, and Marginal to Unsatisfactory, and each rating was given a full description in the Solicitation. Id. § M.3.1. A “Good” rating, for instance, was merited when a “proposal offers some advantages to the Government as well as fully meeting the requirements of the RFP,” and described an approach that “has a high probability of meeting or exceeding the requirements.” Id. A “Satisfactory” rating “indicated that the proposal offers no particular advantages to the Government but does fully meet the ‘minimum’ requirements of the RFP,” and was associated with an approach having “an acceptable probability of meeting the requirements.” Id. An “Outstanding” rating meant, among other things, that “the proposal offers significant advantages to the Government” and had “an excellent probability of exceeding the requirements,” while a “Marginal” rating meant the finding of “minor omissions or misunderstandings of the requirements of the RFP” and “a low probability of meeting some of the requirements of the RFP. ” Id. “Unsatisfactory” was reserved for proposals with “major omissions and/or misunderstandings,” which have “a high probability of not being able to meet a significant number of the requirements of the RFP.” Id.
The risk ratings for the Management sub-factor and its components were based on the amount of “disruption of schedule, increase in cost, or degradation of performance” expected, as well as the amounts of “contractor emphasis” and “Government monitoring” that would be required. These ratings ran
Three risk ratings were used for the PPE subfactor, based on the amount of doubt that an “offeror will successfully perform the required effort.” Id. § M.3.3. A “Low” rating meant little doubt, a “Moderate” rating some doubt, and a “High” rating substantial doubt. Id. If an offeror lacked an identifiable performance record, it would instead receive the neutral “Unknown” rating. Id. Regarding PPE, the Solicitation explained that “more similar PPE, in terms of type and scope of work, will be given greater credit than less similar.” Id. § M.2.2.2. The PPE was to be “evaluated by examining both the offeror’s proposal and questionnaires provided to the Contracting Officer by references selected by the offeror.” Id. The Army was also to “consider the breadth and depth of experience reflected in the offeror’s list of work accomplished,” and would consider information obtained from references, from firms or individuals connected to the list of work supplied by the offeror, and from “other relevant sources available to it.” Id.
As was noted above, the Cost factor was significantly less important than the Quality factor. Offerors were told that this factor “will be assessed for completeness, reasonableness, and realism, but will not be assigned an adjectival or risk rating.” Solicitation § M.2.3, AR Tab 7. The reasonableness assessment entailed a comparison of proposed costs to an Independent Government Estimate (IGE). Id. § M.2.3.2. An assessment of realism considered, among other things, whether proposed costs were “consistent with the unique methods of performance and materials described in the offeror’s Management subfactor proposal.” Id. § M.2.3.3. Sounding this same theme, the Solicitation’s discussion of the Management and Technical element of the Management subfactor explained that “[t]he Government will consider whether the proposal provides any innovative ideas that could improve services, enhance customer service, and/or reduce costs.” Id. § M.2.2.1.1. Offerors were specifically warned, albeit parenthetically, of the interplay between the Management subfactor and the cost realism aspect of the cost analysis: “Deficiencies or weaknesses in cost realism may result from comparable deficiencies or weaknesses in the Management subfactor proposal and may, thus, also result in lower adjectival and higher risk ratings] in the evaluation of that subfactor.” Id. § M.2.3.3.
The evaluation of the Cost factor involved the consideration of two subfactors, Total Estimated Cost (TEC) and cost realism. The former “includes an assessment of the offeror’s total estimated cost and fee,” an evaluation that “will include fairness and reasonableness of the Total Estimated Cost based on the competition associated with this acquisition.” Id. § M.2.3.5. Concerning cost realism, the Solicitation provided:
Cost realism includes the evaluation of variance, balance and allocation. The Government will develop a Most Probable Cost (MPC) based on the Government’s cost realism analysis. The MPC will be determined by adjusting each offeror’s proposal to reflect any additions or reductions in costs to realistic levels based on the results of the Government’s cost realism analysis. The MPC may differ from the proposed cost and will reflect the Government’s best estimate of the cost that is most likely to result from the offeror’s proposal. An of-feror’s eligibility for award may be significantly downgraded if the offeror’s proposed costs and technical approach result in significant cost realism adjustments.
Id. § M.2.3.6 (emphasis added).
Each of the three aspects of cost realism— variance, balance, and allocation — was to be “evaluated as either satisfactory or substandard.” Id. §§ M.2.3.6.1-M.2.3.6.3. Variance was defined as the difference between a proposal’s TEC, claimed by the offeror, and the MPC, as determined by the Army. Offerors were informed that “consideration [will be] given to the percentage of the variance, the dollar amount of the variance, the anticipated impact on contract performance, and the of-feror’s proposed method of performance.” Id § M.2.3.6.1. Balance was to be based on whether “cost is underestimated in some areas or periods and overstated in others,” id. § M.2.3.6.2, and allocation considered the
B. The Source Selection Plan
The Source Selection Plan (SSP) specified that the Source Selection Authority (SSA), Mr. Steve J. McCoy, would appoint a Source Selection Evaluation Board (SSEB) to assist him in the evaluation of proposals. SSP, AR Tab 34 at 6. The SSEB consisted of a Chairman, a Quality Committee, a Cost and Past Performance Committee, the Contracting Officer, the Contract Attorney, and technical advisors. See Attachment E to SSP, AR Tab 34 at 38; see also AR Tab 31 (SSEB appointment letters). While all members of the SSEB were to read the Solicitation and the SSP, it appears that only the individual members of the Quality Committee were charged with reading and evaluating each proposal to recommend ratings for the Management sub-factor and its elements. SSP § III.B(2), AR Tab 34 at 9.
The SSEB will assign adjectival ratings to the elements, and based upon the relatively equal importance of the elements, assign an overall adjectival rating to the overall subfactor. Ratings must be accompanied by a consistent narrative assessment, inclusive of strengths and weaknesses that describes the basis for the assigned ratings.
Id. The proposal risks for the Management subfactor and its elements were to be “assess[ed] and assign[ed]” at this same time. SSP § III.B(3), AR Tab 34 at 9. The SSP also required, for this subfactor and elements, that “[a]ll evaluations shall be supported with detailed rationale.” SSP § III.B(4), AR Tab 34 at 9. The SSP used the same definitions of adjectival and risk ratings as were contained in Solicitation section M, which was also included as attachment C to the SSP. Compare SSP § III.B(5)-(6), AR Tab 34 at 10-11 with Solicitation §§ M.3.1-M.3.2, AR Tab 7 at 52-53; see also Attachment C to SSP, AR Tab 34 at 18.
Concerning the Past Performance and Experience evaluation, SSEB members were to “[djocument [the] individual assessments, conduct a consensus evaluation and prepare a summary report for the [Contracting Officer].” SSP § III.C(2), AR Tab 34 at 11. The performance risk definitions from the Solicitation were repeated in the SSP. See id. § III.C(3) at 11. The SSEB’s Cost factor evaluation language fairly closely tracked the method explained in the Solicitation. Compare SSP § III.D, AR Tab 34 with Attachment C to id.
The SSP explained that SSEB members were to “[participate in group discussions to arrive at consensus ratings of each proposal,” and emphasized that these ratings “shall not be mathematically computed based on the individual evaluations, but rather shall be established by agreement among the SSEB members.” SSP § II.B(5)(d), AR Tab 34 at 8. The SSEB’s consensus evaluation was to be contained in a Proposal Analysis (PA) Report. See id. § II.B(5)(c) at 8. The PA Report was described as a “formal report that fully documents the results of the [SSEB], using objective appraisals and ratings of the Quality and Cost proposals.” Id. § I.F(8) at 4. The “ratings and appraisals” in the PA Report were required to “be the consensus agreement of all SSEB members.” Id. The SSP further explained that one of the
Prepares and submits a comprehensive PA (Proposal Analysis) Report of the results of the evaluation to include a summary; adjectival ratings of the Management sub-factor along with a statement of the specific deficiencies, strengths, and weaknesses with respect to the overall subfactor and each element; the proposal and performance risk assessments; and overall recommendations to the [Contracting Officer]. The report shall be supported by the individual SSEB members’ evaluations and other reports/write-ups on the evaluation process.
Id. § II.B(4)(h) at 7.
As was provided in the Solicitation, the SSA was ultimately responsible for the award decision, as well as for “document[ing] the supporting rationale.” Id. § II.B(l)(f) at 6. According to the SSP, the proposals were to be evaluated by the SSA by assessing the Management subfactor and its five elements, the PPE subfactor and its four elements, and the Cost factor. Id. § IV.A(1). The cost factor assessment was said to involve an evaluation for completeness, reasonableness, and realism. Id. In addition to again omitting any reference to the Total Estimated Cost subfaetor from section M of the Solicitation, see Solicitation § M.2.3.5, AR Tab 7 at 51, the SSP contained this erroneous instruction to the SSA: “Within the Quality factor, the Management subfactor is considered substantially more important than the Past Performance and Experience subfactor. Within both subfactors, the elements are of roughly equal importance.” SSP § IV.A(2), AR Tab 34 at 14.
The SSA shall utilize all evaluation reports and analysis in determining the successful proposal pursuant to the basis for award language shown in both FAR Provision 52.215-1 Instruction to Offerors — Competitive and FAR Part 15.305 Proposal Evaluation. In making the award determination, the SSA can draw on his business judgment and personal knowledge.
Id. § IV.B(l) at 14.
C. The Initial Evaluation
Six proposals were submitted in response to the Solicitation. See AR Tab 76 at 13. After the SSEB performed its individual and consensus evaluations, and briefed the SSA,
The initial award thus came down to a consideration of the Management subfactor evaluations and the cost of each proposal. The SSA found that FCSS’s “proposal re
These two offerors with the lowest-rated Quality factor also proposed the lowest cost. The TEC of FCSS was fourth lowest, just $[XXXX] higher than The Ginn Group’s over ten years, and its MPC was the lowest of all offerors. Id. at 2-5. The SSA concluded, “[a]ny modestly lower costs possible from any of the other offers are more than offset by the significantly higher ratings [FCSS] achieved in the Quality factor (reflecting its superior presentation of understanding of the requirements and innovative approaches to satisfying them),” and he selected FCSS as the Best Value Offeror. Id. at 5. The SSA found FCSS’s proposal to be “clearly superi- or to the other proposals,” and thought “the other proposals would require significant improvement to be competitive with” FCSS’s, justifying an award without discussions. Id.
D. The Initial Protests
On August 12, 2004, the Contracting Officer and other officials signed and approved the Price Negotiation Memorandum memorializing the decision to award the contract to FCSS. See AR Tab 78. Five days later, the Army Contracting Agency approved the award. AR Tab 79. Pre-award notification letters were sent to the unsuccessful offerors on August 26, 2004. AR Tab 81. On September 1, 2004, the contract was awarded to FCSS. AR Tab 85. That same day, post-award notification letters, which also served as written debriefings, were sent to the unsuccessful offerors. AR Tab 87 (letter to KIRA), Tab 88 (letter to TGG). Each letter briefly informed the unsuccessful offeror of the strengths, significant weaknesses, and deficiencies of that offeror’s proposal and revealed its MPC. The letters also explained that FCSS’s “proposal received significantly higher ratings in the Quality factor, showing greater understanding of the requirements, better attention to detail, and more innovations than any of the other proposals.” AR Tabs 87, 88. Perhaps most significantly, the letters revealed FCSS’s MPC of $204,301,121, and identified FCSS’s proposed cost to be $165,766,025. Id.
After receiving the written debriefing, KIRA, TGG, and a third offeror filed protests of the award with the U.S. Government Accountability Office (GAO).
In response to the discrepancy in FCSS’s proposed cost figures, and at the recommendation of the Contracting Officer, the adjectival and risk ratings given to the FCSS proposal for the Management and Technical element and the overall Management subfactor were downgraded from “Good” and “Low” to “Satisfactory” and “Moderate.” AR Tabs 94, 95. Presumably because of the “problems with the information” identified by the Contracting Officer, see AR Tab 95 at 2, on October 5, 2004 the Army took corrective action, staying performance of the contract awarded to FCSS and establishing a competitive range that included FCSS and the three protesting offerors. AR Tab 97; see also id. Tab 105 (Nov. 8, 2004 Prenego-tiation Objective Memorandum determining the offerors in the competitive range).
E. Evaluation of the Final Proposal Revisions
After the SSEB met on several occasions to discuss the strengths, uncertainties, weaknesses, significant weaknesses, and deficiencies previously identified in the evaluations of the various proposals, see AR Tabs 99, 101, 102, letters were sent to each offeror in the competitive range, with a list of the uncertainties, weaknesses, significant weaknesses, and deficiencies of that offeror’s proposal. AR Tabs 115-17. At the bottom of each list, after the identification of “Deficiencies” and set off from the rest of the document by a line running the width of the text, was the identical note for each:
In addition to the above matters, the proposed staffing reflected substantial variances from that which the Government considered necessary. These variances included large numbers of instances of un-derstaffing and a much smaller number of instances of overstaffing.
Id.
Discussions were separately held between representatives of each offeror and members of the SSEB. See AR Tabs 132, 135, 138 (Army’s discussion minutes). During these discussions, the Contracting Officer fielded questions about the specific items from each list of criticisms provided to that offeror, and also followed a written script so that each offeror was provided the same general description of the need to recognize and consider possible understaffing and to justify any reduction in staff needs due to innovations.
On December 21, 2004, the Army issued Amendment 10 to the Solicitation, which invited each offeror in the competitive range to submit a Final Proposal Revision (FPR). AR Tab 149. In this amendment, the Army repeated that “the Quality factor is significantly more important than the Cost factor” and emphasized: “Especially in light of this, revised proposals must show a clear correlation between methods and means for accomplishing or exceeding the requirements of the solicitation as expressed in the Management subfactor proposal and the costs shown in the Cost factor proposal.” Id. at 2. Offerors were informed that the Army “will re-validate PPE information, including any new information that may be available for activities since the original proposal submission date,” and that offerors would be allowed to add two additional references who may submit PPE questionnaires. Id. The amendment also stressed: “As stated in paragraph M.2.2.1.1, the Government will consider any innovative ideas, but proposals should clearly identify and explain ... any cost savings that rely upon those innovations. ” Id. at 2-3 (emphasis added).
Fort Carson Support Services, TGG, and KIRA each timely submitted an FPR.
The Army evaluated the FPRs to determine the Most Probable Cost for each. This involved identifying instances of overstaffing and understaffing, relative to the Independent Government Estimate (“IGE”) for the contract, taking into account labor savings due to proposed innovations. See AR Tabs 197(IGE), 206 (KIRA MPC adjustments), 207 (TGG MPC adjustments), 208 (FCSS adjustments).
On March 30, 2005, the Chairman of the SSEB submitted the Proposal Analysis to the Contracting Officer, reflecting the consensus judgment of the SSEB. AR Tab 209. The PA Report explained that the time between the last caucus meeting concerning an individual FPR — February 17, 2005 — and the final caucus on March 4, 2005, “was used by SSEB members to re-review the initial proposals as necessary and to reflect on whether they had evaluated all the final proposals in the same manner.” Id. at 3. Concerning PPE, the Chairman noted “that the descriptions of the various performance risk assess
As part of the evaluation process for the final revised proposals, the SSEB reconsidered the PPE information in light of the importance of the PPE Subfactor (relatively equal to the Management Subfactor) and in light of the fact that the Management Subfaetor ratings and proposal risk assessments of the offerors were closer than they had been in the initial evaluation round. In this reconsideration, the SSEB found some instances in which it felt that its initial evaluations of PPE had been erroneous and made appropriate adjustments in this round.
Id. at 5 (emphasis added).
The PA Report set out the basis of the MPC calculation for each proposal. Starting with the IGE’s 258 full-time equivalent employees, an additional five were added for KIRA to reflect positions proposed under its method of performance that were not included in the IGE. Id. at 6. This number was then reduced by 6.3 man-years, reflecting credit “for proposed innovations that the SSEB deemed justified.” Id. The SSEB found that other reductions for efficiencies proposed by KIRA “were not supported by documentation in the proposal and, thus, were not reflected in the MPC calculation.” Id. The KIRA proposal was determined to be overstaffed in five positions and understaffed in forty-two, for a net shortage of 30.7 man-years. See id.; see also AR Tab 211 at 39. The KIRA proposal’s MPC over ten years was calculated as $209.3 million — a variance of $[XXX], or [XXX] percent, over its TEC. AR Tab 209 at 7. The SSEB determined that TGG was overstaffed in 8.5 positions and short by 33 man-years, for a net shortage of 24.5 productive man-years relative to the IGE. See id.; see also AR Tab 211 at 59. The total MPC for TGG’s proposal was $[XXXX] million, a variance of $[XXX] million, or 7.4 percent, over its TEC. AR Tab 209 at 8. The FCSS proposal was found to be overstaffed by nineteen positions and understaffed by 94.5 man-years’ worth. Id. The proposal “was credited with 3 man-years for proposed innovations and .5 man years for home office effort.” Id. This resulted in a total MPC of $208.6 million, which was $32.7 million — or 15.4 percent — higher than the FCSS proposal’s TEC. Id.
In discussing the variance of the calculated MPC from the proposals’ estimated costs, the Chairman noted that KIRA’s variance for the first option year was just 5.8 percent or $1.1 million. Id. at 7. This variance “was evaluated as not likely to significantly impact contract performance and reasonable in relation to the offeror’s proposed method of performance.” Id. The higher variance of 10.6 percent from the KIRA proposed cost for the fall-life of the contract, “did raise more of a question about impact on contract performance and reasonableness of estimated costs,” but “was still considered satisfactory.” Id. The TGG proposal’s smaller variance of 7.4 percent based on total MPC “was evaluated as not likely to significantly impact contract performance” and was found satisfactory. Id. at 8. The FCSS variance of 15.7 percent for the first option year and 15.4 percent from its total estimated cost “was evaluated as possible to impact contract performance and raised [a] question in relation to the offeror’s proposed method of performance,” resulting in the SSEB’s finding that this variance was “only marginally satisfactory.” Id.
Concerning the Quality factor, the PA Report contained less information. Despite the statement, quoted above, that the SSEB found that some of “its initial evaluations of PPE had been erroneous,” id. at 5, no risk ratings or adjustments relating to the PPE consensus evaluation are identified. For the Management Subfactor scoring, the briefing charts prepared for the SSA presentation, see AR Tab 211, are referenced as providing the consensus ratings and proposal risk assessments. AR Tab 209 at 3. The Chairman did, however, elaborate on the methodology followed by the SSEB:
As part of the Management Subfaetor evaluation, the SSEB considered whether the offerors’ proposed staffing was eonsis-*582 tent with their proposed means and methods for performing the contract. This item was a 'point of emphasis in the discussion sessions with the offerors in December and, afterwards, was specifically re-emphasized in Amendment 10, paragraph 4b. As a result, the SSEB scrutinized this issue much more closely in the second evaluation, and did not, as it may have in the first evaluation, give benefit of doubt to the offerors. Shortfalls and excesses in proposed staffing thus determined were then used in the evaluation of the Cost Factor, as well as, where appropriate, in assessing the rating and proposal risk assessment for the Management Sub-factor.
Id. at 4 (emphasis added).
The briefing charts used in the SSEB’s presentation of its evaluation results to the SSA identified the strengths and weaknesses found for each proposal under the Management Subfaetor and each corresponding element, and provided the adjectival and proposal risk ratings. See AR Tab 211 (Mar. 31, 2005 FRP Evaluation Briefing). These charts also explained the PPE evaluations and risk ratings, and provided the Cost factor information. Id. The Management Sub-factor adjectival and proposal risk ratings for KIRA improved from the initial “Marginal” and “Medium-High” to “Satisfactory” and “Moderate.” Id. at 32. For the individual elements, the SSEB found several improvements in the KIRA FPR. The ratings in the most important Management and Technical element improved, as did the overall Subfactor scoring, from the initial “Marginal” and “Medium-High” to “Satisfactory” and “Moderate.” AR Tab 211 at 22. The adjectival rating for Human Resources Management improved from “Satisfactory” to “Good,” although the risk rating slipped from “Low” to “Moderate/Low.” Id. at 28. The Phase-In Planning adjectival rating moved from the initial “Good” to “Outstanding.” Id. at 24.
For offeror TGG, the Management Subfactor evaluation also showed an improvement in its FPR. The over-all ratings and those for the Management and Technical element— initially “Marginal” with “Medium-High” risk — were raised by the SSEB to “Satisfactory” and “Moderate” risk. Id. at 42, 52. For the Contract Administration and Human Resources Management elements, TGG’s adjectival ratings improved from “Marginal” to “Satisfactory.” Id. at 46, 48. But TGG’s Quality Control element continued to be rated “Unsatisfactory,” with “High” proposal risk. Id. at 50. The SSEB found this aspect of TGG’s proposal to be a deficiency, because it “lacked sufficient detail to be evaluated,” id. at 51, with the result that TGG “would not be eligible for award.” Id. at 52.
The SSEB slightly reduced the over-all Management Subfactor ratings of FCSS, dropping its evaluation of performance risk from “Low” to “Moderate.” AR Tab 211 at 72. The Management and Technical element was downgraded from the initial “Good” and “Low” risk to “Satisfactory” and “Moderate” risk. Id. at 62. On the other hand, the significantly less important Quality Control element ratings were boosted from “Marginal” and “Moderate” risk to “Good” and “Low” risk. Id. at 70.
Concerning Past Performance and Experience, the SSEB determined that neither KIRA nor TGG had any experience with cost contracts, and changed the risk rating for that element from “Low” to the neutral “Unknown” for both. Id. at 35, 55. The “Low” risk assessment for the other elements of PPE, and for PPE over-all, was retained for these two offerors. Id. at 35-36, 55-56. Two of the four ratings of the PPE elements for FCSS dropped from “Low” to “Moderate” — Managerial, with the comments that “[rjatmgs varied” and “lower scores generally came from Plum Island,” id. at 75; and Cost Control, for which the SSEB noted one of the joint venture members “had problems with only cost contract.” Id. The SSEB gave FCSS an overall PPE Subfactor assessment of “Low/Moderate” risk, compared to the initial “Low” assessment. Id. at 76. The briefing charts also contained the same cost information that was included in the PA Report. See AR Tab 211 at 38-40, 58-60, 78-80, 83.
Two weeks after receiving the SSEB briefing, the SSA issued his second SSDD for this procurement. AR Tab 214 (Apr. 14, 2005 SSDD). The SSA stated that he had “fully considered the merits of the proposals” and “the facts and conclusions provided by” the SSEB. Id. at 1. He determined that KIRA’s proposal offered the best value, and documented his rationale. Id. The SSA stated that, “[e]xcept where noted,” he concurred in the SSEB’s evaluations of the offerors. Id. at 2. The SSA explained that KIRA “substantially revised its proposal for this round,” and he assigned it a “Satisfactory” rating with “Moderate” proposal risk for the Management Subfactor and the Management and Technical element, as did the SSEB. Id. The SSA briefly discussed some of the strengths and weaknesses of the KIRA proposal. He found that KIRA’s PPE information “reflected that it has had experience as a prime contractor, although not in projects of the scope or magnitude to [sic] this contract.” Id. The SSA noted that the proposed main subcontractor of KIRA, however, “has had experience in similar projects,” and the PPE performance risk of “Low” was assigned to KIRA across-the-board. Id.
For TGG, the SSA concurred in the “Satisfactory” rating and “Moderate” proposal risk given the FPR by the SSEB, id., but also concurred in the deficiency identified by the SSEB:
In contrast to [TGG]’s Management strengths, the Quality Control plans provided in its proposal, as revised, lacked detail sufficient to reflect an understanding of the requirements. As a result, the SSEB evaluated this element as deficient, with an unsatisfactory rating and a high proposal risk.
Id. at 3. The performance risk assessment of “Low” was given TGG, for the PPE Subfactor and each element, with the SSA noting the “submitted information reflected experience as a prime contractor, although not of the magnitude of this contract.” Id.
In discussing the FCSS proposal, the SSA briefly discussed its “significant strengths,” but added that it “had an even more significant weakness of a very large shortage in staffing, evaluated as almost a fourth of the total workforce estimated by the Government as necessary for this contract.” Id. The SSA characterized the FCSS past performance information as “showing experience for each of its two joint venturers as a prime, although of considerably less magnitude and narrower scope than this contract.” Id. (emphasis added).
Discussing his best value determination, the SSA explained that this “included consideration of both the stated weights of the evaluation factors and subfactors and the large range of the various rating definitions.” AR Tab 214 at 3. For the Management Sub-factor, the SSA “considered [KIRA]’s proposal to be at the high end of satisfactory.” Id. at 4. Although FCSS’s proposal “was better written than either of the other two, based on its distinctly better maintenance plans,” the SSA agreed with the SSEB that “this advantage was undercut by its significant understaffing, which was almost twice that of [KIRA] and more than twice that of [TGG].” Id. He concluded:
This understaffing raised [a] question of the credibility to be afforded [FCSS]’s Management proposal; i.e., whether [FCSS] really understood the contract requirements. As a result, I considered [FCSS]’s Management subfactor to be only slightly better than [KIRA]’s, with [TGG] below the other two.
Id. Concerning PPE, the other subfactor under Quality, the SSA explained:
Looking beyond the ratings to the substantive information, I considered [KIRA] to be well above the other two. This was based on [KIRA]’s experience as a prime and the higher quality of the ratings by its references. Both [TGG] and [FCSSJs experiences as primes were significantly less relevant than [KIRA]’s, and [FCSS]’s ratings were considerably more mixed than either of the other two offerors.
Overall, then, for the Quality factor, [KIRAJs substantially better PPE than [FCSSJs more than offset [FCSSJs slightly better Management subfactor proposal. I considered [TGGJs PPE to be somewhat better than [FCSSJs, but not enough to offset [FCSSJs considerably better Management subfactor proposal.
Id.
Looking at the Cost factor, the SSA noted that although FCSS’s “Total Estimated Cost was about $11 million lower than [KIRAJs, that difference is mostly attributable to [FCSSJs understaffing,” and recognized that KIRA’s “MPC was only about $700,000 higher than” FCSS’s. Id. at 4. The SSA believed the “minor difference” in MPC over ten years was “essentially insignificant.” Id. Moving to the staffing shortage, the SSA explained that in the initial evaluation, he “gave [FCSS] a substantial credit in manpower based on its assertion of savings anticipated from proposed innovative methods of performance,” but this still left “a significant manpower shortage, which was pointed out to [FCSS] in discussions.” Id. He noted that “[i]n the second round of evaluation, the SSEB reassessed [FCSSJs essentially still undocumented claims of innovation savings and gave them a minimal credit.” Id. The SSA acknowledged that KIRA “also had a significant staffing issue,” based on undocumented alleged efficiencies that progressively grow in the option years. Id. But the KIRA variance was “substantially less” than FCSS’s, and KIRA’s “proposal for the first full contact year showed reasonable staffing,” which the SSA thought were “major distinctions” between the two offerors. Id. at 5. The SSA added that he and the SSEB had “concern about the credibility of [FCSSJs technical approach,” as FCSS’s “large staffing shortfall existed throughout its proposal.” Id.
The SSA thus determined that KIRA’s proposal offered the best value to the government. This decision was confirmed in a Post-Price Negotiation Objective Memorandum signed April 14, 2005, and approved April 21, 2005. AR Tab 223. Pre-award notifications were sent to FCSS and TGG on April 25, 2005. AR Tabs 225, 227. The contract award was executed on June 14, 2005, AR Tab 245; see also id. Tab 244 (June 10, 2005 Congressional Notification), and the post-award notices were sent the next day. AR Tabs 248, 251. On June 16, 2005, the Army held a post-award debriefing for FCSS. See AR Tab 255. Five days later, TGG received its post-award debriefing. See AR Tab 259.
G. Procedural History
The post-award bid protest of FCSS was filed in this Court on June 20, 2005, and the protest of TGG was filed on July 1, 2005. A protective order was issued on June 21, 2005. The cases were consolidated on July 11, 2005.
The government filed the Administrative Record, consisting of fourteen volumes, on July 18, 2005. On several occasions during the pendency of this case, the government and both plaintiffs have moved for leave to file additional documents as part of the Administrative Record, to add recently discovered documents, transcriptions of briefings, or clearer copies of documents previously submitted.
To allow the matter to be set for a normal briefing schedule, the government initially extended the contract of the incumbent contractor through the end of 2005, see Order (June 21, 2005) at 1, and then further extended that contract to June 30, 2006. Tr. at 325. Oral argument was held on March 14, 2006. The Court issued an order on June 1, 2006, granting the government’s and KIRA’s motions for judgment on the administrative record, and denying the plaintiffs’ motions for judgment on the administrative record. This opinion explains the reasoning behind that order.
II. DISCUSSION
A. Standard of Review
Post-award bid protests are heard by this Court under the Tucker Act, as amended by the Administrative Dispute Resolution Act of 1996 (“ADRA”), Pub.L. No. 104-320, §§ 12(a)-(b), 110 Stat. 3870, 3874 (1996). 28 U.S.C. § 1491(b)(1) (2000). This provision requires our Court to follow Administrative Procedure Act (“APA”) standards of review. 28 U.S.C. § 1491(b)(4). Those standards, incorporated by reference, provide that a:
reviewing court shall ... (2) hold unlawful and set aside agency action, findings, and conclusions found to be — [¶] (A) arbitrary, capricious, an abuse of discretion, or otherwise not in accordance with law; [¶] (B) contrary to constitutional right, power, privilege, or immunity; [¶] (C) in excess of statutory jurisdiction, authority, or limitations, or short of statutory right; [¶] (D) without observance of procedure required by law; [¶] (E) unsupported by substantial evidence in a case subject to sections 556 and 557 of this title or otherwise reviewed on the record of an agency hearing provided by statute; or [¶] (F) unwarranted by the facts to the extent that the facts are subject to trial de novo by the reviewing court. In making the foregoing determinations, the court shall review the whole record or those parts of it cited by a party, and due account shall be taken of the rule of prejudicial error.
5 U.S.C. § 706 (2000).
Based on an apparent misreading of the legislative history, see Gulf Group. Inc. v. United States, 61 Fed.Cl. 338, 350 n. 25 (2004), the Supreme Court had determined, before the 1996 enactment of the ADRA, that the de novo review standard of 5 U.S.C. 706(2)(F) does not usually apply in review of informal agency decisions — decisions, that is, such as procurement awards. See Citizens to Preserve Overton Park, Inc. v. Volpe, 401 U.S. 402, 415, 91 S.Ct. 814, 28 L.Ed.2d 136 (1971). Instead, courts in those cases are supposed to apply the standard of 5 U.S.C. § 706(2)(A): whether the agency’s acts were “arbitrary, capricious, an abuse of discretion, or otherwise not in accordance with law.” See Overton Park, 401 U.S. at 416, 91 S.Ct. 814 (citation omitted). The “focal point for judicial review should be the administrative record already in existence, not some new record made initially in the reviewing court.” Camp v. Pitts, 411 U.S. 138, 142, 93 S.Ct. 1241, 36 L.Ed.2d 106 (1973). This applies even where, as here, the matter under review was not the product of a formal hearing. See Florida Power & Light Co. v. Lorion, 470 U.S. 729, 744, 105 S.Ct. 1598, 84 L.Ed.2d 643 (1985).
A motion under RCFC 56.1 for judgment on the administrative record differs from motions for summary judgment under RCFC 56, as the existence of genuine issues of material fact does not preclude judgment on the administrative record. See Bannum, Inc. v. United States, 404 F.3d 1346, 1355-56 (Fed.Cir.2005). Rather, a motion for judgment on the administrative record examines whether the administrative body, given all the disputed and undisputed facts appearing in the record, acted in a manner that was arbitrary, capricious, an abuse of discretion, or otherwise not in accordance with law. See Arch Chemicals, Inc. v. United States, 64 Fed.Cl. 380, 388 (2005). If arbitrary action is found as a matter of law, the Court will then decide the factual question of whether the action was prejudicial to the bid protester. See Bannum, 404 F.3d at 1351-54. Under
Because of the deference courts give to discretionary procurement decisions, “the disappointed bidder bears a ‘heavy burden’ of showing that the award decision ‘had no rational basis.’ ” Impresa Construzioni 238 F.3d at 1333 (quoting Saratoga Dev. Corp. v. United States, 21 F.3d 445, 446 (D.C.Cir.1994)). In particular, the evaluation of proposals for their technical excellence or quality is a process that often requires the special expertise of procurement officials, and thus reviewing courts give the greatest deference possible to these determinations. See E.W. Bliss Co. v. United States, 77 F.3d 445, 449 (Fed.Cir.1996); Arch Chems., 64 Fed.Cl. at 400; Gulf Group, 61 Fed.Cl. at 351; Overstreet Elec. Co. v. United States, 59 Fed.Cl. 99, 102, 108, 117 (2003). Challenges concerning “the minutiae of the procurement process in such matters as technical ratings ... involve discretionary determinations of procurement officials that a court will not second guess.” E.W. Bliss, 77 F.3d at 449. “Naked claims” of disagreement with evaluations, “no matter how vigorous, fall far short of meeting the heavy burden of demonstrating that the findings in question were the product of an irrational process and hence were arbitrary and capricious.” Banknote Corp. of Am. v. United States, 56 Fed.Cl. 377, 384 (2003).
The presence (by the government or inter-venor) or absence (by the protester) of any rational basis for the agency decision must be demonstrated by a preponderance of the evidence. See Gulf Group, 61 Fed.Cl. at 351; Overstreet, 59 Fed.Cl. at 117; Info. Tech. & Appl’ns Corp. v. U.S., 51 Fed.Cl. 340, 346 (2001) (citing GraphicData LLC v. United States, 37 Fed.Cl. 771, 779 (1997)), aff'd, 316 F.3d 1312 (Fed.Cir.2003).
An additional ground for a contract award to be set aside is when the protester can show that “the procurement procedure involved a violation of regulation or procedure.” Impresa Construzioni, 238 F.3d at 1332.
B. Fort Carson Support Services’ Protest
As was recounted above, FCSS was initially awarded the contract at issue after the first round of evaluations. See AJÍ Tab 77. At that time, the PPE ratings of all offerors were considered equal, and the SSA focused on the Management subfactor evaluation and cost. Id. at 2. No other offeror equaled FCSS’s rating of “Good” for the subfaetor, or its “Low” proposal risk rating, and its MPC was the lowest of all offerors. See id. at 2-5. The low MPC resulted from the Army’s decision to adjust FCSS’s estimated costs to reflect only a thirty-nine employee shortfall, although the FCSS proposal was short a net seventy-eight full-time equivalents (FTEs)
In its final proposal, FCSS made modest changes to its staffing levels.
Fort Carson Support Services challenges the award to KIRA on several grounds. First, it contends that under the Solicitation criteria, its proposal was the highest rated as well as lowest priced and thus it should have been awarded the contract. Second, it argues that the Army failed to articulate a rational basis for preferring the KIRA proposal. Third, it disputes the Army’s evaluations of its proposed staffing levels, its past performance and experience, and the cost of its proposal. And finally, it claims that the Army failed to engage in meaningful discussions concerning its proposal.
1. Was the FCSS Proposal, as Evaluated, Entitled to the Contract Award?
The first argument of FCSS is that a best value trade-off was neither necessary nor permitted, as it claims to have received the highest score for Quality, and it proposed the lowest price. FCSS Mot. at 24-31. This sort of argument could, perhaps, have been avoided had the Army followed a source selection plan in which an overall rating were applied to the Quality factor. But instead, the Quality evaluation was “a composite evaluation using all of the subfae-tors and elements associated with this factor” that “evaluates the extent to which an offeror’s proposal demonstrates a sound, practical methodology for meeting or exceeding the overall requirements associated with the PWS.” Solicitation § M.2.2, AR Tab 7 at 49. Unlike the two subfactors and their elements, the evaluation of the Quality factor did not result in a particular rating, label, or category.
Thus, FCSS focuses on the one-word ratings generated by the evaluation process. In the second evaluation, FCSS’s RFP received an overall adjectival rating of “Good” and proposal risk rating of “Moderate” for the Management subfactor; KIRA and TGG’s RFPs each received the adjectival rating of “Satisfactory” and risk rating of “Moderate.” AR Tab 214 at 2-3. For the PPE subfactor, FCSS received an overall performance risk rating of “Low/Moderate,” while KIRA and TGG each received a “Low” rating. Id. at 4. But since the Solicitation provides for just three distinct performance risk ratings
The “Low/Moderate” performance risk rating was the result of the SSEB’s determination that two PPE elements warranted a “Low,” while the other two elements merited just a “Moderate.” See AR Tab 211 at 75-76. Although the Solicitation instructed that its rating descriptions “may not precisely describe the assessment of a specific ... element” and “[i]n that case, evaluators will assess the rating with the description most closely describing the actual assessment,” AR Tab 7 at 52, the SSEB avoided the decision of whether the overall performance risk was closer to “some doubt” or “little doubt,” and attempted to split the difference. Fort Carson Support Services would place emphasis on the words “description most closely describing,” in essence arguing that the SSEB and SSA were required to generate a single label for their risk rating of this subfactor. But the last three words, “the actual assessment,” appear to be the more important. In the course of the evaluation process, the Management subfactor and elements would “receive an overall adjectival and proposal risk assessment rating,” and the PPE subfactor and elements would “receive a performance risk rating.” Solicitation § M.2.2, AR Tab 7 at 49. But these ratings are not themselves the actual evaluation or assessment, but merely tools employed in the assessment. If the purpose of the evaluation were to generate a label used in comparing proposals, there would be no need for the SSA to utilize a “tradeoff process among the factors, subfaetors, and elements, ” Id. § M.1.1, AR Tab 7 at 48 (emphasis added), to determine the best value; there would be no need to delve below the subfactor scores.
But in the evaluation process established for this procurement, the elements matter. The Source Selection Plan required that “[t]he SSA shall utilize all evaluation reports and analysis in determining the successful proposal,” SSP § IV.B(l), AR Tab 34 at 14, not that he shall use just the ratings, let alone just the ratings of the subfaetors. In this context, neither the SSEB nor the SSA
As the SSEB Chairman noted in his PA Report:
[T]he descriptions of the various performance risk assessments are very broad. As a result, offerors which have received the same assessment may not be equal in regard to PPE, and, as a result, the supporting documentation may provide additional relevant insight, if necessary.
AR Tab 209 at 5. The SSA explained that his “determination of the best value awardee included consideration of both the stated weights of the evaluation factors and subfac-tors and the large range of the various rating definitions.” AR Tab 214 at 3 (emphasis added). In other words, proposals of varying quality might be described as offering “no particular advantages to the Government but ... fully meetpng] the ‘minimum’ requirements of the RFP,” or as involving an approach with “an acceptable probability of meeting the requirements,” to take the “Satisfactory” adjectival rating as an example. See Solicitation § M.3.1, AR Tab 7 at 52. And the difference between some of these proposals, and the “Good” ones for which it can be said that “some advantages to the Government” are offered, or that they embody an approach with “a high probability of meeting or exceeding the requirements,” might be small. See id. This seems particularly true for the performance risk ratings, such as the “Low” risk resulting when “little doubt exists that the offeror will successfully perform the required effort.” See Solicitation § M.3.3, AR Tab 7 at 53.
Thus, the evaluation process cannot simply be boiled down to whether two proposals received the same rating for one subfactor, and one was a grade higher in one of the two ratings used for another subfaetor. To be sure, it would be a different matter if the SSA had determined that the FCSS proposal was a full category better than KIRA’s in the adjectival and proposal risk ratings for the Management subfactor and in the overall performance risk rating. Under such circumstances, it would be arbitrary and inconsistent to judge the KIRA proposal to be the highest quality. But those circumstances are not presented here. In the Source Selection Decision, to look just at the labels for the moment, it was determined that KIRA’s evaluated PPE risk of “Low” exceeded the value of FCSS’s risk of “Low/Moderate” (or “Low,” if you will) by more than the FCSS Management subfactor evaluation of “Good” with “Moderate” risk exceeded the KIRA evaluation of “Satisfactory” with “Moderate” risk. Is this, as FCSS argues, necessarily incorrect?
To return to the university example, the use of broad rating categories, without any greater refinements, is like a grading system that uses solely the letter grades A, B, C, D, and F. If an A were earned by a final exam score of 93 or higher, and a B by a score from 86 to 92, then the difference between an A and a B could be smaller than the difference between two Bs — say, if the corresponding scores were 93, 92, and 86. A person who reviews the underlying exam scores will evaluate academic performance differently than a person who lacks access to the raw scores.
Or, to put it another way, since the evaluation process for this procurement did not involve mathematical scoring, consider the use of plus/minus in grading. Without plus/minus, a high B and a low B are seen as the same by one perusing a transcript, and would both appear to be equally below an A. In an evaluation calculating grade-point averages, the A merits a 4.0 and the B a 3.0. But if the grades are further refined, using one common approach to pluses and minuses that
To translate the example to our case, it is not inconsistent with the Solicitation evaluation criteria for the SSA to decide that a KIRA A+ compared to an FCSS Aout-weighs the difference between an FCSS A- and a KIRA B + . Of course, the evaluation process did not use ratings distinctions similar to a plus or minus, nor is there a way of mathematically comparing a “Good” adjectival rating with a “Low” performance risk rating. But the evaluation did involve the SSA using more than the one-word labels— the equivalent of letter grades — to determine which proposal merited the highest Quality factor assessment. He may not have been given pluses and minuses, but he did not need them, as he had the benefit of the equivalent of the exam scores (or, perhaps more analogously, term papers with the professor’s comments) to use in assessing the proposals.
Concerning past performance, the SSA specifically stated that he was “[Booking beyond the [PPE subfactor] ratings to the substantive information” in determining that KIRA’s PPE was “well above the other two.” AR Tab 214 at 4. He based this determination on KIRA’s “experience as a prime and the higher quality of the ratings by its references.” Id. The SSA noted that FCSS “submitted information showing experience for each of its two joint venturers as a prime, although of considerably less magnitude and narrower scope than this contract.” Id. at 3. He found that FCSS’s and TGG’s “experiences as primes were significantly less relevant than [KIRA’s], and [FCSS’s] ratings were considerably more mixed than either of the other two offerors,” and concluded that KIRA had a “substantially better PPE than” FCSS. Id. at 4.
Likewise, the SSA looked beyond the adjectival ratings to determine that KIRA’s Management proposal was “at the high end of satisfactory,” and that FCSS’s Management proposal was “only slightly better than [KIRA]’s.” AR Tab 214 at 4.,Moreover, the SSA determined that although FCSS’s Management proposal was “better written” than the other two offerors’ because of FCSS’s “distinctly better maintenance plans,” “this advantage was undercut by its significant understaffing,” which “raised [the] question of the credibility to be afforded [FCSS’s] Management proposal ... i.e., whether [FCSS] really understood the contract requirements.” AR Tab 214 at 4. The net result of the SSA’s analysis concerning the Quality factor was that “[KIRA]’s substantially better PPE than [FCSS]’s more than offset [FCSS]’s slightly better Management subfactor proposal.” Id. The Court does not find this conclusion to be inconsistent with the Solicitation’s evaluation criteria. As was explained above, the rating categories themselves are broad enough so that the difference between a low “Good” and a high end “Satisfactory” may well be exceeded by the difference between two Past Performance and Experience ratings of “Low” risk. The SSA “articulate[d] a satisfactory explanation for” his decision, Motor Vehicle Mfrs. Ass’n, 463 U.S. at 43, 103 S.Ct. 2856, and the Court accordingly rejects the argument that FCSS received the best Quality factor evaluation of the three offerors.
2. Did the Army Adequately Articulate a Basis for the Award to KIRA ?
In its first ground for protest, discussed above, FCSS erred in positing an evaluation process in which the ultimate deci-sionmaker could not see through the various labels attached to aspects of proposals, but instead was forced to consider all ratings with the same label to be equal. As part of its argument for its second ground for protest, FCSS errs in the opposite direction, by placing undue attention on the raw number of strengths and weaknesses identified for each proposal. In arguing that it was irrational for the Army to have preferred the KIRA proposal to its own, FCSS relies on the SSEB’s caucus determinations that KIRA’s final proposal contained seventeen strengths, one uncertainty, and nineteen weaknesses (including two significant weaknesses), see AR Tab 203, while its own final
When the SSEB briefed the SSA on the results of its FPR evaluations, the strength and weakness gaps between the two offerors narrowed. Considering the strengths and weaknesses the SSEB thought were notable enough to be included in the presentation, KIRA had thirteen strengths to FCSS’s twelve, and five weaknesses to FCSS’s two (including one significant weakness for each). See AR Tab 211 at 23-33, 63-73. In the PA Report, the Chairman of the SSEB explained that “these categories [of strengths and weaknesses] are all very broad and that, as a result, the listings within them are not of equal value or importance,” and that “no attempt has been, or should be, made to establish ratings or risk assessments based solely on the numbers listed in each such category.” AR Tab 209 at 3. After citing these statements of the Chairman, FCSS nonetheless argues that “the total numbers must have significance, otherwise the entire evaluation process ... would be meaningless.” FCSS Mot. at 33.
But the reason that strengths and weaknesses are identified, and must be documented in an evaluation, see 48 C.F.R. § 15.305(a), is to help ensure that the evaluations are thoughtful. When evaluators focus on the strengths, weaknesses, significant weaknesses, and deficiencies of proposals, they are better able to classify and articulate the quality of a proposal. Documentation of this adds transparency to the process, in two ways: it provides evidence that each proposal was carefully considered, and provides a reviewing court with some decisions that are verifiable (in circumstances when the description of a strength or weakness may be flatly contradicted by the portion of the proposal it references). The identification of strengths or weaknesses, however, does not convert the evaluators’ subjective judgment into some objective fact that may be disproved in court, nor does it result in a product that can be mechanically summed or subtracted. To be sure, an agency could probably adopt an evaluation process in which certain strengths and weaknesses are given positive or negative value in advance, and the evaluators merely check-off the number of positive or negative features present in a proposal, and total these up to reach their conclusion. This was not the process adopted here.
The numbers that FCSS points to simply cannot be used in the way it suggests. There is no reason why the “total numbers must have significance,” as FCSS argues, any more than the total number of words used to describe an evaluation is intrinsically significant. And in any event, even if one could view all weaknesses and all strengths to be of equal value, the mere numbers do nothing to advance FCSS’s ease. These strengths and weaknesses all relate to the Management subfactor. See AR Tabs 203, 204, 211. But no one denies that FCSS had the highest rated offer under the subfactor-an adjectival rating of “Good” with “Moderate” risk, compared to the adjectival ratings of “Satisfactory” and “Moderate” risk for the other two offerors. See AR Tab 214 at 3; see also AR Tab 211 at 32, 52, 72. The fact that FCSS had the highest rating for the Management subfactor does not render the award decision irrational, as the PPE subfactor and its elements carried the same weight as the Management subfactor and its components. See AR Tab 214 at 1; AR Tab 7 at 48.
The SSA’s decision was not based on which offeror had the best Management subfaetor rating, but considered the overall Quality factor, taking into account the subfactors and their elements, relative to the less important Cost factor. AR Tab 214 at 4-5. Concerning this decision, FCSS argues that the Army failed to adequately articulate the basis for its best value tradeoff. FCSS Mot. at 35-39; FCSS Reply at 36-37. Portions of the post-award debriefing by the Contracting Officer are quoted to demonstrate the lack of an adequate articulation of the tradeoff decision. See FCSS Mot. at 35-38. But while such debriefings might, on occasion, yield useful party admissions, any new explanations of the evaluation process made post-award are the sort of “post hoc rationalizations” that courts reject when offered by the
In the SSDD, the Source Selection Authority provides a summary of the proposal evaluations based on the relevant factors identified in the Solicitation, and a supporting narrative, see AR Tab 214, as the Federal Acquisition Regulations (FAR) require. See 48 C.F.R. § 15.305(a)(3). The SSA’s rationale for his business judgment and tradeoff decision is also supplied. After explaining that KIRA’s “substantially better PPE than [FCSSj’s more than offset [FCSSj’s slightly better Management subfactor proposal,” the SSA turned to an analysis of the Cost factor. AR Tab 214 at 4. He noted that the $11 million difference between the proposed costs of KIRA and FCSS was “mostly attributable to [FCSSj’s understaffing.” Id. He looked at the Most Probable Cost for the two proposals, as mandated by the FAR, see 48 C.F.R. § 15.404 — l(d)(2)(i), and determined that KIRA’s $700,000 higher MPC was a “minor difference over ten years and in relation to the overall cost of over $200 million is essentially insignificant, certainly not enough to offset [KIRAJ’s advantage in the Quality factor.” AR Tab 214 at 4.
Fort Carson Support Services argues that the Army failed to “articulate what it gained in Quality by awarding to KIRA.” FCSS Mot. at 35; see also id. at 39. But the SSA plainly explained that FCSS’s “overall manpower shortage ... was considered detrimental,” AR Tab 214 at 4, and that he found KIRA’s proposal preferable, despite KIRA’s “significant staffing issue,” because KIRA’s “proposal for the first full contract year showed reasonable staffing, although with some shortfalls, thus allowing a conclusion that its technical approach was credible.” Id. at 4-5. The SSA’s confidence in the staffing level proposed by KIRA for the first full year of performance, and KIRA’s “substantially” smaller variance between proposed and probable costs, relative to FCSS, were the benefits received by the Army in return for the higher price. This explanation satisfies the FAR, as “documentation need not quantify the tradeoffs that led to the decision.” 48 C.F.R. § 15.308 (2005). The record contains an adequate, rational basis for the award to KIRA, and this ground for protest is accordingly denied.
3. The Army’s Evaluation of the FCSS Proposal
After placing undue emphasis on the labels used in the evaluation process, and then on the raw numbers of strengths and weaknesses identified by evaluators, FCSS in its third ground for protest finally gets around to issues that are central to a court’s review of procurement decisions. Fort Carson Support Services argues that the Army’s evaluation of its proposal, in several respects, was arbitrary and capricious, and unsupported by the record. In particular, it challenges the Army’s determinations relating to its staffing levels, FCSS Mot. at 39-53; the Army’s PPE evaluation, id. at 53-60; and the Army’s Cost factor analysis. Id. at 60-68.
Review concerning these issues entails identifying the judgments made by the relevant officials and verifying that the relevant information was considered, the relevant factors were employed, and a satisfactory explanation was articulated. See, e.g., Overton Park, 401 U.S. at 416, 91 S.Ct. 814; Motor Vehicle Mfrs. Ass’n, 463 U.S. at 43, 103 S.Ct. 2856. In this regard, the Court rejects the argument made by KIRA that the Army’s failure to follow the Source Selection Plan would not be a valid basis for a bid protest. See KIRA Br. at 30-31. To the contrary, the SSP provides operating rules for the evaluators, to make the process more transparent and less ad hoc. While the Army adopts this plan, and as an agency is free to change it, the subordinate officials whose activities are controlled by an SSP do not have the power to deviate from it. Unless an element of the SSP is expressly waived by an official with the power to do so, via a valid, articulated
If FCSS can prevail on any of these issues despite the great deference courts must give to agency evaluators of past performance or quality, see E.W. Bliss Co., 77 F.3d at 449; Gulf Group, 61 Fed.Cl. at 351, then the Court must determine whether the arbitrary action was prejudicial to FCSS. Bannum, 404 F.3d at 1351-54. On the factual question of prejudice, the method of evaluation employed by the Army might be problematic. Under a method that relies on numerical scores, the prejudice determination is as simple as comparing the points affected by an arbitrary decision to the spread between the protester and winning bidder. See, e.g., Beta Analytics Int’l, Inc. v. United States, 67 Fed.Cl. 384, 407-08 (2005). The narrative approach followed here, however, makes the government vulnerable to a finding of prejudice whenever a decision that appeared to have influenced the ultimate decision of best value is found arbitrary — that is, is found to rely on false statements, is contradicted by the evidence, violates procurement rules, etc. But the Army accepted this risk in employing this particular evaluation process.
a. Staffing level evaluation.
First, FCSS argues that the Army incorrectly evaluated its staffing levels and failed to properly credit it for proposed innovations. In making this argument, FCSS charges the government with failure to provide a reason for “reducing” the size of the credit for staff positions saved due to innovations that its proposal received in the initial evaluation. FCSS Mot. at 49. It is true that the Army did not specifically identify positions that it no longer felt could be credited against FCSS’s net staffing shortfall, in performing the evaluation of the final proposals. But this is because the initial decision to credit FCSS with several dozen FTEs was not itself a reasoned decision.
In the first round of evaluations, the IGE prepared by the Army determined that 263 employees were needed to perform the contract, spread throughout various classifications. See AR Tab 52(1). The Army then compared the number of employees proposed by the offerors, in each classification, to the IGE to calculate understaffing and overstaff-ing. See AR Tabs 65 (KIRA), 66(TGG), 67 (FCSS). It was determined that the initial proposal of FCSS was understaffed, relative to the IGE, by seventy-eight positions. AR Tab 67 at 2.
When this information was originally presented to the SSA by the SSEB, the number of positions proposed by FCSS appears to have been inaccurately identified as 182, as opposed to the 185 used in the IGE comparison. Compare AR Tab 69 at 36 (July 26, 2004 SSA Briefing) with AR Tab 67 at 2. The net understaffing, identified as a “variance,” as a consequence rose to eighty-one positions. See AR Tab 69 at 36. But the number of additional employees used in the MPC calculation was forty-seven, which purportedly resulted in an increase of $ 1,645,000 for the base year. Id. Since the FCSS proposal that was understaffed by a net eighty-one positions was adjusted only for a shortfall of forty-seven, the SSEB appeared to be implicitly crediting FCSS with thirty-four FTEs.
Two weeks later, the SSA was briefed again by the SSEB, this time resulting in the decision to award a contract to FCSS without discussions. See AR Tab 76. In these briefing materials, FCSS was said to have proposed 185 employees, consistent with the number used in the IGE comparison. Id. at 34. The difference between this staffing level and the IGE, or “variance,” was then calculated to be seventy-eight — the difference between 263 FTEs determined in the IGE and the number proposed by FCSS. Id. But the SSEB also calculated that 224 employees were needed for FCSS to perform the contract, apparently as part of the Army’s MPC calculation. Id. The resulting variance was thus just thirty-nine FTEs. This thirty-nine employee shortfall was then used to calculate the MPC adjustment to the FCSS proposed base period cost, which included the “[ajddition of 39 employees @ $ 2,763,359.” Id.; see also AR Tab 68 at 1.
The briefing materials do not explain how the FTE shortfall was reduced from seventy-eight to thirty-nine — or, in other words, how a thirty-nine employee credit was derived for
In the Price Negotiation Memorandum memorializing the decision to award the contract to FCSS, the Army explained the MPC methodology used in the initial evaluation: “The process tended to normalize the proposals to the IGE [for labor costs] except in the case where [FCSS] proposed efficiencies with convincing discussion which will lead to labor reductions and cost savings.” AR Tab 78 at 4. Discussing FCSS’s proposal, this memorandum reiterated nearly verbatim the language from the SSDD concerning the innovations, efficiencies, and the staffing shortfall, explaining that the latter “may be offset by the innovations proposed.” Id. at 7. The FCSS staffing credit is discussed under the “Cost” subheading: “The Management Committee determined [FCSS] would have to increase their staffing by an overall 39 FTEs based on the proposed efficiencies, in lieu of the full 78 variance compared to the IGE.” Id. No further explanation is given, and there appears to be no documentation of this decision of the SSEB.
After offerors in the competitive range were warned that “proposed staffing reflected substantial variances” that “included large numbers of instances of understaffing and a much smaller number of instances of overstaffing,” AR Tabs 115-17, and informed through Amendment 10 that “the Government will consider any innovative ideas, but proposals should clearly identify and explain ... any cost savings that rely upon those innovations,” AR Tab 149, FCSS submitted its FPR. In this final offer, FCSS did add additional language relating to proposed innovations. A paragraph, for instance, was added to the Executive Summary concerning the “one man and a truck” concept, and the use of communications and management tools that are “productive of cost savings,” although the number of staff hours saved is still deemed “countless.” AR Vol. VIII at revised 3-4. A few sentences are added to the “Work Reception & Dispatch Desk” (“WR & DD”) discussion concerning these same innovations. Id. at revised 24. A couple of lines are added concerning the use of MAXIMO Mobile Suite and Nextel TeleNav-track, see id. at rev. 38, 38a, 39, and a new chart identifies the various phases in which these technologies are to be deployed. Id. at
The FPR also adds a sentence that explains that FCSS “will significantly increase grounds maintenance efficiency and lower costs by preparing the battlefield/grounds prior to each season’s workload, and by working with tenant units to coordinate efforts.” AR Vol. VIII at rev. D-4. Corresponding to PWS C.5.4.2.1.1, regarding grounds mowing and trimming, FCSS elaborates that it can “accomplish the requirements with significantly fewer FTEs than are currently being used” by removing rock piles and rotting landscaping timber, “thoroughly prepar[ing] working areas annually,” and conducting “time and motion studies.” Id. at rev. D-9. Thirty-two pages of spread sheets are added to the Appendix C to Section II of the RFP, which FCSS explained as “its workload takeoff and estimating process, figures, assumptions and fulltime-equivalent (FTE) staffing computations.” App. to FCSS Mot., Ex. C at 5 (describing AR Vol. VIII App. C at rev. 13^4). And FCSS explains in greater detail the process it and its subcontractor followed in estimating the number of employees needed. AR Vol. VI Tab 6 at rev. 1-6.
After reviewing the new material in the FCSS proposal, the evaluator responsible for assessing the staffing levels believed that FCSS “did not show their manpower savings from innovative approaches like [KIRA].” AR Tab 189 at 1. In its evaluation of the final proposals, the Army credited FCSS with three FTEs for innovations — two for carpenter positions, and one for a plumber. See AR Tab 208 at 2-3. A credit of one-half of an FTE was also given to FCSS for the unexplained “home office effort.” AR Tab 211 at 79. KIRA received credit for 6.3 FTEs due to innovations, see AR Tab 211 at 39, although the MPC calculation breakdown shows innovation credits totaling 6.6 FTEs— two for grounds maintenance laborers, one-half for a truck driver, two for carpenters, three-tenths each for an electrician and for a stationary engineer, and 1.5 for the category comprising plumber, steamfitter and pipefit-ter positions. See AR Tab 206 at 1-2; see also Tab 197 (chart showing MPC adjustment for KIRA as 6.6 FTEs for innovations). According to the minutes from the SSEB’s final caucus concerning the Management subfactor, “[t]he evaluator reviewing the staffing stated that the credits against the IGE for innovations equate[d] to fewer for each offeror than originally expected.” AR Tab 196 at 2. The minutes then note the decision to give FCSS a credit of three FTEs, but give a slightly different total for KIRA of 6.5 FTEs. Id.
In the PA Report, the SSEB Chairman explains that “[t]he labor portion of the Most Probable Cost (MPC) was calculated for each offeror based on the analysis of its proposed manning levels in comparison to the IGE and included recognition of any savings that the SSEB considered justified in the proposal as resulting from proposed innovations.” AR Tab 209 at 5. He described the adjustment for KIRA as requiring the addition of five FTEs above the IGE to reflect extra positions needed due to KIRA’s “proposed methodology of performance,” and identifies the innovations credit received by KIRA as “6.3 man-years.” Id. at 6. The Chairman also notes that KIRA’s claim of staff savings due to efficiencies in option years two through nine were rejected by the SSEB as “not supported by documentation in the proposal.” Id. Had this claim been accepted, the additional credits would have grown to 25.7 FTEs by option year 7. See id. Regarding FCSS, the Chairman confirmed that it “was credited with 3 man-years for proposed innovations and .5 man-years for home office effort.” Id. at 8.
In the second round of evaluations, the IGE was based on the determination that employees totaling 258 FTEs were needed. See AR Tab 172(1) at 3; AR Tab 197. Due to the innovations and home office credit of 3.5 FTEs, and based on its proposed workforce totaling 182 man-years, the net under-staffing of FCSS was calculated to be a variance of 72.5 FTEs. AR Tab 211 at 79. The KIRA figure was a shortfall of 30.7 FTEs. Id. at 39. The Source Selection Authority concurred with this analysis. AR Tab 214 at 2. In the SSDD, he discussed the
In the first evaluation, both the SSEB and I gave [FCSS] a substantial credit in manpower based on its assertion of savings anticipated from proposed innovative methods of performance. However, even so, [FCSS] still had a significant manpower shortage, which was pointed out to it in discussions. Further, both in discussions and in later amendment to the Request for Proposals, all of the offerors were strongly cautioned that they needed either to clearly document any claimed savings from innovations or to raise their staffing levels in the areas the Government pointed out as low. In the second round of evaluation, the SSEB reassessed [FCSS]’s essentially still undocumented claims of innovation savings and gave them a minimal credit.
AR Tab 214 at 4.
The Court concludes that the record of the procurement decision demonstrates that the Army’s evaluation of staffing levels was not arbitrary. In the process of compiling the IGE for the work to be performed under the Solicitation, the Army initially concluded that employees totaling 263 FTEs were needed. See AR Tab 52(1). It later revised this figure to 258 FTEs. See AR Tab 172(1). In reviewing the manpower proposed by the offerors relative to the level used in the IGE, the Army credited FCSS and KIRA for staffing positions saved due to innovative approaches to the work, and specifically identified the positions credited. AR Tabs 206, 208. These credit determinations were adopted by the SSEB after consensus meetings, and were used by the SSA in making his best value determination. See AR Tab 209 at 6-8; AR Tab 211 at 39, 79; AR Tab 214 at 4-5.
The staffing level computations are attacked by FCSS in a number of ways. As was noted above, its primary argument appears to be that the Army did not explain why its innovation credit figure was reduced from the initial evaluation. This argument would have some force, were the initial decision itself properly documented and articulated. Had the SSEB initially determined, or relied upon a document which determined, that FCSS’s proposed method would need X fewer employees in one classification, Y fewer in another, Z fewer in a third, etc., then an evaluation of the final proposal that dropped out all discussion of these staff savings could well reflect an arbitrary action. Unfortunately for FCSS, the situation presented in the record is the reverse: it was the second decision that was supported by documentation, not the initial evaluation. And the second staffing and innovation determinations were the ones that influenced the final decision.
In making these determinations, the Army considered the relevant information — it made an IGE, compared the staff levels to those proposed by the offerors, and evaluated the offerors’ proposals to see whether credit for innovations should be given. The Army articulated a reason for the credits being the size they were — it felt that FCSS demonstrated the need for fewer FTEs for the positions of carpenter and plumber, see AR Tab 208 at 2-3, and that KIRA showed it could make do with fewer FTEs for the positions of grounds maintenance laborer, truck driver, carpenter, electrician, stationary engineer, plumber, steamfitter and pipe-fitter. See AR Tab 206 at 1-2. Fort Carson Support Services argues that the record should show the qualifications of the people making these decisions, precisely how they made them, and which proposed innovations were responsible for which credits. See FCSS Reply at 17-18. But such level of detail is not needed to demonstrate that the decision was the product of the relevant information being consulted. While minutiae — such as a description of each individual judgment made by each evaluator concerning the relationship between each position for which an offeror was understaffed and the innovations that offeror proposed — could aid a court in second-guessing the technical judgments of evaluators, this is not the proper role of courts. See, e.g., E.W. Bliss, 77
Fort Carson Support Services takes issue with the SSA’s conclusion that its innovation savings were “essentially still undocumented,” AR Tab 214 at 4, and points to the revisions in its proposal that discuss innovations. FCSS Reply at 18. But the language identified in the FPR does not attempt to quantify the results of any savings due to the innovations, and the spread sheet purporting to calculate the number of FTEs needed is just a spread sheet. See AR Vol. VIII App. C at rev. 13-44. The spread sheet may well demonstrate that FCSS and its subcontractor systematically and carefully calculated the number of employees they felt were needed to do the work, but when this number differed from the Army’s experience, no principle requires the Army to accept the former.
But FCSS argues that if the Army relies on its experience, the result is necessarily irrational when a cost-plus contract is at stake. The Army penalized FCSS for believing it could do the work with fewer employees than were traditionally employed, and FCSS claims the alternative was to pad the proposed staffing numbers. Essentially, FCSS argues that the SSEB’s evaluation of the staffing variance was «reasonable because it would have been “more reasonable to posit that adding manpower just to bulk up its number ... is less rational than maintaining its lower staffing ... with the resultant lower costs.” FCSS Mot. at 50 (emphasis added). The standard for reviewing an agency’s decision in this Court, however, is not whether there exists a hypothetically “more reasonable” decision, but whether the decision actually made is reasonable in light of the record before the Court. See Camp v. Pitts, 411 U.S. 138, 142, 93 S.Ct. 1241, 36 L.Ed.2d 106 (1973); Impresa Construzioni, 238 F.3d at 1332 (adopting APA standards developed by the D.C. Circuit); see also Delta Data Sys. Corp. v. Webster, 744 F.2d 197, 204 (D.C.Cir.1984). The record shows that the Army was not convinced that the methods of performance explained by FCSS warranted a credit for more than 3.5 FTEs. See, e.g., AR Tabs 196, 208, 209, 211. From the perspective of FCSS, perhaps it would not have been rational for it to propose positions it felt it did not need, even in light of the warning about understaffing that all offerors were given, and the identification of the specific work classifications for which its proposal was short. See AR Tabs 117, 147. But just because an agency disagrees with even a reasoned judgment of an offeror does not mean the agency is acting arbitrarily.
As was discussed above, Amendment 10 to the Solicitation informed offerors that “the Government will consider any innovative ideas, but proposals should clearly identify and explain ... any cost savings that rely upon those innovations.” AR Tab 149. A second round of scoring in a negotiated procurement is not per se unlawful, see Dismas Charities, Inc. v. United States, 61 Fed.Cl. 191, 204 (2004) and offerors were on notice that the Army intended to make the award after discussions. AR Tab 7 at 35 (referencing 48 C.F.R. § 52.215-1 Alternate I).
Fort Carson Support Services also argues that it “was not required to show manpower savings, just cost savings.” FCSS Reply at 14. But in the discussions, the Army clearly explained, “you have to establish in your proposals that the numbers and types of people that you propose correlate accurately to the method and means you propose for performing this Contract.” Orange (FCSS) Disc. Tr. (Dec. 15, 2004) at 38. Moreover, the Solicitation explained that the Army would “evaluate the offeror’s methodology for managing the manpower ... required to perform all tasks described in the PWS and its application of personnel ... for implementation of the technical approach for performing the tasks described in the PWS.” Solicitation § M.2.2.1.1, AR Tab 7 at 49. And while FCSS may feel that it adequately demonstrated how it estimated the staffing level needs, by providing the aforementioned workload spread sheet, or that it did a better job than KIRA in demonstrating its manpower requirement, these are the sorts of judgments that receive the greatest deference possible. See, e.g., E.W. Bliss, 77 F.3d at 449. The argument by FCSS that it was irrational for KIRA to have received more innovation FTEs than FCSS did, in light of the Army’s rejection of KIRA’s claimed efficiencies for the out years, see FCSS Mot. at 34, is also unavailing. The Army’s decision that certain staffing savings were not documented by KIRA, see AR Tab 214 at 4, has nothing to do with its determination that others were documented and, if anything, helps demonstrate that the decision was a reasoned one. The Army’s evaluation of the staffing levels and innovations proposed was not arbitrary, capricious, or unlawful.
b. Past Performance and Experience evaluation.
The next objection raised by FCSS concerns the Army’s evaluation of its PPE information. In the initial review of proposals, all offerors were assigned a “Low” risk rating for the PPE subfactor and each element of PPE. See AR Tab 76 at 16, 22, 27, 33, 38, 44, 46; AR Tab 77 at 2. After the Army “re-validate[d] PPE information,” as was announced in Amendment 10, see AR Tab 149 at 2, it determined that FCSS should receive only a “Moderate” risk rating for the Managerial and Cost Control elements of PPE, AR Tab 211 at 75, and an overall PPE risk rating of “Low/Moderate.” Id. at 76. Fort Carson Support Services argues that the worsening of its risk rating in this area was arbitrary and capricious, since the second evaluation was based on the same information as the first one. FCSS Mot. at 56. It also repeats its argument that the use of the hybrid “Low/Moderate” rating was improper. Id.
Although FCSS argues that “re-vali-dat[ing],” per Amendment 10, is different than reevaluating, see FCSS Mot. at 56 n. 54, the Army did not need to inform offerors of its intention to re-evaluate any portion of the proposals. Unless such an action was forbidden by law, by the Solicitation, by the FAR, or by the SSP, the Army was free to review proposals as many times as it felt it needed, so long as it treated all offerors fairly and equally. See Dismas, 61 Fed.Cl. at 202-03; cf. Beta Analytics Int’l, Inc. v. United States, 67 Fed.Cl. 384, 407 (2005) (holding evaluation arbitrary when only one offeror received a second evaluation). Moreover, the Court finds that this language placed offerors on notice that PPE was going to be reevaluated, for it explained “re-validat[ion]” as “including any new information that may be available for activities since the original proposal submission date.” AR Tab 149 at 2. If old, previously available information were not to be used in this process, there would have been no need for this explanation.
Once again, of course, if an initial evaluation of a proposal is properly documented and articulated, and a subsequent evaluation of the same proposal results in a different assessment, the lack of an explanation for the agency’s change of mind could well result in a finding that the second decision was arbitrary. Here, as with the staffing evaluation, there is a problem. The record contains the proper documentation of the individual SSEB member’s pre-caucus evaluation of FCSS’s PPE information, see AR Tab 62, as is called for in the SSP. See AR Tab 34 at 11. This appears to reflect a “Low” risk rating for the Performance and Business Relations elements, and a “Moderate” risk rating for the Managerial and Cost Control elements as well as for PPE overall. AR Tab 62.
In the caucus, the determination was that the information provided by the offeror and obtained by the Government supported the LOW assessment. The caucus determined that the somewhat derogatory reference concerning its Plum Island contract was offset by the labor union issues that existed before its start of work on the contract. Also, evaluation of the offeror’s ability to manage this contract was enhanced by the experience of its proposed subcontractor.
Id.
So the questions to be answered are how and why the assessment of FCSS’s PPE changed after the second evaluation, with the risk ratings for the elements Managerial and Cost Control reverting back to the individual evaluator’s assessment of “Moderate” risk, and the overall risk slipping to “Low/Moderate.” The record shows that search of a government database yielded two new evaluation reports for contracts performed by one of the FCSS joint venture partners. AR Tab 195. The individual evaluator assessing PPE information documented her findings. AR Tab 205. The two new FCSS reports “reflected strong performance through ratings of excellent, good, and outstanding,” although one’s “relevancy to the solicited procurement is questionable except for issues relating to management, customer service, cost control, etc. of which the evaluator rated the contractor highly.” Id. at 2. The evaluator concluded that the new reports “seem to support the Government’s original PPE risk assessment of LOW.” Id. The evaluator also stated that despite “current investigations related to the proposed subcontractor” for FCSS, there was “no basis for downgrading the subcontractor’s PPE.” Id.
The ultimate decisionmaker in the process, the SSA, explained the PPE evaluation of FCSS in the SSDD:
In the PPE subfactor, [FCSS] submitted information showing experience for each of its two joint venturers as a prime, although of considerably less magnitude and narrower scope than this contract. [FCSS] did not submit any PPE information for the joint venture or for its proposed main subcontractor. Considering both the information submitted by the offeror and that elicited by the SSEB from other sources, [FCSS]’s performance risk for the PPE subfactor was rated as low-to-moderate, with two elements moderate and two low.
AR Tab 214 at 3. He later explained that “[Hooking beyond the ratings to the substantive information,” he found KIRA “well above” FCSS and TGG in the PPE subfaetor assessment, as the other two offerors’ “experiences as primes were significantly less relevant than [KIRA]’s, and [FCSS]’s ratings were considerably more mixed that either of the other two offerors.” Id. at 4. In light of this record, can a reviewing court find the assessment of FCSS’s PPE to be arbitrary?
The Court concludes not, particularly given the “greatest deference possible” that must be afforded to the agency’s judgments in this area. See University Research Co., 65 Fed.Cl. at 505. The individual evaluator who considered PPE initially felt that FCSS should receive a “Moderate” performance risk rating for two elements and for the subfactor overall. AR Tab 62. After the SSEB caucus decided, without documentation, to provide FCSS with the same “Low” performance risk rating across-the-board as was given to all offerors, the SSA chose to focus solely on the Management subfactor and Cost factor in the initial evaluation. See AR Tab 77 at 2. After the offerors in the competitive range were given the opportunity to submit additional PPE information, the SSEB re-validated the information for all three offerors in the second round. See AR Tabs 193-95, 205. And in making the source selection decision, the SSA explained that he considered not just the SSEB’s risk ratings but also the PPE information underlying them. AR Tab 214 at 4.
The Court notes that nothing contained in the underlying information and analysis contradicts the SSA’s conclusions concerning PPE. He found FCSS’s experience “of considerably less magnitude and narrower scope” than the contract at issue. Id. at 3. These four contracts ranged from $500,000 to $ 8 million, AR Tab 211 at 74; see also AR Tabs 60, 61, which are indeed much smaller than the procurement at issue, for which the IGE for the first full-year was over $ 19.4 million. See AR Tab 208 at 4. By contrast, the KIRA experience was evaluated based on a $ 27 million contract, a $ 32 million contract, and contracts of its proposed subcontractor that ranged from $117 million to $ 801.8 million. AR Tab 211 at 34; see also AR Tabs 54-56. The TGG contracts evaluated ranged from $100,000 to $25.4 million in size. AR Tab 211 at 54. In re-evaluating the PPE subfactor, the SSA followed the Solicitation, which provided: “The Government will also consider the breadth and depth of experience reflected in the offeror’s
Moreover, the SSA’s decision that FCSS’s ratings “were considerably more mixed than either of the other two offerors,” AR Tab 214 at 4, is not contradicted by the record. A review of the ratings provided in response to the PPE questionnaires (which contained twenty-nine items to be rated) shows that KIRA received ninety-eight “Exceptional” ratings and thirty-six “Very Good” ratings, see AR Tab 56; TGG received fifty-four “Exceptional” ratings, nine “Very Good” ratings, and twenty-one “Satisfactory” ratings, see AR Tab 59; and FCSS received nineteen “Exceptional” ratings, fifty-three ‘Very Good” ratings, twenty-five “Satisfactory” ratings, and four “Marginal” ratings. See AR Tab 60.
c. Cost factor evaluation.
Fort Carson Support Services also disputes the manner in which the Cost factor was used by the Army in the evaluation. See FCSS Mot. at 60-68; FCSS Reply at 30-35. This would not seem to be a particularly fruitful avenue for protest, given that the evaluation methodology of the Solicitation stated that the other factor, Quality, was “significantly more important than Cost.” Solicitation § M.2.1, AR Tab 7 at 48. But FCSS argues that the Solicitation’s evaluation criteria were violated by the Army’s use of the Cost factor and its components.
The Cost factor included two subfactors, Total Estimated Cost, and Cost Realism, which were “of approximately equal importance.” Solicitation § M.2.1.1, AR Tab 7 at 48; see also id. §§ M.2.3.5, M.2.3.6, AR Tab 7 at 51. The TEC subfactor evaluation included “reasonableness,” which meant determining if proposed costs were “unreasonably high or low in relation to the offeror’s technical and management approaches and in comparison with” the IGE. Id. § M.2.3.2, AR Tab
One of FCSS’s arguments is that the Army’s cost variance evaluation was flawed in part because it was based on “a flawed assessment of FCSS’s staffing,” in which “the Army failed to justify giving FCSS only 3.5 FTEs credit for its innovations, which were the same innovations that the Army had determined were worth approximately 34 FTEs in the first evaluation.” FCSS Mot. at 63. The Court rejects this argument, for the reasons explained above concerning the staffing level evaluation. The record demonstrates that the Army followed a reasoned process, and considered the relevant information, to determine the number of innovation credits to allow each offeror. Because FCSS was given just 3 FTEs due to innovations, and 0.5 FTEs for “home office effort,” it was short 72.5 FTEs relative to the IGE. AR Tab 211 at 79. The adjustment for the labor shortfall proposed by FCSS accounts for the entire amount of the increase from its TEC to its MPC — or, in other words, its variance. See AR Tab 208 at 4.
Fort Carson Support Services’ other arguments relating to the Cost factor charge the Army with violating the terms of the Solicitation. The use of the adverb “marginally” to modify the “satisfactory” evaluation of FCSS’s cost variance is challenged, see FCSS Mot. at 65. But nothing in the Solicitation or the SSP forbids the use of such descriptors. See AR Tabs 7, 34. While the Solicitation states that “[vjariance will be evaluated as either satisfactory or substandard,” Solicitation § M.2.3.6.1, AR Tab 7 at 51, variance is a part of the realism analysis, of which the Solicitation also states:
If an offeror’s cost is evaluated as unrealistically low or high compared to the anticipated costs of performance; and/or the offeror fails to recognize and/or mitigate technical, schedule, and cost risks; and/or the offeror fails to adequately explain the proposed costs, the Government will reflect the inherent findings, risks and associated cost impacts in the Cost Realism Analysis.
Id. § M.2.3.3 AR Tab 7 at 51. Thus, the evaluation was designed to be more than the mere generation of the labels “satisfactory or substandard,” and the use of a modifier does not conflict with the Solicitation. The argument is also pressed that variance is just one of three elements of one of two subfactors of Cost, and thus was given a more prominent role in the evaluation than its one-sixth share of the process would warrant. FCSS Mot. at 66-68. But the Solicitation clearly informed prospective offerors that the source selection “decision will not be made by the application of a predefined formula, but by the conduct of a tradeoff process among the factors, sub-factors, and elements identified” in the Solicitation. Solicitation § M.1.1, AR Tab 7 at 48. It was certainly proper for the Army to consider FCSS’s cost variance element evaluation in making the best value tradeoff.
Moreover, offerors were also warned that “[a]n offeror’s eligibility for award may be significantly downgraded if the offeror’s proposed costs and technical approach result in significant cost realism adjustments.” Id. § M.2.3.6, AR Tab 7 at 51 (emphasis added). And the cost variance evaluation specifically involved “consideration given to the percentage of the variance, the dollar amount of the variance, the anticipated impact on contract performance, and the offeror’s proposed method of performance.” Id. § M.2.3.6.1 AR Tab 7 at 51. The variance of FCSS was calculated to be $32.7 million over the life of the contract, and 15.7% of the FCSS proposal’s probable cost — much higher than the other two offerors. See AR Tab 211 at 83. Under the Solicitation’s evaluation methodology, it was hardly arbitrary for the evalua
U- Did the Army Conduct Meaningful Discussions with FCSS?
The final challenge brought by FCSS to the Army’s award of the contract to KIRA concerns alleged violations of an applicable regulation. Fort Carson Support Services argues that the Army failed to engage in meaningful discussions with it, in violation of 48 C.F.R. § 15.306. See FCSS Mot. at 68-84. This provision of the FAR concerns discussions with offerors in the competitive range, and requires that, “[a]t a minimum, the contracting officer must ... indicate to, or discuss with, each offeror still being considered for award, deficiencies, significant weaknesses, and adverse past performance information to which the offeror has not yet had an opportunity to respond.” Discussions are meaningful if “they generally lead offerors into the areas of their proposals requiring amplification or correction, which means that discussions should be as specific as practical considerations permit.” WorldTravelService v. United States, 49 Fed.Cl. 431, 439 (2001) (quoting Advanced Data Concepts, Inc. v. United States, 43 Fed.Cl. 410, 422 (1999), aff'd, 216 F.3d 1054 (Fed.Cir.2000)).
a. Staffing levels.
Two aspects of the discussions concerning its staffing levels are raised by FCSS. It concedes that “during discussions, the Army told FCSS of the areas in which it considered FCSS’s staffing to be either high or low,” and that “the Army told FCSS that it was only identifying those areas where the under- or over-staffing was significant.” FCSS Mot. at 70. Indeed, all offerors were told in writing that their “proposed staffing reflected substantial variances from that which the Government considered necessary,” which “included large numbers of instances of understaffing.” AR Tab 117 at 2; see also AR Tabs 115, 116. As a follow-up to the discussion, each offeror was sent an email confirming the PWS categories for which its proposal was understaffed or overstaffed. AR Tabs 143,146,147. And during the discussion with FCSS, the Army revealed the Most Probable Cost figure for the FCSS proposal and explained that the “gross difference” between this number and FCSS’s proposed cost reflected the net amount of under-staffing. Orange (FCSS) Disc. Tr. (Dec. 15, 2004) at 37.
But FCSS argues that the Army did not more precisely spell out the understaffing amounts, and that “the Army never gave FCSS any reason to believe that its staffing overall constituted a significant weakness or even a weakness or problem.” FCSS Mot. at 70. It frames this argument in light of the fact that it was initially awarded the contract, and thus had no reason to believe the Army found serious problems with its staffing approach. Id. at 70-71. But the Army clearly conveyed to FCSS that it was understaffed for eight of the seventeen work categories specified in the PWS, see AR Tab 13, and overstaffed in just two. AR Tab 147. And the Army informed FCSS that the MPC calculated in the initial evaluation was a figure exceeding FCSS’s proposed costs by more than $ 26.3 million, which was due entirely to net understaffing. See Orange (FCSS) Disc. Tr. (Dec. 15, 2004) at 37; AR Tab 76 at 34. The offeror without question “was placed on notice of significant identified management problems. That renders the discussions meaningful under the FAR.” Cubic Defense Sys., Inc. v. United States, 45 Fed.Cl. 450, 471 (1999) (internal quote omitted). The Army specifically identified areas of understaffing and conveyed the seriousness of the issue, satisfying its obligation under the FAR, as it was “not required to recommend a particular level of staffing.” WorldTravelService, 49 Fed.Cl. at 440 (inter
The other staffing issue concerns FCSS’s Work Reception and Dispatch Desk plan. See FCSS Mot. at 71-83. The only aspect of this plan identified as a weakness during discussions concerned the person to whom calls were to be sent after the desk closed at eleven p.m. See AR Tab 117 at 2 (weakness described as “[possible gap in coverage if plant operator is occupied with plant duties”); Orange (FCSS) Disc. Tr. (Dec. 15, 2004) at 18-19. But in the FPR evaluation, the fact that the WR & DD was to be staffed after 3:30 p.m. was identified as a weakness. See FCSS Mot. at 77; AR Tab 204 (comment number 3650). The only change in this aspect of the plan was to increase by one hour, until midnight, the staffing of the desk. The offeror accuses the Army of a “bait-and-switch” in identifying one weakness to be addressed, but ultimately finding another. FCSS Mot. at 71.
It appears that the SSEB initially found the concept of extended hours for the WR & DD to be a strength, see AR 76 at 31 (listing among strengths: “Good concept on work reception desk manned from 0700 to 2300 hrs”) but then made a complete U-turn and found it to be a weakness. See AR Tab 211 at 63 (listing as a weakness: “Manning of work reception desk until midnight — extra cost.”). One doubts that the marginal addition of the eleventh hour was the Army’s concern. But in any event, the Court does not find the Army’s failure to raise this aspect of the proposal during discussions to be a violation of 48 C.F.R. § 15.306(d)(3). As FCSS’s own brief shows, the Army addressed the staffing of FCSS’s WR & DD plan during discussions. FCSS Mot. at 75-76, 78-83; Orange (FCSS) Disc. Tr. (Dec. 15, 2004) at 18-20, 29-35, 69-70. This sufficed to “generally lead” FCSS “into the areas of [its] proposal ] requiring amplification or correction.” WorldTravelService, 49 Fed.Cl. at 439. Moreover, the FAR provision requires a discussion of “significant weaknesses,” not all weaknesses. 48 C.F.R. § 15.306(d)(3) (2005). And even after the second evaluation, the WR & DD plan was considered merely a weakness, not a significant one. AR Tab 211 at 63.
But more to the point, nothing in the record demonstrates that the extended hours aspect of the WR & DD plan was considered by the Army to be a weakness before the FPRs were reviewed, so discussions concerning weaknesses could not have included it.
b. Adverse past performance information.
The Army’s discussions with FCSS are also faulted for failing to cover certain past performance information that was considered
We did not revise the rule to permit offer-ors to address past performance information to which they have already had an opportunity to respond because the solicitation provides offerors with the opportunity to address problems encountered on previous contracts and related corrective actions. In addition, FAR Subpart 42.15, Contractor performance information, already contains formal rebuttal procedures.
62 Fed.Reg. 51,224, 51,228 (Sept. 30, 1997).
i) Plum Island
The first past performance information that FCSS argues should have been discussed pertains to the operations and maintenance support contract performed by Olgoon-ik, one of the FCSS joint venture partners, at the Plum Island Animal Disease Center. Fort Carson Support Services asked an official at the U.S. Department of Agriculture’s Agricultural Research Service, which ran the Center at the time of the contract, to complete the past performance questionnaire for this procurement, and listed the Plum Island contract as relevant experience. See AR Tab 60; AR Vol. VI, Tab 2 at 12. Olgoonik’s work on the contract had ended on December 31, 2003, and the questionnaire was completed on May 27, 2004. See AR Tab 60.
The respondent for the Plum Island contract provided a rating for twenty-seven of the twenty-nine questions for which a rating was requested. For three questions, he rated performance to be “Very Good,” and for twenty, he rated the performance “Satisfactory.” Id. In response to two questions, the respondent circled both “Satisfactory” and “Marginal,”
The SSEB member responsible for the initial PPE evaluation recommended that FCSS receive a “Moderate” performance risk rating for PPE overall and for the elements Managerial and Cost Control. See Tab 62. The Plum Island ratings were mentioned as a factor in these determinations that “some” doubt existed concerning successful performance. Id. Athough it went undocumented, see AR Tab 209 at 4, the SSEB’s initial consensus was that FCSS should nonetheless receive a “Low” performance risk rating for the PPE subfactor and each element. See AR Tab 76 at 33. The SSEB Chairman later explained that “[t]he caucus determined that the somewhat derogatory reference concerning its Plum Island contract was offset by the labor union issues that existed before its start of work on the contract.” AR Tab 209 at 4. The question
To answer the question, we must first determine whether the information was “adverse.” For the PPE questionnaire, the worst rating that could be provided was “Unsatisfactory.” See AR Tab 19 at 3. The rating of “Marginal” was defined as follows:
Performance does/did not meet some contractual requirements. The contractual performance of the element or sub-element being assessed reflects a serious problem for which the contractor has not yet identified corrective actions. The contractor’s proposed actions appear only marginally effective or were not fully implemented.
Id. Does the failure to meet “some” requirements, or “a serious problem” in performance amount to an adverse rating? There does not appear to be any case law on point. One opinion of our Court held that evaluations needed to be “deemed ‘poor’ or otherwise deficient” to be adverse for the purposes of this FAR provision. Bannum, Inc. v. United States, 60 Fed.Cl. 718, 729 (2004). A GAO decision suggests that a “marginal” rating would be adverse past performance information. See NMS Management, Inc., B-286335, 2000 CPD ¶ 197, 2000 WL 1775243, at *3 (Comp.Gen. Nov.24, 2000). Another suggests that adverse means an evaluation that “reflected a finding of a history of performance problems.” See Standard Communications, Inc., B-296972, 2005 CPD ¶ 200, 2005 WL 3078884, at *6 (Comp.Gen. Nov.1, 2005).
The Court considers this a close question. But it appears that the SSEB’s PPE evaluator was influenced by the Plum Island ratings in determining that “some” doubt of successful performance existed. See AR Tab 62. And the record does not contain any document preexisting the discussions that reflects a consensus opinion that the “Marginal” ratings were mitigated by inherited labor conditions. In light of the above, and the characterization of the evaluation as “somewhat derogatory” — and considering that the definition of “Marginal” includes serious performance problems and performance failures — the Court concludes that the Plum Island evaluation contained adverse past performance information.
But in order for FCSS to have been entitled to notice of this information in the discussions with the Army, FCSS must show that at the time of the procurement process, it had “not yet had an opportunity to respond” to this information. 48 C.F.R. § 15.306(d)(3) (2005). The information pertained to a contract which had expired four months before FCSS asked the respondent to complete the questionnaire. See AR Tab 60. The Plum Island contract was with the federal government, so the FAR provisions entitling the contractor to review of its performance evaluation applied. See 48 C.F.R. §§ 42.1501-03 (2005). This information was submitted at FCSS’s request, and FCSS could have provided an explanation of the Plum Island performance in its proposal. But most importantly, no information in the record shows that Olgoonik had “not yet had an opportunity to respond” to its Plum Island performance evaluation. Fort Carson Support Services argues the contrary position, that nothing in the record establishes that Olgoonik knew of or had the opportunity to this information, and contends that “[t]he burden of compliance with FAR 15.306(d)(3) is not FCSS’s, it is the Army’s.” FCSS Reply at 85-86.
But while the FAR provision indeed imposes a requirement on the Army, FCSS must show a “clear and prejudicial violation” of the regulation to prevail. See Impresa Construzioni, 238 F.3d at 1333. This Court has held that, in bid protests, “the record may be supplemented with (and through discovery a party may seek) relevant information that by its very nature would not be found in an agency record.” Orion Int’l Techs. v. United States, 60 Fed.Cl. 338, 343 (2004); see also Beta Analytics Int’l, Inc. v. United States, 61 Fed.Cl. 223, 225 (2004). Evidence of the lack of an opportunity to respond to past performance information fits this description, and if FCSS possessed any relevant information on this point, FCSS could have supplemented the record with it— in the form of a declaration from the relevant
ii) Kellogg Brown & Root
The other past performance information which FCSS contends should have been a topic of discussion relates to the work of its proposed subcontractor, KBR. On January 26, 2005 — five days after the FPRs were filed, and six weeks after the discussions were conducted — KIRA, acting through its counsel, sent a letter to the Contracting Officer. AR Tab 179.
Without question, this past performance information must be considered “adverse.” According to KIRA, it concerned “deficiencies” and “one of the worst contractor performances in recent memory.” AR Tab 179 at 1, 2. The Chairman of the SSEB described KIRA’s attachment as “a derogatory report.” AR Tab 209 at 4. He also stated that “the report did not indicate an opportunity for the subcontractor to respond to its findings.” Id.
KIRA’s submission of the audit report had its intended effect, as the Chairman acknowledged that “[t]he SSEB considered that information” in evaluating FCSS’s past performance. AR Tab 209 at 4. Under these circumstances, then, adverse past performance information concerning FCSS was used in the procurement process, and FCSS had neither an opportunity to respond to the information nor any discussion of this information with the Army. This is a clear violation of 48 C.F.R. § 15.306(d)(3). But a violation must be not only clear, but prejudicial, if it is to be the basis for overturning a procurement decision. Impresa Construzioni, 238 F.3d at 1333. The record establishes that this violation was not prejudicial to FCSS’s chance to receive the contract award.
The Chairman explained that the SSEB considered the KBR audit report, “but decided that it would not adversely affect [FCSS]’s performance risk assessment.” AR Tab 209 at 4. He stated that the “primary reason” for this decision “was that the SSEB felt that it was not reasonable to use work performed in a foreign, chaotic, wartime situation as a basis for assessing ability to perform in a stateside, garrison context such as under this contract.” Id. He added that the “report reflected disagreement with its findings by the Defense Contract Management
The Court notes that KIRA’s submission of the letter and audit report was certainly contrary to the spirit of competitive bidding, a process the government uses to provide offerors with the incentive to improve the quality and price of their bids, see Arch Chems., 64 Fed.Cl. at 400, not the opportunity to conduct a negative campaign impugning the abilities of one’s competitors. But while the Solicitation does not seem to contemplate such actions, the Court is not aware of any provisions in it or the FAR that would prohibit this behavior, either.
C. The Ginn Group’s Protest
In his Source Selection Decision, the SSA concluded that TGG’s proposal ranked “third of the three offerors.” AR Tab 214 at 4. The Ginn Group’s proposed costs and MPC were the highest of the three. Id. For the “significantly more important” Quality factor, see Solicitation § M.2.1, AR Tab 7 at 48, TGG was also ranked third, as the SSA “considered [TGGj’s PPE to be somewhat better than [FCSSJ’s, but not enough to offset [FCSSj’s considerably better Management subfactor proposal.” AR Tab 214 at 4. The Ginn Group was assigned the adjectival rating of “Satisfactory” and the proposal risk rating of “Moderate” for the Management subfactor overall, id. at 2, but its Quality Control element was evaluated as deficient, with an “Unsatisfactory” adjectival rating and a “High” proposal risk rating. Id. at 3. The SSA explained that “the Quality Control plans provided in its proposal, as revised, lacked detail sufficient to reflect an understanding of the [Solicitation] requirements.” Id.
The Ginn Group raises three grounds for protest.
At the outset, the Court notes that TGG attempts to shoe-horn this matter to fit within the analytical framework employed by this Court in Beta Analytics Int’l, Inc. v. United States, 67 Fed.Cl. 384 (2005). See, e.g., TGG Mot. at 22, 26-32. In Beta Analytics, the evaluations and notes of individual evaluators were relevant to our Court’s review, because the evaluation process employed therein required offerors to be compared on the basis of the average of the scores provided by the individual evaluators. Beta Analytics, 67
The evaluation methodology used in the procurement decision at issue here differs from that employed in Beta Analytics in many critical respects. The SSA’s best value determination was not based on the rigid use of scores supplied by other officials, but instead involved “a tradeoff process among the factors, subfactors, and elements” defined in the Solicitation, and “the exercise of ... sound business judgment.” Solicitation § M.1.1, Tab 7 at 48. The Solicitation expressly provided that this “decision will not be made by the application of a predefined formula.” Id. The Source Selection Plan charged the SSEB to produce “consensus ratings,” and required that these ratings “shall not be mathematically computed based on the individual evaluations, but rather shall be established by agreement among the SSEB members.” AR 34 at 8; see also id. at 9,11. And, as was explained above, far from generating numerical scores of any kind, the evaluation process used adjectival and risk ratings as tools to aid in the assessment of proposals, rather than as ends in themselves.
In an evaluation process in which the ultimate decisionmaker was required to “utilize all evaluation reports and analysis in determining the successful proposal,” AR 34 at 14, and in which he possessed the discretion to “subjectively” reach his own best value determination, see Solicitation § M.1.2, AR Tab 7 at 48, circumstances calling for a review of the judgments of individual evaluators are less likely to be presented. And these circumstances are further limited by the particular types of assessments involved in this matter, as “the amount of deference given to a particular decision will depend on how objective the associated criteria for scoring happened to be.” Beta Analytics, 67 Fed.Cl. at 399. Of course, a Court may always review the rationale of the ultimate decision-maker, or of any other official or group that is expressly relied upon by that decisionmaker, “to see if a necessary element for a score is missing,” id., but it cannot “second guess” the “discretionary determinations of procurement officials” concerning the “minutiae” involved in “technical ratings.” E.W. Bliss, 77 F.3d at 449.
1. Was the Evaluation of the Quality Control Element Arbitrary or Unlawful?
The Quality Control element of TGG’s proposal was evaluated as deficient, disqualifying TGG for award under the SSP. See AR 34 at 3 (explaining that a deficiency “increases the risk of unsuccessful contract performance to an unacceptable level”). Thus, this element is particularly critical to the success of TGG’s protest.
a. The Army’s evaluation of the QC element.
The first argument of TGG concerning the evaluation of the QC element is that the SSA misinterpreted the SSEB’s criticism that TGG’s QC plan “lacked sufficient detail to be evaluated.” TGG Mot. at 14-15 (quoting AR Tab 211 at 51); see also AR Tab 181 at 2; AR Tab 202 at 1. In the Source Selection Decision Document, the SSA explained that “the Quality Control plans provided in [TGG’s] proposal, as revised, lacked detail sufficient to reflect an understanding of the [Solicitation’s] requirements.” AR Tab 214 at 3. But these two findings are not inconsistent. The Solicitation required offerors to “[d]escribe in detail the approach to deficiency prevention and documentation and how the QC system will be used to achieve constant improvement in the performance of all PWS tasks.” Solicitation § L.8.6.5, AR Tab 7 at 43. Offerors were told that “[t]he Government will evaluate how the offeror’s overall approach to quality control will be put into place to meet or exceed the require-
The Ginn Group’s other arguments concerning the evaluation of its QC plan ask the Court to second guess a number of determinations made by the SSEB and adopted by the SSA. When TGG was informed, by letter, that its proposal had been included in the competitive range, it also received a list of discussion topics, organized under the headings “Uncertainties,” ‘Weaknesses,” “Significant Weaknesses,” and “Deficiencies.” AR Tab 116. The Army list of these problems included a weakness in the QC plan: “States critical work will receive [XXXX] but did not provide sufficient detail to enable the evaluation of the proposed method of inspection.” Id. (list at 2). Two deficiencies regarding the QC plan were expressly identified. The first stated that the QC plan “for Grounds Maintenance was copied entirely from the PWS and only covered what is required for Grounds Maintenance and did not provide any detail on how this area would be monitored,” and added the criticism that “[t]he QC plans did not provide details on how they integrate with the offeror’s overall QC programs.” Id. (list at 3) The other deficiency noted that the QC plan “for Service Orders did not explain the proper use of random sampling for service orders and has cited the use of MIL-STD 105E that is an obsolete standard.” Id.
In response to these criticisms, TGG revised its QC proposal, but the Army still regarded TGG’s QC proposal as “Unsatisfactory” with a “High” risk. The Ginn Group takes aim at the specific findings of the SSEB, memorialized in the caucus session minutes. See TGG Mot. at 18-20. In addition to the deficiency, described above, of a “Mack of sufficient detail within the Quality Control plans for evaluation,” the SSEB also found that “the proposed plan does not explain how to determine AQL’s or to perform random sampling,” and that TGG deleted the obsolete standard referenced in its initial proposal “but did not replace it with either sufficient description to explain the program or with another standard, which would have supplied adequate AQL’s.” AR Tab 181 at 1-2.
The Ginn Group argues that its QC plan description was sufficiently detailed. TGG Mot. at 19. It contends that statements such as “random sampling will be used to select and measure a representative group of performance units,” along with a chart specifying the various percentages of performance objectives to be randomly sampled, suffices to show how random sampling will be performed. Id. at 20 (quoting AR Vol. XIV, at Appendix 8-4, 9-4, and 10-4). And it asserts that it needs no standard as a source for AQLs, for it proposes a [XXXXXXXX], Id. at 21. The Ginn Group invites the Court to conduct a “simple review of [the details of] TGG’s Quality Control Plans,” TGG Mot. at 19, and “a comparison of TGG’s and [KIRA]’s Quality Control plans.” Id. at 20. But the sort of judgments that TGG wants reviewed — whether a plan is detailed enough to be evaluated; whether a system for determining AQLs is necessary; whether defect levels are set too low; or whether the manner of performing random sampling is adequately described — are quintessential minutiae of the evaluation process.
b. The discussions concerning the QC proposal.
The Ginn Group also alleges that the Army failed to engage in meaningful discussions about its QC plan, and that “the Army never indicated to TGG the level of concern that the Army had regarding TGG’s quality control plans.” TGG Mot. at 15 (emphasis added). The latter statement is premised on TGG’s description of the attachment to the competitive range letter as just a “list of discussion questions,” id., when the items concerning the QC element were designated as one “Weakness” and two “Deficiencies.” See AR Tab 116. Calling QC plan shortcomings “deficiencies” and a “weakness” seems to adequately communicate a serious level of concern.
As was discussed above, the FAR requires, when discussions are held, that “[a]t a minimum, the contracting officer must ... indicate to, or discuss with, each offeror still being considered for award, deficiencies, [and] significant weaknesses.” 48 C.F.R. § 15.306(d)(3). To be meaningful, discussions must “generally lead offerors into the areas of their proposals requiring amplification or correction,” and “should be as specific as practical considerations permit.” WorldTravelService, 49 Fed.Cl. at 439 (citation omitted). The GAO has described meaningful discussions as not requiring “that an agency must spoon feed a vendor as to each and every item that must be revised or addressed to improve the submission.” Uniband, Inc., B289305, 2002 CPD ¶ 51, 2002 WL 393059, at *11 (Comp.Gen. Feb. 8, 2002) (internal quotations and citations omitted).
According to TGG, “the Army informed TGG during ... discussions that TGG could address the agency’s concerns with TGG’s Quality Control plans” by defining “critical” work, including in each individual plan a description of how its individual QC plans integrate with its overall QC plan, and including AQLs in its proposal. TGG Mot. at 23. The Ginn Group argues that since it made these revisions, but still received an “Unsatisfactory” adjectival rating and a “High” proposal risk rating for the QC element, “there is no conclusion but that the Army misled TGG.” Id. But the Army did not, during the discussions, ever claim that all TGG had to do was make the changes discussed to have a satisfactory QC plan. See Blue (TGG) Disc. Tr. (Dec. 15, 2004) at 13-15; see also AR Tab 130 at 3 (prepared statement for discussions) (the Army tells the offerors “we are not here to tell you how to respond to those deficiencies or weaknesses or how, otherwise, you may improve your proposal”).
The Ginn Group also argues that it was misleading for the Army to have been so specific about problems with TGG’s QC proposal, so that TGG believed it could improve the proposal by merely focusing on the specific areas discussed. See TGG Reply at 8 (citing Analytical & Research Tech., Inc. v. United States, 39 Fed.Cl. 34 (1997)). It “acknowledges that it might now have no basis for protest” if “the Army merely informed TGG of its ultimate purported conclusion, that TGG’s Quality Control plans ‘lacked detail.’” Id. at 8. The Ginn Group contends that since the lack of detail was raised only as relating to the “[XXXXXX]” proposal, see id. at 9 & n. 3, and that it clarified this as requested, the Army cannot use the lack of detail as the reason the QC plan is deficient. But since the Army has no obligation to help an offeror rewrite its proposal, it could not be required, during discussions, to explain how to write a detailed QC proposal. Moreover, it appears that the Army’s concern about the lack of detail in the final proposal related to the random sampling and AQL provisions that were added after discussions occurred. See AR Tab 181. Discussions need not employ clairvoyance to be meaningful.
The Solicitation required offerors to provide the Army with a “practical, straightforward, specific, concise, and complete” QC plan. Solicitation § L.8.6.5 AR Tab 7 at 43. The Army provided TGG with a list of weaknesses, significant weaknesses, and deficiencies identifying areas of the initial proposal that needed improvement. AR Tab 116. If TGG failed to provide sufficient detail in its final proposal, this was due to its own business judgment, as the Army is not obliged to explain the amount of detail TGG’s proposal required. The Ginn Group has failed to demonstrate a clear and prejudicial violation of 48 C.F.R. § 15.306(d)(3).
Although TGG’s failure to prevail on this ground for protest may well moot the protest as a whole — since the deficiency under the QC element stands — the Court will nevertheless consider the remainder of TGG’s arguments, in turn.
2. Was the Army’s Evaluation of the TGG and KIRA Proposals Consistent and in Accordance with the Solicitation?
The Ginn Group also argues that the Army’s evaluation of its FPR was contrary to the terms of the Solicitation, and inconsistent with the Army’s evaluation of KIRA’s FPR. TGG Mot. at 24-31. It criticizes the Army’s overall evaluation of the Management subfactor, based on alleged problems with the evaluation of three elements of that subfactor— Phase-in Planning, Contract Administration, and Human Resources.
a. The Phase-in-Planning element.
In its presentation to the SSA, the SSEB commented that the strengths of TGG’s Phase-in Planning were: “All aspects strong.” AR Tab 211 at 45. The Ginn Group contends that this is the equivalent of finding that the proposal “offer[ed] significant advantages to the Government,” see TGG Mot. at 25, which is part of the description of the adjectival rating “Outstanding.” See Solicitation § M.3.1, AR Tab 7 at 52. Thus, it argues, the Army acted arbitrarily in merely rating this element as “Good.” TGG Mot. at 25. But a Court is hardly in a position to decide whether a proposal should be rated “Outstanding” rather than “Good,” even given that the proposal is strong in all aspects. Considering that a “Good” proposal is one “offering] some advantages to the Government,” how is the Court to decide whether some or significant advantages are
The Ginn Group further argues the Army was inconsistent in its evaluation of this element, as KIRA received an “Outstanding” rating although it was evaluated as having “only” two identified strengths (and no weaknesses). TGG Mot. at 26; see AR Tab 211 at 24-25. But neither the SSP nor the Solicitation employed a methodology whereby the raw numbers of strengths and weaknesses were added or subtracted to determine the rating an offeror received. The Court cannot say that it is irrational for a proposal with a couple of identified strengths to be better than one that was strong in all aspects, as nothing decrees that all strengths are created equal.
b. The Contract Administration element.
The Ginn Group makes a similar argument regarding the Contract Administration element. For this element, it received a “Satisfactory” adjectival rating and a “Moderate” proposal risk rating. AR Tab 211 at 46. Because the SSEB identified one specific strength and no weaknesses relating to this element, id. at 47, TGG argues that these ratings were irrational. A “Satisfactory” proposal was described in the Solicitation as one that “offerfed] no particular advantages to the Government,” Solicitation § M.3.1, AR Tab 7 at 52; a “Moderate” risk rating meant a proposal was “[ljikely to cause some disruption of schedule, increase in cost, or degradation of performance” and would “require [a] continuous level of contractor emphasis.” Id. § M.3.2, AR Tab 7 at 53. The Ginn Group reasons that a strength must be a particular advantage, and that the lack of any identified weaknesses means the conditions for a “Moderate” level of risk are not present.
But this, again, is not the sort of judgment a Court can make. In giving TGG the ratings it did, the Army implicitly found that these descriptions “most closely describ[ed] the actual assessment,” Id. § M.3, AR Tab 7 at 52, and these ratings — as well as the identified strengths and weaknesses — are tools to aid the Army in assessing proposals, not ends in themselves.
The Ginn Group also argues that KIRA’s evaluation under this element was inconsistent with the Solicitation criteria. TGG Mot. at 28-30. KIRA received a “Good” adjectival rating and a “Low” proposal risk rating for Contract Administration. AR Tab 211 at 26. KIRA had three identified strengths, but also what TGG characterized as “a relatively serious weakness.” TGG Mot. at 29 (citing AR Tab 211 at 27). This weakness was described in SSA briefing materials as: “Proposed work schedule is not in agreement with the PWS or the current CBA. Could result in an inability to meet PWS requirements and reduced oversight by the Government.” AR Tab 211 at 27. The Ginn Group argues that this weakness causes KIRA to fail the test for a “Good” rating, which was described as a proposal “fully meeting the requirements of the” Solicitation and having “a high probability of meeting or exceeding the requirements.” Solicitation § M.3.1 AR Tab 7 at 52; see TGG Mot. at 29. But the finding that an aspect of KIRA’s proposal “could result in an inability to meet PWS requirements” does not necessarily conflict with a finding of a high probability of meeting these requirements — a could, after all, may still reflect a relatively low probability. Nor is the finding that something could re-
c. The Human Resources element.
For the Human Resources element, TGG received a “Moderate” proposal risk rating. The Ginn Group argues, as it did concerning the Contract Administration element, that the lack of identified weaknesses for this element, see AR Tab 211 at 49, makes the “Moderate” risk rating irrational. The Court rejects this argument, for the same reasons stated above. The Army did not adopt a method of evaluation in which ratings were mechanistically determined by the raw numbers of strengths and weaknesses, and it is not this Court’s proper function to substitute its judgment for the technical evaluations of the contracting officials. Absent a showing that the evaluation was arbitrary, capricious, or otherwise not in accordance with law, the Court must defer to the Army’s assignment of a particular rating during the evaluation process. See E.W. Bliss, 77 F.3d at 449; Overstreet Elec. Co., 59 Fed.Cl. at 117.
3. The Best Value Determination was not Arbitrary
The Ginn Group’s argument that its proposal was arbitrarily evaluated under the Management subfactor was based on the cumulative effect of the alleged problems with each of the four elements discussed above. TGG Mot. at 30-31. Since the Court does not find anything unlawful or arbitrary in the manner in which these individual elements were evaluated, TGG’s claim concerning the overall subfactor must necessarily fail. The assertion by TGG that the best value determination was arbitrary and capricious is similarly based on its contentions concerning the conduct of the SSEB in evaluating four Management subfactor elements, see id. at 31-32, and is thus rejected for the same reasons.
III. CONCLUSION
For the foregoing reasons, the Court DENIES plaintiffs’ motions for judgment on the administrative record and for injunctive relief, and correspondingly GRANTS defendant’s and intervenor’s cross-motions for judgment on the administrative record. The Clerk of the Court is directed to enter judgment accordingly.
IT IS SO ORDERED.
. The Quality Committee seems, probably because it evaluated only the Management subfactor, to also have been called the "Management Committee.” See Price Negotiation Memorandum, AR Tab 78.
. The major difference between the two is the omission, in the body of the SSP, of any discussion of the Total Estimated Cost subfactor. See Attachment C to SSP, § M.2.3.5, AR Tab 34.
. "Deficiency" and "Weakness" were defined as in the FAR, with an example added for the former. Compare SSP §§ I.F(l) and (3), AR Tab 34 at 3-4 with 48 C.F.R. § 15.001. "Strength” was defined as "[a]n aspect of a proposal that appreciably decreases the risk of unsuccessful contract performance or that represents a significant benefit to the Government.” SSP § I.F(2), AR Tab 34 at 4.
. In the actual evaluations, the SSA followed the relative importance given the subfactors and elements in the Solicitation, and thus this error in the SSP was of no import. See AR Tab 77 at 1; AR Tab 214 at 1.
. The Chairman of the SSEB apparently did not prepare a PA Report for the initial evaluation. See AR Tab 209 V 2.a. (Mar. 30, 2005 Proposal Analysis) (Chairman admits he "did not prepare a separate Proposal Analysis for that stage of the evaluation").
. For purposes of the evaluation, FCSS was given the color "Orange” to mask its identity from the SSA. KIRA was identified as offeror "Brown,” and The Ginn Group as offeror "Blue.” See AR Tab 176 (Jan. 24, 2005 Mem.).
. Under the SSP, "to be eligible for award, a proposal must receive at least a 'Satisfactory’ rating for the Management subfactor (i.e., reflecting that the proposal meets the minimum requirements of the RFP).” SSP § III.B(5), AR Tab 34 at 10.
. Chenega was identified as offeror "Red” in the evaluation.
. Protests challenging FCSS’s HUBZone status were apparently also filed with the Small Business Administration, and were denied. See AR Tab 209 at 1.
. Chenega withdrew from the competition for the contract on January 13, 2005. See AR Tab 209 at 2.
. See AR Tabs 130, 131 (Contracting Officer’s scripts); see also Orange (FCSS) Disc. Tr. (Dec. 15, 2004) at 37-38; Blue (TGG) Disc. Tr. (Dec. 15, 2004) at 16-17; Brown (KIRA) Disc. Tr. (Dec. 16, 2004) at 78-79.
. This amendment also shortened the base period, from the originally-contemplated one-year, see Solicitation § F.2, AR Tab 7 at 14, to four months. AR Tab 149 at 2. As a consequence, the Army used the first option year, rather than the base period, as its basis for calculating MPC in the second evaluation. See, e.g., AR Tab 208 at 4.
. See AR Vol. VI Tabs 4-6 and Vol. VIII (FPR of FCSS); AR Vol. IX Tabs 4-6 and Vol. XI (FPR of KIRA); AR Vol. XII Tabs 4-6 and Vol. XIV (FPR of TGG).
. There was a discrepancy in the presentation of the FCSS MPC information, as the briefing chart showed that the offeror’s FPR was understaffed by ninety-five positions, as opposed to the
. The government ultimately submitted a fifteenth volume, and the Second Revised Volume XV was filed on August 31, 2005.
. It is not clear whether this standard comes from 5 U.S.C. § 706(2)(A), as being “otherwise not in accordance with law,” or under 5 U.S.C. § 706(2)(D), as an example of an action taken "without observance of procedure required by law.” Presumably, 5 U.S.C. § 706(2)(B)-(D) are mere elaborations on the 5 U.S.C. § 706(2)(A) theme, and thus apply in all bid protests.
. In the FPR, FCSS proposed 182 FTEs. See AR Vol. VI, Tab 6. It appears that 185 FTEs were proposed initially. See AR Tab 67.
. See Tab 209 at 5-6 (PA Report explaining ODC adjustments); compare AR Tab 68 with AR Tabs 206, 207, 208. A direct comparison of the two different MPC adjustments is complicated by a change in the baseline, from the base period used in the first evaluation, to the first option year used in the second evaluation. But it appears that the baseline MPC adjustment to FCSS's estimated costs due to understaffing increased from $ 2.76 million to $ 3.04 million, while the deduction for ODC fell from $ 0.89 million to $ 0.08 million. Compare AR Tab 68 with AR Tab 208. Thus, of the baseline MPC increase of $ 1.2 million, ODC contributed $ 0.81 million and understaffing only $ 0.28 million.
. In its Complaint, FCSS raises one additional argument&emdash;that an experienced evaluator who retired from government service between the first and second evaluations was not replaced on the SSEB, placing too much responsibility on the lesser-experienced remaining members. See Complaint ¶ 55. The inexperience of evaluators was mentioned only in passing at oral argument, Tr. (Mar. 14, 2006) at 115, and wisely seems to have been dropped as a separate ground for protest in FCSS’s briefing on the matter&emdash;as the judgment employed by Executive branch officers in selecting such evaluators is not normally a proper matter for court review.
. See AR Tab 210 (SSEB determines that neither the Solicitation nor the SSP "requires a score for the Overall Quality Factor”). The Solicitation stated that the adjectival and proposal
. The neutral rating of "Unknown” is also used, when no performance record had been identified for an element. AR Tab 7 at 53.
. It also argues that there was no basis for worsening its risk rating from “Low” to "Moderate" for the two sub-elements, FCSS Mot. at 27, which will be discussed below.
. Fort Carson Support Services also included in this aspect of its protest a variation on its argument concerning the evaluation of staffing levels, which is considered below. See FCSS Mot. at 33-34.
. Nor is there any explanation of how the IGE comparison that calculated the FCSS proposal to be understaffed by ninety-eight positions and overstaffed by seventeen was converted into the MPC decision that FCSS was understaffed by forty-five positions and overstaffed by six. Compare AR Tab 67 at 2 with AR Tab 76 at 34.
. On another slide, the SSEB stated that ”[t]he staffing proposed cannot sufficiently provide a full-range of coverage for the entire functional areas of the contract.” AR Tab 76 at 32.
. During oral argument, counsel for the government confirmed that the decision to credit FCSS with thirty-nine FTEs due to innovations was “not documented, and the cost analysis wasn’t provided.” Tr. (Mar. 14, 2006) at 208. See also App. to FCSS Mot., Ex. F at 4 (Aug. 31, 2005 letter from Army counsel stating contracting office was "not aware of any documents related to any innovations by FCSS that are not already contained in the record”).
. There was one slight discrepancy, to KIRA's disadvantage, as it appears that the size of the innovations credit calculated by the evaluator was actually 6.6 FTEs, not the 6.3 FTEs figure used in the decision. Compare AR Tabs 197, 206 with AR Tabs 209, 211.
. At least one member of the SSEB seemed to believe that at least some of the perceived under-staffing in FCSS's proposal would be overcome by labor-saving innovations. See AR Tab 100 (Nov. 4, 2004 email from SSEB member stating that FCSS ‘‘needs 16 people less than what [the Army] estimated base [sic] on their way of doing business [using PDAs], In the government’s estimate of personnel we assumed ‘business as usual.’ ’’).
. Fort Carson Support Services attempts to distinguish Dismas on the ground that the reevaluation in that case was done pre-award. FCSS Reply at 19. The Court fails to see the relevance of this distinction. Moreover, the Army discovered that FCSS was not eligible for the initial award because of a deficiency in the FCSS proposal, see AR Tab 95 at 2, an action that was not challenged by FCSS. Thus, the initial award does not merit any special dignity in this review.
. Fort Carson Support Services also objects to the Army's evaluation of its WR & DD plan. A slide from the SSEB’s presentation to the SSA identified this as a weakness under the Management and Technical element, and included the words "extra cost” in the description. AR Tab 211 at 63. During the post-award debriefing, FCSS asked the Contracting Officer whether cost considerations were included in the Management subfactor evaluation, and he explained that no dollar figure was considered, as the note reflected “just somebody’s comment. You know, by them staffing more than what they need to, it's costing the Contract more money.” Orange (FCSS) Post-Award Tr. at 32. When asked whether this entailed an overstaffing determination, the Contracting Officer explained “that would have been reflected in the Most Probable Cost adjustment.” Id. at 33. The protester misinterprets this remark into a claim that the MPC was adjusted upwards for the extra cost of this approach, resulting in a double-counting of the costs, see FCSS Mot. at 52-53. But it is clear that the official was merely speculating that an adjustment, if any, would have been made in the MPC. Since any adjustment for overstaffing would reduce the number of FTEs added for net understaffing, see, e.g., AR Tab 208, and thus reduce the MPC cost adjustment, FCSS’s argument&emdash;based on words ripped from their context&emdash;is particularly inapt.
. Another argument, concerning the FAR requirement that certain adverse past performance information be discussed, see 48 C.F.R. § 15.306(d)(3) (2005), is addressed separately below.
. It appears that the color "Green" corresponds to "Low” risk and the color "Yellow” to "Moderate" risk. See AR Tab 209 at 4 (explaining that "in the original evaluation, the evaluator’s performance risk assessment for Offeror [FCSS] was Moderate”).
. The SSEB member who individually assessed the PPE information for the initial evaluation erred in compiling the summary of PPE questionnaire responses for FCSS, identifying ten “Satisfactory” evaluations for Plum Island as "Marginal.” Compare AR Tab 60 (Plum Island questionnaire nos. 7-11, 13, 14, 26-28) with AR Tab 62. The evaluator classified two blended "Satisfactoiy” and "Marginal” ratings as "Marginal,” compare AR Tab 60 (Plum Island questionnaire nos. 6, 12) with AR Tab 62, but mistakenly described one of these as a score of "marg[inal]-unsat[isfactory].” AR Tab 62 (Section III — Past Performance Evaluation at 4). These errors by one individual do not appear to have influenced the SSEB as a whole, as its initial consensus decision was that FCSS merited a "Low” performance risk rating for the PPE and its elements, see AR Tab 69 at 34, and FCSS itself does not challenge this aspect of the evaluation — instead taking the position that “the record does not identify any alleged errors in the SSEB’s initial evaluation of FCSS’s PPE.” FCSS Reply at 25.
. The Court rejects the argument of FCSS that the use of a “Low/Moderate” risk rating for overall PPE violated the stated evaluation criteria, see FCSS Mot. at 53; FCSS Reply at 21, for the same reasons this argument was rejected in the context of FCSS's claim that it received the highest Quality factor rating. These labels were not the actual assessment, but tools used in the assessment, and the SSA clearly based his decision on the underlying PPE information and analysis, and not the subfactor label. See AR Tab 214 at 4.
. The understaffing adjustment actually exceeds the variance. In the first option year, this adjustment totaled $3,043,187, while the variance was just $3,002,833, due to the net effect of the other adjustments. See AR Tab 208 at 4.
. Fort Carson Support Services also argues that its Cost Variance rating of "marginally satisfactory” is inconsistent with its adjectival ratings of “Good" in the Management and Technical element and in the overall Management subfactor. FCSS Mot. at 64. But FCSS received the lower adjectival rating of "Satisfactory” for the element, not "Good.” See AR Tab 211 at 62. Documents in the record'demonstrate that the SSEB downgraded this adjectival rating, and the proposal risk rating of both the element and the overall subfactor, due to the understafiSng concerns that caused FCSS’s cost variance. See AR Tabs 189, 196.
. The Court does not find that it was reasonable for FCSS to believe that the initial award to it communicated the lack of serious staffing problems. The Army subsequently stayed this award under a corrective action, see AR Tab 97, because of undisclosed “questions regarding the evaluation of proposals and the subsequent award.” Id. If FCSS believed that its staffing level was not problematic, based on its own speculation concerning the reasons behind the corrective action, it is itself responsible for this belief. Such an assumption was a serious mistake by FCSS, for the record amply demonstrates that one of the "questions” that prompted the corrective action related to the SSA having made his tradeoff decision based on erroneous calculations of the FCSS variance — which understated the impact of the understaffing by approximately sixty percent. See AR Tab 95 at 2 (evaluation was based on variances of $10.2 million and six percent, as opposed to $ 26.3 million and fifteen percent); ARTab 76 at 35.
. And FCSS admits that the perceived strengths of its initial proposal were not revealed to it by the Army during the procurement process, see FCSS Mot. at 75 n. 69, so it cannot argue that the Army somehow induced it not to consider improving this aspect of its proposal.
. The SSEB evaluator appears to have counted both of these as "Marginal.” See AR Tab 62 at 2-3 (questions six and twelve).
. A copy of the letter and attachment appear to have been sent to the Army via facsimile machine on January 27, 2005. See ARTab 179.
. The audit report is included, with the KIRA letter, in the record at Tab 179. KIRA's counsel misrepresented this fact to the Court, and asserted that "[t]he audit report in question is not part of the Record.” KIRA Br. at 34.
. KIRA asserts, without citation, that the “report, however, shows that the subcontractor had an opportunity to respond." KIRA Br. at 34. This assertion appears to be false. See AR Tab 179 (Audit Report).
. This conclusion is reinforced by other documentation in the record. On November 19, 2004, after making an inquiry to another Army official, the Contracting Attorney determined that information relating to KBR’s Kuwait and Iraq operations would not be considered in the procurement process, as there was “no basis for downgrading KBR’s PPE rating based on the current investigations.” AR Tab 114A; see also ARTab 205 at 2.
. What TGG calls "other, more minor, examples of arbitrary and capricious behavior by the Army," TGG Mot. at 12 n. 2, contained in its complaint, were not pressed in the motion. See TGG Compl. ¶¶ 25-28 (lack of discussions concerning key personnel), 30-38 (Management and Technical element evaluation); 80-84 (cost realism).
. An AQL, or "acceptable quality level," is "the highest number of defects per 100, highest percentage of defects, or highest number of defects allowed for any service performance indicator.” AR Vol. XIV, Tab 10 at 8-4.
. Due to technical problems, the first portion of the discussion session with TGG, when this part of the statement was read, was not recorded. See AR Tab 132 at 5. The minutes from the discussions state that this written statement was read to TGG. See id. at 1. Since the same statement was read during the KIRA discussions, see Brown (KIRA) Disc. Tr. (Dec. 16, 2004) at 5, and the FCSS discussions, see Orange (FCSS) Disc. Tr. (Dec. 15, 2004) at 5, the Court has no reason to doubt that it was read at the beginning of TGG discussions. See also Blue (TGG) Disc. Tr. (Dec. 15, 2004) at 18 (Contracting Officer tells TGG it "should not construe anything in our earlier