81 Fed. Cl. 463 | Fed. Cl. | 2008
OPINION
This consolidated post-award bid protest ease is before the court on the parties’ cross-motions for judgment on the administrative record. It involves a government-wide acquisition contract (GWAC)
Eight unsuccessful offerors—Serco, Inc. (Serco); CGI Federal Inc.(CGI); STG, Inc. (STG); Artel, Inc. (Artel); Advanced Technology Systems Inc. (ATS); Apptis, Inc. (Apptis); Nortel Government Solutions, Inc. (Nortel); and The Centech Group (Cen-tech)—vigorously assert that the process used to make the awards under this GWAC was arbitrary, capricious, and contrary to law. They contend that they were prejudiced by an array of errors—some broad and systemic, others heterogeneous. Not so, responds defendant and five of the awardees (here as defendant-intervenors), arguing that GSA acted reasonably, and well within its lawful discretion, in making distinctions that were admittedly fine, but, nonetheless, compelled by the breadth and complexity of this procurement.
In the main, plaintiffs are correct. For the reasons that follow, the court concludes that GSA, in attaching talismanic significance to technical calculations that suffer from false precision, made distinctions that, in their own right, likely were arbitrary, capricious and contrary to law, but certainly became so when the agency failed adequately to account for price and to make appropriate tradeoff decisions. Those compounding errors prejudiced the plaintiffs and oblige this court to set aside the awards in question and order appropriate injunctive relief.
I. BACKGROUND
The administrative record in this case reveals the following:
A. The Solicitation
GSA issued the Solicitation on September 29, 2006. The Solicitation advised that GSA “contemplate[d making] approximately 25 to 30 awards ... but reserves the right to place fewer or more awards, depending upon the quality of the proposals received.”
The Solicitation indicated that the procedure for evaluating proposals would begin with an “acceptability review,” to be graded on “pass/fail” basis, focusing on whether a given offer contained all requested information in the appropriate formats. Acceptable proposals were then to be subject to a technical evaluation, a cost/price evaluation and, ultimately, a best value tradeoff. Regarding this last step, the Solicitation advised: “Consistent with FAR 15.101-1, the Government will conduct a ‘best value’ tradeoff, in which differences in non-price factors and evaluated price will be compared between the Offerors in order to determine which Offeror represents the best value to the Government.”
1. Proposal Requirements
The Solicitation indicated that two technical evaluation factors—Past Performance and the Basic Contract Plan—“are approximately equal in importance to each other, and when combined, are significantly more important than Cost or Price.” The Solicitation encouraged offerors to include their “best terms from a technical and cost/price standpoint,” as GSA “intend[ed] to evaluate proposals and award contracts without discussions with Offerors, except for clarifications, as described in FAR 15.306(a).”
a. Past Performance Proposal Requirements
The Solicitation indicated that in evaluating past performance, the agency would “look ‘retrospectively’ and consider the Offer- or’s history of success in delivering high quality service and solutions on contract efforts of similar scope and complexity to those anticipated under the Affiant Contract.” To facilitate this, the Solicitation required offer-ors to present information about contracts comparable to Affiant in two tables. Table 1 was to include contracts the offeror had performed in a “multiple-award environment,” such as GWACs and other forms of multiple award contracts, while Table 2 was to highlight the offeror’s efforts under unique, single-award contracts. Offerors could list up to twenty and fifty efforts in Tables 1 and 2, respectively, and could select up to three of the customers identified in Table 2 for the Government to contact as references. The Solicitation also permitted offerors to submit a narrative summary, not to exceed five pages, of the past performance efforts listed in the tables, including “whatever information [offerors] believe[d] may be relevant to the Government as it evaluates [the] proposal[s.]”
The Solicitation advised offerors that GSA “intends to use reasonable efforts to check approximately ten (10) efforts (total) for each Offeror, selected from Tables 1 and 2, including the three (3) efforts which the Offeror has identified.” However, the agency reserved the “right to check more or fewer efforts, at its discretion,” adding that it could not “guarantee that it will contact any particular effort listed in Table 1 or Table 2, even the efforts specifically listed by the Offer- or.”
b. Basic Contract Plan Proposal Requirements
In contrast to past performance, the purpose of the Affiant “Basic Contract Plan” (BCP) was to allow the agency to “look ‘prospectively’ and consider the Offeror’s level of commitment to the Affiant program and potential for high quality service and solutions.” The Solicitation indicated that in this plan, the offeror should explain “how it will continuously identify, mitigate, manage and control risks within its holistic approach for managing the comprehensive scope of the Affiant program.” Offerors were asked, inter alia, to “identify any gaps or weaknesses in past performance and specifically address them,” and to “convey [the] ability to insure successful performance of all aspects of the Basic Contract and Orders to include all 3 component areas.” The latter were defined as “Infrastructure, Application Services and IT Management Services.” The offerors were further instructed to “demonstrate a clear understanding of the management and performance requirements” of the Solicitation by providing a description of its management approach to three subfactors: resources, pro
The Solicitation also generally required the submission of a subcontract plan; various certifications; evidence of a top secret facility clearance; information concerning any teaming arrangements; and documentation of the offeror’s achievement of small business subcontracting goals.
c. Cost/Price Proposal Requirements
Offerors were required to propose hourly rates for eighty labor categories, by contract year and location of performance, for the potential ten-year term. The labor categories described in the Solicitation ranged from non-technical jobs, such as administrative/clerieal assistants and help-desk specialists, to highly technical jobs such a application systems analyst, computer scientist, data architect, disaster recovery specialist, hardware and software engineers and web designers. These categories were further subdivided in terms of knowledge and skill levels (e.g., entry-level, journeyman, senior, master). The Solicitation instructed offerors how to enter this rate information into a spreadsheet supplied by GSA. Those instructions explained, inter alia, that GSA would use the data in the spreadsheet to calculate loaded hourly rates for each applicable labor category, which, in turn, would be applied, for purposes of price analysis, to the independent Government estimates of hours for each category (which were already in the spreadsheet but protected from access by offerors). By this method, the agency intended to generate a total evaluated price for each bidder.
2. Evaluation Criteria
The evaluation criteria to be applied to proposals were set forth section M of the Solicitation. Those criteria can be described briefly as follows:
a. Past Performance
The Solicitation identified five subfactors within Past Performance, each of which were “approximately equal in importance to each other:” (i) quality of service; (ii) schedule; (iii) cost control; (iv) business relations; and (b) subcontract management/socioeconomic goals.
Under the quality of service subfactor, GSA was to evaluate the offeror’s ability to provide high-quality technological services. The agency was to consider whether, in past contracts, performance risk factors had been identified, mitigated, and managed. Offerors that employed “innovative and unique quality assurance tools and methodologies to ensure efficient and effective design, development and implementation of quality solutions” were to be more highly rated. For the schedule subfactor, the agency evaluated an offeror’s “ability to meet all schedule goals related to completion of the contract, task orders, milestones, delivery schedules and administrative requirements.” With regard to cost control, the agency evaluated the offeror’s “ability to deliver a service at the agreed to price/cost to include their ability to effectively forecast, manage and control contract costs, as well as report and analyze variances.” Under the business relations subfactor, the offeror needed to demonstrate its ability to “integrate and coordinate all activity needed to execute the contract” and to be an “effective business partner,” including such things as the timeliness, completeness, and quality of problem identification; corrective action plans; quality of proposals, change orders, and task order requests; and the contractor’s history of cooperative behavior and customer satisfaction. Finally, the Solicitation indicated that the agency would evaluate the offeror’s “ability to select, trace and manage subcontractors” and whether they had “met their small business utilization goals in the past.”
The BCP technical factor was broken into three subfactors: (i) resources—including internal performance capabilities and the ability to supplement core internal capabilities; (ii) program management, which focused, inter alia, on the offeror’s program management information system and quality control program; and (iii) corporate commitment, in which the offerors were to describe “how the Alliant program will be optimized through its business development, technological innovations, proposal management and contract administration efforts.” According to the Solicitation, these subfactors, as listed above, were “in descending order of importance when determining the overall rating for specific Technical Evaluation Factors.” Although offerors were not told exactly how these subfactors would be weighted for evaluation purposes, the SSP employed the following weighting: 50 percent for resources, 30 percent for program management, and 20 percent for corporate commitment. Additionally, as discussed below, for its evaluations, GSA further subdivided the three sub-factors into differing numbers of “elements” of equal weight, tracking the language of the Solicitation. As revealed by the Solicitation, the plans were also to be evaluated to determine whether they “sufficiently addresse[d] any gaps or weaknesses not addressed in past performance.”
c. Discriminators
The Solicitation described various “discriminators” that would be applied in the event that offerors were “deemed virtually equivalent” in terms of past performance. Thus, the Solicitation indicated that if there was such an equivalency with respect to past performance, offerors would be “evaluated more favorably” if they “demonstrated success in providing integrated IT solutions” involving infrastructure, application services, and management services or if they had “successful performance on efforts that include OCONUS work [work in the continental United States], cost-type contracts, balanced with fixed price contracts, or include significant subcontract management.”
By comparison, the Solicitation did not, in its discussion of the BCP evaluation, list specific “discriminators” to be used when of-ferors were deemed virtually equivalent. Instead, this part of the Solicitation cited qualifications that would lead to proposals being rated “more favorably” or “even more favorably.” For example, in evaluating internal resources, the Solicitation indicated that “[o]fferors that propose Program Managers with proven track records of managing programs similar to the Alliant in scope and magnitude will be evaluated more favorably.” Under this subfactor, it further stated—
An Offeror that can demonstrate that its purchasing policies and practices are efficient and provide adequate protection of the Government’s interests by providing evidence from the cognizant Federal agency ACO [ (administrative contracting officer) ] of successful progress toward an approved purchasing system will be rated more favorably. An Offeror that provides documentation from the cognizant Federal agency ACO as evidence of an approved purchasing system pursuant to FAR 44.305 will be evaluated even more favorably.
Following this pattern, under the program management subfactor, the Solicitation indicated that “[o]fferors providing comprehensive and thorough descriptions that demonstrate a program management information system and is capable of producing timely quality data products will be evaluated more favorably.” Likewise, offerors providing evidence from “a cognizant Federal agency ACO of successful progress toward an approved EVMS [Earned Value Management System] on an ongoing project will be evaluated more favorably,” while those offerors with “an EVMS that has been approved by a cognizant Federal agency will be evaluated even more favorably.” Certain features involving the offerors corporate commitment were likewise cited as being evaluated more favorably.
d. Cost And Price
The Solicitation confirmed that there would be a price analysis and might be a cost analysis. Price analysis was defined as “the
The Solicitation indicated that the agency “anticipates that pricing for this acquisition ■will be based on adequate price competition and therefore does not require submission or certification of cost or pricing data.” Nonetheless, the Solicitation warned that offerors should submit supporting documentation for direct labor, labor escalation, and each indirect cost consistent with their organization’s cost accounting system and be prepared to allow the agency to examine internal books and records so as to “permit an adequate evaluation of the proposed price in accordance with FAR 15.403-3.” It further indicated that “[cjost/price proposals will be evaluated using proposal analysis techniques consistent with FAR 15.404-1 to ensure that proposed direct and indirect rates for each labor category during the base period and the option period are fair, reasonable, and predictable for anticipated work under the Basic Contract.” In this regard, it gave examples of four techniques that might be employed to “ensure a fair and reasonable price,” to wit: (i) “Adequate Price Competition;” (ii) “[cjomparison of proposed Loaded Hourly Labor Rates received in response to the Solicitation;” (in) “[comparison of previously proposed Loaded Hourly Labor Rates for the same or similar labor categories using independent Government Price estimates;” and (iv) “[ojverall evaluated price using independent Government estimates.”
In terms of cost analysis, the Solicitation indicated that the agency might employ a variety of techniques “to ensure a fair and reasonable price,” among which were “[v]eri-fication that the Offeror’s cost submissions are in accordance with the contract cost principles and procedures in FAR Part 31,” and “[verification from an Offeror’s cognizant audit agency that the Offeror’s cost controls and surveillance systems are adequate.”
B. The Evaluation Process
GSA received sixty-six proposals on or before the deadline for submissions, November 17, 2006. After the “acceptability review,” four offerors were eliminated. GSA took the following steps in evaluating the remaining sixty-two proposals, including those of each of the plaintiffs in this matter.
1. The Basic Process
Pursuant to the SSP, the source selection evaluation team (SSET) consisted of the source selection authority, Steven J. Kempf; an SSET chair, who was also the procuring contracting officer, Mary Catherine Beasley; team leaders; team members; and advisors. A cosVprice evaluation team (CPET) and a technical evaluation team (TET) also were formed. The TET, in turn, consisted of a Past Performance evaluation team (PPET) and a Basie Contract Plan evaluation team (BCPET). The PPET consisted of three evaluators; the BCPET had four; and the CPET had three.
The adjectival rating system used by the Alliant TET is set forth in the SSP. As defined by the plan, those rating, as they applied to the BCPs, were as follows: (i) Exceptionally High Confidence [EH]— awarded if the BCP meets “all contractual requirements and exceeds many requirements to the Government’s benefit” and “essentially no reasonable doubt exists that the Offeror will successfully perform the required effort;” (ii) Significant Confidence [S]—awarded if the BCP meets “all contractual requirements and exceeds some requirements to the Government’s benefit” and essentially “[l]ittle doubt exists that the Of-feror will successfully perform the required effort;” (iii) Acceptable Confidence [A]— awarded if the BCP meets “all contractual requirements” and “[s]ome doubt exists that the Offeror will successfully perform the required effort;” (iv) Marginal Confidence [M]—awarded if the BCP “marginally meets contractual requirements (minimal information)” and “[s]ubstantial doubt exists that the Offeror will successfully perform the
The same set of adjectival ratings, albeit with sightly different individual descriptions, were used to evaluate past performance. For the past performance factor only, the Soicitation also isted a “neutral” evaluation, reserved for where “[performance in this area is not appicable to effort assessed, indicating neither a favorable nor unfavorable past performance evaluation.”
2. Past Performance Evaluation and Scoring
As anticipated in the Soicitation, to facilitate the past performance evaluations, GSA retained a poling firm, Calyptus, to conduct telephone interviews of references identified in offerors’ proposals.
Contractor personnel will interview the point-of-contact referenced in the tables, reading the script ... with questions____ the contractor personnel wUl then transcribe the responses to the questions, and transmit a copy of the transcript to the point-of-contact for confirmation that the [transcript] was an accurate record of the conversation. As this task is not subjective, the PCO determined that it was acceptable to have contractor support personnel do the task.
The scripts required the Calyptus personnel to ask open-ended questions, such as “[h]ow effectively did the Offeror conform to the contract requirements?” or “[t]o what extent did the Offeror successfully manage risk under the contract?” Although the Calyptus personnel were encouraged to ask follow-up questions, they were provided minimal guidance as to what to ask. Nor were they provided with any guidance as to the type of information the questions were intended to elicit or the significance that particularly-worded answers might have in assigning adjectival ratings to particular subfactors.
The three members of the PPET reviewed the Calyptus transcripts and, using the information contained therein, assigned adjectival ratings to five subfactors for each proposal, recording those scores on rating sheets, together with their notes.
3. BCP Evaluation And Scoring
To facilitate its work, the four-member BCPET subdivided the three subfactors to be employed in evaluating the plans—resources, program management and corporate commitment—into various “elements.” Each element had equal weight within its subfaetor. The team used a total of 15 elements, apportioned among the three subfactors as follows:
Resources subfactor:
Recruiting/retention and secrel/top secret clearances
Resumes of key personnel
Business systems
Purchasing system
Subcontracting management and supplemental core capabilities
OCONUS capabilities and continuity of operations
Program management subfactor:
Management information system (MIS) Earned value management system (EVMS)
Quality controls, surveillance methodologies
Corporate commitment subfactor:
Corporate structure, organization chart, and lines of authority
Proposal management
Risk mitigation
Business process improvements
Technical innovation
Business development
Each of the four members of the BCPET assigned an adjectival rating to each of the elements, again recording these scores on a rating sheet, together with notes. Unlike past performance, no consensus rating was made on the offeror’s BCP. Those individual ratings for each element—a total of sixty in all (four raters times fifteen elements)—were then converted into numerical scores using the same scale described above (e.g., 5 for EH, etc.). These numbers were then averaged to produce an overall score for each element (e.g., (5 + 4 + 4 + 4)/4 = 4.25). The overall numeric scores for each element were then averaged within each subfactor to produce an overall score for each of the three subfactors, that is, one score for resources, program management and corporate commitment, respectively. The three subfactor scores were then averaged in a weighted calculation—50 percent for resources, 30 percent for program management and 20 percent for corporate commitment—to produce a composite figure enumerated as the “BCP Average Evaluation Score.”
The Past Performance (PP) Evaluation Consensus Score and the BCP Average Evaluation Score were then averaged (added together and divided by two) to produce a
4. Cost And Price Analysis
As described above and in the Solicitation, GSA used a spreadsheet to develop a total estimated price for each offeror, based upon estimated demand over a ten-year term. The agency developed its Alliant labor estimate from various sources of information, including, inter alia, usage rates experienced under existing GSA IT services contracts. The three-member CPET inquired into whether an offeror’s cost structure appeared reasonable, assuming reasonable efficiency by examining offerors’ previously audited rates and verifying their cost submissions.
The evaluation team presumed that because of the competitive nature of the procurement and the number of bidders, the prices received were fair and reasonable.
C. Trade/Off Analyses and the Source Selection Decisions
On May 25, 2007, Ms. Beasley, as the SSET chair, and the leaders of each evaluation team briefed the source selection authority (SSA), Mr. Kempf, on their results to date. At the meeting, the SSET indicated that there was a “natural break” between the 27th and 28th highest technically-ranked of-ferors (derived using the Weighted BCP-PP Average Score). For this presentation, the SSET prepared a number of slides in which it highlighted the strengths and weaknesses of the offerors that were “above” and “below” the “natural break.” There is no indication in this slide presentation that in describing the “natural break,” the SSET focused at all on price. It recommended that the twenty-seven offerors above the “natural break” be considered for award and that the remaining offerors be considered unsuccessful.
Following that briefing, the SSET prepared a report setting forth the final scores and conclusions of the TET and CPET and sixty-two separate exhibits summarizing, inter alia, the trade-off analyses done for each responsive proposal and the SSET’s ultimate conclusion as to whether a particular offeror should receive an award.
The written tradeoff analyses generally followed one of three blueprints, each of which varied somewhat in terms of the level of documentation and analysis. All three forms contained certain baseline features: They compared the individual subfactor adjectival ratings of the evaluated offeror to the ratings of the three comparison offerors. Then, they compared the four offerors in terms of “Past Performance Discriminators” and the “BCP Discriminators,” the results of which were presented in charts and briefly summarized. Under the Past Performance Discriminators section, the agency described how one offeror had more or less of a particular type of work experience than another. The BCP discriminators, on the other hand, focused only on whether the offeror had an approved purchasing system, an approved EVMS, or whether the offeror was in the process of obtaining either one of those systems.
At this point, the written analyses diverged. For some offerors, including five plaintiffs (ATS, NGS, STG, Serco and CGO), the SSET merely added to the basic features described above a boilerplate “Conclusion” that read as follows:
The Source Selection Authority chose to not make an award to this Offeror. [X] was technically ranked [Y] out of 62 Offer-ors. The Government considered the technical merit of the [X] proposal and the proposed price. Though [X]’s proposed price was within the [Z] standard deviation of the mean overall Price of all the Offer-ors, the Government determined that [X]’s technical merit was not worth its lower price when compared with other Offerors of higher technical merit and higher prices.
(Emphasis in original). Notably, the last sentence of this boilerplate was the only mention of price in these trade-off discussions.
The tradeoff analyses for a second group of offerors, which includes three plaintiffs (Artel, Centech and Apptis), went a bit further. In addition to the analysis described above, these trade-off recommendations contained separate comparative discussions for the evaluated offeror vis-a-vis each of the three offerors selected for tradeoff. But, these additional comparisons did little more than parrot back the strengths and weaknesses of the proposals represented in the Solicitation’s evaluation criteria, thus highlighting the same characteristics that had led to the initial overall technical scores and related rankings. These comparisons again focused on price in only a conclusory fashion, typically stating either:
[X]’s proposal price was [Y] more than [Z]’s price. Considering the much greater technical merit of the [X] proposal when compared with the inferior [Zjproposal, as described above, the Government determined that the X proposal was a better value to the Government.
or
[X]’s proposal price was [Y] more than [Z]’s price. Considering this difference in price over a ten-year contract period and the greater technical merit of the [X] proposal when compared with the [Z] proposal, as described above, the Government determined that the [X] proposal was a better value for the Government.
These paragraphs were employed for offer-ors listed inside and outside the presumptive award group.
For the three highest-priced offers in the presumptive award group—those of Man-Tech, SAIC and CSC—the agency conducted a more thorough analysis.
Overall, although ManTech’s evaluated price exceeded two standard deviations on the high-end of Offerors, and was the second highest price of all Offerors, the Government has determined that their high rates were consistent with their technical ability and success in providing Integrated Command, Control, Communications and Intelligence (C4I) support to the Department of Defense, which requires proven experience with enterprise architectures, the latest eybersecurity automated tools, and a cadre of highly-cleared specialized technicians, engineers, and diffieult-to-find enterprise architecture personnel who are the key to C41 project success.
The analysis concluded that “[t]he Government determined that ManTech’s technical merit was worth its higher price when compared with other Offerors of less technical merit with lower prices.”
In a memorandum dated July 26, 2007, the head of the SSET summarized the process used to rank the offerors and provided the results of that process, both with an eye toward their being presented to the SSA. Attached to this memorandum was a chart, the critical portions of which are reproduced below:
*476 AJliant Past Basic Weighted Tech Total Price
Offeror Performance Contract BCP-PP Rank Evaluated Rank
(PP) Plan (BCP) Average Order Price Order
Evaluation Average Score
Consensus Evaluation
Score Score
L-3 Comms. 4.800 4.029 4.415 $41,322,310 52
.45 Harris 4.600 ,4.108 4.354 $38,936,210
¿24 311 >.'5.000 ■3,642 ' '41321 '■-'■ $31;SSS,607
35 3KA 4.600 3.958 4.279 $35,479.005
..-51 3oozAllen 4.200 4.267 4.233 $41,004.846
-’43 GHD “4.400 4.017 '4:208 $38.142,140
28 3AE 4.200 4.213 4.206 $31,805,838
,55 2SC .4.200 4.125 ■A. 163 $47,505,252
1 .41 ITS : 4.750 ■3.533 ‘4,142 ■ $36,520,763
•61 3AIC 4.000 4.213 4.106 10 $50,799,512
-'■14 INDUS Corn. 4.800 '3.400 4,100 11 ■,. .$28,317,285
"44 rASC ' 4.000 '4.183 ■4.'092 12 $38,585,071
61 VtanTech 4.600 3.533 4.067 13 $53,062,367
3SS ’4.400 '3.700 . 4:050 14 .$26,484,059
Lockheed Martin 3.800 4.275 4.038 15 $28,532,027 -15
46 3earinR?oint 4.200 3.804 4.002 16 $40,461,128
¿37 Raytheon. 4.200 ,; 3.796 3,998 17 ’-$34,895,936
■36 Accenture 4.200 ■3,788 31994 18 $34,452,849
, 2 4CI 4.400 3.571 3.985 19 $21,058,803
■ ‘.'33 UNISYS ,'4.200 ■.'3.729 3.965 20 ¿'.$33,255-,709
27 DB.C 4.000 3.925 3.963 21 $31,714,726
46 IBM 4.200 3.679 3.940 22 $39,368,972
■■'l'S «TC 4:400 3.463 3.931 23 -1$30,1631777
_2C :aci 4.000 ■3.854 3.927 24 l $30,537,485
12 ISIS ,4.400 3.433 3.917 25 $27,666,647
*477 ATT 4.000 3.833 3.917 25 $34,226,519 35
AMTI 4.200 3.575 3.888 27 $31,512,216 23
32 Alien 3.800 3.858 3.829 28 $33,007,426
_56 2ENTECH 4.500 3.125 3.813 29 r i
_55 TYBRIN 4.200 3.404 3.802 30 r i
_7 EDS 3.750 3.842 3.796 31 J_L
34 Stanley 4.000 3.538 3.769 32 $33,612,936
25 VlacB 4.400 3.108 3.754 33
57 ARINC 4.200 3.304 3.752 34 r i
47 !GI Federal 4.000 3.496 3.748 35 r i
38 Apptis 4.000 3.492 3.746 36 J_L
S 3wRI 3.S00 3.675 3.738 37 f 1
4.000 3.463 3.731 38 r i
_5< 2NSI 4.000 3.413 3,706 39 r i
jy ?s 4.000 3.404 3.702 40 J_L
_13 3MX 4.400 2.996 3.698 41 [ 1
_2i ATS 3.800 3.538 3.669 42 Í 1
31 ÍCeane 4.Q00 3.329 3.665 43 J_L
_4C AS! 4.200 3.083 3.642 44 f 1
5 3TG 4.000 3.254 3.627 45 r i
It MGS 4.000 3.192 3.596 46 r i
_4S Verizon 4.000 3.142 3.571 47 f 1
11 Alliant Solutions 4.000 3.100 3.550 48 f 1
21 American Sys. 3.400 3.650 3.525 49 J_L
1C T3 Alliance 3.750 3.242 3.496 50 [ 1
22 ARTEL 4.250 2.742 3.496 50 J_L
Lucent Techs. 4.200 2.763 3.481 52 J_L
VtcNeil Techs. 4.000 2.883 3.442 S3 r i _58
_62 nccc 4.200 2.646 3.423 54 J_L
3 ?ROSOFT 4.000 2.763 3.381 55
6 fRAWICK 4,000 2.679 3.340 56 r i
4 BPS 4.000 2.675 3.338 57 r i
17 Abacus 3.600 2.983 3.292 58 f 1
_2S Honeywell 2.800 3.779 33290 59 f 1
_n ?earson Alliant 3.000 3.425 3.213 60 r i
SYS 4.000 2.238 3.119 61 \ 1
42 JOMTek 2.800 2.196 2.498 62 L
Regarding this chart, the memorandum indicated that “[r]ow 30 [corresponding to the 28th-ranked offeror] represents the first natural break detected by the SSET that includes a minimum set of 25 contracts,” adding that “[t]he SSET recommends limiting awards to only those contractors highlighted in ‘green’ [through Alion].” [Green indicated by shading] The memorandum did not explain why the “natural break” had shifted a notch lower since the May 2007 meeting— from between the 27th and 28th ranked of-ferors to between the 28th and 29th ranked offerors.
Around this time, the SSET began finalizing its source selection recommendations for each offeror, which included the trade-off analyses described above. Some of these recommendations are dated July 24, 2007, one is dated July 28, 2007, and the remainder (a majority) are dated July 30, 2007. These memoranda recommended that the same top 28 technically-ranked offerors be extended awards. However, one additional award was contemplated. Thus, the memorandum for EDS recommended that it receive an award because of the breadth of its experience, the fact that it had an approved purchasing system, and the fact that its price was “significantly lower” than those offerors who had received a similar technical ranking.
In a SSD Memorandum dated July 31, 2007, Mr. Kempf concurred with the SSET’s recommendation that twenty-nine contracts be awarded to the top twenty-eight technically-ranked offerors, plus EDS. By way of a prelude, Mr. Kempf admitted that “[t]he Alli-ant tradeoff analysis posed challenges.” Summarizing the process that was used and
I found the evaluated proposals to be arrayed along a continuum reflective of their strengths and weaknesses, in technical quality and price. There is a fairly smooth continuum from best to least technically capable, and from lowest to highest in price. For me, the critical area was where the selection decision approaches the zone of 25-30. Within that zone, I saw a soft “break” point between the proposal ranked 28th overall in technical quality (that of Alion), and the next two proposals (Cen-tech and Tybrin), as these latter two companies’ proposals had progressively less technical quality coupled with significantly higher prices. As explained below, I also found reason to include one additional closely-ranked company (EDS) that offered distinct benefits to the Government. Inclusion of EDS brings the total number of awardees to 29.
With the tentative selection of this group of 29 awardees, my attention turned to the remaining proposals, starting with Centech, Tybrin, Stanley, and so on. As found in the individual company exhibits, prepared at my direction, my consideration took the form of a “trade-off’ analysis. The predicate for this trade-off analysis is that the Government seeks an advantageous mix of technical quality and attractive prices. Once I had provisionally identified the 29 best value proposals for Alliant, I wanted to answer this question: Did any other proposal offer compelling reasons to displace one or more of the apparent awardees[?] As explained in the individual Offeror exhibits, I found the answer to be “no”—and I believe that the group of 29 awardees do represent the best value and most advantageous selections.
Mr. Kempf then commented further on the evaluations of four offerors: Centech, Tyb-rin, EDS and Stanley. The breadth of these comments varied, with more attention paid to Centech and Stanley.
Based upon this decision, Mr. Kempf selected twenty-nine companies for award— again, the top twenty-eight ranked technical offers, plus EDS. By letters dated July 31, 2007, GSA informed the unsuccessful offerors that they had not been awarded a contracted. Following debriefings, beginning in late August of 2007, each of the plaintiffs (except Serco) filed protests with the Government Accountability Office (GAO).
D. Procedural History of this Case
On September 26, 2007, Serco, Inc. (Serco) filed a complaint in this court challenging the
II. DISCUSSION
Before turning to plaintiffs’ claims, we begin with common ground.
A. Standard of Review
The Federal Circuit, in Bannum, Inc. v. United States, 404 F.3d 1346, 1355 (Fed.Cir.2005), instructed that courts must “distinguish ... [a] judgment on the administrative record from a summary judgment requiring the absence of a genuine issue of material fact.” Toward this end, Bannum teaches that two principles commonly associated with summary judgment motions—that the existence of a genuine issue of material fact precludes a grant of summary judgment and that inferences must be weighed in favor of the non-moving party—do not apply in deciding a motion for a judgment on the administrative record. Id. at 1356-57. The
Bannum’s approach to deciding motions for judgment on the administrative record fits well with the limited nature of the review conducted in bid protests. In such cases, this court will enjoin defendant only where an agency’s actions were arbitrary, capricious, an abuse of discretion, or otherwise not in accordance with law. 5 U.S.C. § 706(2)(A) (2006); see also 28 U.S.C. § 1491(b)(4) (2006). By its very definition, this standard recognizes the possibility of a zone of acceptable results in a particular case and requires only that the final decision reached by an agency be the result of a process which “consider^] the relevant factors” and is “within the bounds of reasoned decisionmak-ing.” Baltimore Gas & Elec. Co. v. Natural Res. Def. Council, Inc., 462 U.S. 87, 105, 103 S.Ct. 2246, 76 L.Ed.2d 437 (1983); see Software Testing Solutions, Inc. v. United States, 58 Fed.Cl. 533, 538 (2003); Gulf Group, Inc. v. United States, 56 Fed.Cl. 391, 396 n. 7 (2003). As the focus of this standard is more on the reasonableness of the agency’s result than on its correctness, the court, to apply this standard properly, must exercise restraint in examining information that was not available to the agency. Failing to do so risks converting arbitrary and capricious review into a subtle form of de novo review.
The aggrieved bidder must demonstrate that the challenged agency decision is either irrational or involved a clear violation of applicable statutes and regulations. Banknote Corp., 365 F.3d at 1351, aff'g, 56 Fed.Cl. 377, 380 (2003); see also ARINC Eng’g Servs. v. United States, 77 Fed.Cl. 196, 201 (2007).
B. Alleged Errors in the Evaluation Approach
It is with this basic analytical framework in mind, that we now turn to the specific allegations of error here. Some of those allegations are systematic and cross-cutting, in that, if true, they impact all or virtually all the plaintiffs. Others sweep more narrowly, potentially impacting one or, at most, a few of plaintiffs. For a variety of reasons, the court will consider these claims chronologically, that is, in the order that they arose during the evaluation process. The court will focus first on claims that involve the integrity of the performance data initially generated by the agency. The court will then consider, in turn, alleged errors involving: (i) the reliability of the calculations that gave rise to the technical rankings that were used to identify the “presumptive awardees”; (ii) the agency’s consideration of price and price reasonableness (or failure to do so); and, ultimately, (iii) the propriety of the agency’s tradeoff analyses.
1. Gathering of Past Performance Information—Unequal Treatment
We begin by analyzing the survey process that GSA used to develop past performance information about the offerors. While, in reviewing this process, the court must “accord an agency’s technical evaluation great deference,” Hamilton Sundstrand Power Sys. Inc. v. United States, 75 Fed.Cl. 512, 516 (2007), it must, nonetheless, be sensitive to the agency’s fundamental duty to “[ejnsure that contractors receive impartial, fair and equitable treatment.” FAR § 1.602-2; see also Precision Images, LLC
v. United States, 79 Fed.Cl. 598, 619 (2007); HWA, Inc. v. United States, 78 Fed.Cl. 685, 697 (2007). “This obligation necessarily extends to the equal and impartial evaluation of all proposals,” this court has stated, “for it is well-established ... that a ‘contracting agency must treat all offerors equally, evaluating proposals evenhandedly against common requirements and evaluation criteria.’ ” Hamilton Sundstrand Power Sys., 75 Fed.Cl. at 516 (quoting Banknote Corp. of Am., 56 Fed.Cl. at 383).
They assert that the process by which Calyptus obtained information from past performance references was fatally flawed, leading different offerors to be treated unequally in their past performance evaluations. Their concerns focus on claims that the Calyptus personnel were given inadequate guidance as to the type of past performance information that they should elicit from the references. In this regard, plaintiffs asseverate, inter alia, that the scripts provided to the Calyp-tus employees included questions that were insufficiently tailored to the evaluation criteria and thus were inadequately designed to produce information relevant to those criteria.
As should be obvious, there is nothing inherently wrong with an agency using a survey, telephone interview, or questionnaire to elicit information regarding past performance. To the contrary, a review of agency regulations, see, e.g., 48 C.F.R. § 1815.304-70(d)(3) (1999), not to mention the decisional law, suggests that such surveying has been frequently employed without objection.
Here, the questions employed by Calyptus largely were not designed to produce answers responsive to the past performance evaluation standards. Those standards made critical distinctions based upon the degree to which the offeror exceeded or met contractual requirements; whether the offeror’s track record on a given factor was “outstanding,” “above-average,” “acceptable,” “marginal,” or “adverse;” and the degree of doubt the reference had as to the ability of the offeror to perform successfully under the Affiant contract {e.g., “no,” “little,” “some,” “substantial,” and “significant”). By comparison, the scripts were more generic, with ten of the
That arbitrariness thereby crept into the process is illustrated by the various anomalous ratings that were made. For example, on the cost control subfactor, CENTECH received uniformly positive comments, such as “5 (out of 5),” “excellent for cost control,” “managed costs exceptionally well,” “exceptional,” and “excellent control over cost,” and yet received an S rating. Likewise, STG’s references all positively described its quality of service as “very well,” “top quality,” “exceptional,” “phenomenal success” and “excellent”—yet it too received an S rating. Finally, on the schedule subfactor, CGI received comments of “very effective,” “very successful,” “excellent,” and “very well” and again received an S rating, even though two awar-dees (INDUS and ManTech) with very similar comments managed to receive an EH rating.
It is conceivable that these errors were the result of faulty individual ratings, rather than systematic problems with the survey questions themselves. But, one merely needs to compare the rating standards to the questions in the script to conclude that the latter was a major culprit here. It is also conceivable that unresponsive answers might have been provided even if the questions posed had been more precise. Perhaps tacitly
In a last ditch attempt to resurrect the surveys, defendant observes that we do not know what the evaluations would have yielded if more relevant questions had been asked and the transcripts had been supplied to the past performance references for correction. Perhaps, defendant speculates, plaintiffs would have fared no better or even worse. But, this dim view misses the point, for while the agency would have had considerable discretion in evaluating plaintiffs’ past performance had a proper process been used to develop performance information, it did not have the discretion to employ a survey system that did not generate veriflably accurate and relevant past performance information. Viewed in terms of prejudice, plaintiffs need not demonstrate what would have resulted had a rational evaluation process been employed. Indeed, for them to attempt to do so would require them to engage in substantial speculation—not only as to how GSA would draft a proper survey, but also as to how
2. Calculating the Technical Rankings—False Statistical Precision
In the next stage of the process, GSA took the initial observations that it had accumulated in terms of past performance (PP) and the basic contract plans (BCP) and developed a composite technical score. As the accompanying chart illustrates, the agency took slightly different paths in calculating the two numbers that were averaged to create that technical score, again enumerated as the “Weighted BCP/PP Average Score.” In calculating the past performance component of that score, GSA first developed a consensus adjectival rating for each subfaetor {e.g., S), then converted those subfaetors into whole numbers {e.g., 4) and averaged them into an average PP score {e.g., 4.50). Despite instructions to this effect in the SSP, GSA did not make use of consensus adjectival ratings in calculating the weighted BCP average score. Rather, it first assigned adjectival ratings to each of the elements that made up the three BCP subfactors {e.g., S). It then converted those ratings into whole numbers and developed an average number for each element {e.g., 4.25) and used those averages to calculate a value for each subfactor {e.g., 4.21), which were then weighted to calculate the weighted average BCP score {e.g., 3.125). The average PP score and the weighted average BCP score were then averaged, yet again, to produce the ‘Weighted BCP/PP Average Score,” expressed in numbers carried out to three decimals {e.g., 3.813).
But were these numbers actually significant to three decimals, such that GSA could make rational distinctions based upon the thousandths of a point—as it did? There are strong indications that the answer to this question is no.
Precision of thought is not always reflected in the number of digits found to the right of a decimal point—indeed, as with other constructs, there can be, to paraphrase Holmes, a “kind of precision that obscures.” Regarding statistical accuracy, a prominent treatise on numerical analysis explains that numerical computations may be impacted by “errors in the input data,” adding that “input data can be the result of measurements with a limited accuracy, or real numbers, which must be represented with a fixed number of digits.” Lars Elden & Linde Wittmeyer-Koch, Numerical Analysis: An Introduction 7 (1990). Numerical calculations may also be impacted by rounding errors—those that “arise when computations are performed using a fixed number of digits in the operands.” Id. Introducing accuracy and rounding errors into a calculation does not mean that the product thereof is meaningless; it does mean, though, that the error bound on that calculation must reflect the uncertainty of the inputs, thereby potentially reducing the number of digits in the output that are significant—e.g., those digits to the right of the decimal point. Id. at 12 (“When approximate values are used in computations then, of course, their errors will give rise to errors in the results.”); 16 McGraw-Hill Encyclopedia of Science & Technology 461 (9th ed. 2002) (“Computations cannot improve the precision of the measurement.”). Generally speaking, the result of a computation that involves multiple numbers should be reported at the same precision level as the least precise measure— to the digit precision of the “weakest link,” so to speak.
Here, of course, GSA made a host of fine-line distinctions using the average BCP/PP technical scores, determining, most prominently, which firms were among the “presumptive awardee” group based, ultimately, upon as few as 8 thousandths of a point. Yet, there are strong reasons to suspect that these comparisons were undermined by false precision, particularly since no error propagation analysis of the sort described above was performed. For one thing, there is no indication that the initial numerical data generated in evaluating the offerors’ proposals— data that, in the first instance, was captured in whole numbers (e.g., 3, 4, or 5)—had anywhere near the level of accuracy reflected in the final evaluation scores (e.g., 3.841). Indeed, in assigning adjectival scores to the various subfactors involved here, agency personnel effectively engaged in rounding (performing a computation with one digit in the operand)—they were, in other words, forced to choose one versus another adjectival rating even when a fraction (e.g., 3.8 rather than 4) would have been more accurate. Such imprecision and rounding is most evident in the evaluation of past performance, the accu
In short, there is little doubt in the court’s mind that GSA should have reported fewer significant digits in its findings to better reflect the uncertainty in its technical rankings. Had it done so, many of the observations that were predicated upon the detailed BCP/PP averages, including the notion that there was a “natural break” somewhere along the rankings, would have vanished, requiring GSA to make its award decisions on some other, albeit perhaps less convenient, basis. Indeed, one of the grandest ironies in this ease is the fact that the supposed “natural break point” here kept shifting—as identified by the SSET, it was between the 27th and 28th ranked offerors; as set by the SSA, it was between the 28th and 29th ranked offerors; and, ultimately, after Stanley was given an award, it settled between the 29th and 30th ranked offerors. This should have been a telltale sign that the break point here was not so “natural” and that more caution was warranted in mining information from these statistics. Defendant, for its part, has advanced no reason to convince the court that these statistics could be wielded in the fashion that they were—indeed, despite being offered a specific opportunity to do so, defendant has not provided a single case in which an agency was found properly to base important award decisions on technical rankings carried to the thousandths of a point. And with good reason. At the least, the obvious threat of false precision in the numbers employed here ought to have caused the agency to be more circumspect in relying on its multi-digit technical rankings, let alone in piling Pelion and Ossa upon those rankings by establishing the group of “presumptive awardees” and then, to top everything off, requiring a “compelling” reason to alter it. Other cases, indeed, suggest that agencies have been much more cautious in applying similar statistics, tending to view relatively close technical scores as being essentially equivalent.
In the end, this court concludes that it need not decide whether vel non the false precision here, standing alone, is fatal to the awards made.
3. Flaws in the Price and Price Reasonableness Analyses
After developing the weighted BCP/PP average scores, GSA next purported
Long ago, Congress rejected the notion of giving contracting officers the authority to ignore price considerations in negotiated procurements. See Schoenbrod v. United States, 187 Ct.Cl. 627, 410 F.2d 400, 402-03 (1969); see also Paul v. United States, 371 U.S. 245, 252-53, 83 S.Ct. 426, 9 L.Ed.2d 292 (1963). Any final question in this regard were put to rest by the Competition in Contracting Act of 1984, Pub.L. No. 98-369, 98 Stat. 1175, which unambiguously requires:
In prescribing the evaluation factors to be included in each solicitation for competitive proposals, [the agency] ... shall include cost or price to the Federal Government as an evaluation factor that must be considered in the evaluation of proposals.
41 U.S.C. § 253a(c)(l)(B). The legislative history of CICA indicates that this provision, as well as others designed to promote competition, were designed not only to allow the Federal government to obtain the “best products,” but to do so at the “best prices”— to avoid paying “$436 for an ordinary claw hammer ... where it can be bought for $7.” H.R. Rep. 98-1157, at 18 (1984); see also S. Rep. 98-50, at 32 (1983) (noting that “price” should have a “significant bearing on the selection for award”). “Congress intended for competition to affect the amount of money that the Government pays for goods and services,” two well-known commentators have stated, and “any competitive process that does not require films to compete on the basis of the amount of money that they want, and, in which differences in the amount of money sought cannot affect the outcome of the competition, is not consistent with that intention.” Vernon J. Edwards and Ralph C. Nash, “Price as a ‘Significant’ Evaluation Factor: Has the GAO Misinterpreted CICA?,” 20 No. 8 Nash & Cibinic Rep. ¶ 40 (hereinafter “Price as a Significant Evaluation Factor”).
Giving effect to the statute and its legislative history, the FAR ordains that “[p]rice or cost to the Government shall be evaluated in every source selection.” FAR § 15.304(c)(1); see also Northrwp Grumman Info. Tech., Inc., 2005 C.P.D. ¶ 45, 2005 WL 735939, at *9 (2005). Following this lead, GAO has repeatedly held that price must be a “significant evaluation factor;” that it must be given “meaningful consideration.” See, e.g., MIL Corp., 2005 C.P.D. ¶ 29, 2004 WL 3190217, at *7 (2004); Eurest Support Servs., 2003 C.P.D. ¶ 139, 2001 WL 34118414, at *6 (2001); RTF/TCI/EAI Joint Venture, 98-2 C.P.D. ¶ 162, 1998 WL 911892, at *8 (1998); H.J. Group Verdures, Inc., 92-1 C.P.D. ¶ 203, 1992 WL 48487, at *3 (1992); see also Price as a Significant Evaluation Factor, supra (surveying cases). It follows, a fortiori, that price can neither be a nominal evaluation factor nor relegated to the role of being a mere consideration in determining whether a proposal is eligible for award. See Lockheed Missiles & Space Co., Inc. v. Bentsen, 4 F.3d 955, 959-60 (Fed.Cir.1993); MIL Corp., 2004 WL 3190217, at *7; Electronic Design, Inc., 98-2 C.P.D. 69, 1998 WL 600991, at *5-6 (1998). These are not minor distinctions—an evaluation that fails to give price its due consideration is inconsistent with CICA and cannot serve as a reasonable basis for an award. See MIL Corp., 2004 WL 3190217, at *7; Boeing, Sikorsky Aircraft Support, 97-2 C.P.D. ¶ 91, 1997 WL 611539, at * 10 (1997).
GWACs are not immune from these requirements. To the contrary, the Federal Acquisition Streamlining Act of 1994 (FASA), Pub.L. No. 103-355, 108 Stat. 3243, which codified existing authority for agencies to enter into task and delivery order contracts, did not carve out an exception for such contracts on the premise that price competition would occur for each task order. Thus, FASA’s legislative history made clear that all the CICA competition requirements are to apply to multiple award contracts, stating:
In addition, the conference agreement would provide general authorization for the use of task and delivery order contracts to acquire goods and services other than advi*492 sory and assistance services. The conferees note that this provision is intended as a codification of existing authority to use such contractual vehicles. All otherwise applicable provisions of law would remain applicable to such acquisitions, except to the extent specifically provided in this section. For example, the requirements of [CICA], although they would be inapplicable to the issuance of individual orders under task and delivery order contracts, would continue to apply to the solicitation and award of the contracts themselves.
H.R. Conf. Rep. 103-712, at 18. Relying on these reports, the GAO has concluded that “there is no exception to the requirement set forth in CICA that cost or price to the government be considered in selecting proposals for award because the selected awar-dees will be provided the opportunity to compete for task orders under the awarded contracts.” MIL Corp., 2004 WL 3190217, at *7.
But this begs the question—did GSA adequately consider price in making the award decisions here? To be sure, GSA conducted a variety of statistical reviews that appeared to consider price—for example, it compared the offerors’ prices to the mean price and indicated that it would look “favorably” upon those that were within one standard deviation thereof. Likewise, it compared the prices to the IGCE, and indicated that it would look “favorably” upon those that were below that figure. Yet, ultimately, it made awards to several offerors that did not meet one of these benchmarks and to three offer-ors that did not meet either—CSC, SAIC and ManTech. The latter trio, in fact, offered prices that were ranked 59th, 60th and 61st out of the 62 offerors, respectively. Now, defendant asserts that although it indicated that it would look “favorably” on offers that were below its benchmarks, it did not say what it would do if those thresholds were exceeded. But, one cannot reconcile this bit of ipse dixit with the legal requirement that price be a significant evaluation factor here— for price to be a significant factor, there must be ramifications if a price substantially exceeds the norms. Indeed, here the variances were staggering—the prices of CSC, SAIC and ManTech were more than two standard deviations beyond the mean, suggesting that they were truly outliers. In raw terms, they exceeded the IGCE by about $10 million (and the mean price by nearly $17 million) and were more than double the price of the lowest-priced awardee. And these seemingly high-priced offerors were not alone in receiving awards, as offerors with the 51st and 52nd highest prices—prices that also well exceeded the mean price—managed to make the award list, as well.
Despite strong circumstantial evidence that price did not play a significant role here, defendant steadfastly maintains that price was given all the weight it was due under the law and the Solicitation. Of course, it immediately undercuts those assurances by repeatedly discounting, in its briefs, the need for considering price at this stage of the procurement. Thus, it claims that this is so because there will be “continued price and quality competition among the awardees throughout the life of the program” in the award of task orders—thereby coming within a hair of making an argument that, as mentioned above, has been decisively rejected by the GAO.
Defendant then compounded this mistake by failing to conduct a proper price reasonableness analysis. Before turning to this point, it is best to define the legal parameters of such an analysis.
Prior to 1996, 41 U.S.C. § 401(a) provided that purchases should be made at the “lowest reasonable cost.” That provision, however, was repealed in 1996 by the National Defense Authorization Act for Fiscal Year 1996, Pub.L. No. 104-106, § 4305, 110 Stat. 186, 665. See Ralph C. Nash & John Cibinic, “Price Reasonableness: A Much Misunderstood Term.” 17 No. 4 Nash & Cibinic Rep. ¶22 (2003). Under CICA, however, contracting officers are still required to use “data other than certified cost or pricing data to the extent necessary to determine the reasonableness of the price of [a] contract.” 41 U.S.C. § 254b(d)(l). This requirement is reflected in FAR § 15.402(a), which states that the contracting officer “must” purchase supplies and services at “fair and reasonable prices.” This leitmotif reappears in FAR § 15.404-l(a) which states that “(t]he objective of proposal analysis is to ensure that the final agreed-to price is fair and reasonable,” adding, in subparagraph (1) thereof, that “[t]he contracting officer is responsible for evaluating the reasonableness of the offered prices.” Paragraph (b)(2) of this same provision then states that “[t]he Government may use various price analysis techniques and procedures to ensure a fair and reasonable
Even a cursory reading of these provisions thoroughly rebuts the notion, heavily advanced by several defendant-intervenors, that neither the FAR nor the Solicitation required GSA to preclude any offerors from receiving an award on the basis of its price reasonableness analysis. Per contra. Ultimately, the focus of such an analysis is not on whether the majority—or even a vast majority—of the prices offered are reasonable, but on whether a given awardee’s price is fair and reasonable.
Nor does FAR § 15.404-l(b)(2)(i) indicate otherwise. Indeed, it conspicuously states that adequate price competition “normally” establishes price reasonableness, leaving the distinct impression that there are exceptions to this general rule. Providing a window into those limitations, FAR § 15.404(b)(2)(i) cross-references FAR § 15.403-l(c)(l), which, in turn, defines adequate price competition in the following terms—
A price is based on adequate price competition if—
(i) Two or more responsible offerors, competing independently, submit priced offers that satisfy the Government’s expressed requirements and if—
(A) Award will be made to the offeror whose proposal represents the best value (see 2.101) where price is a substantial factor in source selection; and
(B) There is no finding that the price of the otherwise successful offeror is unreasonable.
FAR § 15.403—l(c)(l)(i); see Erinys Iraq. Ltd. v. United States, 78 Fed.Cl. 518, 531 (2007). This provision again makes clear that, even where there is competition, the focus of the price reasonableness analysis
Of course, here, GSA found that every evaluated price was “fair and reasonable.” Indeed, every price analysis in the record concluded with the following language, haec verba—“In relation to the IGCE for overall price and Mean Overall Price among all of-ferors (Government and Contractor Site combined over 10 years), the Offeror’s Mean Overall Price is in line with adequate price competition and is, therefore, considered fair and reasonable.” What is striking about that sentence is that it is omnipresent. It is found in exhibits in which the agency noted that the offeror’s evaluated price was “below” the IGCE and “within one standard deviation” to the mean and likewise in exhibits in which it was noted that the offeror’s evaluated price was “above” the IGCE and “outside two standard deviations” to the mean. That is to say, GSA wrote that the evaluated prices for offerors were “in line with adequate price competition,” even when its own statistics demonstrated that they were not— even when a price was an outlier that significantly departed from the pattern established by the wide majority of the offers and exceeded every statistical barometer of reasonableness employed. Such a conclusion— awkwardly mismatched with the analysis from which it supposedly springs—cannot be squared with the FAR. It can be obtained only from a price reasonableness analysis that was so unreasonable as to be arbitrary, capricious and contrary to law—one that, truth be told, was no price reasonableness analysis at all. See Multimax, Inc., 2006 C.P.D. ¶ 165, 2006 WL 3300346, at *8 (2006) (agency’s price analysis was unreasonable where “[tjhere is no indication that the agency every reviewed the results of the formula to assure that the prices at the extreme end of the ranges reflected reasonable pricing; rather, the agency mechanistically applied the formula and accepted the results without further analysis.”); Crawford Labs., 97-2 C.P.D. ¶ 63, 1997 WL 532508, at *4 (1997) (protest sustained where agency provided no rational basis to support its reasonableness determination).
In sum, GSA has provided no rational basis for its determination of price reasonableness of contract awards at prices that, at least using the methods GSA employed, appear to be neither fair nor reasonable.
4. Flaws in the Tradeoff Analysis
Lastly, plaintiffs hotly contest GSA’s best-value, tradeoff decisions. To be sure, as noted at the outset, plaintiffs have a significant burden of showing error in that regard because a court must accord considerable deference to an agency’s best-value decision in trading off price with other factors. See Info. Tech. & Apps. Corp. v. United States, 51 Fed.Cl. 340, 343, 346 (2001), aff'd, 316 F.3d 1312 (Fed.Cir.2003); see also E.W. Bliss Co. v. United States, 77 F.3d at 449; Bean Stuyvesant, LLC v. United States, 48 Fed.Cl. 303, 320 (2000); Cubic Def. Sys., Inc. v. United States, 45 Fed.Cl. 450, 458 (1999). In the court’s view, they have met that burden.
The best value process is described in FAR § 15.101-1 as follows:
(a) A tradeoff process is appropriate when it may be in the best interest of the Government to consider award to other than the lowest priced offeror or other than the highest technically rated offeror.
(b) When using a tradeoff process, the following apply:
(1) All evaluation factors and significant subfaetors that will affect contract award and their relative importance shall be clearly stated in the solicitation; and (2) The solicitation shall state whether all evaluation factors other than cost or price, when combined, are significantly more important than, approximately equal to, or significantly less important than cost or price.
(c) This process permits tradeoffs among cost or price and non-cost factors and allows the Government to accept other than the lowest priced proposal. The perceived benefits of the higher priced proposal shall merit the additional cost, and the rationale for tradeoffs must be documented in the file in accordance with 15.406.
The documentation requirement is reiterated in FAR § 15.308, which states that “[t]he source selection decision shall be documented, and the documentation shall include the rationale for any business judgments and tradeoffs made or relied on by the SSA, including benefits associated with additional costs.” See also FAR § 15.305(a). Even where, as here, a solicitation provides that technical criteria are more important than price, an agency must select a lower-priced, lower technically scored proposal if it reasonably decides that the premium associated with selecting the higher-rated proposal is unwarranted. See Magney Grande Distribution, Inc., 2001 C.P.D. ¶ 56, 2001 WL 287038, at *3 (2001); Oshkosh Truck Corp., 93-2 C.P.D. ¶ 115, 1993 WL 335049, at *6 (1993).
In determining whether an agency has complied with these regulations, this court must determine whether the agency’s decisions are grounded in reason. TRW, Inc., 98 F.3d at 1327; Comprehensive Health Servs., Inc. v. United States, 70 Fed.Cl. at 721. Various concepts serve to add a skeletal framework to this inquiry.
First, the regulation requires the agency to make a business judgment as to whether the higher price of an offer is worth the technical benefits its acceptance will afford.
Of course, it is conceivable that the SSA, in his own mind, made such costybenefit comparisons, but merely failed to capture them on paper. But, that too would violate the FAR and its documentation requirements. To comply with the FAR, documentation must “include the rationale for any business judgments and tradeoffs made, including the
Defendant offers up several surrogates for measuring the reasonableness of the tradeoff analyses conducted by GSA. For example, it repeatedly emphasizes the sheer number of pages that the agency dedicated to that purpose—going so far as to include an appendix to its opening brief listing the specific pages that the agency spent on each offeror. But, while having very little space dedicated to tradeoff comparisons might be indicative of a problem, the converse is not true—the court cannot, like an indolent schoolmaster grading a term paper, assess the reasonableness of a tradeoff analysis by the weight of the paper involved. Indeed, an examination of the tradeoff exhibits here reveals that, but for a few select passages, they simply reiterate descriptions of the procurement and adjectival ranking and discriminator information found in the technical evaluation portion of the source selection documents. Again, defendant makes much of the sheer volume of this information. But, if all an agency need do, in order to select a higher-priced offer over a lower-priced one, is to recite that the former has a better technical ranking than the latter, then performing a tradeoff analysis is truly a waste of good paper. A plain reading of the FAR reveals that more is envisioned—the agency must document how the perceived benefits of the higher priced proposal merit paying the additional cost. The decisional law, as well as common sense, suggest that this entails more than what is represented by relisting the adjectival ratings that gave rise to the technical ranking and then flatly concluding, as often was done here, that the government has “determined that [X]’s technical merit was not worth its lower price when compared with other Offer-ors of higher technical merit and higher prices.” Indeed, such formulaic incantations particularly miss the mark where, as here, they are invoked whether the prices compared are close together or a gulf apart.
Finally, some discussion is warranted not about the content of the comparisons, but rather about which firms were compared. Plaintiffs have mounted a multi-pronged assault on the basic formula used by GSA in establishing most of tradeoff pairings. Recall that formula had two parts, to wit: If the evaluated offeror was within the presumptive award group, it was compared to the three highest technically ranked offerors not in the presumptive award group that had a lower price than the evaluated offeror. If the evaluated offeror was not within the presumptive award group, it was compared to the three lowest scoring offerors within the presumptive award group that had a higher price than the evaluated offeror. As can readily be deduced, under this formula, each offeror was initially compared to a minimum of three offerors; in fact, at least on paper, certain offerors with high prices were compared to many others. Plaintiffs, for their part, claim that this formula again placed undue emphasis on the technical rankings and failed adequately to take into account price. They also complain about particular comparisons between them and other offerors that were or were not made. Some go so far as to argue that the agency was compelled to compare all the offerors to each other.
To an extent, the arguments regarding the use of the formula have been overtaken by events. Thus, it makes little sense, at this point, to debate fine points regarding the merits of particular comparisons (e.g., IBM v. STG) when the court already has concluded that the past performance survey information
In sum, the court finds that GSA’s value determinations were also arbitrary, capricious and otherwise contrary to law.
C. Redux
So where does this leave us? In his book, Some Economic Factors in Modem Life, Sir Josiah Charles Stamp described the potential for over-reliance on statistics in what would become known as “Stamp’s Law of Statistics”:
The individual source of statistics may easily be the weakest link. Harold Cox tells a story of his life as a young man ... He quoted some statistics to a Judge, an Englishman, and a very good fellow. [The Judge] said, Cox, when you are bit older, you will not quote statistics with that assurance. The Government are very keen on amassing statistics—they collect them, add them, raise them to the nth power, take the cube root and prepare wonderful diagrams. But what you must never forget is that every one of those figures comes in the first instance from the village watchman, who just puts down what he damn pleases.
Josiah Stamp, Some Economic Factors in Modem Life 258-59 (1929). The situation here, of course, is not as bad. Indeed, the court harbors no doubts that the agency here made a good faith effort to distinguish between the sixty-two offerors which responded to the Solicitation. And no party here has argued otherwise. Yet, on a variety of planes, the agency’s effort came up well short, resulting in award decisions that were arbitrary, capricious and otherwise contrary to law.
Defendant intimates that the court should afford the agency more slack than usual, on account of the size of this procurement and the number of offerors to be evaluated. But, given the extraordinary breadth of discretion already afforded to agencies in government procurements, it is hard to fathom what form a still more relaxed rule of deference might take. Would such a rule permit the adoption of procedures that would allow the agency to
Lastly, a further word about prejudice. All the parties agree that the nature of the prejudice analysis is somewhat different in the context of GWAC than in the ordinary bid protest case. Fundamentally, the question here is not simply whether, but for the errors associated with their respective case, a protester had a substantial chance of receiving a contract award instead of one of the current awardees, see, e.g., Alfa Laval Separation, 175 F.3d at 1367. Rather, the court must also ask whether, but for those errors, the same protester had a substantial chance of receiving a contract award in addition to the other awardees. These twin inquiries take on added dimensions in a case in which the correction of a systemic error might not only improve the protester’s evaluation, but diminish that of a current awardee, or even ehminate that awardee from further consideration altogether (as might be true under a proper price reasonableness analysis). With all of these varied dimensions, and since it beyond peradventure here that the slightest shifting of a single adjectival rating could have significant impact not only on the ranking of a given protester, but also on who they might be compared with in a tradeoff analysis, the court is left with the firm conviction that the combined impact of the errors encountered here clearly prejudiced each of the protesters.
D. Injunctive Relief
Having concluded that the instant procurement was legally flawed and that each plaintiff, to varying degrees, was thereby prejudiced, the court must determine whether plaintiffs have made three additional showings to warrant injunctive relief, to wit, that: (i) they will suffer immediate and irreparable injury; (ii) the public interest would be better served by the relief requested; and (iii) the balance of hardships on all the parties favors them. Idea Intern., Inc. v. United States, 74 Fed.Cl. 129, 137 (2006); Bannum, 60 Fed.Cl. at 730; Seattle Sec. Servs., 45 Fed.Cl. at 571. No one factor is dispositive to the court’s inquiry as “the weakness of the showing regarding one factor may be overborne by the strength of the others.” FMC Corp. v. United States, 3 F.3d 424, 427 (Fed.Cir.1993); see also Seattle Sec. Services, 45 Fed.Cl. at 571. In the instant case, the existence of irreparable injury to plaintiffs, the balancing of harms in favor of the plaintiffs, and the public interest all lead this court to grant injunctive relief to plaintiffs.
1. Irreparable Injury
When assessing irreparable injury, “[t]he relevant inquiry in weighing this factor is whether plaintiff has an adequate remedy in the absence of an injunction.” Magellan Corp. v. United States, 27 Fed.Cl. 446, 447 (1993). Plaintiffs argue that they will suffer irreparable harm if an injunction is not granted, because the only other available relief—the potential for recovery of bid preparation costs—would not compensate them for
2. Balance of Hardships
Under this factor, “the court must consider whether the balance of hardships leans in the plaintiffs’ favor,” requiring “a consideration of the harm to the government and to the intervening defendant.” Reilly’s Wholesale Produce v. United States, 73 Fed.Cl. 705, 715 (2006); see also Heritage of America, LLC v. United States, 77 Fed.Cl. 66, 79 (2007); PGBA, 57 Fed.Cl. at 663. Defendant and defendant-intervenors intimate that enjoining the performance of the contracts would delay implementation of a contract designed to provide state-of-the-art technology to the entire Federal government. But, this court has observed that “‘only in an exceptional ease would [such delay] alone warrant a denial of injunctive relief, or the courts would never grant injunctive relief in bid protests.’ ” PGBA, 57 Fed.Cl. at 663 (quoting Ellsworth Assocs., Inc. v. United States, 45 Fed.Cl. 388, 399 (1999)); see also Reilly’s Wholesale, 73 Fed.Cl. at 716. This is not such an “exceptional case,” for a variety of reasons.
First, defendant has not indicated that setting aside the Alliant awards would work an immediate hardship on agencies looking to fulfill their technology requirements. In fact, the record reveals that the existing technology contracts will not expire until at least December of this year—and, without information indicating otherwise, it is reasonable to assume that those contracts, like most government contracts, may be extended for a period of time should circumstances warrant. See, e.g., PGBA, 57 Fed.Cl. at 663. Second, the delay and administrative burdens associated with setting aside the current awards and requiring the agency to take curative actions are, of course, problems of defendant’s own making. Those ill effects, moreover, may be ameliorated by a properly-framed injunction that, as defendant has repeatedly urged, is narrowly tailored to the problems encountered. See, e.g., Reilly’s Wholesale, 73 Fed.Cl. at 716. In these circumstances, the balance of hardships tilts in the plaintiffs’ favor.
3. Public Interest
Plaintiff also contends that the public interest will be served by granting the requested preliminary injunctive relief. “Clearly, the public interest in honest, open, and fair competition in the procurement process is compromised whenever an agency abuses its discretion in evaluating a contractor’s bid.” PGBA, 57 Fed.Cl. at 663; see also Rotech Healthcare, Inc. v. United States, 71 Fed.Cl. 393, 430 (2006); Cincom Sys., Inc. v. United States, 37 Fed.Cl. 266, 269 (1997); Magellan Corp., 27 Fed.Cl. at 448. In the present case, the public’s interest likewise lies in preserving the integrity of the competitive process, particularly in the context of a massive procurement that will impact the public potentially for a decade.
III. CONCLUSION
In closing, the court must note that the presentations made by the many counsel involved in this case were extraordinarily help-
Based on the foregoing:
1. Plaintiffs’ motions for judgment on the administrative record are GRANTED, in part, and DENIED, in part, and defendant’s cross motion for judgment on the administrative record is DENIED, in part, and GRANTED, in part.
2. Defendant, acting by and through the General Services Administration, is hereby ENJOINED from performing or allowing others to perform on the contract(s) awarded pursuant to Affiant Solicitation No. TQ2006MCB0001. Said parties also must suspend any related activities that may result in additional obligations being incurred by the United States under the contract(s).
3. Defendant, acting by and through the General Services Administration, is hereby ENJOINED from:
a. Relying, in making future award decisions pursuant to the Affiant Solicitation, on the results of the survey conducted by Calyptus, unless defendant, consistent with this opinion, confirms the accuracy of those results and supplements them with information that is responsive to the past performance evaluation criteria specified in the Solicitation;
b. Relying, in making future award decisions pursuant to the Affiant Solieitation, on any BCP/PP combined scores similar to those previously derived herein unless defendant, consistent with this opinion, confirms the accuracy of any such scores and, in particular, determines which digits of such scores are significant enough to be relied upon;
c. Failing, in making future award decisions pursuant to the Affiant Solicitation, to consider price and price reasonableness to the extent required by applicable statutes, the FAR and the Solicitation, as interpreted by this opinion; and
d. Faffing, in making future award decisions pursuant to the Affiant Solicitation, to make and fully document tradeoff decisions to the extent required by the FAR and the Solicitation, as interpreted by this opinion.
4. Nothing herein shall be deemed to prevent defendant and all or some of the plaintiffs from mutually agreeing to resolve this matter in such fashion as they deem appropriate.
5. This opinion shall be published as issued after 3:00 P.M. on March 5, 2008, unless the parties identify protected and/or privileged materials subject to redaction prior to said time and date. Said materials shall be identified with specificity, both in terms of the language to be redacted and the reasons for that redaction (including appropriate citations to authority).
IT IS SO ORDERED.
Appendix A—Technical Acronyms
A Acceptable Confidence
AGO Administrative Contracting Officer
BCP Basic Contract Plan
BCPET Basic Contract Plan Evaluation Team
CICA Competition in Contracting Act
COOP Continuity of Operations
CPET Cost/price Evaluation Team
EH Exceptionally High Confidence
EVMS Earned value management system
FAR Federal Acquisition Regulation
FASA Federal Acquisition Streamlining Act
GWAC Government-wide Acquisition Contract
IGCE Independent Government Cost Estimate
IT Information Technology
L/N Little to No Confidence
MA/IDIQ Multiple-award of indefinite delivery, indefinite quantity
MIS Management Information System
OCONUS Outside the continental United States
PCO Procuring Contracting Officer
PP Past Performance
PPET Past Performance Evaluation Team
PPIRS Past Performance Information Retrieval System
S Significant Confidence
SSA Source Selection Authority
SSD Source Selection Decision
SSDM Source Selection Decision Memorandum
SSET Source Selection Evaluation Team
SSP Source Selection Plan
TET Technical Evaluation Team
TOR Task Order Request
Appendix B—Trade Off List by Offeror
Exhibit C-l L-3 Communications Titan Corporation (L-3)—tradeoff with Stanley, MaeB, Apptis
Exhibit C-2 Harris Corp, Government Communications Systems Division (Harris)—tradeoff with Stanley, MacB, Apptis
Exhibit C-3 SI International, Inc. (SII)—tradeoff with SwRI, Smartronix, STG
Exhibit C—4 System Research and Application Corporation (SRA)-tradeoff with Stanley, SwRI, Serco
Exhibit C-5 Booz Allan Hamilton, Inc. (BAH)—tradeoff with Stanley, MaeB, Apptis
Exhibit C-6 General Dynamics One Source, LLC (GD)-tradeoff with Stanley, Mac B, Apptis
Exhibit C-7 BAE Systems, Inc. (BAE)-ti’adeoff with MacB, SwRI, Smartronix
Exhibit C-8 Computer Sciences Corporation (CSC)—tradeoff with Centech, Tybrin, Stanley
Exhibit C-9 I.T.S. Corporation (ITS)—tradeoff with Stanley, MacB, Apptis
Exhibit C-10 Science Application International Corporation (SAIQ—tradeoff with Centech, Tybrin, Stanley
Exhibit C-ll Indus Corporation (Indus)—tradeoff with SWRI, Smartronix, STG
Exhibit C—12 TASC, Inc. (TASQ—tradeoff with Stanley, MaeB, Apptis
Exhibit C-13 ManTech Advanced Systems International, Inc. (ManTech)—tradeoff with Centech, Tybrin, Stanley
Exhibit C-14 QSS Group, Inc. (QSS)—tradeoff with STG, Lucent, Prosoft
Exhibit C-15 Lockheed Martin Integrated Systems, Inc. (Lockheed)—tradeoff with SwRI, Smartronix, STG
Exhibit C-16 Bearing Point, Ine. (Bearing Point)—tradeoff with Stanley, MaeB, CGI
Exhibit C-17 Raytheon Company (Raytheon)—tradeoff with Stanley, MacB, SwRI
Exhibit C-18 Accenture National Security Services LLC; (Accenture)—tradeoff with Stanley, MacB, SWRI
Exhibit C-19 NCI Information Systems, Inc. (NCI)—tradeoff only with Lucent
Exhibit C-20 Unisys Corporation (Unisys)—tradeoff with MacB, SwRI, Serco
Exhibit C-21 Dynamic Research Corporation (DRC)—tradeoff with MacB, SwRI, Serco
Exhibit C-22 International Business Machines Corporation (IBM)—tradeoff with Stanley, MacB, Apptis
Exhibit C-23 MTC Technologies, Ine. (MTC)—tradeoff with SwRI, Smartronix, STG
Exhibit C-24 CACI, Inc.—Federal (CACI)—tradeoff with SwRI, Smartronix, STO
Exhibit C-25A RS Information Systems, Inc. (RSIS)—tradeoff with SwRI, STG, T3 Alliance
Exhibit C-25B AT & T Government Solutions, Ine. (AT & T)—tradeoff with Stanley, MacB, SwRI
Exhibit C-28 Alion Science & Technology Corporation (Alion)—tradeoff with MacB, SWRI, Serco
Exhibit C-29 The Centech Group (Centech)—tradeoff with ManTech, SAIC, CSC Second
Exhibit C-30 Tybrin Corporation (Tybrin)—tradeoff with ManTech, SAIC, CSC
Exhibit C-31 Electronic Data Systems Corporation (EDS)—tradeoff with STG, Lucent, Prosoft
Exhibit C-32 Stanley Associates, Inc. (Stanley)—tradeoff with AT & T, IBM, Accenture
Exhibit C-33 Macaulay-Brown, Inc. (MacB)—tradeoff with Alion, AT & T, IBM
Exhibit C-34 ARINC Engineering Services, Inc. (ARINC)—tradeoff with ManTech, SAIC, CSC
Exhibit C-35 CGI Federal, Inc. (CGI)—tradeoff with Bearing Point, ManTech, SAIC
Exhibit C-36 Apptis, Inc. (Apptis)—tradeoff with IBM, Bearing Point, ManTech
Exhibit C-37 Southwest Research Institute (SwRI)—tradeoff with Alion, TtAMTI, AT & T
Exhibit C-38 Serco, Inc. (Serco)—tradeoff with Alion, AT & T, IBM
Exhibit C-39 CNS, Inc. (CNSI)—tradeoff with ManTech, SAIC, CSC
Exhibit C-40 Perot Systems Government Services (PS)—tradeoff with ManTech, SAIC, CSC
Exhibit C-41 Smartronix, Inc. (SMX)—tradeoff with Alion, TtAMTI, AT & T
Exhibit 042 Advanced Technology Systems, Inc. (ATS)—tradeoff with Alion, AT & T, IBM
Exhibit C-43 Keane Federal Systems, Inc. (Keane)—tradeoff with Alion, AT & T, IBM
Exhibit C-44 Analytical Services, Inc. (ASI)—tradeoff with IBM, Bearing Point, ManTech
Exhibit 045 STG, Inc.—tradeoff with EDS, Alion, TtAMTI
Exhibit C-46 Nortel Government Solutions Inc. (NGS)—tradeoff with Alion, TtAMTI, AT & T
Exhibit 047 Federal Network Systems, LLC (FNS/Verizon)—tradeoff with Bearing Point, ManTech, SAIC
Exhibit 048 Alliant Solutions, LLC—tradeoff with Alion, TtAMTI, AT & T
Exhibit 049 American Systems Corporation—tradeoff with Alion, TtAMTI, AT & T
Exhibit C-50A Artel, Inc.—tradeoff with Alion, TtAMTI, AT & T
Exhibit C-50B T3 Alliance—tradeoff with Alion, TtAMTI, AT & T
Exhibit C-52 Lucent Technologies, Inc. (Lucent)—tradeoff with EDS, Alion, TtAMTI
Exhibit C-53 McNeil Technologies, Inc. (McNeil)—tradeoff with ManTech, SAIC, CSC
Exhibit C-54 TKC Communications, LLC (TKCC)—no tradeoff as TKCC is highest priced proposal
Exhibit 0-55 Professional Software Engineering, Inc. (Prosoft)—tradeoff with EDS, Alion, TtAMTI
Exhibit 0-56 Trawick & Associates (Trawick)—tradeoff with EDS, Alion, TtAMTI
Exhibit 0-57 Engineering and Professional Services, Inc. (EPS)—tradeoff with EDS, Alion, TtAMTI
Exhibit 0-58 Abacus Technology Corporation (Abacus)—tradeoff with Alion, TtAMTI, AT & T
Exhibit C-59 Honevwell Technology Solutions, Inc. (HTSI)—tradeoff with Alion, AT & T, IBM
Exhibit 0-60 Pearson Alliant—tradeoff with Alion, TtAMTI, AT & T
Exhibit C-61 SYS Technologies, Inc. (SYS)—tradeoff with ManTech, SAIC, CSC
Exhibit 0-62 Communication Technologies, Inc. (COMtek)—tradeoff with IBM, Bearing Point, ManTech
. A listing of the technical acronyms used in this opinion may be found in Appendix A.
. See Knowledge Connections, Inc. v. United States, 79 Fed.Cl. 750, 754 (2007) (describing the procurement process under the Clinger-Cohen Act). According to the record, Alliant and a related GWAC for small businesses are the successors to two existing contracts: (i) Applications 'n Support of Widely-Diverse End User Requirements (ANSWER); and (ii) Millenia, due to expire in December 2008 and April 2009, respectively.
. Originally, the Solicitation indicated that no protest under the Federal Acquisition Regulations (FAR) (48 C.F.R.) would be authorized in connection with the issuance or proposed issuance of a task order, except on the grounds that the order increased the scope, period, or maximum value of the Alliant contract. Moreover, another part of the Solicitation indicated that ”[f]ormal evaluation plans or scoring of quotes or offers” on task orders "are not required.” However, Congress recently enacted section 834(a) of the National Defense Authorization Act for Fiscal Year 2008, Pub.L. No. 110-181, 122 Stat. 3, which amended 10 U.S.C. § 2304(c) to provide, over a three-year test period, for GAO protests of task orders valued in excess of $10,000,000. That provision is effective 120 days after the statute’s enactment (January 28, 2008), id. at § 834(a)(2)(C), leading defendant and certain of the parties to conclude that it will apply to Alliant task orders.
. As recounted in the Source Selection Decision (SSD), the number of expected awards "was based upon market research, responses to the [requests for information], outreach meetings, and discussions with potential customers.” The SSD further indicated that ”[t]his number was arrived at to assure adequate competition under the GWAC, meet the objectives of the procurement, and finally to provide the industry a reasonable basis upon which to bid and formulate both their pricing and basic contract plan." Under the Solicitation, GSA reserved the right, over the term of the contract, to terminate particular awardees (an "off ramp”) or to add additional awardees (an "on ramp”) through a new procurement, as the government’s needs dictate.
. Similar language was contained in the Source Selection Plan (SSP).
. Unbeknownst to the offerors, the SSP indicated that “the Government intends to check nine (9) efforts selected from Tables 1 and 2, including the three efforts which the Offeror has identified.” The SSP further indicated that "[i]n the event that an evaluator is unable to reach the point-of-contact for a targeted effort after three attempts (via phone or email), allowing 24 hours between each attempt (which is considered ‘reasonable’), the reference will be marked as ‘not available’ (N/A) and an alternate effort will be substituted.” The SSP provided detailed rules on how an alternate effort should be selected.
. Consistent with the requirements of the FAR, the SSP indicated that—
If an evaluator receives adverse past performance information from a point-of-contact during an interview, the evaluator will provide the adverse information to the PCI, who will decide whether to contact the Offeror and give the Offeror an opportunity to "rebut” the adverse evaluation. The process and decision will be consistent with FAR 15.306.
Offerors will not have the opportunity to comment on adverse information captured in a past performance system, such as PPIRS, as that opportunity has already been provided at the time of the evaluation [sic] submission in the system.
. In the Solicitation, offerors were notified that "the General Service Administration may use the services of non-government evaluators."
. Apart from the questions themselves, the interviewers received only the following guidance—
Please encourage the reference to be as specific as possible about how the Offeror performed a task or accomplished a milestone. Attempt to identify specific performance metrics that were met or exceeded. For example, the following narrative ... “the Offeror exceeded the performance metric which decreased turnaround time by 25% on average” is better than just stating that the Offeror did a "good” job. If the reference identified problems, be sure to ask what corrective actions were taken, if any, to address the problem. The key is assessing the Offeror’s performance risk, so encourage the responder to describe how something was accomplished or why a certain action took place. Again, please encourage your interviewee to provide you with objective, quantifiable, or specific statements in your transcript whenever possible.
(Emphasis in original). The instructions ended with the following bolded statement—“Please email the transcript of the response to the point-of-contact you interviewed to verify the responses.”
. If the evaluators came across adverse past performance information in the interviews, the established procedures required them to forward the information to the Contracting Officer for him to determine whether to provide the affected offeror with an opportunity to respond to the negative information.
. To the extent that any subfactor received a neutral score, the denominator in this calculation shrunk by one (e.g., if the proposal had four numeric scores and one neutral score, the average was produced by adding the numeric scores and dividing by 4).
. It appears that the process used to establish the BCP rating deviated from the SSP, which provided—
After the evaluations are completed by each evaluator, the team leaders will caucus members to form a consensus evaluation for each Offeror, that is, a single adjectival rating for each Technical Factor for each Offeror. Each team will rank order Offerors for their technical factor. Individual evaluator comments will be consolidated into potential discussion questions. BCP team members will evaluate an Offeror's resources using the specific adjectival ratings in Annex D[] as discriminators.
. The SSP defined “Adequate Price Competition" using the following bullets:
• Two or more responsible Offerors, competing independently, submit priced offers that satisfy the Government’s expressed requirement;
• Award will be made to the Offeror whose proposal represents the best value where price is a substantial factor in source selection; and
• There is no finding that the price of the otherwise successful Offeror is unreasonable. Any finding that the price is unreasonable must be supported by a statement of the facts and approved at a level above the Contracting Officer.
It further indicated that “overall Price analysis will consist of:”
• Adequate Price Competition;
• []In relation to the IGCE for each loaded hourly labor rate price and each Median hourly labor rate price among all offers ...; the Government will identify those rates that vary significantly;
• []In relation to the IGCE for Overall Price and the Median Overall Price among all offer-ors (Government and Contractor Site combined over 10 years); the Government will identify Overall Pricing that varies significantly-
The plan added ”[w]hen the IGCE and Median vary significantly to one another, the Median labor rates and overall Median Price will take precedence for adequate price competition.”
. The record contains only the exhibits (Exhibits Cs) summarizing the analysis as to plaintiffs, the defendant-intervenors and certain other of-ferors.
. ManTech and SAIC had prices well above the IGCE and above two standard deviations from the mean overall price among offerors, and CSC’s price was above the IGCE and fell just
. Mr. Kempf noted that while Centech’s proposal ranked 29th in terms of technical quality, it was one of the highest-priced proposals, at a rank of 56th out 62 offerors. Highlighting particular features of that offer, Mr. Kempf indicated that—
Centech’s lesser technical merit than other Of-ferors chosen for award was centered in its Basic Contract Plan. In some cases, Centech did not have the breadth of resources that other successful Offerors brought to the table. It did not provide information on the size of its workforce and it had a comparatively small number of personnel (161) with security clearances with no comprehensive plan to boost this number nor did it provide any detailed information on its plans for providing a staff with security clearances when needed. Though Centech indicated that it had a "significant” OCONUS presence for a company of its size, its proposal did not demonstrate an understanding or the ability to quickly staff an OCO-NUS task order request (TOR).
Mr. Kempf’s comment regarding Stanley were less kind. Thus, he stated that ”[w]hen compared with other Offeror’s, Stanley’s proposal was, in general, undistinguished,” adding that ”[t]here was nothing remarkable in its area of corporate commitment and its lack of innovation was noted in several evaluation subfactors.” In addition, after summarizing negative comments received from several of Stanley's past performance references, Mr. Kempf stated that ”[t]here was noted risk, based on Stanley’s track record, that Stanley may not have the innovative tools and methodologies to support Alliant customers when compared to other Offerors that employed innovative and very effective quality assurance techniques, including highly effective metrics and quality controls.”
. Several parties attribute the delay in filing their complaints to the fact that it took the GAO a number of weeks to dismiss all the Alliant protests pending before it. On October 1, 2007, GAO notified all the protesters of the filing of Serco’s case and invited them to submit comments regarding whether their respective protests should be dismissed under § 21.11(b) of the GAO's bid protest regulations. The latter provision states: "GAO will dismiss any case where the matter involved is the subject of litigation before, or has been decided on the merits by, a court of competent jurisdiction.” After GAO received those comments, several weeks passed before it dismissed all the protests—the latest such dismissal occurred on October 31, 2007 (Apptis withdrew its protest on November 8, 2007, while still awaiting a dismissal). Of course, nothing prevented any of the plaintiffs from voluntarily dismissing their GAO protests and immediately filing in this court—as some did.
. In an affidavit filed with the court explaining this action, Mr. Kempf admitted that a mistake had been made in assigning an adjectival rating to Stanley’s purchasing system. Correcting that single mistake increased Stanley’s Weighted BCP/PP Average Score from 3.769 to 3.821 and caused GSA to take "corrective action” in making an award to Stanley. Explaining this decision, Mr. Kempf stated—
After this adjustment, I see little difference between the price and technical merit of Stanley when compared to the next highest ranked offeror selected for award: Alion. As I discussed in my Source Selection Decision Memorandum (SSDM), the agency’s announced goal was to award 25 to 30 contracts to the best value proposals. The corrected presentation of data, above, reveals a natural break between Stanley and CENTECH rather than between Alion and CENTECH, as in the original award. CENTECH and Tybrin have less technical merit and a significantly higher price thus not a best value for the Government; each has documented weaknesses or only acceptable Basic Contract plans that do [not] warrant a best value award as discussed in Exhibits C-29 and C-30 of the SSDM.
Mr. Kempf did not comment upon or renege any of the criticisms that he had previously leveled at Stanley, to wit, that its proposal was "undistinguished,” "demonstrated a lack of innovation,” and involved "noted risk” by the SSET. In terms of technical ranking, the new "natural break” described by Mr. Kempf was between Stanley at 3.821 and CENTECH at 3.813.
. Bannum was based upon RCFC 56.1, which was recently abrogated and replaced by RCFC 52.1. However, the latter rule was designed to incorporate the decision in Bannum. See RCFC 52.1, Rules Committee Note (June 20, 2006); see also NVT Techs., Inc. v. United States, 73 Fed.Cl. 459, 462 n. 3 (2006); Bice v. United States, 72 Fed.Cl. 432, 441 (2006).
. In Bannum, the Federal Circuit noted that, in Banknote Corp. of Am. Inc. v. United States, 365 F.3d 1345, 1352-53 (Fed.Cir.2004), it had erroneously conflated the standards under RCFC 56 and 56.1, albeit in dicta. In this regard, the Bannum court stated that—
Although it never reached the factual question of prejudice, the Banknote II court added that it is the trial and appellate courts' task to "determine whether there are any genuine issues of material fact as to whether the agency decision lacked a rational basis or involved a prejudicial violation of applicable statutes or regulations.” This language equates a RCFC 56.1 judgment to a summary judgment under RCFC 56 and is unnecessary to the Banknote II holding. Because the court decided the issue by an interpretation of the solicitation, e.g., making a legal determination, the court in Banknote II did not need to consider whether the trial court overlooked a genuine dispute or improperly considered the facts of that case.
Bannum, 404 F.3d at 1354. Prior decisions of this court have made the same error. See, e.g., JWK Int'l Corp. v. United States, 49 Fed.Cl. 371, 387 (2001), aff'd, 279 F.3d 985 (Fed.Cir.2002). Indeed, while various decisions of this court refer to a "motion for summary judgment on the administrative record,” see, e.g., ManTech Telecom. & Info. Sys. Corp. v. United States, 49 Fed.Cl. 57, 64-65 (2001), aff'd, 30 Fed.Appx. 995 (Fed.Cir.2002), there was no such thing under former RCFC 56.1 (nor is there under RCFC 52.1), as properly construed.
. See Murakami v. United States, 46 Fed.Cl. 731, 734 (2000); Aero Corp., S.A. v. United States, 38 Fed.Cl. 408, 410-11 (1997); see also Florida Power & Light Co. v. Lorion, 470 U.S. 729, 743-44, 105 S.Ct. 1598, 84 L.Ed.2d 643 (1985) ("The task of the reviewing court is to apply the appropriate APA standard of review ... to the agency decision based on the record the agency presents to the reviewing court;” where a record in an arbitrary and capricious review is incomplete, "the proper course, except in rare circumstances, is to remand to the agency for additional investigation or explanation. The reviewing court is not generally empowered to conduct a de novo inquiry into the matter being reviewed and to reach its own conclusions based on such an inquiry.”); FCC v. ITT World Communications, Inc., 466 U.S. 463, 469, 104 S.Ct. 1936, 80 L.Ed.2d 480 (1984); Camp v. Pitts, 411 U.S. 138, 142, 93 S.Ct. 1241, 36 L.Ed.2d 106 (1973) (the focal point of arbitrary and capricious review "should be the administrative record already in existence, not some new record made initially in the reviewing court”). As this court has explained elsewhere, Esch v. Yeutter, 876 F.2d 976 (D.C.Cir.1989)—often cited as authority for "supplementing” the administrative record—-is heavi
. As noted by the Federal Circuit, "[pjrocurement officials have substantial discretion to determine which proposal represents the best value for the government.” E.W. Bliss Co. v. United States, 77 F.3d 445, 449 (Fed.Cir. 1996); see also Galen Med. Assocs., Inc. v. United States, 369 F.3d 1324, 1330 (Fed.Cir.2004); TRW, Inc. v. Unisys Corp., 98 F.3d 1325, 1327-28 (Fed.Cir. 1996); LaBarge Prods., Inc. v. West, 46 F.3d 1547, 1555 (Fed.Cir. 1995) (citing Burroughs Corp. v. United States, 223 Ct.Cl. 53, 617 F.2d 590, 597-98 (1980)); EP Prods., Inc. v. United States, 63 Fed.Cl. 220, 223 (2005), aff'd, 163 Fed.Appx. 892 (Fed.Cir.2006); JWK Int'l Corp. v. United States, 52 Fed.Cl. 650, 655 (2002), aff'd, 56 Fed.Appx. 474 (Fed.Cir.2003).
. In Banknote Corp., the Federal Circuit expounded upon these principles, as follows:
Under the APA standard as applied in ... [bid protest] cases, "a bid award may be set aside if either (1) the procurement official’s decision lacked a rational basis; or (2) the procurement procedure involved a violation of regulation or procedure.” [Impresa Construzioni Geom. Do-menico Garufi v. United States, 238 F.3d 1324, 1332-33 (Fed.Cir.2001)]. When a challenge is brought on the first ground, the test is “whether the contracting agency provided a coherent and reasonable explanation of its exercise of discretion, and the disappointed bidder bears a 'heavy burden' of showing that the award decision had no rational basis.” Id. at 1332-33 (citations omitted). "When a challenge is brought on the second ground, the disappointed bidder must show a clear and prejudicial violation of applicable statutes or regulations.” Id. at 1333.
Banknote Corp., 365 F.3d at 1351; see also Seattle Sec. Servs., Inc. v. United States, 45 Fed.Cl. 560, 566 (2000); Analytical & Research Tech., Inc. v. United States, 39 Fed.Cl. 34, 42 (1997).
. A review of Federal Circuit cases indicates that this prejudice analysis actually comes in two varieties. The first is that described above— namely, the ultimate requirement that a protestor must show prejudice in order to merit relief. A second prejudice analysis focuses more preliminarily on standing. In this regard, the Federal Circuit has held that “because the question of prejudice goes directly to the question of standing, the prejudice issue must be reached before addressing the merits.” Info. Tech. & Applications Corp. v. United States, 316 F.3d 1312, 1319 (Fed.Cir.2003); see also Myers Investigative Sec. Servs., Inc. v. United States, 275 F.3d 1366, 1370 (Fed.Cir.2002); Overstreet Elec. Co., Inc. v. United States, 59 Fed.Cl. 99, 109 (2003). Cases construing this second variation on the prejudice inquiry have held that it requires merely a "viable allegation of agency wrong doing,” with " 'viability' here turning on the reasonableness of the likelihood of prevailing on the prospective bid taking the protestor’s allegations as true.” McKing Consulting Corp. v. United States, 78 Fed.Cl. 715, 721 (2007); see also 210 Earll, LLC v. United States, 77 Fed.Cl. 710, 719 (2006); Textron, Inc. v. United States, 74 Fed.Cl. 277, 285 (2006). Because of the serious nature and breadth of the numerous allegations of error here, it requires little effort to conclude that all the plaintiffs in this case meet this preliminary "standing” threshold—and defendant, to its credit, does not argue otherwise.
. See also Seattle Sec. Servs., 45 Fed.Cl. at 569; Incident Catering Servs., LLC, 2005 CPD ¶ 193, 2005 WL 3193685, at *4 (2005); U.S. Prop. Mgmt. Serv. Corp., 98-1 CPD ¶ 88, 1998 WL 126845, at *4 (1998); Wind Gap Knitwear, Inc., 95-2 C.P.D. ¶ 124, 1995 WL 368421, at *2 (1995).
. The scripts were neither attached to the Solicitation nor otherwise provided to the offerors prior to the award decisions. Accordingly, the complaints made about them are timely. Compare Halter Marine, Inc. v. United States, 56 Fed.Cl. 144, 169 (2003); SWR, Inc., 97-2 C.P.D. ¶ 34 (1997), 1997 WL 422231, at *3 (1997).
. For cases discussing procurements in which surveys were used, see, e.g., Southern Foods, Inc. v. United States, 76 Fed.Cl. 769, 779 (2007); Day & Zimmermann Servs., a Div. of Day & Zimmermann, Inc. v. United States, 38 Fed.Cl. 591, 609 (1997); S3 LTD, 2001 C.P.D. ¶ 165, 2001 WL 1105271, at *5 (2001); Beneco Enters., Inc., 2000 C.P.D. ¶ 175, 1999 WL 33218776, at *2-3 (1999); Cont’l Serv. Co., 97-1 C.P.D. ¶ 9, 1996 WL 753889, at *2-3 (1996); see also Ralph C. Nash & John Cibinic, "Postscript IV: Past Performance Evaluations,” 16 No. 3 Nash & Cibinic Rep. ¶ 14 (2002) ("Data can be obtained from references by questionnaire or telephone interview.”). Indeed, the Office of Federal Procurement Policy (OFPP) has provided guidance on how such surveys should be conducted. See Office of Mgmt. & Budget, OFPP, Best Practices for Collecting and Using Past Performance Information (2000) ("If adequate documentation is not readily available ... then a brief survey with follow up calls, or phone interviews should be used to verify past performance.”); id. at Appendix I (providing a sample survey form).
. See, e.g., Cooperativa Muratori Riuiniti, 2005 C.P.D. ¶ 21, 2005 WL 277303, at *6 (2005) (survey that was not geared to the evaluation criteria held to be arbitrary and capricious); FC Construction Co., Inc., 2001 CPD ¶ 76, 2001 WL 370895, at *5 (2001) (telephone survey not arbitrary capricious, where conducted by contract administrator who read references twenty-six questions, provided the references with the definitions for adjectival ratings, and allowed references to review completed questionnaires); Dismas Charities, Inc., 2003 C.P.D. ¶ 125, 2003 WL 21665202, at *4 (2003) (questionnaires were "materially flawed” and contrary to the FAR when the agency did not provide references with instructions on how to score offeror’s past performance); ENMAX Corp., 99-1 C.P.D. ¶ 102, 1999 WL 335687, at *5 (survey was reasonable where it includes seventeen questions that were designed to correspond with the three technical evaluation factors and required the reference to rate the offeror as "above average,” "average,” "below average,” or "not observed”); see also Maint. Engrs. v. United States, 50 Fed.Cl. 399, 421 (2002); S3 LTD, 2001 WL 1105271, at *5.
. In some instances, it appears that particular awardees benefitted from isolated comments that happened to coincide with the definitions in the adjectival ratings. For example, while INDUS comments were very similar to those of CGI (e.g., "very well,” "effective,” "very good”), one of its references indicated that it "always meet[s] or exceed[s]” its schedules. Similarly, while several of ManTech’s references indicated that it was "very effective,” "really good,” and "very good" in meeting schedules, one of its references reportedly said “[t]hey have pretty much exceeded everything that we have asked them.”
. The distinction drawn in the Solicitation between firm, fixed-price contracts and cost-reimbursement contracts is logical and consistent with the FAR. Under the former 1ype of contract, the contractor absorbs all cost risk, whereas under the latter, the government must reimburse the contractor’s allowable costs. See 48 C.F.R. §§ 16.202-1, 16.301-1. Accordingly, while a contractor’s ability to control costs is critically important to the government in a cost-reimbursement contract, it is of little moment in a firm, fixed-price contract.
. To be sure, one paragraph of the instructions that came with the scripts urged the interviewers to "[a]ttempt to identify specific performance metrics that were met or exceeded.” But, neither this paragraph nor anything else alerted the interviewers that the degree to which those metrics were exceed was critical to whether the offerors would receive an EH (“exceeds many requirements”) or a S rating (“exceed some requirements”). Moreover, while this same paragraph cautioned interviewers that specific answers were better than having the reference just state that the offeror did a "good” job, a review of the transcripts reveals dozens of such generic answers, suggesting that the agency did not enforce this instruction, either.
. At oral argument, defendant speculated that the evidence documenting this confirmation simply did not find its way into the administrative record. The court has several problems with this contention. First, the instructions to Calyptus indicated that the completed transcripts were to be provided to the references via e-mail. The email to the IBM references is in the record, but no other such e-mails appear within the approximately 24,000 pages of the record—at least none that the court has found or defendant has cited. Second, the transcripts that were provided to the IBM references came back heavily marked, yet no other transcripts in the record contain any markings or indications of editing. Finally, defendant could have provided materials supporting its claim on numerous occasions, including after these arguments were raised by various plaintiffs—but did not. In these circumstances, the court is compelled to render its decision based upon the record actually before it—a record that does not support the claim that this corroboration step was performed.
. To be sure, GSA’s failure to follow the SSP in this regard is not an independent basis for setting aside the awards. See, e.g., ManTech Telecomm., 49 Fed.Cl. at 67; Quality Sys., Inc., 89-2 C.P.D. ¶ 197, 1989 WL 241122, at *9 (1989). Yet, its failure to comply with a step that it apparently believed was important to validate the survey results certainly sheds light on whether its conduct of the surveys was arbitraiy and capricious.
. The systemic nature of the errors encountered with respect to the past performance survey obviates the necessity for the court to consider most of the claims by individual plaintiffs that errors were made in converting the comments of their references into adjectival ratings. Some of these claims plainly are misplaced—for example, GSA did not act arbitrarily in assigning neutral ratings to offerors that lacked certain past performance experience. See Metcalf Constr. Co. v. United States, 53 Fed.Cl. 617 (2002). In addition, the court rejects Serco’s argument that GSA failed to provide offerors with the opportunity to rebut "adverse” past performance information. The Solicitation stated that the evaluator was to "provide adverse information to the PCO [procuring contracting officer],” and that the PCO would "decide” whether the offeror should be given the "opportunity to rebut’ the information.” This language tracks FAR § 15.306(a)(2), which states that ”[i]f award will be made without conducting discussions, offerors may be given the opportunity to clarify ... adverse past performance information to which the offeror has not previously had an opportunity to respond.” While the regulation does not define what it means by "adverse" past performance information, the adjectival ratings here defined such information as that which resulted in a past performance rating of "L/N" or "1.” None of plaintiffs received such a rating. Even if they had, neither the FAR nor the Solicitation guaranteed that an offeror would have an opportunity to respond to such information—both talk only in permissive terms. See DynCorp., Int'l, LLC v. United States, 76 Fed.Cl. 528, 540 (2007); see also JWK Int'l Corp., 52 Fed.Cl. at 661. And, certainly, no plaintiff has shown that GSA abused its discretion in failing to provide an opportunity to respond. Finally as to this segment, the court has considered, but rejects, various assertions that GSA treated certain plaintiffs unequally in evaluating their BCPs.
. See Joseph S. Danowsky, "Statistical Report Flaws: A Spotter’s Guide,” 18 No. 8 Acct. & Fin.
Measured numbers are not exact. They depend on our skill and care in making the measurements. Because all measured numbers are approximate, it is important to know how well the reported number represents the actual value. The comparison of a measurement to the actual or true value is defined as accuracy and is typically presented as a percent difference from the true value ... Measurements are only as accurate or precise as the device that measures them. In other words, if water levels are measured with a steel tape and chalk that is only marked in feet, and then the water level is re-measured using an electronic monitoring system that provides measurements that are marked in tenths of an inch, a more accurate measurement is made with the second tape than with the first. If measurements are made over a six-month period using both techniques and then measurements are averaged, the average value will only be as accurate as the measurements made with the first tape (the least precise measuring device). The number of digits and the decimal notation of this average are determined by the
significant digits of each of the measurements. James W. Conrad, Jr., Env. Sci. Deskbook § 1:7 (2007) (emphasis in original). This passage concludes that “[slummary statistics and values can only have as many significant digits as the measurement used with the least number of significant digits.” See also Philip N. Baldwin Jr., Statistics: Know-How Made Easy 30 (1999).
. As discussed above, the record demonstrates that GSA did not comply fully with portions of the SSP that had been designed to improve the accuracy of its preliminary review, such as requiring the contractor to provide the transcript of their interviews to the past performance references for verification.
. See M-Cubed Info. Sys., Inc., 2000 CPD ¶ 74, 2000 WL 656188, at *6 (2000) (agency properly determined that past performance scores of 39.5 and 42.5, as well as technical scores of 88.00 and 87.70, were "technically equal”); The Parks Co., 92-2 C.P.D. ¶ 354, 1992 WL 346769, at *2-3 (1992) (agency properly determined that technical scores of 2,860 and 2,515 points out of the available 3,000 weighted points were "essentially equal technically.”); Correa Enters., Inc., 91-1 C.P.D. ¶ 249, 1991 WL 73075, at *2 (1991) (agency properly determined that technical scores of 91.75 and 90.50 points on a 100 point scale were "essentially technically equal”); DDL Omni Eng'g, 85-2 C.P.D. ¶ 684, 1985 WL 53700, at *5 (1985) (agency reasonably determined that proposals were "substantially equal” despite a 0.35 point difference on a scale of 10); see also Training and Info. Servs., Inc., 87-1 C.P.D. ¶ 266, 1987 WL 96915, at *3 (1987) (protest sustained—0.2 point differential (on a 100 point scale) between protester and awardee, after correction of errors, held to be "an almost insignificant point difference”).
. The court rejects plaintiffs' claim that the BCP/PP averages effectively gave disproportionate weight to the BCP rating. As the court understands them, these claims essentially focus on the fact that there was much more dispersion under the BCP rating than under the past performance rating, so that when the two ratings were averaged, the BCP rating had more impact on where a given offeror would be ranked. But assuming for the sake argument that the dispersion under the BCP rating was bona fide (which remains to be seen), there is nothing about the spread of those rankings that violated the Solicitation's promise that past performance and basic contract ratings would be "approximately equal in importance.”
. The sports-minded reasonably might ask how the situation here differs from the calculation of batting averages. After all, that calculation begins with inputs that are whole numbers, yet ends with averages that are regularly carried to the third decimal place and, if necessary to break a tie, to a fourth place (e.g., George Kell versus Ted Williams in 1949). But, there are two major differences between the present case and the calculations used to determine the batting title. The first is that, unlike this case, the inputs into the batting average equation are certain—e.g., two hits in four at bats—and not approximate measurements or other observations that incorporate some level of imprecision or rounding that necessitates the use of an error bound. Averaging such figures thus does not raise issues concerning error propagation. The second distinction involves the use of the statistics. In baseball, the decision being made based upon the batting averages is defined in terms of the average itself—the player with the highest average wins the batting title. Here, that insularity is lacking. The ultimate question here is not which firm has the highest average technical ranking, but rather which presents the best value to the government. It is the context of the decision, then, that ultimately suggests that caution should be employed in using approximate numbers. Indeed, no baseball purist, certainly not Bill James (the guru of baseball statistics), would point solely to a batting average, carried to whatever number of digits, to resolve who is the better hitter (e.g., Babe Ruth and "Big Dan" Brouthers both have career batting averages of .3421). See Bill James, The New Bill James Historical Baseball Abstract (2003).
. Several offerors argue that GSA created a de facto "competitive range” when it designated the presumptive award group. While the FAR does not specifically define when such a range arises, it indicates that if such a range is established, the agency must notify excluded offerors and provide them with an opportunity for debriefing. FAR § 15.306(c)(3)-(4). Yet, even without a definition, it would appear that the sin qua non of a “competitive range" is that it excludes certain offerors from further consideration. See M.W. Kellogg Co. v. United States, 10 Cl.Ct. 17, 23 (1986). Such was not the case with the "presumptive awardee” group established by GSA.
. See also C.W. Gov't Travel, Inc., 2005 C.P.D. ¶ 139, 2005 WL 1805945, at *4 (2005) ("[t]he statutory requirement that cost to the government be considered in the evaluation and selection of proposals for award is not satisfied by the promise that cost or price will be considered later, during the award of individual task orders”); S.J. Thomas Co., Inc., 99-2 C.P.D. ¶ 73 (1999), 1999 WL 961750, at *3 (1999) (same); SCIENTECH, Inc., 98-1 C.P.D. ¶ 33, 1998 WL 29236, at *5 (1998) (same); Ralph C. Nash & John Cibinic, "Using Best Value to Select GWACs Contractors; A Flawed Procurement,” 19 No. 7 Nash & Cibinic Rep. ¶ 35 (2005).
. While it is neither for defendant nor this court to question why price is relevant here—it is enough that it is—one can conceive of at least two answers to this question. First, it is reasonable to assume that offerors that list reasonable prices under the GWAC will continue to do so when they compete for task orders. Indeed, one of the government's objectives, in conducting this GWAC, is to line up a diverse group of reasonably-priced offerors to bid on those orders. Second, failing to consider prices at this stage would raise the specter that offerors would provide basic contract plans that are economically infeasible (e.g., gold-plated)—plans that have wonderful options, to be sure, but only at prohibitive
. Miguel de Cervantes Saavedra, The Ingenious Hidalgo Don Quixote of La Mancha, Part ii, Chap, xxiv (1615).
. Defendant makes much of the fact that the SSA mentioned price in setting the "natural break point” between the 28th and 29th ranked offerors. But, it essentially ignores the fact that the SSET, in setting the same "natural break point" between the 27th and 28th ranked offer-ors made no mention whatsoever of price. Indeed, if one looks more broadly at the rankings, it is hard to believe that the SSA gave serious consideration to price in setting this "natural break point." The last of his presumptive awar-dees, Alion, had a weighted BCP/PP Average of 3.829 and a price of $33,007,426, while the 33rd ranked offeror, MacB, had a BCP/PP Average of 3.754 and a price of $[]—its technical ranking was 1.9 percent lower than that of Alion, but its price was [] percent lower. Likewise, the 37th ranked offeror, SwRI, had a BCP/PP Average of 3.738 and a price of $[]—its technical ranking was 2.3 percent lower than that of Alion, but its price was [] percent lower. Indeed, had defendant properly accounted for the lack of precision in its technical rankings—leading it likely to conclude that the offerors were much more bunched together—it would have been obliged to give price an even greater role in the award decisions, as the Solicitation plainly advised that "the closer the technical scores of the various proposals are to one another the more important cost or price considerations become in determining the overall best-value for the Government.” That did not occur.
. In the court’s view, the foregoing discussion fully disposes of the argument made by Centech that GSA gave price too much consideration.
. In other contexts, an agency may, for example, cancel a solicitation if its receives unreasonable prices from all the offerors. See FAR §§ 14.404-l(c)(6); see also Nutech Laundry & Textiles, Inc., 2003 C.P.D. ¶ 34, 2003 WL 282208, at *3 (2003).
. Consistent with this view, agencies in other procurements have used the results of a price reasonableness analysis to conclude that an award could not be made to a particular offeror. See, e.g., Crowley American Transp., Inc., 95-1 C.P.D. 277, 1995 WL 366985, at *3 (1995); see also Ralph C. Nash & John Cibinic, "Cost and Price Analysis: Understanding the Terms," 9 No. 1 Nash & Cibinic Rep. ¶ 5 (1995) ("The purpose of a price reasonableness analysis is to ascertain that the Government is not paying too high a price.").
. Compare Moore's Cafeteria Servs. d/b/a MCS Management, 2007 C.P.D. ¶ 99, 2007 WL 1746378, at *4 (2007) (price was fair and reasonable where higher than protester but lower than the third technically acceptable offer); U.S. Dynamics Corp., 2007 C.P.D. ¶ 21, 2006 WL 4043686, at *3 (2006) (price reasonableness established based on "adequate price competition” where there was a 4 percent differential between the price of the offerors); Clearwater Instrumentation, Inc., 2001 C.P.D. ¶ 151, 2001 WL 1047078, at *4 (2001) (price reasonableness analysis was proper where prices proposed by three competitors "were relatively consistent and in line with each other”); ITT Fed. Sys. Int’l Corp., 2001 C.P.D. ¶ 45 (2001), 2001 WL 238559, at *9 (2001).
. At oral argument, defendant asserted that the focus of the price reasonableness analysis should only be on the reasonableness of the profit factor built into a particular price. Under this approach, competition could be viewed as uniformly establishing price reasonableness based on the assumption that an offeror would not propose a price that included an uncompetitive profit margin. Defendant, however, could supply no authority in support of its cramped view of the role of price reasonableness. This is not surprising as its invention runs counter to the entire thrust of the FAR, which clearly is concerned that awards not be made to offers with unreasonable prices, whatever the cause. As once noted by the Court of Claims, “price does not become a reasonable one for the Government to pay simply because it is thought to yield the seller only a minimal profit.” Gibraltar Mfg. Co. v. United States, 212 Ct.Cl. 226, 546 F.2d 386, 390-91 (1976).
. This is not to say that these prices could not be shown to be fair and reasonable under another method, perhaps one of the others listed in FAR § 15.404—1(b)(2), including an analysis of pricing information supplied by the offeror. Indeed, various plaintiffs argue that GSA erred in failing to conduct a cost realism analysis that focused on the various labor categories that were
. See Beautify Prof'l Servs. Corp., 2003 C.P.D. ¶ 178, 2003 WL 22339300, at *4 (2003) ("In a best value procurement, it is the function of the source selection authority to perform a price/ non-price factor tradeoff, that is, to determine whether one proposal’s superiority under the non-price factor is worth a higher price. This tradeoff process allows an agency to accept other than the lowest-priced proposal.”); Johnson Controls World Servs., Inc., 2002 C.P.D. ¶ 88, 2002 WL 1162912, at *4 (2002) (“The propriety of the cost/price-technical tradeoff decision turns not on the difference in the technical scores or ratings per se, but on whether the selection official’s judgment concerning the significance of the difference was reasonable and adequately justified in light of the RFP’s evaluation scheme.”); Opti-Lite Optical, 99-1 C.P.D. ¶ 61, 1999 WL 152145, at *3-4 (1999) ("While adjectival ratings and point scores are useful as guides to decision-making, they generally are not controlling, but rather, must be supported by documentation of the relative differences between the proposals, their strengths, weaknesses and risks, and the basis and reasons for the selection decision"); Cygnus Corp., 97-1 C.P.D. ¶63, 1997 WL 46987, at *11 (1997).
. Indeed, issues of false precision akin to those discussed above can creep into the process if an agency mechanically relies on a purely mathematical price/technical tradeoff methodology. As Messrs. Cibinic and Nash once wrote, "[sjome of these formulas have used a total point system by assigning points to the price, others have assigned dollar values to the non-price factors, and a third has divided the price by the points assigned to non-price factors to come up with a dollars per point figure.” Cibinic & Nash, supra at 715. Decisions reviewing these formulae have found them unobjectionable provided that they were used only as evaluation techniques and not as the actual basis for making best value determinations. See, e.g., Med. Dev. Int’l, 99-1 C.P.D. ¶ 68, 1999 WL 194481, at *6-7 (1999); Opti-Lite Optical, 1999 WL 152145, at *5; Teltara, Inc., 98-2 C.P.D. ¶ 124, 1998 WL 841469, at *4 (1998); Moran Assocs., 91-2 C.P.D. ¶ 495, 1991 WL 296778 (1991); Storage Tech. Corp., 84-2 C.P.D. ¶ 190, 573 N.W.2d 625, 1998 WL 46528, at *3 (1998); Harrison Sys., Ltd., 84-1 C.P.D. ¶ 572, 1984 WL 43532, at *4-5 (1984).
. See, e.g., Magellan Health Servs., 2007 C.P.D. ¶ 81, 2007 WL 1469049, at * 15 (2007); SOS Interpreting Ltd., 2005 C.P.D. ¶ 26, 2005 WL 357422, at *5-6 (2004); Blue Rock Structures, Inc., 2004 C.P.D. ¶ 63, 2004 WL 414581, at *3-4 (2004); Johnson Controls World Servs., 2002 WL 1162912, at *5; AIU N. Am., Inc., 2000 C.P.D. ¶ 39, 2000 WL 255431, at * 7 (2000).
. It is, of course, axiomatic that, while the court may uphold a decision of less than ideal clarity, the court "may not supply a reasoned basis for the agency's action that the agency itself has not given.” Motor Vehicle Mfrs. Assn., 463 U.S. at 43, 103 S.Ct. 2856; see also Bowman Transp., Inc. v. Arkansas-Best Freight Sys., Inc., 419 U.S. 281, 285, 95 S.Ct. 438, 42 L.Ed.2d 447 (1974).
. See Remington Arms Co., Inc., 2006 C.P.D. ¶ 32, 2006 WL 327974, at * 13 (2006) (“The propriety of such a price/technical tradeoff decision turns not on the difference in the technical scores or ratings per se, but on whether the selection official’s judgment concerning the significance of the difference was reasonable and adequately justified in light of the RFP's evaluation scheme.”); Blue Rock Structures, Inc., 2004 WL 414581, at *3 (requiring the tradeoff analysis to furnish an explanation as to why technical advantages warrant a price premium); General Offshore Corp., 92-2 C.P.D. ¶ 335, 1992 WL 79117, at *4 (1992) (“The determining element is not the difference in technical merit, per se, but the contracting agency's judgment concerning the significance of that difference.”); see also A & D Fire Protection, Inc., 2002 C.P.D. ¶ 74, 2002 WL 841331, at *3 (2002); Tecom, Inc., 94-2 C.P.D. ¶ 212, 1994 WL 683269, at *5 (1994); Oshkosh Truck Corp., 1993 WL 335049, at *6.
. As noted above, in performing best value anal-yses, agencies have sometimes divided the offering price by the points assigned to non-price factors to come up with a dollars per point figure. See Cibinic & Nash, supra at 715. Such a calculation here, made using the Weighted BCP/PP Average scores, reveals wide disparities between offerors. At the bottom of the scale, there are awardees such as NCI and plaintiff STG, at $5,284,517 and $[] per point, respectively. On the other end of the spectrum, we find awardees, such as ManTech and SAIC, at $13,414,596 and $12,365,996 per point, respectively—essentially double the numbers for NCI and STG. These figures are informative not because the agency was required to conduct such statistical studies as part of its tradeoff analysis, but because the disparities demonstrate that GSA needed to conduct more careful tradeoff analy-ses, in whatever form they took.
. This is not to say that the tradeoff analysis for ManTech and for certain other high-priced offer-ors necessarily met the requirements for the FAR. While the agency might have provided an adequate explanation of the benefits that it felt were worth paying these higher price, it remains that the agency did not properly determine whether those price were fair and reasonable. Whether that means that the tradeoff analysis themselves were deficient or were based upon a false predicate, i.e., that the prices being analyzed were fair and reasonable, is a distinction without meaning at this point. It would seem that the agency first needs to determine whether a given price is fair and reasonable and then may consider whether it is worth paying any premium represented by that price.
. The plaintiffs making this assertion have cited nothing in support thereof. Requiring the SSA to make comparisons even with no likelihood that a tradeoff would occur makes no sense and runs counter to the discretion afforded to the SSA in making tradeoff decisions.
. Establishing the presumptive group and requiring a compelling reason for disturbing it not only placed further undue reliance on its technical rankings, but arguably straight-jacketed what should be a flexible approach in making best value determinations. See, e.g., Gen. Offshore Corp., 1992 WL 79117, at *4 ("[W]e consistently have stated that evaluation scores are merely guides for the selection official, who must use his judgment to determine what the technical difference between the competing proposals might mean to contract performance, and who must consider what it would cost to take advantage of it.").
. Support also exists for the proposition that the denial of the right to have a bid fairly and lawfully considered constitutes irreparable harm. See Ellsworth Assocs., Inc. v. United States, 45 Fed.Cl. 388, 398-99 (1999) (citing cases); but see Minor Metals, Inc. v. United States, 38 Fed.Cl. 379, 381-82 (1997) (noting "economic harm without more, does not seem to rise to the level of irreparable injury”).