ELECTRONIC ON-RAMP, INC. v. USA, No. 1:2014cv01038 - Document 43 (Fed. Cl. 2015)
Court Description: REPORTED OPINION. Signed by Judge Elaine D. Kaplan. (ca) Copy to parties.
In the United States Court of Federal Claims (Bid Protest) No. 14-1038C (Filed Under Seal: March 18, 2015 | Reissued: March 30, 2015)* ) THE ELECTRONIC ON-RAMP, INC., ) ) Plaintiff, ) ) v. ) ) THE UNITED STATES OF AMERICA, ) ) Defendant, ) and ) ) PREMIER MANAGEMENT ) CORPORATION, ) Defendant-Intervenor. ) ) Post-award bid protest; Best-Value Procurement; FAR Part 15; Deficiency; FAR 15.001. William Thomas Welch, McMahon, Welch & Learned, Reston, VA, for plaintiff. Barbara E. Thomas, Commercial Litigation Branch, Civil Division, United States Department of Justice, Washington D.C., for defendant. With her on the briefs were Robert M. Norway, Martin F. Hockey, Jr., Robert E. Kirschman, Jr., and Joyce R. Branda of the Department of Justice and Of Counsel, Warren M. Rose of the Air Force Legal Operations Agency. Heather A. James, Whiteford, Taylor and Preston, LLP, Washington, D.C., for defendant-intervenor. OPINION AND ORDER KAPLAN, Judge. In this post-award bid protest, Plaintiff Electronic On-Ramp, Inc. (“EOR”) challenges the Air Force’s award of a contract for linguist and analyst support services to defendant-intervenor Premier Management Corporation (“Premier”). Currently before the Court are EOR’s motion for judgment on the administrative record and the government’s cross-motion for judgment on * This Opinion was originally issued under seal, and the parties were given the opportunity to propose redactions by March 27, 2015. Because the Court did not receive any such proposals, the Court reissues its decision without redactions. the administrative record. For the reasons set forth below, EOR’s motion is DENIED, and the government’s cross-motion is GRANTED. BACKGROUND1 I. The Solicitation In May 2013, the Air Force Intelligence, Surveillance, and Reconnaissance Agency (“the agency” or “the government”) issued a solicitation seeking “linguist and analyst support services,” to be performed during a base and two option years pursuant to task orders issued under an indefinite delivery, indefinite quantity contract. Administrative Record (“AR”) 10:211, 243.2 The solicitation sought offers only from disadvantaged small businesses, pursuant to the Small Business Administration’s 8(a) Business Development Program. AR 10:141. The contract would be awarded to the offeror whose proposal represented the best value to the government. AR 10:175. A. Services to be Provided A “performance work statement,” or PWS, included in the solicitation described the tasks that the personnel of the successful offeror would be obligated to perform. These tasks included: translating spoken and written communications into English from various other languages, and vice versa, sometimes in “near real-time,” AR 10:243-45; analyzing translated communications for potentially significant intelligence, which would be reported to agency personnel and used to inform tactical and strategic military decisions, AR 10:245-46; and developing and conducting training programs to teach agency personnel to translate and analyze specified languages. AR 10:246-48. Work under the contract would be performed at agency facilities in a variety of locations, including Fort Meade, Maryland; Royal Air Force (RAF) Mildenhall, United Kingdom; JB Kunia, Hawaii; and Kadena Air Base, Japan, AR 10:243, and would be so timesensitive that the contractor would be required to retain 85 percent of the necessary personnel as of contract award and 100 percent as of 90 days after award. AR 10:178. The solicitation instructed each potential offeror to submit a proposal consisting of five volumes, including a technical volume (Volume II), a past and present performance volume (Volume III), and a price/cost volume (Volume IV). AR 10:166. In the technical volume, offerors were to address two major subject areas: their technical approach and their management approach. AR 10:167-68. 1 The background constitutes findings of fact made by the Court from the administrative record of the procurement filed pursuant to RCFC 52.1(a). See Bannum, Inc. v. United States, 404 F.3d 1346, 1356 (Fed. Cir. 2005) (observing that bid protests “provide for trial on a paper record, allowing fact-finding by the trial court”). 2 The administrative record is paginated sequentially and also divided into tabs. In citing to the administrative record, the Court will designate the tab, followed by the page number. For example, AR 10:211 refers to page 211, which is located in Tab 10. 2 Most pertinent to this case, in connection with the technical approach factor, the agency requested that offerors provide “five (5) samples of current complete curriculum” for five of the foreign-language instruction courses conducted by each offeror, “to include lesson plans, quizzes/progress checks and tests for courses.” AR 10:168. After the issuance of the solicitation, the agency received the following question: “Would a comprehensive course syllabus that covers daily lesson plans for the entire course duration, as well as quizzes, test[s], and student progress be sufficient to meet the sample requirements?” AR 11:317. The agency clarified that such a syllabus would not satisfy the solicitation’s requirements because the term “‘curriculum’ refers to the entire course” conducted by the offeror. Id. After describing the information that offerors were to provide regarding their technical approach, the solicitation listed a variety of topics related to management approach that offerors should also discuss in their technical volume. AR 10:168. The agency identified and asked the offeror to identify transition and execution risks that it considered likely to arise during contract performance and asked how the offeror would mitigate or address each one. See id. In addition, each offeror was asked to “[d]escribe your approach for managing personnel resources across multiple projects at various geographical locations” and for “managing multiple task orders at various geographical locations.” AR 10:168. The agency also requested “a Staffing Plan that addresses your acquisition and retention methods for skilled personnel, including retention during times of limited taskings,” as well as “a Training Plan for skill personnel” that detailed the contractor’s “process for introducing new personnel to the project” and its method for “ensur[ing] that the required training will be accomplished and managed.” Id. B. Evaluation Criteria The solicitation explained the manner in which the agency would evaluate information offerors provided in their proposals. AR 10:174-81. The agency would evaluate proposals using three factors listed here in descending order of importance: technical, past and present performance, and cost/price. AR 10:175. The technical factor would encompass two subfactors, technical approach and management approach, of which the former would carry more weight. AR 10:175-76. In combination, the technical and past and present performance factors would be “significantly more important than cost or price,” so that award might be made “to a higher rated, higher priced offeror, where the decision is consistent with the evaluation factors, and the Source Selection Authority (SSA) reasonably determines that the technical and/or overall business approach and/or past and present performance of the higher price offeror outweighs the cost difference.” AR 10:175-76. In evaluating the technical aspects of offerors’ proposals, the agency would conduct “two distinct but related assessments,” each having equal weight, with respect to both the technical approach subfactor and the management approach subfactor: “the [t]echnical [r]ating and the [t]echnical [r]isk [r]ating.” AR 10:176. The technical rating, expressed by an adjective paired with a corresponding color, would capture the agency’s determinations regarding the extent to which a proposal met the Government’s requirements. Id. The highest possible technical rating of blue/outstanding would be awarded to any proposal that “[met] requirements and indicate[d] an exceptional [a]pproach and understanding of the requirements”; such a proposal would “contain multiple strengths and no deficiencies.” Id. The lowest rating of red/unacceptable 3 would be assigned to any proposal that did “not meet requirements[,] . . . contain[ed] one or more deficiencies,” and was, accordingly, “unawardable.” Id. Along with the technical rating, the agency would assign each proposal a technical risk rating, determined based upon “the identification of weakness(es)” that, in the agency’s estimation, increased the potential “for disruption of schedule, increased costs, . . . degradation of performance, the need for Government oversight, or the likelihood of unsuccessful contract performance.” AR 10:176. Possible risk ratings ranged from low, indicating “little potential to cause disruption of schedule, increased cost or degradation of performance,” to high, indicating a “likel[ihood] to cause significant disruption of schedule, increased cost, or degradation of performance” that was “unlikely” to be “overcome . . . , even with special contractor emphasis and close Government monitoring.” AR 10:176-77. II. Submission and Initial Evaluation of Proposals The agency received proposals from nine offerors, including EOR and the ultimate awardee, Premier. AR 80:9814-15. EOR offered a lower price than any other offeror. Id. In its initial evaluation of EOR’s technical approach, the agency’s Source Selection Evaluation Board (SSEB) assigned EOR a red/unacceptable rating because EOR “did not sufficiently demonstrate understanding/knowledge in some areas and it was deficient in providing current complete course samples.” AR 33:7192. Specifically, EOR’s proposal was deficient because it did not provide “samples of quizzes/progress checks or tests” for its five foreign language course curricula. AR 33:7193. The SSEB also determined that the proposal included several weaknesses not rising to the level of deficiencies: EOR “did not show a sufficient understanding of the roles of the assigned [contractor] personnel within the DoD Intelligence Community . . . and its mission of providing near to real time intelligence to supported units,” nor did EOR “show sufficient knowledge and experience” in “[p]rocessing Signals Intelligence (SIGINT),” “[p]roducing written documentation such as transcripts and reports,” or “implement[ing] . . . training req[uirements] as related to DCGS, DSO, COMPASS CALL, Airborne/Ground Linguists.” Id. Finally, the SSEB noted that EOR’s proposal “failed to comply with page limits,” and that the SSEB had “removed,” and so had not considered, “pages 61 and later” from EOR’s technical volume. Id. The SSEB identified one strength in EOR’s proposal, approving of EOR’s plan to use “a rapid prototype approach” to adapt its training courses to changing needs. AR 33:7191-93. Because of the incompleteness of EOR’s curricula and its “weaknesses in showing understanding/knowledge of items specified in” the solicitation, the SSEB assessed the risk associated with the offeror’s technical approach as moderate. AR 33:7193. As to its management approach, EOR received a technical rating of yellow/marginal and a technical risk rating of high. AR 33:7193-94. The SSEB noted “several weak areas” in EOR’s responses to prompts contained in the solicitation. AR 33:7193. First, EOR “did not sufficiently detail how [it] would deal with transition risks,” including the need to “meet Government staffing requirements for 85% staff at contract award and 100% staff at 90 days.” Id. Second, EOR did not “show how [it] can sufficiently manage personnel at multi-geographical locations 4 and assigned to multiple tasking orders. AR 33:7193-94. Finally, although EOR’s “management hierarchy satisfactorily m[et] Government requirements,” “it d[id] not include an acceptable staffing and training plan for their personnel.” AR 33:7194. Although the SSEB found no strengths or deficiencies in EOR’s management approach, it concluded that EOR’s “significant weaknesses . . . increase[d] the risk of unsuccessful contract performance.” Id. In contrast to EOR, Premier received a technical rating of blue/outstanding and technical risk rating of low risk for both technical and management approaches. AR 36:7407. Both the technical approach and the management approach proposed by Premier offered several strengths, while revealing no weaknesses. AR 34:7197-200. The SSEB assessed Premier’s risk of nonperformance as low. Id. Several other offerors also received technical ratings that, while not as high as Premier’s, far exceeded EOR’s. AR 36:7407. On the past and present performance factor, EOR received a rating of limited confidence. AR 31:7182. EOR identified in its performance volume five contracts performed by either itself or one of its six proposed subcontractors. AR 18:1315-31. The agency’s past performance evaluation team evaluated only two of those contracts, however, because EOR’s past performance volume did not comply with the established page limitations, running fourteen pages over the ten-page limit. AR 31:7171. See AR 18:1310-1333. Once the excess pages were removed, the remaining pages described only two contracts, neither of which was considered relevant by the evaluators. AR 31:7169. Performance reviews obtained by EOR and by the agency provided little to no information regarding the extent to which EOR had in the past successfully performed activities similar to those required by the solicitation. AR 31:7181. The evaluators noted further concerns. EOR’s past performance volume “was of extremely poor quality” and so “rife with typographical errors, incorrect word usage, fragmented or incomplete sentence[s]/thoughts[,] and contradictory statements” that evaluators had difficulty understanding “what was being communicated.” AR 31:7181. The SSEB also noted that the source selection team “was concerned that EOR’s adversarial attitude (noted in the past performance binder and CPARS 1) towards current and previous teammates (including the government)” could adversely impact this contract. AR 31:7181. After evaluation teams had rated the technical and past performance volumes of all offerors, the source selection authority (SSA) reviewed those ratings and weighed them against each other, as well as against each offeror’s price. While the SSA recognized that EOR offered the lowest price, based upon that weighing process, the SSA ranked Premier’s proposal first, with ratings on technical and past performance factors as follows: 1. Premier Management (Premier) Factor 1: Technical Subfactor 1, Technical Approach: Subfactor 2, Management Approach Factor 2, Past Performance Technical Rating Risk Rating Blue/Outstanding Low Risk Blue/Outstanding Low Risk Satisfactory Confidence AR 36:7407. EOR was ranked seventh out of nine offerors, with ratings on technical and past performance factors as follows: 5 7. Electronic On-Ramp (EOR) Factor 1: Technical Subfactor 1, Technical Approach: Technical Rating Red/Unacceptable Subfactor 2, Management Approach Factor 2, Past Performance Yellow/Marginal Limited Confidence Risk Rating Moderate Risk High Risk AR 36:7408. III. Award, Protests, and Corrective Action by Agency Based upon the SSA’s determination that Premier’s proposal represented the best value to the government, the agency awarded the contract to Premier on September 20, 2013. AR 40:7588. Shortly thereafter, two unsuccessful offerors lodged protests with the Small Business Administration (SBA), asserting that Premier did not properly qualify as a business entitled to participate in the 8(a) program. AR 43:7735-37; 45:7740-57. The SBA rejected these protests. AR 61:8547, 8549. While the SBA’s decision was pending, EOR filed a protest before the Government Accountability Office (GAO). AR 51:8203-54. EOR argued, among other things, that (1) in evaluating EOR’s sample curricula, the agency failed to notice that sample tests and quizzes had been submitted for one curriculum, a course in the Tamasheq language, AR 51:8208; (2) the agency had erroneously excluded from consideration several pages of EOR’s technical proposal that had fallen within the page limit and treated the presence of the pages as a weakness in its proposal, AR 51:8214-15; and (3) the two contracts considered by the agency during the past and present performance evaluation should have been deemed relevant. AR 51:8248-49. EOR subsequently withdrew its protest after the agency issued a notice announcing that it would take voluntary corrective action. AR 53:8256. In its notice, the agency stated that it “believes its evaluation of past performance submissions did not comply with the evaluation scheme stated in the solicitation,” and that “there were instances in the technical evaluations where documents or pages within the proposals were not sufficiently evaluated.” AR 52:8255. As a result, the agency would re-evaluate offerors’ past performance and technical volumes. Id. In accordance with its corrective action notice, the agency conducted re-evaluations of offerors’ past and present performance submissions, see AR, Tabs 55, 56, 64, 65, and of EOR’s technical volume. See AR Tab 54. Upon completion of its re-evaluation of EOR’s past performance information the agency reaffirmed EOR’s “limited confidence” rating. AR 64:8593-94. In re-evaluating EOR’s technical information, as pertinent here, the agency re-examined EOR’s sample curricula and noted that the Tamasheq curriculum does, in fact, contain sample quizzes. AR 54:8259. Nevertheless, EOR’s technical volume remained “deficient in this area by not providing samples of quizzes/progress checks or tests for [the other four course samples].” Id. Furthermore, the SSEB noted that the quizzes included in the Tamasheq course consisted of two quizzes with “10 questions each,” a third quiz involving only “translat[ion] [of] a song,” and “a blank progress check sample” with “no actual questions.” Id. The agency considered these 6 sample assessments “very weak.” Overall, EOR’s technical approach ratings remained unchanged. After completing this initial re-evaluation of EOR’s technical proposal, the SSEB conducted a second re-evaluation, considering the arguments advanced by EOR in its protest before the GAO. AR 77:9331-35. The SSEB determined that none of those arguments justified a change in EOR’s technical ratings. AR 77:9335. It also deleted its reference to the weakness of EOR’s Tamasheq quizzes from its explanation of the deficiency of EOR’s proposal, making clear that the absence of tests and quizzes from the other four curricula, not the weakness of the Tamasheq assessments, was responsible for the finding of deficiency. AR 77:9336. After conducting re-evaluations to ensure that all proposals had been submitted precisely as the solicitation dictated, the SSA’s ultimate ranking of offerors remained the same. AR 80:9814-15. IV. Award and Second GAO Protest The agency awarded the contract to Premier on June 17, 2014. AR 81:9842-43. On June 30, 2014, EOR, despite being ranked seventh out of nine offerors, again filed a protest at the GAO. AR 1:1-50. After EOR failed to timely submit its comments on the agency’s report, the GAO dismissed the protest on August 14, 2014. AR 89:9986-87. EOR filed this protest here on October 24, 2014 seeking to permanently enjoin the agency from proceeding with performance of the contract. EOR contends that the evaluation of its proposal as well as Premier’s was unreasonable, arbitrary, and unsupported by the written record. It challenges virtually every rating that the agency assigned to its proposal. Moreover, it contends that the agency failed to follow through on its promised corrective action upon which EOR agreed to withdraw its prior protest. The Court held oral arguments on the parties’ motion on February 10, 2015. For the reasons set forth below, the Court concludes that the government is entitled to judgment on the administrative record because the agency’s decision to rate EOR as red/unacceptable on its technical proposal was neither arbitrary, nor capricious, nor contrary to law. In light of that ruling, EOR was ineligible for an award under the terms of the solicitation, see AR 10:176, and it is unnecessary for the Court to address EOR’s other challenges to the agency’s award decision. DISCUSSION I. Jurisdiction The Court of Federal Claims has “jurisdiction to render judgment on an action by an interested party objecting to . . . a proposed award or the award of a contract or any alleged violation of statute or regulation in connection with a procurement or a proposed procurement.” 28 U.S.C. § 1491(b)(1) (2012). A party is an “interested party” with standing to bring suit under 28 U.S.C. § 1491(b)(1) if the party “is an actual or prospective bidder whose direct economic interest would be affected by the award of the contract.” Orion Tech., Inc. v. United States, 704 7 F.3d 1344, 1348 (Fed. Cir. 2013). An offeror has a direct economic interest if it suffered a competitive injury or prejudice. Myers Investigative & Sec. Servs., Inc. v. United States, 275 F.3d 1366, 1370 (Fed. Cir. 2002) (holding that “prejudice (or injury) is a necessary element of standing”). In a post-award bid protest, the protestor has suffered prejudice if it would have had a “substantial chance” of winning the award “but for the alleged error in the procurement process.” Info. Tech. & Applications Corp. v. United States, 316 F.3d 1312, 1319 (Fed. Cir. 2003). See also Weeks Marine, Inc. v. United States, 575 F.3d 1352, 1359 (Fed. Cir. 2009); Rex Serv. Corp. v. United States, 448 F.3d 1305, 1308 (Fed. Cir. 2006). “In other words, the protestor’s chance of securing the award must not have been insubstantial.” Info. Tech., 316 F.3d at 1319. Neither the government nor Premier disputes that EOR is an “interested party” within the meaning of the statute. EOR was an actual offeror for the contract in question. Although EOR was ranked seventh out of nine offerors, it proposed the lowest price and received slightly higher technical ratings than the other unsuccessful offerors. If all of EOR’s allegations were found meritorious, including its challenge to the assessment of a deficiency in its technical proposal, then “all of the agency’s ratings would need to be redone, and a new best value determination made.” Preferred Sys. Solutions v. United States, 110 Fed. Cl. 48, 57 (2013). Because EOR would have had a not insubstantial chance of securing the award upon a ruling that the agency’s technical evaluation of its proposal was erroneous or that award to Premier was invalid, EOR has established its standing as an interested party. II. Standard for Granting Judgment on the Administrative Record Pursuant to RCFC 52.1, the Court reviews an agency’s procurement decision based on the administrative record. Bannum, Inc., 404 F.3d at 1354. The Court makes “factual findings under RCFC [52.1] from the record evidence as if it were conducting a trial on the record.” Id. at 1357. Thus, “resolution of a motion respecting the administrative record is akin to an expedited trial on the paper record, and the Court must make fact findings where necessary.” Baird v. United States, 77 Fed. Cl. 114, 116 (2007). The court’s inquiry is “whether, given all the disputed and undisputed facts, a party has met its burden of proof based on the evidence in the record.” A&D Fire Prot., Inc. v. United States, 72 Fed. Cl. 126, 131 (2006). Unlike a summary judgment proceeding, genuine issues of material fact will not foreclose judgment on the administrative record. Bannum, Inc., 404 F.3d at 1356. III. Standard of Review in Bid Protest Cases The court reviews challenges to a contract award under the same standards used to evaluate agency action under the Administrative Procedure Act (“APA”), 5 U.S.C. § 706 (2012). See 28 U.S.C. § 1491(b)(4) (stating that “[i]n any action under this subsection, the courts shall review the agency’s decision pursuant to the standards set forth in section 706 of title 5”). To successfully challenge an agency’s procurement decision, a plaintiff must show that the agency’s decision was “arbitrary, capricious, an abuse of discretion, or otherwise not in accordance with law.” 5 U.S.C. § 706(2)(A); Bannum, Inc., 404 F.3d at 1351. “The arbitrary and capricious standard applicable here is highly deferential. This standard requires a reviewing court to sustain 8 an agency action evincing rational reasoning and consideration of relevant factors.” Advanced Data Concepts, Inc. v. United States, 216 F.3d 1054, 1058 (Fed. Cir. 2000) (citing Bowman Transp., Inc. v. Arkansas-Best Freight Sys., Inc., 419 U.S. 281, 285 (1974)). In a bid protest, the disappointed offeror “bears a heavy burden” in attempting to show that a procuring agency’s decision lacked a rational basis. Impresa Construzioni Geom. Domenico Garufi v. United States, 238 F.3d 1324, 1338 (Fed. Cir. 2001). Indeed, such a challenge can succeed only where the agency “entirely failed to consider an important aspect of the problem, offered an explanation for its decision that runs counter to the evidence before the agency, or [the decision] is so implausible that it could not be ascribed to a difference in view or the product of agency expertise.” Ala. Aircraft Indus., Inc.–Birmingham v. United States, 586 F.3d 1372, 1375 (Fed. Cir. 2009) (alteration in original) (quoting Motor Vehicle Mfrs. Ass’n v. State Farm Mut. Auto Ins. Co. (“State Farm”), 463 U.S. 29, 43 (1983)). The Court is particularly deferential to an agency’s technical evaluation. E.W. Bliss Co. v. United States, 77 F.3d 445, 449 (Fed. Cir. 1996) (stating that “such matters as technical ratings” involve “the minutiae of the procurement process . . . that a court will not second guess”). “This is because the evaluation of proposals for their technical excellence or quality is a process that often requires the special expertise of procurement officials, and thus reviewing courts give the greatest deference possible to these determinations.” One Largo Metro, LLC v. United States, 109 Fed. Cl. 39, 74 (2013) (alteration and internal quotations marks omitted) (quoting Beta Analytics Int’l, Inc. v. United States, 67 Fed. Cl. 384, 395 (2009)). Given this highly deferential standard of review, the court’s function is limited to “determin[ing] whether ‘the contracting agency provided a coherent and reasonable explanation of its exercise of discretion.’” Impresa, 238 F.3d at 1332-33 (quoting Saratoga Dev. Corp. v. United States, 21 F.3d 445, 456 (D.C. Cir. 1994)). The agency need only articulate a “rational connection between the facts found and the choice made,” and the court will “uphold a decision of less than ideal clarity if the agency’s path may reasonably be discerned.” State Farm, 463 U.S. at 43. IV. The Government Is Entitled to Judgment on the Administrative Record As described above, the agency found EOR’s technical proposal deficient3 because the curricula that EOR submitted for four of five sample courses were incomplete. Specifically, the agency concluded, EOR did not include samples of the quizzes/progress checks or tests that EOR stated would be administered as part of those courses. EOR claims that this finding was incorrect and contrary to the record. The Court disagrees. As described above, the solicitation required offerors to provide “five (5) samples of current complete curriculum to include lesson plans, quizzes/progress checks and tests for courses.” AR 10:168. After the issuance of the solicitation, the agency clarified that 3 A “deficiency” is a “material failure of a proposal to meet a Government requirement or a combination of significant weaknesses in a proposal that increases the risk of unsuccessful contract performance to an unacceptable level.” FAR 15.001. 9 “‘curriculum’ refers to the entire course” conducted by the offeror, and that “a comprehensive course syllabus that covers daily lesson plans, . . . quizzes, test[s], and student progress” would not satisfy the solicitation’s requirement. AR 11:317. The parties agree that the solicitation required offerors to include any quizzes or tests actually used in its curricula. Furthermore, the solicitation stated that the “curriculum samples [would] be considered as an indicator of technical capability.” AR 10:168. According to the agency, EOR failed to meet a material requirement of the solicitation because it did “not provid[e] samples of quizzes/progress checks or tests for . . . [its] Arabic Refresher, Bambara Advanced, French Basic, and French Intermediate” courses. AR 54:8259. This conclusion was eminently reasonable and clearly supported by the administrative record. EOR stated in each of its curricula that tests and quizzes would be administered as part of each class. For example, for its Arabic Refresher course for advanced students, EOR stated that the course included grammar quizzes, unit tests, weekly exams, and writing assignments. See AR 16:528, 530 (noting that as part of the course objectives, students were required “to score at least 70% on all assessment modalities: unit tests, weekly exams and homework”; “score  at least 70% in the 10 grammar quizzes”; and “score at least 70% on your writing assignment”). EOR contends that “‘Quiz[zes]’ meaning Examination[s]” are included in the curriculum as well as progress reports, a half term examination, a mid-term examination, and a final examination. Pl.’s MJAR 8. The pages identified by EOR as containing quizzes or tests, however, are virtually blank, containing only a single line at the top right hand corner of the page that includes the words “Test/Exam” in Arabic and specifies the number of minutes allowed. AR 16:553, 573, 583, 596, 618, 640, 663, 678, 694, 706, 734, 746, 759, 770. See also LeBlanc Decl. ¶¶ 4-31, Def.’s Unopposed Mot. Supplement AR Ex. B, Dec. 5, 2014, ECF No. 23. These blank pages are plainly not sufficient to meet the specifications of the solicitation that required EOR to actually include the contents of any tests or quizzes for each course in which it intended to administer such quizzes and tests.4 Similarly, in a section of its Bambara Advanced curriculum titled “Testing Schedule Formula,” EOR stated that “[a] quiz will be given every day to test the mastery of previous lessons” and “an exam will be given in the middle and at the end of the semester.” AR 16:783. It further stated that the midterm and final exam would be administered in class by the professor. Id.5 4 On January 27, 2015, the defendant notified the Court that the blank pages in EOR’s Arabic curriculum labeled “Test/Exam” actually contained hidden text that for whatever reason is not apparently visible on the electronic or hard copies of EOR’s proposal. Def.’s Notice 1-2, ECF No. 38. This technical problem may explain but does not excuse EOR’s non-compliance with the solicitation’s requirements. The agency cannot reasonably have been expected to guess that an apparently blank page in EOR’s proposal contained hidden text in the electronic copy which was visible only when the page was highlighted. 5 The syllabus for the Bambara Advanced course also recommended that instructors distribute periodic progress reports each week. The periodic progress report is a form that is to be filled out by the course instructor for each individual student in which the instructor assesses the student’s skills in areas such as simple short conversations, reporting facts about current events, 10 The end of each lesson plan for the Bambara course includes “Exercises,” “TDA,” and “Self-Evaluation,” all of which ask students to recall and use the skills and knowledge gained from the day’s lesson. See AR 16:815-16, 824, 828, 835-37, 841-43, 850-52, 862-63, 867-70, 876-77, 885, 889-90, 894, 898, 902, 908-09, 915, 923. Students are then asked to evaluate themselves on specific skills like, “I can speak about daily activities: YES____ NOT YET____.” EOR contends that the Exercises at the end of each lesson constitute the quizzes that it planned to “give every day to test the mastery of previous lessons” in accordance with the testing schedule and formula. Pl.’s MJAR 1. Moreover, EOR contends, the exercises and progress checks taken together are within the meaning of “quiz/progress check” as used in the solicitation since they “serve the function of testing and evaluation of the student on a daily basis.” Pl.’s Resp. & Reply 1. These contentions are unavailing. First, as noted, EOR’s syllabus for the Bambara Advanced course states that instructors will administer a midterm and final exam to students. EOR does not provide any citation to material in its proposal that the agency should reasonably have construed as a midterm or final exam and the Court has not found any. Further, at the oral argument in this case, counsel for EOR acknowledged that the final and midterm exams for this course (and for the other courses) were “just not there.” See Oral Arg., 11:39:38-11:39:41. It was also not unreasonable for the agency to conclude that EOR’s proposal failed to provide the copies of the daily quizzes it discussed in its syllabus. The daily “exercises” which EOR now contends the agency should have recognized as “quizzes” were not identified as such in EOR’s materials. And, as the record shows, when EOR wants to identify a series of questions as a “quiz,” it knows how to do so—as it did in the Arabic course and in the syllabus for the Bambara course. See, e.g., AR 16:530. Moreover, the agency’s conclusion that the exercises were not quizzes or tests comports with the plain and ordinary meaning of the word “exercise” as used in the educational context, which is “a particular task or set of tasks devised as exercise, practice in a technique, etc.” or a mental activity “esp. as practice to develop a faculty.” Oxford English Reference Dictionary 490 (rev. 2d ed. 2002). By contrast, the use of the word “quiz” or “examination” in this context connotes not a vehicle for a student to practice his skills, but rather a test or assessment which will be scored to measure a student’s knowledge of the course material. EOR’s French Intermediate and French Beginner’s courses proposed the same testing requirements as the Bambara course. Thus, for each course, EOR stated that “a quiz will be given every day to test the mastery of previous lessons. An exam will be given in the middle and understanding speech about basic needs, or reading simple authentic material. AR 16:796-97. In addition, the instructor is to rate students on whether they are below average, average, above average or superior on their pronunciation, fluency, grammar control, and vocabulary retention. Part three of the progress report includes an optional place for the instructor to include an achievement test score, expressed as a percentage. Students were also encouraged to evaluate their own progress at the end of each week. AR 16:783. 11 at the end of the semester.” AR 17:963, 1045. Like the Bambara course, students were also to receive a weekly periodic assessment report assessing a student’s proficiency in the language. EOR contends that it met the requirements of the solicitation because its French courses include “Periodic Assessment Reports” as well as “‘Exercices in French’ meaning quizzes,” and tests. Pl.’s MJAR. 7-8. As with the Advanced Bambara course, however, the tests EOR stated it would administer at the middle and at the end of the semester are absent from the curricula for both Intermediate and Beginner French. The French curricula contain references to the administration of tests, and include placeholders for such tests, but copies of the tests are not included in the course samples. See AR 17:1098, 1030, 1103. Again similar to the Advanced Bambara course, none of the pages of the administrative record that EOR has cited with respect to its two French courses include materials that are identified as quizzes. Further, the word “exercices” in French means “exercises,” not “quizzes.” See Kendall Decl. ¶¶ 4-7. And, as explained above, the agency had ample basis to find that the “exercises” were not the “quizzes” that EOR referred to in its course design. As perhaps its final Hail Mary pass, EOR contends that “[t]o the extent that one or more of the curricula [are] lacking or ‘weak’ . . . there is no question that the source material for quizzes and daily exercises is present in the curricula provided.” Pl.’s Resp. & Reply, Dec. 16, 2014, ECF No. 31; see also Oral Arg., 11:37:06-11:37:35 (stating that EOR intended to create examinations by “using the exercises that are included and then compiling them as types of[,] as a type of midterm as a type of final”). But the solicitation specifically required offerors to include lesson plans, quizzes/progress checks, and tests that were going to be administered in the courses, not material that could be used to create these instruments. Further, to the extent that EOR intended to redeploy the instructional materials it submitted as quizzes/progress checks and tests, it failed to advise the agency of that intention. In short, this argument is unsupported by the record and is, at best, a post hoc rationalization of EOR’s failure to supply the quizzes and tests along with its proposal. Indeed, the deficiencies of EOR’s submission are highlighted by comparing EOR’s curricula to the submission of the contract awardee, Premier.6 For all of its courses, Premier included a short section near the beginning of each curriculum stating that it would use both formal and informal assessments. AR 23:1501; AR 24:2199-200; AR 25:3106-07; AR 26:5465; AR 27:6512. It then supplied sample copies of all of the assessment tools its curricula stated that it would use in the courses. See AR Tabs 23-27. Thus, for its Intermediate and Expedited Pashto courses, Premier provided copies of the formal assessments it intended to use, which consisted of tests and quizzes. For example, the 6 The Court requested supplemental briefs from the parties asking them to identify which parts of the administrative record reflect “quizzes/progress checks and tests” submitted by the contract awardee, Premier. Order, January 16, 2015, ECF No. 36. See Tech. Sys. v. United States, 98 Fed. Cl. 228, 253 (2011) (noting that, in some cases, the court may examine whether features of a proposal that are highly rated are similar to another proposal that is not similarly credited). 12 Intermediate Pashto course included weekly reading and listening tests, as well as “pre-” and “post-tests,” given at the beginning and [at] the end of the course respectively. AR 24:2199, 2322-27, 2475-81, 2546-50, 2687-92, 2246-71 (pre-test), 3079-97 (post-test). The expedited course, in addition to listening and reading tests, also incorporates daily quizzes, vocabulary quizzes, speaking tests, and mid-term tests, all of which were included in Premier’s proposal. AR 25:3106-07, 3178, 3205, 3249 (examples of daily quizzes), 3414-15, 3780-81, 3986-87 (examples of speaking tests), 3417-23, 3786-94, 3988-98 (examples of listening and reading tests); 4302-15, 4317-4340 (mid-term tests). Premier did not propose to use its own formal assessments for three of its curricula. For example, for its Persian-Farsi course, Premier stated that the formal assessment for the course was the DLPT.7 AR 27:6512. Premier explained that “[t]his course is designed as a refresher course. As such, it is highly encouraged that the students [take the] DLPT upon course completion, or as directed by their CLPM or their unit.” Premier provided similar responses for its Arabic Dialect and Persian-Dari refresher courses. See AR 23:1501; AR 26:5465. Premier stated that it would use informal assessments for these courses. These informal assessments would consist of self-check questions, optional homework assignments, and opportunities for students to present questions to and request clarification from the instructor at the end of each day. AR 23:1501; AR 26:5465; AR 27:6512. Premier then went on to include samples of these informal assessments in its curricula. See, e.g., AR 23: 1510, 1514-15, 151819, 1540-41, 1543; AR 26:5487-510, 5549, 5571-72, 5585-86; AR 27:6522-31, 6542-48, 656168, 6576-83, 6829-33. Plaintiff argues that—because Premier used informal rather than formal assessments for these courses—its submission was as deficient as EOR’s. See Pl.’s Reply to Def.’s Resp. to Ct. Order 1-2, Jan. 30, 2015, ECF No. 39. This contention is unpersuasive. The solicitation did not require the use of formal assessments; it simply required that if such assessments were to be used, then sample copies would be provided as part of the offeror’s proposal. Premier outlined the assessment strategy used in each of its courses, provided the materials it had created to implement those strategies, and clearly identified its assessment tools as such in its submission.8 Finally, because the agency’s assignment of a deficiency had a rational basis, the agency acted reasonably in rating EOR as red/unacceptable on its technical proposal. Thus, pursuant to the terms of the solicitation, a proposal containing one or more deficiencies must be assigned a technical rating of red/unacceptable, which makes it ineligible for award. AR 10:176. For that reason, the Court need not address EOR’s other challenges to the ratings assigned to its proposal by the Air Force. See Glenn Defense Marine (Asia), PTE Ltd. v. United States, 720 F.3d 901, 912 (Fed. Cir. 2013) (noting that plaintiff failed to show prejudicial error where it could not 7 DLPT stands for Defense Language Proficiency Test. AR 10:257. 8 Premier did not include copies of DLPTs in its submission. It was not unreasonable for the agency not to assess any deficiency on that basis because the DLPT is created by the Defense Language Institute, used by the United States Department of Defense, and is administered by the government, not Premier. See AR 10:244. 13 show that even if the agency had erred in its past performance ratings, it would have had a substantial chance of winning the award). CONCLUSION As described above, EOR failed to meet a requirement of the solicitation which the agency considered material, making its proposal ineligible for contract award. Consequently, EOR has failed to establish the merit of its claims by showing any error, prejudicial or otherwise, in the procurement process. Accordingly, the government’s motion for judgment on the administrative record is GRANTED. The plaintiff’s motion for judgment on the administrative record is DENIED. Each party shall bear its own costs. IT IS SO ORDERED. s/ Elaine D. Kaplan ELAINE D. KAPLAN Judge 14