Case Information
*1
UNITED STATES DISTRICT COURT FOR THE DISTRICT OF COLUMBIA
FILED
SEP 062011
Clerk, U.S. District and Bankruptcy Courts
H&;R BLOCK, INC., et al.,
Defendants.
Civil Action No. 11-00948 (BAH)
SEALED
MEMORANDUM OPINION AND ORDER
The United States, through the Antitrust Division of the Department of Justice (the "DOJ" or the "plaintiff"), brought this civil case to enjoin the proposed acquisition of a digital do-it-yourself tax preparation company known as TaxACT by H&;R Block, another company that sells digital do-it-yourself tax preparation products as well as provides other tax preparation services. [1] A preliminary injunction hearing in this case is scheduled for September 6, 2011, and the plaintiff has filed a motion in limine to exclude evidence of an email survey commissioned by the defendants and the portions of a defendants' expert opinion that relies upon the survey. For the reasons that follow, the Court denies the motion in limine. [2]
I. BACKGROUND
The DOJ filed this action on May 23, 2011, seeking to enjoin Defendant H&;R Block, Inc. from acquiring Defendant 2SS Holdings, Inc. ("TaxACT"), which sells digital do-it-yourself
*2 tax preparation products marketed under the brand name TaxACT. Compl. ¶ 10. Defendant TA IX, L.P. ("TA"), owns a two-thirds interest in TaxACT. [3] Id. ¶ 11.
According to the Complaint, last year an estimated 35 to 40 million taxpayers filed their taxes using digital do-it-yourself tax preparation products ("Digitial DIY Tax Preparation Products"). Id. ¶ 1. In the U.S. Digital DIY Tax Preparation Product market, the three largest firms collectively have about of the market share. Id. The leading company in the market is Intuit, Inc., the maker of "TurboTax." Id. ¶ 3. H&;R Block's proposed acquisition of TaxACT, if allowed to proceed, would combine the second- and third-largest providers in the market - i.e., H&;R Block and TaxACT, respectively. Id.
The Complaint alleges that TaxACT is a "maverick" competitor that has a history of "disrupting" the Digital DIY Tax Preparation market and has forced its competitors, including H&;R Block and Intuit, "to offer free products and increase the quality of their products for American taxpayers." Id. ¶ 28. The Complaint alleges that TaxACT has aggressively competed with H&;R Block and Intuit by providing high-quality products and services at low cost. See id. ¶¶ 30-40. The DOJ alleges that the acquisition of TaxACT by H&;R Block would reduce competition in the industry and make anticompetitive coordination between the two major remaining market participants - H&;R Block and Intuit - substantially more likely. Id. ¶¶ 40-49. The DOJ alleges that therefore the proposed acquisition violates Section 7 of the Clayton Act, 15 U.S.C. § 18, and accordingly it seeks an injunction blocking H&;R Block from acquiring TaxACT. Id. ¶¶ 53-55.
*3
The defendants dispute that the appropriate product market is Digital DIY Tax Preparation products, but argue that the relevant market instead consists of "all methods of tax preparation for the U.S. federal and state income taxes." Report of Dr. Christine Siegwarth Meyer, DX0017-006, at 2. Furthermore, even assuming arguendo that the plaintiff's alleged market definition were correct, the defendants deny that the transaction would result in anticompetitive effects because, inter alia, "H&;R Block and TaxAct are not close substitutes and the merger is likely to lead to substantial, incremental, merger-specific efficiencies." Joint PreHearing Statement at 4.
A pre-hearing conference in this matter, including oral argument on the motion in limine, was held on September 2, 2011.
II. DISCUSSION
The plaintiff has moved, pursuant to Federal Rules of Evidence 702 and 703, to exclude evidence of an email survey of defendants' customers and to limit defendants' expert opinion to the extent that it relies on this survey (the "2011 email survey"). Pl.'s Mem. in Supp. of Pl.'s Mot. to Exclude the 2011 Litigation Survey and Limit Defs.' Expert Report ("Pl.'s Mem.") at 12. The plaintiff contends the survey's "methodology falls far short of the requirements of Federal Rules of Evidence 703 and 702 because: (1) it fails to ask a question relevant to this proceeding; (2) it suffers from extraordinary non-response bias, with response rates far below what courts have found necessary to establish reliability; and (3) the response options provided are leading and fail to discourage guessing." Id. at 2. For the reasons below, the Court declines to exclude the evidence or to limit the expert's opinion.
*4
A. Standard of Review
"Under Rule 702, a trial court may only admit expert testimony that is both relevant and reliable." Harris v. Koenig, No. 02-618,
Under Rule 703, the facts or data underlying an expert's opinion "need not be admissible in evidence in order for the opinion or inference to be admitted" if the facts or data are "of a type reasonably relied upon by experts in the particular field." Fed. R. Evid. 703. In addition, such "facts or data that are otherwise inadmissible" may be disclosed by the proponent of the opinion if their probative value substantially outweighs any prejudicial effect. Id.
*5
B. Background Regarding the 2011 Email Survey
In late April 2011, following the 2011 tax season, the defendants commissioned Directions Research, a market research firm, to conduct an Internet-based survey of TaxACT's customers. GX 604. According to the defendants, during the investigation of the transaction, the DOJ did not accept the defendants' evidence of an "extremely low switching between the Defendants' products" as an appropriate proxy for diversion - i.e., the government contended that current switching rates are not necessarily predictive of how TaxACT customers would react to a price increase or functionality decrease. Defs.' Mem. in Opp'n to PI.'s Mot. In Limine ("Defs.' Mem.") at 3. The survey was therefore initiated to respond to a question raised by the DOJ - namely, "how TaxAct consumers would react to a price increase, service decrease or functionality decrease in the TaxAct products." Id. The survey asked one primary question: "If you had become dissatisfied with TaxACT's price, functionality or quality, which of these products or services would you have considered using to prepare your federal taxes?" GX 604 (Results of the 2011 email survey); GX623 (App'x 2 to Report of Dr. Dhar). The survey then offered the respondents a list of eleven options, including other products or services, from which to choose and instructed them to select all applicable options. Id. The list of options that respondents were given varied somewhat depending on the respondents' filing status and the payments they had made for their 2011 tax returns. [4] Id. A follow-up question asked the respondents to narrow their selections to a single choice. Id.
*6 The research firm sent over 70,000 surveys via email to a statistically-random selection of TaxACT customers inviting them to participate in the survey. Declaration of Tina Ruddy, dated August 24, 2011 ("Ruddy Decl.") 9, 10. Survey respondents were asked screening questions to determine their membership in one of four categories of customers: (1) those who paid to use TaxAct's products for filing both federal and state tax returns (denominated in the survey report analysis by Directions Research as "Paid Fed/Paid State"); (2) those who paid to use TaxAct's federal return product but not for filing the state return ("Paid Fed/No State"); (3) those who used TaxAct's free product for filing a federal return and paid to use a TaxAct product to file a state return ("Free Fed/Paid State"); and (4) those who used TaxAct's free product for filing a federal return but not for filing the state return ("Free Fed/No State"). [5] GX604 (Results of the 2011 email survey).
A total of 1,089 customers responded to the survey. Id. at 2-3. The response rates for the four categories of customers were: (1) for paid federal / paid state filing (422); (2) for paid federal / no state filing (182); (245); (3) for free federal / paid state filing; and (4) for free federal / no state filing (240). Id.; see also Pl.'s Mem. at 3.
Defendants' expert, Dr. Christine Siegwarth Meyer, summarized the results of the survey as follows: [A] survey of TaxACT customers indicates that few would switch to H&;R Block At Home in the event that they were satisfied with TaxACT. In each of the four groups, the comparable HRB product was neither the first nor second most likely alternative tax products for the respondents. The percentage of TaxACT customers that would switch to H&;R Block At Home ranged from 4 to 10 percent, with a weighted average of only 6 percent. Instead, in each group, pen-and-paper and TurboTax were the two options with the
*7
highest responses. In only one of the four groups was HRB the third response. Instead, Free TaxUSA (Tax Hawk) was a more prevalent choice than HRB for three of the four groups.
Report of Dr. Meyer, DX0017-042 (footnotes omitted). Based upon this analysis of the results, Dr. Meyer opined that:
This further indicates that HRB is not a particularly close competitor to TaxACT. Following the merger, consumers who are dissatisfied with TaxACT will have numerous other choices to which they can and would turn, including TurboTax, pen-and-paper, other software products, and assisted tax preparation. Id.
C. Analysis
In order to admit expert testimony under Rule 702, the Court must find that it is "relevant and reliable." Kumho Tire Co.,
Facts or data underlying an expert's opinion that are "of a type reasonably relied upon by experts in the particular field" need not be admissible in evidence in order for the opinion or inference to be admitted. Fed. R. Evid. 703. Regarding survey results in particular, technical and methodological deficiencies in a survey generally go to the weight of the evidence, not the admissibility, unless the deficiencies are so substantial as to render the survey unreliable. See Univ. of Kan. v. Sinks, No. 06-2341,
*8 As discussed in more detail below, having reviewed the report of defendants' expert, Dr. Christine Siegwarth Meyer, the Court finds that Dr. Meyer's anticipated testimony regarding the 2011 email survey meets the criteria for admissibility. Dr. Meyer is an accomplished economist with a Ph.D. in economics from the Massachusetts Institute of Technology. See DX0017, Ex. 1. Her conclusions regarding the level of competition between TaxACT and H&;R Block, as expressed in her report, are not based solely upon the results of the survey to which the plaintiff objects. See DX0017-022-26, 28-42 (discussing various documents and data to support Dr. Meyer's conclusions regarding competition with manual filing and unilateral effects); Meyer Dep. 175:6-16 ("I think [the 2011 email survey] - it's an important data point. It's not the only data point to make any of the points that I use it to make.") (cited in Defs.' Mem. at 18 n.58).
As for the survey itself, the Court finds that Dr. Meyer's use of the survey as a datum in reaching her conclusions is not unreliable. While the plaintiff has identified cogent concerns about the wording and the methodology of the survey, the Court finds that these concerns go to the weight, not the admissibility, of the evidence, especially in this bench hearing where there is no concern about jury confusion or prejudice.
[6]
See Seaboard Lumber Co. v. United States,
*9
1. Relevance
The Court finds that the survey is relevant because it is probative of the degree to which TaxACT and H&;R Block are competitors, whether the market is defined, as alleged by the plaintiff, to be Digital DIY tax preparation products or more broadly, as alleged by the defendants, to be all tax preparation products and services. The options provided to survey respondents encompassed both competing Digital DIY tax preparation products as well as alternative services.
The plaintiff argues that the survey is irrelevant because Dr. Meyer asserted in her report that "this survey is closer to the concept of a diversion ratio than are data on overall switching between products," PI.'s Mem. at 5 (citing Dr. Meyer's Report at DX0017-024 n.85). According to the plaintiff, diversion refers to "a measured customer reaction to a measured increase in price." Id. Since the phrasing of the survey question conflates customer concerns about price, functionality, and quality, and does not actually ask about a change in price, the plaintiff argues that the survey cannot shed any light on customer reactions to price changes. Id. Even accepting the plaintiff's critique, however, would not obliterate the survey's relevance entirely. Further, Dr. Meyer's report, in the portion cited by the plaintiff, did not state that the survey provided direct evidence of diversion, but rather that "this survey is closer to the concept of a diversion ratio than are data on overall switching between products," which the Court understands as expressing a different idea. While the wording of the survey question may limit its relevance on specific issues, because the survey provides at least some indication of the products and services that compete with TaxACT, the Court cannot say that it does not have "any tendency to make the existence of any fact that is of consequence to the determination of the action more probable or
*10
less probable than it would be without the evidence." Fed. R. Evid. 401 (defining relevant evidence).
2. Reliability
Customer surveys generally are a type of evidence that may be reasonably relied upon by experts in defining markets in antitrust cases. See GX622 (Christine Meyer, "Designing and Using Surveys to Define Relevant Markets," in Economics of Antitrust: Complex Issues in a Dynamic Economy 101, 108 (Lawrence Wu, ed. 2007)); GX624 (Shari Seidman Diamond, "Reference Guide on Survey Research," in Federal Judicial Center, Reference Manual on Scientific Evidence, hereinafter "Diamond," at 234-35).
In determining whether a particular survey may be admissible under Rule 703, courts examine the validity of the survey's methods and ask "Was the poll or survey conducted in accordance with generally accepted survey principles, and were the results used in a statistically correct way?" Diamond at 233-34. The methodological validity of a survey is necessarily a question of degree. As discussed above, technical deficiencies that can be adequately explored on cross-examination generally go to the weight, rather than the admissibility, of the evidence, unless the methodological deficiencies are so sweeping or fundamental as to render the survey wholly unreliable and therefore inadmissible.
a. Response Rate
The plaintiff argues that the survey is not reliable because it suffers from an "extraordinary level of non-response bias" due to its low response rate. Pl.'s Mem. at 7. The plaintiff cites authority indicating that a response rate below "should be regarded with significant caution as a basis for precise quantitative statements from which the sample was drawn." Pl.'s Mem. at 7 (citing Diamond at 245). The plaintiff's rebuttal expert, Dr. Ravi Dhar,
*11 has concluded that the "level of nonresponse . . . is extremely high (more than 98\%)" and that the "extremely low response rates makes it difficult to determine whether the results were impacted by a certain segment who were systematically more likely to respond to the survey (e.g., those who were price sensitive or time insensitive) in relation to those who did not respond." GX 623 (Report of Dr. Dhar at 10).
The defendants respond by citing authority indicating that web-based surveys typically have significantly lower response rates than other types of surveys. Defs.' Mem. at 7-8. They also point to jury trials in which courts have admitted surveys with response rates of
or lower. Defs.' Mem. at 7-8 (citing Sinks,
The defendants further argue that the survey's sample size and response rate comply with industry standards for market research. A declaration from the Vice President of the market research firm that performed the survey affirms that large national companies commonly use similar survey results "to make business and pricing decisions." Ruddy Decl. 3-4. This declaration also states that "[t]he standard, industry-expected response rate for surveys of this kind is generally between " and that, for web-based email surveys, "a response rate above is so improbable as to be considered entirely unavailable." Ruddy Decl. .
While the Court agrees with the plaintiff that the survey's response rate "appears, by any standard, to be quite low," Sinks,
*12
testimony that the survey's sample size and response rate are in line with industry standards for similar surveys establish sufficient reliability to allow admission of the survey evidence. The Court is cognizant of the detailed critique of the survey's response rate presented by the plaintiff's expert, Dr. Dhar, and the Court is fully capable of taking this critique into account in determining how much weight to accord the survey's results in its analysis. See Sinks,
b. Closed-Ended Questions and Discouragement of Guessing
The plaintiff also contends that the survey is fatally flawed because it asks only "closedended, leading questions" and that it failed to discourage guessing by not including a "no opinion" or "I don't know" option.
The plaintiff contends that the survey's closed-ended response options are "severely flawed because they are not exhaustive and fail to take into account that some people may not switch [products] even though they are dissatisfied." Pl.'s Mem. at 9 (citing GX 623, Report of Dr. Dhar, 18-19). The plaintiff also argues that, by providing survey respondents with a list of options from which to choose, the survey "hardly mirrors competition in the marketplace where the Big Three competitors spend millions of dollars annually to get their message in front of potential customers. In contrast, the [survey] counterfactually de-emphasizes the significance of brand and the millions spent building and maintaining it." Id. at 10. In addition, the plaintiff contends that the survey questions are leading because the response options vary depending on the product the respondent stated he or she used during the prior tax year. Id.
The defendants respond that the questions in the survey were not inappropriately leading and that authority cited by the plaintiff's expert actually establishes that "closed-ended questions
*13 are more suitable for assessing choices between well-identified options or obtaining ratings on a clear set of alternatives." Defs.' Mem. at 8-9 (citing Diamond at 253). The defendants also note that the questions were not wholly closed-ended in that "Other" was an option that respondents could select. Id. at 9 .
By providing survey respondents with a pre-selected list of alternative options, rather than letting respondents respond organically, the survey does lead respondents to think about the market for tax preparation services in the same terms that the defendants do, which may have led respondents to select options they otherwise would not have selected. This effect is not so inherently suggestive as to render the survey's results wholly unreliable and therefore inadmissible, however. The survey does not appear to lead respondents to select any particular answer; the response choices included the major market participants under both parties' views of the market; and it also included an "Other" option. Moreover, the survey question cannot be considered to be as "leading" as the questions identified as problematic in the plaintiff's cited authority. [7] Accordingly, this critique goes to the weight of the evidence and not to admissibility.
The same is true of plaintiffs' concerns about the survey's failure to discourage guessing by not including an "I don't know" option. The plaintiff argues that "[s]urveys should explicitly mention that it is completely appropriate for respondents to have no opinion [on] a given question" and that failure to include an "I don't know" opinion skews the results by failing to
*14
provide an adequate answer for respondents who have no opinion on the question. Pl.'s Mem. at 10-11. While this concern is valid and the absence of such an "I don't know" option does diminish the weight the Court should accord this survey, the failure to include this option is partially mitigated here by the inclusion of the "Other" option. The Court does not find that this defect so undermines the survey as to require it to be deemed inadmissible. See Kinetic Concepts,
3. Consideration of All Critiques
In sum, the plaintiff has identified a number of aspects about the methodology and wording of the defendants' 2011 email survey that will impact the weight that the Court gives this survey in its analysis. These defects do not undermine the survey and the expert's reliance on it so overwhelmingly that they render the survey wholly irrelevant or unreliable-and therefore inadmissible. While the admissibility of this survey might be a closer question in the context of a jury trial, since this hearing is not before a jury, "the importance of the trial court's gatekeeper role is significantly diminished . . . because, there being no jury, there is no risk of tainting the trial by exposing a jury to unreliable evidence." Whitehouse Hotel Ltd. Partnership,
*15
III. CONCLUSION
For the reasons explained above, the plaintiff's motion in limine is DENIED. This Memorandum Opinion and Order shall be filed under seal. On or before September 12, 2011, the parties shall advise the Court regarding whether any portion of this Memorandum Opinion should be redacted before public filing because it contains confidential information. In addition, if the parties have not already filed versions of their legal memoranda that may be filed publicly, they shall do so by September 12, 2011.
DATED: September 6, 2011
NOTES
Notes
The Court has subject matter jurisdiction to hear this suit under 15 U.S.C. § 25, and 28 U.S.C. §§ 1331, 1337 and 1345 .
The plaintiff initially filed an additional motion in limine to exclude evidence of defendants' guarantee to maintain TaxACT's current prices for three years following the acquisition. See ECF No. 44. The plaintiff withdrew that motion at oral argument on September 2, 2011.
2nd Story Software, Inc. ("2SS") is a wholly-owned subsidiary of 2SS Holdings, Inc., which is the entity being purchased by H&;R Block. Declaration of Lance Dunn, dated May 27, 2011 ("Dunn Decl."), ¶¶ 2, 4. Both 2SS and 2SS Holdings, Inc. share the same address in Cedar Rapids, Iowa.
The response options varied among four different categories of filers, which are discussed further below. For example, the list of options presented to filers who completed a free federal tax return and no state return were: "I would prepare myself without help," "TurboTax Free Edition," "H&;R Block at Home Free Edition," "Free TaxUSA Free Edition," "Complete Tax Free Basic," "An Accountant," "I would use a product on FFA [i.e., Free File Alliance]," "TaxSlayer Free Edition," "Jackson Hewitt Free Basic," "Tax$imple Free Basic," and "Other." GX604 at 2 .
By contrast to federal tax returns, TaxAct does not provide free state return preparation software but does offer free electronic filing of state returns to those customers who purchase a TaxAct-branded desktop software product available at Staples retail stores. Report of Dr. Meyer 191-92. The Court understands the survey's category for "No State" to cover those respondents who did not purchase TaxAct's state return preparation software from Staples or any other source.
At oral argument and in its briefs, the plaintiff urged the Court to follow United States v. Dentsply Int'l Inc., an antitrust case in which the district court excluded a survey.
In Novartis Consumer Health, Inc. v. Johnson &; Johnson-Merck Consumer Pharms. Co.,
