Katz, Abosch, Windesheim, Gersham & Freedman, P.A. v. Parkway Neuroscience & Spine Institute, LLC

Annotate this Case
Justia Opinion Summary

In this professional malpractice action, the Supreme Court vacated the judgment of the appellate court reversing the summary judgment of the trial court in favor of Defendant because Plaintiff could not prove damages and remanded this case to the circuit court for further proceedings consistent with this opinion, holding that remand was required.

Plaintiff, a medical and surgical practice, retained Defendant, an accounting firm, in 2013 and terminated Defendant's services in 2015. In 2018, Plaintiff sued Defendant to recover damages for lost profits. Plaintiff designated a certificated public accountant (CPA) as an expert witness, who used the "before-and-after" method to calculate Plaintiff's lost profits. In 2021, the CPA issued updated calculations reflecting "normalizing adjustments" that she had made. The trial court excluded the CPA's expert testimony based on its application of the Daubert-Rochkind factors. The appellate court reversed the trial court's exclusion of the CPA's testimony and remanded the case for further proceedings. The Supreme Court remanded the cause, holding that the trial court erred in its consideration of the normalizing adjustments as reflecting on the reliability of the CPA's methodology, as opposed to the credibility of the CPA herself.

Download PDF
Katz, Abosch, Windesheim, Gershman & Freedman, P.A., et al. v. Parkway Neuroscience and Spine Institute, LLC, No. 30, September Term, 2022. Opinion by Biran, J. EXPERT WITNESSES – ADMISSIBILITY OF EXPERT TESTIMONY – MARYLAND RULE 5-702 – LIMITED REMAND – Respondent filed a lawsuit against Petitioners alleging accountant malpractice and related claims. In the course of discovery, Respondent designated an expert to provide an opinion concerning Respondent’s lost profits resulting from Petitioners’ alleged torts. Petitioners moved to exclude the testimony of the proffered expert under Daubert v. Merrell Dow Pharmaceuticals, Inc., 509 U.S. 579 (1993), and Rochkind v. Stevenson, 471 Md. 1 (2020). After conducting a DaubertRochkind hearing, the circuit court granted Petitioners’ motion to exclude the proffered expert testimony. The Supreme Court of Maryland held that much of the trial court’s consideration of the Daubert-Rochkind factors was appropriate, including the trial court’s assessment of how the expert’s choice of data, assumptions, and other inputs affected the reliability of her methodology. However, the trial court erred when it considered the expert’s “normalizing adjustments” that recategorized certain expenses from one year to another, as reflecting on the reliability of the expert’s methodology. The Court ordered a limited remand to the circuit court under Maryland Rule 8-604(d)(1) so that the trial court may decide to admit or exclude the expert’s testimony without consideration of the normalizing adjustments as reflecting on the reliability of the expert’s methodology. Circuit Court for Howard County Case No.: C-13-CV-18-000181 Argued: May 4, 2023 IN THE SUPREME COURT OF MARYLAND* No. 30 September Term, 2022 KATZ, ABOSCH, WINDESHEIM, GERSHMAN & FREEDMAN, P.A., ET AL. v. PARKWAY NEUROSCIENCE AND SPINE INSTITUTE, LLC Fader, C.J. Watts Hotten Booth Biran Gould Eaves, JJ. Pursuant to the Maryland Uniform Electronic Legal Materials Act (§§ 10-1601 et seq. of the State Government Article) this document is authentic. 2023-08-30 12:44-04:00 Opinion by Biran, J. Booth, J., concurs. Gould, J., concurs in part and dissents in part. Watts, J., dissents. Filed: August 30, 2023 Gregory Hilton, Clerk * At the November 8, 2022 general election, the voters of Maryland ratified a constitutional amendment changing the name of the Court of Appeals of Maryland to the Supreme Court of Maryland. The name change took effect on December 14, 2022. When this Court adopted the Daubert1 expert testimony admissibility standard in Rochkind v. Stevenson, 471 Md. 1 (2020), we embraced a regime that prizes the reliability of an expert’s methodology over its general acceptance. We empowered trial judges to protect juries from junk science while also broadening the range of possibly admissible opinions beyond just those dominant among practitioners. We asked judges to engage with the science without playing amateur scientist, and we promised the deference appropriate to courts administering a flexible approach to analyzing the admissibility of expert testimony. This case requires us to reflect on that flexibility and deference. Parkway Neuroscience and Spine Institute, LLC (“PNSI”, the Respondent here) is a medical and surgical practice that began to expand in 2011 and needed accounting help. In 2013, PNSI retained accounting firm Katz, Abosch, Windesheim, Gershman & Freedman, P.A. and, specifically, Mark Rapson, who specialized in medical practice accounting (we shall refer to the firm and Mr. Rapson, the Petitioners here, collectively as “KatzAbosch”). Within a few years after retaining KatzAbosch, PNSI began to disintegrate; members of the practice began leaving in 2015, and by the middle of 2016, only two members remained of the nine who had been in place at the end of 2014. PNSI terminated KatzAbosch’s services in 2015. PNSI alleges that malpractice by KatzAbosch caused the mass exodus of its members. In 2018, PNSI sued KatzAbosch in the Circuit Court for Howard County to recover damages for lost profits. To establish those damages, PNSI designated certified 1 See Daubert v. Merrell Dow Pharmaceuticals, Inc., 509 U.S. 579 (1993). public accountant Meghan Cardell as an expert witness. She used the widely accepted “before-and-after” method to calculate PNSI’s lost profits, choosing 2015 as a “baseline” period against which she would compare the actual profits in subsequent years through 2019, and adding up the differences to arrive at an estimate of what profits PNSI missed out on due to KatzAbosch’s alleged harmful conduct. A few weeks before the June 2021 Daubert-Rochkind hearing, Ms. Cardell issued updated calculations reflecting some “normalizing adjustments” she had made; although PNSI’s accounting records had not changed since her initial analysis, Ms. Cardell reviewed PNSI’s financial information again and noticed some payments that had been categorized in the wrong years. She reallocated those payments to the years she believed to be correct and updated her calculations. Those two issues – Ms. Cardell’s choice of 2015 as the “before” in her “before-andafter” analysis and her June 2021 normalizing updates – rose to the top of the trial court’s mind in the Daubert-Rochkind hearing. The trial court noted speculative and insufficiently substantiated judgment calls that Ms. Cardell had made in arriving at the 2015 benchmark. Among other things, the trial court wondered why Ms. Cardell had chosen 2015 (a profitable year) rather than, say, an average that included the several (unprofitable) years prior to the alleged harm event. The trial court also was concerned about Ms. Cardell’s inability to articulate industry standards relating to the concept of “economic impact” and to the proper treatment of owner draws. In addition to these points, the trial court commented several times about Ms. Cardell’s June 2021 normalizing adjustments, which negatively affected its opinion of Ms. Cardell’s reliability. Essentially, the court did not understand why it had taken Ms. Cardell 2 so long to notice the errors. The court discussed these adjustments when considering the Daubert factors relating to a methodology’s error rate and to whether the field of expertise claimed by the expert is known to reach reliable results for the type of opinion the expert would give. Based on its application of the Daubert-Rochkind factors, the trial court excluded Ms. Cardell’s testimony, leading to summary judgment in favor of KatzAbosch because PNSI could not prove damages. PNSI appealed, and the Appellate Court of Maryland2 held that the circuit court abused its discretion in finding Ms. Cardell’s methodology unreliable. As to the 2015 baseline choice, the Appellate Court agreed with PNSI that the choice was a question of data (and thus a factual question for the jury) rather than of methodology. With respect to the normalizing adjustments, the Appellate Court said that Daubert’s “error rate” factor must be understood as the rate of unknown errors in the methodology employed, not as an “error correction rate,” or else courts would create incentives against experts disclosing and explaining errors they made. The intermediate appellate court reversed the trial court’s exclusion of Ms. Cardell’s expert testimony and remanded for consistent proceedings. KatzAbosch petitioned this Court for further review. As we explain more fully below, the choice or calculation of the inputs to a methodology can be a part of the methodology itself, and we reject an unduly rigid dividing 2 At the November 8, 2022 general election, the voters of Maryland ratified a constitutional amendment changing the name of the Court of Special Appeals of Maryland to the Appellate Court of Maryland. The name change took effect on December 14, 2022. 3 line between “data” and “methodology” that binds courts to admit methodologically questionable analyses cloaked as data. To the extent the trial court considered how Ms. Cardell’s choice of data, assumptions, and other inputs affected the reliability of her methodology, the trial court’s Daubert-Rochkind analysis was proper. However, the trial court erred in its consideration of the normalizing adjustments as reflecting on the reliability of Ms. Cardell’s methodology, as opposed to the credibility (or reliability) of Ms. Cardell herself. After a careful review of the record, we determine that the fair and prudent course of action at this point is to remand the case to the circuit court to decide whether to admit or exclude Ms. Cardell’s testimony without consideration of the June 2021 normalizing adjustments as reflecting on the reliability of Ms. Cardell’s methodology. I Background A. Facts During the period relevant to this case, PNSI was a Western Maryland and Pennsylvania mixed medical practice that diagnosed, managed, and treated disorders of the brain, spine, and peripheral nervous system. It employed neurosurgeons, interventional and non-interventional pain physicians, neurologists, physicians’ assistants, and support staff. The practice had operated since 1998. Beginning in 2011, PNSI expanded, hiring more physicians and support staff. These efforts caused PNSI to spend more on salaries and build-out expenses without offsetting revenue. At the end of 2014, the practice had nine member-owners, all physicians. 4 1. The Engagement of KatzAbosch None of these members, however, were accounting or finance experts. In early 2013, PNSI’s long-time accountant advised PNSI that the practice had outgrown his firm’s services and recommended that the practice retain a new accounting firm to help guide PNSI through its growth and expansion process. So PNSI began searching for a firm that specialized in medical practice accounting and finance. In October 2013, PNSI retained KatzAbosch to provide tax, accounting, and financial advice and services, as well as to provide “expert business and financial guidance and direction to help PNSI continue to grow its practice.” The engagement included analyzing PNSI’s general ledger and financial statements, making recommendations concerning PNSI’s financial affairs, and designing and administering a new member compensation model. Mr. Rapson (chair of KatzAbosch’s Medical Services Group) was responsible for the account. PNSI’s 2012 Operating Agreement provided the compensation terms for its member-physicians. First, each member was to maintain a capital account on the books of the practice, with a minimum balance that PNSI’s Board of Managers (the “Board”) would determine annually. Second, members were to be paid a monthly “draw” (determined by the Board) at the start of each year – in essence a form of salary. Third, revenue received by the practice for hospital and trauma calls would be distributed to the member who had taken the call, with PNSI functioning as a pass-through entity. These pass-through payments, along with the monthly draws, comprised “guaranteed payments” to the members. Finally, each quarter, the Board would distribute any excess cash flow (“distributions”) in proportion to each member’s ownership stake. 5 In early 2014, KatzAbosch designed and proposed a new compensation model, and PNSI adopted it. According to PNSI, KatzAbosch’s model did not reserve funds for known build-out-related expenses concerning one of PNSI’s locations that would be coming due later that year as well as other significant expenses. PNSI alleges that, despite these expenses coming due, KatzAbosch directed PNSI to make almost $1 million in quarterly distributions to members between July and October 2014. 2. Termination of KatzAbosch and the Departure of Most of PNSI’s Members PNSI alleges that, as a result of KatzAbosch’s erroneous advice, PNSI almost ran out of money by the end of 2014. According to PNSI, the practice had to use over $660,000 from its line of credit in the fourth quarter of 2014 alone. In January 2015, KatzAbosch disclosed to PNSI its precarious financial situation. While PNSI had had almost $1 million in cash on hand at the start of 2014, it had less than $40,000 by the start of 2015. PNSI terminated KatzAbosch’s services in the spring of 2015. The practice then began to come apart. PNSI alleges that members – who had personally guaranteed loans the practice had taken – were increasingly distressed about the financial condition of the practice and their personal liability. Starting in mid-2015, members began to withdraw from the practice, taking with them patients and the associated revenue streams. By mid-2016, only two members – Dr. Brian Holmes and Dr. Neil O’Malley – remained of the nine who had been in place at the end of 2014. KatzAbosch claims that Dr. Holmes and Dr. O’Malley received more compensation from the practice after they were the only two members remaining (thereby depressing estimates of PNSI’s 6 profits post-exodus), although PNSI explains this as the result of those two doctors taking more trauma calls and thus receiving more in guaranteed payments. B. The Lawsuit In 2018, PNSI sued KatzAbosch in the Circuit Court for Howard County, alleging claims for accountant malpractice, negligent misrepresentation, breach of contract, and unjust enrichment. PNSI initially sought damages for lost profits, settlement amounts paid to a departing member, written-off amounts owed by two departing members, fees paid to KatzAbosch, and prejudgment interest, all totaling $9,456,035. The bulk of the claimed damages was the alleged lost profits. 1. Meghan Cardell, PNSI’s Damages Expert In July 2019, PNSI designated Meghan Cardell as an expert witness expected to testify that “PNSI suffered and will continue to suffer significant financial losses resulting from the mass exodus of members and staff, loss of patients, replacement of long-time established physicians with newer-practicing physicians, and litigation with withdrawing members, as identified in the Calculation of Damages[.]” Ms. Cardell is a certified public accountant and a certified fraud examiner. At the time of her designation by PNSI as an expert in this case, Ms. Cardell was Director for Disputes and Investigations at the Washington, D.C. accounting firm Alvarez & Marsal, where she had worked since 2014. Before that, she was a senior associate in forensics at Veris Consulting from 2011 to 2014. Ms. Cardell used the “before-and-after” methodology of calculating PNSI’s lost profits. Under this approach, the expert compares profits before and after a damaging event, 7 the former being the “benchmark” or “base” period and the latter being the “loss” period. Ms. Cardell initially calculated PNSI’s damages in May 2019, as follows: Damages (less prejudgment interest) (Calculated in May 2019) Total lost profits $8,520,744 Dr. DeMarco Settlement $89,421 Dr. Sullivan Amount Owed $84,836 KatzAbosch Fees $182,010 Total: $8,877,012[3] Prejudgment interest Lost profits $503,999 DeMarco & Sullivan $20,185 KatzAbosch Fees $54,839 Total prejudgment interest: $579,023 Total Damages $9,456,035 The lost profits total included figures for 2016 through 2025. Ms. Cardell selected 2015 as the base year. She then used actual figures to calculate the lost profits for 2016, 2017, and 2018; she used the trend line from those three years to project lost profits through 2025, discounting the figures by 20% in 2019-22 and 50% in 2023-25 to reflect a trend of declining annual lost profits as the practice would gain doctors, patients, and income in the future. On July 26, 2019, KatzAbosch moved to strike PNSI’s claim and to exclude Ms. Cardell’s testimony. At a hearing in June 2020, the trial court (the Honorable Richard S. Bernhardt, specially assigned) denied KatzAbosch’s motion, but expressed concern about admitting Ms. Cardell’s calculations at trial if she continued to project lost profits for 2020 3 This total amount is one dollar greater than the sum of the four amounts that precede it. It is not clear whether this discrepancy was due to a rounding error or some other issue. 8 through 2025 related to the 2015-16 departure of physicians, despite the changed landscape created by the COVID-19 pandemic. On May 5, 2021, KatzAbosch renewed its motion to strike PNSI’s lost profit claims and to exclude Ms. Cardell’s testimony under this Court’s recently adopted DaubertRochkind standard for expert testimony, disputing the factual basis of Ms. Cardell’s Calculation of Damages report. Ms. Cardell updated her Calculation of Damages on May 17, 2021. PNSI no longer sought damages for 2020 or beyond, and Ms. Cardell’s new calculations omitted those years. The updated calculations featured a revised damages calculation of $7,335,447: Damages (less prejudgment interest) – Updated May 17, 2021 (dropping 2020 and onward; concerning only 2016-19 pre-pandemic; using finalized actual 2019 figures) Total lost profits $8,520,744 $5,789,521 Dr. DeMarco Settlement $89,421 Dr. Sullivan Amount Owed $84,836 KatzAbosch Fees $182,010 Total: $8,877,012 $6,145,789 Prejudgment interest Lost profits $503,999 $1,078,968 DeMarco & Sullivan $20,185 $37,630 KatzAbosch Fees $54,839 $73,060 Total prejudgment interest: $579,023 $1,189,658 Total Damages $9,456,035 $7,335,447 9 On May 27, 2021, the trial court scheduled a full-day, in-person evidentiary hearing on KatzAbosch’s renewed motion for June 30, 2021. On June 11, 2021, Ms. Cardell updated her calculations again: Damages (less prejudgment interest) – Updated June 11, 2021 Total lost profits $8,520,744 $5,789,521 $4,956,080 Dr. DeMarco Settlement $89,421 Dr. Sullivan Amount Owed $84,836 KatzAbosch Fees $182,010 Total: $8,877,012 $6,145,789 $5,312,348 Prejudgment interest Lost profits $503,999 $1,078,968 $890,837 DeMarco & Sullivan $20,185 $37,630 $38,346 KatzAbosch Fees $54,839 $73,060 $73,808 Total prejudgment interest: $579,023 $1,189,658 $1,002,991 Total Damages $9,456,035 $7,335,447 $6,315,339 Notably, the June 2021 updates eliminated the loss from 2016 (meaning Ms. Cardell had now found that year to be profitable) and showed lost profits only from 2017-19. These changes reflected “normalizing adjustments” discussed in detail below. 2. The Daubert-Rochkind Hearing The trial court conducted a Daubert-Rochkind hearing on June 30, 2021. The crux of PNSI’s argument was that KatzAbosch’s concerns with Ms. Cardell dealt less with her methodology and more with the assumptions she had made, and those assumptions went to weight rather than admissibility. The court declined either to accept or to reject Ms. Cardell’s qualifications as an expert, although it eventually said her experience (or lack thereof) with “niche” medical practices bore somewhat on the reliability of her testimony, independent of the question of her qualification as an expert witness. 10 a. Issues Raised in the Hearing i. 2015 as the Base Year After describing why other methodologies to determine contractual damages and lost business value (and, relatedly, lost future profits) were not applicable here, Ms. Cardell described the before-and-after methodology that she chose to apply. She testified that this method “in [her] experience is the most commonly used and most commonly accepted methodology of measuring lost profits,” that she had used it in her career “many, many times,” and that she had seen it used by other experts in and out of litigation settings just as frequently. She oriented her analysis around the harm event of seven of nine doctors withdrawing from the practice within a short time; the first doctor left in June 2015, a few more at the end of 2015, and a few more by June 2016. So calendar year 2015 was her benchmark “before” period, and the “after” period (or “loss” period) began in calendar year 2016 and ran through 2019. She described this as a “conservative proxy” for the practice’s future earnings, had so many members not left. The business had been profitable in 2015, and PNSI had been investing to grow; Ms. Cardell said “[t]he practice had sort of hit its stride in 2015.” She said medical businesses are not subject to swings of consumer preference that might have destabilized the results of those investments, and the medical specialty industry was projected to grow because of aging Baby Boomers and the prevalence of chronic illness. The trial court articulated some concerns with Ms. Cardell’s methodology. First, the court asked why she had chosen just one year (2015) rather than a benchmark of multiple years before the harm event; Ms. Cardell said the benchmark could be an average or could 11 be one year, particularly if there were past years that would not be representative of what future earnings would look like. Second, the court noted that 2015 had been the most profitable year since 2010,4 and the court did not understand why 2015, rather than a slower year, should become the “new normal,” as there might have been post-harm years that also would have been a “squeeze.” Ms. Cardell believed 2014 was not entirely unusual but would have been unrepresentative because 2015 was the year when PNSI had emerged from the phase of significant expenditures (and reduced profits) in order to grow the practice. Third, the court questioned how Ms. Cardell knew that PNSI had truly hit its stride in 2015 and that there would not have been significant future problems caused by overexpansion – personality conflicts between the new doctors, sub-par new facilities, etc. She answered that the practice had been profitable in 2016 even as the exodus had begun, showing that it could still be profitable despite some amount of dysfunction and confirming that it would have continued to generate profits in the future without the full exodus. ii. Other Data and Assumptions (Reimbursement Rates) Ms. Cardell assumed that, but for the harm event, PNSI’s profits in 2017, 2018, and 2019 would have at least equaled the 2015 profits. Ms. Cardell had not analyzed changes in medical service reimbursement rates (which had been decreasing) during the relevant years, although the market research she examined showed that medical industry revenues were expected to grow because of expanded access to Medicare, Medicaid, and private health insurance under the Affordable Care Act. The core of the trial court’s concern was 4 As measured by Ms. Cardell, PNSI had had a $22,000 profit in 2010, losses in 2011-14, and a $321,751 profit in 2015. 12 whether Ms. Cardell had looked “at all of the various revenue threads that come up with the fabric of [PNSI’s] yearly income” or whether she had just been “looking at numbers without understanding from further records or where those numbers came from[.]” Ms. Cardell said that her analysis was based on revenues. iii. Correction of Calculations Ms. Cardell described normalizing adjustments she had made to her calculations earlier in that month relating to trauma and on-call payments. In 2016, PNSI had made payments to its physicians based on trauma and on-call revenues the practice had received in 2015. Ms. Cardell initially included those payments as expenses of the practice for 2016. After Ms. Cardell discovered that the practice had received the associated revenues in 2015, she removed that expense from 2016 and added it to 2015. This reduced by half PNSI’s adjusted income in base year 2015 by increasing adjusted expenses. Ms. Cardell’s $395,010 downward adjustment in expenses for 2016 increased adjusted income for that year, making 2016 profitable (compared to her earlier May 2021 estimates for 2016, which had shown a loss) and contributing to the elimination of any loss in profits in 2016. There were also trauma and on-call adjustments for 2017, 2018, and 2019. In those years, Dr. Holmes and Dr. O’Malley earned income based on their trauma and on-call services, and that money flowed to the practice for normal pass-through purposes, but they chose not to pay themselves those funds; rather, they effectively loaned those amounts to the practice in order to increase its cash on hand. In the past, those guaranteed payments would have been recorded as expenses, so Ms. Cardell included them as such via the June 2021 adjustments. Her first updated calculations (issued May 2021) had included upward 13 trauma and on-call expense adjustments for years 2017 ($129,573), 2018 ($767,573), and 2019 ($767,573). Each of these upward adjustments had the effect of increasing adjusted expenses and therefore reducing adjusted income, making the lost profits figures (and thus PNSI’s alleged damages) larger for each year. Ms. Cardell’s second updated calculations (issued June 2021) decreased the upward expense adjustments for 2017 (to $77,000) but increased the upward expense adjustments for 2018 (to $842,400) and 2019 (to $1,233,617) in comparison to her first updated calculations. For all three years, the post-update upward adjustments still increased adjusted expenses, decreased adjusted income, and increased lost profits (and damages). Although the information that led to these adjustments had long been available to her, Ms. Cardell explained that she looked again at the numbers ahead of the DaubertRochkind hearing and identified a payment that looked like it belonged to a different year. A PNSI accounting manager confirmed this intuition, and so Ms. Cardell made the judgment that the changes relating to trauma and on-call payments/expenses were needed. The trial court expressed concern that Ms. Cardell’s calculations might be liable to additional such changes; Ms. Cardell testified that she did not believe any other adjustments were necessary. KatzAbosch’s counsel attempted to characterize Ms. Cardell’s updates to her calculations (from May 2021 to June 2021) as a “fifty percent error rate” because she had made revisions between the first and second updates. 14 iv. Treatment of Member Draws The trial court noted two different ways that a business could treat member draws when calculating profits. To illustrate the point, the court referred to a hypothetical limited liability corporation that, after paying all its expenses other than potential draws to its owner, has $100 in cash. On the one hand, the owner could withdraw the $100 as salary and the business would show no profit that year. On the other hand, the business could be considered to have $100 in profits regardless of whether the owner withdraws it from the business that year. Ms. Cardell said that she believed the latter was the correct way to look at it, but there was no industry standard on this issue one way or the other. The trial court expressed skepticism that there was no industry standard, given that the classification of draws (owner salary) as profits or expenses “seems like a pretty basic issue that [is] capable of rearing its head in every case in which … an owner’s draw is possible.” v. Lack of Member-Specific Lost Profits Calculation Ms. Cardell was “unable to parse out” the financial effects of any particular member’s departure from the practice. She could not recall whether she had looked at specific collections for Dr. Holmes and Dr. O’Malley, her analysis having been based on overall revenues for the practice. Defense counsel presented her with PNSI financial records showing that those doctors’ post-exodus revenue collections declined from 2016 through 2019. b. The Trial Court’s Ruling PNSI acknowledged that “reasonable minds can differ” as to whether its profits in 2015 was the appropriate benchmark against which to measure the practice’s future profits. 15 However, PNSI argued that none of the Daubert-Rochkind factors militated toward excluding Ms. Cardell. The court ruled from the bench, granting KatzAbosch’s motion to exclude Ms. Cardell’s testimony, based on the Daubert-Rochkind standard. i. The Court’s Overall Sense The trial court began its ruling by acknowledging the wide acceptance of the beforeand-after methodology for calculating lost profits. The court then highlighted its prime concerns with Ms. Cardell’s testimony: speculation, ipse dixit “judgment calls,” helpfulness to the jury, and information Ms. Cardell had failed to consider. First, the court found that Ms. Cardell’s selection of profitable 2015 as a benchmark (rather than unprofitable 2011-14, marginally profitable 2010, or an average) was speculation, especially given that she had no specialization in a niche practice like PNSI. So, too, was Ms. Cardell’s analysis that, after years of unprofitability, 2015 marked the critical turning point and that, in the court’s words, “the rocket had left the launch pad and was going straight up.” The speculation finding was bolstered by the fact that Ms. Cardell relied heavily on direct oral communications with PNSI and its employees rather than information more concretely assessable by the jury. Second, the court pointed to unreliable and ipse dixit “judgment calls.” The court discerned possible bias and some degree of unreliability in the judgment calls that Ms. Cardell made without some clear authority as to (1) the June 2021 normalizing adjustments (which the court said Ms. Cardell could have made as part of her original calculations); (2) “whether owner draws reduce profits or not”; and (3) Ms. Cardell’s subjective 16 understanding of the term “economic impact.”5 These judgment calls took on special importance because, the court said, either Ms. Cardell lacked awareness of industry standards or no industry standards existed, heightening her reliance on unverified sources of information and her own say-so. Third, the court questioned how helpful Ms. Cardell’s testimony would be to the jury. Ms. Cardell chose not to disaggregate revenues according to each individual departing physician, and instead she considered lost profits in an “all or nothing” manner, rendering her opinion “only helpful if the jury accepts that each and every doctor of the seven … left solely because … of the acts of [KatzAbosch].” Relatedly, the trial court stated that “[n]ot every doctor, necessarily is going to make or contribute the same amount of money each and every year. There was nothing presented in the Report to, that considered whether or not the income, the revenue generating ability of the remaining doctors or the leaving doctors was considered [sic]. It was just, let’s take 2015 and go from there. Let’s just use the numbers as we get them without examining whether or not, you know, the doctors who have income would have had that income and things of that nature. And again, there’s nothing about her training to me that would qualify her to make those assumptions that all these numbers would not be affected by passage of time and the doctor’s passage of time.” 5 Ms. Cardell testified that she considers “economic impact” in her work. When the court asked her what “economic impact” means, Ms. Cardell began her response by saying, “the way I think about it is…” This raised a red flag for the trial court: “[Y]ou start off with, ‘the way I think of it’. What does your industry consider it to be?” Ms. Cardell could not provide a specific industry definition for the term. 17 Fourth, the court noted that Ms. Cardell had failed to consider certain important pieces of information, including: (1) whether Dr. Holmes and Dr. O’Malley had other income streams; and (2) how changing insurance reimbursement rates were influencing the practice’s profitability. ii. The Daubert Factors The trial court then considered the Daubert-Rochkind factors: (1) Testing – The court acknowledged that a “before-and-after” analysis was appropriate and testable generally, but questioned the testability of Ms. Cardell’s judgment calls, including why she had chosen 2015 as the base year and the meaning of “economic impact.” (2) Peer review – The court did not find this factor relevant. (3) Rate of error – The court did not find this factor relevant as it is normally applied, where there is some known rate of false positive or false negative results. But the court did note concern with Ms. Cardell’s June 2021 pre-hearing updates, as those changes were not caused by a change in the facts. (4) Standards and controls – The court said “there was very little evidence of any standards of controls that exist,” in particular on economic impact and the treatment of owner draws in profit computations. 18 (5) General acceptance – The court acknowledged that the beforeand-after analysis to measure lost profits was generally accepted. (6) Purpose: prior research or litigation – Although the court expressed that there is “no inherent negativity” to experts who develop their opinions for litigation purposes compared to those who develop their expertise for independent research, the court did express concerns about Ms. Cardell’s heavy reliance on oral communications with PNSI personnel. (7) Unjustifiable extrapolation from an accepted premise – The court said: “This applies if the … premise that we’re talking about from which the unfounded conclusions roles [sic] would be the acceptance of 2015 as the benchmark[.]” (8) Accounting for obvious alternative explanations – The court found that Ms. Cardell “clearly” had not accounted for alternative explanations – in particular, she had failed to consider doctorspecific revenue generation and changing reimbursement rates’ effects on revenue. (9) Care here as in professional non-litigation work – The court observed as to this factor: “I don’t find that to be applicable, I don’t know what to say about that and I have no reason to think that she blew this off as an inconsequential project, I mean she took this very seriously.” 19 (10) Field known to reach reliable results for this type of opinion – The court “incorporate[d] everything [it had] said,” noting in particular the June 2021 updates (driven by subjective reasons rather than newly revealed facts) as “mak[ing] the whole reliability even that much more suspect.” In sum, the court said, PNSI had failed to meet its burdens primarily as to Ms. Cardell’s reliability and secondarily as to Ms. Cardell’s usefulness to the jury, although the court described the “usefulness” finding as a “very, very slight factor.” On the following day, the trial court issued a written supplement to its oral ruling in which it discussed the persuasive value of CDW LLC, et al. v. NETech Corp., 906 F. Supp. 2d 815 (S.D. Ind. 2012). In CDW, the expert determined lost profits by using the “yardstick method” – i.e., he compared the subject business branch’s profits to the plaintiff-business’s other branches. The CDW Court noted that “[a]n expert’s choice in data sampling is at the heart of his methodology. A yardstick approach is an acceptably reliable method under Daubert for calculating lost profits only if the benchmarks (or yardsticks) are sufficiently comparable that they may be used as accurate predictors of what the target would have done.” CDW, 906 F. Supp. 2d at 824 (internal citation omitted). “Absent the requisite showing of comparability, a damage model that predicts either the presence or absence of future profits is impermissibly speculative and conjectural.” Id. (internal quotation marks and citation omitted). The trial court found the before-and-after and yardstick methods sufficiently similar and wrote that “the expert’s choice of a benchmark in CDW is analogous to the expert’s choice of 2015 as the benchmark in the instant case. … Thus 20 [this] Court cites CDW as having persuasive value in finding that [PNSI] has not satisfied Maryland Rule 5-702 and this Court’s decision to exclude this expert on Daubert grounds.” The parties subsequently filed a stipulation of dismissal, agreeing (among other things) that PNSI could not prove a prima facie case with respect to its claims for accountant malpractice, negligent misrepresentation, and breach of contract, in light of the trial court’s exclusion of Ms. Cardell’s expert testimony. The trial court entered summary judgment in favor of KatzAbosch and dismissed the case in its entirety. 3. Appeal PNSI appealed to the Appellate Court of Maryland. In a reported opinion, the Appellate Court held that the circuit court had abused its discretion in finding Ms. Cardell’s methodology unreliable under Daubert-Rochkind. Parkway Neuroscience and Spine Institute, LLC v. Katz, Abosch, Windesheim, Gershman & Freedman, P.A., et al., 255 Md. App. 596, 623-37 (2022). The Appellate Court discerned error in the trial court’s criticisms of Ms. Cardell’s testimony based on: (1) the selection of 2015 as the base year; (2) insurance reimbursement rates; (3) standards for member draws; (4) the June 2021 updates to her calculations; and (5) the lack of individual per-doctor lost profit figures. Id. at 623-37. The Appellate Court also interpreted the trial court’s comments regarding Ms. Cardell’s lack of experience with respect to specialty medical practices as a finding that she lacked the requisite qualifications to be accepted as an expert. See id. at 622-23. Specifically, the Appellate Court concluded that “[a]nalyzing reimbursement rates, selecting the base year, and relying on data from and conversations with PNSI all are issues with the soundness of the data” rather than with the reliability of Ms. Cardell’s 21 methodology. Id. at 627 n.9. The court further opined that “[w]hether Ms. Cardell should have deducted the member draws from the projected profits is something [KatzAbosch] can attack during cross-examination before a jury – but not at the Daubert-Rochkind hearing.” Id. at 633. The Appellate Court considered the proper treatment of member draws to be “a fact-laden issue – involving credibility, not reliability.” Id. As to the June 2021 updates, the Appellate Court determined that the trial court misapplied the “known or potential rate of error” Daubert factor, because “[t]he error rate that Daubert speaks of is the rate of unknown errors in the methodology employed, not an error correction rate.” Id. at 634 (internal quotation marks and citation omitted) (emphasis in original). To find methodological unreliability in an expert’s decision to correct errors in her earlier analysis “would be a disincentive to disclose and explain errors.” Id. at 635. Finally, regarding Ms. Cardell’s failure to calculate member-specific lost profits, the Appellate Court concluded that the circuit court looked improperly to causation when the purpose of the Daubert hearing was to assess the reliability of Ms. Cardell’s methodology – not whether PNSI had met its burden of proof on causation. Id. at 636. The Appellate Court reversed the circuit court’s decisions excluding Ms. Cardell, striking PNSI’s lost profits claim, and granting summary judgment. The Appellate Court remanded the case to the circuit court for proceedings consistent with its opinion. Id. at 639. KatzAbosch then petitioned this Court for a writ of certiorari, which we granted on January 20, 2023. Katz, Abosch, Windesheim, Gershman & Freedman, P.A. v. Parkway Neuroscience and Spine Institute, LLC, 482 Md. 534 (2023). KatzAbosch presents the 22 following question for our review: “Did the [Appellate Court] err in finding that the trial court abused its discretion in excluding expert testimony on lost profits?” II Standard of Review Appellate courts review a trial court’s decision concerning the admissibility of expert testimony under Maryland Rule 5-702 for abuse of discretion. See Rochkind, 471 Md. at 10-11; State v. Matthews, 479 Md. 278, 305-06 (2022). As we said in Matthews: Under this standard, an appellate court does “not reverse simply because the ... court would not have made the same ruling.” Devincentz v. State, 460 Md. 518, 550 (2018) (internal quotation marks and citation omitted). “Rather, the trial court’s decision must be well removed from any center mark imagined by the reviewing court and beyond the fringe of what that court deems minimally acceptable.” Id. (internal quotation marks and citation omitted); see also Williams v. State, 457 Md. 551, 563 (2018) (“An abuse of discretion occurs where no reasonable person would take the view adopted by the circuit court.”); Jenkins v. State, 375 Md. 284, 295-96 (2003) (“Abuse occurs when a trial judge exercises discretion in an arbitrary or capricious manner or when he or she acts beyond the letter or reason of the law.”). Matthews, 479 Md. at 305-06; see also Abruquah v. State, 483 Md. 637, 652 n.5 (2023) (an abuse-of-discretion analysis requires a reviewing court to determine the “outer bounds of what is acceptable expert evidence”). As the Supreme Court has explained, “the law grants a [trial] court the same broad latitude when it decides how to determine reliability as it enjoys in respect to its ultimate reliability determination.” Kumho Tire Co. v. Carmichael, 526 U.S. 137, 142 (1999) (emphasis in original). 23 III Discussion A. From Frye-Reed to Daubert-Rochkind 1. Frye-Reed Starting with Maryland’s 1978 adoption of the D.C. Circuit’s 1923 Frye general acceptance test, Maryland courts deciding the admissibility of expert testimony predicated on a novel scientific principle or discovery would determine whether the scientific principles or discovery were generally accepted in the relevant scientific community. Reed v. State, 283 Md. 374 (1978); see Rochkind, 471 Md. at 13 (collecting cases under the FryeReed regime). In 1993, in Daubert v. Merrell Dow Pharmaceuticals, Inc., the Supreme Court held that Federal Rule of Evidence 702 superseded Frye. The Court provided a non-exclusive list of factors that may be pertinent when determining whether the scientific testimony at issue is not only relevant but reliable. 509 U.S. 579, 589, 593-94 (1993). “The Daubert analysis, according to the Supreme Court, was more flexible than the ‘uncompromising [Frye] general acceptance test’ and gave trial courts greater discretion to admit scientific expert testimony that is relevant and founded on sound principles, even though novel or controversial.” Rochkind, 471 Md. at 14 (quoting Daubert, 509 U.S. at 596) (additional internal quotation omitted). Daubert allowed trial courts to admit a broader range of scientific testimony than would have been possible under Frye – including, for example, minority opinions within a field. General Electric Co. v. Joiner, 522 U.S. 136, 142 (1997). 24 But it did not uncritically throw the doors open to expert testimony, and instead it required the trial court to act as a “gatekeeper.” Id. Under the Frye-Reed regime, the proponent of the evidence typically showed the general acceptance of the scientific expert’s methodology “by surveying scientific publications, judicial decisions, or practical applications, or by presenting testimony from scientists as to the attitudes of their fellow scientists.” 1 MCCORMICK ON EVID. § 203.1 (8th ed. 2020) (Standards for admitting scientific evidence–The generalacceptance requirement). “Daubert, by contrast, refocus[ed] the attention away from acceptance of a given methodology … and centers on the reliability of the methodology used to reach a particular result.” Rochkind, 471 Md. at 31. While Daubert originally focused solely on the expert’s methodology (as opposed to their conclusions), the Joiner Court recognized that conclusions and methodology “are not entirely distinct from one another. Trained experts commonly extrapolate from existing data. But nothing in either Daubert or the Federal Rules of Evidence requires a district court to admit opinion evidence that is connected to existing data only by the ipse dixit of the expert. A court may conclude that there is simply too great an analytical gap between the data and the opinion proffered.” Joiner, 522 U.S. at 146. 25 2. Maryland Rule 5-702 This Court adopted Maryland Rule 5-702 in 1994, shortly after the Supreme Court issued Daubert but well before this Court fully embraced Daubert’s approach in Rochkind.6 Under Rule 5-702, [e]xpert testimony may be admitted, in the form of an opinion or otherwise, if the court determines that the testimony will assist the trier of fact to understand the evidence or to determine a fact in issue. In making that determination, the court shall determine (1) whether the witness is qualified as an expert by knowledge, skill, experience, training, or education, (2) the appropriateness of the expert testimony on the particular subject, and (3) whether a sufficient factual basis exists to support the expert testimony. The third “sufficient factual basis” prong includes two sub-factors. First, the expert must have available an adequate supply of data. Second, the expert must use a reliable methodology in analyzing that data. Matthews, 479 Md. at 309; Roy v. Dackman, 445 Md. 23, 42-43 (2015). Absent either of these factors, an expert opinion is “mere speculation or conjecture.” Matthews, 479 Md. at 309 (quoting Rochkind, 471 Md. at 22). 3. Rochkind After the consensus among the states had shifted to the Daubert regime, Maryland courts followed suit in 2020. Rochkind pointed to the problem with Frye that Daubert was 6 In the 40 years after Reed, Maryland courts experienced a “jurisprudential drift: the Frye-Reed standard announced in 1978 slowly morphed into a ‘Frye-Reed Plus’ standard, implicitly and explicitly relying on and adopting several Daubert principles.” Rochkind, 471 Md. at 5. 26 meant to solve: the court sought not just pure liberalization but instead to correct errors in both directions. “[U]sing acceptance as the only measure of reliability presents a conundrum: a generally accepted methodology may produce ‘bad science’ and be admitted, while a methodology not yet accepted may be excluded, even if it produces ‘good science.’” Rochkind, 471 Md. at 30. Maryland’s adoption of Daubert’s focus on reliability would “streamline the evaluation of scientific expert testimony under Rule 5-702.” Id. at 35. Under Rochkind, trial courts “should consider a number of factors in determining whether the proffered expert testimony is sufficiently reliable to be provided to the trier of fact.” Matthews, 479 Md. at 310. They are: (1) whether a theory or technique can be (and has been) tested; (2) whether a theory or technique has been subjected to peer review and publication; (3) whether a particular scientific technique has a known or potential rate of error; (4) the existence and maintenance of standards and controls; … (5) whether a theory or technique is generally accepted[;] […] (6) whether experts are proposing to testify about matters growing naturally and directly out of research they have conducted independent of the litigation, or whether they have developed their opinions expressly for purposes of testifying; (7) whether the expert has unjustifiably extrapolated from an accepted premise to an unfounded conclusion; (8) whether the expert has adequately accounted for obvious alternative explanations; 27 (9) whether the expert is being as careful as he [or she] would be in his [or her] regular professional work outside his [or her] paid litigation consulting; and (10) whether the field of expertise claimed by the expert is known to reach reliable results for the type of opinion the expert would give. Matthews, 479 Md. at 310-11 (summarizing Rochkind) (internal citations omitted). The Rochkind Court added several overarching items of guidance in adopting the Daubert standard: 1. The reliability inquiry is flexible. 2. Trial courts must focus “solely on principles and methodology, not on the conclusions that they generate,” although those are not entirely distinct and thus a trial court must consider the relationship between the two. 3. A trial court need not “admit opinion evidence that is connected to existing data only by the ipse dixit of the expert”; rather, a court may conclude that there is simply too great an analytical gap between the data and the opinion proffered. 4. All of the Daubert factors are relevant in the reliability inquiry, but none is dispositive, and a trial court may apply some, all, or none depending on the particular expert testimony at issue. 5. Rochkind did “not upend [the] trial court’s gatekeeping function. ‘Vigorous cross-examination, presentation of contrary evidence, and careful instruction 28 on the burden of proof are the traditional and appropriate means of attacking shaky but admissible evidence.’” See Matthews, 479 Md. at 311-12 (summarizing the Rochkind Court’s observations). 4. Matthews In Matthews, this Court for the first time post-Rochkind addressed under the Daubert standard whether a trial court erred in deciding the admissibility of expert testimony. Matthews, 479 Md. at 284. In attempting to solve a murder, the State enlisted the help of an FBI scientist who used a technique known as “reverse projection photogrammetry” to estimate the height of a suspect who had been caught on video but whose face was indiscernible. Id. at 288-89. The FBI scientist determined that the suspect on the video was approximately 5’8” tall, plus or minus two-thirds of an inch; defendant Matthews was approximately 5’9”. Id. But the FBI scientist’s expert report noted that, due to several variables, “the degree of uncertainty in this measurement could be significantly greater,” and at a pretrial hearing, the expert testified that she could not scientifically quantify several variables that might lead to uncertainty greater than two-thirds of an inch. Id. Nevertheless, the trial court admitted the expert testimony, and the jury convicted Matthews of the murder. Id. at 297, 304. The intermediate appellate court reversed Matthews’s conviction, reasoning that the inability to quantify the effect of the variables noted by the expert made her height measurement unreliable. The court perceived an “analytical gap” between the underlying data and the expert’s conclusion, and therefore held that the trial court abused its discretion in admitting the expert opinion testimony. Id. at 304-05. 29 This Court reinstated Matthews’s conviction, concluding that the expert’s methodology was reliable and finding no analytical gap in the expert’s proffered testimony. Matthews had argued that only one Daubert-Rochkind factor warranted exclusion of the testimony: the error rate, driven by the expert’s inability to provide an overall margin of error for her height estimate. This Court explained that “it is not sufficient to point to an unknown degree of uncertainty/error rate that applies to an expert opinion and claim that a trial court is necessarily stripped of discretion to admit that opinion.” Matthews, 479 Md. at 314. Instead, we distinguished between “uncertainty inherent in an expert’s methodology” and “uncertainty that applies to an expert’s conclusions following the application of a reliable methodology.” Id. at 315. Under the former circumstance of inherent uncertainty, a trial court more likely would exclude expert testimony due to “the unacceptably high risk of an inaccurate conclusion being reached in every case where the technique is used.” Id. at 316. However, the latter scenario, involving uncertainty in an expert’s conclusion following the application of a reliable methodology, “is generally less problematic than where an expert has applied a technique that is unreliable in every instance in which it is used.” Id. In Matthews, it was undisputed that the expert’s methodology was reliable, and that the uncertainty related to the expert’s conclusion following the application of that methodology. Thus, we recognized, the trial court was not required to exclude the testimony due to an inherently unreliable methodology. Id. at 317. However, we “emphasize[d] that just because the trial court was not required to exclude [an expert’s] testimony when [the expert] acknowledged an unknown degree of uncertainty, it does not 30 follow that the trial court was required to admit it.” Id. We further explained that, if the uncertainty applies to the expert’s conclusions, the “trial court should determine whether the uncertainty in the expert’s conclusions is the product of an analytical gap in the expert’s analysis and/or whether the uncertainty ultimately renders the opinion unhelpful to the trier of fact.” Id. at 314-15. If either of those circumstances exist, the trial court acts within its discretion in excluding the proffered testimony. Id. In Matthews’s case, we perceived no analytical gap in the expert’s testimony, where there was no disconnect between the results of the expert’s analysis and the expert’s opinion. Id. at 318. There was nothing illogical about the expert’s explanation that her analysis showed the subject’s height was 5’8” plus or minus two-thirds of an inch, although the margin of error might be greater based on other variables she could not quantify. Id. at 318. We also concluded that the trial court had acted within its discretion in finding that the expert’s testimony would “assist the trier of fact to understand the evidence or to determine a fact in issue” – in other words, that it was helpful, as required by the text of Rule 5-702. See id. at 319-23. First, the expert had explained her analysis in detail. Matthews, 479 Md. at 319. Second, the expert had explained why, even with the unknown degree of uncertainty attributable to certain variables, she remained comfortable with her height estimate, including the two-thirds inch margin of error. Id. at 319-20. Finally, the expert herself had stood in the same spot and position as the subject in the image, and was able to opine that the subject of the image was just slightly shorter than the expert herself, who was between 5’9” and 5’10”. Id. at 320. Together, these factors allowed the trial court 31 to reasonably conclude that the expert’s opinion would help the jury, despite the expert’s acknowledged uncertainty. Id. at 320-21. B. Applying Daubert-Rochkind Here The case centers on the primacy – and boundaries – of methodological reliability in the Daubert-Rochkind analysis. As we explain below, the trial court acted within its discretion in considering most of its points of concern at the Daubert-Rochkind hearing. However, the court made a significant error when it relied on Ms. Cardell’s June 2021 normalizing adjustments regarding trauma and on-call pass-through payments as a basis for excluding her testimony. We shall order a limited remand under Maryland Rule 8-604(d)(1) for the trial court to revisit its ruling without consideration of the June 2021 normalizing adjustments as reflecting on the reliability of Ms. Cardell’s methodology. 1. The Relationship Between Data and Methodology An expert witness generally arrives at an opinion by choosing a methodology, selecting data to which to apply the chosen methodology, and drawing conclusions based on the results of the application of the methodology. The reliability of the expert’s methodology is a core focus of a trial court under Rule 5-702. Without reliable methods (in addition to an adequate supply of data), an expert’s opinion lacks sufficient factual basis to support it and is instead “mere speculation or conjecture.” Matthews, 479 Md. at 309. The disagreement here largely is over the precise boundary between data and methodology when assessing the factual basis of an expert’s testimony. There are two competing visions. KatzAbosch and the trial court offer one vision, in which input-related choices an expert makes can be so central to a straightforward methodology that those choices 32 implicate the reliability of the methodology itself. As the trial court stated in its supplemental written memorandum, “[a]n expert’s choice in data sampling is at the heart of his methodology.” (Quoting CDW, 906 F. Supp. 2d at 824.) The before-and-after method for measuring a company’s lost profits is as widely accepted as it is simple: identify a benchmark period before an alleged harm event and compare the company’s profits during that benchmark period to the company’s profits in subsequent periods. The play in the joints arises out of the choices the expert makes in identifying the benchmark period and thereby deciding what will be the “before” period and what will be the “after” period, and in the choices the expert makes about how she will measure “profits.” Should the expert use pre- or post-“guaranteed payment” figures? Should she apply the before-and-after method to a medical practice’s financial performance without regard to changes in reimbursement rates between years? In KatzAbosch’s and the trial court’s view, these considerations concerning data differ from consideration of the soundness of the data that an expert uses in applying their methodology, or the soundness of the data from which an expert extrapolates after completing the application of the methodology.7 7 Consider the use of inputs where their selection does not depend on judgment calls in the way that the trial court viewed the selection of the benchmark year in this case. For example, imagine hypothetically that KatzAbosch did not dispute that 2015 was the proper benchmark year and agreed that Ms. Cardell used all the correct sources of data in her calculation of lost profits, but KatzAbosch called the veracity of some of the data into question (perhaps suggesting that bookkeepers at PNSI had not accurately classified various expenses in the company’s books and records). In that instance, KatzAbosch would have been complaining about the soundness of the data to which Ms. Cardell applied a concededly reliable methodology. The veracity of the data in this hypothetical is a factual question that the expert has no real impact on; instead, it is up to the jury to decide whether the expert’s reliable methodology has been applied to accurate data or whether the 33 PNSI and the Appellate Court offer a more rigid vision, arguing that anything dealing with the inputs is a “data” question to be siloed off from the methodological reliability analysis and instead left to the jury.8 They acknowledge that the line between data and methodology (and, for that matter, conclusions) is blurry, but they draw the line sharply in this case, arguing that only the core before-and-after subtraction counts as methodology, and the rest is data (e.g., why 2015?) or conclusion. In this regard, PNSI relies primarily on the Seventh Circuit’s decision in Manpower, Inc. v. Ins. Co. of Pa., 732 F.3d 796 (7th Cir. 2013), noting that we cited Manpower in Matthews.9 In Manpower, the federal district court excluded insured-plaintiff Manpower’s accounting expert, whom Manpower needed to establish business interruption damages in its suit against the insurer-defendant. The insurer-defendant challenged the reliability of the expert’s methodology, and the district court found that the expert had followed the master insurance policy’s “straightforward” methodology for his calculations. Manpower, 732 F.3d at 801-02. But the calculations’ reliability nevertheless “turn[ed] on whether [the expert] used reliable methods when selecting the numbers used in his calculations.” Id. at proponent has offered “garbage in, garbage out” expert testimony to be discredited even after being admitted. 8 For example: “Whether Ms. Cardell failed to consider reimbursement rates is not an issue with the methodology – the before-and-after method. Rather, it is an issue with the soundness of the data she used to reach her conclusion.” Parkway Neuroscience, 255 Md. App. at 627. 9 See Matthews, 479 Md. at 316 (quoting Manpower, 732 F.3d at 806: “The district court usurps the role of the jury, and therefore abuses its discretion, if it unduly scrutinizes the quality of the expert’s data and conclusions rather than the reliability of the methodology the expert employed.”). 34 801. The district court found that the answer to that question was no. The expert used a shorter base period (from which he extrapolated a relatively high growth rate) because of recent corporate acquisitions, new policies, and new managers at Manpower, and the expert did not consider other indicators that may have affected the growth rate. Id. The district court also criticized the expert’s reliance on conversations with the company managers. Id. But the district court primarily found that the expert’s “‘analysis [broke] down’ at his choice of growth rate.” Id. In the district court’s view, “[o]nly a more thorough analysis of the reasons for the growth would have supported [the expert’s] choice of a projected growth rate.” Id. The Seventh Circuit reversed the district court’s exclusion of the expert, agreeing with Manpower that “the district court exercised its gatekeeping role under Daubert with too much vigor.” Id. at 805. The federal appellate court opined that reliability is primarily a question of the expert’s methodology, not the data that is the input or the conclusions that are the output. Id. at 806. The Manpower Court acknowledged the blurriness of the line between undue scrutiny of an expert’s data and proper scrutiny of an expert’s methodology, writing that “[t]he critical inquiry is whether there is a connection between the data employed and the opinion offered.” Id. Opinions are properly excluded when they are connected to existing data “only by the ipse dixit of the expert.” Id. More specifically, the Manpower Court said that the district court should have stopped its inquiry when it found the methodology reliable, but instead it “drilled down to a third level in order to assess the quality of the data inputs [the expert] selected” and then took issue with that data selection. Id. at 807. The expert’s data selection was “substantially 35 more nuanced and principled than the district court’s characterization reflects,” and so the ultimate opinion was not ipse dixit but rather “reasoned and founded on data.” Id. at 809. Each of the district court’s criticisms of the expert’s reliability “was a comment on the soundness of the factual underpinnings of his calculation” and the district court’s rulings could even have been something of a “roadmap for [the defendant-insurer’s] crossexamination of [the expert].” Id. “But the district court supplanted that adversarial process with its admissibility determination” and “set the bar too high and therefore abused its discretion.” Id. The Manpower Court’s understanding of the line between data and methodology, of course, does not bind this Court, and its rigidity is not in keeping with the approach taken by other federal appellate courts. Instead, many federal courts have explained the Daubert standard in ways that reject this sharp line and acknowledge that problems with data and data selection (which itself can involve its own methodology) can bear on admissibility before the judge and not just weight before the jury. See, e.g., EEOC v. Freeman, 778 F.3d 463, 467, 472 (4th Cir. 2015);10 Elcock v. Kmart Corp., 233 F.3d 734, 10 In Freeman, plaintiff EEOC’s industrial psychologist expert witness delivered a report based on a database riddled with errors, introducing even more errors in an updated analysis; the “sheer number of mistakes and omissions” rendered the expert’s analysis “outside the range where experts might reasonably differ” under Kumho Tire, and so the district court did not abuse its discretion in excluding the proffered expert’s testimony as unreliable. The EEOC unsuccessfully argued that “the issue of the reliability of an expert’s data is always a question of fact for the jury, except perhaps in some theoretical, rare case.” Freeman, 778 F.3d at 472 (Agee, J., concurring). 36 755-56 (3d Cir. 2000);11 In re Mirena IUS Levonorgestrel-Related Products Liab. Litig. (No. II), 982 F.3d 113, 123 (2d Cir. 2020).12 Consider as one example Rink v. Cheminova, an Eleventh Circuit case in which the appellate court found no abuse of discretion in the district court’s exclusion of a putative class’s expert. 400 F.3d 1286, 1293-94 (11th Cir. 2005). The plaintiffs maintained that Cheminova’s pesticide malathion defectively contained elevated levels of isomalathion, which makes malathion particularly toxic to humans and which, they claimed, was created by exposure of the malathion supply to temperatures above 77 degrees Fahrenheit during storage at sites in Texas, Georgia, and Florida. Id. at 1289. The plaintiffs’ expert estimated the isomalathion content of pesticides used in Tampa; he began with National Weather Service temperature readings near the malathion storage sites, using the recorded high and low temperatures as upper and lower limits of the probable exposure temperatures. Id. Because there was evidence that the inside of the Texas storage facility was actually 18 degrees warmer than the outside ambient air temperature, he added 18 degrees to the 11 In Elcock, the Third Circuit held that the district court abused its discretion in admitting the expert’s economic damages model, which relied on empirical assumptions not supported by the record, fearing that a jury would be “likely to adopt the gross figure advanced by a witness who has been presented as an expert.” 233 F.3d at 755-56. 12 In Mirena, the Second Circuit rejected the plaintiffs’ contention that a trial court erred by taking a “hard look” at their expert’s methodology: “[A]n expert’s methodology must be reliable at every step of the way, and in deciding whether a step in an expert’s analysis is unreliable, the district court should undertake a rigorous examination of the facts on which the expert relies, the method by which the expert draws an opinion from those facts, and how the expert applies the facts and methods to the case at hand.” 982 F.3d at 123 (internal quotation marks and citation omitted) (emphasis added by the Mirena Court). 37 upper plausibility limits. Id. at 1290. He then averaged the upper and lower limits to find the most probable exposure temperature. Id. He took that temperature and plugged it into an equation to calculate the level of toxic isomalathion as a function of time and temperature. Id. Importantly, the plaintiffs’ theory hinged on the isomalathion content of the pesticides used in Tampa, Florida, which included not just the malathion that had initially been stored in Texas, but also malathion that had been initially stored at sites in Florida and Georgia. See id. at 1289-90. The district court excluded the expert’s testimony for several reasons, including the expert’s “method of extrapolating data from one site [Texas] to another [Florida and Georgia] without making particularized findings which accounted for the differences in conditions and length of storage at each site.” Id. at 1290. On appeal, the plaintiffs argued that the district court had improperly taken issue with the expert’s temperature data and not his methodology. The Eleventh Circuit disagreed: While [the plaintiffs] suggest that the only methodology at issue was [the expert’s] use of an equation to determine isomalathion levels, this argument belies the fact that [the expert] employed two methodologies: first, he employed certain methods of extrapolation and transposition to arrive at temperature data; and second, he inserted the temperature data into an equation to arrive at the level of isomalathion in the [pesticide]. As we have explained, the district court’s exclusion of [the expert] was based on its rejection of his methodology to derive temperature data, not the data itself. Thus, the district court’s exclusion of [the expert] can be distinguished from a situation in which an exclusion is based on a district court’s refusal to credit hard data arrived at by unassailable methods.… Here, the data [the expert] produced was driven by the methodology he used, and thus the district court’s inquiry into how he arrived at the data is not inappropriate 38 considering that the district court is charged with evaluating an expert’s methodology. Id. at 1293.13 The Rink Court took an appropriately holistic view of the methodology being applied, examining the production of the equation’s inputs as part of the process itself. A narrower understanding of the methodology – something along the lines of “use this formula to calculate the isomalathion content of this pesticide” – might have called for allowing the expert to testify at trial, where opposing counsel could have impeached him by showing the weaknesses in his Tampa storage temperature input selection. We agree with the Eleventh Circuit that, under a proper application of Daubert, the trial court in Rink acted as an appropriate gatekeeper. Another useful case to consider is In re Wholesale Grocery Products Antitrust Litigation, an Eighth Circuit antitrust case in which the district court had excluded an expert witness who sought to establish the plaintiff’s injury by selecting a competitive benchmark. 946 F.3d 995 (8th Cir. 2019). The district court excluded as unreliable the benchmark the expert selected, because the expert’s choice of a non-independent chain grocery store (Stop & Shop) was premised on an unfounded assumption that independent retailers’ charges 13 The Rink Court contrasted the case before it with Quiet Tech. DC-8, Inc. v. HurelDubois UK Ltd., 326 F.3d 1333 (11th Cir. 2003), where the district court had admitted testimony of Hurel-Dubois’s expert on computational fluid dynamics. Rink, 400 F.3d at 1293. After losing at trial, Quiet Tech argued that the district court abused its discretion by admitting the other side’s expert testimony. The Eleventh Circuit disagreed, observing that Quiet Tech “does not argue that it is improper to conduct a [computational fluid dynamics] study using the sorts of aerodynamic data that [the expert] employed, but rather that the specific numbers that [the expert] used were wrong. Thus, the alleged flaws in [the expert’s] analysis are of a character that impugn the accuracy of his results, not the general scientific validity of his methods.” Quiet Tech, 326 F.3d at 1343-44. This case resembles Rink, whereas the hypothetical we pose above in footnote 7 resembles Quiet Tech. 39 followed the same pattern as Stop & Shop’s. Id. at 999. The plaintiff argued on appeal that benchmarking is a recognized tool for establishing antitrust injury, and once the expert chose a benchmark, there was nothing left for the district court to analyze under Daubert; what remained were questions of fact for the jury. Id. at 1001. The Eighth Circuit rejected this argument “because it is the foundation of the assumption underlying the application of the method employed by [the expert] on these facts … that led the district court to its conclusion that [the expert’s] testimony should be excluded under Rule 702.” Id. “The district court did not go ‘miles beyond’ its appropriate role in this case, as [the plaintiff] argues, but rather held that the reasoning underlying [the expert’s] testimony was not on solid footing because the assumption upon which the report relied was insufficient to validate his opinion.” Id. at 1002. “[T]he district court held that ultimately these analyses were a house of cards of sorts and our own analysis reveals no abuse of discretion in this conclusion. At its base level, the core assumption of the analysis was by the ipse dixit of [the expert].” Id. Here too, we see the same arguments as in Manpower and in the case before us: undue scrutiny and a tunnel vision focus on methodology. But the Eighth Circuit declined to accept the benchmark selection without question and instead approved of the district court’s inquiry into the assumptions underlying the expert’s choice of benchmark. This is a long way from Manpower’s approach. In short, whether an expert’s methodology is sufficiently reliable to admit the expert’s testimony at trial will sometimes require a trial court to consider data and assumptions that the expert has employed in deciding threshold points relating to the methodology. Manpower’s rigid separation of “data” and “methodology” misses this grey 40 area and creates a categorical rule when the Daubert-Rochkind regime calls for flexibility and deference. When this Court cited Manpower in Matthews, it did so to demonstrate the proposition that where a sufficient factual basis exists under Rule 5-702(3) and DaubertRochkind – that is, where an expert has applied a reliable methodology to an adequate supply of data – courts should not exclude an expert merely because the expert’s particular conclusions may be inaccurate, but rather should only exclude expert testimony that is “mere speculation or conjecture.” Matthews, 479 Md. at 316. The Appellate Court in this case relied on the part of Manpower to which we do not subscribe, categorizing the 2015 base year choice, the failure to consider reimbursement rates, etc., as arguable defects in the soundness of Ms. Cardell’s data, rather than defects of her methodology. See Parkway Neuroscience, 255 Md. App. at 627-28 & n.9. The Appellate Court was right to center methodology in its analysis of the expert opinion’s reliability, a critical aspect (along with the data’s adequacy) of the opinion’s factual basis without which the opinion would be “mere speculation or conjecture.” Matthews, 479 Md. at 309 (citing Rochkind, 471 Md. at 22); see Parkway Neuroscience, 255 Md. App. at 629 & n.11. But the intermediate appellate court, relying on Manpower, took an overly rigid approach in analyzing the relationship between data and methodology, holding that the trial court’s analysis of Ms. Cardell’s opinion went to her data rather than to her methodology. Just as the U.S. Supreme Court has noted the blurred line between methodology and conclusions, Joiner, 522 U.S. at 146, we note the sometimes blurred line between data and methodology. Trial courts must not transmute all questions of data’s provenance or veracity 41 into questions of methodology, just as they must not – under Joiner – transmute all disagreements with conclusions into disagreements with methodology. But by the same token, trial courts should not wear “methodology blinders” and deny the existence of some limited overlap between data and methodology (and between methodology and conclusions). Determining whether a dispute concerning expert testimony implicates the soundness of data or soundness of methodology is precisely the type of matter that calls for the exercise of a trial court’s discretion. Here, the trial court explicitly noted speculative, insufficiently substantiated judgment calls that were central to Ms. Cardell’s application of the before-and-after method. In the trial court’s estimation, Ms. Cardell exercised subjective judgment in settling on the 2015 base year – that is, it was impossible to test the validity of that decision. Relatedly, the trial court was troubled by the effect on the base year determination of Ms. Cardell’s treatment of member draws as expenses without understanding whether that decision was consistent with any industry standard. In addition, the trial court faulted Ms. Cardell for failing to appropriately factor in confounding variables (in particular, declining insurance reimbursements) into her methodology. The court determined these judgment calls veered toward speculation and conjecture and ate away at the factual basis for Ms. Cardell’s opinions, because even though many of the problems had to do with the “inputs” to the before-and-after method, the selection of those inputs is central to the reliability of the method itself. We rely on trial courts that conduct Daubert-Rochkind hearings to determine where the line between data and methodology is in the specific cases before them, and whether 42 the proffered expert’s choices relating to data, assumptions, and other inputs implicate the reliability of the expert’s methodology. In this part of the trial court’s analysis, the court did just that. That is, we discern no error in the trial court’s analysis of Daubert-Rochkind factors one and two and four through nine.14,15 14 The trial court discussed Ms. Cardell’s decision not to account for the doctors’ individual revenue generation capabilities as part of the court’s discussion of Daubert factor eight (accounting for obvious alternative explanations). As discussed below, we think the court’s concerns on that front go more to the reliability of the before-and-aftermethod when used to calculate lost profits of a limited liability company with a small number of revenue generators, rather than a failure to account for obvious alternative explanations. 15 The trial court said that, at the Daubert hearing stage, it was not the court’s responsibility to determine whether Ms. Cardell qualified as an expert or not. And the court explicitly stated that it was not deciding one way or the other whether Ms. Cardell was, in fact, qualified to render an expert opinion at trial. However, the court did note Ms. Cardell’s lack of experience with specialty medical practices and her reliance on oral communications with PNSI staff (rather than universally available, concrete sources of information), and her reliance on her own experiences and judgment rather than on industry standards. In the court’s mind, these points undermined Ms. Cardell’s overall reliability. We do not discern any abuse of discretion in the court’s reliance on these points. However, it is important to acknowledge that there is nothing inherently problematic with an expert’s use of information provided orally or in writing by the proponent of her testimony. Experts routinely rely on such information in arriving at their opinions. Nor is it uncommon for experts to apply their subjective judgment, based on their training and experience, in formulating their opinions. The trial court did not indicate that it believed it was per se improper for an expert to rely on oral communications with the proponent of her testimony or to make subjective judgment calls. Rather, the trial court’s point was that, in the absence of other sources of information that tended to support the reliability of Ms. Cardell’s methodology – such as industry standards – and given her inability to adequately explain why she made the judgment calls she did, Ms. Cardell’s reliance on oral communications with PNSI and on her subjective judgment detracted from the reliability of her methodology. 43 2. The Federal Analog to Maryland Rule 5-702 The direction of analogous Federal Rule 702 confirms our understanding of meaningful gatekeeping as to an expert opinion’s factual basis. In a May 2022 report, the Advisory Committee on Evidence Rules wrote that the Committee resolved to respond to the fact that many courts have declared that the reliability requirements set forth in Rule 702(b) and (d) – that the expert has relied on sufficient facts or data and has reliably applied a reliable methodology – are questions of weight and not admissibility, and more broadly that the expert testimony is presumed to be admissible. These statements misstate Rule 702, because its admissibility requirements must be established to a court by a preponderance of the evidence. The Committee concluded that in a fair number of cases, the courts have found expert testimony admissible even though the proponent has not satisfied the Rule 702(b) and (d) requirements by a preponderance of the evidence – essentially treating these questions as ones of weight rather than admissibility…. COMM. ON RULES OF PRAC. AND PROC. OF THE JUD. CONF. OF THE U.S., REP. OF THE ADVISORY COMM. ON EVIDENCE RULES 6 (May 15, 2022), available at https://perma.cc/PK3B-Q8G5. Absent any contrary Congressional action, Federal Rule 702 will officially reflect this reality come December 1, 2023, when a set of amendments will take effect to provide that the proponent of expert testimony must meet Rule 702’s standards – including related to the testimony’s factual basis – by a preponderance of evidence, or else the testimony is inadmissible and may not go to the jury: A witness who is qualified as an expert by knowledge, skill, experience, training, or education may testify in the form of an opinion or otherwise if the proponent demonstrates to the court that it is more likely than not that: (a) the expert’s scientific, technical, or other specialized knowledge will help the trier of fact to understand the evidence or to determine a fact in issue; 44 (b) the testimony is based on sufficient facts or data; (c) the testimony is the product of reliable principles and methods; and (d) the expert’s opinion reflects a reliable application of the principles and methods to the facts of the case. U.S. SUPREME COURT, ORDER 4 (April 24, 2023), available at https://perma.cc/RU2SKEYM (most relevant new language emphasized). The change emphasizing the preponderance standard “specifically was made necessary by the courts that have failed to apply correctly the reliability requirements of [Federal Rule 702].” FED. R. EVID. 702 advisory committee’s note to 2023 amendment. “[M]any courts have held that the critical questions of the sufficiency of an expert’s basis, and the application of the expert’s methodology, are questions of weight and not admissibility. These rulings are an incorrect application of Rules 702 and 104(a).” Id.; see Sardis v. Overhead Door Corp., 10 F.4th 268, 283-84 (4th Cir. 2021) (observing that the then-proposed amendments to Federal Rule 702 would make explicit the preponderance of evidence standard of admissibility to the rule’s sufficiency of basis and reliability analyses; confirming that these rule revisions and clarifications “clearly echo[] the existing law on the issue” from Daubert, Kumho Tire, and Rule 702 itself). The new amendments comprehend that some challenges to expert testimony will, in fact, go to weight rather than admissibility. For example, if the court finds it more likely than not that an expert has a sufficient basis to support an opinion, the fact that the expert has not read every single study that exists will raise a question of weight and not admissibility. But this does not mean, as certain courts have held, that arguments about the sufficiency of an expert’s basis always go to weight and not admissibility. Rather it means that once the court has found it more likely 45 than not that the admissibility requirement has been met, any attack by the opponent will go only to the weight of the evidence. FED. R. EVID. 702 advisory comm. note to 2023 amendment (emphasis added). Indeed, they do not require federal courts to “nitpick an expert’s opinion in order to reach a perfect expression of what the basis and methodology can support,” but instead seek to block “claims that are unsupported by the expert’s basis and methodology.” Id. 3. The Trial Court’s Error Although we discern no error in most of the trial court’s application of the DaubertRochkind factors, we are constrained to conclude that the court erred in one respect. The trial court viewed Ms. Cardell’s June 2021 updates, which the court discussed in the context of Daubert-Rochkind factors three (known or potential rate of error) and 10 (whether the field of expertise is known to reach reliable results for the projected type of expert opinion), as implicating the reliability of her methodology. To be sure, the court observed that the error rate factor did not apply “as it was ordinarily considered,” e.g., concrete and numerical probabilities of correctness in DNA testing. Still, the court was troubled by the timing of Ms. Cardell’s updates – without new information and “for fully subjective reasons” – which, the court believed, reflected negatively on her methodology. This analysis missed the mark. In fact, there was new information of a sort: Ms. Cardell noticed something she had not noticed before on first examination. She then sought clarification and revised her opinion, just as a doctor might order a biopsy and diagnose a patient with skin cancer if the doctor had missed a mole upon first examination of the patient. Catching something peculiar the second time around neither undermines the 46 adequacy of the data (the patient’s skin) nor the court’s understanding of the expert’s methodology (examining the patient’s skin for disease indicators). Notably, when discussing the tenth Daubert factor – whether the field of expertise is known to reach reliable results for the type of opinion the expert would give – the trial court specifically referenced the June 2021 adjustments, stating that “the mere fact the findings changed in June of this year for fully subjective reasons, it had nothing to do with any new information. It kind of makes the whole reliability even that much more suspect.” This comment, along with the court’s other references to the June 2021 adjustments, leaves us with the abiding concern that, in this instance, the trial court strayed from its gatekeeping role and that this error was significant in the court’s overall analysis.16 Ms. Cardell’s decision to make the adjustments relating to trauma/on-call payments in 2016 did not implicate the reliability of her methodology. At most, it went to the care with which she 16 The trial court also erred when it stated that Ms. Cardell’s lost-profit calculations would not be helpful to the jury. The court reached this conclusion based on Ms. Cardell’s failure to consider each departing physician’s projected revenue generation in her calculation of lost profits. The court’s point was that, if the jury concluded that one (or more) of the departing physicians left for reasons other than KatzAbosch’s negligence, a lost-profits analysis that included that physician’s projected revenue would overstate the practice’s lost profits and therefore not help the jury to accurately calculate damages. The court stated, however, that its finding with respect to helpfulness was a “very, very slight factor” in its decision to exclude Ms. Cardell’s testimony. We confirmed in Matthews that a trial court has discretion under Maryland Rule 5-702 to exclude an expert’s opinion based on a reliable methodology if the court nevertheless concludes that the expert’s testimony would not be helpful to the jury. See Matthews, 479 Md. at 320-21. Here, however, there was no way to know before the trial whether the jury would conclude that any of the physicians left for reasons other than KatzAbosch’s negligence. The trial court’s conclusion that Ms. Cardell’s testimony would not be helpful, therefore, was speculative and not a proper ground for exclusion. 47 applied her methodology, which is a matter to be explored on cross-examination before the jury (if Ms. Cardell’s testimony is otherwise found to be sufficiently reliable).17 In Matthews, we “emphasize[d] that just because the trial court was not required to exclude [the expert’s] testimony when [the expert] acknowledged an unknown degree of uncertainty, it does not follow that the trial court was required to admit it.” 479 Md. at 317. Similarly, we do not believe that, based on the record before the trial court in this case, the court was required to exclude Ms. Cardell’s testimony. Rather, based on the factors that the trial court appropriately considered in this case, it was within the trial court’s discretion to admit or exclude Ms. Cardell’s testimony. Having carefully reviewed the record, we conclude that the fair and prudent course of action at this point is to order a limited remand to the circuit court under Maryland Rule 8-604(d)(1) so that the trial court may decide to admit or exclude Ms. Cardell’s testimony without consideration of her June 2021 normalizing adjustments as reflecting on the reliability of Ms. Cardell’s methodology.18 The trial court may make that decision based on the existing record or, in its discretion, may allow further examination of Ms. Cardell and/or other witnesses before issuing a new 17 The trial court elsewhere in its analysis of the Daubert factors said that it had “no reason to think that [Ms. Cardell] blew this off as an inconsequential project, I mean she took this very seriously.” 18 We do not mean to suggest that, in every case where an appellate court concludes that part of a trial court’s Daubert ruling was based on proper factors and another part was not, a limited remand to the trial court is necessary. We expect that, in many cases, it will be clear from the record whether the trial court would have admitted or excluded the expert testimony without consideration of a factor that is later determined to have been improper on appeal. We encourage trial courts to make such matters explicit on the record when possible. 48 ruling.19 In the trial court’s discretion, it also may permit the parties to submit additional written and oral arguments prior to issuing its ruling. The trial court should provide a written explanation of its decision. We shall retain jurisdiction over this case. After the trial court issues its decision on remand, we shall issue an appropriate Order. 19 The trial court’s criticism of Ms. Cardell’s failure to consider the remaining and departing doctors’ inherent ability to generate revenue when discussing Daubert factor eight (accounting for obvious alternative explanations) is more a criticism of the reliability of the before-and-after methodology itself as applied to a limited liability company with a small number of revenue-generating members, than it is a criticism of Ms. Cardell’s failure to account for obvious alternative explanations in applying the before-and-after methodology. In any business with a small number of revenue-generating members, the firm’s overall performance in a particular year may well be the result of some members having unusually strong or unusually weak years, compared to their individual performances over time. As Ms. Cardell described the before-and-after method of calculating lost profits, there is no consideration of how individual revenue generators perform over time. Rather, Ms. Cardell testified, “under the before and after methodology what happens is that the damages expert looks at the [company’s] performance in two different periods. The first being what we called the benchmark period or the before period. And that’s the period that is unaffected by whatever the alleged harm event, breach is. And then that is [compared] to the after period or the loss period, which is the period that is affected by whatever the alleged harm, breach, or event is. This methodology essentially calculates what the [company’s] profits would have been but for again that alleged breach, harm event.” The trial court did not question the general acceptance of the before-and-after analysis as a reliable methodology for calculating lost profits. Nor did the trial court explicitly state that the before-and-after methodology, as described by Ms. Cardell, cannot reliably be applied to a limited liability company with a small number of revenuegenerating members. However, that seems to be the import of the concerns the trial court articulated when discussing Ms. Cardell’s failure to consider the individual revenuegenerating abilities of the physicians who departed, as well as the two physicians who remained at PNSI. On remand, in its discretion, the trial court may give additional consideration to whether the before-and-after methodology, as a general matter, is a reliable methodology for calculating lost profits of an entity such as PNSI. We express no view concerning the answer to that question. 49 IV Conclusion The trial court, in its gatekeeping role under Daubert-Rochkind, acted within its discretion in analyzing the data and other inputs and assumptions that implicated the reliability of Ms. Cardell’s methodology. However, the court improperly considered Ms. Cardell’s June 2021 normalizing adjustments relating to trauma/on-call payments as reflecting on the reliability of Ms. Cardell’s methodology. We order a limited remand to the circuit court to allow that court to decide whether to admit or exclude Ms. Cardell’s expert testimony without such consideration of the June 2021 normalizing adjustments. JUDGMENT OF THE APPELLATE COURT OF MARYLAND VACATED; CASE REMANDED TO THAT COURT WITH THE DIRECTION TO REMAND THE CASE TO THE CIRCUIT COURT FOR HOWARD COUNTY, WITHOUT AFFIRMING OR REVERSING THE JUDGMENT OF THE CIRCUIT COURT, FOR FURTHER PROCEEDINGS CONSISTENT WITH THIS OPINION. COSTS TO ABIDE. 50 Circuit Court for Howard County Case No. C-13-CV-18-000181 Argued: May 4, 2023 IN THE SUPREME COURT OF MARYLAND* No. 30 September Term, 2022 KATZ, ABOSCH, WINDESHEIM, GERSHMAN & FREEDMAN, P.A., ET AL. v. PARKWAY NEUROSCIENCE AND SPINE INSTITUTE, LLC Fader, C.J., Watts, Hotten, Booth, Biran, Gould, Eaves, JJ. Concurring Opinion by Booth, J. Filed: August 30, 2023 * During the November 8, 2022 general election, the voters of Maryland ratified a constitutional amendment changing the name of the Court of Appeals of Maryland to the Supreme Court of Maryland. The name change took effect on December 14, 2022. I agree with and join the Majority’s well-written opinion in this case. I write separately to respond to the Majority’s invitation to “reflect on [the] flexibility and deference” due to courts analyzing the admissibility of expert testimony. Maj. Slip Op. at 1. I have observed that our traditional formulation of the abuse of discretion standard is not the best or most accurate way of describing our abuse of discretion review in the context of reviewing expert testimony admissibility determinations, and suggest that this Court reformulate the definition of our abuse of discretion standard in the context of appellate review of expert witness testimony admissibility determinations. Since this Court’s adoption of Daubert in 2020, we have reviewed for abuse of discretion four cases involving trial courts’ decisions to admit or preclude expert testimony: State v. Matthews, 479 Md. 278 (2022); Abruquah v. State, 483 Md. 637 (2023); Oglesby v. Baltimore School Associates, No. 26 Sept. Term, 2022, 2023 WL 4755689 (July 26, 2023); and this case. The Court’s decisions concerning the admissibility of expert testimony in Matthews and Abruquah were not unanimous ones.1 Not only have there been 1 In State v. Matthews, 479 Md. 278, 325–40 (2022) (Watts, J., dissenting), Justice Watts filed a dissenting opinion stating that she would have upheld the Appellate Court’s conclusion that the trial court abused its discretion in admitting the State’s expert witness concerning her opinions involving reverse projection photogrammetry. In Abruquah v. State, 483 Md. 637, 699–711 (2023), Justice Hotten filed a dissenting opinion, joined by Justice Gould and Justice Eaves, in which those justices would have determined that the trial court did not abuse its discretion in permitting the unqualified opinions expressed at trial by the firearms expert. In a separate dissenting opinion, Justice Gould criticized the Majority’s application of the abuse of discretion standard of review, asserting that the Majority “sidestep[ped] the deferential standard of review by recasting its decision as establishing the outer bounds of what is acceptable expert evidence in [the] area [of firearms identification].” Abruquah, 483 Md. at 712 (quotations omitted) (Gould, J. dissenting). different views concerning the circuit courts’ discretion in making expert witness admissibility determinations among jurists on this Court, it is also notable that, in three of the four above-referenced cases, this Court reached opposite conclusions from the Appellate Court of Maryland undertaking the same review.2 In Matthews—our maiden voyage in appellate review of a trial court’s decision to admit expert testimony under the application of the Daubert-Rochkind factors—we “reaffirmed” the sentiments expressed in Rochkind, that “it is still the rare case in which a Maryland trial court’s exercise of discretion to admit or deny expert testimony will be overturned.” 479 Md. at 306. Given our post-Rochkind batting average, I am not sure that sentiment holds true. I was one of the members of this Court who voted in Rochkind to adopt the Daubert standard. I joined the majority opinions written by my colleagues in Matthews, Abruquah, Oglesby, and the instant case, and I agree with our analysis in each of them. That said, with some time to reflect on the Court’s application of the abuse of discretion standard in the context of appellate review of a trial court’s decision to admit or deny expert testimony 2 See Matthews v. State, 249 Md. App. 509 (2021) (reversing a criminal defendant’s second-degree murder conviction after determining that the circuit court’s decision to admit the photogrammetry expert’s opinion about the height of a suspect captured in a surveillance video was inadmissible, and error was not harmless); Oglesby v. Baltimore Sch. Assocs., No. 130 Sept. Term, 2021, 2022 WL 3211044 (App. Ct. Md. Aug. 9, 2022) (holding that the circuit court did not abuse its discretion in excluding a plaintiff’s causation expert in a lead paint case, or err in granting summary judgment in light of the exclusion of the causation testimony); Parkway Neuroscience and Spine Inst., LLC v. Katz Abosch, Windesheim, Gershman & Freeman, P.A., et. al., 255 Md. App. 596 (2022) (holding that the trial court abused its discretion in excluding Ms. Cardell’s opinions concerning lost profits, and therefore, erred in granting summary judgment in favor of the defendants). We do not know whether the Appellate Court would have considered the application of the Daubert-Rochkind factors in the same manner as this Court in Abruquah because this Court granted certiorari while the case was pending in that court. See Abruquah, 483 Md. at 652. 2 in its consideration and application of the Daubert-Rochkind factors, I have some unease about our recitation of our traditional abuse of discretion formulation, which we developed and apply in other contexts. I observe that this formulation appears to be inconsistent with the abuse of discretion standard employed by the federal courts in the Daubert context, as well as the careful and searching examination that this Court is conducting in reviewing these cases. For the reasons expressed below, I believe that when this Court applies abuse of discretion when reviewing expert witness admissibility determinations, we should articulate an abuse of discretion standard that is in line with the federal courts’ formulation and that reflects this Court’s current practice. When we adopted the Daubert standard, we adopted it in full. That necessarily includes the federal courts’ application of the abuse of discretion standard. I explain my reasoning more fully below. A. Rochkind—This Court’s Decision to Adopt Daubert In 2020, Maryland joined the supermajority of states that adopted Daubert3 as the standard for admission of expert testimony. See Rochkind v. Stevenson, 471 Md. 1 (2020). The change to the Daubert standard shifted the focus of the analysis from the general acceptance of a methodology under Frye-Reed4 to the reliability of a methodology. Rochkind, 471 Md. at 5, 31; Daubert, 509 U.S. at 594–95. In doing so, this Court adopted the Supreme Court’s Daubert trilogy for analyzing admissibility of expert opinions. See Daubert, 509 U.S. 579, 588–89 (replacing the “general acceptance” test under Frye, with 3 Daubert v. Merrell Dow Pharmaceuticals, Inc., 509 U.S. 579 (1993). 4 Before the adoption of Daubert in Maryland, Frye-Reed was the prevailing standard for admissibility of expert evidence. See Reed v. State, 283 Md. 374 (1978). 3 the more flexible standard in determining whether scientific evidence is reliable and admissible); General Electric Co. v. Joiner, 522 U.S. 136, 139 (1997) (clarifying that the proper scope of appellate review of a trial court’s rulings on expert admissibility is “abuse of discretion”); Kumho Tire Co., Ltd. v. Carmichael, 526 U.S. 137, 141–42 (1999) (clarifying that the trial judge’s Rule 702 gatekeeping duties apply to all expert testimony, whether such testimony is based upon scientific, technical, or other specialized knowledge). In adopting the Daubert standard, we expressed doubt that adopting the standard would “upend Maryland evidence law[,]” and we observed that, by adopting Daubert, “Maryland courts will be able to ‘draw from and contribute to the broad base of case law grappling with scientific testimony.’” Rochkind, 471 Md. at 34–35 (quoting Savage v. State, 455 Md. 138, 185 (2017) (Adkins, J., concurring)). Prior to Rochkind, Maryland appellate courts reviewed a trial court’s decision concerning the admissibility of expert testimony under two different standards of review— conducting a de novo review of a trial court’s determinations under Frye-Reed and reviewing the trial court’s determinations under Rule 5-702 for abuse of discretion. See Rochkind, 471 Md. at 37. With the change to Daubert, we recognized that appellate review of a trial court’s decision to admit or exclude expert opinion testimony would be the abuse of discretion standard. Id. In abrogating the Frye-Reed standard in favor of Daubert, “we reiterated that a trial court’s ruling to admit or to exclude expert witness testimony ‘will seldom constitute a ground for reversal.’” State v. Matthews, 479 Md. 278, 306 (quoting Rochkind, 471 Md. at 10 (quoting Roy v. Dackman, 445 Md. 23, 38–39 (2015))). We did 4 not, however, grapple with the fact that the formulation of the abuse of discretion standard that we have traditionally applied to evidentiary rulings and similar decisions is different from the formulation employed by federal courts applying the Daubert standard. As discussed below, since the transition to the Daubert standard, we have conducted searching and exacting reviews of the record. B. This Court’s Post-Rochkind Appellate Review of Circuit Courts’ Expert Testimony Rulings 1. State v. Matthews Having set the stage in Rochkind for transition from a de novo standard of review under Frye-Reed to an abuse of discretion standard under Daubert, we undertook our first appellate review of a trial court’s Daubert-Rochkind analysis in Matthews. 479 Md. at 278. Acknowledging that our review was for an abuse of discretion, the Court relied on cases outside the expert testimony context and incorporated some of our traditional definitions of that standard, stating that an appellate court does ‘not reverse simply because the . . . court would not have made the same ruling.’ Devincentz v. State, 460 Md. 518, 550 (2018). ‘Rather, the trial court’s decision must be well removed from any center mark imagined by the reviewing court and beyond the fringe of what that court deems minimally acceptable.’ Id.; see also Williams v. State, 457 Md. 551, 563 (2018) (“An abuse of discretion occurs where no reasonable person would take the view adopted by the circuit court.”); Jenkins v. State, 375 Md. 284, 295–96 (2003) (“Abuse occurs when a trial judge exercises discretion in an arbitrary or capricious manner or when he or she acts beyond the letter or reason of the law.”). Matthews, 479 Md. at 305–06 (cleaned up). In addition, we “reaffirmed” the sentiments expressed in Rochkind, stating that “it is still the rare case in which a Maryland trial court’s exercise of discretion to admit or deny expert testimony will be overturned.” Id. at 306. 5 In Matthews, we held that the trial court did not abuse its discretion by failing to exclude a photogrammetry expert’s testimony due to the expert’s inability to provide a margin of error that accounted for several potential variables relating to height estimates of a criminal defendant in connection with a reverse projection photogrammetry analysis. Id. at 325. We determined that “[t]here [was] no dispute that [the expert’s] methodology was reliable.” Id. at 313. “Nor was there any analytical gap in [the expert’s] proffered testimony.” Id. Instead, we concluded that “[t]he unknown degree of uncertainty concerning the accuracy of [the expert’s] height estimate went to the weight the jury should give to the expert testimony, not to its admissibility.” Id. (footnote omitted). The Majority opinion in this case describes our careful and searching examination of the record in Matthews, see Maj. Slip Op. at 29–32, which I need not repeat here. 2. Abruquah v. State Our second case involving the application of Daubert-Rochkind was Abruquah v. State. In that case, we were asked to determine whether, in a murder case, the trial court erred in permitting a firearms expert to testify, over objection, that “each of the four bullets and the bullet fragment . . . ‘at some point’ ‘had been fired’ from or through ‘the Taurus revolver[]’”—the gun that had been located at the criminal defendant’s home and was the alleged murder weapon. 483 Md. at 680. The expert “testified neither that his opinion was offered to any particular level of certainty nor that it was subject to qualifications or caveats.” Id. In connection with our review, we “discuss[ed] general background on the firearms identification methodology employed by the State’s expert witness, criticisms of the 6 methodology, studies of the methodology, the testimony presented to the circuit court, and caselaw from other jurisdictions.” Id. at 653; 656–79. We then considered the trial court’s application of the Daubert-Rochkind factors. See id. at 680–98. After undertaking this review, based on the evidence presented at the hearings, we held that the circuit court did not abuse its discretion in ruling that the firearms expert “could testify about firearms identification generally,” as well as “his examination of the bullets and bullet fragments found at the crime scene, his comparison of that evidence to bullets known to have been fired from the [criminal defendant’s] Taurus revolver, and whether the patterns and markings on the crime scene bullets are consistent or inconsistent with the patterns and markings on the known bullets.” Id. at 698. With regard to the circuit court’s decision to “permit[] the State’s expert witness to opine without qualification that the crime scene bullets were fired from [the defendant’s] firearm[,]” we determined that the circuit court abused its discretion because the studies and other information in the record do not “support the use of [the firearms identification methodology] to reliably opine without qualification that the bullets of unknown origin were fired from the particular firearm.” Id. at 694–95, 698. We explained that “[b]ecause the court’s error was not harmless beyond a reasonable doubt,” we reversed the circuit court’s ruling on the defendant’s motion in limine, vacated his conviction, and remanded the case for a new trial. Id. at 698. In undertaking our review, we acknowledged that we were applying an abuse of discretion standard. Id. at 652. We also observed that in Matthews, we applied the same “frequently described” formulation of that standard from our case law—namely, that an abuse of discretion occurs “when ‘no reasonable person would take the view adopted by 7 the circuit court’ or when a decision is ‘well removed from any center mark imagined by the reviewing court and beyond the fringe of what the court deems minimally acceptable.’” Abruquah, 483 Md. at 652 n.5 (quoting Matthews, 479 Md. at 305 (first quoting Williams, 457 Md. at 563, and next quoting Devincentz, 460 Md. at 550)). Recognizing that the description of this standard as articulated in our case law really did not fit with the trial court’s careful and thoughtful approach to the application of the Daubert factors, we stated: In our view, the application of those descriptions to a trial court’s application of a newly adopted standard, such as that adopted by this Court in Rochkind as applicable to the admissibility of expert testimony, is somewhat unfair. In this case, in the absence of additional caselaw from this Court implementing the newly adopted standard, the circuit court acted deliberately and thoughtfully in approaching, analyzing, and resolving the question before it. This Court’s majority has come to a different conclusion concerning the outer bounds of what is acceptable expert evidence in this area. Abruquah, 483 Md. at 652 n.5. 3. Oglesby v. Baltimore School Associates Last month, this Court issued its opinion Oglesby v. Baltimore School Associates. Oglesby was a lead paint case in which the plaintiff sued the owners and managers of an apartment building for negligence, negligent misrepresentation, and a violation of the Maryland Consumer Protection Act. Oglesby, 2023 WL 4755689 at *1. We were asked to consider whether the circuit court erred in granting summary judgment in favor of the property owners after the circuit court excluded the plaintiff’s expert testimony to establish that the property in question was a source of lead exposure and a significant factor 8 contributing to the plaintiff’s alleged injuries, including a loss of IQ points.5 Id. The plaintiff’s causation expert, Steven Elliot Caplan, M.D., “concluded that [the plaintiff’s] likely exposure to lead at the property was a significant contributing factor to bringing about the [alleged] cognitive deficiencies and impairments[,]” and to a loss of IQ points. Id. The property owners contended that Dr. Caplan lacked a sufficient factual basis for his causation opinions, and that his methodology used to calculate the plaintiff’s IQ loss was not reliable or generally accepted. Id. at *2. After a hearing at which the circuit court considered the property owners’ motions to preclude expert opinions and motions for summary judgment, the court granted both motions. Id. The circuit court agreed with the property owners that Dr. Caplan lacked a sufficient factual basis for his opinions, and that, without Dr. Caplan’s testimony as to causation, the plaintiff was unable to establish a prima facie case for negligence. Id. After the Appellate Court affirmed the circuit court’s judgment, we granted certiorari to determine whether the circuit court erred in ruling that Dr. Caplan’s opinions lacked a sufficient factual basis. Id. In undertaking our review, we once again described our abuse of discretion standard of review “as occurring ‘where no reasonable person would take the view adopted by the trial court,’ or when ‘the decision under consideration 5 In Oglesby, the trial court analyzed the proffered expert testimony under the FryeReed standard because the trial court proceedings took place before this Court issued its decision in Rochkind, and, therefore, Oglesby did not involve review of the trial court’s application of the Daubert-Rochkind factors. However, the parties petitioned and briefed this Court for review of the trial court’s determination under the post-Rochkind abuse of discretion standard, and this Court conducted a careful and searching review of the record. 9 is well removed from any center mark imagined by the reviewing court and beyond the fringe of what that court deems minimally acceptable[.]’” Id. at *12 (cleaned up). After conducting a “careful review of the record,” we determined that “Dr. Caplan’s opinions had a sufficient factual basis,” and that “the circuit court resolved genuine disputes of material fact.” Id. at *2. We further concluded that Dr. Caplan “had more than an adequate supply of data from which to form the opinion and the methodology he employed was reliable.” Id. As such, we held that the circuit court “abused its discretion in granting the motion to preclude and in determining that Dr. Caplan’s testimony that [the plaintiff’s] exposure to lead at the property was a significant contributing factor to her injures was inadmissible.” Id. In undertaking our review, we determined that the court resolved certain facts in the property owners’ favor, such as whether there was lead paint present at the property at the time that the plaintiff lived there, whether the plaintiff had come into contact with lead at the property, through peeling or chipping paint, and whether the plaintiff was potentially exposed to lead at other locations and the condition of those properties. Id. at *16. We stated, “[i]n short, a trial court is not permitted to resolve disputes of material fact in determining whether a sufficient factual basis exists to support an expert’s opinion. Doing so is a clear abuse of discretion.” Id. We proceeded to discuss in detail the evidence produced by the plaintiff, and the data that Dr. Caplan relied upon to support the causation links necessary to avoid summary judgment in a lead paint case. Id. at *17–22. After conducting our own review of the evidence, we stated that: 10 Under the particular circumstances of this case, given the large quantity of data that [the expert] had available and reviewed and the nature of the challenge to the admissibility of his testimony, a remand for further proceedings as to Dr. Caplan’s opinion that [the plaintiff’s] exposure to lead at the property was a substantial contributing factor to her injuries (other than IQ loss) is not warranted. Id. at *22. Concerning Dr. Caplan’s testimony that the plaintiff suffered an IQ loss of 3-4 points, in reliance upon particular lead paint studies—the Lanphear study and the Canfield study—we discussed our Court’s previous case law in lead paint cases involving experts’ reliance upon the Lanphear study. Id. at *23–26. We concluded that “[t]he bottom line is that these cases demonstrate that experts have been permitted to rely on the Lanphear study and extrapolate from its findings and render an opinion that an individual suffered a specified loss of IQ points as a result of exposure to lead.” Id. at *26. We observed that, “although our case law demonstrates that it is permissible for an expert to rely on the Lanphear study to offer an opinion that exposure to lead resulted in a specific loss of IQ points, the record in this case shows that neither Dr. Caplan’s report nor his deposition testimony fully explains the basis for his calculations under either” the Lanphear study or the Canfield study. Id. at *27. We stated that “without a full explanation of his methodology, including the reasons for his choices, even though it is possible to discern the basis of the calculations, we cannot determine whether it was an abuse of discretion for the circuit court to preclude this aspect of Dr. Caplan’s testimony.” Id. We remanded the case for a Daubert-Rochkind hearing “for the circuit court to determine whether the calculations that Dr. Caplan employed using the Lanphear study are 11 reliable and to assess Dr. Caplan’s use of the Canfield study and the reliability of his methodology with respect to it, should [the plaintiff] seek to introduce evidence concerning her IQ loss at trial.” Id. at *28. 4. The Case at Hand This case is our fourth opportunity since Rochkind to review expert witness admissibility determinations for abuse of discretion. In undertaking this review, we state that with the adoption of Daubert, “we promised the deference appropriate to courts administering a flexible approach to analyzing the admissibility of expert testimony[]” and that “[t]his case requires us to reflect on that flexibility and deference.” Maj. Slip Op. at 1. In connection with that deferential review, we once again recite our traditional formulation of the abuse of discretion standard as described in our case law—such as upholding a trial court’s decision unless it is “well removed from any center mark imagined by the reviewing court and beyond the fringe of what that court deems minimally acceptable” or “no reasonable person would take the view adopted by the circuit court.” Maj. Slip Op. at 23 (citations omitted). In the context of our review in this case, we have considered the trial court’s stated reason for its decision to exclude Ms. Cardell as an expert witness alongside our own searching review of the record. In doing so, we carefully lay out the contents of the record, including the data Ms. Cardell used in her before-and-after method calculations, how Ms. Cardell reached her conclusions, the issues raised at the hearing, and the trial court’s stated reasons for excluding Ms. Cardell as an expert. Maj. Slip Op. at 7–21. We ultimately agree with the trial court that there were “speculative, insufficiently substantiated judgment 12 calls that were central to Ms. Cardell’s application of the before-and-after method.” Maj. Slip Op. at 42. Specifically, this Court concludes the trial court did not abuse its discretion in determining that Ms. Cardell’s “judgment calls” regarding the selection of the 2015 base year, treatment of member draws, and failure to factor in confounding variables in her methodology were matters bearing on the reliability of the methodology and that each weighed against admissibility. Maj. Slip Op. at 42. We explain that: We rely on trial courts that conduct Daubert-Rochkind hearings to determine where the line between data and methodology is in the specific cases before them, and whether the proffered expert’s choices relating to data, assumptions, and other inputs implicate the reliability of the expert’s methodology. In this part of the trial court’s analysis, the court did just that. Maj. Slip Op. at 42–43. Although this Court “discern[s] no error in most of the trial court’s application of the Daubert-Rochkind factors,” we conclude that the trial court erred in its determination that Ms. Cardell’s treatment of the 2021 updates implicated the reliability of her methodology. Maj. Slip Op. at 46. The trial court based this conclusion on its observation that Ms. Cardell made this update without new information and “‘for fully subjective reasons.’” Id. In determining that the trial court abused its discretion in weighing this information against admissibility, we carefully look at the record and conclude that Ms. Cardell was basing her decision on new information because she “noticed something she had not noticed before on first examination[,]” and “[c]atching something peculiar the second time around neither undermines the adequacy of the data . . . nor the court’s understanding of the expert’s methodology[.]” Maj. Slip Op. at 46–47. In other words, this Court rejects the trial court’s determination that Ms. Cardell’s June 2021 13 adjustments were made for “fully subjective reasons” based on our own review of the record. Maj. Slip Op. at 46. We go on to explain that Ms. Cardell’s decision to make the adjustments “did not implicate the reliability of her methodology. At most, it went to the care with which she applied her methodology, which is a matter to be explored on crossexamination before the jury (if Ms. Cardell’s testimony is otherwise found to be sufficiently reliable).” Maj. Slip Op. at 47–48. We also hold that “[t]he trial court . . . erred when it stated that Ms. Cardell’s lostprofit calculations would not be helpful to the jury.” Maj. Slip Op. at 47, n.16. The trial court’s reasoning for weighing this factor against admission was because “if the jury concluded that one (or more) of the departing physicians left for reasons other than KatzAbosch’s negligence, a lost-profits analysis that included that physician’s projected revenue would overstate the practice’s lost profits and therefore not help the jury to accurately calculate damages.” Id. On review, we determine that “there was no way to know before the trial whether the jury would conclude that any of the physicians left for reasons other than KatzAbosch’s negligence[,]” and, therefore, the trial court’s conclusion “was speculative and not a proper ground for exclusion.” Id. Importantly, this Court points out that there is overlap between matters of reliability and credibility, as well as data and methodology, which can create challenges for trial courts applying Daubert. Maj. Slip Op. at 32–43. As such, making admissibility determinations on these types of matters is necessarily fact-specific, both before the trial court and on appellate review. These questions often require the reviewing court to do more than simply accept a trial court’s justification for an admissibility determination as a 14 matter within the trial court’s discretion. Indeed, when conducting a review of reliability determinations, a reviewing court may need to look carefully at the record, including the underlying facts of the case, methodology employed, data relied upon, and contents of expert testimony and reports to determine if the trial court abused its discretion. This is exactly what this Court does here, and this is in line with how federal courts apply the abuse of discretion standard in the Daubert context. I take no issue with the manner in which we have undertaken our review in this case, or in Matthews, Abruquah, and Oglesby, as I joined the majority opinions authored by my colleagues in each of these cases. In the aftermath of Rochkind, I do not challenge the nature of our review for an abuse of discretion, but I question whether our reliance upon our frequently described formulation of this standard, in fact, reflects the nature and type of review that we are undertaking in these cases and that federal courts have employed in undertaking the same exercise. C. “Abuse of Discretion” as a Context-Dependent Standard In the context of appellate review, reviewing courts use one of three well-known standards of review. An appellate court reviews questions of law de novo. Questions of fact are reviewed under the clearly erroneous standard. Finally, matters that are committed to the discretion of the trial court—such as whether to admit expert testimony—are reviewed for an “abuse of discretion.” Of the three standards of review, the abuse of discretion standard is typically considered to be the most deferential. However, it has been described as “famously slippery,” and has been understood to have different meanings and applications in different contexts. Zervos v. Verizon, 252 F.3d 163, 168 n.4 (2d Cir. 2001); 15 Gasperini v. Center for Humanities, Inc., 149 F.3d 137, 141 (2d Cir. 1998) (noting that abuse of discretion “may have different meanings in different contexts” depending on why the decision is within the trial court’s discretion); Lawson Prods., Inc. v. Avnet, Inc., 782 F.2d 1429, 1438 (7th Cir. 1986) (describing abuse of discretion as a term “that seems to escape easy definition”). In attempting to explain the difference between the abuse of discretion standard and clearly erroneous standard, the Seventh Circuit has observed: Abuse of discretion is conventionally regarded as a more deferential standard than clear error, though whether there is any real or consistent difference has been questioned. The alternative view is that both standards denote a range, rather than a point, that the ranges overlap and maybe coincide, and that the actual degree of scrutiny in a particular case depends on the particulars of that case rather than on the label affixed to the standard of appellate review. Haugh v. Jones & Laughlin Steel Corp., 949 F.2d 914, 916–17 (7th Cir. 1991). The type of ruling under review informs how appellate courts apply the abuse of discretion standard. Dissenting in Calderon v. Thompson, Justice David Souter wrote that “the variety of subjects left to discretionary decision requires caution in synthesizing abuse of discretion cases.” 523 U.S. 538, 567 (1998) (Souter, J. dissenting) (citations omitted). In the same vein, Judge Henry J. Friendly recommended that when applying the abuse of discretion standard, appellate courts “must carefully scrutinize the nature of the trial court’s determination and decide whether that court’s superior opportunities of observation or other reasons of policy require greater deference than would be accorded to its formulations of law or its application of law to the facts.” Henry J. Friendly, Indiscretion About Discretion, 31 Emory L.J. 747, 784 (1982). In the decades since Daubert was decided, 16 federal courts have shaped the abuse of discretion standard in reviewing admissibility determinations of expert witness testimony. D. The Federal Courts’ Description of the Abuse of Discretion Standard in the Daubert Context As discussed, after Daubert, the Supreme Court made clear that the abuse of discretion standard is the standard that a reviewing court must apply in undertaking a review of a trial court’s decision to admit or exclude expert testimony. See Joiner, 522 U.S. at 136; Kumho, 526 U.S. at 142 (explaining that “the law grants a [trial] court the same broad latitude when it decides how to determine reliability as it enjoys in respect to its ultimate reliability determination[]”). Although the standard was established by these cases, it does not always appear to be applied as the hands-off deferential approach that often accompanies appellate review for an abuse of discretion. Instead, it involves a more searching and careful examination of the records, data, studies, testimony, and trial court’s analysis of the same. Writing for the American Bar Association, lawyers David F. Herr and Morgan L. Holcomb observed that after [Joiner] the proper standard of review is clearly the “abuse-ofdiscretion” standard. However, even Joiner itself appeared to apply quite searching review. As Justice Stevens pointed out in his partially concurring opinion, Part III of the majority opinion went beyond announcing the appropriate standard of review, and carefully examined the basis on which the trial court excluded the expert evidence. Joiner, 522 U.S. at 151–55 (Steven[s], J.[,] concurring). This careful examination has not gone unnoticed; for example, one commentator noted, “while they don’t say so, some appellate opinions come close to a de novo style of review of the proffered expert testimony.” Michael J. Saks, et al., Annotated Reference Manual on Scientific Evidence Second 17 23 (West 2004) (citing Mathis v. Exxon Corp., 302 F.3d 448 (5th Cir. 2002) (noting that the trial court had failed to “offer any reasons in support of admitting the expert’s testimony,” but nonetheless holding that the appellate court did not have to remand and could carry out the review itself on appeal: “Because admissibility is a legal question—one ill-suited to remand and further explication by the [trial] court—we will decide the question in this case without remand.”)). David F. Herr & Morgan L. Holcomb, Opinion and Expert Testimony in Federal and State Courts, Am. L. Inst. – Am. Bar Ass’n 629, 642 (Jan. 2007). Herr and Holcomb go on to catalog additional examples of federal circuit courts conducting “searching” reviews of admissibility determinations in the expert witness context. Id. (citing United States v. Mitchell, 365 F.3d 215 (3d Cir. 2004) (conducting a searching analysis of the admission of fingerprint evidence, and concluding that admission of the prosecution’s expert was not error); Lauzon v. Senco Prods., Inc., 270 F.3d 681 (8th Cir. 2001) (conducting a fairly searching review, and ultimately holding that “[t]hrough examination of the record in light of the requirements of Daubert and its progeny, ineluctably we are led to conclude the [trial] court’s exclusion of the testimony was an abuse of discretion and fell outside the spirit of admissibility as set forth in Federal Rule of Evidence 702.”); Pride v. BIC Corp., 218 F.3d 566, 578 (6th Cir. 2000) (noting review of Daubert ruling was for abuse of discretion, but conducting “de novo review of the record” to conclude that the “record supports the [trial] court’s finding on this issue”)). The authors also note the reluctance of appellate courts to remand these issues because appellate courts tend to make admissibility determinations themselves. Id. at 642–43. Herr and Holcomb’s description of how federal courts apply the abuse of discretion standard in the expert witness context—like this Court’s application of the standard in 18 Matthews, Abruquah, Oglesby, and here—feels out of sync with the even more deferential approach that we take when reviewing other types of discretionary decisions by trial courts and our “no reasonable person” and “well removed from the center mark” articulation of the standard. In the Daubert context, various federal circuits have supplied their own descriptions of the nature of the abuse of discretion standard, a few of which I highlight here. In its description of the standard, the First Circuit explains that a trial court abuses its discretion “when a material factor deserving significant weight is ignored, when an improper factor is relied upon, or when all proper and no improper factors are assessed, but the court makes a serious mistake in weighing them.” Lawes v. CSA Architects and Engineers, LLP, 963 F.3d 72, 90 (1st Cir. 2020) (quotations omitted). The Court notes that the abuse of discretion “standard is not monolithic: within it, embedded findings of facts are reviewed for clear error, questions of law are reviewed de novo, and judgment calls are subjected to classic abuse-of-discretion review.” Id. (internal quotations omitted). Accordingly, the reviewing court will reverse a trial court’s decision if it determines the trial court “committed a material error of law or a meaningful error in judgment.” Id. (quotations omitted). The Tenth Circuit has described the standard as follows: We review de novo the question of whether the [trial] court applied the proper legal test in admitting an expert’s testimony. Though the [trial] court has discretion in how it conducts the gatekeeper function, we have recognized that it has no discretion to avoid performing the gatekeeper function. Therefore, we review de novo the question of whether the [trial] court applied the proper standard and actually performed its gatekeeper role in the first instance. We then review the trial court’s actual application of the 19 standard in deciding whether to admit or exclude an expert’s testimony for abuse of discretion. The trial court’s broad discretion applies both in deciding how to assess an expert’s reliability, including what procedures to utilize in making that assessment, as well as in making the ultimate determination of reliability. Accordingly, we will not disturb the [trial] court’s ruling unless it is arbitrary, capricious, whimsical or manifestly unreasonable or when we are convinced that the [trial] court made a clear error of judgment or exceeded the bounds of permissible choice in the circumstances. Dodge v. Cotter Corp., 328 F.3d 1212, 1223 (10th Cir. 2003) (quotations omitted) (cleaned up). That court has also emphasized that a “natural requirement of the gatekeeper function is the creation of a sufficiently developed record[,]” and has observed that “Kumho and Daubert make it clear that the [trial] court must, on the record, make some kind of reliability determination.” Id. at 1223 (cleaned up) (citations omitted). Thus, the Tenth Circuit has held “when faced with a party’s objection, a [trial] court must adequately demonstrate by specific findings on the record that it has performed its duty as gatekeeper.” Id. (cleaned up). The court has stated that: Without specific findings or discussion on the record, it is impossible on appeal to determine whether the [trial] court carefully and meticulously reviewed the proffered scientific evidence or simply made an off-the-cuff decision to admit the expert testimony. In the absence of such findings, we must conclude that the court abused its discretion in admitting such testimony. Id. (quotations omitted). The Seventh Circuit describes a “two-step standard of review in cases challenging a [trial] court’s admission or exclusion of the testimony of an expert.” C.W. ex rel. Wood v. Textron, Inc., 807 F.3d 827, 835 (7th Cir. 2015). First, the court “review[s] de novo a [trial] court’s application of the Daubert framework.” Id. (citing United States v. Brumley, 20 217 F.3d 905, 911 (7th Cir. 2000)). If the reviewing court determines that “the [trial] court properly adhered to the Daubert framework,” then it reviews the “decision to exclude (or not to exclude) expert testimony for abuse of discretion.” Id. The Eleventh Circuit explained the rationale behind the abuse of discretion standard in Daubert, as well as that Court’s application of the standard, in United States v. Brown, 415 F.3d 1257 (11th Cir. 2005). That Court explained that the standard is proper because the trial court is “[i]mmersed in the case as it unfolds” and is “more familiar with the procedural and factual details and is in a better position to decide Daubert issues.” Id. at 1266. That court explained: The rules relating to Daubert issues are not precisely calibrated and must be applied in case-specific evidentiary circumstances that often defy generalization. And we don’t want to denigrate the importance of the trial and encourage appeals of rulings relating to the testimony of expert witnesses. All of this explains why the task of evaluating the reliability of expert testimony is uniquely entrusted to the [trial] court under Daubert, and why we give the [trial] court considerable leeway in the execution of its duty. Id. (quotations omitted). However, the court goes on to clarify, [t]o be sure, review under an abuse of discretion standard does entail review, and granting considerable leeway is not the same thing as abdicating appellate responsibility. We will reverse when the [trial] court’s Daubert ruling does amount to an abuse of discretion that affected the outcome of a trial. An abuse of discretion can occur where the [trial] court applies the wrong law, follows the wrong procedure, bases its decision on clearly erroneous facts, or commits a clear error in judgment. In the Daubert context, if one of those types of serious error occur, we may conclude that the [trial] court has not properly fulfilled its role as gatekeeper. We have also found that a [trial] court abuses its discretion where it fails to act as a gatekeeper by essentially abdicating its gatekeeping role. Id. (quotations omitted). 21 E. Some Concluding Thoughts on Our Standard of Review in the Daubert Context When we made the decision in Rochkind to adopt Daubert, we did so with the knowledge that there is a “broad base of case law” that went along with the new standard. Rochkind, 471 Md. at 34–35 (quoting Savage, 455 Md. at 185 (Adkins, J., concurring)). However, in describing the abuse of discretion standard of review, we transplanted our formulation of the standard we apply in other contexts. See, e.g., Devincentz, 460 Md. at 539, 550 (applying the abuse of discretion standard to the review of a trial court’s decision to admit or exclude a character witness’s opinion); Williams, 457 Md. at 562–63 (applying the abuse of discretion standard to the review of a trial court’s decision to admit relevant evidence); Jenkins, 375 Md. at 295–96 (applying the abuse of discretion standard to the review of a trial court’s decision to deny a motion for a new trial). In my view, our traditional formulation of the abuse of discretion standard—as invoking the articulation of “no reasonable person” or “well removed from any center mark imagined by a reviewing court and beyond the fringe of what the court deems minimally acceptable”—is not the best or most accurate way of describing our abuse of discretion review in the DaubertRochkind context. See Williams, 457 Md. at 563; Devincentz, 460 Md. at 550. I believe that this Court should re-formulate a definition of our abuse of discretion standard in the context of appellate review of a trial court’s application of the DaubertRochkind factors when analyzing whether to admit or exclude expert testimony. In doing so, we should look to the federal courts’ articulation of the standard of review in the same manner that we will look to the federal courts’ jurisprudence in its application. I agree with 22 the observation made by Judge Friendly and various federal circuits that the abuse of discretion standard is contextual. See, e.g., Gasperini, 149 F.3d at 141 (noting that abuse of discretion may have different meanings in different contexts depending on why the decision is within the trial court’s discretion); Lawes, 963 F.3d at 90 (explaining that the abuse of discretion standard is not monolithic). As is clearly reflected in our appellate review of the trial court’s expert admissibility determinations in Matthews, Abruquah, Oglesby, and the instant case, our abuse of discretion standard of review involves a searching and careful examination of the record, data, studies, testimony, and trial court’s analysis of the same. Our traditional formulation of the abuse of discretion standard does not comport with the type of review that we are undertaking here. I believe that the parties, their counsel and experts, trial courts and our Appellate Court would all benefit from this Court re-formulating our description of the abuse of discretion standard of review that is undertaken in this context and, in so doing, utilize the federal courts’ discussion of that standard. Our standard should reflect that the hallmark of the abuse of discretion standard regardless of context is the high degree of deference to a court exercising discretionary authority. Joiner, 522 U.S. at 137, 143. “Accordingly, we will not disturb the [trial] court’s ruling unless it is arbitrary, capricious, whimsical, or manifestly unreasonable or when we are convinced that the [trial] court made a clear error of judgment or exceeded the bounds of permissible choice in the circumstances.” Dodge, 328 F.3d at 1223 (quotations omitted). But this broad discretion “is not the same thing as abdicating appellate responsibility.” Brown, 415 F.3d at 1266. In the context of admissibility of expert testimony, this means 23 that while trial courts are given leeway on matters within the range of their discretion, on review, the appellate court will find an abuse of discretion where the trial court’s rulings are “manifestly erroneous.” Joiner, 522 U.S. at 142. That is, when applying the abuse of discretion standard in the expert witness context, we reverse only where there is “a meaningful error in judgment.” Lawes, 963 F.3d at 90 (emphasis added). In determining whether the trial court has made such a meaningful error, reviewing courts may need to conduct a searching review of the record. With these points in mind, our standard should reflect that the “[abuse of discretion] standard is not monolithic: within it, embedded findings of fact are reviewed for clear error, questions of law are reviewed de novo, and judgment calls are subjected to a classic abuseof-discretion review.” Lawes, 963 F.3d at 90 (quotations omitted). A reviewing court may find a trial court abused its discretion where the trial court (1) “applies the wrong law;” (2) “follows the wrong procedure;” (3) “bases its decision on clearly erroneous facts;” (4) “commits a clear error in judgment;” or (5) abdicates its gatekeeping role altogether. Brown, 415 F.3d at 1266; see also Lawes, 963 F.3d at 90. Although trial courts have discretion for determining the best procedure for the proper resolution of any expert admissibility issues, see Kumho, 526 U.S. at 152, trial judges should be reminded that a reviewing court will find an abuse of discretion where they fail to act as a gatekeeper by essentially abdicating their gatekeeping role, or where the trial court assumes the role of the jury. Part of this gatekeeping role is developing a clear record of the court’s basis for its determination to admit or exclude expert testimony. See Dodge, 328 F.3d at 1223. 24 Our formulation of the standard should also take into account that appellate review of expert admissibility determinations encompasses many types of challenges, including whether: expert testimony is necessary or appropriate; the expert is qualified; the expert’s area of expertise is unreliable; the expert’s conclusions are invalid or unreliable; the testimony exceeds the expert’s area of expertise; and the expert testimony is relevant. In my view, “the actual degree of scrutiny in a particular case [may] depend[] on the particulars of that case rather than on the label affixed to the standard of appellate review.” Haugh, 949 F.2d at 916–17; see also Brown, 415 F.3d at 1265. A re-formulated standard should reflect this flexibility. As far as our statements in Rochkind and Matthews that it will be the “rare case in which a Maryland trial court’s exercise of discretion to admit or deny expert testimony will be overturned,” I suppose only time will tell. Matthews, 479 Md. at 286, 306 (emphasis added). 25 Circuit Court for Howard County Case No. C-13- CV-18-000181 Argued: May 4, 2023 IN THE SUPREME COURT OF MARYLAND* No. 30 September Term, 2022 __________________________________ KATZ, ABOSCH, WINDESHEIM, GERSHMAN & FREEDMAN, P.A., ET AL. v. PARKWAY NEUROSCIENCE AND SPINE INSTITUTE, LLC __________________________________ Fader, C.J., Watts, Hotten, Booth, Biran, Gould, Eaves, JJ. __________________________________ Concurring and Dissenting Opinion by Gould, J. __________________________________ Filed: August 30, 2023 *During the November 8, 2022 general election, the voters of Maryland ratified a constitutional amendment changing the name of the Court of Appeals to the Supreme Court of Maryland. The name change took effect on December 14, 2022 I respectfully concur in part and dissent in part from the Majority’s well-written and thorough opinion. As I understand it, the Majority concludes that the trial court’s analysis of the various factors under Daubert-Rochkind could very well support the trial court’s discretionary decision to exclude Ms. Cardell’s testimony, but the Majority is concerned that the trial court may have unduly considered Ms. Cardell’s updated calculations in assessing those factors. I agree with the Majority that Ms. Cardell’s updated calculations and her reasons for them do not go to the reliability of her methodology, but are instead grist for the cross-examination mill. I write separately because I would not remand for the trial court to reconsider its analysis. In my view, Ms. Cardell’s methodology is so fundamentally flawed as to constitute “the rare case” in which a trial court’s admission of expert testimony would have been an abuse of discretion. See State v. Matthews, 479 Md. 278, 306 (2022). Parkway Neuroscience and Spine Institute, LLC’s (“PNSI”) theory of the case was that, beginning in 2013, Katz, Abosch, Windesheim, Gershman & Freedman, P.A.’s and Mark Rapson’s (collectively, “KatzAbosch”) negligent advice caused seven of the nine member-physicians, who were still with PNSI at the end of 2014, to leave the practice by mid-2016, and that their departures caused damages to the company in the form of lost profits. Ms. Cardell’s job was to quantify the lost profits caused by those departures. She utilized the “before-and-after” method. As the Majority aptly notes, despite the seemingly straightforward nature of this method, as an expert, Ms. Cardell was required to make decisions requiring judgment and to sufficiently explain those decisions. See Maj. slip op. at 33, 42-43 & n.15. Ms. Cardell assumed that, if not for the departure of member-physicians, PNSI’s profits in 2017 through 2019 would have been, at a minimum, the same as in 2015. So, according to Ms. Cardell’s theory, to calculate the lost profits for, say, 2017, you take the difference between the profit achieved in 2015 ($321,751) and the loss incurred in 2017 ($1,515,436) and, voila, you have $1,837,186 of lost profits for 2017.1 Ms. Cardell used the same analysis for 2018 and 2019 to arrive at a total lost profits of $4,956,080 for all three years. Ms. Cardell thus attributed the entire change in profitability at PNSI to the departure of the member-physicians. It is that assumption that requires the exclusion of her testimony. A critical component of the “before-and-after” method for calculating lost profits is accounting or controlling for confounding variables—changes in conditions that might have affected the plaintiff’s profitability during the “loss period.” Accounting for confounding variables may be considered, in the lost profits context, akin to “whether the expert has adequately accounted for obvious alternative explanations” under the DaubertRochkind rubric. See Rochkind v. Stevenson, 471 Md. 1, 36 (2020). Confounding variables could include, for instance, the prices of goods and services, consumer demand and preferences, technological developments, personnel turnover, and broader industry and 1 Due to rounding, the numbers do not add up exactly. 2 economic conditions. The failure to account for relevant confounding variables has routinely resulted in the exclusion of such testimony.2, 3 Here, the circuit court determined that Ms. Cardell’s failure to evaluate revenue generation by individual member-physicians and failure to consider insurance reimbursement rates favored exclusion. In my view, Ms. Cardell’s wholesale failure to evaluate and consider those and other relevant factors required the exclusion of her testimony. 2 See, e.g., E. Auto Distribs., Inc. v. Peugeot Motors of Am., Inc., 795 F.2d 329 (4th Cir. 1986) (holding expert testimony properly excluded because it failed to account for changes in efficiency of sales force and demand for products); Craftsmen Limousine, Inc. v. Ford Motor Co., 363 F.3d 761 (8th Cir. 2004) (holding expert testimony should have been excluded because it failed to analyze the effects of general economic conditions and increased competition); Schiller & Schmidt, Inc. v. Nordisco Corp., 969 F.2d 410 (7th Cir. 1992) (holding expert testimony inadequate because it failed to distinguish between the damages caused by the lawful entry of a powerful competitor and those caused by the competitor’s misconduct); Isaksen v. Vermont Castings, Inc., 825 F.2d 1158, 1165 (7th Cir. 1987) (holding expert testimony inadequate when the plaintiff “made no effort to establish how much of the loss was due to [the defendant’s unlawful] activity as distinct from unrelated business factors” and “[a]ll [the plaintiff] did to prove damages was to compare his average profits for several years before and several years during the period of unlawful activity”); Farley Transp. Co. v. Santa Fe Trail Transp. Co., 786 F.2d 1342 (9th Cir. 1985) (holding expert testimony inadequate because it failed to distinguish between losses caused by lawful and unlawful competition); Coleman Motor Co. v. Chrysler Corp., 525 F.2d 1338 (3rd Cir. 1975) (holding expert testimony inadequate for failing to account for lawful competition, market changes, and changes in the character of the neighborhood where plaintiff’s auto dealership was located). 3 Asking experts to provide analysis that implicates relationships between profits and confounding variables does not infringe upon the jury’s responsibility to determine whether the defendant caused the alleged harm. Here, Ms. Cardell was not called upon to determine as part of her analysis whether KatzAbosch’s allegedly poor financial advice caused the member-physicians to depart. Rather, she was asked to quantify the lost profits caused by those departures. 3 It is inconceivable that 100 percent of the changes in profitability between the benchmark year of 2015 and the loss years of 2017, 2018, and 2019 were attributable to the departures of the member-physicians. Stated differently, it’s inconceivable that had none of the member-physicians departed, PNSI’s profits would have remained exactly the same. PNSI’s own allegations in its complaint establish this very point: the company’s “overhead, expenses and revenue generated by the Members’ practices during the relevant time period differed markedly by specialty.” That alone indicates that the impact on the company’s profitability from the departure of a particular member-physician would depend on the specific costs and revenue structures of that member-physician’s practice. As the trial court noted, Ms. Cardell virtually ignored changes in reimbursement rates. To the extent she considered them at all, she did not analyze how such changes would have affected the total revenues had each of the doctors stayed. Instead, she merely speculated that the observed decrease in actual reimbursement rates—which defendant’s counsel presented and she acknowledged on cross-examination—would have been offset by general increases in demand for medical services.4 4 That may or may not be so, but her speculation only reinforces why an expert must drill down on the figures. Which specialties were expected to experience increased demand? The high profit margin specialties? The low profit margin specialties? What marginal expenses would PNSI have incurred in association with that increased demand? Did PNSI have the capacity to accommodate such increased demand, and if so, by how much? How many additional services would PNSI have had to provide to offset the decreased reimbursement rates? That is, where was the break-even point? Ms. Cardell’s mere speculation that decreased reimbursement rates would have been offset by increased demand for services is no substitute for an actual analysis that accounts for confounding variables. 4 You do not have to be an expert to know that expenses of any business are subject to fluctuations. Indeed, accounting for such fluctuations was factored into PNSI’s compensation model. PNSI’s complaint alleges that a new compensation model developed and recommended by KatzAbosch was approved and adopted by its members. Here, PNSI alleges that under this new model: in determining Member cash flow distributions, each Member regardless of specialty was allocated an equal portion of 2/3 of [PNSI’s] net increase or decrease in expenses (the “fixed portion”) and a portion of the remaining 1/3 of [PNSI’s] net increase or decrease in expenses based on the Member’s net collections for the period (the “variable portion”). Thus, PNSI’s own complaint shows that its revenues and expenses were not static. Indeed, PNSI assumed as much in restructuring its compensation model. There is no evidence Ms. Cardell analyzed the extent to which changes in expenses or increased demand for services could have affected profits had the seven member-physicians remained. It would be as if a lost profits analysis of a gas station failed to account for changes in the wholesale and retail prices of gasoline. Ms. Cardell’s failure to make any attempt to account for such confounding variables requires the exclusion of her testimony. Accordingly, I dissent to that part of the Majority’s opinion that requires a remand for further proceedings. I would reverse the judgment of the Appellate Court. 5 Circuit Court for Howard County Case No. C-13-CV-18-000181 Argued: May 4, 2023 IN THE SUPREME COURT OF MARYLAND* No. 30 September Term, 2022 ______________________________________ KATZ, ABOSCH, WINDESHEIM, GERSHMAN & FREEDMAN, P.A., AND MARK E. RAPSON v. PARKWAY NEUROSCIENCE AND SPINE INSTITUTE, LLC ______________________________________ Fader, C.J. Watts Hotten Booth Biran Gould Eaves, JJ. ______________________________________ Dissenting Opinion by Watts, J. ______________________________________ Filed: August 30, 2023 *At the November 8, 2022 general election, the voters of Maryland ratified a constitutional amendment changing the name of the Court of Appeals of Maryland to the Supreme Court of Maryland. The name change took effect on December 14, 2022. Respectfully, I dissent. By and large, I agree with much of the Majority’s wellwritten opinion, but I disagree with its disposition of the case. I agree with the Majority that, in some instances, the choices made by an expert and the information that an expert inputs to a methodology can be part of the methodology. See Maj. Slip Op. at 3. I also agree that, in determining whether expert testimony has a sufficient factual basis, this Court should not adopt “an unduly rigid dividing line between ‘data’ and ‘methodology[.]’” Maj. Slip Op. at 3-4. Additionally, I agree that, in assessing the testimony of Meghan Cardell, an expert for Parkway Neuroscience and Spine Institute, LLC (“PNSI”), Respondent, the Circuit Court for Howard County abused its discretion in its “consideration of the [June 2021] normalizing adjustments as reflecting on the reliability of Ms. Cardell’s methodology, as opposed to the credibility (or reliability) of Ms. Cardell herself.” Maj. Slip Op. at 4. I part ways with the Majority as to the outcome of the case, though. Rather than remand the case to the circuit court for it to decide whether to admit Ms. Cardell’s testimony without considering the June 2021 normalizing adjustments as affecting the reliability of her methodology, I would affirm the decision of the Appellate Court of Maryland. Like the Appellate Court, I would conclude that Ms. Cardell’s expert opinion was admissible and that, in assessing the reliability of her methodology, the circuit court abused its discretion by determining that choices of data and other inputs rendered her methodology unreliable. See Parkway Neuroscience & Spine Inst., LLC v. Katz, Abosch, Windesheim, Gershman & Freedman, P.A., 255 Md. App. 596, 630, 283 A.3d 753, 773 (2022). In my view, in assessing the reliability of Ms. Cardell’s methodology, the circuit court took issue with matters such as the soundness of the data that Ms. Cardell used and the conclusions she reached, the nature of her experience, and whether she was biased or not. Unlike the Majority, I would not order a limited remand regarding the admissibility of Ms. Cardell’s expert testimony. See Maj. Slip Op. at 50. Instead, I would affirm the judgment of the Appellate Court, which reversed and remanded the case for further proceedings, with Ms. Cardell’s testimony admissible. See Parkway, 255 Md. App. at 639, 283 A.3d at 778. Determining the amount that a business would have earned or lost in any given period will generally require both an in-depth analysis of the business’s revenues and expenses over a period of time and looking at meaningful events affecting the business, such as employee changes and other data, culminating in projections made through use of a methodology deemed reliable for calculating profit or loss. In this case, it is undisputed that the method for calculation of lost profit that Ms. Cardell used—the before-and-after method—is generally accepted “as a reliable methodology.” Maj. Slip Op. at 49 n.19. Here, however, in an analysis of the factors set forth in Daubert v. Merrell Dow Pharms., Inc., 509 U.S. 579, 593-94 (1993) and Rochkind v. Stevenson, 471 Md. 1, 35-36, 236 A.3d 630, 650 (2020), the circuit court took issue with the quality of the data that Ms. Cardell relied on, not the reliability of the methodology that she used, and, in ruling, faulted her level of experience, and questioned her impartiality. To briefly recap the circuit court’s ruling, at the conclusion of the hearing on KatzAbosch’s1 motion to exclude Ms. Cardell’s testimony, the circuit court granted the motion, finding that PNSI did not meet the burden 1 Petitioners, Abosch, Windesheim, Gershman & Freedman, P.A., and Mark E. Rapson, together, are referred to as “KatzAbosch.” -2- of establishing that Ms. Cardell’s testimony was admissible. The circuit court summarized its major concerns with Ms. Cardell’s testimony before delving into the Daubert-Rochkind factors. A primary concern of the circuit court was an absence specialized knowledge, experience, or training in Ms. Cardell’s background. The circuit court stated that PNSI is a medical practice that required a distinct knowledge set, which Ms. Cardell lacked. Because of this, the circuit court questioned Ms. Cardell’s judgment in designating 2015 as the base year for calculating lost profits, although the circuit court found that her choice of the before-and-after method was correct. Finding this base period to be “speculative[,]” the circuit court concluded that subjectivity affected Ms. Cardell’s rationale. The circuit court took issue with the overall quality of data used in Ms. Cardell’s methodology, particularly the circumstance that, in its view, “she relie[d] entirely on representations from within the business and the Plaintiffs[,]” i.e., PNSI. Finding her testimony to contain “inherent bias” due to her “support” for PNSI’s claims, the circuit court critiqued Ms. Cardell’s judgment calls, including her decision to make a normalizing adjustment to the calculations in June 2021 when there had been no change in the underlying facts. The circuit court noted what it called Ms. Cardell’s failure to examine the individual economic impact of each doctor’s departure, stating that “[n]ot every doctor[] is going to make or contribute the same amount of money each . . . year.” This contributed to the -3- circuit court’s conclusion that Ms. Cardell’s opinion would not be helpful to the jury.2 Continuing to express concern about Ms. Cardell’s “judgment calls,” the circuit court faulted her failure to consider data that would have provided alternative explanations for lost profits, such as changes in insurance reimbursement rates. Another problem, according to the circuit court, was Ms. Cardell’s decision to consider member draws as reductions of profit, i.e., expenses, in her calculations without identifying an objective industry standard for such a classification. It was only after the circuit court set forth the above reasoning for excluding Ms. Cardell’s testimony that it turned to the Daubert-Rochkind factors, indicating that perhaps some of the factors may not apply. The circuit court quoted our statement in Rochkind, 471 Md. at 37, 236 A.3d at 651, that “[a] trial court may apply some, all, or none of the factors depending on the particular expert testimony at issue” (citation omitted), then added: “[T]hat means that not each and every factor has to be resolved, and not each and every factor of the ten have to be resolved against the witness.” (Internal quotation marks omitted). In addressing the factors, the circuit court repeated much of the reasoning that it had already discussed. Of the 10 Daubert-Rochkind factors, the circuit court appeared to determine that 7 factors weighed against the admissibility of Ms. Cardell’s testimony: (1) 2 The circuit court reiterated this conclusion in its final holding that PNSI had not met its burden under Daubert and Rochkind to show Ms. Cardell’s testimony’s reliability and usefulness to the jury. However, the circuit court promptly limited that holding, stating that “usefulness is a very, very slight factor and I don’t want to dwell on it because I don’t want to create reversible error[.]” -4- testability, (2) the rate of error, (3) standards and controls, (4) the purpose of the opinion, (5) unjustifiable extrapolation from an accepted premise, (6) accounting for obvious alternative explanations, and (7) the ability of the field to reach reliable results for the type of opinion. The circuit court appeared to determine that 2 factors were neutral or not applicable: (1) peer review, and (2) being as careful as the expert would be in non-litigation work. Although the circuit court did not expressly address the factor as to general acceptance, the circuit court agreed that the before-and-after method was the appropriate procedure to use. Of the 7 factors that appeared to weigh against admissibility, 2 factors— the rate of error and the ability of the field to reach reliable results for the type of opinion— were construed against the admissibility of Ms. Cardell’s testimony based on the June 2021 update, i.e., the normalizing adjustment, as to which the Majority concludes that the circuit court erred. See Maj. Slip Op. at 50. In discussing the other 5 factors that appeared to weigh against admissibility, the circuit court criticized Ms. Cardell’s choice of 2015 as the base year, her reliance on information from PNSI, and her failure to consider other information, such as changing insurance reimbursement rates, as an alternative explanation for lost profits—all of these pertained to the soundness of the data that Ms. Cardell used in formulating her opinion and the circuit court’s disagreement with her conclusions, or the court’s previous findings that Ms. Cardell lacked the experience necessary to accurately determine the base year for PNSI’s type of medical practice, under the before-and-after method, and that her testimony was inherently biased. To be sure, the use of data cannot always be surgically excised from the reliability of a methodology. That said, while “sometimes blurred[,]” Maj. Slip Op. at -5- 41, there is a distinction between data and methodology. Under Maryland Rule 5-702(3), in determining whether expert testimony is admissible, a trial court must assess “whether a sufficient factual basis exists to support the expert testimony.” “‘[S]ufficient factual basis’ includes two sub-elements: (1) an adequate supply of data; and (2) a reliable methodology.” Rochkind, 471 Md. at 22, 236 A.3d at 642 (citation omitted). Different definitions apply to these two different elements. Among other things, for there to be an adequate supply of data, there cannot be “too great an analytical gap between the data and the opinion proffered.” Id. at 36, 236 A.3d at 651 (cleaned up). Meanwhile, “to constitute reliable methodology, an expert opinion must provide a sound reasoning process for inducing its conclusion from the factual data and must have an adequate theory or rational explanation of how the factual data led to the expert’s conclusion.” Id. at 32, 236 A.3d at 648-49 (cleaned up). Our case law shows that there is a distinction between data and methodology. Of course, a given circumstance cannot always be neatly categorized as one or the other and might instead pertain to both. Here, however, it is clear that the circuit court’s analysis did not implicate the reliability of the widely accepted before-and-after method of calculating lost profits and exceeded a review of the soundness of Ms. Cardell’s reasoning process. Although the circuit court appeared to fault the soundness of the process that Ms. Cardell employed, the circuit court’s underlying criticism was with the data Ms. Cardell relied on, her background and experience, and whether she was too closely associated with PNSI. Although KatzAbosch contended that Ms. Cardell’s testimony was unreliable because she mistreated member draws and insurance reimbursement rates, which in turn -6- influenced her to incorrectly identify 2015 as the benchmark year, Ms. Cardell’s testimony at the hearing established an evidentiary basis for categorizing member draws as expenses or profits based on the nature of the business, and her research and consultation with another expert in damages. She then followed that approach by examining the different types of member draws and categorizing them based on the factual circumstances. Nothing in the record supports the conclusion that this was the wrong approach or that this approach was unreliable. Rather, the dispute arises from KatzAbosch challenging the result or opinion that Ms. Cardell reached.3 I agree with the Appellate Court that determining whether to label member draws as expenses or profits is a fact-laden decision suitable for cross-examination. See Parkway, 255 Md. App. at 633, 283 A.3d at 774. The record demonstrates that Ms. Cardell explained the “judgment calls” that she made in selecting 2015 as the base year. Unlike in CDW LLC v. NETech Corp., 906 F. Supp. 2d 815, 823-24 (S.D. Ind. 2012)—where the expert for the plaintiff technology company offered no explanation for his use of the plaintiff’s other branches in the Great Lakes region as a benchmark for determining how much the plaintiff’s Indianapolis branch would have grown but for the defendant’s alleged wrongful conduct—here, Ms. Cardell 3 In my view, categorizing all member draws as profits would have resulted in a higher estimate of lost profits—so, Ms. Cardell’s decision on this point is consistent with her other choices designed to arrive at reasonable but conservative estimates of PNSI’s lost profits. And, contrary to KatzAbosch’s contention, it is unlikely that treating all member draws as profits would have affected Ms. Cardell’s choice of 2015 as the benchmark, because such a classification would not have changed the underlying facts that she relied on in that decision: the unlikelihood of expenses repeating in the future, the growth of the business and its revenue trends, and the growth of the market. -7- explicitly described her process for concluding that 2015 was the appropriate benchmark.4 She testified about PNSI’s revenue trends, the growth-associated costs, her market analysis, and the profit in 2016, despite doctors leaving, as the basis for her conclusion that 2015 reflected a reasonable prediction for the future of PNSI absent the doctors’ departures. Ms. Cardell’s description of her method is not the kind of unsupported conjecture excluded in CDW, where the expert: (1) “offer[ed] no reason” for choosing the plaintiff’s “other branches as appropriately comparable to [its] Indianapolis branch[,]” for which he was calculating lost profits; (2) “avoid[ed] explaining how another branch’s experience in any year” was an appropriate point of comparison; (3) failed to establish that the market forces for the Indianapolis branch and other branches were comparable; (4) used the average of all of the other branches’ growths in revenue, even though there were “wide variations in branch performance from year to year” and, thus, “the cumulative growth of 4 The circuit court supplemented its ruling from the bench with a written order discussing the persuasive value of a decision from the United States District Court for the Southern District of Indiana, CDW, 906 F. Supp. 2d 815. The circuit court reasoned that the base year established by Ms. Cardell was akin to the benchmark established by the excluded expert in CDW, which the District Court in that case had rejected as too speculative because it was based on average earnings at other branches of the affected business that were not sufficiently comparable to the branch that allegedly suffered lost profits. In CDW, the District Court excluded the expert’s testimony on lost profits calculated under a different method of calculating lost profits—the yardstick method. See CDW, id. at 823-25. The District Court noted that the yardstick method requires use of benchmarks “sufficiently comparable that they may be used as accurate predictors of what the target would have done.” Id. at 824. However, the District Court concluded that the expert did not select such a benchmark when he averaged the growth rates of plaintiff’s other branches in the Great Lakes region without any explanation of why those branches, much less their average, was appropriate for comparison. See id. The District Court also held that other information analyzed by the expert tended to demonstrate that his analysis was “way off base.” Id. at 825. -8- all the included branches d[id] not approximate the growth of any individual branch”; and (5) analyzed, as an alternative measure of lost profits, information that tended to show that his predicted revenue growth was “way off base.” Id. at 824-25 (cleaned up). In contrast, here, Ms. Cardell explained why 2015 and not 2014 was the appropriate base year, described how her market research led her to incorporate market forces into her analysis, and indicated that PNSI’s profitability, its period of growth, and other factors led her to believe that the profitability in 2015 would have been replicated in the ensuing years but for the doctors’ departures, and no other evidence undermined her analysis. This is not a case of the expert arriving at a conclusion by ipse dixit—as discussed above, Ms. Cardell fully explained her reasoning for selecting 2015 as the base year and classifying some member draws as expenses based on her years of experience as a certified public accountant conducting similar analyses, doing market research, and consulting another expert. The circuit court rejected Ms. Cardell’s conclusion about 2015 being the appropriate base year by faulting her data, reasoning that she received too much input from PNSI, and questioning the nature of her experience. Ms. Cardell’s use of PNSI’s accounting data and consultation with its staff did not render her methodology unreliable. As noted by the Appellate Court, “‘[t]hese are the kinds of data that accountants normally rely on in calculating future earnings [and lost profits].’” Parkway, 255 Md. App. at 630, 283 A.3d at 772-73 (quoting Manpower, Inc. v. Ins. Co. of Pa., 732 F.3d 796, 809 (7th Cir. 2013)) (second alteration in original). It is difficult to imagine what other data might be more reliable or relevant to determining the trajectory of a particular business. Indeed, even in CDW, 906 F. Supp. 2d at 822, the -9- District Court was unpersuaded by the defendant’s contention that the plaintiff’s expert “blindly accepted as accurate, and without any verification, the figures supplied to him by” the plaintiff. The District Court concluded that CDW was unlike cases in which “experts were merely ‘parroting’ unsubstantiated views of their clients[.]” Id. The same logic applies here. To the extent that the circuit court expressed concern about Ms. Cardell potentially being biased, this clearly was a matter for the jury’s consideration. As we observed in Falik v. Hornage, 413 Md. 163, 179, 991 A.2d 1234, 1244 (2010), “[a]n expert’s testimony can have a compelling effect on a jury[,]” and “[t]hat is why, especially with expert witnesses, wide latitude must be given a cross-examiner in exploring a witness’s bias or motivation in testifying[.]” (Cleaned up). The circuit court undeniably questioned Ms. Cardell’s “experience for this particular matter,” which was a consideration for the court in assessing the reliability of her methodology. Like the Appellate Court, I would reject the premise that “knowledge of the accounting principles . . . required to analyze profits for a . . . medical practice differs so markedly from that required for other small businesses . . . that the proffered [certified public accountant] must demonstrate particularized experience[.]” Parkway, 255 Md. App. at 622, 283 A.3d at 768 (citations omitted). In addition, consideration of an expert’s level of experience is not one of the Daubert-Rochkind factors adopted by this Court. Although the list of factors has been described as not being “exhaustive[,]” State v. Matthews, 479 Md. 278, 314, 277 A.3d 991, 1012 (2022) (citations omitted), Maryland Rule 5-702(1) provides that, in assessing the admissibility of expert testimony, a trial “court shall - 10 - determine[] whether the witness is qualified as an expert by knowledge, skill, experience, training, or education[.]” In this case, the circuit court did not do that. My concern is that the Daubert-Rochkind factors should not be used to exclude an expert’s testimony based on the expert’s experience or training without making a finding under Maryland Rule 5702(1). Finally, although the circuit court emphasized that its conclusion that Ms. Cardell’s testimony would not be useful to the jury was “a very, very slight factor” in its ruling, one of the reasons that the circuit court expressed for finding that Ms. Cardell’s testimony would not be useful was that the jury might not find that all of the departing doctors left because of KatzAbosch’s actions, which it reasoned would render Ms. Cardell’s all-ornothing lost profits estimate unhelpful. Aside from this being another criticism of the data that Ms. Cardell relied on, there is a fundamental problem with the circuit court’s conclusion that Ms. Cardell’s testimony was not helpful because the jury might not accept PNSI’s contention that all of the doctors left because of KatzAbosch. “[A] trial court is not permitted to resolve disputes of material fact in determining whether a sufficient factual basis exists to support an expert’s opinion. Doing so is a clear abuse of discretion.” Jamaiya Oglesby v. Balt. Sch. Assocs., et al., No. 26, Sept. Term, 2022, 2023 WL 4755689, at *16 (Md. July 26, 2023). Under the circuit court’s logic, in Gen. Elec. Co. v. Joiner, 522 U.S. 136, 146-47 (1997), the trial court could have excluded the expert testimony, not because of any analytical gap between the data and the opinion, but instead because the jury might have found that the plaintiff never encountered the chemicals that allegedly contributed to his cancer. - 11 - Under the circumstances of the limited remand due solely to the circuit court’s consideration the 2021 normalizing adjustment, the circuit court would be free to continue to consider matters such as Ms. Cardell’s lack of specialized experience, whether she may be inherently biased, whether she relied too heavily on data from PNSI, and whether she relied on information that a jury might reject.5 All of these circumstances go to the weight to be given to an expert’s testimony, not admissibility. All of them are for the jury, not the circuit court, to consider. The limited remand will essentially be a redo of the motion hearing, giving the circuit court an opportunity to once again take these matters into account as the record establishes that these considerations reflect what the circuit court thinks of Ms. Cardell’s testimony. In determining admissibility of expert testimony, a trial court is “a gatekeeper,” not “an armed guard[,]” and, when expert “testimony rests upon good grounds, based on what is known, it should be tested by the adversary process – competing expert testimony and active cross-examination – rather than excluded from jurors’ scrutiny for fear that they will not grasp its complexities or satisfactorily weigh its inadequacies.’” Matthews, 479 Md. at 322-23, 277 A.3d at 1017 (quoting Ruiz-Troche v. Pepsi Cola of Puerto Rico Bottling Co., 161 F.3d 77, 85-86 (1st Cir. 1998)) (internal quotation marks omitted). In this case, the better course of action is to affirm the decision of the Appellate Court, which correctly reversed and remanded for further proceedings in 5 In my view, it was not for lack of effort or acumen that the circuit court veered off course by taking these considerations into account—to the contrary, this is a complicated case with a lot of moving parts. The record reflects that the circuit court set forth its ruling on the motion to exclude, carefully laid out what it genuinely perceived to be issues with Ms. Cardell’s testimony, turned to thoroughly addressing the Daubert-Rochkind factors, and attempted to render the best possible decision that it could under the circumstances. - 12 - which Ms. Cardell’s testimony would be admissible. See Parkway, 255 Md. App. at 639, 283 A.3d at 778. For the above reasons, respectfully, I dissent. - 13 -
Primary Holding

In this professional malpractice action, the Supreme Court held that the trial court erred in excluding the expert testimony of a witness.


Disclaimer: Justia Annotations is a forum for attorneys to summarize, comment on, and analyze case law published on our site. Justia makes no guarantees or warranties that the annotations are accurate or reflect the current state of law, and no annotation is intended to be, nor should it be construed as, legal advice. Contacting Justia or any attorney through this site, via web form, email, or otherwise, does not create an attorney-client relationship.

Some case metadata and case summaries were written with the help of AI, which can produce inaccuracies. You should read the full case before relying on it for legal research purposes.

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.