Katz, Abosch, etc., P.A. v. Parkway Neuroscience ( 2023 )


Menu:
  • Katz, Abosch, Windesheim, Gershman & Freedman, P.A., et al. v. Parkway Neuroscience
    and Spine Institute, LLC, No. 30, September Term, 2022. Opinion by Biran, J.
    EXPERT WITNESSES – ADMISSIBILITY OF EXPERT TESTIMONY –
    MARYLAND RULE 5-702 – LIMITED REMAND – Respondent filed a lawsuit against
    Petitioners alleging accountant malpractice and related claims. In the course of discovery,
    Respondent designated an expert to provide an opinion concerning Respondent’s lost
    profits resulting from Petitioners’ alleged torts. Petitioners moved to exclude the testimony
    of the proffered expert under Daubert v. Merrell Dow Pharmaceuticals, Inc., 
    509 U.S. 579
    (1993), and Rochkind v. Stevenson, 
    471 Md. 1
     (2020). After conducting a Daubert-
    Rochkind hearing, the circuit court granted Petitioners’ motion to exclude the proffered
    expert testimony. The Supreme Court of Maryland held that much of the trial court’s
    consideration of the Daubert-Rochkind factors was appropriate, including the trial court’s
    assessment of how the expert’s choice of data, assumptions, and other inputs affected the
    reliability of her methodology. However, the trial court erred when it considered the
    expert’s “normalizing adjustments” that recategorized certain expenses from one year to
    another, as reflecting on the reliability of the expert’s methodology. The Court ordered a
    limited remand to the circuit court under Maryland Rule 8-604(d)(1) so that the trial court
    may decide to admit or exclude the expert’s testimony without consideration of the
    normalizing adjustments as reflecting on the reliability of the expert’s methodology.
    Circuit Court for Howard County
    Case No.: C-13-CV-18-000181
    Argued: May 4, 2023
    IN THE SUPREME COURT
    OF MARYLAND*
    No. 30
    September Term, 2022
    KATZ, ABOSCH, WINDESHEIM,
    GERSHMAN & FREEDMAN, P.A., ET AL.
    v.
    PARKWAY NEUROSCIENCE
    AND SPINE INSTITUTE, LLC
    Fader, C.J.
    Watts
    Hotten
    Booth
    Biran
    Gould
    Eaves,
    JJ.
    Opinion by Biran, J.
    Booth, J., concurs.
    Gould, J., concurs in part and dissents in part.
    Pursuant to the Maryland Uniform Electronic Legal Materials                   Watts, J., dissents.
    Act (§§ 10-1601 et seq. of the State Government Article) this
    document is authentic.
    2023-08-30
    12:44-04:00                                             Filed: August 30, 2023
    Gregory Hilton, Clerk
    * At the November 8, 2022 general election, the voters of Maryland ratified a constitutional
    amendment changing the name of the Court of Appeals of Maryland to the Supreme Court
    of Maryland. The name change took effect on December 14, 2022.
    When this Court adopted the Daubert1 expert testimony admissibility standard in
    Rochkind v. Stevenson, 
    471 Md. 1
     (2020), we embraced a regime that prizes the reliability
    of an expert’s methodology over its general acceptance. We empowered trial judges to
    protect juries from junk science while also broadening the range of possibly admissible
    opinions beyond just those dominant among practitioners. We asked judges to engage with
    the science without playing amateur scientist, and we promised the deference appropriate
    to courts administering a flexible approach to analyzing the admissibility of expert
    testimony. This case requires us to reflect on that flexibility and deference.
    Parkway Neuroscience and Spine Institute, LLC (“PNSI”, the Respondent here) is
    a medical and surgical practice that began to expand in 2011 and needed accounting help.
    In 2013, PNSI retained accounting firm Katz, Abosch, Windesheim, Gershman &
    Freedman, P.A. and, specifically, Mark Rapson, who specialized in medical practice
    accounting (we shall refer to the firm and Mr. Rapson, the Petitioners here, collectively as
    “KatzAbosch”). Within a few years after retaining KatzAbosch, PNSI began to
    disintegrate; members of the practice began leaving in 2015, and by the middle of 2016,
    only two members remained of the nine who had been in place at the end of 2014. PNSI
    terminated KatzAbosch’s services in 2015.
    PNSI alleges that malpractice by KatzAbosch caused the mass exodus of its
    members. In 2018, PNSI sued KatzAbosch in the Circuit Court for Howard County to
    recover damages for lost profits. To establish those damages, PNSI designated certified
    1
    See Daubert v. Merrell Dow Pharmaceuticals, Inc., 
    509 U.S. 579
     (1993).
    public accountant Meghan Cardell as an expert witness. She used the widely accepted
    “before-and-after” method to calculate PNSI’s lost profits, choosing 2015 as a “baseline”
    period against which she would compare the actual profits in subsequent years through
    2019, and adding up the differences to arrive at an estimate of what profits PNSI missed
    out on due to KatzAbosch’s alleged harmful conduct. A few weeks before the June 2021
    Daubert-Rochkind hearing, Ms. Cardell issued updated calculations reflecting some
    “normalizing adjustments” she had made; although PNSI’s accounting records had not
    changed since her initial analysis, Ms. Cardell reviewed PNSI’s financial information again
    and noticed some payments that had been categorized in the wrong years. She reallocated
    those payments to the years she believed to be correct and updated her calculations.
    Those two issues – Ms. Cardell’s choice of 2015 as the “before” in her “before-and-
    after” analysis and her June 2021 normalizing updates – rose to the top of the trial court’s
    mind in the Daubert-Rochkind hearing. The trial court noted speculative and insufficiently
    substantiated judgment calls that Ms. Cardell had made in arriving at the 2015 benchmark.
    Among other things, the trial court wondered why Ms. Cardell had chosen 2015 (a
    profitable year) rather than, say, an average that included the several (unprofitable) years
    prior to the alleged harm event. The trial court also was concerned about Ms. Cardell’s
    inability to articulate industry standards relating to the concept of “economic impact” and
    to the proper treatment of owner draws.
    In addition to these points, the trial court commented several times about Ms.
    Cardell’s June 2021 normalizing adjustments, which negatively affected its opinion of Ms.
    Cardell’s reliability. Essentially, the court did not understand why it had taken Ms. Cardell
    2
    so long to notice the errors. The court discussed these adjustments when considering the
    Daubert factors relating to a methodology’s error rate and to whether the field of expertise
    claimed by the expert is known to reach reliable results for the type of opinion the expert
    would give.
    Based on its application of the Daubert-Rochkind factors, the trial court excluded
    Ms. Cardell’s testimony, leading to summary judgment in favor of KatzAbosch because
    PNSI could not prove damages.
    PNSI appealed, and the Appellate Court of Maryland2 held that the circuit court
    abused its discretion in finding Ms. Cardell’s methodology unreliable. As to the 2015
    baseline choice, the Appellate Court agreed with PNSI that the choice was a question of
    data (and thus a factual question for the jury) rather than of methodology. With respect to
    the normalizing adjustments, the Appellate Court said that Daubert’s “error rate” factor
    must be understood as the rate of unknown errors in the methodology employed, not as an
    “error correction rate,” or else courts would create incentives against experts disclosing
    and explaining errors they made. The intermediate appellate court reversed the trial court’s
    exclusion of Ms. Cardell’s expert testimony and remanded for consistent proceedings.
    KatzAbosch petitioned this Court for further review.
    As we explain more fully below, the choice or calculation of the inputs to a
    methodology can be a part of the methodology itself, and we reject an unduly rigid dividing
    2
    At the November 8, 2022 general election, the voters of Maryland ratified a
    constitutional amendment changing the name of the Court of Special Appeals of Maryland
    to the Appellate Court of Maryland. The name change took effect on December 14, 2022.
    3
    line between “data” and “methodology” that binds courts to admit methodologically
    questionable analyses cloaked as data. To the extent the trial court considered how Ms.
    Cardell’s choice of data, assumptions, and other inputs affected the reliability of her
    methodology, the trial court’s Daubert-Rochkind analysis was proper. However, the trial
    court erred in its consideration of the normalizing adjustments as reflecting on the
    reliability of Ms. Cardell’s methodology, as opposed to the credibility (or reliability) of
    Ms. Cardell herself. After a careful review of the record, we determine that the fair and
    prudent course of action at this point is to remand the case to the circuit court to decide
    whether to admit or exclude Ms. Cardell’s testimony without consideration of the June
    2021 normalizing adjustments as reflecting on the reliability of Ms. Cardell’s methodology.
    I
    Background
    A. Facts
    During the period relevant to this case, PNSI was a Western Maryland and
    Pennsylvania mixed medical practice that diagnosed, managed, and treated disorders of the
    brain, spine, and peripheral nervous system. It employed neurosurgeons, interventional and
    non-interventional pain physicians, neurologists, physicians’ assistants, and support staff.
    The practice had operated since 1998. Beginning in 2011, PNSI expanded, hiring more
    physicians and support staff. These efforts caused PNSI to spend more on salaries and
    build-out expenses without offsetting revenue. At the end of 2014, the practice had nine
    member-owners, all physicians.
    4
    1. The Engagement of KatzAbosch
    None of these members, however, were accounting or finance experts. In early
    2013, PNSI’s long-time accountant advised PNSI that the practice had outgrown his firm’s
    services and recommended that the practice retain a new accounting firm to help guide
    PNSI through its growth and expansion process. So PNSI began searching for a firm that
    specialized in medical practice accounting and finance. In October 2013, PNSI retained
    KatzAbosch to provide tax, accounting, and financial advice and services, as well as to
    provide “expert business and financial guidance and direction to help PNSI continue to
    grow its practice.” The engagement included analyzing PNSI’s general ledger and financial
    statements, making recommendations concerning PNSI’s financial affairs, and designing
    and administering a new member compensation model. Mr. Rapson (chair of KatzAbosch’s
    Medical Services Group) was responsible for the account.
    PNSI’s 2012 Operating Agreement provided the compensation terms for its
    member-physicians. First, each member was to maintain a capital account on the books of
    the practice, with a minimum balance that PNSI’s Board of Managers (the “Board”) would
    determine annually. Second, members were to be paid a monthly “draw” (determined by
    the Board) at the start of each year – in essence a form of salary. Third, revenue received
    by the practice for hospital and trauma calls would be distributed to the member who had
    taken the call, with PNSI functioning as a pass-through entity. These pass-through
    payments, along with the monthly draws, comprised “guaranteed payments” to the
    members. Finally, each quarter, the Board would distribute any excess cash flow
    (“distributions”) in proportion to each member’s ownership stake.
    5
    In early 2014, KatzAbosch designed and proposed a new compensation model, and
    PNSI adopted it. According to PNSI, KatzAbosch’s model did not reserve funds for known
    build-out-related expenses concerning one of PNSI’s locations that would be coming due
    later that year as well as other significant expenses. PNSI alleges that, despite these
    expenses coming due, KatzAbosch directed PNSI to make almost $1 million in quarterly
    distributions to members between July and October 2014.
    2. Termination of KatzAbosch and the Departure of Most of PNSI’s Members
    PNSI alleges that, as a result of KatzAbosch’s erroneous advice, PNSI almost ran
    out of money by the end of 2014. According to PNSI, the practice had to use over $660,000
    from its line of credit in the fourth quarter of 2014 alone. In January 2015, KatzAbosch
    disclosed to PNSI its precarious financial situation. While PNSI had had almost $1 million
    in cash on hand at the start of 2014, it had less than $40,000 by the start of 2015. PNSI
    terminated KatzAbosch’s services in the spring of 2015.
    The practice then began to come apart. PNSI alleges that members – who had
    personally guaranteed loans the practice had taken – were increasingly distressed about the
    financial condition of the practice and their personal liability. Starting in mid-2015,
    members began to withdraw from the practice, taking with them patients and the associated
    revenue streams. By mid-2016, only two members – Dr. Brian Holmes and Dr. Neil
    O’Malley – remained of the nine who had been in place at the end of 2014. KatzAbosch
    claims that Dr. Holmes and Dr. O’Malley received more compensation from the practice
    after they were the only two members remaining (thereby depressing estimates of PNSI’s
    6
    profits post-exodus), although PNSI explains this as the result of those two doctors taking
    more trauma calls and thus receiving more in guaranteed payments.
    B. The Lawsuit
    In 2018, PNSI sued KatzAbosch in the Circuit Court for Howard County, alleging
    claims for accountant malpractice, negligent misrepresentation, breach of contract, and
    unjust enrichment. PNSI initially sought damages for lost profits, settlement amounts paid
    to a departing member, written-off amounts owed by two departing members, fees paid to
    KatzAbosch, and prejudgment interest, all totaling $9,456,035. The bulk of the claimed
    damages was the alleged lost profits.
    1. Meghan Cardell, PNSI’s Damages Expert
    In July 2019, PNSI designated Meghan Cardell as an expert witness expected to
    testify that “PNSI suffered and will continue to suffer significant financial losses resulting
    from the mass exodus of members and staff, loss of patients, replacement of long-time
    established physicians with newer-practicing physicians, and litigation with withdrawing
    members, as identified in the Calculation of Damages[.]” Ms. Cardell is a certified public
    accountant and a certified fraud examiner. At the time of her designation by PNSI as an
    expert in this case, Ms. Cardell was Director for Disputes and Investigations at the
    Washington, D.C. accounting firm Alvarez & Marsal, where she had worked since 2014.
    Before that, she was a senior associate in forensics at Veris Consulting from 2011 to 2014.
    Ms. Cardell used the “before-and-after” methodology of calculating PNSI’s lost
    profits. Under this approach, the expert compares profits before and after a damaging event,
    7
    the former being the “benchmark” or “base” period and the latter being the “loss” period.
    Ms. Cardell initially calculated PNSI’s damages in May 2019, as follows:
    Damages (less prejudgment interest)
    (Calculated in May 2019)
    Total lost profits                          $8,520,744
    Dr. DeMarco Settlement                      $89,421
    Dr. Sullivan Amount Owed                    $84,836
    KatzAbosch Fees                             $182,010
    Total:                                      $8,877,012[3]
    Prejudgment interest
    Lost profits                              $503,999
    DeMarco & Sullivan                        $20,185
    KatzAbosch Fees                           $54,839
    Total prejudgment interest:               $579,023
    Total Damages                               $9,456,035
    The lost profits total included figures for 2016 through 2025. Ms. Cardell selected
    2015 as the base year. She then used actual figures to calculate the lost profits for 2016,
    2017, and 2018; she used the trend line from those three years to project lost profits through
    2025, discounting the figures by 20% in 2019-22 and 50% in 2023-25 to reflect a trend of
    declining annual lost profits as the practice would gain doctors, patients, and income in the
    future.
    On July 26, 2019, KatzAbosch moved to strike PNSI’s claim and to exclude Ms.
    Cardell’s testimony. At a hearing in June 2020, the trial court (the Honorable Richard S.
    Bernhardt, specially assigned) denied KatzAbosch’s motion, but expressed concern about
    admitting Ms. Cardell’s calculations at trial if she continued to project lost profits for 2020
    3
    This total amount is one dollar greater than the sum of the four amounts that
    precede it. It is not clear whether this discrepancy was due to a rounding error or some
    other issue.
    8
    through 2025 related to the 2015-16 departure of physicians, despite the changed landscape
    created by the COVID-19 pandemic.
    On May 5, 2021, KatzAbosch renewed its motion to strike PNSI’s lost profit claims
    and to exclude Ms. Cardell’s testimony under this Court’s recently adopted Daubert-
    Rochkind standard for expert testimony, disputing the factual basis of Ms. Cardell’s
    Calculation of Damages report.
    Ms. Cardell updated her Calculation of Damages on May 17, 2021. PNSI no longer
    sought damages for 2020 or beyond, and Ms. Cardell’s new calculations omitted those
    years. The updated calculations featured a revised damages calculation of $7,335,447:
    Damages (less prejudgment interest) – Updated May 17, 2021
    (dropping 2020 and onward; concerning only 2016-19 pre-pandemic; using finalized
    actual 2019 figures)
    Total lost profits                         $8,520,744 $5,789,521
    Dr. DeMarco Settlement                     $89,421
    Dr. Sullivan Amount Owed                   $84,836
    KatzAbosch Fees                            $182,010
    Total:                                     $8,877,012 $6,145,789
    Prejudgment interest
    Lost profits                             $503,999   $1,078,968
    DeMarco & Sullivan                       $20,185    $37,630
    KatzAbosch Fees                          $54,839    $73,060
    Total prejudgment interest:              $579,023   $1,189,658
    Total Damages                              $9,456,035 $7,335,447
    9
    On May 27, 2021, the trial court scheduled a full-day, in-person evidentiary hearing
    on KatzAbosch’s renewed motion for June 30, 2021. On June 11, 2021, Ms. Cardell
    updated her calculations again:
    Damages (less prejudgment interest) – Updated June 11, 2021
    Total lost profits                        $8,520,744 $5,789,521 $4,956,080
    Dr. DeMarco Settlement                    $89,421
    Dr. Sullivan Amount Owed                  $84,836
    KatzAbosch Fees                           $182,010
    Total:                                    $8,877,012 $6,145,789 $5,312,348
    Prejudgment interest
    Lost profits                            $503,999     $1,078,968 $890,837
    DeMarco & Sullivan                      $20,185      $37,630      $38,346
    KatzAbosch Fees                         $54,839      $73,060      $73,808
    Total prejudgment interest:             $579,023     $1,189,658 $1,002,991
    Total Damages                             $9,456,035 $7,335,447 $6,315,339
    Notably, the June 2021 updates eliminated the loss from 2016 (meaning Ms. Cardell had
    now found that year to be profitable) and showed lost profits only from 2017-19. These
    changes reflected “normalizing adjustments” discussed in detail below.
    2. The Daubert-Rochkind Hearing
    The trial court conducted a Daubert-Rochkind hearing on June 30, 2021. The crux
    of PNSI’s argument was that KatzAbosch’s concerns with Ms. Cardell dealt less with her
    methodology and more with the assumptions she had made, and those assumptions went
    to weight rather than admissibility. The court declined either to accept or to reject Ms.
    Cardell’s qualifications as an expert, although it eventually said her experience (or lack
    thereof) with “niche” medical practices bore somewhat on the reliability of her testimony,
    independent of the question of her qualification as an expert witness.
    10
    a. Issues Raised in the Hearing
    i.   2015 as the Base Year
    After describing why other methodologies to determine contractual damages and
    lost business value (and, relatedly, lost future profits) were not applicable here, Ms. Cardell
    described the before-and-after methodology that she chose to apply. She testified that this
    method “in [her] experience is the most commonly used and most commonly accepted
    methodology of measuring lost profits,” that she had used it in her career “many, many
    times,” and that she had seen it used by other experts in and out of litigation settings just
    as frequently. She oriented her analysis around the harm event of seven of nine doctors
    withdrawing from the practice within a short time; the first doctor left in June 2015, a few
    more at the end of 2015, and a few more by June 2016. So calendar year 2015 was her
    benchmark “before” period, and the “after” period (or “loss” period) began in calendar year
    2016 and ran through 2019. She described this as a “conservative proxy” for the practice’s
    future earnings, had so many members not left.
    The business had been profitable in 2015, and PNSI had been investing to grow;
    Ms. Cardell said “[t]he practice had sort of hit its stride in 2015.” She said medical
    businesses are not subject to swings of consumer preference that might have destabilized
    the results of those investments, and the medical specialty industry was projected to grow
    because of aging Baby Boomers and the prevalence of chronic illness.
    The trial court articulated some concerns with Ms. Cardell’s methodology. First, the
    court asked why she had chosen just one year (2015) rather than a benchmark of multiple
    years before the harm event; Ms. Cardell said the benchmark could be an average or could
    11
    be one year, particularly if there were past years that would not be representative of what
    future earnings would look like. Second, the court noted that 2015 had been the most
    profitable year since 2010,4 and the court did not understand why 2015, rather than a slower
    year, should become the “new normal,” as there might have been post-harm years that also
    would have been a “squeeze.” Ms. Cardell believed 2014 was not entirely unusual but
    would have been unrepresentative because 2015 was the year when PNSI had emerged
    from the phase of significant expenditures (and reduced profits) in order to grow the
    practice. Third, the court questioned how Ms. Cardell knew that PNSI had truly hit its stride
    in 2015 and that there would not have been significant future problems caused by
    overexpansion – personality conflicts between the new doctors, sub-par new facilities, etc.
    She answered that the practice had been profitable in 2016 even as the exodus had begun,
    showing that it could still be profitable despite some amount of dysfunction and confirming
    that it would have continued to generate profits in the future without the full exodus.
    ii.   Other Data and Assumptions (Reimbursement Rates)
    Ms. Cardell assumed that, but for the harm event, PNSI’s profits in 2017, 2018, and
    2019 would have at least equaled the 2015 profits. Ms. Cardell had not analyzed changes
    in medical service reimbursement rates (which had been decreasing) during the relevant
    years, although the market research she examined showed that medical industry revenues
    were expected to grow because of expanded access to Medicare, Medicaid, and private
    health insurance under the Affordable Care Act. The core of the trial court’s concern was
    4
    As measured by Ms. Cardell, PNSI had had a $22,000 profit in 2010, losses in
    2011-14, and a $321,751 profit in 2015.
    12
    whether Ms. Cardell had looked “at all of the various revenue threads that come up with
    the fabric of [PNSI’s] yearly income” or whether she had just been “looking at numbers
    without understanding from further records or where those numbers came from[.]” Ms.
    Cardell said that her analysis was based on revenues.
    iii.   Correction of Calculations
    Ms. Cardell described normalizing adjustments she had made to her calculations
    earlier in that month relating to trauma and on-call payments. In 2016, PNSI had made
    payments to its physicians based on trauma and on-call revenues the practice had received
    in 2015. Ms. Cardell initially included those payments as expenses of the practice for 2016.
    After Ms. Cardell discovered that the practice had received the associated revenues in 2015,
    she removed that expense from 2016 and added it to 2015. This reduced by half PNSI’s
    adjusted income in base year 2015 by increasing adjusted expenses. Ms. Cardell’s
    $395,010 downward adjustment in expenses for 2016 increased adjusted income for that
    year, making 2016 profitable (compared to her earlier May 2021 estimates for 2016, which
    had shown a loss) and contributing to the elimination of any loss in profits in 2016.
    There were also trauma and on-call adjustments for 2017, 2018, and 2019. In those
    years, Dr. Holmes and Dr. O’Malley earned income based on their trauma and on-call
    services, and that money flowed to the practice for normal pass-through purposes, but they
    chose not to pay themselves those funds; rather, they effectively loaned those amounts to
    the practice in order to increase its cash on hand. In the past, those guaranteed payments
    would have been recorded as expenses, so Ms. Cardell included them as such via the June
    2021 adjustments. Her first updated calculations (issued May 2021) had included upward
    13
    trauma and on-call expense adjustments for years 2017 ($129,573), 2018 ($767,573), and
    2019 ($767,573). Each of these upward adjustments had the effect of increasing adjusted
    expenses and therefore reducing adjusted income, making the lost profits figures (and thus
    PNSI’s alleged damages) larger for each year. Ms. Cardell’s second updated calculations
    (issued June 2021) decreased the upward expense adjustments for 2017 (to $77,000) but
    increased the upward expense adjustments for 2018 (to $842,400) and 2019 (to $1,233,617)
    in comparison to her first updated calculations. For all three years, the post-update upward
    adjustments still increased adjusted expenses, decreased adjusted income, and increased
    lost profits (and damages).
    Although the information that led to these adjustments had long been available to
    her, Ms. Cardell explained that she looked again at the numbers ahead of the Daubert-
    Rochkind hearing and identified a payment that looked like it belonged to a different year.
    A PNSI accounting manager confirmed this intuition, and so Ms. Cardell made the
    judgment that the changes relating to trauma and on-call payments/expenses were needed.
    The trial court expressed concern that Ms. Cardell’s calculations might be liable to
    additional such changes; Ms. Cardell testified that she did not believe any other
    adjustments were necessary.
    KatzAbosch’s counsel attempted to characterize Ms. Cardell’s updates to her
    calculations (from May 2021 to June 2021) as a “fifty percent error rate” because she had
    made revisions between the first and second updates.
    14
    iv.   Treatment of Member Draws
    The trial court noted two different ways that a business could treat member draws
    when calculating profits. To illustrate the point, the court referred to a hypothetical limited
    liability corporation that, after paying all its expenses other than potential draws to its
    owner, has $100 in cash. On the one hand, the owner could withdraw the $100 as salary
    and the business would show no profit that year. On the other hand, the business could be
    considered to have $100 in profits regardless of whether the owner withdraws it from the
    business that year. Ms. Cardell said that she believed the latter was the correct way to look
    at it, but there was no industry standard on this issue one way or the other. The trial court
    expressed skepticism that there was no industry standard, given that the classification of
    draws (owner salary) as profits or expenses “seems like a pretty basic issue that [is] capable
    of rearing its head in every case in which … an owner’s draw is possible.”
    v.    Lack of Member-Specific Lost Profits Calculation
    Ms. Cardell was “unable to parse out” the financial effects of any particular
    member’s departure from the practice. She could not recall whether she had looked at
    specific collections for Dr. Holmes and Dr. O’Malley, her analysis having been based on
    overall revenues for the practice. Defense counsel presented her with PNSI financial
    records showing that those doctors’ post-exodus revenue collections declined from 2016
    through 2019.
    b. The Trial Court’s Ruling
    PNSI acknowledged that “reasonable minds can differ” as to whether its profits in
    2015 was the appropriate benchmark against which to measure the practice’s future profits.
    15
    However, PNSI argued that none of the Daubert-Rochkind factors militated toward
    excluding Ms. Cardell. The court ruled from the bench, granting KatzAbosch’s motion to
    exclude Ms. Cardell’s testimony, based on the Daubert-Rochkind standard.
    i.   The Court’s Overall Sense
    The trial court began its ruling by acknowledging the wide acceptance of the before-
    and-after methodology for calculating lost profits. The court then highlighted its prime
    concerns with Ms. Cardell’s testimony: speculation, ipse dixit “judgment calls,”
    helpfulness to the jury, and information Ms. Cardell had failed to consider.
    First, the court found that Ms. Cardell’s selection of profitable 2015 as a benchmark
    (rather than unprofitable 2011-14, marginally profitable 2010, or an average) was
    speculation, especially given that she had no specialization in a niche practice like PNSI.
    So, too, was Ms. Cardell’s analysis that, after years of unprofitability, 2015 marked the
    critical turning point and that, in the court’s words, “the rocket had left the launch pad and
    was going straight up.” The speculation finding was bolstered by the fact that Ms. Cardell
    relied heavily on direct oral communications with PNSI and its employees rather than
    information more concretely assessable by the jury.
    Second, the court pointed to unreliable and ipse dixit “judgment calls.” The court
    discerned possible bias and some degree of unreliability in the judgment calls that Ms.
    Cardell made without some clear authority as to (1) the June 2021 normalizing adjustments
    (which the court said Ms. Cardell could have made as part of her original calculations);
    (2) “whether owner draws reduce profits or not”; and (3) Ms. Cardell’s subjective
    16
    understanding of the term “economic impact.”5 These judgment calls took on special
    importance because, the court said, either Ms. Cardell lacked awareness of industry
    standards or no industry standards existed, heightening her reliance on unverified sources
    of information and her own say-so.
    Third, the court questioned how helpful Ms. Cardell’s testimony would be to the
    jury. Ms. Cardell chose not to disaggregate revenues according to each individual departing
    physician, and instead she considered lost profits in an “all or nothing” manner, rendering
    her opinion “only helpful if the jury accepts that each and every doctor of the seven … left
    solely because … of the acts of [KatzAbosch].” Relatedly, the trial court stated that “[n]ot
    every doctor, necessarily is going to make or contribute the same amount of money each
    and every year. There was nothing presented in the Report to, that considered whether or
    not the income, the revenue generating ability of the remaining doctors or the leaving
    doctors was considered [sic]. It was just, let’s take 2015 and go from there. Let’s just use
    the numbers as we get them without examining whether or not, you know, the doctors who
    have income would have had that income and things of that nature. And again, there’s
    nothing about her training to me that would qualify her to make those assumptions that all
    these numbers would not be affected by passage of time and the doctor’s passage of time.”
    5
    Ms. Cardell testified that she considers “economic impact” in her work. When the
    court asked her what “economic impact” means, Ms. Cardell began her response by saying,
    “the way I think about it is…” This raised a red flag for the trial court: “[Y]ou start off
    with, ‘the way I think of it’. What does your industry consider it to be?” Ms. Cardell could
    not provide a specific industry definition for the term.
    17
    Fourth, the court noted that Ms. Cardell had failed to consider certain important
    pieces of information, including: (1) whether Dr. Holmes and Dr. O’Malley had other
    income streams; and (2) how changing insurance reimbursement rates were influencing the
    practice’s profitability.
    ii.         The Daubert Factors
    The trial court then considered the Daubert-Rochkind factors:
    (1)      Testing – The court acknowledged that a “before-and-after”
    analysis was appropriate and testable generally, but questioned the
    testability of Ms. Cardell’s judgment calls, including why she had
    chosen 2015 as the base year and the meaning of “economic
    impact.”
    (2)      Peer review – The court did not find this factor relevant.
    (3)      Rate of error – The court did not find this factor relevant as it is
    normally applied, where there is some known rate of false positive
    or false negative results. But the court did note concern with Ms.
    Cardell’s June 2021 pre-hearing updates, as those changes were
    not caused by a change in the facts.
    (4)      Standards and controls – The court said “there was very little
    evidence of any standards of controls that exist,” in particular on
    economic impact and the treatment of owner draws in profit
    computations.
    18
    (5)   General acceptance – The court acknowledged that the before-
    and-after analysis to measure lost profits was generally accepted.
    (6)   Purpose: prior research or litigation – Although the court
    expressed that there is “no inherent negativity” to experts who
    develop their opinions for litigation purposes compared to those
    who develop their expertise for independent research, the court did
    express concerns about Ms. Cardell’s heavy reliance on oral
    communications with PNSI personnel.
    (7)   Unjustifiable extrapolation from an accepted premise – The
    court said: “This applies if the … premise that we’re talking about
    from which the unfounded conclusions roles [sic] would be the
    acceptance of 2015 as the benchmark[.]”
    (8)   Accounting for obvious alternative explanations – The court
    found that Ms. Cardell “clearly” had not accounted for alternative
    explanations – in particular, she had failed to consider doctor-
    specific revenue generation and changing reimbursement rates’
    effects on revenue.
    (9)   Care here as in professional non-litigation work – The court
    observed as to this factor: “I don’t find that to be applicable, I don’t
    know what to say about that and I have no reason to think that she
    blew this off as an inconsequential project, I mean she took this
    very seriously.”
    19
    (10)   Field known to reach reliable results for this type of opinion –
    The court “incorporate[d] everything [it had] said,” noting in
    particular the June 2021 updates (driven by subjective reasons
    rather than newly revealed facts) as “mak[ing] the whole reliability
    even that much more suspect.”
    In sum, the court said, PNSI had failed to meet its burdens primarily as to Ms.
    Cardell’s reliability and secondarily as to Ms. Cardell’s usefulness to the jury, although the
    court described the “usefulness” finding as a “very, very slight factor.”
    On the following day, the trial court issued a written supplement to its oral ruling in
    which it discussed the persuasive value of CDW LLC, et al. v. NETech Corp., 
    906 F. Supp. 2d 815
     (S.D. Ind. 2012). In CDW, the expert determined lost profits by using the “yardstick
    method” – i.e., he compared the subject business branch’s profits to the plaintiff-business’s
    other branches. The CDW Court noted that “[a]n expert’s choice in data sampling is at the
    heart of his methodology. A yardstick approach is an acceptably reliable method under
    Daubert for calculating lost profits only if the benchmarks (or yardsticks) are sufficiently
    comparable that they may be used as accurate predictors of what the target would have
    done.” CDW, 
    906 F. Supp. 2d at 824
     (internal citation omitted). “Absent the requisite
    showing of comparability, a damage model that predicts either the presence or absence of
    future profits is impermissibly speculative and conjectural.” 
    Id.
     (internal quotation marks
    and citation omitted). The trial court found the before-and-after and yardstick methods
    sufficiently similar and wrote that “the expert’s choice of a benchmark in CDW is
    analogous to the expert’s choice of 2015 as the benchmark in the instant case. … Thus
    20
    [this] Court cites CDW as having persuasive value in finding that [PNSI] has not satisfied
    Maryland Rule 5-702 and this Court’s decision to exclude this expert on Daubert grounds.”
    The parties subsequently filed a stipulation of dismissal, agreeing (among other
    things) that PNSI could not prove a prima facie case with respect to its claims for
    accountant malpractice, negligent misrepresentation, and breach of contract, in light of the
    trial court’s exclusion of Ms. Cardell’s expert testimony. The trial court entered summary
    judgment in favor of KatzAbosch and dismissed the case in its entirety.
    3. Appeal
    PNSI appealed to the Appellate Court of Maryland. In a reported opinion, the
    Appellate Court held that the circuit court had abused its discretion in finding Ms. Cardell’s
    methodology unreliable under Daubert-Rochkind. Parkway Neuroscience and Spine
    Institute, LLC v. Katz, Abosch, Windesheim, Gershman & Freedman, P.A., et al., 
    255 Md. App. 596
    , 623-37 (2022). The Appellate Court discerned error in the trial court’s criticisms
    of Ms. Cardell’s testimony based on: (1) the selection of 2015 as the base year;
    (2) insurance reimbursement rates; (3) standards for member draws; (4) the June 2021
    updates to her calculations; and (5) the lack of individual per-doctor lost profit figures. Id.
    at 623-37. The Appellate Court also interpreted the trial court’s comments regarding Ms.
    Cardell’s lack of experience with respect to specialty medical practices as a finding that
    she lacked the requisite qualifications to be accepted as an expert. See id. at 622-23.
    Specifically, the Appellate Court concluded that “[a]nalyzing reimbursement rates,
    selecting the base year, and relying on data from and conversations with PNSI all are issues
    with the soundness of the data” rather than with the reliability of Ms. Cardell’s
    21
    methodology. Id. at 627 n.9. The court further opined that “[w]hether Ms. Cardell should
    have deducted the member draws from the projected profits is something [KatzAbosch]
    can attack during cross-examination before a jury – but not at the Daubert-Rochkind
    hearing.” Id. at 633. The Appellate Court considered the proper treatment of member draws
    to be “a fact-laden issue – involving credibility, not reliability.” Id.
    As to the June 2021 updates, the Appellate Court determined that the trial court
    misapplied the “known or potential rate of error” Daubert factor, because “[t]he error rate
    that Daubert speaks of is the rate of unknown errors in the methodology employed, not an
    error correction rate.” Id. at 634 (internal quotation marks and citation omitted) (emphasis
    in original). To find methodological unreliability in an expert’s decision to correct errors
    in her earlier analysis “would be a disincentive to disclose and explain errors.” Id. at 635.
    Finally, regarding Ms. Cardell’s failure to calculate member-specific lost profits,
    the Appellate Court concluded that the circuit court looked improperly to causation when
    the purpose of the Daubert hearing was to assess the reliability of Ms. Cardell’s
    methodology – not whether PNSI had met its burden of proof on causation. Id. at 636.
    The Appellate Court reversed the circuit court’s decisions excluding Ms. Cardell,
    striking PNSI’s lost profits claim, and granting summary judgment. The Appellate Court
    remanded the case to the circuit court for proceedings consistent with its opinion. Id. at
    639.
    KatzAbosch then petitioned this Court for a writ of certiorari, which we granted on
    January 20, 2023. Katz, Abosch, Windesheim, Gershman & Freedman, P.A. v. Parkway
    Neuroscience and Spine Institute, LLC, 
    482 Md. 534
     (2023). KatzAbosch presents the
    22
    following question for our review: “Did the [Appellate Court] err in finding that the trial
    court abused its discretion in excluding expert testimony on lost profits?”
    II
    Standard of Review
    Appellate courts review a trial court’s decision concerning the admissibility of
    expert testimony under Maryland Rule 5-702 for abuse of discretion. See Rochkind, 471
    Md. at 10-11; State v. Matthews, 
    479 Md. 278
    , 305-06 (2022). As we said in Matthews:
    Under this standard, an appellate court does “not reverse simply because the
    ... court would not have made the same ruling.” Devincentz v. State, 
    460 Md. 518
    , 550 (2018) (internal quotation marks and citation omitted). “Rather, the
    trial court’s decision must be well removed from any center mark imagined
    by the reviewing court and beyond the fringe of what that court deems
    minimally acceptable.” 
    Id.
     (internal quotation marks and citation omitted);
    see also Williams v. State, 
    457 Md. 551
    , 563 (2018) (“An abuse of discretion
    occurs where no reasonable person would take the view adopted by the
    circuit court.”); Jenkins v. State, 
    375 Md. 284
    , 295-96 (2003) (“Abuse occurs
    when a trial judge exercises discretion in an arbitrary or capricious manner
    or when he or she acts beyond the letter or reason of the law.”).
    Matthews, 479 Md. at 305-06; see also Abruquah v. State, 
    483 Md. 637
    , 652 n.5 (2023)
    (an abuse-of-discretion analysis requires a reviewing court to determine the “outer bounds
    of what is acceptable expert evidence”). As the Supreme Court has explained, “the law
    grants a [trial] court the same broad latitude when it decides how to determine reliability
    as it enjoys in respect to its ultimate reliability determination.” Kumho Tire Co. v.
    Carmichael, 
    526 U.S. 137
    , 142 (1999) (emphasis in original).
    23
    III
    Discussion
    A. From Frye-Reed to Daubert-Rochkind
    1. Frye-Reed
    Starting with Maryland’s 1978 adoption of the D.C. Circuit’s 1923 Frye general
    acceptance test, Maryland courts deciding the admissibility of expert testimony predicated
    on a novel scientific principle or discovery would determine whether the scientific
    principles or discovery were generally accepted in the relevant scientific community. Reed
    v. State, 
    283 Md. 374
     (1978); see Rochkind, 471 Md. at 13 (collecting cases under the Frye-
    Reed regime).
    In 1993, in Daubert v. Merrell Dow Pharmaceuticals, Inc., the Supreme Court held
    that Federal Rule of Evidence 702 superseded Frye. The Court provided a non-exclusive
    list of factors that may be pertinent when determining whether the scientific testimony at
    issue is not only relevant but reliable. 
    509 U.S. 579
    , 589, 593-94 (1993). “The Daubert
    analysis, according to the Supreme Court, was more flexible than the ‘uncompromising
    [Frye] general acceptance test’ and gave trial courts greater discretion to admit scientific
    expert testimony that is relevant and founded on sound principles, even though novel or
    controversial.” Rochkind, 471 Md. at 14 (quoting Daubert, 
    509 U.S. at 596
    ) (additional
    internal quotation omitted). Daubert allowed trial courts to admit a broader range of
    scientific testimony than would have been possible under Frye – including, for example,
    minority opinions within a field. General Electric Co. v. Joiner, 
    522 U.S. 136
    , 142 (1997).
    24
    But it did not uncritically throw the doors open to expert testimony, and instead it required
    the trial court to act as a “gatekeeper.” 
    Id.
    Under the Frye-Reed regime, the proponent of the evidence typically showed the
    general acceptance of the scientific expert’s methodology “by surveying scientific
    publications, judicial decisions, or practical applications, or by presenting testimony from
    scientists as to the attitudes of their fellow scientists.” 1 MCCORMICK ON
    EVID. § 203.1 (8th ed. 2020) (Standards for admitting scientific evidence–The general-
    acceptance requirement). “Daubert, by contrast, refocus[ed] the attention away from
    acceptance of a given methodology … and centers on the reliability of the methodology
    used to reach a particular result.” Rochkind, 471 Md. at 31.
    While Daubert originally focused solely on the expert’s methodology (as opposed
    to their conclusions), the Joiner Court recognized that conclusions and methodology “are
    not entirely distinct from one another. Trained experts commonly extrapolate from existing
    data. But nothing in either Daubert or the Federal Rules of Evidence requires a district
    court to admit opinion evidence that is connected to existing data only by the ipse dixit of
    the expert. A court may conclude that there is simply too great an analytical gap between
    the data and the opinion proffered.” Joiner, 
    522 U.S. at 146
    .
    25
    2. Maryland Rule 5-702
    This Court adopted Maryland Rule 5-702 in 1994, shortly after the Supreme Court
    issued Daubert but well before this Court fully embraced Daubert’s approach in Rochkind.6
    Under Rule 5-702,
    [e]xpert testimony may be admitted, in the form of an opinion or otherwise,
    if the court determines that the testimony will assist the trier of fact to
    understand the evidence or to determine a fact in issue. In making that
    determination, the court shall determine
    (1) whether the witness is qualified as an expert by knowledge, skill,
    experience, training, or education,
    (2) the appropriateness of the expert testimony on the particular subject,
    and
    (3) whether a sufficient factual basis exists to support the expert
    testimony.
    The third “sufficient factual basis” prong includes two sub-factors. First, the expert must
    have available an adequate supply of data. Second, the expert must use a reliable
    methodology in analyzing that data. Matthews, 479 Md. at 309; Roy v. Dackman, 
    445 Md. 23
    , 42-43 (2015). Absent either of these factors, an expert opinion is “mere speculation or
    conjecture.” Matthews, 479 Md. at 309 (quoting Rochkind, 471 Md. at 22).
    3. Rochkind
    After the consensus among the states had shifted to the Daubert regime, Maryland
    courts followed suit in 2020. Rochkind pointed to the problem with Frye that Daubert was
    6
    In the 40 years after Reed, Maryland courts experienced a “jurisprudential drift:
    the Frye-Reed standard announced in 1978 slowly morphed into a ‘Frye-Reed Plus’
    standard, implicitly and explicitly relying on and adopting several Daubert principles.”
    Rochkind, 471 Md. at 5.
    26
    meant to solve: the court sought not just pure liberalization but instead to correct errors in
    both directions. “[U]sing acceptance as the only measure of reliability presents a
    conundrum: a generally accepted methodology may produce ‘bad science’ and be admitted,
    while a methodology not yet accepted may be excluded, even if it produces ‘good
    science.’” Rochkind, 471 Md. at 30. Maryland’s adoption of Daubert’s focus on reliability
    would “streamline the evaluation of scientific expert testimony under Rule 5-702.” Id. at
    35.
    Under Rochkind, trial courts “should consider a number of factors in determining
    whether the proffered expert testimony is sufficiently reliable to be provided to the trier of
    fact.” Matthews, 479 Md. at 310. They are:
    (1) whether a theory or technique can be (and has been) tested;
    (2) whether a theory or technique has been subjected to peer review and
    publication;
    (3) whether a particular scientific technique has a known or potential
    rate of error;
    (4) the existence and maintenance of standards and controls; …
    (5) whether a theory or technique is generally accepted[;]
    […]
    (6) whether experts are proposing to testify about matters growing
    naturally and directly out of research they have conducted
    independent of the litigation, or whether they have developed their
    opinions expressly for purposes of testifying;
    (7) whether the expert has unjustifiably extrapolated from an accepted
    premise to an unfounded conclusion;
    (8) whether the expert has adequately accounted for obvious alternative
    explanations;
    27
    (9) whether the expert is being as careful as he [or she] would be in his
    [or her] regular professional work outside his [or her] paid litigation
    consulting; and
    (10) whether the field of expertise claimed by the expert is known to
    reach reliable results for the type of opinion the expert would give.
    Matthews, 479 Md. at 310-11 (summarizing Rochkind) (internal citations omitted). The
    Rochkind Court added several overarching items of guidance in adopting the Daubert
    standard:
    1. The reliability inquiry is flexible.
    2. Trial courts must focus “solely on principles and methodology, not on the
    conclusions that they generate,” although those are not entirely distinct and
    thus a trial court must consider the relationship between the two.
    3. A trial court need not “admit opinion evidence that is connected to existing
    data only by the ipse dixit of the expert”; rather, a court may conclude that
    there is simply too great an analytical gap between the data and the opinion
    proffered.
    4. All of the Daubert factors are relevant in the reliability inquiry, but none is
    dispositive, and a trial court may apply some, all, or none depending on the
    particular expert testimony at issue.
    5. Rochkind did “not upend [the] trial court’s gatekeeping function. ‘Vigorous
    cross-examination, presentation of contrary evidence, and careful instruction
    28
    on the burden of proof are the traditional and appropriate means of attacking
    shaky but admissible evidence.’”
    See Matthews, 479 Md. at 311-12 (summarizing the Rochkind Court’s observations).
    4. Matthews
    In Matthews, this Court for the first time post-Rochkind addressed under the
    Daubert standard whether a trial court erred in deciding the admissibility of expert
    testimony. Matthews, 479 Md. at 284. In attempting to solve a murder, the State enlisted
    the help of an FBI scientist who used a technique known as “reverse projection
    photogrammetry” to estimate the height of a suspect who had been caught on video but
    whose face was indiscernible. Id. at 288-89. The FBI scientist determined that the suspect
    on the video was approximately 5’8” tall, plus or minus two-thirds of an inch; defendant
    Matthews was approximately 5’9”. Id. But the FBI scientist’s expert report noted that, due
    to several variables, “the degree of uncertainty in this measurement could be significantly
    greater,” and at a pretrial hearing, the expert testified that she could not scientifically
    quantify several variables that might lead to uncertainty greater than two-thirds of an inch.
    Id. Nevertheless, the trial court admitted the expert testimony, and the jury convicted
    Matthews of the murder. Id. at 297, 304.
    The intermediate appellate court reversed Matthews’s conviction, reasoning that the
    inability to quantify the effect of the variables noted by the expert made her height
    measurement unreliable. The court perceived an “analytical gap” between the underlying
    data and the expert’s conclusion, and therefore held that the trial court abused its discretion
    in admitting the expert opinion testimony. Id. at 304-05.
    29
    This Court reinstated Matthews’s conviction, concluding that the expert’s
    methodology was reliable and finding no analytical gap in the expert’s proffered testimony.
    Matthews had argued that only one Daubert-Rochkind factor warranted exclusion of the
    testimony: the error rate, driven by the expert’s inability to provide an overall margin of
    error for her height estimate. This Court explained that “it is not sufficient to point to an
    unknown degree of uncertainty/error rate that applies to an expert opinion and claim that a
    trial court is necessarily stripped of discretion to admit that opinion.” Matthews, 479 Md.
    at 314. Instead, we distinguished between “uncertainty inherent in an expert’s
    methodology” and “uncertainty that applies to an expert’s conclusions following the
    application of a reliable methodology.” Id. at 315. Under the former circumstance of
    inherent uncertainty, a trial court more likely would exclude expert testimony due to “the
    unacceptably high risk of an inaccurate conclusion being reached in every case where the
    technique is used.” Id. at 316. However, the latter scenario, involving uncertainty in an
    expert’s conclusion following the application of a reliable methodology, “is generally less
    problematic than where an expert has applied a technique that is unreliable in every
    instance in which it is used.” Id.
    In Matthews, it was undisputed that the expert’s methodology was reliable, and that
    the uncertainty related to the expert’s conclusion following the application of that
    methodology. Thus, we recognized, the trial court was not required to exclude the
    testimony due to an inherently unreliable methodology. Id. at 317. However, we
    “emphasize[d] that just because the trial court was not required to exclude [an expert’s]
    testimony when [the expert] acknowledged an unknown degree of uncertainty, it does not
    30
    follow that the trial court was required to admit it.” Id. We further explained that, if the
    uncertainty applies to the expert’s conclusions, the “trial court should determine whether
    the uncertainty in the expert’s conclusions is the product of an analytical gap in the expert’s
    analysis and/or whether the uncertainty ultimately renders the opinion unhelpful to the trier
    of fact.” Id. at 314-15. If either of those circumstances exist, the trial court acts within its
    discretion in excluding the proffered testimony. Id.
    In Matthews’s case, we perceived no analytical gap in the expert’s testimony, where
    there was no disconnect between the results of the expert’s analysis and the expert’s
    opinion. Id. at 318. There was nothing illogical about the expert’s explanation that her
    analysis showed the subject’s height was 5’8” plus or minus two-thirds of an inch, although
    the margin of error might be greater based on other variables she could not quantify. Id. at
    318.
    We also concluded that the trial court had acted within its discretion in finding that
    the expert’s testimony would “assist the trier of fact to understand the evidence or to
    determine a fact in issue” – in other words, that it was helpful, as required by the text of
    Rule 5-702. See id. at 319-23. First, the expert had explained her analysis in detail.
    Matthews, 479 Md. at 319. Second, the expert had explained why, even with the unknown
    degree of uncertainty attributable to certain variables, she remained comfortable with her
    height estimate, including the two-thirds inch margin of error. Id. at 319-20. Finally, the
    expert herself had stood in the same spot and position as the subject in the image, and was
    able to opine that the subject of the image was just slightly shorter than the expert herself,
    who was between 5’9” and 5’10”. Id. at 320. Together, these factors allowed the trial court
    31
    to reasonably conclude that the expert’s opinion would help the jury, despite the expert’s
    acknowledged uncertainty. Id. at 320-21.
    B. Applying Daubert-Rochkind Here
    The case centers on the primacy – and boundaries – of methodological reliability in
    the Daubert-Rochkind analysis. As we explain below, the trial court acted within its
    discretion in considering most of its points of concern at the Daubert-Rochkind hearing.
    However, the court made a significant error when it relied on Ms. Cardell’s June 2021
    normalizing adjustments regarding trauma and on-call pass-through payments as a basis
    for excluding her testimony. We shall order a limited remand under Maryland Rule
    8-604(d)(1) for the trial court to revisit its ruling without consideration of the June 2021
    normalizing adjustments as reflecting on the reliability of Ms. Cardell’s methodology.
    1. The Relationship Between Data and Methodology
    An expert witness generally arrives at an opinion by choosing a methodology,
    selecting data to which to apply the chosen methodology, and drawing conclusions based
    on the results of the application of the methodology. The reliability of the expert’s
    methodology is a core focus of a trial court under Rule 5-702. Without reliable methods
    (in addition to an adequate supply of data), an expert’s opinion lacks sufficient factual basis
    to support it and is instead “mere speculation or conjecture.” Matthews, 479 Md. at 309.
    The disagreement here largely is over the precise boundary between data and methodology
    when assessing the factual basis of an expert’s testimony. There are two competing visions.
    KatzAbosch and the trial court offer one vision, in which input-related choices an
    expert makes can be so central to a straightforward methodology that those choices
    32
    implicate the reliability of the methodology itself. As the trial court stated in its
    supplemental written memorandum, “[a]n expert’s choice in data sampling is at the heart
    of his methodology.” (Quoting CDW, 
    906 F. Supp. 2d at 824
    .) The before-and-after method
    for measuring a company’s lost profits is as widely accepted as it is simple: identify a
    benchmark period before an alleged harm event and compare the company’s profits during
    that benchmark period to the company’s profits in subsequent periods. The play in the
    joints arises out of the choices the expert makes in identifying the benchmark period and
    thereby deciding what will be the “before” period and what will be the “after” period, and
    in the choices the expert makes about how she will measure “profits.” Should the expert
    use pre- or post-“guaranteed payment” figures? Should she apply the before-and-after
    method to a medical practice’s financial performance without regard to changes in
    reimbursement rates between years? In KatzAbosch’s and the trial court’s view, these
    considerations concerning data differ from consideration of the soundness of the data that
    an expert uses in applying their methodology, or the soundness of the data from which an
    expert extrapolates after completing the application of the methodology.7
    7
    Consider the use of inputs where their selection does not depend on judgment calls
    in the way that the trial court viewed the selection of the benchmark year in this case. For
    example, imagine hypothetically that KatzAbosch did not dispute that 2015 was the proper
    benchmark year and agreed that Ms. Cardell used all the correct sources of data in her
    calculation of lost profits, but KatzAbosch called the veracity of some of the data into
    question (perhaps suggesting that bookkeepers at PNSI had not accurately classified
    various expenses in the company’s books and records). In that instance, KatzAbosch would
    have been complaining about the soundness of the data to which Ms. Cardell applied a
    concededly reliable methodology. The veracity of the data in this hypothetical is a factual
    question that the expert has no real impact on; instead, it is up to the jury to decide whether
    the expert’s reliable methodology has been applied to accurate data or whether the
    33
    PNSI and the Appellate Court offer a more rigid vision, arguing that anything
    dealing with the inputs is a “data” question to be siloed off from the methodological
    reliability analysis and instead left to the jury.8 They acknowledge that the line between
    data and methodology (and, for that matter, conclusions) is blurry, but they draw the line
    sharply in this case, arguing that only the core before-and-after subtraction counts as
    methodology, and the rest is data (e.g., why 2015?) or conclusion. In this regard, PNSI
    relies primarily on the Seventh Circuit’s decision in Manpower, Inc. v. Ins. Co. of Pa., 
    732 F.3d 796
     (7th Cir. 2013), noting that we cited Manpower in Matthews.9
    In Manpower, the federal district court excluded insured-plaintiff Manpower’s
    accounting expert, whom Manpower needed to establish business interruption damages in
    its suit against the insurer-defendant. The insurer-defendant challenged the reliability of
    the expert’s methodology, and the district court found that the expert had followed the
    master insurance policy’s “straightforward” methodology for his calculations. Manpower,
    
    732 F.3d at 801-02
    . But the calculations’ reliability nevertheless “turn[ed] on whether [the
    expert] used reliable methods when selecting the numbers used in his calculations.” 
    Id.
     at
    proponent has offered “garbage in, garbage out” expert testimony to be discredited even
    after being admitted.
    8
    For example: “Whether Ms. Cardell failed to consider reimbursement rates is not
    an issue with the methodology – the before-and-after method. Rather, it is an issue with
    the soundness of the data she used to reach her conclusion.” Parkway Neuroscience, 255
    Md. App. at 627.
    9
    See Matthews, 479 Md. at 316 (quoting Manpower, 
    732 F.3d at
    806: “The district
    court usurps the role of the jury, and therefore abuses its discretion, if it unduly scrutinizes
    the quality of the expert’s data and conclusions rather than the reliability of the
    methodology the expert employed.”).
    34
    801. The district court found that the answer to that question was no. The expert used a
    shorter base period (from which he extrapolated a relatively high growth rate) because of
    recent corporate acquisitions, new policies, and new managers at Manpower, and the expert
    did not consider other indicators that may have affected the growth rate. 
    Id.
     The district
    court also criticized the expert’s reliance on conversations with the company managers. 
    Id.
    But the district court primarily found that the expert’s “‘analysis [broke] down’ at his
    choice of growth rate.” 
    Id.
     In the district court’s view, “[o]nly a more thorough analysis of
    the reasons for the growth would have supported [the expert’s] choice of a projected growth
    rate.” 
    Id.
    The Seventh Circuit reversed the district court’s exclusion of the expert, agreeing
    with Manpower that “the district court exercised its gatekeeping role under Daubert with
    too much vigor.” 
    Id. at 805
    . The federal appellate court opined that reliability is primarily
    a question of the expert’s methodology, not the data that is the input or the conclusions that
    are the output. 
    Id. at 806
    . The Manpower Court acknowledged the blurriness of the line
    between undue scrutiny of an expert’s data and proper scrutiny of an expert’s methodology,
    writing that “[t]he critical inquiry is whether there is a connection between the data
    employed and the opinion offered.” 
    Id.
     Opinions are properly excluded when they are
    connected to existing data “only by the ipse dixit of the expert.” 
    Id.
    More specifically, the Manpower Court said that the district court should have
    stopped its inquiry when it found the methodology reliable, but instead it “drilled down to
    a third level in order to assess the quality of the data inputs [the expert] selected” and then
    took issue with that data selection. 
    Id. at 807
    . The expert’s data selection was “substantially
    35
    more nuanced and principled than the district court’s characterization reflects,” and so the
    ultimate opinion was not ipse dixit but rather “reasoned and founded on data.” 
    Id. at 809
    .
    Each of the district court’s criticisms of the expert’s reliability “was a comment on the
    soundness of the factual underpinnings of his calculation” and the district court’s rulings
    could even have been something of a “roadmap for [the defendant-insurer’s] cross-
    examination of [the expert].” 
    Id.
     “But the district court supplanted that adversarial process
    with its admissibility determination” and “set the bar too high and therefore abused its
    discretion.” 
    Id.
    The Manpower Court’s understanding of the line between data and methodology,
    of course, does not bind this Court, and its rigidity is not in keeping with the approach
    taken by other federal appellate courts. Instead, many federal courts have explained the
    Daubert standard in ways that reject this sharp line and acknowledge that problems with
    data and data selection (which itself can involve its own methodology) can bear on
    admissibility before the judge and not just weight before the jury. See, e.g., EEOC v.
    Freeman, 
    778 F.3d 463
    , 467, 472 (4th Cir. 2015);10 Elcock v. Kmart Corp., 
    233 F.3d 734
    ,
    10
    In Freeman, plaintiff EEOC’s industrial psychologist expert witness delivered a
    report based on a database riddled with errors, introducing even more errors in an updated
    analysis; the “sheer number of mistakes and omissions” rendered the expert’s analysis
    “outside the range where experts might reasonably differ” under Kumho Tire, and so the
    district court did not abuse its discretion in excluding the proffered expert’s testimony as
    unreliable. The EEOC unsuccessfully argued that “the issue of the reliability of an expert’s
    data is always a question of fact for the jury, except perhaps in some theoretical, rare case.”
    Freeman, 
    778 F.3d at 472
     (Agee, J., concurring).
    36
    755-56 (3d Cir. 2000);11 In re Mirena IUS Levonorgestrel-Related Products Liab. Litig.
    (No. II), 
    982 F.3d 113
    , 123 (2d Cir. 2020).12
    Consider as one example Rink v. Cheminova, an Eleventh Circuit case in which the
    appellate court found no abuse of discretion in the district court’s exclusion of a putative
    class’s expert. 
    400 F.3d 1286
    , 1293-94 (11th Cir. 2005). The plaintiffs maintained that
    Cheminova’s pesticide malathion defectively contained elevated levels of isomalathion,
    which makes malathion particularly toxic to humans and which, they claimed, was created
    by exposure of the malathion supply to temperatures above 77 degrees Fahrenheit during
    storage at sites in Texas, Georgia, and Florida. 
    Id. at 1289
    . The plaintiffs’ expert estimated
    the isomalathion content of pesticides used in Tampa; he began with National Weather
    Service temperature readings near the malathion storage sites, using the recorded high and
    low temperatures as upper and lower limits of the probable exposure temperatures. 
    Id.
    Because there was evidence that the inside of the Texas storage facility was actually
    18 degrees warmer than the outside ambient air temperature, he added 18 degrees to the
    11
    In Elcock, the Third Circuit held that the district court abused its discretion in
    admitting the expert’s economic damages model, which relied on empirical assumptions
    not supported by the record, fearing that a jury would be “likely to adopt the gross figure
    advanced by a witness who has been presented as an expert.” 
    233 F.3d at 755-56
    .
    12
    In Mirena, the Second Circuit rejected the plaintiffs’ contention that a trial court
    erred by taking a “hard look” at their expert’s methodology: “[A]n expert’s methodology
    must be reliable at every step of the way, and in deciding whether a step in an expert’s
    analysis is unreliable, the district court should undertake a rigorous examination of the
    facts on which the expert relies, the method by which the expert draws an opinion from
    those facts, and how the expert applies the facts and methods to the case at hand.” 982 F.3d
    at 123 (internal quotation marks and citation omitted) (emphasis added by the Mirena
    Court).
    37
    upper plausibility limits. Id. at 1290. He then averaged the upper and lower limits to find
    the most probable exposure temperature. Id. He took that temperature and plugged it into
    an equation to calculate the level of toxic isomalathion as a function of time and
    temperature. Id. Importantly, the plaintiffs’ theory hinged on the isomalathion content of
    the pesticides used in Tampa, Florida, which included not just the malathion that had
    initially been stored in Texas, but also malathion that had been initially stored at sites in
    Florida and Georgia. See id. at 1289-90.
    The district court excluded the expert’s testimony for several reasons, including the
    expert’s “method of extrapolating data from one site [Texas] to another [Florida and
    Georgia] without making particularized findings which accounted for the differences in
    conditions and length of storage at each site.” Id. at 1290. On appeal, the plaintiffs argued
    that the district court had improperly taken issue with the expert’s temperature data and not
    his methodology. The Eleventh Circuit disagreed:
    While [the plaintiffs] suggest that the only methodology at issue was [the
    expert’s] use of an equation to determine isomalathion levels, this argument
    belies the fact that [the expert] employed two methodologies: first, he
    employed certain methods of extrapolation and transposition to arrive at
    temperature data; and second, he inserted the temperature data into an
    equation to arrive at the level of isomalathion in the [pesticide]. As we have
    explained, the district court’s exclusion of [the expert] was based on its
    rejection of his methodology to derive temperature data, not the data itself.
    Thus, the district court’s exclusion of [the expert] can be distinguished from
    a situation in which an exclusion is based on a district court’s refusal to credit
    hard data arrived at by unassailable methods.… Here, the data [the expert]
    produced was driven by the methodology he used, and thus the district
    court’s inquiry into how he arrived at the data is not inappropriate
    38
    considering that the district court is charged with evaluating an expert’s
    methodology.
    Id. at 1293.13 The Rink Court took an appropriately holistic view of the methodology being
    applied, examining the production of the equation’s inputs as part of the process itself. A
    narrower understanding of the methodology – something along the lines of “use this
    formula to calculate the isomalathion content of this pesticide” – might have called for
    allowing the expert to testify at trial, where opposing counsel could have impeached him
    by showing the weaknesses in his Tampa storage temperature input selection. We agree
    with the Eleventh Circuit that, under a proper application of Daubert, the trial court in Rink
    acted as an appropriate gatekeeper.
    Another useful case to consider is In re Wholesale Grocery Products Antitrust
    Litigation, an Eighth Circuit antitrust case in which the district court had excluded an expert
    witness who sought to establish the plaintiff’s injury by selecting a competitive benchmark.
    
    946 F.3d 995
     (8th Cir. 2019). The district court excluded as unreliable the benchmark the
    expert selected, because the expert’s choice of a non-independent chain grocery store (Stop
    & Shop) was premised on an unfounded assumption that independent retailers’ charges
    13
    The Rink Court contrasted the case before it with Quiet Tech. DC-8, Inc. v. Hurel-
    Dubois UK Ltd., 
    326 F.3d 1333
     (11th Cir. 2003), where the district court had admitted
    testimony of Hurel-Dubois’s expert on computational fluid dynamics. Rink, 
    400 F.3d at 1293
    . After losing at trial, Quiet Tech argued that the district court abused its discretion by
    admitting the other side’s expert testimony. The Eleventh Circuit disagreed, observing that
    Quiet Tech “does not argue that it is improper to conduct a [computational fluid dynamics]
    study using the sorts of aerodynamic data that [the expert] employed, but rather that the
    specific numbers that [the expert] used were wrong. Thus, the alleged flaws in [the
    expert’s] analysis are of a character that impugn the accuracy of his results, not the general
    scientific validity of his methods.” Quiet Tech, 
    326 F.3d at 1343-44
    . This case resembles
    Rink, whereas the hypothetical we pose above in footnote 7 resembles Quiet Tech.
    39
    followed the same pattern as Stop & Shop’s. Id. at 999. The plaintiff argued on appeal that
    benchmarking is a recognized tool for establishing antitrust injury, and once the expert
    chose a benchmark, there was nothing left for the district court to analyze under Daubert;
    what remained were questions of fact for the jury. Id. at 1001. The Eighth Circuit rejected
    this argument “because it is the foundation of the assumption underlying the application of
    the method employed by [the expert] on these facts … that led the district court to its
    conclusion that [the expert’s] testimony should be excluded under Rule 702.” Id. “The
    district court did not go ‘miles beyond’ its appropriate role in this case, as [the plaintiff]
    argues, but rather held that the reasoning underlying [the expert’s] testimony was not on
    solid footing because the assumption upon which the report relied was insufficient to
    validate his opinion.” Id. at 1002. “[T]he district court held that ultimately these analyses
    were a house of cards of sorts and our own analysis reveals no abuse of discretion in this
    conclusion. At its base level, the core assumption of the analysis was by the ipse dixit of
    [the expert].” Id. Here too, we see the same arguments as in Manpower and in the case
    before us: undue scrutiny and a tunnel vision focus on methodology. But the Eighth Circuit
    declined to accept the benchmark selection without question and instead approved of the
    district court’s inquiry into the assumptions underlying the expert’s choice of benchmark.
    This is a long way from Manpower’s approach.
    In short, whether an expert’s methodology is sufficiently reliable to admit the
    expert’s testimony at trial will sometimes require a trial court to consider data and
    assumptions that the expert has employed in deciding threshold points relating to the
    methodology. Manpower’s rigid separation of “data” and “methodology” misses this grey
    40
    area and creates a categorical rule when the Daubert-Rochkind regime calls for flexibility
    and deference.
    When this Court cited Manpower in Matthews, it did so to demonstrate the
    proposition that where a sufficient factual basis exists under Rule 5-702(3) and Daubert-
    Rochkind – that is, where an expert has applied a reliable methodology to an adequate
    supply of data – courts should not exclude an expert merely because the expert’s particular
    conclusions may be inaccurate, but rather should only exclude expert testimony that is
    “mere speculation or conjecture.” Matthews, 479 Md. at 316.
    The Appellate Court in this case relied on the part of Manpower to which we do not
    subscribe, categorizing the 2015 base year choice, the failure to consider reimbursement
    rates, etc., as arguable defects in the soundness of Ms. Cardell’s data, rather than defects
    of her methodology. See Parkway Neuroscience, 255 Md. App. at 627-28 & n.9. The
    Appellate Court was right to center methodology in its analysis of the expert opinion’s
    reliability, a critical aspect (along with the data’s adequacy) of the opinion’s factual basis
    without which the opinion would be “mere speculation or conjecture.” Matthews, 479 Md.
    at 309 (citing Rochkind, 471 Md. at 22); see Parkway Neuroscience, 255 Md. App. at 629
    & n.11. But the intermediate appellate court, relying on Manpower, took an overly rigid
    approach in analyzing the relationship between data and methodology, holding that the trial
    court’s analysis of Ms. Cardell’s opinion went to her data rather than to her methodology.
    Just as the U.S. Supreme Court has noted the blurred line between methodology and
    conclusions, Joiner, 
    522 U.S. at 146
    , we note the sometimes blurred line between data and
    methodology. Trial courts must not transmute all questions of data’s provenance or veracity
    41
    into questions of methodology, just as they must not – under Joiner – transmute all
    disagreements with conclusions into disagreements with methodology. But by the same
    token, trial courts should not wear “methodology blinders” and deny the existence of some
    limited overlap between data and methodology (and between methodology and
    conclusions). Determining whether a dispute concerning expert testimony implicates the
    soundness of data or soundness of methodology is precisely the type of matter that calls
    for the exercise of a trial court’s discretion.
    Here, the trial court explicitly noted speculative, insufficiently substantiated
    judgment calls that were central to Ms. Cardell’s application of the before-and-after
    method. In the trial court’s estimation, Ms. Cardell exercised subjective judgment in
    settling on the 2015 base year – that is, it was impossible to test the validity of that decision.
    Relatedly, the trial court was troubled by the effect on the base year determination of Ms.
    Cardell’s treatment of member draws as expenses without understanding whether that
    decision was consistent with any industry standard. In addition, the trial court faulted Ms.
    Cardell for failing to appropriately factor in confounding variables (in particular, declining
    insurance reimbursements) into her methodology. The court determined these judgment
    calls veered toward speculation and conjecture and ate away at the factual basis for Ms.
    Cardell’s opinions, because even though many of the problems had to do with the “inputs”
    to the before-and-after method, the selection of those inputs is central to the reliability of
    the method itself.
    We rely on trial courts that conduct Daubert-Rochkind hearings to determine where
    the line between data and methodology is in the specific cases before them, and whether
    42
    the proffered expert’s choices relating to data, assumptions, and other inputs implicate the
    reliability of the expert’s methodology. In this part of the trial court’s analysis, the court
    did just that. That is, we discern no error in the trial court’s analysis of Daubert-Rochkind
    factors one and two and four through nine.14,15
    14
    The trial court discussed Ms. Cardell’s decision not to account for the doctors’
    individual revenue generation capabilities as part of the court’s discussion of Daubert
    factor eight (accounting for obvious alternative explanations). As discussed below, we
    think the court’s concerns on that front go more to the reliability of the before-and-after-
    method when used to calculate lost profits of a limited liability company with a small
    number of revenue generators, rather than a failure to account for obvious alternative
    explanations.
    15
    The trial court said that, at the Daubert hearing stage, it was not the court’s
    responsibility to determine whether Ms. Cardell qualified as an expert or not. And the court
    explicitly stated that it was not deciding one way or the other whether Ms. Cardell was, in
    fact, qualified to render an expert opinion at trial. However, the court did note Ms. Cardell’s
    lack of experience with specialty medical practices and her reliance on oral
    communications with PNSI staff (rather than universally available, concrete sources of
    information), and her reliance on her own experiences and judgment rather than on industry
    standards. In the court’s mind, these points undermined Ms. Cardell’s overall reliability.
    We do not discern any abuse of discretion in the court’s reliance on these points.
    However, it is important to acknowledge that there is nothing inherently problematic
    with an expert’s use of information provided orally or in writing by the proponent of her
    testimony. Experts routinely rely on such information in arriving at their opinions. Nor is
    it uncommon for experts to apply their subjective judgment, based on their training and
    experience, in formulating their opinions. The trial court did not indicate that it believed it
    was per se improper for an expert to rely on oral communications with the proponent of
    her testimony or to make subjective judgment calls. Rather, the trial court’s point was that,
    in the absence of other sources of information that tended to support the reliability of Ms.
    Cardell’s methodology – such as industry standards – and given her inability to adequately
    explain why she made the judgment calls she did, Ms. Cardell’s reliance on oral
    communications with PNSI and on her subjective judgment detracted from the reliability
    of her methodology.
    43
    2. The Federal Analog to Maryland Rule 5-702
    The direction of analogous Federal Rule 702 confirms our understanding of
    meaningful gatekeeping as to an expert opinion’s factual basis. In a May 2022 report, the
    Advisory Committee on Evidence Rules wrote that
    the Committee resolved to respond to the fact that many courts have declared
    that the reliability requirements set forth in Rule 702(b) and (d) – that the
    expert has relied on sufficient facts or data and has reliably applied a reliable
    methodology – are questions of weight and not admissibility, and more
    broadly that the expert testimony is presumed to be admissible. These
    statements misstate Rule 702, because its admissibility requirements must be
    established to a court by a preponderance of the evidence. The Committee
    concluded that in a fair number of cases, the courts have found expert
    testimony admissible even though the proponent has not satisfied the Rule
    702(b) and (d) requirements by a preponderance of the evidence – essentially
    treating these questions as ones of weight rather than admissibility….
    COMM. ON RULES OF PRAC. AND PROC. OF THE JUD. CONF. OF THE U.S., REP. OF THE
    ADVISORY COMM. ON EVIDENCE RULES                     6   (May    15,   2022),    available   at
    https://perma.cc/PK3B-Q8G5.
    Absent any contrary Congressional action, Federal Rule 702 will officially reflect
    this reality come December 1, 2023, when a set of amendments will take effect to provide
    that the proponent of expert testimony must meet Rule 702’s standards – including related
    to the testimony’s factual basis – by a preponderance of evidence, or else the testimony is
    inadmissible and may not go to the jury:
    A witness who is qualified as an expert by knowledge, skill, experience,
    training, or education may testify in the form of an opinion or otherwise if
    the proponent demonstrates to the court that it is more likely than not that:
    (a) the expert’s scientific, technical, or other specialized knowledge will
    help the trier of fact to understand the evidence or to determine a fact
    in issue;
    44
    (b) the testimony is based on sufficient facts or data;
    (c) the testimony is the product of reliable principles and methods; and
    (d) the expert’s opinion reflects a reliable application of the principles
    and methods to the facts of the case.
    U.S. SUPREME COURT, ORDER 4 (April 24, 2023), available at https://perma.cc/RU2S-
    KEYM (most relevant new language emphasized). The change emphasizing the
    preponderance standard “specifically was made necessary by the courts that have failed to
    apply correctly the reliability requirements of [Federal Rule 702].” FED. R. EVID. 702
    advisory committee’s note to 2023 amendment. “[M]any courts have held that the critical
    questions of the sufficiency of an expert’s basis, and the application of the expert’s
    methodology, are questions of weight and not admissibility. These rulings are an incorrect
    application of Rules 702 and 104(a).” Id.; see Sardis v. Overhead Door Corp., 
    10 F.4th 268
    , 283-84 (4th Cir. 2021) (observing that the then-proposed amendments to Federal Rule
    702 would make explicit the preponderance of evidence standard of admissibility to the
    rule’s sufficiency of basis and reliability analyses; confirming that these rule revisions and
    clarifications “clearly echo[] the existing law on the issue” from Daubert, Kumho Tire, and
    Rule 702 itself).
    The new amendments comprehend that some challenges to expert testimony will,
    in fact, go to weight rather than admissibility.
    For example, if the court finds it more likely than not that an expert has a
    sufficient basis to support an opinion, the fact that the expert has not read
    every single study that exists will raise a question of weight and not
    admissibility. But this does not mean, as certain courts have held, that
    arguments about the sufficiency of an expert’s basis always go to weight and
    not admissibility. Rather it means that once the court has found it more likely
    45
    than not that the admissibility requirement has been met, any attack by the
    opponent will go only to the weight of the evidence.
    FED. R. EVID. 702 advisory comm. note to 2023 amendment (emphasis added). Indeed,
    they do not require federal courts to “nitpick an expert’s opinion in order to reach a perfect
    expression of what the basis and methodology can support,” but instead seek to block
    “claims that are unsupported by the expert’s basis and methodology.” 
    Id. 3
    . The Trial Court’s Error
    Although we discern no error in most of the trial court’s application of the Daubert-
    Rochkind factors, we are constrained to conclude that the court erred in one respect. The
    trial court viewed Ms. Cardell’s June 2021 updates, which the court discussed in the
    context of Daubert-Rochkind factors three (known or potential rate of error) and 10
    (whether the field of expertise is known to reach reliable results for the projected type of
    expert opinion), as implicating the reliability of her methodology. To be sure, the court
    observed that the error rate factor did not apply “as it was ordinarily considered,” e.g.,
    concrete and numerical probabilities of correctness in DNA testing. Still, the court was
    troubled by the timing of Ms. Cardell’s updates – without new information and “for fully
    subjective reasons” – which, the court believed, reflected negatively on her methodology.
    This analysis missed the mark. In fact, there was new information of a sort: Ms.
    Cardell noticed something she had not noticed before on first examination. She then sought
    clarification and revised her opinion, just as a doctor might order a biopsy and diagnose a
    patient with skin cancer if the doctor had missed a mole upon first examination of the
    patient. Catching something peculiar the second time around neither undermines the
    46
    adequacy of the data (the patient’s skin) nor the court’s understanding of the expert’s
    methodology (examining the patient’s skin for disease indicators).
    Notably, when discussing the tenth Daubert factor – whether the field of expertise
    is known to reach reliable results for the type of opinion the expert would give – the trial
    court specifically referenced the June 2021 adjustments, stating that “the mere fact the
    findings changed in June of this year for fully subjective reasons, it had nothing to do with
    any new information. It kind of makes the whole reliability even that much more suspect.”
    This comment, along with the court’s other references to the June 2021 adjustments, leaves
    us with the abiding concern that, in this instance, the trial court strayed from its gatekeeping
    role and that this error was significant in the court’s overall analysis.16 Ms. Cardell’s
    decision to make the adjustments relating to trauma/on-call payments in 2016 did not
    implicate the reliability of her methodology. At most, it went to the care with which she
    16
    The trial court also erred when it stated that Ms. Cardell’s lost-profit calculations
    would not be helpful to the jury. The court reached this conclusion based on Ms. Cardell’s
    failure to consider each departing physician’s projected revenue generation in her
    calculation of lost profits. The court’s point was that, if the jury concluded that one (or
    more) of the departing physicians left for reasons other than KatzAbosch’s negligence, a
    lost-profits analysis that included that physician’s projected revenue would overstate the
    practice’s lost profits and therefore not help the jury to accurately calculate damages. The
    court stated, however, that its finding with respect to helpfulness was a “very, very slight
    factor” in its decision to exclude Ms. Cardell’s testimony.
    We confirmed in Matthews that a trial court has discretion under Maryland Rule
    5-702 to exclude an expert’s opinion based on a reliable methodology if the court
    nevertheless concludes that the expert’s testimony would not be helpful to the jury. See
    Matthews, 479 Md. at 320-21. Here, however, there was no way to know before the trial
    whether the jury would conclude that any of the physicians left for reasons other than
    KatzAbosch’s negligence. The trial court’s conclusion that Ms. Cardell’s testimony would
    not be helpful, therefore, was speculative and not a proper ground for exclusion.
    47
    applied her methodology, which is a matter to be explored on cross-examination before the
    jury (if Ms. Cardell’s testimony is otherwise found to be sufficiently reliable).17
    In Matthews, we “emphasize[d] that just because the trial court was not required to
    exclude [the expert’s] testimony when [the expert] acknowledged an unknown degree of
    uncertainty, it does not follow that the trial court was required to admit it.” 479 Md. at 317.
    Similarly, we do not believe that, based on the record before the trial court in this case, the
    court was required to exclude Ms. Cardell’s testimony. Rather, based on the factors that
    the trial court appropriately considered in this case, it was within the trial court’s discretion
    to admit or exclude Ms. Cardell’s testimony. Having carefully reviewed the record, we
    conclude that the fair and prudent course of action at this point is to order a limited remand
    to the circuit court under Maryland Rule 8-604(d)(1) so that the trial court may decide to
    admit or exclude Ms. Cardell’s testimony without consideration of her June 2021
    normalizing adjustments as reflecting on the reliability of Ms. Cardell’s methodology.18
    The trial court may make that decision based on the existing record or, in its discretion,
    may allow further examination of Ms. Cardell and/or other witnesses before issuing a new
    17
    The trial court elsewhere in its analysis of the Daubert factors said that it had “no
    reason to think that [Ms. Cardell] blew this off as an inconsequential project, I mean she
    took this very seriously.”
    18
    We do not mean to suggest that, in every case where an appellate court concludes
    that part of a trial court’s Daubert ruling was based on proper factors and another part was
    not, a limited remand to the trial court is necessary. We expect that, in many cases, it will
    be clear from the record whether the trial court would have admitted or excluded the expert
    testimony without consideration of a factor that is later determined to have been improper
    on appeal. We encourage trial courts to make such matters explicit on the record when
    possible.
    48
    ruling.19 In the trial court’s discretion, it also may permit the parties to submit additional
    written and oral arguments prior to issuing its ruling. The trial court should provide a
    written explanation of its decision.
    We shall retain jurisdiction over this case. After the trial court issues its decision on
    remand, we shall issue an appropriate Order.
    19
    The trial court’s criticism of Ms. Cardell’s failure to consider the remaining and
    departing doctors’ inherent ability to generate revenue when discussing Daubert factor
    eight (accounting for obvious alternative explanations) is more a criticism of the reliability
    of the before-and-after methodology itself as applied to a limited liability company with a
    small number of revenue-generating members, than it is a criticism of Ms. Cardell’s failure
    to account for obvious alternative explanations in applying the before-and-after
    methodology. In any business with a small number of revenue-generating members, the
    firm’s overall performance in a particular year may well be the result of some members
    having unusually strong or unusually weak years, compared to their individual
    performances over time. As Ms. Cardell described the before-and-after method of
    calculating lost profits, there is no consideration of how individual revenue generators
    perform over time. Rather, Ms. Cardell testified, “under the before and after methodology
    what happens is that the damages expert looks at the [company’s] performance in two
    different periods. The first being what we called the benchmark period or the before period.
    And that’s the period that is unaffected by whatever the alleged harm event, breach is. And
    then that is [compared] to the after period or the loss period, which is the period that is
    affected by whatever the alleged harm, breach, or event is. This methodology essentially
    calculates what the [company’s] profits would have been but for again that alleged breach,
    harm event.”
    The trial court did not question the general acceptance of the before-and-after
    analysis as a reliable methodology for calculating lost profits. Nor did the trial court
    explicitly state that the before-and-after methodology, as described by Ms. Cardell, cannot
    reliably be applied to a limited liability company with a small number of revenue-
    generating members. However, that seems to be the import of the concerns the trial court
    articulated when discussing Ms. Cardell’s failure to consider the individual revenue-
    generating abilities of the physicians who departed, as well as the two physicians who
    remained at PNSI. On remand, in its discretion, the trial court may give additional
    consideration to whether the before-and-after methodology, as a general matter, is a
    reliable methodology for calculating lost profits of an entity such as PNSI. We express no
    view concerning the answer to that question.
    49
    IV
    Conclusion
    The trial court, in its gatekeeping role under Daubert-Rochkind, acted within its
    discretion in analyzing the data and other inputs and assumptions that implicated the
    reliability of Ms. Cardell’s methodology. However, the court improperly considered Ms.
    Cardell’s June 2021 normalizing adjustments relating to trauma/on-call payments as
    reflecting on the reliability of Ms. Cardell’s methodology. We order a limited remand to
    the circuit court to allow that court to decide whether to admit or exclude Ms. Cardell’s
    expert testimony without such consideration of the June 2021 normalizing adjustments.
    JUDGMENT OF THE APPELLATE COURT OF
    MARYLAND VACATED; CASE REMANDED TO
    THAT COURT WITH THE DIRECTION TO
    REMAND THE CASE TO THE CIRCUIT COURT
    FOR   HOWARD       COUNTY,  WITHOUT
    AFFIRMING OR REVERSING THE JUDGMENT
    OF THE CIRCUIT COURT, FOR FURTHER
    PROCEEDINGS CONSISTENT WITH THIS
    OPINION. COSTS TO ABIDE.
    50
    Circuit Court for Howard County
    Case No. C-13-CV-18-000181
    Argued: May 4, 2023
    IN THE SUPREME COURT
    OF MARYLAND*
    No. 30
    September Term, 2022
    KATZ, ABOSCH, WINDESHEIM,
    GERSHMAN & FREEDMAN, P.A., ET AL.
    v.
    PARKWAY NEUROSCIENCE
    AND SPINE INSTITUTE, LLC
    Fader, C.J.,
    Watts,
    Hotten,
    Booth,
    Biran,
    Gould,
    Eaves,
    JJ.
    Concurring Opinion by Booth, J.
    Filed: August 30, 2023
    * During the November 8, 2022 general election,
    the voters of Maryland ratified a constitutional
    amendment changing the name of the Court of
    Appeals of Maryland to the Supreme Court of
    Maryland. The name change took effect on
    December 14, 2022.
    I agree with and join the Majority’s well-written opinion in this case. I write
    separately to respond to the Majority’s invitation to “reflect on [the] flexibility and
    deference” due to courts analyzing the admissibility of expert testimony. Maj. Slip Op.
    at 1. I have observed that our traditional formulation of the abuse of discretion standard is
    not the best or most accurate way of describing our abuse of discretion review in the context
    of reviewing expert testimony admissibility determinations, and suggest that this Court re-
    formulate the definition of our abuse of discretion standard in the context of appellate
    review of expert witness testimony admissibility determinations.
    Since this Court’s adoption of Daubert in 2020, we have reviewed for abuse of
    discretion four cases involving trial courts’ decisions to admit or preclude expert testimony:
    State v. Matthews, 
    479 Md. 278
     (2022); Abruquah v. State, 
    483 Md. 637
     (2023); Oglesby
    v. Baltimore School Associates, No. 26 Sept. Term, 2022, 
    2023 WL 4755689
     (July 26,
    2023); and this case.     The Court’s decisions concerning the admissibility of expert
    testimony in Matthews and Abruquah were not unanimous ones.1 Not only have there been
    1
    In State v. Matthews, 
    479 Md. 278
    , 325–40 (2022) (Watts, J., dissenting), Justice
    Watts filed a dissenting opinion stating that she would have upheld the Appellate Court’s
    conclusion that the trial court abused its discretion in admitting the State’s expert witness
    concerning her opinions involving reverse projection photogrammetry.
    In Abruquah v. State, 
    483 Md. 637
    , 699–711 (2023), Justice Hotten filed a
    dissenting opinion, joined by Justice Gould and Justice Eaves, in which those justices
    would have determined that the trial court did not abuse its discretion in permitting the
    unqualified opinions expressed at trial by the firearms expert. In a separate dissenting
    opinion, Justice Gould criticized the Majority’s application of the abuse of discretion
    standard of review, asserting that the Majority “sidestep[ped] the deferential standard of
    review by recasting its decision as establishing the outer bounds of what is acceptable
    expert evidence in [the] area [of firearms identification].” Abruquah, 483 Md. at 712
    (quotations omitted) (Gould, J. dissenting).
    different views concerning the circuit courts’ discretion in making expert witness
    admissibility determinations among jurists on this Court, it is also notable that, in three of
    the four above-referenced cases, this Court reached opposite conclusions from the
    Appellate Court of Maryland undertaking the same review.2 In Matthews—our maiden
    voyage in appellate review of a trial court’s decision to admit expert testimony under the
    application of the Daubert-Rochkind factors—we “reaffirmed” the sentiments expressed
    in Rochkind, that “it is still the rare case in which a Maryland trial court’s exercise of
    discretion to admit or deny expert testimony will be overturned.” 479 Md. at 306. Given
    our post-Rochkind batting average, I am not sure that sentiment holds true.
    I was one of the members of this Court who voted in Rochkind to adopt the Daubert
    standard. I joined the majority opinions written by my colleagues in Matthews, Abruquah,
    Oglesby, and the instant case, and I agree with our analysis in each of them. That said,
    with some time to reflect on the Court’s application of the abuse of discretion standard in
    the context of appellate review of a trial court’s decision to admit or deny expert testimony
    2
    See Matthews v. State, 
    249 Md. App. 509
     (2021) (reversing a criminal defendant’s
    second-degree murder conviction after determining that the circuit court’s decision to
    admit the photogrammetry expert’s opinion about the height of a suspect captured in a
    surveillance video was inadmissible, and error was not harmless); Oglesby v. Baltimore
    Sch. Assocs., No. 130 Sept. Term, 2021, 
    2022 WL 3211044
     (App. Ct. Md. Aug. 9, 2022)
    (holding that the circuit court did not abuse its discretion in excluding a plaintiff’s causation
    expert in a lead paint case, or err in granting summary judgment in light of the exclusion
    of the causation testimony); Parkway Neuroscience and Spine Inst., LLC v. Katz Abosch,
    Windesheim, Gershman & Freeman, P.A., et. al., 
    255 Md. App. 596
     (2022) (holding that
    the trial court abused its discretion in excluding Ms. Cardell’s opinions concerning lost
    profits, and therefore, erred in granting summary judgment in favor of the defendants). We
    do not know whether the Appellate Court would have considered the application of the
    Daubert-Rochkind factors in the same manner as this Court in Abruquah because this Court
    granted certiorari while the case was pending in that court. See Abruquah, 483 Md. at 652.
    2
    in its consideration and application of the Daubert-Rochkind factors, I have some unease
    about our recitation of our traditional abuse of discretion formulation, which we developed
    and apply in other contexts. I observe that this formulation appears to be inconsistent with
    the abuse of discretion standard employed by the federal courts in the Daubert context, as
    well as the careful and searching examination that this Court is conducting in reviewing
    these cases. For the reasons expressed below, I believe that when this Court applies abuse
    of discretion when reviewing expert witness admissibility determinations, we should
    articulate an abuse of discretion standard that is in line with the federal courts’ formulation
    and that reflects this Court’s current practice. When we adopted the Daubert standard, we
    adopted it in full. That necessarily includes the federal courts’ application of the abuse of
    discretion standard. I explain my reasoning more fully below.
    A. Rochkind—This Court’s Decision to Adopt Daubert
    In 2020, Maryland joined the supermajority of states that adopted Daubert3 as the
    standard for admission of expert testimony. See Rochkind v. Stevenson, 
    471 Md. 1
     (2020).
    The change to the Daubert standard shifted the focus of the analysis from the general
    acceptance of a methodology under Frye-Reed4 to the reliability of a methodology.
    Rochkind, 471 Md. at 5, 31; Daubert, 
    509 U.S. at
    594–95. In doing so, this Court adopted
    the Supreme Court’s Daubert trilogy for analyzing admissibility of expert opinions. See
    Daubert, 
    509 U.S. 579
    , 588–89 (replacing the “general acceptance” test under Frye, with
    3
    Daubert v. Merrell Dow Pharmaceuticals, Inc., 
    509 U.S. 579
     (1993).
    4
    Before the adoption of Daubert in Maryland, Frye-Reed was the prevailing
    standard for admissibility of expert evidence. See Reed v. State, 
    283 Md. 374
     (1978).
    3
    the more flexible standard in determining whether scientific evidence is reliable and
    admissible); General Electric Co. v. Joiner, 
    522 U.S. 136
    , 139 (1997) (clarifying that the
    proper scope of appellate review of a trial court’s rulings on expert admissibility is “abuse
    of discretion”); Kumho Tire Co., Ltd. v. Carmichael, 
    526 U.S. 137
    , 141–42 (1999)
    (clarifying that the trial judge’s Rule 702 gatekeeping duties apply to all expert testimony,
    whether such testimony is based upon scientific, technical, or other specialized
    knowledge).
    In adopting the Daubert standard, we expressed doubt that adopting the standard
    would “upend Maryland evidence law[,]” and we observed that, by adopting Daubert,
    “Maryland courts will be able to ‘draw from and contribute to the broad base of case law
    grappling with scientific testimony.’” Rochkind, 471 Md. at 34–35 (quoting Savage v.
    State, 
    455 Md. 138
    , 185 (2017) (Adkins, J., concurring)).
    Prior to Rochkind, Maryland appellate courts reviewed a trial court’s decision
    concerning the admissibility of expert testimony under two different standards of review—
    conducting a de novo review of a trial court’s determinations under Frye-Reed and
    reviewing the trial court’s determinations under Rule 5-702 for abuse of discretion. See
    Rochkind, 471 Md. at 37. With the change to Daubert, we recognized that appellate review
    of a trial court’s decision to admit or exclude expert opinion testimony would be the abuse
    of discretion standard. Id. In abrogating the Frye-Reed standard in favor of Daubert, “we
    reiterated that a trial court’s ruling to admit or to exclude expert witness testimony ‘will
    seldom constitute a ground for reversal.’” State v. Matthews, 
    479 Md. 278
    , 306 (quoting
    Rochkind, 471 Md. at 10 (quoting Roy v. Dackman, 
    445 Md. 23
    , 38–39 (2015))). We did
    4
    not, however, grapple with the fact that the formulation of the abuse of discretion standard
    that we have traditionally applied to evidentiary rulings and similar decisions is different
    from the formulation employed by federal courts applying the Daubert standard. As
    discussed below, since the transition to the Daubert standard, we have conducted searching
    and exacting reviews of the record.
    B.     This Court’s Post-Rochkind Appellate Review of Circuit Courts’ Expert
    Testimony Rulings
    1. State v. Matthews
    Having set the stage in Rochkind for transition from a de novo standard of review
    under Frye-Reed to an abuse of discretion standard under Daubert, we undertook our first
    appellate review of a trial court’s Daubert-Rochkind analysis in Matthews. 479 Md. at 278.
    Acknowledging that our review was for an abuse of discretion, the Court relied on cases
    outside the expert testimony context and incorporated some of our traditional definitions
    of that standard, stating that
    an appellate court does ‘not reverse simply because the . . . court would not
    have made the same ruling.’ Devincentz v. State, 
    460 Md. 518
    , 550 (2018).
    ‘Rather, the trial court’s decision must be well removed from any center mark
    imagined by the reviewing court and beyond the fringe of what that court
    deems minimally acceptable.’ Id.; see also Williams v. State, 
    457 Md. 551
    ,
    563 (2018) (“An abuse of discretion occurs where no reasonable person
    would take the view adopted by the circuit court.”); Jenkins v. State, 
    375 Md. 284
    , 295–96 (2003) (“Abuse occurs when a trial judge exercises discretion
    in an arbitrary or capricious manner or when he or she acts beyond the letter
    or reason of the law.”).
    Matthews, 479 Md. at 305–06 (cleaned up). In addition, we “reaffirmed” the sentiments
    expressed in Rochkind, stating that “it is still the rare case in which a Maryland trial court’s
    exercise of discretion to admit or deny expert testimony will be overturned.” Id. at 306.
    5
    In Matthews, we held that the trial court did not abuse its discretion by failing to
    exclude a photogrammetry expert’s testimony due to the expert’s inability to provide a
    margin of error that accounted for several potential variables relating to height estimates
    of a criminal defendant in connection with a reverse projection photogrammetry analysis.
    Id. at 325. We determined that “[t]here [was] no dispute that [the expert’s] methodology
    was reliable.” Id. at 313. “Nor was there any analytical gap in [the expert’s] proffered
    testimony.”     Id.   Instead, we concluded that “[t]he unknown degree of uncertainty
    concerning the accuracy of [the expert’s] height estimate went to the weight the jury should
    give to the expert testimony, not to its admissibility.” Id. (footnote omitted). The Majority
    opinion in this case describes our careful and searching examination of the record in
    Matthews, see Maj. Slip Op. at 29–32, which I need not repeat here.
    2. Abruquah v. State
    Our second case involving the application of Daubert-Rochkind was Abruquah v.
    State. In that case, we were asked to determine whether, in a murder case, the trial court
    erred in permitting a firearms expert to testify, over objection, that “each of the four bullets
    and the bullet fragment . . . ‘at some point’ ‘had been fired’ from or through ‘the Taurus
    revolver[]’”—the gun that had been located at the criminal defendant’s home and was the
    alleged murder weapon. 483 Md. at 680. The expert “testified neither that his opinion was
    offered to any particular level of certainty nor that it was subject to qualifications or
    caveats.” Id.
    In connection with our review, we “discuss[ed] general background on the firearms
    identification methodology employed by the State’s expert witness, criticisms of the
    6
    methodology, studies of the methodology, the testimony presented to the circuit court, and
    caselaw from other jurisdictions.” Id. at 653; 656–79. We then considered the trial court’s
    application of the Daubert-Rochkind factors. See id. at 680–98. After undertaking this
    review, based on the evidence presented at the hearings, we held that the circuit court did
    not abuse its discretion in ruling that the firearms expert “could testify about firearms
    identification generally,” as well as “his examination of the bullets and bullet fragments
    found at the crime scene, his comparison of that evidence to bullets known to have been
    fired from the [criminal defendant’s] Taurus revolver, and whether the patterns and
    markings on the crime scene bullets are consistent or inconsistent with the patterns and
    markings on the known bullets.” Id. at 698. With regard to the circuit court’s decision to
    “permit[] the State’s expert witness to opine without qualification that the crime scene
    bullets were fired from [the defendant’s] firearm[,]” we determined that the circuit court
    abused its discretion because the studies and other information in the record do not “support
    the use of [the firearms identification methodology] to reliably opine without qualification
    that the bullets of unknown origin were fired from the particular firearm.” Id. at 694–95,
    698. We explained that “[b]ecause the court’s error was not harmless beyond a reasonable
    doubt,” we reversed the circuit court’s ruling on the defendant’s motion in limine, vacated
    his conviction, and remanded the case for a new trial. Id. at 698.
    In undertaking our review, we acknowledged that we were applying an abuse of
    discretion standard. Id. at 652. We also observed that in Matthews, we applied the same
    “frequently described” formulation of that standard from our case law—namely, that an
    abuse of discretion occurs “when ‘no reasonable person would take the view adopted by
    7
    the circuit court’ or when a decision is ‘well removed from any center mark imagined by
    the reviewing court and beyond the fringe of what the court deems minimally acceptable.’”
    Abruquah, 483 Md. at 652 n.5 (quoting Matthews, 479 Md. at 305 (first quoting Williams,
    
    457 Md. at 563
    , and next quoting Devincentz, 
    460 Md. at 550
    )). Recognizing that the
    description of this standard as articulated in our case law really did not fit with the trial
    court’s careful and thoughtful approach to the application of the Daubert factors, we stated:
    In our view, the application of those descriptions to a trial court’s application
    of a newly adopted standard, such as that adopted by this Court in Rochkind
    as applicable to the admissibility of expert testimony, is somewhat unfair. In
    this case, in the absence of additional caselaw from this Court implementing
    the newly adopted standard, the circuit court acted deliberately and
    thoughtfully in approaching, analyzing, and resolving the question before it.
    This Court’s majority has come to a different conclusion concerning the outer
    bounds of what is acceptable expert evidence in this area.
    Abruquah, 483 Md. at 652 n.5.
    3. Oglesby v. Baltimore School Associates
    Last month, this Court issued its opinion Oglesby v. Baltimore School Associates.
    Oglesby was a lead paint case in which the plaintiff sued the owners and managers of an
    apartment building for negligence, negligent misrepresentation, and a violation of the
    Maryland Consumer Protection Act. Oglesby, 
    2023 WL 4755689
     at *1. We were asked
    to consider whether the circuit court erred in granting summary judgment in favor of the
    property owners after the circuit court excluded the plaintiff’s expert testimony to establish
    that the property in question was a source of lead exposure and a significant factor
    8
    contributing to the plaintiff’s alleged injuries, including a loss of IQ points.5 
    Id.
     The
    plaintiff’s causation expert, Steven Elliot Caplan, M.D., “concluded that [the plaintiff’s]
    likely exposure to lead at the property was a significant contributing factor to bringing
    about the [alleged] cognitive deficiencies and impairments[,]” and to a loss of IQ points.
    
    Id.
     The property owners contended that Dr. Caplan lacked a sufficient factual basis for his
    causation opinions, and that his methodology used to calculate the plaintiff’s IQ loss was
    not reliable or generally accepted. Id. at *2. After a hearing at which the circuit court
    considered the property owners’ motions to preclude expert opinions and motions for
    summary judgment, the court granted both motions. Id. The circuit court agreed with the
    property owners that Dr. Caplan lacked a sufficient factual basis for his opinions, and that,
    without Dr. Caplan’s testimony as to causation, the plaintiff was unable to establish a prima
    facie case for negligence. Id.
    After the Appellate Court affirmed the circuit court’s judgment, we granted
    certiorari to determine whether the circuit court erred in ruling that Dr. Caplan’s opinions
    lacked a sufficient factual basis. Id. In undertaking our review, we once again described
    our abuse of discretion standard of review “as occurring ‘where no reasonable person
    would take the view adopted by the trial court,’ or when ‘the decision under consideration
    5
    In Oglesby, the trial court analyzed the proffered expert testimony under the Frye-
    Reed standard because the trial court proceedings took place before this Court issued its
    decision in Rochkind, and, therefore, Oglesby did not involve review of the trial court’s
    application of the Daubert-Rochkind factors. However, the parties petitioned and briefed
    this Court for review of the trial court’s determination under the post-Rochkind abuse of
    discretion standard, and this Court conducted a careful and searching review of the record.
    9
    is well removed from any center mark imagined by the reviewing court and beyond the
    fringe of what that court deems minimally acceptable[.]’” Id. at *12 (cleaned up).
    After conducting a “careful review of the record,” we determined that “Dr. Caplan’s
    opinions had a sufficient factual basis,” and that “the circuit court resolved genuine
    disputes of material fact.” Id. at *2. We further concluded that Dr. Caplan “had more than
    an adequate supply of data from which to form the opinion and the methodology he
    employed was reliable.” Id. As such, we held that the circuit court “abused its discretion
    in granting the motion to preclude and in determining that Dr. Caplan’s testimony that [the
    plaintiff’s] exposure to lead at the property was a significant contributing factor to her
    injures was inadmissible.” Id.
    In undertaking our review, we determined that the court resolved certain facts in the
    property owners’ favor, such as whether there was lead paint present at the property at the
    time that the plaintiff lived there, whether the plaintiff had come into contact with lead at
    the property, through peeling or chipping paint, and whether the plaintiff was potentially
    exposed to lead at other locations and the condition of those properties. Id. at *16. We
    stated, “[i]n short, a trial court is not permitted to resolve disputes of material fact in
    determining whether a sufficient factual basis exists to support an expert’s opinion. Doing
    so is a clear abuse of discretion.” Id.
    We proceeded to discuss in detail the evidence produced by the plaintiff, and the
    data that Dr. Caplan relied upon to support the causation links necessary to avoid summary
    judgment in a lead paint case. Id. at *17–22. After conducting our own review of the
    evidence, we stated that:
    10
    Under the particular circumstances of this case, given the large quantity of
    data that [the expert] had available and reviewed and the nature of the
    challenge to the admissibility of his testimony, a remand for further
    proceedings as to Dr. Caplan’s opinion that [the plaintiff’s] exposure to lead
    at the property was a substantial contributing factor to her injuries (other than
    IQ loss) is not warranted.
    Id. at *22.
    Concerning Dr. Caplan’s testimony that the plaintiff suffered an IQ loss of 3-4
    points, in reliance upon particular lead paint studies—the Lanphear study and the Canfield
    study—we discussed our Court’s previous case law in lead paint cases involving experts’
    reliance upon the Lanphear study. Id. at *23–26. We concluded that “[t]he bottom line is
    that these cases demonstrate that experts have been permitted to rely on the Lanphear study
    and extrapolate from its findings and render an opinion that an individual suffered a
    specified loss of IQ points as a result of exposure to lead.” Id. at *26.
    We observed that, “although our case law demonstrates that it is permissible for an
    expert to rely on the Lanphear study to offer an opinion that exposure to lead resulted in a
    specific loss of IQ points, the record in this case shows that neither Dr. Caplan’s report nor
    his deposition testimony fully explains the basis for his calculations under either” the
    Lanphear study or the Canfield study. Id. at *27. We stated that “without a full explanation
    of his methodology, including the reasons for his choices, even though it is possible to
    discern the basis of the calculations, we cannot determine whether it was an abuse of
    discretion for the circuit court to preclude this aspect of Dr. Caplan’s testimony.” Id.
    We remanded the case for a Daubert-Rochkind hearing “for the circuit court to
    determine whether the calculations that Dr. Caplan employed using the Lanphear study are
    11
    reliable and to assess Dr. Caplan’s use of the Canfield study and the reliability of his
    methodology with respect to it, should [the plaintiff] seek to introduce evidence concerning
    her IQ loss at trial.” Id. at *28.
    4. The Case at Hand
    This case is our fourth opportunity since Rochkind to review expert witness
    admissibility determinations for abuse of discretion. In undertaking this review, we state
    that with the adoption of Daubert, “we promised the deference appropriate to courts
    administering a flexible approach to analyzing the admissibility of expert testimony[]” and
    that “[t]his case requires us to reflect on that flexibility and deference.” Maj. Slip Op. at 1.
    In connection with that deferential review, we once again recite our traditional
    formulation of the abuse of discretion standard as described in our case law—such as
    upholding a trial court’s decision unless it is “well removed from any center mark imagined
    by the reviewing court and beyond the fringe of what that court deems minimally
    acceptable” or “no reasonable person would take the view adopted by the circuit court.”
    Maj. Slip Op. at 23 (citations omitted).
    In the context of our review in this case, we have considered the trial court’s stated
    reason for its decision to exclude Ms. Cardell as an expert witness alongside our own
    searching review of the record. In doing so, we carefully lay out the contents of the record,
    including the data Ms. Cardell used in her before-and-after method calculations, how Ms.
    Cardell reached her conclusions, the issues raised at the hearing, and the trial court’s stated
    reasons for excluding Ms. Cardell as an expert. Maj. Slip Op. at 7–21. We ultimately
    agree with the trial court that there were “speculative, insufficiently substantiated judgment
    12
    calls that were central to Ms. Cardell’s application of the before-and-after method.” Maj.
    Slip Op. at 42. Specifically, this Court concludes the trial court did not abuse its discretion
    in determining that Ms. Cardell’s “judgment calls” regarding the selection of the 2015 base
    year, treatment of member draws, and failure to factor in confounding variables in her
    methodology were matters bearing on the reliability of the methodology and that each
    weighed against admissibility. Maj. Slip Op. at 42.
    We explain that:
    We rely on trial courts that conduct Daubert-Rochkind hearings to determine
    where the line between data and methodology is in the specific cases before
    them, and whether the proffered expert’s choices relating to data,
    assumptions, and other inputs implicate the reliability of the expert’s
    methodology. In this part of the trial court’s analysis, the court did just that.
    Maj. Slip Op. at 42–43. Although this Court “discern[s] no error in most of the trial court’s
    application of the Daubert-Rochkind factors,” we conclude that the trial court erred in its
    determination that Ms. Cardell’s treatment of the 2021 updates implicated the reliability of
    her methodology. Maj. Slip Op. at 46. The trial court based this conclusion on its
    observation that Ms. Cardell made this update without new information and “‘for fully
    subjective reasons.’” Id. In determining that the trial court abused its discretion in
    weighing this information against admissibility, we carefully look at the record and
    conclude that Ms. Cardell was basing her decision on new information because she
    “noticed something she had not noticed before on first examination[,]” and “[c]atching
    something peculiar the second time around neither undermines the adequacy of the data . . .
    nor the court’s understanding of the expert’s methodology[.]” Maj. Slip Op. at 46–47. In
    other words, this Court rejects the trial court’s determination that Ms. Cardell’s June 2021
    13
    adjustments were made for “fully subjective reasons” based on our own review of the
    record. Maj. Slip Op. at 46. We go on to explain that Ms. Cardell’s decision to make the
    adjustments “did not implicate the reliability of her methodology. At most, it went to the
    care with which she applied her methodology, which is a matter to be explored on cross-
    examination before the jury (if Ms. Cardell’s testimony is otherwise found to be
    sufficiently reliable).” Maj. Slip Op. at 47–48.
    We also hold that “[t]he trial court . . . erred when it stated that Ms. Cardell’s lost-
    profit calculations would not be helpful to the jury.” Maj. Slip Op. at 47, n.16. The trial
    court’s reasoning for weighing this factor against admission was because “if the jury
    concluded that one (or more) of the departing physicians left for reasons other than
    KatzAbosch’s negligence, a lost-profits analysis that included that physician’s projected
    revenue would overstate the practice’s lost profits and therefore not help the jury to
    accurately calculate damages.” Id. On review, we determine that “there was no way to
    know before the trial whether the jury would conclude that any of the physicians left for
    reasons other than KatzAbosch’s negligence[,]” and, therefore, the trial court’s conclusion
    “was speculative and not a proper ground for exclusion.” Id.
    Importantly, this Court points out that there is overlap between matters of reliability
    and credibility, as well as data and methodology, which can create challenges for trial
    courts applying Daubert.      Maj. Slip Op. at 32–43.       As such, making admissibility
    determinations on these types of matters is necessarily fact-specific, both before the trial
    court and on appellate review. These questions often require the reviewing court to do
    more than simply accept a trial court’s justification for an admissibility determination as a
    14
    matter within the trial court’s discretion. Indeed, when conducting a review of reliability
    determinations, a reviewing court may need to look carefully at the record, including the
    underlying facts of the case, methodology employed, data relied upon, and contents of
    expert testimony and reports to determine if the trial court abused its discretion. This is
    exactly what this Court does here, and this is in line with how federal courts apply the
    abuse of discretion standard in the Daubert context.
    I take no issue with the manner in which we have undertaken our review in this case,
    or in Matthews, Abruquah, and Oglesby, as I joined the majority opinions authored by my
    colleagues in each of these cases. In the aftermath of Rochkind, I do not challenge the
    nature of our review for an abuse of discretion, but I question whether our reliance upon
    our frequently described formulation of this standard, in fact, reflects the nature and type
    of review that we are undertaking in these cases and that federal courts have employed in
    undertaking the same exercise.
    C. “Abuse of Discretion” as a Context-Dependent Standard
    In the context of appellate review, reviewing courts use one of three well-known
    standards of review. An appellate court reviews questions of law de novo. Questions of
    fact are reviewed under the clearly erroneous standard. Finally, matters that are committed
    to the discretion of the trial court—such as whether to admit expert testimony—are
    reviewed for an “abuse of discretion.” Of the three standards of review, the abuse of
    discretion standard is typically considered to be the most deferential. However, it has been
    described as “famously slippery,” and has been understood to have different meanings and
    applications in different contexts. Zervos v. Verizon, 
    252 F.3d 163
    , 168 n.4 (2d Cir. 2001);
    15
    Gasperini v. Center for Humanities, Inc., 
    149 F.3d 137
    , 141 (2d Cir. 1998) (noting that
    abuse of discretion “may have different meanings in different contexts” depending on why
    the decision is within the trial court’s discretion); Lawson Prods., Inc. v. Avnet, Inc., 
    782 F.2d 1429
    , 1438 (7th Cir. 1986) (describing abuse of discretion as a term “that seems to
    escape easy definition”). In attempting to explain the difference between the abuse of
    discretion standard and clearly erroneous standard, the Seventh Circuit has observed:
    Abuse of discretion is conventionally regarded as a more deferential standard
    than clear error, though whether there is any real or consistent difference has
    been questioned. The alternative view is that both standards denote a range,
    rather than a point, that the ranges overlap and maybe coincide, and that the
    actual degree of scrutiny in a particular case depends on the particulars of
    that case rather than on the label affixed to the standard of appellate review.
    Haugh v. Jones & Laughlin Steel Corp., 
    949 F.2d 914
    , 916–17 (7th Cir. 1991).
    The type of ruling under review informs how appellate courts apply the abuse of
    discretion standard. Dissenting in Calderon v. Thompson, Justice David Souter wrote that
    “the variety of subjects left to discretionary decision requires caution in synthesizing abuse
    of discretion cases.” 
    523 U.S. 538
    , 567 (1998) (Souter, J. dissenting) (citations omitted).
    In the same vein, Judge Henry J. Friendly recommended that when applying the abuse of
    discretion standard, appellate courts “must carefully scrutinize the nature of the trial court’s
    determination and decide whether that court’s superior opportunities of observation or
    other reasons of policy require greater deference than would be accorded to its formulations
    of law or its application of law to the facts.” Henry J. Friendly, Indiscretion About
    Discretion, 
    31 Emory L.J. 747
    , 784 (1982). In the decades since Daubert was decided,
    16
    federal courts have shaped the abuse of discretion standard in reviewing admissibility
    determinations of expert witness testimony.
    D. The Federal Courts’ Description of the Abuse of Discretion Standard in the
    Daubert Context
    As discussed, after Daubert, the Supreme Court made clear that the abuse of
    discretion standard is the standard that a reviewing court must apply in undertaking a
    review of a trial court’s decision to admit or exclude expert testimony. See Joiner, 
    522 U.S. at 136
    ; Kumho, 
    526 U.S. at 142
     (explaining that “the law grants a [trial] court the
    same broad latitude when it decides how to determine reliability as it enjoys in respect to
    its ultimate reliability determination[]”).
    Although the standard was established by these cases, it does not always appear to
    be applied as the hands-off deferential approach that often accompanies appellate review
    for an abuse of discretion. Instead, it involves a more searching and careful examination
    of the records, data, studies, testimony, and trial court’s analysis of the same. Writing for
    the American Bar Association, lawyers David F. Herr and Morgan L. Holcomb observed
    that
    after [Joiner] the proper standard of review is clearly the “abuse-of-
    discretion” standard. However, even Joiner itself appeared to apply quite
    searching review. As Justice Stevens pointed out in his partially concurring
    opinion, Part III of the majority opinion went beyond announcing the
    appropriate standard of review, and carefully examined the basis on which
    the trial court excluded the expert evidence. Joiner, 
    522 U.S. at
    151–55
    (Steven[s], J.[,] concurring).
    This careful examination has not gone unnoticed; for example, one
    commentator noted, “while they don’t say so, some appellate opinions come
    close to a de novo style of review of the proffered expert testimony.” Michael
    J. Saks, et al., Annotated Reference Manual on Scientific Evidence Second
    17
    23 (West 2004) (citing Mathis v. Exxon Corp., 
    302 F.3d 448
     (5th Cir. 2002)
    (noting that the trial court had failed to “offer any reasons in support of
    admitting the expert’s testimony,” but nonetheless holding that the appellate
    court did not have to remand and could carry out the review itself on appeal:
    “Because admissibility is a legal question—one ill-suited to remand and
    further explication by the [trial] court—we will decide the question in this
    case without remand.”)).
    David F. Herr & Morgan L. Holcomb, Opinion and Expert Testimony in Federal and State
    Courts, Am. L. Inst. – Am. Bar Ass’n 629, 642 (Jan. 2007). Herr and Holcomb go on to
    catalog additional examples of federal circuit courts conducting “searching” reviews of
    admissibility determinations in the expert witness context. 
    Id.
     (citing United States v.
    Mitchell, 
    365 F.3d 215
     (3d Cir. 2004) (conducting a searching analysis of the admission of
    fingerprint evidence, and concluding that admission of the prosecution’s expert was not
    error); Lauzon v. Senco Prods., Inc., 
    270 F.3d 681
     (8th Cir. 2001) (conducting a fairly
    searching review, and ultimately holding that “[t]hrough examination of the record in light
    of the requirements of Daubert and its progeny, ineluctably we are led to conclude the
    [trial] court’s exclusion of the testimony was an abuse of discretion and fell outside the
    spirit of admissibility as set forth in Federal Rule of Evidence 702.”); Pride v. BIC Corp.,
    
    218 F.3d 566
    , 578 (6th Cir. 2000) (noting review of Daubert ruling was for abuse of
    discretion, but conducting “de novo review of the record” to conclude that the “record
    supports the [trial] court’s finding on this issue”)). The authors also note the reluctance of
    appellate courts to remand these issues because appellate courts tend to make admissibility
    determinations themselves. 
    Id.
     at 642–43.
    Herr and Holcomb’s description of how federal courts apply the abuse of discretion
    standard in the expert witness context—like this Court’s application of the standard in
    18
    Matthews, Abruquah, Oglesby, and here—feels out of sync with the even more deferential
    approach that we take when reviewing other types of discretionary decisions by trial courts
    and our “no reasonable person” and “well removed from the center mark” articulation of
    the standard.
    In the Daubert context, various federal circuits have supplied their own descriptions
    of the nature of the abuse of discretion standard, a few of which I highlight here. In its
    description of the standard, the First Circuit explains that a trial court abuses its discretion
    “when a material factor deserving significant weight is ignored, when an improper factor
    is relied upon, or when all proper and no improper factors are assessed, but the court makes
    a serious mistake in weighing them.” Lawes v. CSA Architects and Engineers, LLP, 
    963 F.3d 72
    , 90 (1st Cir. 2020) (quotations omitted). The Court notes that the abuse of
    discretion “standard is not monolithic: within it, embedded findings of facts are reviewed
    for clear error, questions of law are reviewed de novo, and judgment calls are subjected to
    classic abuse-of-discretion review.” 
    Id.
     (internal quotations omitted). Accordingly, the
    reviewing court will reverse a trial court’s decision if it determines the trial court
    “committed a material error of law or a meaningful error in judgment.” 
    Id.
     (quotations
    omitted).
    The Tenth Circuit has described the standard as follows:
    We review de novo the question of whether the [trial] court applied the
    proper legal test in admitting an expert’s testimony. Though the [trial] court
    has discretion in how it conducts the gatekeeper function, we have
    recognized that it has no discretion to avoid performing the gatekeeper
    function. Therefore, we review de novo the question of whether the [trial]
    court applied the proper standard and actually performed its gatekeeper role
    in the first instance. We then review the trial court’s actual application of the
    19
    standard in deciding whether to admit or exclude an expert’s testimony for
    abuse of discretion. The trial court’s broad discretion applies both in
    deciding how to assess an expert’s reliability, including what procedures to
    utilize in making that assessment, as well as in making the ultimate
    determination of reliability. Accordingly, we will not disturb the [trial]
    court’s ruling unless it is arbitrary, capricious, whimsical or manifestly
    unreasonable or when we are convinced that the [trial] court made a clear
    error of judgment or exceeded the bounds of permissible choice in the
    circumstances.
    Dodge v. Cotter Corp., 
    328 F.3d 1212
    , 1223 (10th Cir. 2003) (quotations omitted) (cleaned
    up). That court has also emphasized that a “natural requirement of the gatekeeper function
    is the creation of a sufficiently developed record[,]” and has observed that “Kumho and
    Daubert make it clear that the [trial] court must, on the record, make some kind of reliability
    determination.” 
    Id. at 1223
     (cleaned up) (citations omitted). Thus, the Tenth Circuit has
    held “when faced with a party’s objection, a [trial] court must adequately demonstrate by
    specific findings on the record that it has performed its duty as gatekeeper.” 
    Id.
     (cleaned
    up). The court has stated that:
    Without specific findings or discussion on the record, it is impossible on
    appeal to determine whether the [trial] court carefully and meticulously
    reviewed the proffered scientific evidence or simply made an off-the-cuff
    decision to admit the expert testimony. In the absence of such findings, we
    must conclude that the court abused its discretion in admitting such
    testimony.
    
    Id.
     (quotations omitted).
    The Seventh Circuit describes a “two-step standard of review in cases challenging
    a [trial] court’s admission or exclusion of the testimony of an expert.” C.W. ex rel. Wood
    v. Textron, Inc., 
    807 F.3d 827
    , 835 (7th Cir. 2015). First, the court “review[s] de novo a
    [trial] court’s application of the Daubert framework.” 
    Id.
     (citing United States v. Brumley,
    20
    
    217 F.3d 905
    , 911 (7th Cir. 2000)). If the reviewing court determines that “the [trial] court
    properly adhered to the Daubert framework,” then it reviews the “decision to exclude (or
    not to exclude) expert testimony for abuse of discretion.” 
    Id.
    The Eleventh Circuit explained the rationale behind the abuse of discretion standard
    in Daubert, as well as that Court’s application of the standard, in United States v. Brown,
    
    415 F.3d 1257
     (11th Cir. 2005). That Court explained that the standard is proper because
    the trial court is “[i]mmersed in the case as it unfolds” and is “more familiar with the
    procedural and factual details and is in a better position to decide Daubert issues.” 
    Id. at 1266
    . That court explained:
    The rules relating to Daubert issues are not precisely calibrated and must
    be applied in case-specific evidentiary circumstances that often defy
    generalization. And we don’t want to denigrate the importance of the trial
    and encourage appeals of rulings relating to the testimony of expert
    witnesses. All of this explains why the task of evaluating the reliability of
    expert testimony is uniquely entrusted to the [trial] court under Daubert,
    and why we give the [trial] court considerable leeway in the execution of
    its duty.
    
    Id.
     (quotations omitted). However, the court goes on to clarify,
    [t]o be sure, review under an abuse of discretion standard does entail review,
    and granting considerable leeway is not the same thing as abdicating
    appellate responsibility. We will reverse when the [trial] court’s Daubert
    ruling does amount to an abuse of discretion that affected the outcome of a
    trial. An abuse of discretion can occur where the [trial] court applies the
    wrong law, follows the wrong procedure, bases its decision on clearly
    erroneous facts, or commits a clear error in judgment. In the Daubert
    context, if one of those types of serious error occur, we may conclude that
    the [trial] court has not properly fulfilled its role as gatekeeper. We have also
    found that a [trial] court abuses its discretion where it fails to act as a
    gatekeeper by essentially abdicating its gatekeeping role.
    
    Id.
     (quotations omitted).
    21
    E.     Some Concluding Thoughts on Our Standard of Review in the
    Daubert Context
    When we made the decision in Rochkind to adopt Daubert, we did so with the
    knowledge that there is a “broad base of case law” that went along with the new standard.
    Rochkind, 471 Md. at 34–35 (quoting Savage, 
    455 Md. at 185
     (Adkins, J., concurring)).
    However, in describing the abuse of discretion standard of review, we transplanted our
    formulation of the standard we apply in other contexts. See, e.g., Devincentz, 
    460 Md. at 539, 550
     (applying the abuse of discretion standard to the review of a trial court’s decision
    to admit or exclude a character witness’s opinion); Williams, 
    457 Md. at
    562–63 (applying
    the abuse of discretion standard to the review of a trial court’s decision to admit relevant
    evidence); Jenkins, 
    375 Md. at
    295–96 (applying the abuse of discretion standard to the
    review of a trial court’s decision to deny a motion for a new trial). In my view, our
    traditional formulation of the abuse of discretion standard—as invoking the articulation of
    “no reasonable person” or “well removed from any center mark imagined by a reviewing
    court and beyond the fringe of what the court deems minimally acceptable”—is not the
    best or most accurate way of describing our abuse of discretion review in the Daubert-
    Rochkind context. See Williams, 
    457 Md. at 563
    ; Devincentz, 
    460 Md. at 550
    .
    I believe that this Court should re-formulate a definition of our abuse of discretion
    standard in the context of appellate review of a trial court’s application of the Daubert-
    Rochkind factors when analyzing whether to admit or exclude expert testimony. In doing
    so, we should look to the federal courts’ articulation of the standard of review in the same
    manner that we will look to the federal courts’ jurisprudence in its application. I agree with
    22
    the observation made by Judge Friendly and various federal circuits that the abuse of
    discretion standard is contextual. See, e.g., Gasperini, 
    149 F.3d at 141
     (noting that abuse
    of discretion may have different meanings in different contexts depending on why the
    decision is within the trial court’s discretion); Lawes, 963 F.3d at 90 (explaining that the
    abuse of discretion standard is not monolithic).
    As is clearly reflected in our appellate review of the trial court’s expert admissibility
    determinations in Matthews, Abruquah, Oglesby, and the instant case, our abuse of
    discretion standard of review involves a searching and careful examination of the record,
    data, studies, testimony, and trial court’s analysis of the same. Our traditional formulation
    of the abuse of discretion standard does not comport with the type of review that we are
    undertaking here. I believe that the parties, their counsel and experts, trial courts and our
    Appellate Court would all benefit from this Court re-formulating our description of the
    abuse of discretion standard of review that is undertaken in this context and, in so doing,
    utilize the federal courts’ discussion of that standard.
    Our standard should reflect that the hallmark of the abuse of discretion standard
    regardless of context is the high degree of deference to a court exercising discretionary
    authority. Joiner, 
    522 U.S. at 137, 143
    . “Accordingly, we will not disturb the [trial] court’s
    ruling unless it is arbitrary, capricious, whimsical, or manifestly unreasonable or when we
    are convinced that the [trial] court made a clear error of judgment or exceeded the bounds
    of permissible choice in the circumstances.” Dodge, 
    328 F.3d at 1223
     (quotations omitted).
    But this broad discretion “is not the same thing as abdicating appellate responsibility.”
    Brown, 
    415 F.3d at 1266
    . In the context of admissibility of expert testimony, this means
    23
    that while trial courts are given leeway on matters within the range of their discretion, on
    review, the appellate court will find an abuse of discretion where the trial court’s rulings
    are “manifestly erroneous.” Joiner, 
    522 U.S. at 142
    . That is, when applying the abuse of
    discretion standard in the expert witness context, we reverse only where there is “a
    meaningful error in judgment.” Lawes, 963 F.3d at 90 (emphasis added). In determining
    whether the trial court has made such a meaningful error, reviewing courts may need to
    conduct a searching review of the record.
    With these points in mind, our standard should reflect that the “[abuse of discretion]
    standard is not monolithic: within it, embedded findings of fact are reviewed for clear error,
    questions of law are reviewed de novo, and judgment calls are subjected to a classic abuse-
    of-discretion review.” Lawes, 963 F.3d at 90 (quotations omitted). A reviewing court may
    find a trial court abused its discretion where the trial court (1) “applies the wrong law;” (2)
    “follows the wrong procedure;” (3) “bases its decision on clearly erroneous facts;” (4)
    “commits a clear error in judgment;” or (5) abdicates its gatekeeping role altogether.
    Brown, 
    415 F.3d at 1266
    ; see also Lawes, 963 F.3d at 90. Although trial courts have
    discretion for determining the best procedure for the proper resolution of any expert
    admissibility issues, see Kumho, 
    526 U.S. at 152
    , trial judges should be reminded that a
    reviewing court will find an abuse of discretion where they fail to act as a gatekeeper by
    essentially abdicating their gatekeeping role, or where the trial court assumes the role of
    the jury. Part of this gatekeeping role is developing a clear record of the court’s basis for
    its determination to admit or exclude expert testimony. See Dodge, 
    328 F.3d at 1223
    .
    24
    Our formulation of the standard should also take into account that appellate review
    of expert admissibility determinations encompasses many types of challenges, including
    whether: expert testimony is necessary or appropriate; the expert is qualified; the expert’s
    area of expertise is unreliable; the expert’s conclusions are invalid or unreliable; the
    testimony exceeds the expert’s area of expertise; and the expert testimony is relevant. In
    my view, “the actual degree of scrutiny in a particular case [may] depend[] on the
    particulars of that case rather than on the label affixed to the standard of appellate review.”
    Haugh, 
    949 F.2d at
    916–17; see also Brown, 
    415 F.3d at 1265
    . A re-formulated standard
    should reflect this flexibility.
    As far as our statements in Rochkind and Matthews that it will be the “rare case in
    which a Maryland trial court’s exercise of discretion to admit or deny expert testimony will
    be overturned,” I suppose only time will tell. Matthews, 479 Md. at 286, 306 (emphasis
    added).
    25
    Circuit Court for Howard County
    Case No. C-13- CV-18-000181
    Argued: May 4, 2023                                 IN THE SUPREME COURT
    OF MARYLAND*
    No. 30
    September Term, 2022
    __________________________________
    KATZ, ABOSCH, WINDESHEIM,
    GERSHMAN & FREEDMAN, P.A.,
    ET AL.
    v.
    PARKWAY NEUROSCIENCE AND
    SPINE INSTITUTE, LLC
    __________________________________
    Fader, C.J.,
    Watts,
    Hotten,
    Booth,
    Biran,
    Gould,
    Eaves,
    JJ.
    __________________________________
    Concurring and Dissenting Opinion by
    Gould, J.
    __________________________________
    Filed: August 30, 2023
    *During the November 8, 2022 general election, the voters of Maryland ratified a
    constitutional amendment changing the name of the Court of Appeals to the Supreme Court
    of Maryland. The name change took effect on December 14, 2022
    I respectfully concur in part and dissent in part from the Majority’s well-written and
    thorough opinion. As I understand it, the Majority concludes that the trial court’s analysis
    of the various factors under Daubert-Rochkind could very well support the trial court’s
    discretionary decision to exclude Ms. Cardell’s testimony, but the Majority is concerned
    that the trial court may have unduly considered Ms. Cardell’s updated calculations in
    assessing those factors. I agree with the Majority that Ms. Cardell’s updated calculations
    and her reasons for them do not go to the reliability of her methodology, but are instead
    grist for the cross-examination mill. I write separately because I would not remand for the
    trial court to reconsider its analysis. In my view, Ms. Cardell’s methodology is so
    fundamentally flawed as to constitute “the rare case” in which a trial court’s admission of
    expert testimony would have been an abuse of discretion. See State v. Matthews, 
    479 Md. 278
    , 306 (2022).
    Parkway Neuroscience and Spine Institute, LLC’s (“PNSI”) theory of the case was
    that, beginning in 2013, Katz, Abosch, Windesheim, Gershman & Freedman, P.A.’s and
    Mark Rapson’s (collectively, “KatzAbosch”) negligent advice caused seven of the nine
    member-physicians, who were still with PNSI at the end of 2014, to leave the practice by
    mid-2016, and that their departures caused damages to the company in the form of lost
    profits. Ms. Cardell’s job was to quantify the lost profits caused by those departures. She
    utilized the “before-and-after” method. As the Majority aptly notes, despite the seemingly
    straightforward nature of this method, as an expert, Ms. Cardell was required to make
    decisions requiring judgment and to sufficiently explain those decisions. See Maj. slip op.
    at 33, 42-43 & n.15.
    Ms. Cardell assumed that, if not for the departure of member-physicians, PNSI’s
    profits in 2017 through 2019 would have been, at a minimum, the same as in 2015. So,
    according to Ms. Cardell’s theory, to calculate the lost profits for, say, 2017, you take the
    difference between the profit achieved in 2015 ($321,751) and the loss incurred in 2017
    ($1,515,436) and, voila, you have $1,837,186 of lost profits for 2017.1 Ms. Cardell used
    the same analysis for 2018 and 2019 to arrive at a total lost profits of $4,956,080 for all
    three years. Ms. Cardell thus attributed the entire change in profitability at PNSI to the
    departure of the member-physicians. It is that assumption that requires the exclusion of
    her testimony.
    A critical component of the “before-and-after” method for calculating lost profits is
    accounting or controlling for confounding variables—changes in conditions that might
    have affected the plaintiff’s profitability during the “loss period.”       Accounting for
    confounding variables may be considered, in the lost profits context, akin to “whether the
    expert has adequately accounted for obvious alternative explanations” under the Daubert-
    Rochkind rubric. See Rochkind v. Stevenson, 
    471 Md. 1
    , 36 (2020). Confounding variables
    could include, for instance, the prices of goods and services, consumer demand and
    preferences, technological developments, personnel turnover, and broader industry and
    1
    Due to rounding, the numbers do not add up exactly.
    2
    economic conditions. The failure to account for relevant confounding variables has
    routinely resulted in the exclusion of such testimony.2, 3
    Here, the circuit court determined that Ms. Cardell’s failure to evaluate revenue
    generation by individual member-physicians and failure to consider insurance
    reimbursement rates favored exclusion. In my view, Ms. Cardell’s wholesale failure to
    evaluate and consider those and other relevant factors required the exclusion of her
    testimony.
    2
    See, e.g., E. Auto Distribs., Inc. v. Peugeot Motors of Am., Inc., 
    795 F.2d 329
     (4th
    Cir. 1986) (holding expert testimony properly excluded because it failed to account for
    changes in efficiency of sales force and demand for products); Craftsmen Limousine, Inc.
    v. Ford Motor Co., 
    363 F.3d 761
     (8th Cir. 2004) (holding expert testimony should have
    been excluded because it failed to analyze the effects of general economic conditions and
    increased competition); Schiller & Schmidt, Inc. v. Nordisco Corp., 
    969 F.2d 410
     (7th Cir.
    1992) (holding expert testimony inadequate because it failed to distinguish between the
    damages caused by the lawful entry of a powerful competitor and those caused by the
    competitor’s misconduct); Isaksen v. Vermont Castings, Inc., 
    825 F.2d 1158
    , 1165 (7th
    Cir. 1987) (holding expert testimony inadequate when the plaintiff “made no effort to
    establish how much of the loss was due to [the defendant’s unlawful] activity as distinct
    from unrelated business factors” and “[a]ll [the plaintiff] did to prove damages was to
    compare his average profits for several years before and several years during the period of
    unlawful activity”); Farley Transp. Co. v. Santa Fe Trail Transp. Co., 
    786 F.2d 1342
     (9th
    Cir. 1985) (holding expert testimony inadequate because it failed to distinguish between
    losses caused by lawful and unlawful competition); Coleman Motor Co. v. Chrysler Corp.,
    
    525 F.2d 1338
     (3rd Cir. 1975) (holding expert testimony inadequate for failing to account
    for lawful competition, market changes, and changes in the character of the neighborhood
    where plaintiff’s auto dealership was located).
    3
    Asking experts to provide analysis that implicates relationships between profits
    and confounding variables does not infringe upon the jury’s responsibility to determine
    whether the defendant caused the alleged harm. Here, Ms. Cardell was not called upon to
    determine as part of her analysis whether KatzAbosch’s allegedly poor financial advice
    caused the member-physicians to depart. Rather, she was asked to quantify the lost profits
    caused by those departures.
    3
    It is inconceivable that 100 percent of the changes in profitability between the
    benchmark year of 2015 and the loss years of 2017, 2018, and 2019 were attributable to
    the departures of the member-physicians. Stated differently, it’s inconceivable that had
    none of the member-physicians departed, PNSI’s profits would have remained exactly the
    same. PNSI’s own allegations in its complaint establish this very point: the company’s
    “overhead, expenses and revenue generated by the Members’ practices during the relevant
    time period differed markedly by specialty.” That alone indicates that the impact on the
    company’s profitability from the departure of a particular member-physician would depend
    on the specific costs and revenue structures of that member-physician’s practice.
    As the trial court noted, Ms. Cardell virtually ignored changes in reimbursement
    rates. To the extent she considered them at all, she did not analyze how such changes
    would have affected the total revenues had each of the doctors stayed. Instead, she merely
    speculated that the observed decrease in actual reimbursement rates—which defendant’s
    counsel presented and she acknowledged on cross-examination—would have been offset
    by general increases in demand for medical services.4
    4
    That may or may not be so, but her speculation only reinforces why an expert must
    drill down on the figures. Which specialties were expected to experience increased
    demand? The high profit margin specialties? The low profit margin specialties? What
    marginal expenses would PNSI have incurred in association with that increased demand?
    Did PNSI have the capacity to accommodate such increased demand, and if so, by how
    much? How many additional services would PNSI have had to provide to offset the
    decreased reimbursement rates? That is, where was the break-even point? Ms. Cardell’s
    mere speculation that decreased reimbursement rates would have been offset by increased
    demand for services is no substitute for an actual analysis that accounts for confounding
    variables.
    4
    You do not have to be an expert to know that expenses of any business are subject
    to fluctuations.   Indeed, accounting for such fluctuations was factored into PNSI’s
    compensation model. PNSI’s complaint alleges that a new compensation model developed
    and recommended by KatzAbosch was approved and adopted by its members. Here, PNSI
    alleges that under this new model:
    in determining Member cash flow distributions, each Member regardless of
    specialty was allocated an equal portion of 2/3 of [PNSI’s] net increase or
    decrease in expenses (the “fixed portion”) and a portion of the remaining 1/3
    of [PNSI’s] net increase or decrease in expenses based on the Member’s net
    collections for the period (the “variable portion”).
    Thus, PNSI’s own complaint shows that its revenues and expenses were not static.
    Indeed, PNSI assumed as much in restructuring its compensation model. There is no
    evidence Ms. Cardell analyzed the extent to which changes in expenses or increased
    demand for services could have affected profits had the seven member-physicians
    remained. It would be as if a lost profits analysis of a gas station failed to account for
    changes in the wholesale and retail prices of gasoline.
    Ms. Cardell’s failure to make any attempt to account for such confounding variables
    requires the exclusion of her testimony. Accordingly, I dissent to that part of the Majority’s
    opinion that requires a remand for further proceedings. I would reverse the judgment of
    the Appellate Court.
    5
    Circuit Court for Howard County
    Case No. C-13-CV-18-000181
    Argued: May 4, 2023
    IN THE SUPREME COURT
    OF MARYLAND*
    No. 30
    September Term, 2022
    ______________________________________
    KATZ, ABOSCH, WINDESHEIM,
    GERSHMAN & FREEDMAN, P.A., AND
    MARK E. RAPSON
    v.
    PARKWAY NEUROSCIENCE
    AND SPINE INSTITUTE, LLC
    ______________________________________
    Fader, C.J.
    Watts
    Hotten
    Booth
    Biran
    Gould
    Eaves,
    JJ.
    ______________________________________
    Dissenting Opinion by Watts, J.
    ______________________________________
    Filed: August 30, 2023
    *At the November 8, 2022 general election, the voters of Maryland ratified a constitutional
    amendment changing the name of the Court of Appeals of Maryland to the Supreme Court
    of Maryland. The name change took effect on December 14, 2022.
    Respectfully, I dissent. By and large, I agree with much of the Majority’s well-
    written opinion, but I disagree with its disposition of the case. I agree with the Majority
    that, in some instances, the choices made by an expert and the information that an expert
    inputs to a methodology can be part of the methodology. See Maj. Slip Op. at 3. I also
    agree that, in determining whether expert testimony has a sufficient factual basis, this Court
    should not adopt “an unduly rigid dividing line between ‘data’ and ‘methodology[.]’” Maj.
    Slip Op. at 3-4. Additionally, I agree that, in assessing the testimony of Meghan Cardell,
    an expert for Parkway Neuroscience and Spine Institute, LLC (“PNSI”), Respondent, the
    Circuit Court for Howard County abused its discretion in its “consideration of the [June
    2021] normalizing adjustments as reflecting on the reliability of Ms. Cardell’s
    methodology, as opposed to the credibility (or reliability) of Ms. Cardell herself.” Maj.
    Slip Op. at 4. I part ways with the Majority as to the outcome of the case, though.
    Rather than remand the case to the circuit court for it to decide whether to admit Ms.
    Cardell’s testimony without considering the June 2021 normalizing adjustments as
    affecting the reliability of her methodology, I would affirm the decision of the Appellate
    Court of Maryland. Like the Appellate Court, I would conclude that Ms. Cardell’s expert
    opinion was admissible and that, in assessing the reliability of her methodology, the circuit
    court abused its discretion by determining that choices of data and other inputs rendered
    her methodology unreliable. See Parkway Neuroscience & Spine Inst., LLC v. Katz,
    Abosch, Windesheim, Gershman & Freedman, P.A., 
    255 Md. App. 596
    , 630, 
    283 A.3d 753
    , 773 (2022). In my view, in assessing the reliability of Ms. Cardell’s methodology,
    the circuit court took issue with matters such as the soundness of the data that Ms. Cardell
    used and the conclusions she reached, the nature of her experience, and whether she was
    biased or not. Unlike the Majority, I would not order a limited remand regarding the
    admissibility of Ms. Cardell’s expert testimony. See Maj. Slip Op. at 50. Instead, I would
    affirm the judgment of the Appellate Court, which reversed and remanded the case for
    further proceedings, with Ms. Cardell’s testimony admissible. See Parkway, 255 Md. App.
    at 639, 283 A.3d at 778.
    Determining the amount that a business would have earned or lost in any given
    period will generally require both an in-depth analysis of the business’s revenues and
    expenses over a period of time and looking at meaningful events affecting the business,
    such as employee changes and other data, culminating in projections made through use of
    a methodology deemed reliable for calculating profit or loss. In this case, it is undisputed
    that the method for calculation of lost profit that Ms. Cardell used—the before-and-after
    method—is generally accepted “as a reliable methodology.” Maj. Slip Op. at 49 n.19.
    Here, however, in an analysis of the factors set forth in Daubert v. Merrell Dow
    Pharms., Inc., 
    509 U.S. 579
    , 593-94 (1993) and Rochkind v. Stevenson, 
    471 Md. 1
    , 35-36,
    
    236 A.3d 630
    , 650 (2020), the circuit court took issue with the quality of the data that Ms.
    Cardell relied on, not the reliability of the methodology that she used, and, in ruling, faulted
    her level of experience, and questioned her impartiality. To briefly recap the circuit court’s
    ruling, at the conclusion of the hearing on KatzAbosch’s1 motion to exclude Ms. Cardell’s
    testimony, the circuit court granted the motion, finding that PNSI did not meet the burden
    1
    Petitioners, Abosch, Windesheim, Gershman & Freedman, P.A., and Mark E.
    Rapson, together, are referred to as “KatzAbosch.”
    -2-
    of establishing that Ms. Cardell’s testimony was admissible. The circuit court summarized
    its major concerns with Ms. Cardell’s testimony before delving into the Daubert-Rochkind
    factors. A primary concern of the circuit court was an absence specialized knowledge,
    experience, or training in Ms. Cardell’s background. The circuit court stated that PNSI is
    a medical practice that required a distinct knowledge set, which Ms. Cardell lacked.
    Because of this, the circuit court questioned Ms. Cardell’s judgment in designating 2015
    as the base year for calculating lost profits, although the circuit court found that her choice
    of the before-and-after method was correct. Finding this base period to be “speculative[,]”
    the circuit court concluded that subjectivity affected Ms. Cardell’s rationale.
    The circuit court took issue with the overall quality of data used in Ms. Cardell’s
    methodology, particularly the circumstance that, in its view, “she relie[d] entirely on
    representations from within the business and the Plaintiffs[,]” i.e., PNSI. Finding her
    testimony to contain “inherent bias” due to her “support” for PNSI’s claims, the circuit
    court critiqued Ms. Cardell’s judgment calls, including her decision to make a normalizing
    adjustment to the calculations in June 2021 when there had been no change in the
    underlying facts.
    The circuit court noted what it called Ms. Cardell’s failure to examine the individual
    economic impact of each doctor’s departure, stating that “[n]ot every doctor[] is going to
    make or contribute the same amount of money each . . . year.” This contributed to the
    -3-
    circuit court’s conclusion that Ms. Cardell’s opinion would not be helpful to the jury.2
    Continuing to express concern about Ms. Cardell’s “judgment calls,” the circuit court
    faulted her failure to consider data that would have provided alternative explanations for
    lost profits, such as changes in insurance reimbursement rates. Another problem, according
    to the circuit court, was Ms. Cardell’s decision to consider member draws as reductions of
    profit, i.e., expenses, in her calculations without identifying an objective industry standard
    for such a classification.
    It was only after the circuit court set forth the above reasoning for excluding Ms.
    Cardell’s testimony that it turned to the Daubert-Rochkind factors, indicating that perhaps
    some of the factors may not apply. The circuit court quoted our statement in Rochkind,
    471 Md. at 37, 236 A.3d at 651, that “[a] trial court may apply some, all, or none of the
    factors depending on the particular expert testimony at issue” (citation omitted), then
    added: “[T]hat means that not each and every factor has to be resolved, and not each and
    every factor of the ten have to be resolved against the witness.” (Internal quotation marks
    omitted).
    In addressing the factors, the circuit court repeated much of the reasoning that it had
    already discussed. Of the 10 Daubert-Rochkind factors, the circuit court appeared to
    determine that 7 factors weighed against the admissibility of Ms. Cardell’s testimony: (1)
    2
    The circuit court reiterated this conclusion in its final holding that PNSI had not
    met its burden under Daubert and Rochkind to show Ms. Cardell’s testimony’s reliability
    and usefulness to the jury. However, the circuit court promptly limited that holding, stating
    that “usefulness is a very, very slight factor and I don’t want to dwell on it because I don’t
    want to create reversible error[.]”
    -4-
    testability, (2) the rate of error, (3) standards and controls, (4) the purpose of the opinion,
    (5) unjustifiable extrapolation from an accepted premise, (6) accounting for obvious
    alternative explanations, and (7) the ability of the field to reach reliable results for the type
    of opinion. The circuit court appeared to determine that 2 factors were neutral or not
    applicable: (1) peer review, and (2) being as careful as the expert would be in non-litigation
    work. Although the circuit court did not expressly address the factor as to general
    acceptance, the circuit court agreed that the before-and-after method was the appropriate
    procedure to use. Of the 7 factors that appeared to weigh against admissibility, 2 factors—
    the rate of error and the ability of the field to reach reliable results for the type of opinion—
    were construed against the admissibility of Ms. Cardell’s testimony based on the June 2021
    update, i.e., the normalizing adjustment, as to which the Majority concludes that the circuit
    court erred. See Maj. Slip Op. at 50.
    In discussing the other 5 factors that appeared to weigh against admissibility, the
    circuit court criticized Ms. Cardell’s choice of 2015 as the base year, her reliance on
    information from PNSI, and her failure to consider other information, such as changing
    insurance reimbursement rates, as an alternative explanation for lost profits—all of these
    pertained to the soundness of the data that Ms. Cardell used in formulating her opinion and
    the circuit court’s disagreement with her conclusions, or the court’s previous findings that
    Ms. Cardell lacked the experience necessary to accurately determine the base year for
    PNSI’s type of medical practice, under the before-and-after method, and that her testimony
    was inherently biased. To be sure, the use of data cannot always be surgically excised from
    the reliability of a methodology. That said, while “sometimes blurred[,]” Maj. Slip Op. at
    -5-
    41, there is a distinction between data and methodology. Under Maryland Rule 5-702(3),
    in determining whether expert testimony is admissible, a trial court must assess “whether
    a sufficient factual basis exists to support the expert testimony.” “‘[S]ufficient factual
    basis’ includes two sub-elements: (1) an adequate supply of data; and (2) a reliable
    methodology.” Rochkind, 471 Md. at 22, 236 A.3d at 642 (citation omitted).
    Different definitions apply to these two different elements. Among other things, for
    there to be an adequate supply of data, there cannot be “too great an analytical gap between
    the data and the opinion proffered.” Id. at 36, 236 A.3d at 651 (cleaned up). Meanwhile,
    “to constitute reliable methodology, an expert opinion must provide a sound reasoning
    process for inducing its conclusion from the factual data and must have an adequate theory
    or rational explanation of how the factual data led to the expert’s conclusion.” Id. at 32,
    236 A.3d at 648-49 (cleaned up).
    Our case law shows that there is a distinction between data and methodology. Of
    course, a given circumstance cannot always be neatly categorized as one or the other and
    might instead pertain to both. Here, however, it is clear that the circuit court’s analysis did
    not implicate the reliability of the widely accepted before-and-after method of calculating
    lost profits and exceeded a review of the soundness of Ms. Cardell’s reasoning process.
    Although the circuit court appeared to fault the soundness of the process that Ms. Cardell
    employed, the circuit court’s underlying criticism was with the data Ms. Cardell relied on,
    her background and experience, and whether she was too closely associated with PNSI.
    Although KatzAbosch contended that Ms. Cardell’s testimony was unreliable
    because she mistreated member draws and insurance reimbursement rates, which in turn
    -6-
    influenced her to incorrectly identify 2015 as the benchmark year, Ms. Cardell’s testimony
    at the hearing established an evidentiary basis for categorizing member draws as expenses
    or profits based on the nature of the business, and her research and consultation with
    another expert in damages. She then followed that approach by examining the different
    types of member draws and categorizing them based on the factual circumstances. Nothing
    in the record supports the conclusion that this was the wrong approach or that this approach
    was unreliable. Rather, the dispute arises from KatzAbosch challenging the result or
    opinion that Ms. Cardell reached.3 I agree with the Appellate Court that determining
    whether to label member draws as expenses or profits is a fact-laden decision suitable for
    cross-examination. See Parkway, 255 Md. App. at 633, 283 A.3d at 774.
    The record demonstrates that Ms. Cardell explained the “judgment calls” that she
    made in selecting 2015 as the base year. Unlike in CDW LLC v. NETech Corp., 
    906 F. Supp. 2d 815
    , 823-24 (S.D. Ind. 2012)—where the expert for the plaintiff technology
    company offered no explanation for his use of the plaintiff’s other branches in the Great
    Lakes region as a benchmark for determining how much the plaintiff’s Indianapolis branch
    would have grown but for the defendant’s alleged wrongful conduct—here, Ms. Cardell
    3
    In my view, categorizing all member draws as profits would have resulted in a
    higher estimate of lost profits—so, Ms. Cardell’s decision on this point is consistent with
    her other choices designed to arrive at reasonable but conservative estimates of PNSI’s lost
    profits. And, contrary to KatzAbosch’s contention, it is unlikely that treating all member
    draws as profits would have affected Ms. Cardell’s choice of 2015 as the benchmark,
    because such a classification would not have changed the underlying facts that she relied
    on in that decision: the unlikelihood of expenses repeating in the future, the growth of the
    business and its revenue trends, and the growth of the market.
    -7-
    explicitly described her process for concluding that 2015 was the appropriate benchmark.4
    She testified about PNSI’s revenue trends, the growth-associated costs, her market
    analysis, and the profit in 2016, despite doctors leaving, as the basis for her conclusion that
    2015 reflected a reasonable prediction for the future of PNSI absent the doctors’ departures.
    Ms. Cardell’s description of her method is not the kind of unsupported conjecture
    excluded in CDW, where the expert: (1) “offer[ed] no reason” for choosing the plaintiff’s
    “other branches as appropriately comparable to [its] Indianapolis branch[,]” for which he
    was calculating lost profits; (2) “avoid[ed] explaining how another branch’s experience in
    any year” was an appropriate point of comparison; (3) failed to establish that the market
    forces for the Indianapolis branch and other branches were comparable; (4) used the
    average of all of the other branches’ growths in revenue, even though there were “wide
    variations in branch performance from year to year” and, thus, “the cumulative growth of
    4
    The circuit court supplemented its ruling from the bench with a written order
    discussing the persuasive value of a decision from the United States District Court for the
    Southern District of Indiana, CDW, 
    906 F. Supp. 2d 815
    . The circuit court reasoned that
    the base year established by Ms. Cardell was akin to the benchmark established by the
    excluded expert in CDW, which the District Court in that case had rejected as too
    speculative because it was based on average earnings at other branches of the affected
    business that were not sufficiently comparable to the branch that allegedly suffered lost
    profits.
    In CDW, the District Court excluded the expert’s testimony on lost profits
    calculated under a different method of calculating lost profits—the yardstick method. See
    CDW, id. at 823-25. The District Court noted that the yardstick method requires use of
    benchmarks “sufficiently comparable that they may be used as accurate predictors of what
    the target would have done.” Id. at 824. However, the District Court concluded that the
    expert did not select such a benchmark when he averaged the growth rates of plaintiff’s
    other branches in the Great Lakes region without any explanation of why those branches,
    much less their average, was appropriate for comparison. See id. The District Court also
    held that other information analyzed by the expert tended to demonstrate that his analysis
    was “way off base.” Id. at 825.
    -8-
    all the included branches d[id] not approximate the growth of any individual branch”; and
    (5) analyzed, as an alternative measure of lost profits, information that tended to show that
    his predicted revenue growth was “way off base.” Id. at 824-25 (cleaned up).
    In contrast, here, Ms. Cardell explained why 2015 and not 2014 was the appropriate
    base year, described how her market research led her to incorporate market forces into her
    analysis, and indicated that PNSI’s profitability, its period of growth, and other factors led
    her to believe that the profitability in 2015 would have been replicated in the ensuing years
    but for the doctors’ departures, and no other evidence undermined her analysis. This is not
    a case of the expert arriving at a conclusion by ipse dixit—as discussed above, Ms. Cardell
    fully explained her reasoning for selecting 2015 as the base year and classifying some
    member draws as expenses based on her years of experience as a certified public accountant
    conducting similar analyses, doing market research, and consulting another expert. The
    circuit court rejected Ms. Cardell’s conclusion about 2015 being the appropriate base year
    by faulting her data, reasoning that she received too much input from PNSI, and
    questioning the nature of her experience.
    Ms. Cardell’s use of PNSI’s accounting data and consultation with its staff did not
    render her methodology unreliable. As noted by the Appellate Court, “‘[t]hese are the
    kinds of data that accountants normally rely on in calculating future earnings [and lost
    profits].’” Parkway, 255 Md. App. at 630, 283 A.3d at 772-73 (quoting Manpower, Inc. v.
    Ins. Co. of Pa., 
    732 F.3d 796
    , 809 (7th Cir. 2013)) (second alteration in original). It is
    difficult to imagine what other data might be more reliable or relevant to determining the
    trajectory of a particular business. Indeed, even in CDW, 
    906 F. Supp. 2d at 822
    , the
    -9-
    District Court was unpersuaded by the defendant’s contention that the plaintiff’s expert
    “blindly accepted as accurate, and without any verification, the figures supplied to him by”
    the plaintiff. The District Court concluded that CDW was unlike cases in which “experts
    were merely ‘parroting’ unsubstantiated views of their clients[.]” 
    Id.
     The same logic
    applies here.
    To the extent that the circuit court expressed concern about Ms. Cardell potentially
    being biased, this clearly was a matter for the jury’s consideration. As we observed in
    Falik v. Hornage, 
    413 Md. 163
    , 179, 
    991 A.2d 1234
    , 1244 (2010), “[a]n expert’s testimony
    can have a compelling effect on a jury[,]” and “[t]hat is why, especially with expert
    witnesses, wide latitude must be given a cross-examiner in exploring a witness’s bias or
    motivation in testifying[.]” (Cleaned up).
    The circuit court undeniably questioned Ms. Cardell’s “experience for this particular
    matter,” which was a consideration for the court in assessing the reliability of her
    methodology. Like the Appellate Court, I would reject the premise that “knowledge of the
    accounting principles . . . required to analyze profits for a . . . medical practice differs so
    markedly from that required for other small businesses . . . that the proffered [certified
    public accountant] must demonstrate particularized experience[.]” Parkway, 255 Md. App.
    at 622, 283 A.3d at 768 (citations omitted). In addition, consideration of an expert’s level
    of experience is not one of the Daubert-Rochkind factors adopted by this Court. Although
    the list of factors has been described as not being “exhaustive[,]” State v. Matthews, 
    479 Md. 278
    , 314, 
    277 A.3d 991
    , 1012 (2022) (citations omitted), Maryland Rule 5-702(1)
    provides that, in assessing the admissibility of expert testimony, a trial “court shall
    - 10 -
    determine[] whether the witness is qualified as an expert by knowledge, skill, experience,
    training, or education[.]” In this case, the circuit court did not do that. My concern is that
    the Daubert-Rochkind factors should not be used to exclude an expert’s testimony based
    on the expert’s experience or training without making a finding under Maryland Rule 5-
    702(1).
    Finally, although the circuit court emphasized that its conclusion that Ms. Cardell’s
    testimony would not be useful to the jury was “a very, very slight factor” in its ruling, one
    of the reasons that the circuit court expressed for finding that Ms. Cardell’s testimony
    would not be useful was that the jury might not find that all of the departing doctors left
    because of KatzAbosch’s actions, which it reasoned would render Ms. Cardell’s all-or-
    nothing lost profits estimate unhelpful. Aside from this being another criticism of the data
    that Ms. Cardell relied on, there is a fundamental problem with the circuit court’s
    conclusion that Ms. Cardell’s testimony was not helpful because the jury might not accept
    PNSI’s contention that all of the doctors left because of KatzAbosch. “[A] trial court is
    not permitted to resolve disputes of material fact in determining whether a sufficient factual
    basis exists to support an expert’s opinion. Doing so is a clear abuse of discretion.”
    Jamaiya Oglesby v. Balt. Sch. Assocs., et al., No. 26, Sept. Term, 2022, 
    2023 WL 4755689
    ,
    at *16 (Md. July 26, 2023). Under the circuit court’s logic, in Gen. Elec. Co. v. Joiner, 
    522 U.S. 136
    , 146-47 (1997), the trial court could have excluded the expert testimony, not
    because of any analytical gap between the data and the opinion, but instead because the
    jury might have found that the plaintiff never encountered the chemicals that allegedly
    contributed to his cancer.
    - 11 -
    Under the circumstances of the limited remand due solely to the circuit court’s
    consideration the 2021 normalizing adjustment, the circuit court would be free to continue
    to consider matters such as Ms. Cardell’s lack of specialized experience, whether she may
    be inherently biased, whether she relied too heavily on data from PNSI, and whether she
    relied on information that a jury might reject.5 All of these circumstances go to the weight
    to be given to an expert’s testimony, not admissibility. All of them are for the jury, not the
    circuit court, to consider. The limited remand will essentially be a redo of the motion
    hearing, giving the circuit court an opportunity to once again take these matters into
    account as the record establishes that these considerations reflect what the circuit court
    thinks of Ms. Cardell’s testimony. In determining admissibility of expert testimony, a trial
    court is “a gatekeeper,” not “an armed guard[,]” and, when expert “testimony rests upon
    good grounds, based on what is known, it should be tested by the adversary process –
    competing expert testimony and active cross-examination – rather than excluded from
    jurors’ scrutiny for fear that they will not grasp its complexities or satisfactorily weigh its
    inadequacies.’” Matthews, 479 Md. at 322-23, 277 A.3d at 1017 (quoting Ruiz-Troche v.
    Pepsi Cola of Puerto Rico Bottling Co., 
    161 F.3d 77
    , 85-86 (1st Cir. 1998)) (internal
    quotation marks omitted). In this case, the better course of action is to affirm the decision
    of the Appellate Court, which correctly reversed and remanded for further proceedings in
    5
    In my view, it was not for lack of effort or acumen that the circuit court veered off
    course by taking these considerations into account—to the contrary, this is a complicated
    case with a lot of moving parts. The record reflects that the circuit court set forth its ruling
    on the motion to exclude, carefully laid out what it genuinely perceived to be issues with
    Ms. Cardell’s testimony, turned to thoroughly addressing the Daubert-Rochkind factors,
    and attempted to render the best possible decision that it could under the circumstances.
    - 12 -
    which Ms. Cardell’s testimony would be admissible. See Parkway, 255 Md. App. at 639,
    283 A.3d at 778.
    For the above reasons, respectfully, I dissent.
    - 13 -
    

Document Info

Docket Number: 30-22

Judges: Biran

Filed Date: 8/30/2023

Precedential Status: Precedential

Modified Date: 8/30/2023

Authorities (31)

Ruiz-Troche v. Pepsi Cola of Puerto Rico Bottling Co. , 161 F.3d 77 ( 1998 )

Dodge v. Cotter Corporation , 328 F.3d 1212 ( 2003 )

Rink v. Cheminova, Inc. , 400 F.3d 1286 ( 2005 )

United States v. Ronald Keith Brown , 415 F.3d 1257 ( 2005 )

William Gasperini v. The Center for Humanities, Inc., Doing ... , 149 F.3d 137 ( 1998 )

Quiet Technology DC-8, Inc., Quiet Technology, Inc. v. ... , 326 F.3d 1333 ( 2003 )

Mathis v. Exxon Corporation , 302 F.3d 448 ( 2002 )

Manpower, Incorporated v. Insurance Company of the State , 732 F.3d 796 ( 2013 )

Bethie Pride v. Bic Corporation Societe Bic, S.A. , 218 F.3d 566 ( 2000 )

Coleman Motor Co., a Pennsylvania Corporation v. Chrysler ... , 525 F.2d 1338 ( 1975 )

Schiller & Schmidt, Incorporated v. Nordisco Corporation , 969 F.2d 410 ( 1992 )

United States v. Byron Mitchell , 365 F.3d 215 ( 2004 )

EEOC v. Freeman , 778 F.3d 463 ( 2015 )

Carmelita Elcock v. Kmart Corporation , 233 F.3d 734 ( 2000 )

Craftsmen Limousine, Inc. v. Ford Motor Co. , 363 F.3d 761 ( 2004 )

United States v. Bob Brumley , 217 F.3d 905 ( 2000 )

Fred Lauzon v. Senco Products, Inc. , 270 F.3d 681 ( 2001 )

John Haugh v. Jones & Laughlin Steel Corporation , 949 F.2d 914 ( 1991 )

C.W. & E.W. v. Textron, Inc. , 807 F.3d 827 ( 2015 )

Greggar S. Isaksen D/B/A Applewood Stove Works v. Vermont ... , 825 F.2d 1158 ( 1987 )

View All Authorities »