top of page

Financial Data Test: Only ChatGPT Passes as Perplexity and Gemini Flop

One of the promises of AI is that it can take the grunt work off our plates. As a finance professional, I wanted to test that promise by asking AI tools to gather basic financial data from public companies—something I do often when preparing projections or conducting quick benchmarks.


I assumed this would be an easy, after all, this data is public, structured, and relatively unambiguous. My Assumptions were wrong.


Step 1: Perplexity Left Me… Perplexed

I started with Perplexity, expecting it to be the fastest way to gather public company financials. I typed a basic prompt:

“Can you put together for me in Excel the quarterly financial performance of a certain public company?”

To my surprise, Perplexity immediately chose Intel—without me ever specifying a company—and gave me a pre-filled answer regarding Q2 2025 with a comparison to Q2 2024. It conveniently offered an excel file to download (only to let me know that I can't download it in this "environment" perhaps because I was using the free version).


perplexity response to a request for quarterly financial statements of a certain company, instead it decided I meant Intel

So I got specific:

“Please pull P&L data for Credo Technology Group Holding Ltd (Nasdaq: CRDO) going back 8 quarters. Please put it in a table.”

Perplexity’s results were perplexing . It only managed to extract Revenue and EPS (and one quarter of Net Income). The revenue figure was technically accurate for the most recent quarter, but it mislabeled the period as “Q4 2024” instead of “Quarter ended April 30, 2025.” EPS figures were just… off. And there was no explanation as to where they came from.


It’s worth noting: Credo has a fiscal year ending in April, which may have confused the model. But that doesn’t explain the lack of effort on the rest of the financials.


perplexity AI failing a basic data request for financial results of Credo Technology group CRDO)

Thinking perhaps Perplexity struggled because CRDO isn’t as well-known, I gave it another shot with Intel—its unsolicited favorite.

“Pull P&L data for Intel Corp going back 8 quarters. Please put it in a table.”

This time it populated all 8 quarters. But once I fact-checked the numbers against Yahoo Finance, I found clear discrepancies. Some quarters were accurate. Some were totally off. Perplexity said it used “rounded historical data” — but these weren’t rounding errors. The actual data is freely and easily available.


Perplexity provides inaccurate financial data based on basic data gathering prompt.

Intel Corp. P&L for the three months ended Sep 28 2024
Taken from Intel's financial reports as reported to the SEC

The miss was really bad:

Quarter

Metric

Perplexity

Actual

% Difference

Q3 2024

Net Income (B)

-0.61

-16.9

+2670%

Q3 2024

EPS

-0.38

-3.88

+921%


Step 2: Gemini Tries (and Mostly Fails)


Next I tried Google’s Gemini with the same exact question I did for perplexity: “Can you put together for me in Excel the quarterly financial performance of a certain public company?” To its credit, Gemini noticed that I hadn’t specified a company and asked for clarification so I asked it to provide me with 8 quarters of financial data for Credo.


Google Gemini fails to provide accurate data on a simple data gathering prompt to assemble quarterly data from Credo a public company.
The quarters indicated in red are wrong.

Gemini quickly produced a financial table (above) but unfortunately with inaccurate historical data forcing me to go to the SEC’s EDGAR system to manually fact-check it as if AI never existed.


Actual SEC reported quarterly P&L results for CRDO:

Quarter Ended

Total Revenue

Gross Profit

Quarter Ended

Total Revenue

Gross Profit

Operating Income (Loss)

Net Income (Loss)

Diluted EPS

Operating Income (Loss)

Net Income (Loss)

3-May-25

            170,025

            114,188

3-May-25

            170,025

            114,188

               33,788

               36,588

                    0.20

               33,788

               36,588

1-Feb-25

            135,002

               85,926

1-Feb-25

            135,002

               85,926

               26,194

               29,360

                    0.16

               26,194

               29,360

2-Nov-24

               72,034

               45,512

2-Nov-24

               72,034

               45,512

                 4,474

               (4,225)

                  (0.03)

                 4,474

               (4,225)

3-Aug-24

               59,714

               37,283

3-Aug-24

               59,714

               37,283

                 5,533

               (9,540)

                    0.06

                 5,533

               (9,540)

27-Apr-24

               60,782

               39,966

27-Apr-24

               60,782

               39,966

               (7,881)

             (10,477)

                  (0.06)

               (7,881)

             (10,477)

27-Jan-24

               53,058

               32,558

27-Jan-24

               53,058

               32,558

               (5,911)

                     428

                         -  

               (5,911)

                     428

28-Oct-23

               44,035

               26,117

28-Oct-23

               44,035

               26,117

               (8,875)

               (6,623)

                  (0.04)

               (8,875)

               (6,623)



The most recent two quarters were accurate, but 4 out of the 6 older ones were far off. So I tried to help Gemini fix itself by showing it it's wrong.


Google gemini cannot find the financial statements in a PDF of financial statements.

I provided a link to a 10-Q filing. It wouldn’t read it. I uploaded a PDF version. It claimed it didn’t contain financials. Eventually, I spoon-fed the financial data by copying and pasting it into the chat —and only then did it acknowledge its mistake.


Google Gemini finally corrects its previously wrong numbers it hallucinated


In short: the AI hallucinated confidently until cornered with evidence.


Step 3: ChatGPT Gets It Mostly Right


Finally, I tested ChatGPT. I uploaded the same requests. ChatGPT parsed the structure properly and produced a table that was far more accurate than either Perplexity or Gemini. The only consistent issue was how it labeled the quarter-end dates.


For example, ChatGPT marked a quarter ending November 2, 2024 as “Oct 2024.” Not a dealbreaker—but a reminder that even with helpful AI, the output needs review.


Observations and Takeaways

This was supposed to be an easy test. Public financials. Simple formatting. No interpretation required.


Instead, I got:

  • One tool (Perplexity) that mixed accuracy with invention

  • Another (Gemini) that confidently misreported and refused to learn

  • And only one (ChatGPT) that came close to usable output


Even more curious: the more recent quarters were usually accurate across the board. Older quarters? That’s where hallucinations crept in. It’s as if the AIs were more likely to guess than to search.


This isn’t just an annoyance—it’s part of a broader trend known as AI hallucinations  and alignment faking. These are moments when a model would rather give a plausible (but wrong) answer rather than research for the correct answer or admit it doesn’t know.


Final Verdict

For financial data gathering, ChatGPT was the clear winner—but even it needs human review.

This exercise reminded me of something important: AI is not Google. It’s a language prediction machine. Sometimes it knows; sometimes it confidently fakes it.

If you’re a finance professional looking to use AI for data gathering or projections, don’t blindly trust the output. Ask yourself:

  • Where did this data come from?

  • Does it match official filings?

  • Is the formatting actually usable?

Until these tools get better at citing reliable sources and avoiding hallucinations, they’ll be useful aides, not replacements.

bottom of page