AI Limitations: Why ChatGPT Falls Short for Investment Research

Created On:
May 25, 2025

Introduction to ChatGPT's Popularity

ChatGPT experienced one of the fastest consumer adoption curves in tech history, reaching 400 million users by March 2025. Its intuitive interface and broad capabilities made it a go-to tool for students, professionals, and everyday users. It’s now widely used for general research, writing assistance, coding, and summarizing complex topics. With continuous updates and integrations, ChatGPT has become embedded in daily workflows across industries.

The application of AI in financial research sees ChatGPT being increasingly used to summarize SEC filings and earnings calls transcripts. This promises significant time savings from hours of reading dense 100-page 10-Ks or skimming through 4 or 5 transcripts. Analysts and investors rely on it to expedite diligence and generate informed investment questions.

Trust but verify.

The allure of leveraging Generative AI, such as ChatGPT, for investment research is undeniable, promising unprecedented speed and data synthesis. However, a prudent approach of 'trust but verify' is paramount, as several inherent limitations can lead to significant analytical pitfalls.

Critical Limitations: Why ChatGPT Falls Short for Investment Research

There are four major problems with using ChatGPT for investing research:

Reliance on Unreliable Secondary Sources vs. Primary Data

The quality of your output is directly related to the quality of your inputs or sources. We often say, “garbage in, garbage out.”

Pulling sources from Investing.com, MSNBC, and a random Seeking Alpha article. All of these articles are written by journalists or freelancers for clicks and ads, rather than being helpful to investors or traders. These articles are also known for their high error rates, as well as misleading or biased narratives.

Crucially, ChatGPT lacks direct, real-time access to essential primary sources:

These primary sources are the foundation for all research, as these documents are legally mandated and audited by multiple parties.

General-Purpose Design & "Creativity" Leading to Hallucinations

ChatGPT is a consumer-grade, general-purpose AI wrapper built on top of OpenAI’s large language model; it’s a software application that leverages the OpenAI LLM.

Being both general-purpose and focused on consumers, ChatGPT must cater to a wide range of users, including programmers, copywriters, individuals seeking therapy, language learners, image generators, travelers planning itineraries, and hobbyist chefs searching for new recipes, among others.

A lot of those use cases require ChatGPT to be creative. By allowing ChatGPT to be creative, it opens up the possibility & increases the probability of hallucinations (i.e., taking facts and making fiction). Sadly, as investors, we cannot invest in fiction.

This is a first-principles issue with using ChatGPT: creative work and highly accurate factual work inherently conflict.

When you build any AI wrapper, you have to set the “degree of randomness” parameters: temperature, top-K, and top-P. If those are set too “tight,” then ChatGPT can’t write excellent copy, tell you a fantastic recipe that works with the 10 ingredients in your fridge, or generate cartoon-like images, etc.

So, ChatGPT has set the parameters to be reasonably creative. However, reasonably creative means a certain level of inaccuracies from time to time.

ChatGPT for Financial Research

Inability to Execute Specialized Financial Workflows & Calculations

Investment research demands specific, often nuanced, workflows that general-purpose AI like ChatGPT isn't equipped to handle reliably:

  • Accurate Entity Recognition: These workflows may seem common sense to a human investor, but they can be complicated for ChatGPT. A straightforward workflow involves identifying the correct company or ticker being referenced in the user query (see an interesting research paper on this issue).
  • Financial Arithmetic: One of the most essential workflows for an investor is simple arithmetic. An example is calculating the missing 4 quarter by subtracting quarterly numbers from the annual number or calculating the margins or year-over-year change between specific metrics.

    The most critical workflow that ChatGPT gets incorrect is arithmetic. As a reminder, LLMs are text generation models that create text by predicting the next token, using a statistical understanding of language. This process involves analyzing a sequence of input tokens and generating a probability distribution over all possible following tokens, then selecting a token based on this distribution. If the model has not seen 'a $1.897 billion revenue + a $2.319 billion revenue', there’s a good chance that the arithmetic result is incorrect.

    To solve this problem, ChatGPT would need access to functions or tools for numerical storage and financial calculations, including arithmetic, ratios, and year-over-year or quarter-over-year calculations.
  • Contextual Financial Acumen: Investing isn't just about numbers; it's about understanding what those numbers mean in context (identifying non-recurring items, comparing metrics across companies, understanding industry-specific KPIs). ChatGPT lacks this deep contextual financial acumen.

Constrained Context Window & Fragmented Document Analysis

ChatGPT’s LLM models have a context window of 128,000 to 200,000 tokens. With a bit of math, this translates to approximately 210–260 pages of written text (equivalent to two 10-Ks) that can be inserted into the chat input box at a time. Therefore, there is a limit to the amount of information ChatGPT can analyze at one time. This does not even include the prompt instructions that need to be passed to the model. Fundamental analysis is not just a matter of two documents and done.

An alternative method is to upload a few 10-Ks into ChatGPT.

  • It will then slice each PDF into small chunks (approximately 500-1000 words each), convert each chunk to a vector embedding, and store those chunks in a temporary database.
  • When you ask a question, the model will try to find only the chunks with semantic similarity to your question.
  • These similar chunks will be fed into the LLM model.


This method introduces significant analytical risks:

  • Loss of Narrative Cohesion: The overall story and the interconnectedness of ideas across different sections of a lengthy document can be lost when it is divided into chunks (i.e., the model found chunk #5 but did not include the background from chunk #4).
  • Incomplete Information Retrieval: This chunking method only surfaces bits of info that are similar to the question that will be used as data, but misses vital context from unretrieved preceding or succeeding chunks.
  • Difficulty Surfacing Contrarian Data: Chunking based on query similarity may fail to identify opposing viewpoints or critical data points necessary for a balanced analysis.
  • Impaired Cross-Document Analysis: Additionally, the model will have limitations in understanding connections between topics or companies unless these connections are explicitly specified.


And if the uploaded documents do not contain the data, then your analysis is out of luck.

Bridging the Gap: How can Enlo AI help your investing workflow?

We understand that your time is valuable, and in a world drowning in data, it’s essential to surface accurate and actionable insights.


Unlike consumer-grade, general-purpose AI that requires you to provide context, Enlo AI operates on a vast, curated repository of a decade's worth of SEC filings, transcripts, presentations, and historical financials for over 1,300 top companies tradable on US stock exchanges. This means you're not just summarizing isolated PDFs; you're querying an interconnected financial knowledge base.


We have parsed, organized, and pre-processed the filings into tens of millions of data points; sometimes transforming presentation images into actionable insights. You gain insights grounded exclusively in primary source data; no need to sift through blogs or forums.


Our system doesn't just use one LLM; we employ an average of seven specialized models per chat
, each with parameters tuned for factual accuracy. Each step of our 'Enlo Algorithm' is optimized for financial analysis, from reasoning to double verification and arithmetic checks.


Crucially, every datapoint is cited directly to its source document, giving you instant verification and the ability to dive deeper with confidence – a level of transparency and reliability essential for professional decision-making.


Ready to experience the future of financial research? Learn more about Enlo AI.

Don Tran
Authored by:
Don Tran

Contact Us

You can reach us at admin@enlo.ai

Powerful Tools. Clear Signals. Smarter Decisions.