
AI In Tax: Using LLMs To Work Smarter With Spreadsheets
Despite the growing presence of AI and large language models (LLMs) within tax departments, spreadsheets continue to play a central role in the daily work of tax professionals. While tax departments are embracing digital transformation, many interim processes—like extracting data from ERP systems or reconciling values—still flow through spreadsheets. And despite the promise of intelligent tax tools, this reality isn't going away anytime soon.
But with great reliance comes significant risk. High-profile cases have shown that spreadsheet errors can cost companies millions. A comprehensive academic study found that a staggering 94% of financial spreadsheets contain errors. These aren't just innocent typos but often result in compliance breaches, miscalculations, and flawed reporting.
LLMs may offer real help in managing spreadsheets. To use them effectively and responsibly, tax professionals must understand both what these AI tools can do and where they fall short.
How LLMs Process Spreadsheets
While most LLMs can read common spreadsheet formats like CSV or Excel and answer questions about the data, they differ in two important ways: how much data they can handle at once (known as the context window) and how they process the data.
GPT‑4o has a context window of 128,000 tokens, which limits how much information it can process in a single interaction. When you upload a spreadsheet to ChatGPT powered by GPT‑4o, the model doesn't read the file directly. Instead, it uploads the file to a secure, temporary environment that includes tools like Python and data science libraries. In this setup, GPT‑4o behaves like a Python programmer: it writes and runs code to explore your spreadsheet. It then turns the results of that code into clear, human-readable explanations. If you ask for a chart, GPT‑4o generates the code to create it and shows you the result.
Claude 3.5 Sonnet takes a different approach. It reads spreadsheet content directly as text, interpreting headers, rows, and columns without writing or running code. It currently doesn't support chart generation or code execution, but it has a much larger context window—up to 200,000 tokens—which allows it to handle larger datasets in a single session and generate longer, more detailed responses without losing earlier information.
Based on their characteristics, GPT‑4o may be the better choice for tasks that involve complex data manipulation, calculations, or visualizations. Claude, on the other hand, is excellent for exploring and interpreting large, text-based tables, identifying patterns, and summarizing structured data, especially when working with large volumes of content that don't require advanced computation.
But What About Limitations?
LLMs have some limitations when working with spreadsheets, and the most significant hurdle is context window constraints. Think of an LLM's context window as its short-term memory or the amount of information that can be processed in a single interaction. This information is measured in tokens, which are not the same as words. A token typically represents a few characters or parts of words. For example, 1,000 tokens is roughly equivalent to 750 words of English text.
Each LLM has a different context window size. GPT‑4o, for instance, has a context window of 128,000 tokens. Now consider a large spreadsheet with 10 columns and 100,000 rows—that's 1 million cells. If we estimate an average of 3 tokens per cell, the total token count would be around 3 million tokens, which far exceeds the capacity of any current model, including GPT‑4o.
Even uploading a portion of such a file can push the model beyond its limit. For example, 10 columns × 20,000 rows equals 200,000 cells. At 3 tokens per cell, that's approximately 600,000 tokens, not even counting the extra tokens needed for headers, formatting, or file structure. Since GPT‑4o can only process 128,000 tokens at once, only a small fraction of that spreadsheet can be 'seen' and processed at any given time.
When you upload a spreadsheet to GPT‑4o, the model can only interact with the data that fits within the active context window. It doesn't see the entire file all at once but just the portion that fits within that token limit. For example, if you ask, 'What is the deductible VAT amount listed in row 7,000?' but the model only received the first 5,000 rows, it won't be able to answer because it never saw that row in the first place.
It's also important to understand that the context window includes the entire conversation, not just your current question and the data. As the session continues and more prompts and responses are exchanged, the model may start dropping earlier parts of the conversation to stay within the 128,000-token limit. That means key data, such as the original file content, can be silently dropped as the conversation grows. This can lead to incomplete or incorrect answers, especially when your new question relies on information the model has already "forgotten."
Another limitation is that LLMs are sequence-based models. They read spreadsheets as a linear stream of text and not as a structured, two-dimensional grid. That means they can misinterpret structural relationships and cross-sheet references between cells. LLMs don't automatically recognize that cell D20 contains a formula like =SUM(A20:C20). Similarly, they may not realize that a chart on "Sheet1" is pulling data from a table on "Sheet2,' unless this relationship is clearly described in the prompt.
Finally, LLMs don't truly 'understand' tax law. While they've been trained on large volumes of publicly available tax-related content, they lack the deeper legal reasoning and jurisdiction-specific knowledge that professionals rely on. They can easily make obvious mistakes like not flagging penalties or entertainment expenses as not eligible for input VAT deduction because they are not aware of country-specific rules, unless such rules are explicitly stated in the prompt. As a result, they can produce plausible but incorrect answers if relied on without expert review.
How to Use LLMs Effectively with Spreadsheets
When using LLMs to work with spreadsheets, you'll get the best results by running them within platforms designed for data tasks, such as Python notebooks, Excel plugins, or Copilot-style interfaces. These tools allow the LLM to interact with your spreadsheet by generating Excel formulas or Python code based on your instructions. For example, you might say: 'Write a formula to pull client names from Sheet2 where the VAT IDs match those names." The tool then generates the appropriate formula, and the spreadsheet executes it just like any standard formula.
When dealing with large spreadsheets, another effective strategy is to break the data into smaller, manageable sections and ask the model to analyze each part separately. This approach helps keep the information within the model's memory limits. Once you've gathered insights from each section, you can combine them manually or with the help of a follow-up AI prompt.
Another powerful method is to ask the LLM to write code to process your spreadsheet. You can then run that code in a separate environment (like a Jupyter notebook), and feed just the summarized results back into the model. This allows the LLM to focus on interpreting the findings, generating explanations, or drafting summaries without being overwhelmed by the raw data.
Spreadsheets Are Here to Stay
Spreadsheets aren't going anywhere. They are too flexible, too accessible, and too deeply ingrained in tax operations to disappear. AI and LLMs will continue to transform the way we work with them, but they won't replace them.
Looking ahead, we can expect smarter tools that make spreadsheets more AI-friendly. Innovations like TableLLM and SheetCompressor are paving the way. Though still in the research phase and not yet integrated into mainstream commercial tools, they signal a promising future. TableLLM is a specialized language model trained specifically to understand and reason over tabular data. Unlike general-purpose LLMs that treat tables as plain text, TableLLM recognizes the two-dimensional structure of rows, columns, and cell relationships. SheetCompressor, developed as part of Microsoft's SpreadsheetLLM project, uses AI-driven summarization techniques to drastically reduce spreadsheet size before passing the data to an LLM. It results in up to 90% fewer tokens, while preserving the original structure and key insights.
Beyond TableLLM and SheetCompressor, the field of spreadsheet-focused AI is expanding rapidly. Experimental tools like SheetMind, SheetAgent, and TableTalk explore everything from conversational spreadsheet editing to autonomous multi-step operations. As these technologies mature, AI-powered tax departments won't move away from spreadsheets but will use them in smarter, faster, and more efficient ways.
The opinions expressed in this article are those of the author and do not necessarily reflect the views of any organizations with which the author is affiliated.

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles

Washington Post
31 minutes ago
- Washington Post
Defense Department to begin using Grok, Musk's controversial AI Model
The U.S. Department of Defense will begin using Grok, the AI chatbot built by Elon Musk's start-up xAI, the company announced in a post on Monday. The announcement came as Grok unveiled what it called 'Grok for Government,' a suite that allows agencies and federal offices to adopt its chatbots for their specific uses. President Donald Trump has encouraged more rapid adoption of artificial intelligence tools since taking office in January. Elon Musk was a member of the Trump administration, overseeing the U.S. DOGE Service, until he departed in late May. He has since become a critic of Trump's signature tax and spending legislation. On Monday, xAI said its products would now be 'available to purchase via the General Services Administration (GSA) schedule,' allowing 'every federal government department, agency, or office' to buy them. Grok came under fire last week after launching into an antisemitic rant and invoking Adolf Hitler after it was a programmed to be less politically correct. The incident prompted the company to say it would improve its model. A day later, xAI unveiled a sweeping update that it claimed put Grok on the cutting edge of AI development. The Defense Department did not immediately respond to a request for comment. This is a developing story and will be updated.


Fast Company
34 minutes ago
- Fast Company
How to build a 'superhuman workplace' where AI amplifies human potential
In his compelling keynote address at Fast Company's recent New World of Work Summit in partnership with HP, renowned futurist Bernard Marr challenges organizations to stop simply bolting AI onto existing processes and instead fundamentally reimagine work with humans at the center. Drawing from real-world examples across medicine, Formula One, and executive decision-making, Marr demonstrates how AI agents and generative tools can amplify uniquely human capabilities like creativity and emotional intelligence—creating what he calls the 'superhuman workplace.' With research 1 showing 85% of knowledge workers feel fulfilled when they have adequate technical support, his insights offer a roadmap for leaders navigating what he describes as an unprecedented 'hyper-evolution cycle' in business and technology. HP conducted an online survey in partnership with Edelman Data X Intelligence that fielded between February 19 and March 7, 2025 among N=1000 US knowledge workers to explore the drivers of fulfillment at work. ↩︎
Yahoo
39 minutes ago
- Yahoo
ThroughPut and Aankhen launch breakthrough SKU-level tariff management decision acceleration product
Leveraging CFO-approved data for proactive & targeted management of global & local material strategy PALO ALTO, Calif., July 14, 2025 /PRNewswire/ -- the supply chain decision intelligence pioneer, today announced a strategic partnership with Aankhen, a financial supply chain software platform integrating procurement, logistics, and finance costs through digital twin technology, to deliver the most advanced Tariffs financial mitigation solution and bridge the gap between operational efficiency and financial liquidity across global supply chains. The joint offering combines ThroughPut's real-time AI-driven supply chain decision intelligence and inventory optimization capabilities with Aankhen's fiChains™ financial supply chain digital twin technology to provide actionability into the flow of capital across complex supply chains and establish a sustainable and continuously scalable global solution to address Tariffs uncertainty. By integrating these powerful solutions, Aankhen's fiChains™ solution will provide financial costs at a SKU level to ThroughPut to optimize inventory, pricing, and profitability and ThroughPut's Decision Intelligence Platform will empower organizations with strategic and holistic AI-powered, prioritized recommendations to unlock working capital, increase procurement agility, and reduce operational risk - all without increasing complexity or manual intervention. "Since the tariffs started, ThroughPut's customers have been asking for an advanced tariff management solution that provides the best "new price" estimate of an item," said Ali Raza, CEO and Founder of "A lot of companies need to re-examine every item that they sell and source, especially whether it makes sense to stress-test pricing or eliminate the offering altogether. After searching the entire market, we realized Aankhen was the only company that had these capabilities. Current ThroughPut users can leverage Aankhen's fiChains offering to enable proactiveness and accelerate reactiveness. Aankhen CEO Subhash Chowdary has a track record of transforming supply chains for companies like Nestle, Lockheed, Dell, and other leading supply chains. We have started our Go-To-Market process to get this joint-offering to the right decision makers. If you are a CFO who wants to focus on the right items and get the numbers right, we can make an impact before the August 1st deadline". A New Paradigm for Operational and Financial Synergy As businesses contend with tighter margins, volatile demand, and fragile supplier ecosystems, supply chain leaders are increasingly seeking intelligent solutions that offer not just visibility, but actionable foresight and seamless automation capabilities. This partnership delivers on that demand by integrating: AI-powered demand sensing, forecasting, and inventory rebalancing Financial supply chain visibility and automation across procurement, logistics, and finance Automation of cross-functional processes to reduce manual effort and improve decision-making agility Gain in-context visibility to financial facts across your Physical Supply Chain Improve Agility & Resilience with forward looking visibility Plan with forward looking financial visibility Ensure planned financial outcomes with certainty Automate financial cross-functional dependencies across Finance, Procurement, Logistics, Product Engineering, Suppliers, Carriers to mitigate Tariffs Accelerate time-to-decision, time-to-change, time-to-implement This unified approach is designed to eliminate silos, reduce procurement cycle times, and de-risk supplier relationships – while preserving cash flow and meeting fulfillment targets more reliably. Designed for Impact, Built for Scale The integrated solution is especially suited for capital-intensive sectors like automotive, aerospace, industrials, energy, and pharmaceuticals – industries where has historically helped customers reduce delays, rebalance inventories, improve supplier responsiveness, and boost profitability. Customers leveraging ThroughPut's AI capabilities have seen measurable improvements in operational throughput, supply-demand alignment, and working capital optimization. The partnership with fiChains enhances these outcomes by synchronizing financial flows with physical supply chains and automating decision-critical processes. "At Aankhen, we believe the future of procurement is autonomous, intelligent, and trust-driven," said Subhash Chowdary, Founder and CEO of fiChains. "ThroughPut brings operational foresight. We bring the execution muscle. Together, we're enabling enterprises to act faster – and with confidence – across sourcing, logistics, finance, and supply chain resilience." The partnership underscores both companies' shared mission to help customers navigate complex supply chain realities with purpose-built technologies that deliver results – without costly rip-and-replace implementations or long onboarding cycles, a hallmark of ThroughPut's rapid time-to-value approach. The integrated ThroughPut–fiChains solution is available globally and deployable via major cloud environments. Interested enterprises can request early access or schedule a demo at and Media Contact: Subhash@ About ThroughPut Inc. is a Silicon Valley-based Supply Chain AI-powered Decision Intelligence leader that puts Industrial material flow on Autopilot by leveraging existing Enterprise Data to achieve superior Business, Operations, Financial and Sustainability Results. ThroughPut improves material flow and free-cash-flow across the entire end-to-end value chain far faster than leading contemporary and legacy solutions. The founding team is led by seasoned serial entrepreneurs with real-world AI, Supply Chain, Manufacturing, Transportation and Operational experience, from the shop-floor to the top-floor, at leading Fortune 500 Industrial Companies & pioneering Enterprise Technology companies. About Aankhen Inc. Aankhen is a Silicon Valley-based Financial Supply Chain Visibility solutions provider. Aankhen's Financial Supply Chain innovation, fiChains™ mitigates Tariffs financial impact. fiChains™ uses pre-trained Digital Assistants (DAs) to automate cross-functional collaboration required across Finance, Planning, Procurement, Logistics, Product Engineering, Suppliers and Carriers to compute financial facts to support and implement fact-based executive and operational decisions to mitigate Tariffs impact. Find out more at or View original content to download multimedia: SOURCE ThroughPut Inc. Error in retrieving data Sign in to access your portfolio Error in retrieving data Error in retrieving data Error in retrieving data Error in retrieving data