Back to Archive

LA-AI Insights: How AI Reads (and Why It Stops)

Your weekly AI news and updates from Lower Alabama

Monday, October 27, 2025

Share this newsletter:


When you send text to an AI, it doesn't read it the way you do. It doesn't see "hello" as a word. It breaks everything down into small, digestible pieces called tokens. Think of it like Scrabble tiles. Your sentence isn't one big chunk. It's dozens or sometimes hundreds of individual tiles the AI has to work through.

So where does the word even come from? It matters more than you'd think. Linguists borrowed the term decades ago when they started breaking language into pieces. Computer scientists took it from there. When you break text into the smallest meaningful units a computer can understand, what do you call them? Tokens. A piece that represents the whole in the same way a poker chip represents money or a subway token represents one ride.

In AI, tokens are how the model actually thinks. Your text doesn't stay as text. It gets converted into numerical representations, and those representations are tokens. (Want to understand what happens to those numerical representations? We dive deeper into embeddings and vectors in How AI Learns.) Different models tokenize—convert text into numerical values—differently. The same sentence might be 10 tokens in Claude and 15 in ChatGPT. It's not random. The tokenizer was trained on data to find the most efficient chunks, so the model figures out which combinations appear most often and treats those as single tokens.

If you're using a subscription like ChatGPT Plus, you're not directly paying per token. But token efficiency still matters because it affects how well the AI can actually help you in a single response. For people using APIs to build something, though? Tokens become your most important metric.

This is why shorter prompts aren't always better. I know that sounds backwards. But a vague 50-token prompt that requires three follow-ups actually costs more than a clear 100-token prompt that nails it the first time. The AI understands you better, generates a better answer, no regenerations needed.

You're typing into ChatGPT and suddenly it just stops. Dead. You're sitting there refreshing, wondering if something broke. What actually happened? You hit your context window limit.

Your context window is your model's working memory. The latest models have dramatically larger windows. GPT-5 handles 400,000 tokens, while Gemini 2.5 Pro and Grok 4 can work with 2 million tokens. Older versions like GPT-3.5 still use 16,000, so it depends on which model you're using. A single page of text is roughly 350 to 500 tokens. That 40-page research paper you want to analyze? Ten thousand tokens gone like that. Once you're near the limit, earlier parts of the conversation fall out of context. The model can't reference them anymore. It's not a glitch. It's the model hitting its actual ceiling.

Here's a quick reference of the major models and their context windows as of October 2025:

Model Name

Provider

Max Context Window

Release Date

GPT-5

OpenAI

400,000 tokens

Aug 7, 2025

Gemini 2.5 Pro

Google (DeepMind)

2,000,000 tokens

June 17, 2025

Grok 4 Fast

xAI

2,000,000 tokens

Sep 2025

Claude 3 Opus

Anthropic

200,000 tokens

Mar 4, 2024

Sonar Pro

Perplexity AI

200,000 tokens

Jan 21, 2025

GPT-4 Turbo

OpenAI

128,000 tokens

Nov 2023

GPT-3.5

OpenAI

16,000 tokens

Mar 2023

So, how do you work with a context window? Structure your inputs deliberately. Think about what the AI actually needs, not what sounds impressive. Break large documents into sections. Regenerate prompts if they're not working instead of just pushing through.

Tokens are how AI reads. Your context window is how much it can hold. Understanding that changes everything about how you actually work with these tools—especially if you're building something that has to scale.

**As a reminder, all newsletter content is archived on the LA-AI Blog page



Upcoming Events

We're shifting our Mobile meetup to 11:30 AM – 1:00 PM this Friday due to Halloween—earlier start, free lunch, and you'll actually have your afternoon back and be home in time for trick-or-treating. Come hungry, leave smarter. @ Innovation Portal. See you there. 📆 Friday, October 31st ⏰ 11:30AM-1PM 📍Innovation Portal, 358 St. Louis St., Mobile, AL

Invalid Date

Innovation Portal. 358 St. Louis St. Mobile, AL

It's the first Friday of the month, which means it's time for the LA-AI Fairhope meetup. Join us for a fun afternoon where we will have a very special guest.

Invalid Date

314 Magnolia Ave., Fairhope, AL

This Week in AI

1. Anthropic Introduces Skills for Custom Claude Tasks

Anthropic has launched Skills, allowing users to create custom task-specific capabilities for Claude that can be trained and deployed for specialized workflows. This modular approach enables organizations to build tailored AI solutions without extensive technical expertise, democratizing AI customization for enterprise applications. The development signals a shift toward configurable AI systems where businesses can adapt general-purpose models to specific operational needs, potentially accelerating AI adoption across diverse industry verticals and use cases.

2. With new acquisition, OpenAI signals plans to integrate deeper into the OS

OpenAI's acquisition of the team behind Apple's Shortcuts automation app represents a strategic pivot toward operating system-level AI integration. This move signals OpenAI's intention to embed AI capabilities directly into core computing workflows rather than remaining a standalone application. The acquisition provides critical automation expertise that could enable AI agents to perform complex, multi-step tasks across native OS functions. This development positions OpenAI to compete more directly with Apple and Microsoft's integrated AI strategies, fundamentally changing how users interact with their devices through intelligent automation.

Ars TechnicaRead more

3. Anthropic’s Claude chatbot is getting a ‘memory’ upgrade

Anthropic's introduction of persistent memory capabilities to Claude represents a fundamental advancement in AI assistant functionality, enabling context retention across conversations and sessions. This feature transforms Claude from a stateless interaction model to a personalized AI companion that learns and adapts to individual user preferences and workflows. The memory system allows for more sophisticated long-term reasoning and relationship building between users and AI systems. This development signals the maturation of conversational AI from reactive tools to proactive, context-aware assistants that can maintain meaningful continuity over time.

The VergeRead more

4. Lawsuit: Reddit caught Perplexity “red-handed” stealing data from Google results

Reddit's lawsuit against Perplexity for allegedly scraping data from Google search results highlights critical tensions around AI training data ownership and access rights. The case involves sophisticated detection methods where Reddit embedded tracking markers to catch unauthorized data harvesting. This legal action could establish precedent for how AI companies can access and utilize web content for model training and real-time responses. The outcome will significantly impact the economics of AI development, potentially forcing companies to negotiate formal licensing agreements for training data and fundamentally reshaping the data ecosystem powering modern AI systems.

Ars TechnicaRead more

5. Google’s Quantum Computer Makes a Big Technical Leap

Google announces a significant quantum computing breakthrough, advancing capabilities that could eventually revolutionize AI model training and optimization. The development represents progress toward quantum-classical hybrid systems that might solve complex optimization problems currently limiting AI scalability and efficiency. While practical applications remain years away, this advancement positions Google strategically for the convergence of quantum and AI technologies, potentially providing unprecedented computational advantages for machine learning workloads and reinforcing Google's position as a fundamental technology infrastructure provider.

New York TimesRead more

6. OpenAI Unveils Atlas Web Browser Built to Work Closely With ChatGPT

OpenAI launches Atlas, an AI-native web browser that integrates ChatGPT directly into web browsing experiences, marking a strategic challenge to Google's Chrome dominance. The browser enables conversational interaction with web content, automated task completion, and seamless AI assistance across online activities. This represents OpenAI's expansion beyond chat interfaces into fundamental internet infrastructure, potentially reshaping how users discover, process, and interact with web-based information while establishing new competitive dynamics in the browser market.

New York TimesRead more

7. Real-time Audio Deepfakes Have Arrived

IEEE Spectrum reports breakthrough achievement in real-time audio deepfake generation, enabling live voice impersonation during phone conversations with minimal latency. The technology utilizes advanced neural voice synthesis and real-time processing capabilities that can clone voices from brief audio samples within seconds. This development fundamentally alters cybersecurity threat landscapes, particularly for voice-based authentication systems and social engineering attacks, while raising critical questions about audio evidence reliability in legal and business contexts that organizations must immediately address.

IEEE SpectrumRead more

8. Creating AI that matters

MIT researchers present comprehensive framework for developing AI systems that address fundamental societal challenges rather than incremental commercial applications. The research emphasizes methodologies for ensuring AI development focuses on high-impact problems including healthcare accessibility, climate change mitigation, and educational equity. The framework provides strategic guidance for research institutions and technology companies to prioritize AI projects with measurable societal benefits while maintaining technical excellence and commercial viability, potentially influencing funding priorities and development strategies across the academic and corporate AI ecosystem.

MIT NewsRead more

9. Italy’s publishers file a complaint against Google’s AI Overviews over news visibility

Italian publishers' formal complaint against Google's AI Overviews marks a critical test case for AI platform liability regarding content attribution and publisher revenue models. The challenge centers on how AI systems present information without driving traffic to original sources, potentially reshaping the economic relationship between AI platforms and content creators. This regulatory action could establish precedents for AI platform obligations across global markets, influencing how AI companies structure information presentation and revenue-sharing agreements with traditional media organizations.

The DecoderRead more

Know someone who would enjoy this newsletter?

Forward this email or share the link below

https://la-ai.io/newsletter/view/2025-10-27
Subscribe to LA-AI Newsletter

Join Our AI Community

Get weekly insights on AI innovations and exclusive updates on LA-AI events

Subscribe Now