top of page

Why Does Formatting Look Right in AI Chat but Break in Other Systems?

  • May 3
  • 2 min read
Why does formatting (like tables) look correct in AI chat but break when written to another system?

Let’s break down what’s actually happening.

User comparing formatted content in an AI chat window with a broken table layout in another system on a laptop screen.

What’s really going on

When you use an AI tool to generate formatted content, you’re seeing a preview, not the final stored version. That preview is optimized for readability in the chat interface—not for how another system will store it. Once you send that content elsewhere, a second process kicks in.


The two-step problem

There are always two steps involved:

1. Generation (AI model): The AI produces content that looks structured—often using markdown-like formatting.

2. Conversion (target system): The destination system converts that output into its own internal format (often JSON-based or proprietary).

That second step is where things break.


Why formatting falls apart

Most systems don’t store content the way AI generates it. They:

  • don’t accept raw markdown directly

  • sanitize or strip unsupported formatting

  • convert everything into structured schemas

This is especially noticeable with:

  • tables

  • mixed formatting

  • nested content

  • inline markup

So even if the output looks perfect in chat, it may not survive conversion.


Why automation feels more reliable

This is where the difference becomes clear.

Automation and system-native tools:

  • write directly in the system’s format

  • avoid interpretation

  • produce consistent results

AI-generated workflows:

  • generate first, then translate

  • rely on conversion layers

  • introduce variability

Same goal, different reliability.


Why interaction flow feels unpredictable

Another common frustration:

Trying to control how AI guides the next step.

Even if you want:

  • fixed options

  • required confirmations

  • strict sequences

Most AI interfaces today:

  • respond dynamically

  • may ask follow-up questions

  • don’t enforce rigid flows

You can guide behavior—but not fully control it.


What to do today

If your use case involves:

Structured updates (tables, records, forms)→ Use automation or system-native tools

Drafting, summarizing, or logic→ Use AI interfaces

End-to-end workflows→ Combine both

  • AI generates the content

  • the system writes it reliably


The bigger pattern

AI is doing two jobs:

  • making content understandable to humans

  • attempting to translate it for systems

Most systems only reliably handle the second job when it’s done directly.


Takeaway

If formatting breaks, it’s usually not an AI failure.

It’s the gap between: how AI presents content and how systems store it


AI makes things look right. Your system decides whether they actually are.

Comments


bottom of page