Codex Automation: Generating i18n Support Across 18 Languages
Experience report on using Codex to automatically generate internationalization and README files in 18 languages, reducing weeks of work to 30 minutes.
How Codex Automated 18-Language i18n Support in 30 Minutes — And What It Means for Developers
One prompt. 30 minutes. 8.01 million tokens. 2,000 lines of code. A week's worth of work — done.
If you've ever been responsible for internationalizing a software project, you already know the pain. Translating UI strings, updating README files, maintaining consistency across a dozen languages, and keeping everything in sync as the codebase evolves — it's a grind. It's important work, but it rarely feels like progress. It feels like maintenance.
That's exactly what makes a recent post from developer @cholf5 on X/Twitter so striking. In a single interaction with OpenAI Codex, they issued one natural-language instruction:
"Provide i18n and README in 18 mainstream languages."
Codex ran for 30 minutes, consumed 8.01 million tokens (approximately $2.85), and produced 2,000 new lines of code — covering internationalization files and documentation across 18 languages. Work that, as @cholf5 put it, would have taken a human developer "at least a week" without even accounting for review cycles, handoffs, or translation inconsistencies.
Let's unpack why this matters, what's happening under the hood, and how you can apply this kind of AI-driven automation to your own projects.
Why i18n Is a Developer Bottleneck (And Why It Shouldn't Be)
Internationalization — abbreviated as i18n (because there are 18 letters between the "i" and the "n") — is one of those engineering tasks that's deceptively complex. On the surface, it seems like a translation problem. In practice, it's a software architecture problem layered on top of a linguistics problem.
A complete i18n implementation typically involves:
- Extracting all user-facing strings from the codebase into locale files (e.g.,
en.json,zh.json,de.json) - Translating each string accurately, accounting for context, tone, and cultural nuance
- Maintaining parallel structures across all locale files so keys stay consistent
- Updating documentation (like README files) to reflect multilingual support
- Testing to ensure no strings are missed, broken, or incorrectly displayed in RTL (right-to-left) languages like Arabic or Hebrew
When you're targeting just 2–3 languages, this is manageable. When you're targeting 18 mainstream languages — which might include English, Simplified Chinese, Traditional Chinese, Spanish, French, German, Japanese, Korean, Portuguese, Russian, Italian, Dutch, Arabic, Turkish, Polish, Vietnamese, Thai, and Hindi — the complexity multiplies dramatically.
For a single developer or a small team, a full i18n pass across 18 languages historically meant:
- Days of string extraction and file scaffolding
- Coordination with translators or translation services
- Multiple review rounds for accuracy
- Manual documentation updates for each language
This is precisely the kind of high-volume, highly structured, yet cognitively repetitive task where large language models genuinely excel.
What Codex Actually Did — And How to Replicate It
The magic of the @cholf5 example isn't just that Codex translated text. It's that Codex understood the full scope of a multilingual software project from a single sentence and executed it end-to-end.
Here's a breakdown of what likely happened during those 30 minutes and 8 million tokens:
1. Locale File Generation
Codex would have scanned the existing codebase to identify all user-facing strings, then generated structured locale files for each of the 18 target languages. A typical output might look like:
// locales/zh-CN.json
{
"welcome_message": "欢迎使用我们的应用程序",
"login_button": "登录",
"error_not_found": "页面未找到",
"settings_title": "设置"
}
// locales/de.json
{
"welcome_message": "Willkommen in unserer Anwendung",
"login_button": "Anmelden",
"error_not_found": "Seite nicht gefunden",
"settings_title": "Einstellungen"
}
Multiply this pattern across hundreds of string keys and 18 languages, and you start to appreciate the volume involved — and why this consumed 2,000 lines.
2. README Localization
Beyond code, Codex generated localized README files for each language — a task that's often overlooked but critically important for open-source projects and international developer communities.
A multilingual README directory might look like:
/docs
README.md ← English (default)
README.zh-CN.md ← Simplified Chinese
README.zh-TW.md ← Traditional Chinese
README.ja.md ← Japanese
README.ko.md ← Korean
README.de.md ← German
README.fr.md ← French
README.es.md ← Spanish
README.ar.md ← Arabic (RTL)
...
Each file isn't just a direct translation — a well-executed localization adapts the tone, examples, and formatting to be natural for native speakers of each language.
3. Structural Consistency and Key Management
One of the hardest parts of maintaining i18n files is ensuring key parity — that every language file contains the same set of keys, even if some translations are pending. Codex handled this systematically, producing files with consistent structure that a developer can immediately validate and integrate.
The Real Takeaway: Rethinking What "Developer Work" Means in 2025
The @cholf5 example isn't just a productivity story. It's a signal about the nature of AI-assisted development.
Consider the economics: $2.85 in API costs replaced what could easily be $1,500–$5,000 in developer or translator time for a task of this scope. Even accounting for human review and cleanup of the AI-generated output, the efficiency gain is transformative.
But the more important shift is conceptual. Tasks that developers previously avoided — not because they were technically hard, but because they were tedious and time-consuming — are now viable with a single prompt. This changes project scoping, MVP planning, and what's realistic for a solo developer or small team to ship.
Here are a few similar use cases where this pattern applies directly:
- Auto-generating OpenAPI/Swagger documentation in multiple languages
- Creating locale-aware test suites that validate UI strings across environments
- Scaffolding multi-language marketing copy from a single source-of-truth document
- Building locale configuration files for cloud infrastructure across regions
The common thread: tasks with high structure, high volume, and well-defined rules — exactly what LLMs are optimized for.
Getting Started: Run Your Own i18n Automation
If you want to replicate this workflow, here's a minimal starting point using the OpenAI API:
import openai
prompt = """
You are an expert software localization engineer.
Given the following English locale file, generate equivalent
locale files for these 18 languages: zh-CN, zh-TW, ja, ko,
es, fr, de, pt, ru, it, nl, ar, tr, pl, vi, th, hi.
Return each file as clearly labeled JSON.
English source:
{
"welcome": "Welcome to our app",
"login": "Log in",
"logout": "Log out",
"settings": "Settings"
}
"""
response = openai.chat.completions.create(
model="gpt-4o",
messages=[{"role": "user", "content": prompt}],
max_tokens=8000
)
print(response.choices[0].message.content)
For larger codebases, you'd want to chunk your string extraction, run multiple API calls, and pipe the results into your existing i18n framework — whether that's react-i18next, i18next, Flutter's ARB format, or Android's strings.xml.
Conclusion: The $2.85 That Changed the Math
What @cholf5 demonstrated isn't a party trick. It's a glimpse at how the economics of software development are being rewritten in real time.
Thirty minutes. $2.85. 18 languages. 2,000 lines.
For any developer who has ever stared down a localization backlog and felt defeated before starting, this is the news you've been waiting for. AI-powered automation — through tools like Codex, GPT-4o, and emerging agent frameworks — is making global-ready software accessible to builders who previously couldn't afford the time or resources to pursue it.
The bottleneck is no longer capability. It's imagination.
If your project isn't internationalized yet, you might just be one prompt away from changing that.
Originally inspired by @cholf5 on X/Twitter. Published on ClawList.io — your resource hub for AI automation and OpenClaw skills.
Tags
Related Articles
Vercel's React Best Practices as Reusable Skill
Vercel distilled 10 years of React expertise into a skill, demonstrating how organizations should package internal best practices as reusable AI agent skills.
Building Commercial Apps with Claude Opus
Experience sharing on rapid app development using Claude Opus as a CTO, product manager, and designer combined.
AI-Powered Product Marketing with Video and Social Media
Guide on using AI to create product advertisement videos, user testimonials, and product images for social media marketing campaigns.