An autonomous agent powered by Google's Gemini 3 Flash model for creating novels, books, and short story collections.
- 🤖 Autonomous Writing: The agent plans and executes creative writing tasks independently
- 📚 Multiple Formats: Create novels, books, or short story collections
- ⚡ Real-Time Streaming: See the agent's thinking and writing appear as it's generated
- 💾 Smart Context Management: Automatically compresses context when approaching token limits
- 🔄 Recovery Mode: Resume interrupted work from saved context summaries
- 📊 Token Monitoring: Real-time tracking of token usage with automatic optimization
- 🛠️ Tool Use: Agent can create projects, write files, and manage its workspace
- 🧠 Advanced Thinking: Uses Gemini's thinking mode for better reasoning
We recommend using uv for fast Python package management:
# Install uv (if you don't have it)
curl -LsSf https://astral.sh/uv/install.sh | sh- Install dependencies:
Using uv (recommended):
uv pip install -r requirements.txtOr using pip:
pip install -r requirements.txt- Configure your API key:
Create a .env file with your API key:
# Copy the example file
cp env.example .env
# Edit .env and add your API key
# The file should contain:
GEMINI_API_KEY=your-api-key-hereGet your Gemini API key from: https://aistudio.google.com/app/apikey
Run with an inline prompt:
# Using uv (recommended)
uv run writer.py "Create a collection of 5 sci-fi short stories about AI"
# Or using python directly
python writer.py "Create a collection of 5 sci-fi short stories about AI"Or run interactively:
uv run writer.py
# or: python writer.pyThen enter your prompt when asked.
If the agent is interrupted or you want to continue previous work:
uv run writer.py --recover output/my_project/.context_summary_20250107_143022.md
# or: python writer.py --recover output/my_project/.context_summary_20250107_143022.mdThe agent has access to three tools:
- create_project: Creates a project folder to organize the writing
- write_file: Writes markdown files with three modes:
create: Creates a new file (fails if exists)append: Adds content to an existing fileoverwrite: Replaces the entire file content
- compress_context: Automatically triggered to manage context size
- The agent receives your prompt
- It reasons about the task using Gemini's thinking mode
- It decides which tools to call and executes them
- It reviews the results and continues until the task is complete
- Maximum 300 iterations with automatic context compression
- Token Limit: 1,000,000 tokens (Gemini's large context window)
- Auto-Compression: Triggers at 900,000 tokens (90% of limit)
- Backups: Automatic context summaries every 50 iterations
- Recovery: All summaries saved with timestamps for resumption
kimi-writer/
├── writer.py # Main agent
├── tools/
│ ├── __init__.py # Tool registry
│ ├── writer.py # File writing tool
│ ├── project.py # Project management tool
│ └── compression.py # Context compression tool
├── utils.py # Utilities (token counting, etc.)
├── requirements.txt # Python dependencies
├── env.example # Example configuration
├── .gitignore # Git ignore rules
└── README.md # This file
# Generated during use:
output/ # All AI-generated projects go here
├── your_project_name/ # Created by the agent
│ ├── chapter_01.md # Written by the agent
│ ├── chapter_02.md
│ └── .context_summary_*.md # Auto-saved context summaries
└── another_project/
└── ...
uv run writer.py "Write a mystery novel set in Victorian London with 10 chapters"uv run writer.py "Create 7 interconnected sci-fi short stories exploring the theme of memory"uv run writer.py "Write a comprehensive guide to Python programming with 15 chapters"Watch the agent think and write in real-time:
- 🧠 Thinking Stream: See the agent's thought process as it plans (Gemini's thinking mode)
- 💬 Content Stream: Watch stories being written character by character
- 🔧 Tool Call Progress: Live updates when generating large content
- ⚡ No Waiting: Immediate feedback - no more staring at a blank screen
The agent displays its progress: Iteration X/300
Real-time token usage: Current tokens: 45,234/1,000,000 (4.5%)
Press Ctrl+C to interrupt. The agent will save the current context for recovery.
-
Be Specific: Clear prompts get better results
- Good: "Create a 5-chapter romance novel set in modern Tokyo"
- Less good: "Write something interesting"
-
Let It Work: The agent works autonomously - it will plan and execute the full task
-
Recovery is Easy: If interrupted, just use the
--recoverflag with the latest context summary -
Check Progress: Generated files appear in real-time in the project folder
Make sure you have created a .env file in the project root with your API key:
GEMINI_API_KEY=your-actual-api-key-here- Verify your API key is correct in the
.envfile - Get your API key from: https://aistudio.google.com/app/apikey
Check write permissions in the current directory
The agent can run up to 300 iterations. For very complex tasks, this is normal. Check the project folder to see progress.
The agent automatically compresses context at 900K tokens. If you see compression messages, the system is working correctly.
- Model: gemini-3-flash-preview
- Thinking Level: HIGH (for better reasoning)
- Temperature: 1.0
- Context Window: 1,000,000 tokens
- Max Iterations: 300
- Compression Threshold: 900,000 tokens
You can customize these settings in writer.py.
MIT License with Attribution Requirement - see LICENSE file for details.
Commercial Use: If you use this software in a commercial product, you must provide clear attribution to Pietro Schirano (@Doriandarko).
API Usage: This project uses the Google Gemini API. Please refer to Google's terms of service for API usage guidelines.
- Created by: Pietro Schirano (@Doriandarko)
- Powered by: Google's Gemini 3 Flash model
- Repository: https://github.com/Doriandarko/gemini-writer