Automated Telegram Digest Generator powered by AI
Telebrief collects messages from your Telegram channels (in any language), generates AI-powered summaries, and delivers beautiful daily digests directly to your Telegram account. Group digests by channel or by AI-detected topics. Supports multiple AI providers: OpenAI, Ollama (local), and Anthropic. Output language is configurable (default: Russian).
- π Multi-language Support - Reads channels in ANY language (English, Russian, Ukrainian, Chinese, etc.)
- π Configurable Output Language - All UI labels, summaries, and bot messages in any language (default: Russian)
- π€ Multi-Provider AI - Supports OpenAI, Ollama (local), and Anthropic for summarization
- β° Scheduled & On-Demand - Daily automatic digests + instant generation via bot commands
- π Private Channel Support - Access your private chats and channels
- π Digest Modes - Group by channel (default) or by AI-detected topics like News, Events, Sport
- π¨ Smart Formatting - Markdown with emojis, bullet points, and clickable channel links
- π¨ Long Message Splitting - Digests that exceed Telegram's 4096-character limit are automatically split into sequential messages instead of being truncated
- π Secure - Single-user only, credentials stored safely
- π§Ή Auto-cleanup - Automatically removes old digest messages
Before you begin, you'll need:
-
Python 3.14+ - Download Python
-
Telegram App Credentials - Get from my.telegram.org
api_idandapi_hash
-
Telegram Bot Token - Create via @BotFather
- Send
/newbotto create a new bot - Save the bot token
- Send
-
AI Provider API Key (one of the following):
- OpenAI: Get from platform.openai.com
- Anthropic: Get from console.anthropic.com
- Ollama: No API key needed - install locally
-
Your Telegram User ID - Get from @userinfobot
- Send
/startto get your ID
- Send
Telebrief can be run in Docker for easy deployment. No Python installation required on host!
The image is published to GitHub Container Registry on every release:
# Pull the latest image
docker pull ghcr.io/belaytzev/telebrief:latestAvailable tags: latest, X.Y (minor), X.Y.Z (patch).
# 1. Create Telegram session (REQUIRED - one-time setup)
./create_session.sh
# 2. Start the service
docker compose up -d
# 3. View logs
docker compose logs -f telebriefThe docker-compose.yml uses the pre-built GHCR image by default, so no local build step is needed. If you want to build from source instead, replace the image: line with build: ..
Important: You must create the Telegram session file BEFORE running Docker. The script uses Docker itself, so no additional dependencies needed.
Open Telegram and message your bot:
| Command | Description |
|---|---|
/start |
Show welcome message and available commands |
/help |
Display help message with all commands |
/digest |
Generate and send digest for last 24 hours (uses configured digest_mode) |
/status |
Show configuration, next scheduled run, and system info |
/cleanup |
Manually delete old digest messages |
Telebrief supports two digest modes configured via digest_mode in config.yaml.
Groups summaries by source channel with clickable channel links:
# π Daily Digest β May 2, 2026
## π― Overview
Today's main themes: AI tooling dominated with Anthropic's Claude Opus 4.7
release, crypto markets rallied on spot ETF approvals, EU finalized
amendments to the AI Act.
---
## π» TechCrunch
- π **Claude Opus 4.7 released**: 1M context window, faster output
- π€ **OpenAI GPT-6 leak**: Multimodal benchmarks surface early
- π± **Apple Vision Pro 2**: Rumored Q3 launch with lighter frame
## π° Crypto News
- π **Bitcoin hits $89K**: Spot ETF inflows reach record high
- β οΈ **SEC settles with Ripple**: Final ruling closes 6-year case
- π **Ethereum Pectra upgrade**: Mainnet activation confirmed
---
π **Stats**: 20 channels, 1,847 messages processedGroups summaries by AI-detected topics. You define topic groups in config.yaml:
digest_mode: "digest"
digest_groups:
- name: "Events"
description: "Conferences, meetups, releases, launches, announcements"
- name: "News"
description: "Politics, economy, world affairs, breaking news"
- name: "Sport"
description: "Sports results, transfers, tournaments, matches"Messages that don't match any defined group are placed into an automatic "Other" category.
All labels (header, statistics, bot commands) follow the configured
output_language. The example above uses English; setoutput_language: "Russian"(or any other language) to change the output.
When multiple channels cover the same event, the grouper normally produces one bullet point per channel. Enable dedup_topics to instruct the AI to keep only the most informative description and merge the source attributions:
settings:
digest_mode: "digest"
dedup_topics: true # default: false
digest_groups:
- name: "Tech"
description: "Technology news and releases"With deduplication enabled, if TechCrunch and HackerNews both report the same product launch, the digest will contain a single bullet point with source: "TechCrunch, HackerNews" instead of two separate entries.
Note:
dedup_topicshas no effect indigest_mode: "channel"β deduplication only applies during topic-based grouping.
Each channel entry supports two optional overrides in addition to the required id and name fields.
Override the global settings.lookback_hours for a specific channel. Useful when some channels post infrequently and need a wider collection window, or when you want a tighter window for high-volume channels.
channels:
- id: "@breaking_news"
name: "Breaking News"
# no lookback_hours β uses the global settings.lookback_hours
- id: "@weekly_digest"
name: "Weekly Newsletter"
lookback_hours: 168 # look back 7 days for this channel only
- id: -1001234567890
name: "High Volume Channel"
lookback_hours: 6 # only last 6 hours for this channellookback_hours must be a positive integer. If omitted or set to null, the global value is used.
Append extra instructions to the AI system prompt when summarizing a specific channel. Use this to guide tone, focus, or format for channels that need special treatment.
channels:
- id: "@cryptonews"
name: "Crypto News"
prompt_extra: "Focus only on price movements and regulatory news. Ignore opinion pieces."
- id: "@jobboard"
name: "Job Board"
prompt_extra: "Extract only senior engineering roles. Format as a list: Role β Company β Link."prompt_extra is appended verbatim to the channel's summarization system prompt. Leave it empty (or omit the field) for standard behavior.
By default, Telebrief generates digests on demand without storing raw messages. You can enable a persistent storage layer that saves every collected message to a database for historical access or external LLM workflows.
Storage is disabled by default and opt-in via config.yaml.
No extra setup required. Messages are saved to a local SQLite file.
storage:
enabled: true
backend: sqlite
path: data/messages.db # relative to project rootWhen running in Docker, the data/ directory is already mounted as a volume in docker-compose.yml, so the database persists across container restarts.
Use PostgreSQL for multi-host deployments or when you need concurrent read access to the message store.
storage:
enabled: true
backend: postgres
url: "postgresql://user:pass@host:5432/dbname"asyncpg is included in the standard dependencies and is installed automatically by uv sync. No extra install step is needed.
Both backends create the same logical schema on first run (table and index are created automatically β no manual migration needed):
| Column | Type | Description |
|---|---|---|
channel_name |
text | Channel name from your config |
sender |
text | Message author |
text |
text | Message body |
timestamp |
text / timestamptz | Message timestamp |
link |
text | Telegram message link |
has_media |
bool / integer | Whether the message has media |
media_type |
text | Media type string |
collected_at |
text / timestamptz | When the row was inserted |
Note: Storage is append-only. Overlapping lookback_hours windows across runs will produce duplicate rows for messages collected in both windows.
Telebrief exposes four hook surfaces that let you customise behaviour via config.yaml without modifying core logic. All new fields are optional β existing configs run unchanged.
A filter chain runs after message collection and before storage and summarization. Dropped messages never reach the AI or the database.
Built-in filters live in src/extensions/filters.py:
| Filter | Purpose |
|---|---|
KeywordFilter |
Keep/drop messages by keyword substring (case-insensitive) |
RegexFilter |
Keep or drop messages matching a regex pattern |
MinLengthFilter |
Drop messages shorter than a character threshold |
Configure a global filter chain under settings.filters. Each entry needs a class_path (dotted import path) and an optional config dict passed as keyword arguments to the constructor:
settings:
filters:
- class_path: src.extensions.filters.KeywordFilter
config:
include: ["job", "hiring", "remote"]
exclude: ["nsfw"]
- class_path: src.extensions.filters.MinLengthFilter
config:
min_chars: 30Override the global chain for a single channel by adding filters: under that channel entry. Set filters: [] to disable filtering for that channel entirely, or provide a different list to replace the global chain for that channel only:
channels:
- id: "@jobboard"
name: "Job Board"
filters:
- class_path: src.extensions.filters.RegexFilter
config:
pattern: "senior|staff|principal"
mode: "include"Write your own filter by implementing the MessageFilter Protocol:
from __future__ import annotations
from src.extensions.filters import MessageFilter
from src.config_loader import ChannelConfig
from src.collector import Message
class MyFilter:
name = "my_filter"
def __init__(self, custom_param: str = "") -> None:
self.custom_param = custom_param
async def filter(self, channel: ChannelConfig, messages: list[Message]) -> list[Message]:
return [m for m in messages if self.custom_param in (m.text or "")]Then reference it in config.yaml:
settings:
filters:
- class_path: mypackage.mymodule.MyFilter
config:
custom_param: "important"The base prompt template lives in src/prompts/base_summary.txt. You can point to a custom template file or plug in a custom PromptComposer class.
prompts:
base_template: src/prompts/base_summary.txt # path to template file
composer: "" # empty = built-in DefaultComposerThe built-in DefaultComposer assembles the final system prompt in this order (empty parts are skipped):
base template (with {language} substituted)
+ group.prompt_extra (if channel belongs to a group with prompt_extra set)
+ channel.prompt_extra (if non-empty)
To use a custom composer, implement the PromptComposer Protocol and set composer to its dotted path:
from src.config_loader import ChannelConfig, DigestGroupConfig
from src.extensions.prompts import PromptComposer
class MyComposer:
def __init__(self, base_template: str, language: str) -> None:
self._base = base_template
self._language = language
def compose(self, channel: ChannelConfig, group: DigestGroupConfig | None) -> str:
return f"{self._base}\nRespond in {self._language}."Note: The constructor must accept
(base_template: str, language: str)as its first two positional arguments. A mismatched signature raises aTypeErrorat startup with a descriptive message.
prompts:
composer: mypackage.mymodule.MyComposerChannels can be bound to a digest_groups entry. The group's prompt_extra is then injected into every channel in that group, before the channel's own prompt_extra.
settings:
digest_groups:
- name: "Jobs"
description: "Job listings and hiring announcements"
prompt_extra: "Extract only role title, company, and link. Format as a list."
channels:
- id: "@techleads_jobs"
name: "Tech Jobs"
group: Jobs # must match a digest_groups name or "Other"
prompt_extra: "Focus on senior and staff-level positions only."Channels without a group field (or group: null) use the base template and their own prompt_extra only.
When storage is enabled (storage.enabled: true), the StorageBackend exposes a query_messages read API for external tooling:
from src.storage import SQLiteBackend
from datetime import datetime, timezone
backend = SQLiteBackend("data/messages.db")
await backend.initialize()
messages = await backend.query_messages(
channel_name="TechCrunch", # the configured channels[*].name (NOT the @id)
since=datetime(2026, 4, 1, tzinfo=timezone.utc),
until=datetime(2026, 4, 30, tzinfo=timezone.utc),
limit=500,
)All parameters are optional. channel_name matches the human-readable channels[*].name value from config.yaml (this is the value persisted to the channel_name column at collection time); omit it to query across all channels. Renaming a channel in config will change the value stored for new rows β historical rows keep the old name. Results are ordered by timestamp descending and capped at limit (default 1000, must be β₯ 1).
This project uses uv for package management.
# Install development dependencies
uv sync --extra dev
# Run all tests
uv run pytest tests/ -v
# Type checking
uv run mypy src/
# Linting
uv tool run ruff check src/ tests/
# Auto-format code
make formatQ: Can I change the output language?
A: Yes! Set output_language in config.yaml to any language (e.g., "English", "Spanish", "Chinese").
Q: How many channels can I monitor? A: Tested up to 50 channels. Performance depends on message volume.
Q: Can multiple users receive digests? A: Currently single-user only. Multi-user support would require database and additional auth logic.
Q: Does it work with group chats?
A: Yes! Add group chat IDs to config.yaml the same way as channels.
Q: How do I switch to topic-based digests?
A: Set digest_mode: "digest" in config.yaml and define your digest_groups. Each group has a name and description that guides the AI classification. An implicit "Other" group catches anything that doesn't match.
Q: Can I customize the digest format?
A: Yes! Edit src/formatter.py to change Markdown structure, emojis, and sections.
Q: How much does it cost to run? A: With OpenAI GPT-5-nano: ~$0.30/month. With Ollama: free (runs locally). Anthropic pricing varies by model.
Q: Can I use a local AI model?
A: Yes! Set ai_provider: "ollama" in config.yaml and install Ollama on your machine.
Built with:
- Telethon - Telegram User API
- python-telegram-bot - Bot API
- OpenAI API - AI Summarization (OpenAI provider)
- Ollama - Local AI Summarization
- Anthropic API - AI Summarization (Anthropic provider)
- APScheduler - Task Scheduling