Miniflux with AI
This project fetches RSS subscription content from Miniflux via API and utilizes a large language model (LLM) to generate summaries, translations, etc. The configuration file allows for easy customization and addition of LLM agents.
config.yml
file.pip install -r requirements.txt
The repository includes a template configuration file: config.sample.yml
. Modify the config.yml
to set up:
style_block
parameter controls whether the output is formatted as a code block in Markdown).Example config.yml
:
# INFO、DEBUG、WARN、ERROR
log_level: "INFO"
miniflux:
base_url: https://your.server.com
api_key: Miniflux API key here
llm:
base_url: http://host.docker.internal:11434/v1
api_key: ollama
model: llama3.1:latest
# timeout: 60
# max_workers: 4
agents:
summary:
title: "💡AI 摘要"
prompt: "Please summarize the content of the article under 50 words in Chinese. Do not add any additional Character、markdown language to the result text. 请用不超过50个汉字概括文章内容。结果文本中不要添加任何额外的字符、Markdown语言。"
style_block: true
deny_list:
- https://xxxx.net
allow_list:
translate:
title: "🌐AI 翻译"
prompt: "You are a highly skilled translation engine with expertise in the news media sector. Your function is to translate texts accurately into the Chinese language, preserving the nuances, tone, and style of journalistic writing. Do not add any explanations or annotations to the translated text."
style_block: false
deny_list:
allow_list:
- https://www.xxx.com/
The project includes a docker-compose.yml
file for easy deployment:
version: '3.3'
services:
miniflux_ai:
container_name: miniflux_ai
image: ghcr.io/qetesh/miniflux-ai:latest
restart: always
environment:
TZ: Asia/Shanghai
volumes:
- ./config.yml:/app/config.yml
To start the service, run:
docker-compose up -d
config.yml
is properly configured.python main.py
Feel free to fork this repository and submit pull requests. Contributions are welcome!
This project is licensed under the MIT License.