Compare commits
No commits in common. "master" and "main" have entirely different histories.
9 changed files with 5 additions and 721 deletions
14
.env.example
14
.env.example
|
@ -1,14 +0,0 @@
|
||||||
# GTS Server Configuration
|
|
||||||
GTS_SERVER_URL=https://your-gts-instance.tld
|
|
||||||
GTS_ACCESS_TOKEN=your_gts_access_token_here
|
|
||||||
|
|
||||||
# Processing Configuration
|
|
||||||
MAX_POSTS_PER_RUN=75
|
|
||||||
DELAY_BETWEEN_REQUESTS=1
|
|
||||||
LOG_LEVEL=INFO
|
|
||||||
|
|
||||||
# RSS Configuration
|
|
||||||
RSS_URLS_FILE=/app/rss_feeds.txt
|
|
||||||
|
|
||||||
# Optional: Monitoring
|
|
||||||
# HEALTHCHECK_URL=https://hc-ping.com/your-uuid-here
|
|
29
.gitignore
vendored
29
.gitignore
vendored
|
@ -1,29 +0,0 @@
|
||||||
# Secrets and local config
|
|
||||||
.env
|
|
||||||
rss_feeds.txt
|
|
||||||
|
|
||||||
# Data directory
|
|
||||||
data/
|
|
||||||
*.json
|
|
||||||
|
|
||||||
# Python
|
|
||||||
__pycache__/
|
|
||||||
*.pyc
|
|
||||||
*.pyo
|
|
||||||
|
|
||||||
# Docker
|
|
||||||
.dockerignore
|
|
||||||
|
|
||||||
# Logs
|
|
||||||
*.log
|
|
||||||
|
|
||||||
# Editor files
|
|
||||||
.vscode/
|
|
||||||
.idea/
|
|
||||||
*.swp
|
|
||||||
*.swo
|
|
||||||
*~
|
|
||||||
|
|
||||||
# OS files
|
|
||||||
.DS_Store
|
|
||||||
Thumbs.db
|
|
24
Dockerfile
24
Dockerfile
|
@ -1,24 +0,0 @@
|
||||||
# Dockerfile
|
|
||||||
FROM python:3.11-slim
|
|
||||||
|
|
||||||
# Set working directory
|
|
||||||
WORKDIR /app
|
|
||||||
|
|
||||||
# Copy and install requirements
|
|
||||||
COPY requirements.txt .
|
|
||||||
RUN pip install --no-cache-dir -r requirements.txt
|
|
||||||
|
|
||||||
# Create data directory
|
|
||||||
RUN mkdir -p /app/data
|
|
||||||
|
|
||||||
# Create non-root user
|
|
||||||
RUN useradd -r -u 1000 holmirdas
|
|
||||||
|
|
||||||
# Set ownership
|
|
||||||
RUN chown -R holmirdas:holmirdas /app
|
|
||||||
|
|
||||||
# Switch to non-root user
|
|
||||||
USER holmirdas
|
|
||||||
|
|
||||||
# Default command (will be overridden by docker-compose)
|
|
||||||
CMD ["python", "gts_holmirdas.py"]
|
|
20
LICENSE
20
LICENSE
|
@ -1,21 +1,9 @@
|
||||||
MIT License
|
MIT License
|
||||||
|
|
||||||
Copyright (c) 2025 Matthias
|
Copyright (c) 2024 oliverpifferi
|
||||||
|
|
||||||
Permission is hereby granted, free of charge, to any person obtaining a copy
|
Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions:
|
||||||
of this software and associated documentation files (the "Software"), to deal
|
|
||||||
in the Software without restriction, including without limitation the rights
|
|
||||||
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
|
|
||||||
copies of the Software, and to permit persons to whom the Software is
|
|
||||||
furnished to do so, subject to the following conditions:
|
|
||||||
|
|
||||||
The above copyright notice and this permission notice shall be included in all
|
The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software.
|
||||||
copies or substantial portions of the Software.
|
|
||||||
|
|
||||||
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
|
||||||
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
||||||
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
|
|
||||||
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
||||||
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
|
||||||
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
|
||||||
SOFTWARE.
|
|
306
README.md
306
README.md
|
@ -1,305 +1 @@
|
||||||
# GTS-HolMirDas 🚀
|
https://keyoxide.org/AF953733C09F6368797CA49E2F4968B5F7988AA5
|
||||||
|
|
||||||
RSS-based content discovery for **[GoToSocial](https://codeberg.org/superseriousbusiness/gotosocial)** instances.
|
|
||||||
|
|
||||||
Automatically discovers and federates content from RSS feeds across the Fediverse, helping small GoToSocial instances populate their federated timeline without relying on traditional relays.
|
|
||||||
|
|
||||||
*Inspired by the original [HolMirDas](https://github.com/aliceif/HolMirDas) for Misskey by [@aliceif](https://github.com/aliceif) ([@aliceif@mkultra.x27.one](https://mkultra.x27.one/@aliceif)), this GoToSocial adaptation extends the RSS-to-ActivityPub concept with enhanced Docker deployment and multi-instance processing.*
|
|
||||||
|
|
||||||
## Features
|
|
||||||
|
|
||||||
- 📡 **Multi-Instance RSS Discovery** - Fetches content from configurable RSS feeds across Fediverse instances
|
|
||||||
- ⚡ **Efficient Processing** - Configurable rate limiting and duplicate detection
|
|
||||||
- 🔧 **Production Ready** - Environment-based config, Docker deployment, health monitoring
|
|
||||||
- 📊 **Comprehensive Statistics** - Runtime metrics, content processing, and federation growth tracking
|
|
||||||
- 🐳 **Containerized** - Simple Docker Compose deployment
|
|
||||||
- 📁 **File-based Configuration** - Easy RSS feed management via text files
|
|
||||||
|
|
||||||
## How it Works
|
|
||||||
|
|
||||||
**GTS-HolMirDas** reads RSS feeds from various Fediverse instances and uses GoToSocial's search API to federate the discovered content. This approach:
|
|
||||||
|
|
||||||
- Maintains proper ActivityPub federation (posts remain interactive)
|
|
||||||
- Respects rate limits and instance policies
|
|
||||||
- Provides better content discovery for small instances
|
|
||||||
- Works alongside tools like FediFetcher for comprehensive federation
|
|
||||||
|
|
||||||
## Quick Start
|
|
||||||
|
|
||||||
```bash
|
|
||||||
# Clone the repository
|
|
||||||
git clone https://git.klein.ruhr/matthias/gts-holmirdas
|
|
||||||
cd gts-holmirdas
|
|
||||||
|
|
||||||
# Copy configuration templates
|
|
||||||
cp .env.example .env
|
|
||||||
cp rss_feeds.example.txt rss_feeds.txt
|
|
||||||
|
|
||||||
# Edit configuration
|
|
||||||
nano .env # Add your GTS credentials
|
|
||||||
nano rss_feeds.txt # Customize RSS feeds
|
|
||||||
|
|
||||||
# Deploy
|
|
||||||
docker compose up -d
|
|
||||||
|
|
||||||
# Monitor
|
|
||||||
docker compose logs -f
|
|
||||||
```
|
|
||||||
# Performance Scaling & Configuration
|
|
||||||
|
|
||||||
## 🚀 RSS Feed Optimization (v1.1.0+)
|
|
||||||
|
|
||||||
GTS-HolMirDas supports URL parameters to dramatically increase content discovery without additional API calls.
|
|
||||||
|
|
||||||
### RSS Feed Limits
|
|
||||||
|
|
||||||
Most Mastodon-compatible instances support the `?limit=X` parameter:
|
|
||||||
|
|
||||||
```
|
|
||||||
# Default behavior (20 posts per feed)
|
|
||||||
https://mastodon.social/tags/homelab.rss
|
|
||||||
|
|
||||||
# Increased limits (up to 100 posts per feed)
|
|
||||||
https://mastodon.social/tags/homelab.rss?limit=50
|
|
||||||
https://fosstodon.org/tags/docker.rss?limit=100
|
|
||||||
```
|
|
||||||
|
|
||||||
**Supported limits:** 20 (default), 50, 75, 100 (instance-dependent)
|
|
||||||
|
|
||||||
### Performance Impact
|
|
||||||
|
|
||||||
| Configuration | Posts/Run | API Calls | Processing Time |
|
|
||||||
|---------------|-----------|-----------|-----------------|
|
|
||||||
| Standard (limit=20) | ~100 posts | 30+ feeds | 2-5 minutes |
|
|
||||||
| Optimized (limit=50) | ~300 posts | 30+ feeds | 5-10 minutes |
|
|
||||||
| Maximum (limit=100) | ~600 posts | 30+ feeds | 8-15 minutes |
|
|
||||||
|
|
||||||
## ⚙️ Configuration Tuning
|
|
||||||
|
|
||||||
### Environment Variables
|
|
||||||
|
|
||||||
```env
|
|
||||||
# Processing Configuration
|
|
||||||
MAX_POSTS_PER_RUN=75 # Increase for higher limits
|
|
||||||
DELAY_BETWEEN_REQUESTS=1 # Balance speed vs. server load
|
|
||||||
RSS_URLS_FILE=/app/rss_feeds.txt
|
|
||||||
|
|
||||||
# Recommended combinations:
|
|
||||||
# Conservative: MAX_POSTS_PER_RUN=40, limit=50
|
|
||||||
# Balanced: MAX_POSTS_PER_RUN=75, limit=100
|
|
||||||
# Aggressive: MAX_POSTS_PER_RUN=100, limit=100
|
|
||||||
```
|
|
||||||
|
|
||||||
### RSS Feed Strategy
|
|
||||||
|
|
||||||
```
|
|
||||||
# Progressive scaling approach:
|
|
||||||
# 1. Start with mixed limits to test performance
|
|
||||||
# 2. Increase gradually based on server capacity
|
|
||||||
# 3. Monitor GoToSocial memory usage
|
|
||||||
|
|
||||||
# Example progression:
|
|
||||||
https://mastodon.social/tags/homelab.rss?limit=50
|
|
||||||
https://fosstodon.org/tags/selfhosting.rss?limit=75
|
|
||||||
https://chaos.social/tags/docker.rss?limit=100
|
|
||||||
```
|
|
||||||
|
|
||||||
## 📊 Monitoring & Optimization
|
|
||||||
|
|
||||||
### Performance Metrics
|
|
||||||
|
|
||||||
The statistics output shows real-time performance:
|
|
||||||
|
|
||||||
```
|
|
||||||
📊 GTS-HolMirDas Run Statistics:
|
|
||||||
⏱️ Runtime: 0:08:42
|
|
||||||
📄 Total posts processed: 487
|
|
||||||
🌐 Current known instances: 3150
|
|
||||||
➕ New instances discovered: +45
|
|
||||||
📡 RSS feeds processed: 102
|
|
||||||
⚡ Posts per minute: 56.0
|
|
||||||
```
|
|
||||||
|
|
||||||
### Optimization Guidelines
|
|
||||||
|
|
||||||
**Memory Usage:**
|
|
||||||
- Monitor GoToSocial memory consumption during runs
|
|
||||||
- Each 100 additional posts ≈ ~2-5MB additional RAM
|
|
||||||
- Recommended: 1GB+ RAM for aggressive configurations
|
|
||||||
|
|
||||||
**Processing Time:**
|
|
||||||
- Scales linearly with `MAX_POSTS_PER_RUN × number_of_feeds`
|
|
||||||
- Duplicate detection becomes more important at scale
|
|
||||||
- Consider running frequency vs. content volume
|
|
||||||
|
|
||||||
**Federation Growth:**
|
|
||||||
- Higher limits = more diverse instance discovery
|
|
||||||
- Expect 20-50+ new instances per optimized run
|
|
||||||
- Balance discovery rate with storage capacity
|
|
||||||
|
|
||||||
### Troubleshooting High-Volume Setups
|
|
||||||
|
|
||||||
**If processing takes too long:**
|
|
||||||
```env
|
|
||||||
MAX_POSTS_PER_RUN=50 # Reduce from 75/100
|
|
||||||
DELAY_BETWEEN_REQUESTS=2 # Increase from 1
|
|
||||||
```
|
|
||||||
|
|
||||||
**If GoToSocial uses too much memory:**
|
|
||||||
- Reduce RSS feed count temporarily
|
|
||||||
- Lower `?limit=` parameters to 50 instead of 100
|
|
||||||
- Increase run frequency instead of volume
|
|
||||||
|
|
||||||
**If duplicate detection is slow:**
|
|
||||||
- Storage cleanup: `docker-compose exec gts-holmirdas rm -f /app/data/processed_urls.json`
|
|
||||||
- This forces fresh state tracking (posts will be reprocessed once)
|
|
||||||
|
|
||||||
## 🎯 Best Practices
|
|
||||||
|
|
||||||
### Scaling Strategy
|
|
||||||
|
|
||||||
1. **Start Conservative:** `limit=50`, `MAX_POSTS_PER_RUN=40`
|
|
||||||
2. **Monitor Performance:** Check RAM usage and processing time
|
|
||||||
3. **Scale Gradually:** Increase to `limit=75`, then `limit=100`
|
|
||||||
4. **Optimize Mix:** Use different limits per instance based on quality
|
|
||||||
|
|
||||||
### Instance Selection
|
|
||||||
|
|
||||||
**High-quality instances for aggressive limits:**
|
|
||||||
```
|
|
||||||
# Tech-focused instances (good signal-to-noise ratio)
|
|
||||||
https://fosstodon.org/tags/homelab.rss?limit=100
|
|
||||||
https://infosec.exchange/tags/security.rss?limit=100
|
|
||||||
|
|
||||||
# General instances (moderate limits recommended)
|
|
||||||
https://mastodon.social/tags/technology.rss?limit=50
|
|
||||||
```
|
|
||||||
|
|
||||||
**Performance tip:** Specialized instances often have higher content quality at scale than general-purpose instances.
|
|
||||||
|
|
||||||
## Configuration
|
|
||||||
|
|
||||||
### Environment Variables (.env)
|
|
||||||
|
|
||||||
```bash
|
|
||||||
# GTS Server Configuration
|
|
||||||
GTS_SERVER_URL=https://your-gts-instance.tld
|
|
||||||
GTS_ACCESS_TOKEN=your_gts_access_token
|
|
||||||
|
|
||||||
# Processing Configuration
|
|
||||||
MAX_POSTS_PER_RUN=25 # Posts per feed per run
|
|
||||||
DELAY_BETWEEN_REQUESTS=1 # Seconds between API calls
|
|
||||||
LOG_LEVEL=INFO # Logging verbosity
|
|
||||||
|
|
||||||
# RSS Configuration
|
|
||||||
RSS_URLS_FILE=/app/rss_feeds.txt # Path to RSS feeds file
|
|
||||||
|
|
||||||
# Optional: Monitoring
|
|
||||||
HEALTHCHECK_URL=https://hc-ping.com/your-uuid
|
|
||||||
```
|
|
||||||
|
|
||||||
### RSS Feeds (rss_feeds.txt)
|
|
||||||
|
|
||||||
```
|
|
||||||
# Example RSS feeds - customize for your interests
|
|
||||||
# homelab
|
|
||||||
https://mastodon.social/tags/homelab.rss
|
|
||||||
https://fosstodon.org/tags/homelab.rss
|
|
||||||
|
|
||||||
# selfhosting
|
|
||||||
https://mastodon.social/tags/selfhosting.rss
|
|
||||||
https://infosec.exchange/tags/selfhosting.rss
|
|
||||||
|
|
||||||
# Add your preferred instances and hashtags
|
|
||||||
```
|
|
||||||
|
|
||||||
## Access Token Setup
|
|
||||||
|
|
||||||
1. Login to your GoToSocial instance
|
|
||||||
2. Go to Settings → Applications
|
|
||||||
3. Create new application with scopes: `read`, `read:search`, `read:statuses`
|
|
||||||
4. Copy the access token to your `.env` file
|
|
||||||
|
|
||||||
## Statistics Output
|
|
||||||
|
|
||||||
```
|
|
||||||
📊 GTS-HolMirDas Run Statistics:
|
|
||||||
⏱️ Runtime: 0:04:14
|
|
||||||
📄 Total posts processed: 45
|
|
||||||
🌐 Current known instances: 2519
|
|
||||||
➕ New instances discovered: +3
|
|
||||||
📡 RSS feeds processed: 25
|
|
||||||
⚡ Posts per minute: 10.6
|
|
||||||
```
|
|
||||||
|
|
||||||
## Resource Requirements
|
|
||||||
|
|
||||||
- **Memory**: ~200-500MB depending on feed count
|
|
||||||
- **CPU**: Minimal (mostly I/O bound)
|
|
||||||
- **Storage**: <100MB for application, plus log storage
|
|
||||||
- **Network**: Depends on RSS feed count and frequency
|
|
||||||
|
|
||||||
## Deployment Options
|
|
||||||
|
|
||||||
### Docker Compose (Recommended)
|
|
||||||
```bash
|
|
||||||
docker compose up -d
|
|
||||||
```
|
|
||||||
|
|
||||||
### Standalone Docker
|
|
||||||
```bash
|
|
||||||
docker build -t gts-holmirdas .
|
|
||||||
docker run -d --env-file .env \
|
|
||||||
-v ./data:/app/data \
|
|
||||||
-v ./gts_holmirdas.py:/app/gts_holmirdas.py:ro \
|
|
||||||
-v ./rss_feeds.txt:/app/rss_feeds.txt:ro \
|
|
||||||
gts-holmirdas
|
|
||||||
```
|
|
||||||
|
|
||||||
## Monitoring
|
|
||||||
|
|
||||||
- **Logs**: `docker compose logs -f`
|
|
||||||
- **Health**: Optional Healthchecks.io integration
|
|
||||||
- **Statistics**: Built-in runtime and performance metrics
|
|
||||||
- **Resource Usage**: Docker stats or container monitoring tools
|
|
||||||
|
|
||||||
## Troubleshooting
|
|
||||||
|
|
||||||
### Common Issues
|
|
||||||
|
|
||||||
- **No posts processed**: Check access token permissions and RSS feed URLs
|
|
||||||
- **Rate limiting errors**: Increase `DELAY_BETWEEN_REQUESTS` or reduce feed count
|
|
||||||
- **High memory usage**: Reduce `MAX_POSTS_PER_RUN` or feed frequency
|
|
||||||
- **Container won't start**: Verify `.env` file format and required variables
|
|
||||||
|
|
||||||
### Debug Mode
|
|
||||||
|
|
||||||
```bash
|
|
||||||
# Enable debug logging
|
|
||||||
echo "LOG_LEVEL=DEBUG" >> .env
|
|
||||||
docker compose restart gts-holmirdas
|
|
||||||
```
|
|
||||||
|
|
||||||
## Contributing
|
|
||||||
|
|
||||||
1. Fork the repository
|
|
||||||
2. Create a feature branch
|
|
||||||
3. Make your changes
|
|
||||||
4. Test thoroughly
|
|
||||||
5. Submit a pull request
|
|
||||||
|
|
||||||
## Related Projects
|
|
||||||
|
|
||||||
- [FediFetcher](https://github.com/nanos/fedifetcher) - Fetches missing replies and posts
|
|
||||||
- [GoToSocial](https://github.com/superseriousbusiness/gotosocial) - Lightweight ActivityPub server
|
|
||||||
- [slurp](https://github.com/VyrCossont/slurp) - Import posts from other instances
|
|
||||||
|
|
||||||
## License
|
|
||||||
|
|
||||||
MIT License - see LICENSE file for details.
|
|
||||||
|
|
||||||
## Acknowledgments
|
|
||||||
|
|
||||||
- Inspired by [HolMirDas](https://github.com/aliceif/HolMirDas) by [@aliceif](https://github.com/aliceif) ([@aliceif@mkultra.x27.one](https://mkultra.x27.one/@aliceif)) - the original RSS-to-ActivityPub concept
|
|
||||||
- Built for the GoToSocial community
|
|
||||||
- RSS-to-ActivityPub approach inspired by Fediverse discovery challenges
|
|
32
compose.yml
32
compose.yml
|
@ -1,32 +0,0 @@
|
||||||
services:
|
|
||||||
gts-holmirdas:
|
|
||||||
build: .
|
|
||||||
container_name: gts-holmirdas
|
|
||||||
restart: unless-stopped
|
|
||||||
|
|
||||||
env_file:
|
|
||||||
- .env
|
|
||||||
|
|
||||||
volumes:
|
|
||||||
- ./data:/app/data
|
|
||||||
- ./gts_holmirdas.py:/app/gts_holmirdas.py:ro
|
|
||||||
- ./rss_feeds.txt:/app/rss_feeds.txt:ro
|
|
||||||
|
|
||||||
# Run every 3 hours (balanced frequency)
|
|
||||||
entrypoint: >
|
|
||||||
sh -c "
|
|
||||||
while true; do
|
|
||||||
echo 'Starting GTS-HolMirDas run...'
|
|
||||||
python gts_holmirdas.py
|
|
||||||
echo 'GTS-HolMirDas run completed. Sleeping for 1 hour...'
|
|
||||||
sleep 3600
|
|
||||||
done
|
|
||||||
"
|
|
||||||
|
|
||||||
# Resource limits
|
|
||||||
deploy:
|
|
||||||
resources:
|
|
||||||
limits:
|
|
||||||
memory: 512M
|
|
||||||
reservations:
|
|
||||||
memory: 256M
|
|
281
gts_holmirdas.py
281
gts_holmirdas.py
|
@ -1,281 +0,0 @@
|
||||||
#!/usr/bin/env python3
|
|
||||||
"""
|
|
||||||
GTS-HolMirDas: RSS-based content discovery for GoToSocial
|
|
||||||
|
|
||||||
Inspired by HolMirDas by @aliceif:
|
|
||||||
- GitHub: https://github.com/aliceif/HolMirDas
|
|
||||||
- Fediverse: @aliceif@mkultra.x27.one
|
|
||||||
|
|
||||||
This GoToSocial adaptation extends the original RSS-to-ActivityPub concept
|
|
||||||
with Docker deployment, multi-instance processing, and comprehensive monitoring.
|
|
||||||
"""
|
|
||||||
|
|
||||||
import os
|
|
||||||
import sys
|
|
||||||
import time
|
|
||||||
import json
|
|
||||||
import logging
|
|
||||||
import requests
|
|
||||||
import feedparser
|
|
||||||
from datetime import timedelta
|
|
||||||
from urllib.parse import quote_plus
|
|
||||||
|
|
||||||
class GTSHolMirDas:
|
|
||||||
def __init__(self):
|
|
||||||
"""Initialize the RSS fetcher with configuration"""
|
|
||||||
self.config = {
|
|
||||||
"server_url": os.getenv("GTS_SERVER_URL", "https://your-gts-instance"),
|
|
||||||
"access_token": os.getenv("GTS_ACCESS_TOKEN", ""),
|
|
||||||
"max_posts_per_run": int(os.getenv("MAX_POSTS_PER_RUN", "25")),
|
|
||||||
"delay_between_requests": int(os.getenv("DELAY_BETWEEN_REQUESTS", "2")),
|
|
||||||
"healthcheck_url": os.getenv("HEALTHCHECK_URL", ""),
|
|
||||||
"log_level": os.getenv("LOG_LEVEL", "INFO")
|
|
||||||
}
|
|
||||||
|
|
||||||
# Setup logging FIRST
|
|
||||||
logging.basicConfig(
|
|
||||||
level=getattr(logging, self.config["log_level"]),
|
|
||||||
format='%(asctime)s - %(levelname)s - %(message)s'
|
|
||||||
)
|
|
||||||
self.logger = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
# Load RSS URLs from file or environment
|
|
||||||
rss_urls_file = os.getenv("RSS_URLS_FILE")
|
|
||||||
if rss_urls_file and os.path.exists(rss_urls_file):
|
|
||||||
# Load from file
|
|
||||||
try:
|
|
||||||
with open(rss_urls_file, 'r') as f:
|
|
||||||
self.config["rss_urls"] = [
|
|
||||||
line.strip() for line in f
|
|
||||||
if line.strip() and not line.startswith('#')
|
|
||||||
]
|
|
||||||
self.logger.info(f"Loaded {len(self.config['rss_urls'])} RSS URLs from file: {rss_urls_file}")
|
|
||||||
except Exception as e:
|
|
||||||
self.logger.error(f"Could not load RSS URLs from file {rss_urls_file}: {e}")
|
|
||||||
self.config["rss_urls"] = []
|
|
||||||
else:
|
|
||||||
# Fallback to environment variable
|
|
||||||
self.config["rss_urls"] = [
|
|
||||||
url.strip() for url in os.getenv("RSS_URLS", "").split(",")
|
|
||||||
if url.strip()
|
|
||||||
]
|
|
||||||
if self.config["rss_urls"]:
|
|
||||||
self.logger.info(f"Loaded {len(self.config['rss_urls'])} RSS URLs from environment")
|
|
||||||
|
|
||||||
# Load processed URLs from persistent storage
|
|
||||||
self.processed_urls_file = "/app/data/processed_urls.json"
|
|
||||||
self.processed_urls = self.load_processed_urls()
|
|
||||||
|
|
||||||
# Statistics tracking
|
|
||||||
self.previous_instances = getattr(self, 'previous_instances', 0)
|
|
||||||
|
|
||||||
def load_processed_urls(self):
|
|
||||||
"""Load previously processed URLs and instance count from file"""
|
|
||||||
try:
|
|
||||||
if os.path.exists(self.processed_urls_file):
|
|
||||||
with open(self.processed_urls_file, 'r') as f:
|
|
||||||
data = json.load(f)
|
|
||||||
# Load previous instance count for statistics
|
|
||||||
self.previous_instances = data.get('previous_instances', 0)
|
|
||||||
return set(data.get('processed_urls', []))
|
|
||||||
except Exception as e:
|
|
||||||
self.logger.warning(f"Could not load processed URLs: {e}")
|
|
||||||
|
|
||||||
return set()
|
|
||||||
|
|
||||||
def save_processed_urls(self, current_instances=None):
|
|
||||||
"""Save processed URLs and current instance count to file"""
|
|
||||||
try:
|
|
||||||
os.makedirs(os.path.dirname(self.processed_urls_file), exist_ok=True)
|
|
||||||
data = {
|
|
||||||
'processed_urls': list(self.processed_urls),
|
|
||||||
'last_updated': time.time()
|
|
||||||
}
|
|
||||||
# Save current instance count for next run
|
|
||||||
if current_instances is not None and current_instances != 'unknown':
|
|
||||||
data['previous_instances'] = current_instances
|
|
||||||
|
|
||||||
with open(self.processed_urls_file, 'w') as f:
|
|
||||||
json.dump(data, f, indent=2)
|
|
||||||
except Exception as e:
|
|
||||||
self.logger.error(f"Could not save processed URLs: {e}")
|
|
||||||
|
|
||||||
def fetch_rss_urls(self, rss_url):
|
|
||||||
"""Fetch URLs from RSS feed"""
|
|
||||||
try:
|
|
||||||
self.logger.info(f"Fetching RSS feed: {rss_url}")
|
|
||||||
|
|
||||||
# Parse RSS feed
|
|
||||||
feed = feedparser.parse(rss_url)
|
|
||||||
|
|
||||||
if feed.bozo:
|
|
||||||
self.logger.warning(f"RSS feed may have issues: {rss_url}")
|
|
||||||
|
|
||||||
# Extract URLs from entries
|
|
||||||
urls = []
|
|
||||||
for entry in feed.entries:
|
|
||||||
if hasattr(entry, 'link'):
|
|
||||||
urls.append(entry.link)
|
|
||||||
|
|
||||||
self.logger.info(f"Found {len(urls)} URLs in RSS feed")
|
|
||||||
return urls
|
|
||||||
|
|
||||||
except Exception as e:
|
|
||||||
self.logger.error(f"Error fetching RSS feed {rss_url}: {e}")
|
|
||||||
return []
|
|
||||||
|
|
||||||
def lookup_post(self, post_url):
|
|
||||||
"""Look up a post URL using GTS search API"""
|
|
||||||
try:
|
|
||||||
# Prepare search API call
|
|
||||||
search_url = f"{self.config['server_url']}/api/v2/search"
|
|
||||||
params = {
|
|
||||||
'q': post_url,
|
|
||||||
'type': 'statuses',
|
|
||||||
'resolve': 'true',
|
|
||||||
'limit': 1
|
|
||||||
}
|
|
||||||
headers = {
|
|
||||||
'Authorization': f'Bearer {self.config["access_token"]}',
|
|
||||||
'Content-Type': 'application/json'
|
|
||||||
}
|
|
||||||
|
|
||||||
# Make API call
|
|
||||||
response = requests.get(
|
|
||||||
search_url,
|
|
||||||
params=params,
|
|
||||||
headers=headers,
|
|
||||||
timeout=30
|
|
||||||
)
|
|
||||||
|
|
||||||
if response.status_code == 200:
|
|
||||||
results = response.json()
|
|
||||||
if results.get('statuses') or results.get('accounts'):
|
|
||||||
self.logger.info(f"Successfully looked up: {post_url}")
|
|
||||||
return True
|
|
||||||
else:
|
|
||||||
self.logger.warning(f"No results for: {post_url}")
|
|
||||||
return False
|
|
||||||
else:
|
|
||||||
self.logger.error(f"API error {response.status_code} for {post_url}: {response.text}")
|
|
||||||
return False
|
|
||||||
|
|
||||||
except requests.exceptions.RequestException as e:
|
|
||||||
self.logger.error(f"Error looking up {post_url}: {e}")
|
|
||||||
return False
|
|
||||||
|
|
||||||
def process_feeds(self):
|
|
||||||
"""Process all configured RSS feeds"""
|
|
||||||
total_processed = 0
|
|
||||||
|
|
||||||
# Record start time for statistics
|
|
||||||
self.start_time = time.time()
|
|
||||||
|
|
||||||
# Ping healthcheck start
|
|
||||||
self.ping_healthcheck("/start")
|
|
||||||
|
|
||||||
try:
|
|
||||||
for rss_url in self.config["rss_urls"]:
|
|
||||||
if not rss_url.strip():
|
|
||||||
continue
|
|
||||||
|
|
||||||
self.logger.info(f"Processing feed: {rss_url}")
|
|
||||||
|
|
||||||
# Get URLs from RSS
|
|
||||||
urls = self.fetch_rss_urls(rss_url)
|
|
||||||
|
|
||||||
# Filter out already processed URLs
|
|
||||||
new_urls = [url for url in urls if url not in self.processed_urls]
|
|
||||||
|
|
||||||
if not new_urls:
|
|
||||||
self.logger.info("No new URLs to process")
|
|
||||||
continue
|
|
||||||
|
|
||||||
# Rate limiting: max posts per run
|
|
||||||
urls_to_process = new_urls[:self.config["max_posts_per_run"]]
|
|
||||||
|
|
||||||
self.logger.info(f"Processing {len(urls_to_process)} new URLs")
|
|
||||||
|
|
||||||
for url in urls_to_process:
|
|
||||||
if self.lookup_post(url):
|
|
||||||
self.processed_urls.add(url)
|
|
||||||
total_processed += 1
|
|
||||||
|
|
||||||
# Rate limiting: delay between requests
|
|
||||||
time.sleep(self.config["delay_between_requests"])
|
|
||||||
|
|
||||||
# Calculate runtime
|
|
||||||
end_time = time.time()
|
|
||||||
runtime_seconds = end_time - self.start_time
|
|
||||||
runtime_formatted = str(timedelta(seconds=int(runtime_seconds)))
|
|
||||||
|
|
||||||
# Get current instance count
|
|
||||||
try:
|
|
||||||
instance_info = requests.get(f"{self.config['server_url']}/api/v1/instance",
|
|
||||||
headers={'Authorization': f'Bearer {self.config["access_token"]}'},
|
|
||||||
timeout=10)
|
|
||||||
if instance_info.status_code == 200:
|
|
||||||
current_instances = instance_info.json().get('stats', {}).get('domain_count', 'unknown')
|
|
||||||
else:
|
|
||||||
current_instances = 'unknown'
|
|
||||||
except Exception as e:
|
|
||||||
self.logger.error(f"Failed to get instance count: {e}")
|
|
||||||
current_instances = 'unknown'
|
|
||||||
|
|
||||||
# Calculate new instances (if we have previous data)
|
|
||||||
new_instances = 'unknown'
|
|
||||||
if self.previous_instances > 0 and current_instances != 'unknown':
|
|
||||||
new_instances = current_instances - self.previous_instances
|
|
||||||
|
|
||||||
# Print comprehensive statistics
|
|
||||||
print(f"\n📊 GTS-HolMirDas Run Statistics:")
|
|
||||||
print(f" ⏱️ Runtime: {runtime_formatted}")
|
|
||||||
print(f" 📄 Total posts processed: {total_processed}")
|
|
||||||
print(f" 🌐 Current known instances: {current_instances}")
|
|
||||||
if new_instances != 'unknown' and new_instances > 0:
|
|
||||||
print(f" ➕ New instances discovered: +{new_instances}")
|
|
||||||
elif new_instances == 0:
|
|
||||||
print(f" ➕ New instances discovered: +0")
|
|
||||||
print(f" 📡 RSS feeds processed: {len(self.config['rss_urls'])}")
|
|
||||||
if runtime_seconds > 60:
|
|
||||||
print(f" ⚡ Posts per minute: {total_processed / (runtime_seconds / 60):.1f}")
|
|
||||||
|
|
||||||
self.save_processed_urls(current_instances)
|
|
||||||
|
|
||||||
# Ping healthcheck success
|
|
||||||
self.ping_healthcheck("")
|
|
||||||
|
|
||||||
except Exception as e:
|
|
||||||
self.logger.error(f"Error during processing: {e}")
|
|
||||||
# Ping healthcheck failure
|
|
||||||
self.ping_healthcheck("/fail")
|
|
||||||
raise
|
|
||||||
|
|
||||||
def ping_healthcheck(self, endpoint=""):
|
|
||||||
"""Ping healthchecks.io for monitoring"""
|
|
||||||
if not self.config.get("healthcheck_url"):
|
|
||||||
return
|
|
||||||
|
|
||||||
try:
|
|
||||||
url = self.config["healthcheck_url"] + endpoint
|
|
||||||
requests.get(url, timeout=10)
|
|
||||||
except Exception as e:
|
|
||||||
self.logger.warning(f"Failed to ping healthcheck: {e}")
|
|
||||||
|
|
||||||
def main():
|
|
||||||
"""Main entry point"""
|
|
||||||
try:
|
|
||||||
fetcher = GTSHolMirDas()
|
|
||||||
|
|
||||||
# Validate required config
|
|
||||||
if not fetcher.config["access_token"]:
|
|
||||||
raise ValueError("GTS_ACCESS_TOKEN environment variable is required")
|
|
||||||
|
|
||||||
fetcher.process_feeds()
|
|
||||||
|
|
||||||
except Exception as e:
|
|
||||||
logging.error(f"Fatal error: {e}")
|
|
||||||
raise
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
|
||||||
main()
|
|
|
@ -1,3 +0,0 @@
|
||||||
requests==2.31.0
|
|
||||||
feedparser==6.0.10
|
|
||||||
urllib3==2.0.7
|
|
|
@ -1,17 +0,0 @@
|
||||||
# Example RSS feeds - customize for your interests
|
|
||||||
|
|
||||||
# Add ?limit=X parameter to increase posts per feed (default: 20, max: 100)
|
|
||||||
# Higher limits = more content discovery, but longer processing time
|
|
||||||
# Performance tip: Start with limit=50, then increase to 100 if needed
|
|
||||||
|
|
||||||
# homelab (up to 100 posts per feed)
|
|
||||||
https://mastodon.social/tags/homelab.rss # 20 posts/feed (default)
|
|
||||||
https://fosstodon.org/tags/homelab.rss?limit=50 # 50 posts/feed
|
|
||||||
|
|
||||||
# selfhosting (up to 100 posts per feed)
|
|
||||||
https://mastodon.social/tags/selfhosting.rss?limit=100 # 100 posts/feed
|
|
||||||
https://infosec.exchange/tags/selfhosting.rss?limit=100 # 100 posts/feed
|
|
||||||
|
|
||||||
# docker (up to 100 posts per feed)
|
|
||||||
https://social.tchncs.de/tags/docker.rss?limit=100 # 100 posts/feed
|
|
||||||
https://fosstodon.org/tags/docker.rss?limit=100 # 100 posts/feed
|
|
Loading…
Add table
Add a link
Reference in a new issue