Where technology speaks, and you feel the echo
A comprehensive keyword tracking platform that monitors technology trends across Hacker News, Reddit, and YouTube. Analyze keyword popularity, track trends, and discover insights across multiple tech communities.
Demo
Click the image above to watch the full demo
Features
Real-time Analysis
- Multi-platform tracking: Hacker News, Reddit, YouTube
- Trend visualization: Monthly data trends with interactive charts
- Cross-platform insights: Compare keyword performance across platforms
- Top contributors: Identify key voices in each community
Data Sources
- Hacker News: 2-year historical dataset with stories and comments
- Reddit: All-time search across subreddits
- YouTube: Video metadata, views, and engagement metrics
- Discord: Community data (limited to accessible groups)
Key Metrics
- Total mentions and unique authors
- Monthly trend analysis
- Interaction counts (upvotes, comments, views)
- Platform-specific insights
🚀 Application Ports
BuzzScope consists of multiple independent Streamlit applications, each running on a different port:
Port 8501 - Historical Analysis
- Application:
app_simple_historical.py - Purpose: Historical data analysis and trend visualization
- Features:
- Search keywords across Hacker News, Reddit, YouTube, and Discord
- View historical trends and patterns
- Cross-platform comparison
- Pre-cached data for fast loading
- Access:
http://localhost:8501 - Run:
streamlit run app_simple_historical.py --server.port 8501
Port 8502 - New Keyword Testing
- Application:
app_new_keyword_test.py - Purpose: Test and analyze new keywords in real-time
- Features:
- Real-time data collection for new keywords
- Interactive trend analysis
- Platform-specific insights
- Access:
http://localhost:8502 - Run:
streamlit run app_new_keyword_test.py --server.port 8502
Port 8503 - Hot Post Monitor
- Application:
app_hot_post_monitor.py - Purpose: Real-time hot post detection and email notifications
- Features:
- Automatic detection of trending posts (last 24 hours)
- Email notifications with HTML formatting
- Configurable heat score thresholds
- Multi-keyword monitoring
- Access:
http://localhost:8503 - Run:
streamlit run app_hot_post_monitor.py --server.port 8503
Port 8506 - Multi-Agent Analytics
- Application:
app_multi_agent_analytics.py - Purpose: AI-powered collaborative analysis using multiple LLM agents
- Features:
- Multi-agent collaborative analysis
- Trend detection and insights
- Content generation for different platforms
- Parallel agent execution with convergence
- Access:
http://localhost:8506 - Run:
streamlit run app_multi_agent_analytics.py --server.port 8506
Note: All applications can run simultaneously on different ports. They are independent and share data through the cache directory.
Quick Start
Prerequisites
- Python 3.8+
- API keys for YouTube and OpenAI (Reddit uses public API, no key needed)
Installation
# Clone the repository git clone https://github.com/witch-Judy/BuzzScope.git cd BuzzScope # Install dependencies pip install -r requirements.txt # Configure environment (optional) cp env.example .env # Edit .env with your API keys
Run the Applications
# Historical Analysis (Port 8501) streamlit run app_simple_historical.py --server.port 8501 # New Keyword Testing (Port 8502) streamlit run app_new_keyword_test.py --server.port 8502 # Hot Post Monitor (Port 8503) streamlit run app_hot_post_monitor.py --server.port 8503 # Multi-Agent Analytics (Port 8506) streamlit run app_multi_agent_analytics.py --server.port 8506
Visit the respective URLs to access each application.
Usage
Pre-loaded Keywords
The application comes with pre-analyzed data for:
ai- Artificial Intelligence discussionsiot- Internet of Things topicsmqtt- MQTT protocol discussionsunified_namespace- Industrial automation concepts
New Keyword Analysis
- Enter keywords in the sidebar (comma-separated)
- Click "Collect Missing Data" for new keywords
- View real-time analysis and trends
- Explore cross-platform insights
Features Overview
- Trend Analysis: Monthly mention trends with raw data tables
- Platform Comparison: Side-by-side keyword performance
- Top Contributors: Most active users per platform
- Interactive Charts: Zoom, filter, and explore data
Data Collection
Hacker News
- Uses 2-year historical parquet dataset
- Analyzes stories, comments, and user interactions
- Covers 7.6M+ records from 2022-2024
- Searches across all subreddits
- Collects posts, comments, and metadata
- Uses Reddit's public JSON API (no authentication required)
YouTube
- Searches video titles and descriptions
- Collects view counts, likes, and comments
- Uses YouTube Data API v3
Architecture
BuzzScope/
├── app_new_keyword_test.py # Main Streamlit application
├── src/
│ ├── collectors/ # Data collection modules
│ ├── analyzers/ # Analysis engines
│ └── visualization/ # Chart generation
├── data/
│ ├── cache/ # Processed data cache
│ └── Hackernews_raw/ # Hacker News historical data
└── requirements.txt # Dependencies
API Setup
No API key required! Reddit uses a public JSON API that doesn't require authentication.
- Accessible via:
https://www.reddit.com/r/{subreddit}/new.json - No credentials needed
- Rate limit: ~60 requests per minute (automatically handled)
YouTube API
- Go to Google Cloud Console
- Enable YouTube Data API v3
- Create API key
- Add to
.env:
YOUTUBE_API_KEY=your_api_key
Hacker News Data
The application requires a 2-year Hacker News dataset for full functionality:
Option 1: Download from Google
- Search for "Hacker News 2 years dataset" on Google
- Download the parquet file to
./data/Hackernews_raw/
Option 2: Contact the Author
- The author can provide the dataset upon request
- Contact via GitHub issues or email
Option 3: Use API Only
- The application will work with real-time API data only
- Limited to recent posts and comments
Technical Details
Data Processing
- Caching: Pre-processed metrics for fast loading
- Charts: Pre-generated HTML charts for instant display
- Storage: JSON-based cache with Parquet for large datasets
- Performance: Optimized for real-time analysis
Supported Platforms
- Hacker News: Official API + 2-year historical dataset (7.6M+ records)
- Reddit: Public JSON API - No API key required! Uses
https://www.reddit.com/r/{subreddit}/new.json - YouTube: Data API v3 with quota management (requires API key)
- Discord: Community data (requires access to specific groups)
Contributing
- Fork the repository
- Create a feature branch
- Make your changes
- Submit a pull request
License
MIT License - see LICENSE file for details.
📖 Project Story: How I Built My Own Tech Radar (BuzzScope)
From Data Streams to Daily Insight
I'm always looking for better ways to keep up with what's happening in tech. So I built a small system — BuzzScope, my personal tech radar.
It currently has 3 key functions:
1️⃣ Historical Insight
Search any keyword, and BuzzScope finds the most popular posts and active contributors from multiple platforms (Hacker News, Reddit, YouTube, and even selected Discord servers).
Behind the scenes, historical data from HN and Discord are archived in Parquet format, which allows columnar access and faster queries across millions of records. Reddit and YouTube, on the other hand, are fetched on demand for real-time trending results.
2️⃣ Daily Pulse
It automatically gathers trending posts from the past 24 hours (based on thresholds) and sends them to my inbox — quick and actionable.
3️⃣ AI Analysis & Re-distribution
Then, multiple LLM agents collaboratively analyze the results to detect short-term trends and even generate tailored repost content for different platforms.
Each agent works in parallel rounds — seeing global context but only refining its own results — until the network "converges."
Support
For questions or issues:
- Create an issue on GitHub
- Check the troubleshooting section
- Review the documentation
Start tracking technology trends today
