Macrocosmos MCP
Using Macrocosmos MCP with Claude Desktop or Cursor
Macrocosmos MCP
Using Macrocosmos MCP with Claude Desktop or Cursor
Macrocosmos MCP (Model Context Protocol) allows you to integrate with Data Universe APIs directly into Claude for Desktop, Cursor, or your custom LLM pipeline. Query X (Twitter) and Reddit data on demand from your AI environment!
Prerequisites
Python 3.10+
uvpackage managerClaude Desktop or Cursor installed
Install UV Package Manager
curl -LsSf https://astral.sh/uv/install.sh | shOr via pip:
pip3 install uvQuickstart
Get your API key from Macrocosmos. There is a free tier with $5 of credits to start.
Install
uvusing the command above or see the uv repo for additional install methods.
Configure Claude Desktop
Run the following command to open your Claude configuration file:
Update with this configuration:
Open Claude Desktop and look for the hammer icon — this confirms your MCP server is running. You'll now have SN13 tools available inside Claude.
Configure Cursor
Option 1: Via UI (Recommended)
Go to Cursor Settings
Navigate to MCP settings and select Add New Global MCP Server
Enter the configuration details
Option 2: Manual JSON
Add the same configuration:
⚠️ Note: In some cases, manually editing this file doesn't activate the MCP server in Cursor. If this happens, use the UI method above for best results.
Use Agent Mode
In Cursor, make sure you're using Agent Mode in the chat. Agents have the ability to use any MCP tool — including custom ones and those from SN13.
Available Tools
Quick Query Tool
query_on_demand_data - Real-time Social Media Queries
Fetch real-time data from X (Twitter) and Reddit. Best for quick queries up to 1,000 results.
source
string
REQUIRED. Platform: 'X' or 'REDDIT' (case-sensitive)
usernames
list
Up to 5 usernames. For X: @ is optional. Not available for Reddit
keywords
list
Up to 5 keywords/hashtags. For Reddit: subreddit names (e.g., 'r/MachineLearning')
start_date
string
ISO format (e.g., '2024-01-01T00:00:00Z'). Defaults to 24h ago
end_date
string
ISO format. Defaults to now
limit
int
Max results 1-1000. Default: 10
keyword_mode
string
'any' (default) or 'all' for strict matching
Example prompts:
"What has @elonmusk been posting about today?"
"Get me the latest posts from r/bittensor about dTAO"
"Fetch 50 tweets about #AI from the last week"
Large-Scale Collection Tools (Gravity)
Use Gravity tools when you need to collect large datasets over 7 days (more than 1,000 results).
create_gravity_task - Start 7-Day Data Collection
tasks
list
REQUIRED. List of task objects
name
string
Optional name for the task
email
string
Email for notification when complete
Task object structure:
⚠️ Important: For X (Twitter), topics MUST start with
#or$(e.g.,#ai,$BTC). Plain keywords are rejected!
get_gravity_task_status - Monitor Collection Progress
gravity_task_id
string
REQUIRED. The task ID from create_gravity_task
include_crawlers
bool
Include detailed stats per crawler. Default: True
Returns: Task status, crawler IDs, records_collected, bytes_collected
build_dataset - Build Dataset from Collected Data
crawler_id
string
REQUIRED. Get from get_gravity_task_status
max_rows
int
Max rows to include. Default: 10,000
email
string
Email for notification when ready
⚠️ Warning: Building a dataset will STOP the crawler and de-register it from the network.
get_dataset_status - Get Download Links
dataset_id
string
REQUIRED. The dataset ID from build_dataset
Returns: Build status, download URLs for Parquet files when complete
cancel_gravity_task - Stop Data Collection
gravity_task_id
string
REQUIRED. The task ID to cancel
cancel_dataset - Cancel Build or Purge Dataset
dataset_id
string
REQUIRED. The dataset ID to cancel/purge
Example Workflows
Quick Query (On-Demand)
Large Dataset Collection (Gravity)
Example Prompts
On-Demand Queries
"What has the president of the U.S. been saying over the past week on X?"
"Fetch me information about what people are posting on r/politics today"
"Please analyze posts from @elonmusk for the last week"
"Get me 100 tweets about #Bittensor and analyze the sentiment"
Large-Scale Collection
"Create a gravity task to collect data about #AI from Twitter"
"Start a 7-day collection of $BTC tweets with keyword 'ETF'"
"Check how many records my gravity task has collected"
"Build a dataset with 10,000 rows from my crawler"
Supported Platforms
X (Twitter)
✅ Yes
✅ Yes
N/A
❌ No
✅ Yes
✅ Yes
Troubleshooting
If you encounter any issues:
Ensure you're using Python 3.10+
Verify uv is installed: Run
uv --versionCheck your API key: Ensure
MC_APIis set correctlyRestart the application: After config changes, restart Claude Desktop or Cursor
For more on MCPs, refer to the official MCP documentation.
Last updated
