|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
""" |
|
|
Deduplicate MCP clients by name and capabilities. |
|
|
|
|
|
This script takes the existing mcp-clients dataset and creates a deduplicated view |
|
|
where rows are unique by (name, capabilities). When multiple versions exist for |
|
|
the same (name, capabilities), they are concatenated with "," in a versions column. |
|
|
|
|
|
It also strips " (via mcp-remote X.Y.Z)" suffixes from client names and reports |
|
|
the mcp-remote versions encountered. |
|
|
|
|
|
Usage: |
|
|
# Push deduplicated data to the same repo (evalstate/mcp-clients) under 'deduplicated' split |
|
|
uv run deduplicate_clients.py --push-to-hub |
|
|
|
|
|
# Read from raw, push to dedup split in same repo |
|
|
uv run deduplicate_clients.py --source-split raw --target-split deduplicated --push-to-hub |
|
|
|
|
|
# Output to local file |
|
|
uv run deduplicate_clients.py -o deduplicated.ndjson |
|
|
""" |
|
|
|
|
|
import sys |
|
|
import json |
|
|
import re |
|
|
import argparse |
|
|
from datetime import datetime |
|
|
from collections import defaultdict |
|
|
|
|
|
from datasets import Dataset, Features, Value, load_dataset |
|
|
|
|
|
|
|
|
|
|
|
MCP_REMOTE_PATTERN = re.compile(r'\s*\(via mcp-remote [\d.]+\)') |
|
|
|
|
|
|
|
|
def strip_mcp_remote_suffix(name): |
|
|
"""Remove the '(via mcp-remote X.Y.Z)' suffix from client name.""" |
|
|
match = MCP_REMOTE_PATTERN.search(name) |
|
|
if match: |
|
|
|
|
|
suffix = match.group(0) |
|
|
version_match = re.search(r'mcp-remote ([\d.]+)', suffix) |
|
|
if version_match: |
|
|
version = version_match.group(1) |
|
|
return MCP_REMOTE_PATTERN.sub('', name).strip(), version |
|
|
return name, None |
|
|
|
|
|
|
|
|
def normalize_capabilities(caps): |
|
|
"""Normalize capabilities for comparison.""" |
|
|
if caps is None: |
|
|
return None |
|
|
if isinstance(caps, str): |
|
|
if caps == '{}': |
|
|
return {} |
|
|
try: |
|
|
return json.loads(caps) |
|
|
except Exception: |
|
|
return caps |
|
|
return caps |
|
|
|
|
|
|
|
|
def capabilities_to_string(caps): |
|
|
"""Convert capabilities to JSON string for storage.""" |
|
|
if caps is None: |
|
|
return '' |
|
|
if isinstance(caps, dict): |
|
|
return json.dumps(caps, sort_keys=True) |
|
|
return str(caps) |
|
|
|
|
|
|
|
|
def create_dataset_from_deduplicated(deduplicated_list, features=None): |
|
|
"""Create a Hugging Face Dataset from the deduplicated list.""" |
|
|
if features is None: |
|
|
|
|
|
features = Features({ |
|
|
'name': Value('string'), |
|
|
'versions': Value('string'), |
|
|
'capabilities': Value('string'), |
|
|
'last_seen': Value('string'), |
|
|
}) |
|
|
|
|
|
records = [] |
|
|
for client in deduplicated_list: |
|
|
records.append({ |
|
|
'name': client['name'], |
|
|
'versions': client['versions'], |
|
|
'capabilities': capabilities_to_string(client['capabilities']), |
|
|
'last_seen': client['last_seen'], |
|
|
}) |
|
|
|
|
|
return Dataset.from_list(records, features=features) |
|
|
|
|
|
|
|
|
def push_to_hub(deduplicated_list, repo_id, split=None, token=None, private=False): |
|
|
"""Push the deduplicated dataset to Hugging Face Hub.""" |
|
|
dataset = create_dataset_from_deduplicated(deduplicated_list) |
|
|
|
|
|
if split: |
|
|
print(f"Pushing dataset to https://huggingface.co/datasets/{repo_id} (config: deduplicated, split: {split})", file=sys.stderr) |
|
|
else: |
|
|
print(f"Pushing dataset to https://huggingface.co/datasets/{repo_id} (config: deduplicated)", file=sys.stderr) |
|
|
|
|
|
|
|
|
|
|
|
dataset.push_to_hub( |
|
|
repo_id=repo_id, |
|
|
config_name="deduplicated", |
|
|
split=split, |
|
|
token=token, |
|
|
private=private, |
|
|
commit_message=f"Update deduplicated clients ({datetime.now().strftime('%Y-%m-%d %H:%M')})", |
|
|
) |
|
|
|
|
|
print(f"Successfully pushed {len(deduplicated_list):,} deduplicated records to {repo_id}", file=sys.stderr) |
|
|
|
|
|
|
|
|
def main(): |
|
|
parser = argparse.ArgumentParser( |
|
|
description="Deduplicate MCP clients by name and capabilities", |
|
|
formatter_class=argparse.RawDescriptionHelpFormatter, |
|
|
epilog=""" |
|
|
Examples: |
|
|
# Push deduplicated data to same repo under 'deduplicated' split |
|
|
%(prog)s --push-to-hub |
|
|
|
|
|
# Use custom source/target splits |
|
|
%(prog)s --source-split raw --target-split dedup --push-to-hub |
|
|
|
|
|
# Output to local file |
|
|
%(prog)s -o deduplicated.ndjson |
|
|
|
|
|
# Push to a different repo |
|
|
%(prog)s --target-repo my-org/mcp-clients-dedup --push-to-hub |
|
|
""" |
|
|
) |
|
|
parser.add_argument("-o", "--output", help="Output file path (default: stdout)") |
|
|
parser.add_argument("--source-repo", default="evalstate/mcp-clients", |
|
|
help="Source repository ID (default: evalstate/mcp-clients)") |
|
|
parser.add_argument("--source-config", default="raw", |
|
|
help="Source config name (default: raw)") |
|
|
parser.add_argument("--source-split", default=None, |
|
|
help="Source split name (optional, defaults to config name)") |
|
|
parser.add_argument("--target-repo", default="evalstate/mcp-clients", |
|
|
help="Target repository ID (default: evalstate/mcp-clients)") |
|
|
parser.add_argument("--target-split", default="deduplicated", |
|
|
help="Target split name (default: deduplicated)") |
|
|
parser.add_argument("--push-to-hub", action="store_true", |
|
|
help="Push the resulting dataset to Hugging Face Hub") |
|
|
parser.add_argument("--token", default=None, |
|
|
help="HF token (defaults to HF_TOKEN env var or cached token)") |
|
|
parser.add_argument("--private", action="store_true", |
|
|
help="Create/push to a private repository") |
|
|
args = parser.parse_args() |
|
|
|
|
|
|
|
|
source_split = args.source_split or args.source_config |
|
|
|
|
|
print(f"[{datetime.now().isoformat()}] Loading dataset from {args.source_repo} (config: {args.source_config}, split: {source_split})...", file=sys.stderr) |
|
|
|
|
|
|
|
|
ds = load_dataset(args.source_repo, args.source_config, split=source_split) |
|
|
|
|
|
print(f"[{datetime.now().isoformat()}] Loaded {len(ds):,} client records", file=sys.stderr) |
|
|
|
|
|
|
|
|
mcp_remote_versions = defaultdict(int) |
|
|
|
|
|
|
|
|
groups = defaultdict(list) |
|
|
|
|
|
for row in ds: |
|
|
|
|
|
clean_name, mcp_version = strip_mcp_remote_suffix(row['name']) |
|
|
|
|
|
if mcp_version: |
|
|
mcp_remote_versions[mcp_version] += 1 |
|
|
|
|
|
|
|
|
caps = normalize_capabilities(row['capabilities']) |
|
|
caps_key = capabilities_to_string(caps) |
|
|
|
|
|
key = (clean_name, caps_key) |
|
|
groups[key].append({ |
|
|
'name': clean_name, |
|
|
'version': row['version'], |
|
|
'capabilities': caps, |
|
|
'last_seen': row['last_seen'], |
|
|
}) |
|
|
|
|
|
print(f"[{datetime.now().isoformat()}] Found {len(groups):,} unique (name, capabilities) combinations", file=sys.stderr) |
|
|
|
|
|
|
|
|
if mcp_remote_versions: |
|
|
total_with_mcp_remote = sum(mcp_remote_versions.values()) |
|
|
print(f"[{datetime.now().isoformat()}] MCP-remote version summary:", file=sys.stderr) |
|
|
print(f" Total records with mcp-remote suffix: {total_with_mcp_remote:,}", file=sys.stderr) |
|
|
print(f" Unique mcp-remote versions: {len(mcp_remote_versions)}", file=sys.stderr) |
|
|
print(f" Versions encountered:", file=sys.stderr) |
|
|
for version in sorted(mcp_remote_versions.keys()): |
|
|
print(f" - mcp-remote {version}: {mcp_remote_versions[version]:,} records", file=sys.stderr) |
|
|
print() |
|
|
|
|
|
|
|
|
deduplicated = [] |
|
|
|
|
|
for (name, caps_key), records in groups.items(): |
|
|
|
|
|
latest_record = max(records, key=lambda x: x['last_seen']) |
|
|
|
|
|
|
|
|
versions = sorted(set(r['version'] for r in records)) |
|
|
versions_str = ','.join(versions) |
|
|
|
|
|
deduplicated.append({ |
|
|
'name': name, |
|
|
'versions': versions_str, |
|
|
'capabilities': latest_record['capabilities'], |
|
|
'last_seen': latest_record['last_seen'], |
|
|
}) |
|
|
|
|
|
|
|
|
deduplicated.sort(key=lambda x: x['last_seen'], reverse=True) |
|
|
|
|
|
print(f"[{datetime.now().isoformat()}] Created {len(deduplicated):,} deduplicated records", file=sys.stderr) |
|
|
|
|
|
|
|
|
if args.push_to_hub: |
|
|
push_to_hub(deduplicated, args.target_repo, split=args.target_split, token=args.token, private=args.private) |
|
|
return |
|
|
|
|
|
|
|
|
if args.output: |
|
|
out_file = open(args.output, 'w') |
|
|
else: |
|
|
out_file = sys.stdout |
|
|
|
|
|
|
|
|
for client in deduplicated: |
|
|
out_file.write(json.dumps(client) + '\n') |
|
|
|
|
|
if args.output: |
|
|
out_file.close() |
|
|
print(f"Output written to: {args.output}", file=sys.stderr) |
|
|
|
|
|
|
|
|
if __name__ == '__main__': |
|
|
main() |
|
|
|