Shiowo2's picture
Update README.md
5219c80 verified
---
license: odbl
language:
- id
tags:
- sports
- location
- lantitute
- longtitude
---
# This is The initial dataset we scraped from open maps
## this dataset has not been `cleaned` yet be aware!
```python
# requirements
!pip install requests
```
## script
```python
import csv
import time
import requests
from urllib.parse import quote
OUT_CSV = "jabodetabek_sports_osm.csv"
BBOX = (-6.80, 106.30, -5.90, 107.20)
OVERPASS_URL = "https://overpass-api.de/api/interpreter"
WIKIDATA_ENTITY_URL = "https://www.wikidata.org/wiki/Special:EntityData/{qid}.json"
FETCH_WIKIDATA_IMAGES = True
HEADERS = {"User-Agent": "jabodetabek-sports-scraper/1.0 (contact: yourname@example.com)"}
def osm_browse_link(osm_type: str, osm_id: int) -> str:
return f"https://www.openstreetmap.org/{osm_type}/{osm_id}"
def commons_file_url(filename: str, width: int = 1600) -> str:
fn = filename.strip()
if fn.lower().startswith("file:"):
fn = fn.split(":", 1)[1]
return f"https://commons.wikimedia.org/wiki/Special:FilePath/{quote(fn)}?width={width}"
def extract_image_link(tags: dict) -> str:
img = tags.get("image")
if img:
if img.startswith("http"):
return img
return commons_file_url(img)
wm = tags.get("wikimedia_commons")
if wm:
return commons_file_url(wm)
qid = tags.get("wikidata")
if FETCH_WIKIDATA_IMAGES and qid and qid.upper().startswith("Q"):
try:
r = requests.get(WIKIDATA_ENTITY_URL.format(qid=qid), headers=HEADERS, timeout=30)
if r.status_code == 200:
data = r.json()
ent = data.get("entities", {}).get(qid.upper(), {})
claims = ent.get("claims", {})
p18 = claims.get("P18", [])
if p18:
filename = p18[0]["mainsnak"]["datavalue"]["value"]
return commons_file_url(filename)
except Exception:
pass
return ""
def compose_address(tags: dict) -> str:
if "addr:full" in tags:
return tags["addr:full"]
parts = []
street = tags.get("addr:street")
houseno = tags.get("addr:housenumber")
if street and houseno:
parts.append(f"{street} {houseno}")
elif street:
parts.append(street)
for k in ("addr:neighbourhood", "addr:suburb", "addr:village"):
if tags.get(k):
parts.append(tags[k])
for k in ("addr:city", "addr:municipality", "addr:county"):
if tags.get(k):
parts.append(tags[k])
for k in ("addr:province", "addr:state"):
if tags.get(k):
parts.append(tags[k])
if tags.get("addr:postcode"):
parts.append(tags["addr:postcode"])
return ", ".join(parts)
def build_types(tags: dict) -> str:
bits = []
if "leisure" in tags:
bits.append(f"leisure:{tags['leisure']}")
if "amenity" in tags:
bits.append(f"amenity:{tags['amenity']}")
if "sport" in tags:
bits.append(f"sport:{tags['sport']}")
return ", ".join(bits)
def fetch_overpass(bbox):
s, w, n, e = bbox
leisure_regex = "^(sports_centre|fitness_centre|stadium|pitch|swimming_pool|track)$"
query = f"""
[out:json][timeout:180];
(
node["leisure"~"{leisure_regex}"]({s},{w},{n},{e});
way["leisure"~"{leisure_regex}"]({s},{w},{n},{e});
relation["leisure"~"{leisure_regex}"]({s},{w},{n},{e});
// Any feature explicitly tagged with sport=*, but avoid retail shops
node["sport"]["shop"!~".*"]({s},{w},{n},{e});
way["sport"]["shop"!~".*"]({s},{w},{n},{e});
relation["sport"]["shop"!~".*"]({s},{w},{n},{e});
);
out center tags;
"""
r = requests.post(OVERPASS_URL, data={"data": query}, headers=HEADERS, timeout=180)
r.raise_for_status()
return r.json().get("elements", [])
def element_coords(el) -> tuple[float, float]:
if el["type"] == "node":
return el.get("lat"), el.get("lon")
c = el.get("center") or {}
return c.get("lat"), c.get("lon")
def main():
elements = fetch_overpass(BBOX)
seen = set()
rows = []
for el in elements:
el_type = el.get("type")
el_id = el.get("id")
tags = el.get("tags", {}) or {}
key = (el_type, el_id)
if key in seen:
continue
seen.add(key)
lat, lon = element_coords(el)
if lat is None or lon is None:
continue
name = tags.get("name") or "(Unnamed)"
addr = compose_address(tags)
types = build_types(tags)
osm_link = osm_browse_link(el_type, el_id)
image_link = extract_image_link(tags)
likely_sporty = (
"leisure" in tags and tags["leisure"] in
{"sports_centre", "fitness_centre", "stadium", "pitch", "swimming_pool", "track"}
) or ("sport" in tags)
if not likely_sporty:
continue
rows.append({
"name": name,
"address": addr,
"lat": lat,
"lng": lon,
"types": types,
"osm_link": osm_link,
"image_link": image_link,
"osm_type": el_type,
"osm_id": el_id,
})
fieldnames = ["name", "address", "lat", "lng", "types", "osm_link", "image_link", "osm_type", "osm_id"]
with open(OUT_CSV, "w", newline="", encoding="utf-8") as f:
w = csv.DictWriter(f, fieldnames=fieldnames)
w.writeheader()
for row in rows:
w.writerow(row)
print(f"Saved {len(rows)} places to {OUT_CSV}")
if __name__ == "__main__":
main()
```