Datasets:
File size: 5,631 Bytes
4609f08 bd2e7b8 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 5219c80 4609f08 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 |
---
license: odbl
language:
- id
tags:
- sports
- location
- lantitute
- longtitude
---
# This is The initial dataset we scraped from open maps
## this dataset has not been `cleaned` yet be aware!
```python
# requirements
!pip install requests
```
## script
```python
import csv
import time
import requests
from urllib.parse import quote
OUT_CSV = "jabodetabek_sports_osm.csv"
BBOX = (-6.80, 106.30, -5.90, 107.20)
OVERPASS_URL = "https://overpass-api.de/api/interpreter"
WIKIDATA_ENTITY_URL = "https://www.wikidata.org/wiki/Special:EntityData/{qid}.json"
FETCH_WIKIDATA_IMAGES = True
HEADERS = {"User-Agent": "jabodetabek-sports-scraper/1.0 (contact: yourname@example.com)"}
def osm_browse_link(osm_type: str, osm_id: int) -> str:
return f"https://www.openstreetmap.org/{osm_type}/{osm_id}"
def commons_file_url(filename: str, width: int = 1600) -> str:
fn = filename.strip()
if fn.lower().startswith("file:"):
fn = fn.split(":", 1)[1]
return f"https://commons.wikimedia.org/wiki/Special:FilePath/{quote(fn)}?width={width}"
def extract_image_link(tags: dict) -> str:
img = tags.get("image")
if img:
if img.startswith("http"):
return img
return commons_file_url(img)
wm = tags.get("wikimedia_commons")
if wm:
return commons_file_url(wm)
qid = tags.get("wikidata")
if FETCH_WIKIDATA_IMAGES and qid and qid.upper().startswith("Q"):
try:
r = requests.get(WIKIDATA_ENTITY_URL.format(qid=qid), headers=HEADERS, timeout=30)
if r.status_code == 200:
data = r.json()
ent = data.get("entities", {}).get(qid.upper(), {})
claims = ent.get("claims", {})
p18 = claims.get("P18", [])
if p18:
filename = p18[0]["mainsnak"]["datavalue"]["value"]
return commons_file_url(filename)
except Exception:
pass
return ""
def compose_address(tags: dict) -> str:
if "addr:full" in tags:
return tags["addr:full"]
parts = []
street = tags.get("addr:street")
houseno = tags.get("addr:housenumber")
if street and houseno:
parts.append(f"{street} {houseno}")
elif street:
parts.append(street)
for k in ("addr:neighbourhood", "addr:suburb", "addr:village"):
if tags.get(k):
parts.append(tags[k])
for k in ("addr:city", "addr:municipality", "addr:county"):
if tags.get(k):
parts.append(tags[k])
for k in ("addr:province", "addr:state"):
if tags.get(k):
parts.append(tags[k])
if tags.get("addr:postcode"):
parts.append(tags["addr:postcode"])
return ", ".join(parts)
def build_types(tags: dict) -> str:
bits = []
if "leisure" in tags:
bits.append(f"leisure:{tags['leisure']}")
if "amenity" in tags:
bits.append(f"amenity:{tags['amenity']}")
if "sport" in tags:
bits.append(f"sport:{tags['sport']}")
return ", ".join(bits)
def fetch_overpass(bbox):
s, w, n, e = bbox
leisure_regex = "^(sports_centre|fitness_centre|stadium|pitch|swimming_pool|track)$"
query = f"""
[out:json][timeout:180];
(
node["leisure"~"{leisure_regex}"]({s},{w},{n},{e});
way["leisure"~"{leisure_regex}"]({s},{w},{n},{e});
relation["leisure"~"{leisure_regex}"]({s},{w},{n},{e});
// Any feature explicitly tagged with sport=*, but avoid retail shops
node["sport"]["shop"!~".*"]({s},{w},{n},{e});
way["sport"]["shop"!~".*"]({s},{w},{n},{e});
relation["sport"]["shop"!~".*"]({s},{w},{n},{e});
);
out center tags;
"""
r = requests.post(OVERPASS_URL, data={"data": query}, headers=HEADERS, timeout=180)
r.raise_for_status()
return r.json().get("elements", [])
def element_coords(el) -> tuple[float, float]:
if el["type"] == "node":
return el.get("lat"), el.get("lon")
c = el.get("center") or {}
return c.get("lat"), c.get("lon")
def main():
elements = fetch_overpass(BBOX)
seen = set()
rows = []
for el in elements:
el_type = el.get("type")
el_id = el.get("id")
tags = el.get("tags", {}) or {}
key = (el_type, el_id)
if key in seen:
continue
seen.add(key)
lat, lon = element_coords(el)
if lat is None or lon is None:
continue
name = tags.get("name") or "(Unnamed)"
addr = compose_address(tags)
types = build_types(tags)
osm_link = osm_browse_link(el_type, el_id)
image_link = extract_image_link(tags)
likely_sporty = (
"leisure" in tags and tags["leisure"] in
{"sports_centre", "fitness_centre", "stadium", "pitch", "swimming_pool", "track"}
) or ("sport" in tags)
if not likely_sporty:
continue
rows.append({
"name": name,
"address": addr,
"lat": lat,
"lng": lon,
"types": types,
"osm_link": osm_link,
"image_link": image_link,
"osm_type": el_type,
"osm_id": el_id,
})
fieldnames = ["name", "address", "lat", "lng", "types", "osm_link", "image_link", "osm_type", "osm_id"]
with open(OUT_CSV, "w", newline="", encoding="utf-8") as f:
w = csv.DictWriter(f, fieldnames=fieldnames)
w.writeheader()
for row in rows:
w.writerow(row)
print(f"Saved {len(rows)} places to {OUT_CSV}")
if __name__ == "__main__":
main()
``` |