| | from __future__ import annotations |
| |
|
| | import argparse |
| | import os |
| | from pathlib import Path |
| | from typing import Dict, List, Tuple |
| |
|
| | |
| | try: |
| | from dotenv import load_dotenv |
| | load_dotenv() |
| | except Exception: |
| | pass |
| |
|
| | import pandas as pd |
| | from datasets import Dataset, DatasetDict, Image, Features, Value |
| | from huggingface_hub import HfApi |
| | from huggingface_hub.errors import HfHubHTTPError |
| |
|
| |
|
| | def _find_splits(data_dir: Path) -> List[Tuple[str, Path]]: |
| | out = [] |
| | for split in ("train", "validation", "test"): |
| | sd = data_dir / split |
| | if (sd / "metadata.csv").exists(): |
| | out.append((split, sd)) |
| | return out |
| |
|
| |
|
| | def _normalize_columns(df: pd.DataFrame) -> pd.DataFrame: |
| | |
| | lower_map = {c.lower(): c for c in df.columns} |
| |
|
| | |
| | img_col_candidates = [ |
| | "image", "file_name", "filename", "path", "filepath", "file" |
| | ] |
| | img_col = next((lower_map[c] for c in img_col_candidates if c in lower_map), None) |
| | if img_col is None: |
| | raise ValueError(f"Could not find an image path column among: {df.columns.tolist()}") |
| |
|
| | |
| | lat_candidates = ["latitude", "lat", "Latitude", "LAT"] |
| | lat_col = None |
| | for c in lat_candidates: |
| | if c.lower() in lower_map: |
| | lat_col = lower_map[c.lower()] |
| | break |
| | if lat_col is None: |
| | raise ValueError("Latitude column not found (expected one of latitude/lat)") |
| |
|
| | |
| | lon_candidates = ["longitude", "lon", "Longitude", "LON", "long"] |
| | lon_col = None |
| | for c in lon_candidates: |
| | if c.lower() in lower_map: |
| | lon_col = lower_map[c.lower()] |
| | break |
| | if lon_col is None: |
| | raise ValueError("Longitude column not found (expected one of longitude/lon/long)") |
| |
|
| | out_df = pd.DataFrame({ |
| | "image": df[img_col].astype(str), |
| | "latitude": pd.to_numeric(df[lat_col], errors="coerce"), |
| | "longitude": pd.to_numeric(df[lon_col], errors="coerce"), |
| | }) |
| | out_df = out_df.dropna(subset=["latitude", "longitude"]).reset_index(drop=True) |
| | return out_df |
| |
|
| |
|
| | def _resolve_paths(df: pd.DataFrame, split_dir: Path) -> pd.DataFrame: |
| | paths = [] |
| | for p in df["image"].tolist(): |
| | pth = Path(p) |
| | if pth.is_absolute() and pth.exists(): |
| | paths.append(str(pth)) |
| | continue |
| | |
| | pth2 = (split_dir / p).resolve() |
| | if pth2.exists(): |
| | paths.append(str(pth2)) |
| | continue |
| | |
| | |
| | paths.append(str(p)) |
| | df = df.copy() |
| | df["image"] = paths |
| | return df |
| |
|
| |
|
| | def build_datasetdict(data_dir: Path) -> DatasetDict: |
| | splits = _find_splits(data_dir) |
| | if not splits: |
| | raise SystemExit(f"No splits found under {data_dir}. Expected metadata.csv in train/validation/test.") |
| |
|
| | feats = Features({ |
| | "image": Image(), |
| | "latitude": Value("float64"), |
| | "longitude": Value("float64"), |
| | }) |
| |
|
| | dd: Dict[str, Dataset] = {} |
| | for split, sd in splits: |
| | csv_path = sd / "metadata.csv" |
| | df = pd.read_csv(csv_path) |
| | df = _normalize_columns(df) |
| | df = _resolve_paths(df, sd) |
| |
|
| | ds = Dataset.from_dict(df.to_dict(orient="list"), features=feats) |
| | dd[split] = ds |
| | print(f"Split {split}: {len(ds)} rows") |
| |
|
| | return DatasetDict(dd) |
| |
|
| |
|
| | def push_to_hub(ds: DatasetDict, repo_id: str, private: bool, max_shard_size: str) -> None: |
| | |
| | token = os.environ.get("HUGGINGFACE_HUB_TOKEN") or os.environ.get("HF_TOKEN") |
| | if not token: |
| | print("[auth] 未检测到 Token。请在 .env 设置 HUGGINGFACE_HUB_TOKEN=hf_xxx,或设置环境变量 HF_TOKEN/HUGGINGFACE_HUB_TOKEN。") |
| | print("[auth] 也可以先运行: python -c \"from huggingface_hub import login; login('hf_xxx')\"") |
| | try: |
| | api = HfApi(token=token) |
| | api.create_repo(repo_id=repo_id, repo_type="dataset", exist_ok=True, private=private) |
| | ds.push_to_hub(repo_id, private=private, max_shard_size=max_shard_size, token=token) |
| | print(f"Pushed to https://huggingface.co/datasets/{repo_id}") |
| | except HfHubHTTPError as e: |
| | if hasattr(e, "response") and getattr(e.response, "status_code", None) == 401: |
| | print("[auth] 401 Unauthorized:请检查 Token 是否有效、是否具备 write 权限、是否属于 LarryD123 账号。") |
| | print("[auth] 建议:\n - 在 https://huggingface.co/settings/tokens 重新生成 write Token\n - 将其写入项目根目录 .env (HUGGINGFACE_HUB_TOKEN=hf_xxx)\n - 重新运行上传命令") |
| | raise |
| |
|
| |
|
| | def main(): |
| | ap = argparse.ArgumentParser(description="Build and push a 3-column Image+GPS dataset to Hugging Face.") |
| | ap.add_argument("--data-dir", type=Path, required=True, help="Folder containing split subfolders (train/validation/test)") |
| | ap.add_argument("--repo-id", type=str, required=False, help="<user>/<dataset_name> on Hugging Face") |
| | ap.add_argument("--private", type=str, default="false", help="true/false for private dataset") |
| | ap.add_argument("--max-shard-size", type=str, default="500MB", help="Shard size for HF push") |
| | ap.add_argument("--dry-run", action="store_true", help="Build locally without pushing to Hub") |
| | args = ap.parse_args() |
| |
|
| | ds = build_datasetdict(args.data_dir) |
| | print(ds) |
| |
|
| | if args.dry_run: |
| | print("Dry run: not pushing to hub.") |
| | return |
| |
|
| | if not args.repo_id: |
| | raise SystemExit("--repo-id is required unless --dry-run is set") |
| |
|
| | private = str(args.private).lower() in ("1", "true", "yes", "y") |
| | push_to_hub(ds, args.repo_id, private=private, max_shard_size=args.max_shard_size) |
| |
|
| |
|
| | if __name__ == "__main__": |
| | main() |
| |
|