Files changed (3) hide show
  1. extract_odia_ocr_gemini.py +282 -0
  2. inference.ipynb +485 -0
  3. upload_to_hf.py +257 -0
extract_odia_ocr_gemini.py ADDED
@@ -0,0 +1,282 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """
2
+ Extract Odia OCR text from benchmark dataset images using Gemini.
3
+
4
+ This script:
5
+ 1) Reads images recursively from benchmark_dataset/images (or a custom directory)
6
+ 2) Sends each image to Gemini for OCR
7
+ 3) Appends each result row immediately to a CSV file to avoid losing progress
8
+ """
9
+
10
+ from __future__ import annotations
11
+
12
+ import argparse
13
+ import csv
14
+ import os
15
+ from pathlib import Path
16
+ from typing import Any, Iterable
17
+
18
+
19
+ DEFAULT_PROMPT = (
20
+ "You are an OCR assistant for Odia text.\n"
21
+ "Extract all visible Odia text from this image exactly as written.\n"
22
+ "Return only the extracted text, without translation or explanation."
23
+ )
24
+
25
+ SUPPORTED_EXTENSIONS = {".png", ".jpg", ".jpeg", ".webp", ".bmp", ".tiff", ".tif"}
26
+
27
+
28
+ def load_dotenv(dotenv_path: Path) -> dict[str, str]:
29
+ """Parse a simple .env file (KEY=VALUE lines)."""
30
+ values: dict[str, str] = {}
31
+ if not dotenv_path.exists():
32
+ return values
33
+
34
+ for raw_line in dotenv_path.read_text(encoding="utf-8").splitlines():
35
+ line = raw_line.strip()
36
+ if not line or line.startswith("#") or "=" not in line:
37
+ continue
38
+
39
+ key, value = line.split("=", 1)
40
+ key = key.strip()
41
+ value = value.strip().strip("'").strip('"')
42
+ if key:
43
+ values[key] = value
44
+
45
+ return values
46
+
47
+
48
+ def iter_image_paths(images_dir: Path) -> Iterable[Path]:
49
+ """Yield all supported image files under images_dir recursively."""
50
+ for path in sorted(images_dir.rglob("*")):
51
+ if path.is_file() and path.suffix.lower() in SUPPORTED_EXTENSIONS:
52
+ yield path
53
+
54
+
55
+ def call_gemini_ocr(
56
+ image_path: Path,
57
+ client: Any,
58
+ model: str,
59
+ prompt: str,
60
+ ) -> str:
61
+ """Call Gemini with prompt + image using official google-genai SDK."""
62
+ try:
63
+ from PIL import Image
64
+ except ImportError as exc:
65
+ raise RuntimeError("Missing dependency: pillow. Install with `pip install pillow`.") from exc
66
+
67
+ image = Image.open(image_path).convert("RGB")
68
+ response = client.models.generate_content(
69
+ model=model,
70
+ contents=[prompt, image],
71
+ )
72
+ output_text = (response.text or "").strip()
73
+ if not output_text:
74
+ raise RuntimeError("Empty OCR output in Gemini response")
75
+ return output_text
76
+
77
+
78
+ def normalize_stored_path(path_str: str, project_root: Path) -> str:
79
+ """Normalize CSV image_path for stable matching and dedup."""
80
+ raw = str(path_str).strip()
81
+ if not raw:
82
+ return ""
83
+ p = Path(raw)
84
+ if p.is_absolute():
85
+ try:
86
+ return str(p.resolve().relative_to(project_root))
87
+ except ValueError:
88
+ return str(p.resolve())
89
+ return raw
90
+
91
+
92
+ def load_existing_rows_by_path(output_csv: Path, project_root: Path) -> dict[str, dict[str, str]]:
93
+ """Load CSV rows keyed by normalized image path (latest row wins)."""
94
+ rows_by_path: dict[str, dict[str, str]] = {}
95
+ if not output_csv.exists():
96
+ return rows_by_path
97
+
98
+ with output_csv.open("r", encoding="utf-8", newline="") as f:
99
+ reader = csv.DictReader(f)
100
+ for row in reader:
101
+ key = normalize_stored_path(row.get("image_path", ""), project_root)
102
+ if not key:
103
+ continue
104
+ rows_by_path[key] = {
105
+ "image_path": key,
106
+ "extracted_odia_text": row.get("extracted_odia_text", "") or "",
107
+ "status": row.get("status", "") or "",
108
+ "error": row.get("error", "") or "",
109
+ }
110
+ return rows_by_path
111
+
112
+
113
+ def image_path_key(image_path: Path, project_root: Path) -> str:
114
+ """Use project-relative path for CSV storage and deduplication."""
115
+ resolved = image_path.resolve()
116
+ try:
117
+ return str(resolved.relative_to(project_root))
118
+ except ValueError:
119
+ return str(resolved)
120
+
121
+
122
+ def ensure_output_header(output_csv: Path, append_mode: bool) -> None:
123
+ """Ensure CSV header exists when creating a new output file."""
124
+ output_csv.parent.mkdir(parents=True, exist_ok=True)
125
+ if append_mode and output_csv.exists():
126
+ return
127
+ with output_csv.open("w", encoding="utf-8", newline="") as f:
128
+ writer = csv.writer(f)
129
+ writer.writerow(["image_path", "extracted_odia_text", "status", "error"])
130
+
131
+
132
+ def write_rows(output_csv: Path, rows_by_path: dict[str, dict[str, str]]) -> None:
133
+ """Rewrite CSV from rows map to keep one row per image path."""
134
+ output_csv.parent.mkdir(parents=True, exist_ok=True)
135
+ with output_csv.open("w", encoding="utf-8", newline="") as f:
136
+ writer = csv.DictWriter(
137
+ f,
138
+ fieldnames=["image_path", "extracted_odia_text", "status", "error"],
139
+ )
140
+ writer.writeheader()
141
+ writer.writerows(rows_by_path.values())
142
+ f.flush()
143
+
144
+
145
+ def main() -> None:
146
+ project_root = Path(__file__).parent.parent
147
+ dotenv_values = load_dotenv(project_root / ".env")
148
+
149
+ default_images_dir = (
150
+ dotenv_values.get("IMAGE_FOLDER_PATH")
151
+ or str(project_root / "benchmark_dataset" / "images")
152
+ )
153
+ default_output_csv = (
154
+ dotenv_values.get("OUTPUT_CSV_PATH")
155
+ or str(project_root / "benchmark_dataset" / "gemini_ocr_output.csv")
156
+ )
157
+ default_api_key = dotenv_values.get("GEMINI_API_KEY") or os.getenv(
158
+ "GEMINI_API_KEY", ""
159
+ )
160
+
161
+ parser = argparse.ArgumentParser(
162
+ description="Extract Odia OCR text from benchmark images using Gemini"
163
+ )
164
+ parser.add_argument(
165
+ "--model",
166
+ type=str,
167
+ default="gemini-3-flash-preview",
168
+ help="Gemini model name",
169
+ )
170
+ parser.add_argument(
171
+ "--prompt",
172
+ type=str,
173
+ default=DEFAULT_PROMPT,
174
+ help="Prompt used for OCR extraction",
175
+ )
176
+ parser.add_argument(
177
+ "--limit",
178
+ type=int,
179
+ default=None,
180
+ help="Optional max number of images to process",
181
+ )
182
+ parser.add_argument(
183
+ "--no-resume",
184
+ action="store_true",
185
+ help="Do not skip already processed image paths in output CSV",
186
+ )
187
+ args = parser.parse_args()
188
+
189
+ if not default_api_key:
190
+ raise ValueError(
191
+ "Gemini API key missing. Set GEMINI_API_KEY in .env or environment."
192
+ )
193
+
194
+ try:
195
+ from google import genai
196
+ except ImportError as exc:
197
+ raise RuntimeError(
198
+ "Missing dependency: google-genai. Install with `pip install google-genai`."
199
+ ) from exc
200
+
201
+ client = genai.Client(api_key=default_api_key)
202
+
203
+ images_dir = Path(default_images_dir).resolve()
204
+ output_csv = Path(default_output_csv).resolve()
205
+
206
+ if not images_dir.exists():
207
+ raise FileNotFoundError(f"Images directory not found: {images_dir}")
208
+
209
+ all_images = list(iter_image_paths(images_dir))
210
+ if args.limit is not None:
211
+ all_images = all_images[: max(args.limit, 0)]
212
+
213
+ if not all_images:
214
+ print(f"No images found under: {images_dir}")
215
+ return
216
+
217
+ rows_by_path: dict[str, dict[str, str]] = {}
218
+ processed_success_paths: set[str] = set()
219
+ previous_error_rows = 0
220
+ if not args.no_resume:
221
+ rows_by_path = load_existing_rows_by_path(output_csv, project_root)
222
+ processed_success_paths = {
223
+ p for p, row in rows_by_path.items() if (row.get("status", "").strip().lower() == "ok")
224
+ }
225
+ previous_error_rows = sum(
226
+ 1 for row in rows_by_path.values() if row.get("status", "").strip().lower() == "error"
227
+ )
228
+ else:
229
+ ensure_output_header(output_csv, append_mode=False)
230
+
231
+ # Deduplicate/normalize existing CSV content on each resume run.
232
+ if not args.no_resume and output_csv.exists():
233
+ write_rows(output_csv, rows_by_path)
234
+
235
+ existing_keys = set(processed_success_paths)
236
+
237
+ to_process = [p for p in all_images if image_path_key(p, project_root) not in existing_keys]
238
+ total = len(to_process)
239
+ if total == 0:
240
+ print("No new images to process. Output CSV is already up to date.")
241
+ return
242
+
243
+ print(f"Found {len(all_images)} images in total")
244
+ print(f"Already processed successfully: {len(processed_success_paths)}")
245
+ if previous_error_rows:
246
+ print(f"Previous error rows available to retry: {previous_error_rows}")
247
+ print(f"Processing now: {total}")
248
+ print(f"Writing incremental results to: {output_csv}")
249
+ for idx, image_path in enumerate(to_process, start=1):
250
+ image_str = image_path_key(image_path, project_root)
251
+ status = "ok"
252
+ extracted_text = ""
253
+ err = ""
254
+
255
+ try:
256
+ extracted_text = call_gemini_ocr(
257
+ image_path=image_path,
258
+ client=client,
259
+ model=args.model,
260
+ prompt=args.prompt,
261
+ )
262
+ except Exception as exc: # noqa: BLE001
263
+ status = "error"
264
+ err = str(exc)
265
+
266
+ # Upsert: keep a single latest row per image path.
267
+ rows_by_path[image_str] = {
268
+ "image_path": image_str,
269
+ "extracted_odia_text": extracted_text,
270
+ "status": status,
271
+ "error": err,
272
+ }
273
+ write_rows(output_csv, rows_by_path)
274
+
275
+ print(f"[{idx}/{total}] {status}: {image_str}")
276
+
277
+ print("\nDone.")
278
+ print(f"Final CSV: {output_csv}")
279
+
280
+
281
+ if __name__ == "__main__":
282
+ main()
inference.ipynb ADDED
@@ -0,0 +1,485 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cells": [
3
+ {
4
+ "cell_type": "code",
5
+ "execution_count": 3,
6
+ "id": "384230de",
7
+ "metadata": {},
8
+ "outputs": [
9
+ {
10
+ "name": "stdout",
11
+ "output_type": "stream",
12
+ "text": [
13
+ "[INFO] Loading dataset...\n",
14
+ "\n",
15
+ "[MODEL] OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3\n",
16
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 0\n",
17
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 1\n",
18
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 2\n",
19
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 3\n",
20
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 4\n",
21
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 5\n",
22
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 6\n",
23
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 7\n",
24
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 8\n",
25
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 9\n",
26
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 10\n",
27
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 11\n",
28
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 12\n",
29
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 13\n",
30
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 14\n",
31
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 15\n",
32
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 16\n",
33
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 17\n",
34
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 18\n",
35
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 19\n",
36
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 20\n",
37
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 21\n",
38
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 22\n",
39
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 23\n",
40
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 24\n",
41
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 25\n",
42
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 26\n",
43
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 27\n",
44
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 28\n",
45
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 29\n",
46
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 30\n",
47
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 31\n",
48
+ "✗ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 32 failed: 400 Client Error: Bad Request for url: http://117.18.102.44:34537/v1/chat/completions\n",
49
+ "✗ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 33 failed: 400 Client Error: Bad Request for url: http://117.18.102.44:34537/v1/chat/completions\n",
50
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 34\n",
51
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 35\n",
52
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 36\n",
53
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 37\n",
54
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 38\n",
55
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 39\n",
56
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 40\n",
57
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 41\n",
58
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 42\n",
59
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 43\n",
60
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 44\n",
61
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 45\n",
62
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 46\n",
63
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 47\n",
64
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 48\n",
65
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 49\n",
66
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 50\n",
67
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 51\n",
68
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 52\n",
69
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 53\n",
70
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 54\n",
71
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 55\n",
72
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 56\n",
73
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 57\n",
74
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 58\n",
75
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 59\n",
76
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 60\n",
77
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 61\n",
78
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 62\n",
79
+ "✗ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 63 failed: 400 Client Error: Bad Request for url: http://117.18.102.44:34537/v1/chat/completions\n",
80
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 64\n",
81
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 65\n",
82
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 66\n",
83
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 67\n",
84
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 68\n",
85
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 69\n",
86
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 70\n",
87
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 71\n",
88
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 72\n",
89
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 73\n",
90
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 74\n",
91
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 75\n",
92
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 76\n",
93
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 77\n",
94
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 78\n",
95
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 79\n",
96
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 80\n",
97
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 81\n",
98
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 82\n",
99
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 83\n",
100
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 84\n",
101
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 85\n",
102
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 86\n",
103
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 87\n",
104
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 88\n",
105
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 89\n",
106
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 90\n",
107
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 91\n",
108
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 92\n",
109
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 93\n",
110
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 94\n",
111
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 95\n",
112
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 96\n",
113
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 97\n",
114
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 98\n",
115
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 99\n",
116
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 100\n",
117
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 101\n",
118
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 102\n",
119
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 103\n",
120
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 104\n",
121
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 105\n",
122
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 106\n",
123
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 107\n",
124
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 108\n",
125
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 109\n",
126
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 110\n",
127
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 111\n",
128
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 112\n",
129
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 113\n",
130
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 114\n",
131
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 115\n",
132
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 116\n",
133
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 117\n",
134
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 118\n",
135
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 119\n",
136
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 120\n",
137
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 121\n",
138
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 122\n",
139
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 123\n",
140
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 124\n",
141
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 125\n",
142
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 126\n",
143
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 127\n",
144
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 128\n",
145
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 129\n",
146
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 130\n",
147
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 131\n",
148
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 132\n",
149
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 133\n",
150
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 134\n",
151
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 135\n",
152
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 136\n",
153
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 137\n",
154
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 138\n",
155
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 139\n",
156
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 140\n",
157
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 141\n",
158
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 142\n",
159
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 143\n",
160
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 144\n",
161
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 145\n",
162
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 146\n",
163
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 147\n",
164
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 148\n",
165
+ "✗ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 149 failed: 400 Client Error: Bad Request for url: http://117.18.102.44:34537/v1/chat/completions\n",
166
+ "✗ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 150 failed: 400 Client Error: Bad Request for url: http://117.18.102.44:34537/v1/chat/completions\n",
167
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 151\n",
168
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 152\n",
169
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 153\n",
170
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 154\n",
171
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 155\n",
172
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 156\n",
173
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 157\n",
174
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 158\n",
175
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 159\n",
176
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 160\n",
177
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 161\n",
178
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 162\n",
179
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 163\n",
180
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 164\n",
181
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 165\n",
182
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 166\n",
183
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 167\n",
184
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 168\n",
185
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 169\n",
186
+ "✗ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 170 failed: 400 Client Error: Bad Request for url: http://117.18.102.44:34537/v1/chat/completions\n",
187
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 171\n",
188
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 172\n",
189
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 173\n",
190
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 174\n",
191
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 175\n",
192
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 176\n",
193
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 177\n",
194
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 178\n",
195
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 179\n",
196
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 180\n",
197
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 181\n",
198
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 182\n",
199
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 183\n",
200
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 184\n",
201
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 185\n",
202
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 186\n",
203
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 187\n",
204
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 188\n",
205
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 189\n",
206
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 190\n",
207
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 191\n",
208
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 192\n",
209
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 193\n",
210
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 194\n",
211
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 195\n",
212
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 196\n",
213
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 197\n",
214
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 198\n",
215
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 199\n",
216
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 200\n",
217
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 201\n",
218
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 202\n",
219
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 203\n",
220
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 204\n",
221
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 205\n",
222
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 206\n",
223
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 207\n",
224
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 208\n",
225
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 209\n",
226
+ "✗ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 210 failed: ('Connection aborted.', TimeoutError('timed out'))\n",
227
+ "✗ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 211 failed: ('Connection aborted.', TimeoutError('timed out'))\n",
228
+ "✗ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 212 failed: ('Connection aborted.', TimeoutError('timed out'))\n",
229
+ "✗ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 213 failed: ('Connection aborted.', TimeoutError('timed out'))\n",
230
+ "✗ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 214 failed: ('Connection aborted.', TimeoutError('timed out'))\n",
231
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 215\n",
232
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 216\n",
233
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 217\n",
234
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 218\n",
235
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 219\n",
236
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 220\n",
237
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 221\n",
238
+ "✓ OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3 | sample 222\n"
239
+ ]
240
+ }
241
+ ],
242
+ "source": [
243
+ "import os\n",
244
+ "import re\n",
245
+ "import time\n",
246
+ "import base64\n",
247
+ "import requests\n",
248
+ "import pandas as pd\n",
249
+ "\n",
250
+ "from datasets import load_dataset\n",
251
+ "from PIL import Image\n",
252
+ "from io import BytesIO\n",
253
+ "from rapidfuzz.distance import Levenshtein\n",
254
+ "\n",
255
+ "\n",
256
+ "# -----------------------------------------\n",
257
+ "# Server Config\n",
258
+ "# -----------------------------------------\n",
259
+ "SERVER_IP = \"117.18.102.44\"\n",
260
+ "PORT = 34537\n",
261
+ "\n",
262
+ "MODELS = [\n",
263
+ " \"OdiaGenAIOCR/odia-ocr-qwen-finetuned_v3\"\n",
264
+ "]\n",
265
+ "\n",
266
+ "PROMPT = \"Transcribe all the Odia text from this image exactly as it appears. Return only the extracted odia text, nothing else.\"\n",
267
+ "\n",
268
+ "OUTPUT_CSV = \"ocr_benchmark_results.csv\"\n",
269
+ "\n",
270
+ "\n",
271
+ "# -----------------------------------------\n",
272
+ "# Image -> Base64\n",
273
+ "# -----------------------------------------\n",
274
+ "def encode_image(image):\n",
275
+ " buffer = BytesIO()\n",
276
+ " image.save(buffer, format=\"PNG\")\n",
277
+ " return base64.b64encode(buffer.getvalue()).decode()\n",
278
+ "\n",
279
+ "\n",
280
+ "# -----------------------------------------\n",
281
+ "# Metrics\n",
282
+ "# -----------------------------------------\n",
283
+ "def compute_cer(gt, pred):\n",
284
+ " gt = str(gt)\n",
285
+ " pred = str(pred)\n",
286
+ " if len(gt) == 0:\n",
287
+ " return 0.0\n",
288
+ " return Levenshtein.distance(gt, pred) / len(gt)\n",
289
+ "\n",
290
+ "\n",
291
+ "def has_value(value):\n",
292
+ " return pd.notna(value) and str(value).strip() != \"\"\n",
293
+ "\n",
294
+ "\n",
295
+ "# -----------------------------------------\n",
296
+ "# OCR Server Inference\n",
297
+ "# -----------------------------------------\n",
298
+ "def ocr_predict(image, model):\n",
299
+ " url = f\"http://{SERVER_IP}:{PORT}/v1/chat/completions\"\n",
300
+ " img_base64 = encode_image(image)\n",
301
+ "\n",
302
+ " payload = {\n",
303
+ " \"model\": model,\n",
304
+ " \"temperature\": 0,\n",
305
+ " \"max_tokens\": 1024,\n",
306
+ " \"messages\": [\n",
307
+ " {\n",
308
+ " \"role\": \"user\",\n",
309
+ " \"content\": [\n",
310
+ " {\n",
311
+ " \"type\": \"image_url\",\n",
312
+ " \"image_url\": {\n",
313
+ " \"url\": f\"data:image/png;base64,{img_base64}\"\n",
314
+ " },\n",
315
+ " },\n",
316
+ " {\n",
317
+ " \"type\": \"text\",\n",
318
+ " \"text\": PROMPT,\n",
319
+ " },\n",
320
+ " ],\n",
321
+ " }\n",
322
+ " ],\n",
323
+ " }\n",
324
+ "\n",
325
+ " resp = requests.post(\n",
326
+ " url,\n",
327
+ " headers={\n",
328
+ " \"Content-Type\": \"application/json\",\n",
329
+ " \"Authorization\": \"Bearer EMPTY\",\n",
330
+ " },\n",
331
+ " json=payload,\n",
332
+ " timeout=180,\n",
333
+ " )\n",
334
+ " resp.raise_for_status()\n",
335
+ " return resp.json()[\"choices\"][0][\"message\"][\"content\"].strip()\n",
336
+ "\n",
337
+ "\n",
338
+ "# -----------------------------------------\n",
339
+ "# Results table helpers\n",
340
+ "# -----------------------------------------\n",
341
+ "def model_slug(model_name: str) -> str:\n",
342
+ " return re.sub(r\"[^a-zA-Z0-9]+\", \"_\", model_name).strip(\"_\").lower()\n",
343
+ "\n",
344
+ "\n",
345
+ "def model_pred_col(model_name: str) -> str:\n",
346
+ " return f\"{model_slug(model_name)}_pred\"\n",
347
+ "\n",
348
+ "\n",
349
+ "def model_cer_col(model_name: str) -> str:\n",
350
+ " return f\"{model_slug(model_name)}_cer\"\n",
351
+ "\n",
352
+ "\n",
353
+ "def init_or_load_results(ds):\n",
354
+ " base_df = pd.DataFrame(\n",
355
+ " {\n",
356
+ " \"id\": list(range(len(ds))),\n",
357
+ " \"ground_truth\": [row[\"ground_truth\"] for row in ds],\n",
358
+ " }\n",
359
+ " )\n",
360
+ "\n",
361
+ " if os.path.exists(OUTPUT_CSV):\n",
362
+ " existing = pd.read_csv(OUTPUT_CSV)\n",
363
+ " if \"id\" not in existing.columns:\n",
364
+ " existing = base_df.copy()\n",
365
+ " else:\n",
366
+ " for col in [\"id\", \"ground_truth\"]:\n",
367
+ " if col not in existing.columns:\n",
368
+ " existing[col] = base_df[col]\n",
369
+ " existing = existing.sort_values(\"id\", kind=\"stable\")\n",
370
+ " existing = existing.drop_duplicates(subset=[\"id\"], keep=\"last\")\n",
371
+ " existing = base_df[[\"id\", \"ground_truth\"]].merge(\n",
372
+ " existing.drop(columns=[\"ground_truth\"], errors=\"ignore\"),\n",
373
+ " on=\"id\",\n",
374
+ " how=\"left\",\n",
375
+ " )\n",
376
+ " return existing\n",
377
+ "\n",
378
+ " return base_df\n",
379
+ "\n",
380
+ "\n",
381
+ "def save_results(df):\n",
382
+ " df = df.sort_values(\"id\", kind=\"stable\")\n",
383
+ " df.to_csv(OUTPUT_CSV, index=False, encoding=\"utf-8\")\n",
384
+ "\n",
385
+ "\n",
386
+ "# -----------------------------------------\n",
387
+ "# Benchmark Pipeline\n",
388
+ "# -----------------------------------------\n",
389
+ "def run_benchmark(\n",
390
+ " hf_repo=\"Iftesha/odia-ocr-benchmark\",\n",
391
+ " split=\"train\",\n",
392
+ " max_rows=None,\n",
393
+ " skip_existing=True,\n",
394
+ "):\n",
395
+ " print(\"[INFO] Loading dataset...\")\n",
396
+ " ds = load_dataset(hf_repo, split=split)\n",
397
+ "\n",
398
+ " if max_rows:\n",
399
+ " ds = ds.select(range(max_rows))\n",
400
+ "\n",
401
+ " results_df = init_or_load_results(ds)\n",
402
+ "\n",
403
+ " for model in MODELS:\n",
404
+ " pred_col = model_pred_col(model)\n",
405
+ " cer_col = model_cer_col(model)\n",
406
+ "\n",
407
+ " if pred_col not in results_df.columns:\n",
408
+ " results_df[pred_col] = \"\"\n",
409
+ " if cer_col not in results_df.columns:\n",
410
+ " results_df[cer_col] = \"\"\n",
411
+ "\n",
412
+ " print(f\"\\n[MODEL] {model}\")\n",
413
+ " print(f\"[COLUMNS] {pred_col}, {cer_col}\")\n",
414
+ "\n",
415
+ " for i, row in enumerate(ds):\n",
416
+ " pred_done = has_value(results_df.at[i, pred_col])\n",
417
+ " cer_done = has_value(results_df.at[i, cer_col])\n",
418
+ " if skip_existing and pred_done and cer_done:\n",
419
+ " print(f\"↷ {model} | sample {i} already exists\")\n",
420
+ " continue\n",
421
+ "\n",
422
+ " try:\n",
423
+ " image = row[\"image\"]\n",
424
+ " if not isinstance(image, Image.Image):\n",
425
+ " image = image.convert(\"RGB\")\n",
426
+ "\n",
427
+ " gt = row[\"ground_truth\"]\n",
428
+ " pred = ocr_predict(image, model)\n",
429
+ " cer = compute_cer(gt, pred)\n",
430
+ "\n",
431
+ " results_df.at[i, pred_col] = pred\n",
432
+ " results_df.at[i, cer_col] = cer\n",
433
+ " print(f\"✓ {model} | sample {i}\")\n",
434
+ "\n",
435
+ " except Exception as e:\n",
436
+ " print(f\"✗ {model} | sample {i} failed: {e}\")\n",
437
+ " # Keep empty value for failed inference so retries can fill it later.\n",
438
+ " results_df.at[i, pred_col] = \"\"\n",
439
+ " results_df.at[i, cer_col] = \"\"\n",
440
+ " time.sleep(1)\n",
441
+ "\n",
442
+ " # Persist after each sample for long-running jobs.\n",
443
+ " save_results(results_df)\n",
444
+ "\n",
445
+ " save_results(results_df)\n",
446
+ " print(f\"\\n[INFO] Saved comparison CSV to {OUTPUT_CSV}\")\n",
447
+ "\n",
448
+ "\n",
449
+ "# -----------------------------------------\n",
450
+ "# Run Benchmark\n",
451
+ "# -----------------------------------------\n",
452
+ "run_benchmark()"
453
+ ]
454
+ },
455
+ {
456
+ "cell_type": "code",
457
+ "execution_count": null,
458
+ "id": "3b3e360e",
459
+ "metadata": {},
460
+ "outputs": [],
461
+ "source": []
462
+ }
463
+ ],
464
+ "metadata": {
465
+ "kernelspec": {
466
+ "display_name": "Python (vlm-poetry)",
467
+ "language": "python",
468
+ "name": "vlm-poetry"
469
+ },
470
+ "language_info": {
471
+ "codemirror_mode": {
472
+ "name": "ipython",
473
+ "version": 3
474
+ },
475
+ "file_extension": ".py",
476
+ "mimetype": "text/x-python",
477
+ "name": "python",
478
+ "nbconvert_exporter": "python",
479
+ "pygments_lexer": "ipython3",
480
+ "version": "3.14.2"
481
+ }
482
+ },
483
+ "nbformat": 4,
484
+ "nbformat_minor": 5
485
+ }
upload_to_hf.py ADDED
@@ -0,0 +1,257 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """
2
+ Upload Odia OCR Benchmark Dataset to HuggingFace Hub
3
+
4
+ Converts local images + metadata.csv to HuggingFace Dataset format and pushes.
5
+ """
6
+
7
+ import argparse
8
+ from pathlib import Path
9
+
10
+ from datasets import Dataset, Features, Value, Image
11
+ from huggingface_hub import HfApi
12
+ import pandas as pd
13
+
14
+
15
+ BENCHMARK_DIR = Path(__file__).parent.parent / "benchmark_dataset"
16
+ CSV_PATH = BENCHMARK_DIR / "final_hf.csv"
17
+
18
+
19
+ def resolve_image_path(raw_path: str) -> Path:
20
+ """Resolve image paths from metadata across common path styles."""
21
+ p = Path(str(raw_path).strip())
22
+
23
+ # 1) Already absolute
24
+ if p.is_absolute():
25
+ return p
26
+
27
+ # 2) Relative to project root: benchmark_dataset/images/...
28
+ if p.parts and p.parts[0] == "benchmark_dataset":
29
+ return (BENCHMARK_DIR.parent / p).resolve()
30
+
31
+ # 3) Relative to benchmark dir: images/...
32
+ return (BENCHMARK_DIR / p).resolve()
33
+
34
+
35
+ def load_local_dataset() -> Dataset:
36
+ """Load local images and metadata into a HuggingFace Dataset."""
37
+ print(f"Loading metadata from {CSV_PATH}...")
38
+
39
+ if not CSV_PATH.exists():
40
+ raise FileNotFoundError(f"Metadata CSV not found: {CSV_PATH}")
41
+
42
+ df = pd.read_csv(CSV_PATH)
43
+ print(f"Found {len(df)} samples in metadata")
44
+
45
+ # Normalize image paths, supporting:
46
+ # - images/...
47
+ # - benchmark_dataset/images/...
48
+ # - absolute paths
49
+ df["image_path"] = df["image_path"].apply(lambda p: str(resolve_image_path(p)))
50
+
51
+ # Verify images exist
52
+ missing = []
53
+ for idx, row in df.iterrows():
54
+ if not Path(row["image_path"]).exists():
55
+ missing.append(row["image_path"])
56
+
57
+ if missing:
58
+ print(f"Warning: {len(missing)} images not found:")
59
+ for p in missing[:5]:
60
+ print(f" - {p}")
61
+ if len(missing) > 5:
62
+ print(f" ... and {len(missing) - 5} more")
63
+
64
+ # Filter out missing images
65
+ df = df[df["image_path"].apply(lambda p: Path(p).exists())]
66
+ print(f"Continuing with {len(df)} valid samples")
67
+
68
+ if "id" not in df.columns:
69
+ raise ValueError(
70
+ f"Required column 'id' not found in {CSV_PATH}. "
71
+ "Please add an 'id' column before upload."
72
+ )
73
+
74
+ # Create dataset with Image feature
75
+ features = Features({
76
+ "id": Value("int64"),
77
+ "image": Image(),
78
+ "ground_truth": Value("string"),
79
+ "category": Value("string"),
80
+
81
+ })
82
+
83
+ # Rename image_path to image for HF Dataset
84
+ data = {
85
+ "id": df["id"].tolist(),
86
+ "image": df["image_path"].tolist(),
87
+ "ground_truth": df["ground_truth"].tolist(),
88
+ "category": df["category"].tolist(),
89
+
90
+ }
91
+
92
+ dataset = Dataset.from_dict(data, features=features)
93
+ print(f"Created HuggingFace Dataset with {len(dataset)} samples")
94
+
95
+ return dataset
96
+
97
+
98
+ def push_to_hub(dataset: Dataset, repo_id: str, private: bool = False):
99
+ """Push dataset to HuggingFace Hub."""
100
+ print(f"\nPushing to HuggingFace Hub: {repo_id}")
101
+ print(f"Private: {private}")
102
+
103
+ dataset.push_to_hub(
104
+ repo_id,
105
+ private=private,
106
+ commit_message="Upload Odia OCR benchmark dataset",
107
+ )
108
+
109
+ print(f"\nDataset uploaded to: https://huggingface.co/datasets/{repo_id}")
110
+
111
+
112
+ def push_dataset_card(repo_id: str, card_content: str):
113
+ """Upload dataset card as README.md to HuggingFace Hub."""
114
+ api = HfApi()
115
+ api.upload_file(
116
+ path_or_fileobj=card_content.encode("utf-8"),
117
+ path_in_repo="README.md",
118
+ repo_id=repo_id,
119
+ repo_type="dataset",
120
+ commit_message="Add dataset card README",
121
+ )
122
+ print(f"Dataset card uploaded: https://huggingface.co/datasets/{repo_id}/blob/main/README.md")
123
+
124
+
125
+ def create_dataset_card(repo_id: str):
126
+ """Create a dataset card (README.md) for HuggingFace."""
127
+ card_content = f"""---
128
+ license: cc-by-4.0
129
+ task_categories:
130
+ - image-to-text
131
+ language:
132
+ - or
133
+ tags:
134
+ - ocr
135
+ - odia
136
+ - oriya
137
+ - indic
138
+ - benchmark
139
+ size_categories:
140
+ - n<1K
141
+ ---
142
+
143
+ # Odia OCR Benchmark Dataset
144
+
145
+ ## Description
146
+
147
+ A curated benchmark dataset for evaluating OCR models on Odia (Oriya) text recognition.
148
+ Contains handwritten, printed, scene text, newspaper, books, and digital categories,
149
+ including both short samples and long-text examples for OCR evaluation.
150
+
151
+ ## Dataset Structure
152
+
153
+ - **id**: Unique identifier for each sample
154
+ - **image**: The input image (PIL Image)
155
+ - **ground_truth**: The correct Odia text transcription
156
+ - **category**: Type of text (handwritten, printed, scene_text, newspaper, books, digital)
157
+
158
+
159
+
160
+ ## Usage
161
+
162
+ ```python
163
+ from datasets import load_dataset
164
+
165
+ dataset = load_dataset("{repo_id}")
166
+
167
+ # Access a sample
168
+ sample = dataset["train"][0]
169
+ sample_id = sample["id"]
170
+ image = sample["image"]
171
+ text = sample["ground_truth"]
172
+ ```
173
+
174
+ ## Categories
175
+
176
+ | Category | Description |
177
+ | ------------- | ----------------------------------------------------- |
178
+ | handwritten | Handwritten Odia text (word/short phrase level) |
179
+ | printed | Printed/typed Odia text |
180
+ | scene_text | Text in natural scenes (signboards, posters, etc.) |
181
+ | newspaper | Odia newspaper clippings (including long text) |
182
+ | books | Scanned Odia book pages (including long text) |
183
+ | digital | Screenshots from Odia digital content |
184
+
185
+ ## Sources
186
+
187
+ - `OdiaGenAIOCR/odia-ocr-merged` (handwritten)
188
+ - `darknight054/indic-mozhi-ocr` with config `oriya` (printed)
189
+ - `darknight054/indicstr12-crops` with config `odia` (scene_text)
190
+ - `newspaper`: Odia newspaper scans/clippings
191
+ - `books`: Odia book page images
192
+ - `digital`: odia digital content
193
+
194
+ ## Notes
195
+
196
+ - Includes long-text samples for paragraph-level OCR evaluation.
197
+ - The `source` field records origin for each sample.
198
+
199
+ ## License
200
+
201
+ CC-BY-4.0
202
+ """
203
+ return card_content
204
+
205
+
206
+ def main():
207
+ parser = argparse.ArgumentParser(
208
+ description="Upload Odia OCR benchmark dataset to HuggingFace Hub"
209
+ )
210
+ parser.add_argument(
211
+ "--repo",
212
+ type=str,
213
+ required=True,
214
+ help="HuggingFace repo ID (e.g., 'username/odia-ocr-benchmark')",
215
+ )
216
+ parser.add_argument(
217
+ "--private",
218
+ action="store_true",
219
+ help="Make the dataset private",
220
+ )
221
+ parser.add_argument(
222
+ "--dry-run",
223
+ action="store_true",
224
+ help="Load and validate dataset without uploading",
225
+ )
226
+
227
+ args = parser.parse_args()
228
+
229
+ print("=" * 60)
230
+ print("Upload Odia OCR Benchmark to HuggingFace")
231
+ print("=" * 60)
232
+
233
+ # Load local dataset
234
+ dataset = load_local_dataset()
235
+
236
+ # Show sample
237
+ print("\nSample from dataset:")
238
+ sample = dataset[0]
239
+ print(f" id: {sample['id']}")
240
+ print(f" ground_truth: {sample['ground_truth']}")
241
+ print(f" category: {sample['category']}")
242
+
243
+
244
+ if args.dry_run:
245
+ print("\n[DRY RUN] Dataset validated. Not uploading.")
246
+ return
247
+
248
+ # Push to hub
249
+ push_to_hub(dataset, args.repo, private=args.private)
250
+
251
+ # Push dataset card
252
+ card_content = create_dataset_card(args.repo)
253
+ push_dataset_card(args.repo, card_content)
254
+
255
+
256
+ if __name__ == "__main__":
257
+ main()