|
|
--- |
|
|
dataset_info: |
|
|
features: |
|
|
- name: id |
|
|
dtype: string |
|
|
- name: difficulty |
|
|
dtype: string |
|
|
- name: code |
|
|
dtype: string |
|
|
- name: render_light |
|
|
dtype: image |
|
|
- name: render_dark |
|
|
dtype: image |
|
|
- name: photo |
|
|
dtype: image |
|
|
splits: |
|
|
- name: easy |
|
|
num_bytes: 3082543834 |
|
|
num_examples: 700 |
|
|
- name: medium |
|
|
num_bytes: 902595780 |
|
|
num_examples: 200 |
|
|
- name: hard |
|
|
num_bytes: 500128560 |
|
|
num_examples: 100 |
|
|
download_size: 4481644051 |
|
|
dataset_size: 4485268174 |
|
|
configs: |
|
|
- config_name: default |
|
|
data_files: |
|
|
- split: easy |
|
|
path: data/easy-* |
|
|
- split: medium |
|
|
path: data/medium-* |
|
|
- split: hard |
|
|
path: data/hard-* |
|
|
license: mit |
|
|
task_categories: |
|
|
- image-to-text |
|
|
- text-generation |
|
|
tags: |
|
|
- code |
|
|
- ocr |
|
|
pretty_name: CodeOCR |
|
|
size_categories: |
|
|
- 1K<n<10K |
|
|
--- |
|
|
|
|
|
--- |
|
|
pretty_name: "CodeOCR Dataset (Python Code Images + Ground Truth)" |
|
|
license: mit |
|
|
language: |
|
|
- en |
|
|
task_categories: |
|
|
- image-to-text |
|
|
tags: |
|
|
- ocr |
|
|
- code |
|
|
- python |
|
|
- leetcode |
|
|
- synthetic |
|
|
- computer-vision |
|
|
size_categories: |
|
|
- 1K<n<10K |
|
|
--- |
|
|
|
|
|
# CodeOCR Dataset (Python Code Images + Ground Truth) |
|
|
|
|
|
This dataset is designed for **Optical Character Recognition (OCR) of source code**. |
|
|
Each example pairs **Python code (ground-truth text)** with **image renderings** of that code (light/dark themes) and a **real photo**. |
|
|
|
|
|
## Dataset Summary |
|
|
|
|
|
- **Language:** Python (text ground truth), images of code |
|
|
- **Splits:** `easy`, `medium`, `hard` |
|
|
- **Total examples:** 1,000 |
|
|
- `easy`: 700 |
|
|
- `medium`: 200 |
|
|
- `hard`: 100 |
|
|
- **Modalities:** image + text |
|
|
|
|
|
### What is “ground truth” here? |
|
|
|
|
|
The `code` field is **exactly the content of `gt.py`** used to generate the synthetic renderings. |
|
|
During dataset creation, code is normalized to ensure stable GT properties: |
|
|
|
|
|
- UTF-8 encoding |
|
|
- newline normalization to **LF (`\n`)** |
|
|
- tabs expanded to **4 spaces** |
|
|
- syntax checked with Python `compile()` (syntax/indentation correctness) |
|
|
|
|
|
This makes the dataset suitable for training/evaluating OCR models that output **plain code text**. |
|
|
|
|
|
--- |
|
|
|
|
|
## Data Fields |
|
|
|
|
|
Each row contains: |
|
|
|
|
|
- `id` *(string)*: sample identifier (e.g., `easy_000123`) |
|
|
- `difficulty` *(string)*: `easy` / `medium` / `hard` |
|
|
- `code` *(string)*: **ground-truth Python code** |
|
|
- `render_light` *(image)*: synthetic rendering (light theme) |
|
|
- `render_dark` *(image)*: synthetic rendering (dark theme) |
|
|
- `photo` *(image)*: real photo of the code |
|
|
|
|
|
--- |
|
|
|
|
|
## How to Use |
|
|
|
|
|
### Load with 🤗 Datasets |
|
|
|
|
|
```python |
|
|
from datasets import load_dataset |
|
|
|
|
|
ds = load_dataset("maksonchek/codeocr-dataset") |
|
|
print(ds) |
|
|
print(ds["easy"][0].keys()) |
|
|
``` |
|
|
|
|
|
### Access code and images |
|
|
|
|
|
```python |
|
|
ex = ds["easy"][0] |
|
|
|
|
|
# Ground-truth code |
|
|
print(ex["code"][:500]) |
|
|
|
|
|
# Images are stored as `datasets.Image` features. |
|
|
render = ex["render_light"] |
|
|
print(render) |
|
|
``` |
|
|
|
|
|
If your environment returns image dicts with local paths: |
|
|
|
|
|
```python |
|
|
from PIL import Image |
|
|
|
|
|
img = Image.open(ex["render_light"]["path"]) |
|
|
img.show() |
|
|
``` |
|
|
|
|
|
Real photo (always present in this dataset): |
|
|
|
|
|
```python |
|
|
from PIL import Image |
|
|
|
|
|
photo = Image.open(ex["photo"]["path"]) |
|
|
photo.show() |
|
|
``` |
|
|
|
|
|
--- |
|
|
|
|
|
## Dataset Creation |
|
|
|
|
|
### 1) Code selection |
|
|
Python solutions were collected from an open-source repository of LeetCode solutions (MIT licensed). |
|
|
|
|
|
### 2) Normalization to produce stable GT |
|
|
The collected code is written into `gt.py` after: |
|
|
|
|
|
- newline normalization to LF |
|
|
- tab expansion to 4 spaces |
|
|
- basic cleanup (no hidden control characters) |
|
|
- Python syntax check via `compile()` |
|
|
|
|
|
### 3) Synthetic rendering |
|
|
Synthetic images are generated from the normalized `gt.py` in: |
|
|
|
|
|
- light theme (`render_light`) |
|
|
- dark theme (`render_dark`) |
|
|
|
|
|
### 4) Real photos |
|
|
Real photos are manually captured and linked **for every sample**. |
|
|
|
|
|
--- |
|
|
|
|
|
## Statistics (high-level) |
|
|
|
|
|
Average code length by difficulty (computed on this dataset): |
|
|
|
|
|
- `easy`: ~27 lines, ~669 chars |
|
|
- `medium`: ~36 lines, ~997 chars |
|
|
- `hard`: ~55 lines, ~1767 chars |
|
|
|
|
|
(Exact values may vary if the dataset is extended.) |
|
|
|
|
|
--- |
|
|
|
|
|
## Intended Use |
|
|
|
|
|
- OCR for programming code |
|
|
- robust text extraction from screenshot-like renders and real photos |
|
|
- benchmarking OCR pipelines for code formatting / indentation preservation |
|
|
|
|
|
### Not Intended Use |
|
|
|
|
|
- generating or re-distributing problem statements |
|
|
- competitive programming / cheating use-cases |
|
|
|
|
|
--- |
|
|
|
|
|
## Limitations |
|
|
|
|
|
- Code is checked for **syntax correctness**, but not necessarily for runtime correctness. |
|
|
- Rendering style is controlled and may differ from real-world photos. |
|
|
|
|
|
--- |
|
|
|
|
|
## License & Attribution |
|
|
|
|
|
This dataset is released under the **MIT License**. |
|
|
|
|
|
The included solution code is derived from **kamyu104/LeetCode-Solutions** (MIT License): |
|
|
https://github.com/kamyu104/LeetCode-Solutions |
|
|
|
|
|
If you use this dataset in academic work, please cite the dataset and credit the original solution repository. |
|
|
|
|
|
--- |
|
|
|
|
|
## Citation |
|
|
|
|
|
### BibTeX |
|
|
|
|
|
```bibtex |
|
|
@dataset{codeocr_leetcode_2025, |
|
|
author = {Maksonchek}, |
|
|
title = {CodeOCR Dataset (Python Code Images + Ground Truth)}, |
|
|
year = {2025}, |
|
|
publisher = {Hugging Face}, |
|
|
url = {https://huggingface.co/datasets/maksonchek/codeocr-dataset} |
|
|
} |
|
|
``` |
|
|
|