summaryrefslogtreecommitdiff
path: root/text_recognizer/data/iam_synthetic_paragraphs.py
diff options
context:
space:
mode:
Diffstat (limited to 'text_recognizer/data/iam_synthetic_paragraphs.py')
-rw-r--r--text_recognizer/data/iam_synthetic_paragraphs.py31
1 files changed, 19 insertions, 12 deletions
diff --git a/text_recognizer/data/iam_synthetic_paragraphs.py b/text_recognizer/data/iam_synthetic_paragraphs.py
index b9cf90d..f253427 100644
--- a/text_recognizer/data/iam_synthetic_paragraphs.py
+++ b/text_recognizer/data/iam_synthetic_paragraphs.py
@@ -7,20 +7,11 @@ from loguru import logger as log
import numpy as np
from PIL import Image
+from text_recognizer.data.base_data_module import BaseDataModule, load_and_print_info
from text_recognizer.data.base_dataset import (
BaseDataset,
convert_strings_to_labels,
)
-from text_recognizer.data.base_data_module import BaseDataModule, load_and_print_info
-from text_recognizer.data.iam_paragraphs import (
- get_dataset_properties,
- get_transform,
- get_target_transform,
- NEW_LINE_TOKEN,
- IAMParagraphs,
- IMAGE_SCALE_FACTOR,
- resize_image,
-)
from text_recognizer.data.emnist_mapping import EmnistMapping
from text_recognizer.data.iam import IAM
from text_recognizer.data.iam_lines import (
@@ -28,6 +19,15 @@ from text_recognizer.data.iam_lines import (
load_line_crops_and_labels,
save_images_and_labels,
)
+from text_recognizer.data.iam_paragraphs import (
+ get_dataset_properties,
+ get_target_transform,
+ get_transform,
+ IAMParagraphs,
+ IMAGE_SCALE_FACTOR,
+ NEW_LINE_TOKEN,
+ resize_image,
+)
PROCESSED_DATA_DIRNAME = (
@@ -146,7 +146,10 @@ def generate_synthetic_paragraphs(
)
if len(paragraph_label) > paragraphs_properties["label_length"]["max"]:
log.info(
- "Label longer than longest label in original IAM paragraph dataset - hence dropping."
+ (
+ "Label longer than longest label in original IAM paragraph dataset"
+ " - hence dropping."
+ )
)
continue
@@ -160,7 +163,10 @@ def generate_synthetic_paragraphs(
or paragraph_crop.width > max_paragraph_shape[1]
):
log.info(
- "Crop larger than largest crop in original IAM paragraphs dataset - hence dropping"
+ (
+ "Crop larger than largest crop in original IAM paragraphs dataset"
+ " - hence dropping"
+ )
)
continue
@@ -213,4 +219,5 @@ def generate_random_batches(
def create_synthetic_iam_paragraphs() -> None:
+ """Creates and prints IAM Synthetic Paragraphs dataset."""
load_and_print_info(IAMSyntheticParagraphs)