2014-09-26 04:19:41 -07:00
|
|
|
|
#!/usr/bin/env python3
|
|
|
|
|
# Reimplement ocrPage.sh as Python
|
|
|
|
|
|
|
|
|
|
import sys
|
2014-09-26 04:43:15 -07:00
|
|
|
|
import os.path
|
2015-02-13 13:41:14 -08:00
|
|
|
|
import fileinput
|
|
|
|
|
import re
|
2014-09-26 04:19:41 -07:00
|
|
|
|
from parse import parse
|
2015-03-10 14:28:38 -07:00
|
|
|
|
import PyPDF2 as pypdf
|
|
|
|
|
import shutil
|
2014-10-08 03:21:28 -07:00
|
|
|
|
|
2014-11-14 02:06:23 -08:00
|
|
|
|
from subprocess import Popen, check_call, PIPE, CalledProcessError, \
|
|
|
|
|
TimeoutExpired
|
2014-10-08 03:21:28 -07:00
|
|
|
|
try:
|
|
|
|
|
from subprocess import DEVNULL
|
|
|
|
|
except ImportError:
|
|
|
|
|
import os
|
|
|
|
|
DEVNULL = open(os.devnull, 'wb')
|
|
|
|
|
|
2014-09-26 04:19:41 -07:00
|
|
|
|
|
2014-10-10 00:35:49 -07:00
|
|
|
|
from ruffus import transform, suffix, merge, active_if, regex, jobs_limit, \
|
|
|
|
|
mkdir, formatter
|
2014-10-08 03:21:28 -07:00
|
|
|
|
import ruffus.cmdline as cmdline
|
2015-02-20 17:20:48 -08:00
|
|
|
|
from .hocrtransform import HocrTransform
|
|
|
|
|
|
2014-10-08 03:21:28 -07:00
|
|
|
|
|
2014-10-08 03:54:06 -07:00
|
|
|
|
basedir = os.path.dirname(os.path.realpath(__file__))
|
2014-10-08 03:21:28 -07:00
|
|
|
|
|
|
|
|
|
parser = cmdline.get_argparse(
|
|
|
|
|
prog="ocrpage",
|
|
|
|
|
description="Run OCR and related jobs on a single page of a PDF file")
|
|
|
|
|
|
|
|
|
|
parser.add_argument(
|
2015-02-08 19:51:41 -08:00
|
|
|
|
'input_pdf', # Implemented
|
2014-10-08 03:21:28 -07:00
|
|
|
|
help="PDF file containing the page to be OCRed")
|
|
|
|
|
parser.add_argument(
|
2015-02-08 19:51:41 -08:00
|
|
|
|
'page_info', # Implemented
|
2014-10-08 03:21:28 -07:00
|
|
|
|
help="Various characteristics of the page to be OCRed")
|
|
|
|
|
parser.add_argument(
|
2015-02-08 19:51:41 -08:00
|
|
|
|
'num_pages', # Unused
|
2014-10-08 03:21:28 -07:00
|
|
|
|
help="Total number of page of the PDF file (required for logger)")
|
|
|
|
|
parser.add_argument(
|
2015-02-08 19:51:41 -08:00
|
|
|
|
'tmp_fld', # Implemented
|
2014-10-08 03:21:28 -07:00
|
|
|
|
help="Folder where the temporary files should be placed")
|
|
|
|
|
parser.add_argument(
|
2015-02-08 19:51:41 -08:00
|
|
|
|
'verbosity', type=int, # Superseded
|
2014-10-08 03:21:28 -07:00
|
|
|
|
help="Requested verbosity")
|
|
|
|
|
parser.add_argument(
|
2015-02-08 19:51:41 -08:00
|
|
|
|
'language', # Implemented
|
2014-10-08 03:21:28 -07:00
|
|
|
|
help="Language of the file to be OCRed")
|
|
|
|
|
parser.add_argument(
|
2015-02-08 19:51:41 -08:00
|
|
|
|
'keep_tmp', type=int, # Not implemented
|
2014-10-08 03:21:28 -07:00
|
|
|
|
help="Keep the temporary files after processing (helpful for debugging)")
|
|
|
|
|
parser.add_argument(
|
2015-02-08 19:51:41 -08:00
|
|
|
|
'preprocess_deskew', type=int, # Implemented
|
2014-10-08 03:21:28 -07:00
|
|
|
|
help="Deskew the page to be OCRed")
|
|
|
|
|
parser.add_argument(
|
2015-02-08 19:51:41 -08:00
|
|
|
|
'preprocess_clean', type=int, # Implemented
|
2014-10-08 03:21:28 -07:00
|
|
|
|
help="Clean the page to be OCRed")
|
|
|
|
|
parser.add_argument(
|
2015-02-08 19:51:41 -08:00
|
|
|
|
'preprocess_cleantopdf', type=int, # Implemented
|
2014-10-08 03:21:28 -07:00
|
|
|
|
help="Put the cleaned paged in the OCRed PDF")
|
|
|
|
|
parser.add_argument(
|
2015-03-27 18:32:55 -07:00
|
|
|
|
'oversampling_dpi', type=int, # Implemented
|
2014-10-08 03:21:28 -07:00
|
|
|
|
help="Oversampling resolution in dpi")
|
|
|
|
|
parser.add_argument(
|
2015-02-08 19:51:41 -08:00
|
|
|
|
'pdf_noimg', type=int, # implemented
|
2014-10-08 03:21:28 -07:00
|
|
|
|
help="Generate debug PDF pages with only the OCRed text and no image")
|
|
|
|
|
parser.add_argument(
|
2015-02-08 19:51:41 -08:00
|
|
|
|
'force_ocr', type=int, # Implemented
|
2014-10-08 03:21:28 -07:00
|
|
|
|
help="Force to OCR, even if the page already contains fonts")
|
|
|
|
|
parser.add_argument(
|
2015-02-08 19:51:41 -08:00
|
|
|
|
'skip_text', type=int, # Implemented
|
2014-10-08 03:21:28 -07:00
|
|
|
|
help="Skip OCR on pages that contain fonts and include the page anyway")
|
2015-02-20 15:26:33 -08:00
|
|
|
|
parser.add_argument(
|
|
|
|
|
'skip_big', type=int,
|
|
|
|
|
help="Skip OCR for pages that are very large")
|
2015-03-10 14:28:38 -07:00
|
|
|
|
parser.add_argument(
|
|
|
|
|
'exact_image', type=int,
|
|
|
|
|
help="Use original page from PDF without re-rendering")
|
2014-10-08 03:21:28 -07:00
|
|
|
|
parser.add_argument(
|
2015-02-08 19:51:41 -08:00
|
|
|
|
'tess_cfg_files', default='', nargs='*', # Implemented
|
2014-10-08 03:21:28 -07:00
|
|
|
|
help="Tesseract configuration")
|
2014-11-13 16:53:26 -08:00
|
|
|
|
parser.add_argument(
|
|
|
|
|
'--deskew-provider', choices=['imagemagick', 'leptonica'],
|
|
|
|
|
default='leptonica')
|
2015-02-11 03:13:07 -08:00
|
|
|
|
parser.add_argument(
|
|
|
|
|
'--page-renderer', choices=['pdftoppm', 'ghostscript'],
|
|
|
|
|
default='ghostscript')
|
2014-10-08 03:21:28 -07:00
|
|
|
|
|
|
|
|
|
|
|
|
|
|
options = parser.parse_args()
|
|
|
|
|
|
2015-03-24 22:46:33 -07:00
|
|
|
|
_logger, _logger_mutex = cmdline.setup_logging(__name__, options.log_file,
|
|
|
|
|
options.verbose)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
class WrappedLogger:
|
|
|
|
|
|
|
|
|
|
def __init__(self, my_logger, my_mutex):
|
|
|
|
|
self.logger = my_logger
|
|
|
|
|
self.mutex = my_mutex
|
|
|
|
|
|
|
|
|
|
def log(self, *args, **kwargs):
|
|
|
|
|
with self.mutex:
|
|
|
|
|
self.logger.log(*args, **kwargs)
|
|
|
|
|
|
|
|
|
|
def debug(self, *args, **kwargs):
|
|
|
|
|
with self.mutex:
|
|
|
|
|
self.logger.debug(*args, **kwargs)
|
|
|
|
|
|
|
|
|
|
def info(self, *args, **kwargs):
|
|
|
|
|
with self.mutex:
|
|
|
|
|
self.logger.info(*args, **kwargs)
|
|
|
|
|
|
|
|
|
|
def warning(self, *args, **kwargs):
|
|
|
|
|
with self.mutex:
|
|
|
|
|
self.logger.warning(*args, **kwargs)
|
|
|
|
|
|
|
|
|
|
def error(self, *args, **kwargs):
|
|
|
|
|
with self.mutex:
|
|
|
|
|
self.logger.error(*args, **kwargs)
|
|
|
|
|
|
|
|
|
|
def critical(self, *args, **kwargs):
|
|
|
|
|
with self.mutex:
|
|
|
|
|
self.logger.critical(*args, **kwargs)
|
|
|
|
|
|
|
|
|
|
log = WrappedLogger(_logger, _logger_mutex)
|
2014-09-26 04:19:41 -07:00
|
|
|
|
|
|
|
|
|
|
2014-09-26 04:43:15 -07:00
|
|
|
|
def pdf_get_pageinfo(infile, page, width_pt, height_pt):
|
2014-09-26 04:19:41 -07:00
|
|
|
|
pageinfo = {}
|
2014-09-26 04:43:15 -07:00
|
|
|
|
pageinfo['pageno'] = page
|
|
|
|
|
pageinfo['width_inches'] = width_pt / 72.0
|
|
|
|
|
pageinfo['height_inches'] = height_pt / 72.0
|
|
|
|
|
pageinfo['images'] = []
|
2014-09-26 04:19:41 -07:00
|
|
|
|
|
2015-02-08 17:38:27 -08:00
|
|
|
|
p_pdftotext = Popen(['pdftotext', '-f', str(page), '-l', str(page),
|
|
|
|
|
'-raw', '-nopgbrk', infile, '-'],
|
|
|
|
|
close_fds=True, stdout=PIPE, stderr=PIPE,
|
|
|
|
|
universal_newlines=True)
|
|
|
|
|
text, _ = p_pdftotext.communicate()
|
|
|
|
|
if len(text.strip()) > 0:
|
2014-09-26 04:19:41 -07:00
|
|
|
|
pageinfo['has_text'] = True
|
|
|
|
|
else:
|
|
|
|
|
pageinfo['has_text'] = False
|
|
|
|
|
|
2015-04-08 21:55:23 -07:00
|
|
|
|
pdf = pypdf.PdfFileReader(infile)
|
|
|
|
|
page = pdf.pages[page - 1]
|
|
|
|
|
|
|
|
|
|
if not '/XObject' in page['/Resources']:
|
|
|
|
|
# Missing /XObject means no images or possibly corrupt PDF
|
|
|
|
|
return pageinfo
|
|
|
|
|
|
|
|
|
|
for xobj in page['/Resources']['/XObject']:
|
|
|
|
|
# PyPDF2 returns the keys as an iterator
|
|
|
|
|
pdfimage = page['/Resources']['/XObject'][xobj]
|
|
|
|
|
if pdfimage['/Subtype'] != '/Image':
|
|
|
|
|
continue
|
|
|
|
|
image = {}
|
|
|
|
|
image['width'] = pdfimage['/Width']
|
|
|
|
|
image['height'] = pdfimage['/Height']
|
2014-09-26 04:19:41 -07:00
|
|
|
|
image['dpi_w'] = image['width'] / pageinfo['width_inches']
|
|
|
|
|
image['dpi_h'] = image['height'] / pageinfo['height_inches']
|
|
|
|
|
image['dpi'] = (image['dpi_w'] * image['dpi_h']) ** 0.5
|
|
|
|
|
pageinfo['images'].append(image)
|
|
|
|
|
|
2015-02-08 20:05:54 -08:00
|
|
|
|
if pageinfo['images']:
|
|
|
|
|
xres = max(image['dpi_w'] for image in pageinfo['images'])
|
|
|
|
|
yres = max(image['dpi_h'] for image in pageinfo['images'])
|
|
|
|
|
pageinfo['xres'], pageinfo['yres'] = xres, yres
|
|
|
|
|
pageinfo['width_pixels'] = \
|
|
|
|
|
int(round(xres * pageinfo['width_inches']))
|
|
|
|
|
pageinfo['height_pixels'] = \
|
|
|
|
|
int(round(yres * pageinfo['height_inches']))
|
2014-09-27 15:03:07 -07:00
|
|
|
|
|
2015-03-27 18:32:55 -07:00
|
|
|
|
if options.oversampling_dpi > 0:
|
|
|
|
|
rx, ry = options.oversampling_dpi, options.oversampling_dpi
|
|
|
|
|
else:
|
|
|
|
|
rx, ry = pageinfo['xres'], pageinfo['yres']
|
|
|
|
|
pageinfo['xres_render'], pageinfo['yres_render'] = rx, ry
|
|
|
|
|
|
2014-09-26 04:19:41 -07:00
|
|
|
|
return pageinfo
|
|
|
|
|
|
2014-10-08 03:21:28 -07:00
|
|
|
|
pageno, width_pt, height_pt = map(int, options.page_info.split(' ', 3))
|
|
|
|
|
pageinfo = pdf_get_pageinfo(options.input_pdf, pageno, width_pt, height_pt)
|
|
|
|
|
|
2015-03-24 22:46:33 -07:00
|
|
|
|
if not pageinfo['images']:
|
|
|
|
|
# If the page has no images, then it contains vector content or text
|
|
|
|
|
# or both. It seems quite unlikely that one would find meaningful text
|
|
|
|
|
# from rasterizing vector content. So skip the page.
|
|
|
|
|
log.info(
|
|
|
|
|
"Page {0} has no images - skipping OCR".format(pageno)
|
|
|
|
|
)
|
|
|
|
|
elif pageinfo['has_text']:
|
|
|
|
|
s = "Page {0} already has text! – {1}"
|
|
|
|
|
|
|
|
|
|
if not options.force_ocr and not options.skip_text:
|
|
|
|
|
log.error(s.format(pageno,
|
|
|
|
|
"aborting (use -f or -s to force OCR)"))
|
|
|
|
|
sys.exit(1)
|
|
|
|
|
elif options.force_ocr:
|
|
|
|
|
log.info(s.format(pageno,
|
|
|
|
|
"rasterizing text and running OCR anyway"))
|
|
|
|
|
elif options.skip_text:
|
|
|
|
|
log.info(s.format(pageno,
|
|
|
|
|
"skipping all processing on this page"))
|
2015-02-08 19:51:18 -08:00
|
|
|
|
|
2015-02-08 20:05:54 -08:00
|
|
|
|
ocr_required = pageinfo['images'] and \
|
|
|
|
|
(options.force_ocr or
|
|
|
|
|
(not (pageinfo['has_text'] and options.skip_text)))
|
2015-02-08 19:51:18 -08:00
|
|
|
|
|
2015-02-20 15:26:33 -08:00
|
|
|
|
if ocr_required and options.skip_big:
|
|
|
|
|
area = pageinfo['width_inches'] * pageinfo['height_inches']
|
|
|
|
|
pixel_count = pageinfo['width_pixels'] * pageinfo['height_pixels']
|
|
|
|
|
if area > (11.0 * 17.0) or pixel_count > (300.0 * 300.0 * 11 * 17):
|
|
|
|
|
ocr_required = False
|
2015-03-24 22:46:33 -07:00
|
|
|
|
log.info(
|
2015-02-20 15:26:33 -08:00
|
|
|
|
"Page {0} is very large; skipping due to -b".format(pageno))
|
|
|
|
|
|
2014-09-26 04:19:41 -07:00
|
|
|
|
|
2015-03-24 22:46:33 -07:00
|
|
|
|
def re_symlink(input_file, soft_link_name, log=log):
|
2014-10-10 00:35:49 -07:00
|
|
|
|
"""
|
|
|
|
|
Helper function: relinks soft symbolic link if necessary
|
|
|
|
|
"""
|
|
|
|
|
# Guard against soft linking to oneself
|
|
|
|
|
if input_file == soft_link_name:
|
2015-03-24 22:46:33 -07:00
|
|
|
|
log.debug("Warning: No symbolic link made. You are using " +
|
2014-10-10 00:35:49 -07:00
|
|
|
|
"the original data directory as the working directory.")
|
|
|
|
|
return
|
|
|
|
|
|
|
|
|
|
# Soft link already exists: delete for relink?
|
|
|
|
|
if os.path.lexists(soft_link_name):
|
|
|
|
|
# do not delete or overwrite real (non-soft link) file
|
|
|
|
|
if not os.path.islink(soft_link_name):
|
|
|
|
|
raise Exception("%s exists and is not a link" % soft_link_name)
|
|
|
|
|
try:
|
|
|
|
|
os.unlink(soft_link_name)
|
|
|
|
|
except:
|
2015-03-24 22:46:33 -07:00
|
|
|
|
log.debug("Can't unlink %s" % (soft_link_name))
|
2015-02-11 10:24:10 -08:00
|
|
|
|
|
|
|
|
|
if not os.path.exists(input_file):
|
|
|
|
|
raise Exception("trying to create a broken symlink to %s" % input_file)
|
|
|
|
|
|
2015-03-24 22:46:33 -07:00
|
|
|
|
log.debug("os.symlink(%s, %s)" % (input_file, soft_link_name))
|
2014-10-10 00:35:49 -07:00
|
|
|
|
|
2014-10-11 17:48:56 -07:00
|
|
|
|
# Create symbolic link using absolute path
|
2014-10-10 00:35:49 -07:00
|
|
|
|
os.symlink(
|
2014-10-11 17:48:56 -07:00
|
|
|
|
os.path.abspath(input_file),
|
2014-10-10 00:35:49 -07:00
|
|
|
|
soft_link_name
|
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
@jobs_limit(1)
|
|
|
|
|
@mkdir(options.tmp_fld)
|
|
|
|
|
@transform([options.input_pdf],
|
|
|
|
|
formatter(),
|
2014-10-10 01:30:19 -07:00
|
|
|
|
os.path.join(options.tmp_fld, "original{ext[0]}"))
|
2014-10-10 00:35:49 -07:00
|
|
|
|
def setup_working_directory(input_file, soft_link_name):
|
2015-03-24 22:46:33 -07:00
|
|
|
|
log.debug("Linking %(input_file)s -> %(soft_link_name)s" % locals())
|
2014-11-13 15:58:36 -08:00
|
|
|
|
try:
|
2015-03-24 22:46:33 -07:00
|
|
|
|
re_symlink(input_file, soft_link_name)
|
2014-11-13 15:58:36 -08:00
|
|
|
|
except FileExistsError:
|
|
|
|
|
pass
|
2014-10-10 00:35:49 -07:00
|
|
|
|
|
|
|
|
|
|
2015-03-10 14:28:38 -07:00
|
|
|
|
@active_if(not ocr_required or (ocr_required and options.exact_image))
|
2015-02-08 17:26:16 -08:00
|
|
|
|
@transform(setup_working_directory,
|
|
|
|
|
formatter(),
|
2015-03-10 14:28:38 -07:00
|
|
|
|
os.path.join(options.tmp_fld, '%04i.page.pdf' % pageno))
|
|
|
|
|
def extract_single_page(
|
2015-02-08 17:26:16 -08:00
|
|
|
|
input_file,
|
|
|
|
|
output_file):
|
|
|
|
|
args_pdfseparate = [
|
|
|
|
|
'pdfseparate',
|
|
|
|
|
'-f', str(pageinfo['pageno']), '-l', str(pageinfo['pageno']),
|
|
|
|
|
input_file,
|
|
|
|
|
output_file
|
|
|
|
|
]
|
|
|
|
|
check_call(args_pdfseparate)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
@active_if(ocr_required)
|
2015-02-11 03:13:07 -08:00
|
|
|
|
@active_if(options.page_renderer == 'pdftoppm')
|
2014-10-10 01:30:19 -07:00
|
|
|
|
@transform(setup_working_directory,
|
|
|
|
|
formatter(),
|
|
|
|
|
"{path[0]}/%04i.pnm" % pageno)
|
2014-10-08 03:21:28 -07:00
|
|
|
|
def unpack_with_pdftoppm(
|
|
|
|
|
input_file,
|
|
|
|
|
output_file):
|
|
|
|
|
force_ppm = True
|
2014-10-10 01:54:16 -07:00
|
|
|
|
allow_jpeg = False
|
2014-09-26 04:19:41 -07:00
|
|
|
|
|
|
|
|
|
colorspace = 'color'
|
|
|
|
|
compression = 'deflate'
|
|
|
|
|
output_format = 'tiff'
|
|
|
|
|
if all(image['comp'] == 1 for image in pageinfo['images']):
|
|
|
|
|
if all(image['bpc'] == 1 for image in pageinfo['images']):
|
|
|
|
|
colorspace = 'mono'
|
|
|
|
|
compression = 'deflate'
|
2014-09-26 04:43:15 -07:00
|
|
|
|
elif not any(image['color'] == 'color'
|
|
|
|
|
for image in pageinfo['images']):
|
2014-09-26 04:19:41 -07:00
|
|
|
|
colorspace = 'gray'
|
|
|
|
|
|
2014-10-10 01:54:16 -07:00
|
|
|
|
if allow_jpeg and \
|
|
|
|
|
all(image['enc'] == 'jpeg' for image in pageinfo['images']):
|
2014-09-26 04:19:41 -07:00
|
|
|
|
output_format = 'jpeg'
|
|
|
|
|
|
|
|
|
|
args_pdftoppm = [
|
|
|
|
|
'pdftoppm',
|
2014-09-26 04:43:15 -07:00
|
|
|
|
'-f', str(pageinfo['pageno']), '-l', str(pageinfo['pageno']),
|
2015-03-27 18:32:55 -07:00
|
|
|
|
'-rx', str(pageinfo['xres_render']),
|
|
|
|
|
'-ry', str(pageinfo['yres_render'])
|
2014-09-26 04:19:41 -07:00
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
if not force_ppm:
|
|
|
|
|
if output_format == 'tiff':
|
|
|
|
|
args_pdftoppm.append('-tiff')
|
2014-10-10 01:54:16 -07:00
|
|
|
|
if False and compression:
|
2014-09-26 04:19:41 -07:00
|
|
|
|
args_pdftoppm.append('-tiffcompression')
|
|
|
|
|
args_pdftoppm.append(compression)
|
|
|
|
|
elif output_format == 'jpeg':
|
|
|
|
|
args_pdftoppm.append('-jpeg')
|
|
|
|
|
|
|
|
|
|
if colorspace == 'mono':
|
|
|
|
|
args_pdftoppm.append('-mono')
|
|
|
|
|
elif colorspace == 'gray':
|
|
|
|
|
args_pdftoppm.append('-gray')
|
|
|
|
|
|
2014-10-08 03:21:28 -07:00
|
|
|
|
args_pdftoppm.extend([str(input_file)])
|
2014-09-26 04:19:41 -07:00
|
|
|
|
|
2014-10-17 17:28:31 -07:00
|
|
|
|
# Ask pdftoppm to write the binary output to stdout; therefore set
|
|
|
|
|
# universal_newlines=False
|
2014-10-08 03:21:28 -07:00
|
|
|
|
p = Popen(args_pdftoppm, close_fds=True, stdout=open(output_file, 'wb'),
|
2014-10-10 00:35:49 -07:00
|
|
|
|
stderr=PIPE, universal_newlines=False)
|
2014-10-08 03:21:28 -07:00
|
|
|
|
_, stderr = p.communicate()
|
|
|
|
|
if stderr:
|
2014-10-17 17:28:31 -07:00
|
|
|
|
# Because universal_newlines=False, stderr is bytes(), so we must
|
|
|
|
|
# manually convert it to str for logging
|
2014-11-13 15:58:57 -08:00
|
|
|
|
from codecs import decode
|
2015-03-24 22:46:33 -07:00
|
|
|
|
log.error(decode(stderr, sys.getdefaultencoding(), 'ignore'))
|
2014-10-10 01:07:46 -07:00
|
|
|
|
if p.returncode != 0:
|
2014-10-10 01:54:16 -07:00
|
|
|
|
raise CalledProcessError(p.returncode, args_pdftoppm)
|
|
|
|
|
|
|
|
|
|
|
2015-02-08 17:26:16 -08:00
|
|
|
|
@active_if(ocr_required)
|
2015-02-11 03:13:07 -08:00
|
|
|
|
@transform(unpack_with_pdftoppm, suffix(".pnm"), ".png")
|
|
|
|
|
def convert_to_png(input_file, output_file):
|
2014-10-10 01:54:16 -07:00
|
|
|
|
args_convert = [
|
|
|
|
|
'convert',
|
|
|
|
|
input_file,
|
|
|
|
|
output_file
|
|
|
|
|
]
|
|
|
|
|
check_call(args_convert)
|
2014-09-27 15:03:07 -07:00
|
|
|
|
|
|
|
|
|
|
2015-02-11 03:13:07 -08:00
|
|
|
|
@active_if(ocr_required)
|
|
|
|
|
@active_if(options.page_renderer == 'ghostscript')
|
|
|
|
|
@transform(setup_working_directory,
|
|
|
|
|
formatter(),
|
|
|
|
|
"{path[0]}/%04i.png" % pageno)
|
|
|
|
|
def unpack_with_ghostscript(
|
|
|
|
|
input_file,
|
|
|
|
|
output_file):
|
2015-02-11 03:32:00 -08:00
|
|
|
|
device = 'png16m' # 24-bit
|
|
|
|
|
if all(image['comp'] == 1 for image in pageinfo['images']):
|
|
|
|
|
if all(image['bpc'] == 1 for image in pageinfo['images']):
|
|
|
|
|
device = 'pngmono'
|
|
|
|
|
elif not any(image['color'] == 'color'
|
|
|
|
|
for image in pageinfo['images']):
|
|
|
|
|
device = 'pnggray'
|
|
|
|
|
|
2015-02-11 03:13:07 -08:00
|
|
|
|
args_gs = [
|
|
|
|
|
'gs',
|
|
|
|
|
'-dBATCH', '-dNOPAUSE',
|
|
|
|
|
'-dFirstPage=%i' % pageno,
|
|
|
|
|
'-dLastPage=%i' % pageno,
|
2015-02-11 03:32:00 -08:00
|
|
|
|
'-sDEVICE=%s' % device,
|
2015-02-11 03:13:07 -08:00
|
|
|
|
'-o', output_file,
|
2015-03-27 18:32:55 -07:00
|
|
|
|
'-r{0}x{1}'.format(
|
|
|
|
|
str(pageinfo['xres_render']), str(pageinfo['yres_render'])),
|
2015-02-11 03:13:07 -08:00
|
|
|
|
input_file
|
|
|
|
|
]
|
|
|
|
|
|
|
|
|
|
p = Popen(args_gs, close_fds=True, stdout=PIPE, stderr=PIPE,
|
|
|
|
|
universal_newlines=True)
|
|
|
|
|
stdout, stderr = p.communicate()
|
2015-03-24 22:46:33 -07:00
|
|
|
|
if stdout:
|
|
|
|
|
log.info(stdout)
|
|
|
|
|
if stderr:
|
|
|
|
|
log.error(stderr)
|
2015-02-11 03:13:07 -08:00
|
|
|
|
|
|
|
|
|
try:
|
|
|
|
|
f = open(output_file)
|
|
|
|
|
except FileNotFoundError:
|
|
|
|
|
raise
|
|
|
|
|
else:
|
|
|
|
|
f.close()
|
|
|
|
|
|
|
|
|
|
|
2015-02-08 17:26:16 -08:00
|
|
|
|
@active_if(ocr_required)
|
2014-11-13 16:53:26 -08:00
|
|
|
|
@active_if(options.preprocess_deskew != 0
|
|
|
|
|
and options.deskew_provider == 'imagemagick')
|
2015-02-11 03:13:07 -08:00
|
|
|
|
@transform(convert_to_png, suffix(".png"), ".deskewed.png")
|
2014-10-10 00:35:49 -07:00
|
|
|
|
def deskew_imagemagick(input_file, output_file):
|
2014-09-27 15:03:07 -07:00
|
|
|
|
args_convert = [
|
|
|
|
|
'convert',
|
2014-10-10 00:35:49 -07:00
|
|
|
|
input_file,
|
2014-09-27 15:03:07 -07:00
|
|
|
|
'-deskew', '40%',
|
|
|
|
|
'-gravity', 'center',
|
2014-10-10 00:35:49 -07:00
|
|
|
|
'-extent', '{width_pixels}x{height_pixels}'.format(**pageinfo),
|
2014-10-10 01:54:16 -07:00
|
|
|
|
'+repage',
|
2014-10-10 00:35:49 -07:00
|
|
|
|
output_file
|
2014-09-27 15:03:07 -07:00
|
|
|
|
]
|
|
|
|
|
|
2014-10-10 00:35:49 -07:00
|
|
|
|
p = Popen(args_convert, close_fds=True, stdout=PIPE, stderr=PIPE,
|
|
|
|
|
universal_newlines=True)
|
|
|
|
|
stdout, stderr = p.communicate()
|
|
|
|
|
|
2015-03-24 22:46:33 -07:00
|
|
|
|
if stdout:
|
|
|
|
|
log.info(stdout)
|
|
|
|
|
if stderr:
|
|
|
|
|
log.error(stderr)
|
2014-09-27 15:03:07 -07:00
|
|
|
|
|
2014-10-10 01:07:46 -07:00
|
|
|
|
if p.returncode != 0:
|
2014-10-10 01:54:16 -07:00
|
|
|
|
raise CalledProcessError(p.returncode, args_convert)
|
2014-10-10 01:07:46 -07:00
|
|
|
|
|
2014-09-27 15:03:07 -07:00
|
|
|
|
|
2015-02-08 17:26:16 -08:00
|
|
|
|
@active_if(ocr_required)
|
2014-11-13 16:53:26 -08:00
|
|
|
|
@active_if(options.preprocess_deskew != 0
|
|
|
|
|
and options.deskew_provider == 'leptonica')
|
2015-02-11 03:13:07 -08:00
|
|
|
|
@transform(convert_to_png, suffix(".png"), ".deskewed.png")
|
2014-11-13 16:53:26 -08:00
|
|
|
|
def deskew_leptonica(input_file, output_file):
|
|
|
|
|
from .leptonica import deskew
|
2015-03-24 22:46:33 -07:00
|
|
|
|
deskew(input_file, output_file,
|
|
|
|
|
min(pageinfo['xres'], pageinfo['yres']))
|
2014-11-13 16:53:26 -08:00
|
|
|
|
|
|
|
|
|
|
2015-02-08 17:26:16 -08:00
|
|
|
|
@active_if(ocr_required)
|
|
|
|
|
@active_if(options.preprocess_clean != 0)
|
2015-02-11 03:13:07 -08:00
|
|
|
|
@merge([unpack_with_pdftoppm, unpack_with_ghostscript,
|
|
|
|
|
deskew_imagemagick, deskew_leptonica],
|
2014-11-14 00:23:22 -08:00
|
|
|
|
os.path.join(options.tmp_fld, "%04i.for_clean.pnm" % pageno))
|
|
|
|
|
def select_image_for_cleaning(infiles, output_file):
|
|
|
|
|
input_file = infiles[-1]
|
|
|
|
|
args_convert = [
|
|
|
|
|
'convert',
|
|
|
|
|
input_file,
|
|
|
|
|
output_file
|
|
|
|
|
]
|
|
|
|
|
check_call(args_convert)
|
|
|
|
|
|
|
|
|
|
|
2015-02-08 17:26:16 -08:00
|
|
|
|
@active_if(ocr_required)
|
2014-11-14 00:23:22 -08:00
|
|
|
|
@active_if(options.preprocess_clean != 0)
|
|
|
|
|
@transform(select_image_for_cleaning, suffix(".pnm"), ".cleaned.pnm")
|
|
|
|
|
def clean_unpaper(input_file, output_file):
|
2014-09-27 15:03:07 -07:00
|
|
|
|
args_unpaper = [
|
|
|
|
|
'unpaper',
|
2014-11-14 00:23:22 -08:00
|
|
|
|
'--dpi', str(int(round((pageinfo['xres'] * pageinfo['yres']) ** 0.5))),
|
2014-09-27 15:03:07 -07:00
|
|
|
|
'--mask-scan-size', '100',
|
|
|
|
|
'--no-deskew',
|
|
|
|
|
'--no-grayfilter',
|
|
|
|
|
'--no-blackfilter',
|
|
|
|
|
'--no-mask-center',
|
|
|
|
|
'--no-border-align',
|
2014-11-14 00:23:22 -08:00
|
|
|
|
input_file,
|
|
|
|
|
output_file
|
2014-09-27 15:03:07 -07:00
|
|
|
|
]
|
|
|
|
|
|
2014-11-14 00:23:22 -08:00
|
|
|
|
p = Popen(args_unpaper, close_fds=True, stdout=PIPE, stderr=PIPE,
|
|
|
|
|
universal_newlines=True)
|
|
|
|
|
stdout, stderr = p.communicate()
|
|
|
|
|
|
2015-03-24 22:46:33 -07:00
|
|
|
|
if stdout:
|
2015-07-22 02:57:13 -07:00
|
|
|
|
log.info(stdout)
|
2015-03-24 22:46:33 -07:00
|
|
|
|
if stderr:
|
2015-07-22 02:57:13 -07:00
|
|
|
|
log.error(stderr)
|
2014-11-14 00:23:22 -08:00
|
|
|
|
|
|
|
|
|
if p.returncode != 0:
|
|
|
|
|
raise CalledProcessError(p.returncode, args_unpaper)
|
|
|
|
|
|
|
|
|
|
|
2015-02-08 17:26:16 -08:00
|
|
|
|
@active_if(ocr_required)
|
2015-02-11 03:13:07 -08:00
|
|
|
|
@transform(clean_unpaper, suffix(".cleaned.pnm"), ".cleaned.png")
|
|
|
|
|
def cleaned_to_png(input_file, output_file):
|
2014-11-14 00:23:22 -08:00
|
|
|
|
args_convert = [
|
|
|
|
|
'convert',
|
|
|
|
|
input_file,
|
|
|
|
|
output_file
|
|
|
|
|
]
|
|
|
|
|
check_call(args_convert)
|
2014-09-26 04:19:41 -07:00
|
|
|
|
|
|
|
|
|
|
2015-02-08 17:26:16 -08:00
|
|
|
|
@active_if(ocr_required)
|
2015-02-11 03:13:07 -08:00
|
|
|
|
@merge([unpack_with_ghostscript, convert_to_png, deskew_imagemagick,
|
|
|
|
|
deskew_leptonica, cleaned_to_png],
|
|
|
|
|
os.path.join(options.tmp_fld, "%04i.for_ocr.png" % pageno))
|
2014-10-10 00:35:49 -07:00
|
|
|
|
def select_ocr_image(infiles, output_file):
|
2015-07-22 02:57:13 -07:00
|
|
|
|
re_symlink(infiles[-1], output_file)
|
2014-10-10 00:35:49 -07:00
|
|
|
|
|
|
|
|
|
|
2014-11-14 02:06:23 -08:00
|
|
|
|
hocr_template = '''<?xml version="1.0" encoding="UTF-8"?>
|
|
|
|
|
<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Transitional//EN"
|
|
|
|
|
"http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd">
|
|
|
|
|
<html xmlns="http://www.w3.org/1999/xhtml" xml:lang="en" lang="en">
|
|
|
|
|
<head>
|
|
|
|
|
<title></title>
|
|
|
|
|
<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
|
|
|
|
|
<meta name='ocr-system' content='tesseract 3.02.02' />
|
|
|
|
|
<meta name='ocr-capabilities' content='ocr_page ocr_carea ocr_par ocr_line ocrx_word'/>
|
|
|
|
|
</head>
|
|
|
|
|
<body>
|
|
|
|
|
<div class='ocr_page' id='page_1' title='image "x.tif"; bbox 0 0 {0} {1}; ppageno 0'>
|
|
|
|
|
<div class='ocr_carea' id='block_1_1' title="bbox 0 1 {0} {1}">
|
|
|
|
|
<p class='ocr_par' dir='ltr' id='par_1' title="bbox 0 1 {0} {1}">
|
2015-03-24 22:46:33 -07:00
|
|
|
|
<span class='ocr_line' id='line_1' title="bbox 0 1 {0} {1}"><span class='ocrx_word' id='word_1' title="bbox 0 1 {0} {1}"> </span>
|
2014-11-14 02:06:23 -08:00
|
|
|
|
</span>
|
|
|
|
|
</p>
|
|
|
|
|
</div>
|
|
|
|
|
</div>
|
|
|
|
|
</body>
|
|
|
|
|
</html>'''
|
|
|
|
|
|
|
|
|
|
|
2015-02-08 17:26:16 -08:00
|
|
|
|
@active_if(ocr_required)
|
2015-02-11 03:13:07 -08:00
|
|
|
|
@transform(select_ocr_image, suffix(".for_ocr.png"), ".hocr")
|
2014-10-08 03:21:28 -07:00
|
|
|
|
def ocr_tesseract(
|
|
|
|
|
input_file,
|
|
|
|
|
output_file):
|
|
|
|
|
|
|
|
|
|
args_tesseract = [
|
|
|
|
|
'tesseract',
|
|
|
|
|
'-l', options.language,
|
|
|
|
|
input_file,
|
|
|
|
|
output_file,
|
|
|
|
|
'hocr',
|
|
|
|
|
options.tess_cfg_files
|
|
|
|
|
]
|
|
|
|
|
p = Popen(args_tesseract, close_fds=True, stdout=PIPE, stderr=PIPE,
|
|
|
|
|
universal_newlines=True)
|
2014-11-14 02:06:23 -08:00
|
|
|
|
try:
|
|
|
|
|
stdout, stderr = p.communicate(timeout=180)
|
|
|
|
|
except TimeoutExpired:
|
|
|
|
|
p.kill()
|
|
|
|
|
stdout, stderr = p.communicate()
|
|
|
|
|
# Generate a HOCR file with no recognized text if tesseract times out
|
|
|
|
|
# Temporary workaround to hocrTransform not being able to function if
|
|
|
|
|
# it does not have a valid hOCR file.
|
|
|
|
|
with open(output_file, 'w', encoding="utf-8") as f:
|
|
|
|
|
f.write(hocr_template.format(pageinfo['width_pixels'],
|
|
|
|
|
pageinfo['height_pixels']))
|
|
|
|
|
else:
|
2015-03-24 22:46:33 -07:00
|
|
|
|
if stdout:
|
2015-07-22 02:57:13 -07:00
|
|
|
|
log.info(stdout)
|
2015-03-24 22:46:33 -07:00
|
|
|
|
if stderr:
|
2015-07-22 02:57:13 -07:00
|
|
|
|
log.error(stderr)
|
2014-10-10 00:35:49 -07:00
|
|
|
|
|
2014-11-14 02:06:23 -08:00
|
|
|
|
if p.returncode != 0:
|
|
|
|
|
raise CalledProcessError(p.returncode, args_tesseract)
|
2014-10-10 01:07:46 -07:00
|
|
|
|
|
2015-02-11 10:24:10 -08:00
|
|
|
|
if os.path.exists(output_file + '.html'):
|
2015-03-10 14:28:38 -07:00
|
|
|
|
# Tesseract 3.02 appends suffix ".html" on its own (.hocr.html)
|
|
|
|
|
shutil.move(output_file + '.html', output_file)
|
2015-02-11 10:24:10 -08:00
|
|
|
|
elif os.path.exists(output_file + '.hocr'):
|
2015-03-10 14:28:38 -07:00
|
|
|
|
# Tesseract 3.03 appends suffix ".hocr" on its own (.hocr.hocr)
|
|
|
|
|
shutil.move(output_file + '.hocr', output_file)
|
|
|
|
|
|
|
|
|
|
# Tesseract inserts source filename into hocr file without escaping
|
|
|
|
|
# it. This could break the XML parser. Rewrite the hocr file,
|
|
|
|
|
# replacing the filename with a space.
|
|
|
|
|
regex_nested_single_quotes = re.compile(
|
|
|
|
|
r"""title='image "([^"]*)";""")
|
|
|
|
|
with fileinput.input(files=(output_file,), inplace=True) as f:
|
|
|
|
|
for line in f:
|
|
|
|
|
line = regex_nested_single_quotes.sub(
|
|
|
|
|
r"""title='image " ";""", line)
|
|
|
|
|
print(line, end='') # fileinput.input redirects stdout
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
@active_if(ocr_required and not options.exact_image)
|
2015-02-11 03:13:07 -08:00
|
|
|
|
@merge([unpack_with_ghostscript, convert_to_png,
|
|
|
|
|
deskew_imagemagick, deskew_leptonica, cleaned_to_png],
|
2015-02-11 10:23:45 -08:00
|
|
|
|
os.path.join(options.tmp_fld, "%04i.image_for_pdf" % pageno))
|
2014-11-14 00:23:22 -08:00
|
|
|
|
def select_image_for_pdf(infiles, output_file):
|
|
|
|
|
if options.preprocess_clean != 0 and options.preprocess_cleantopdf != 0:
|
|
|
|
|
input_file = infiles[-1]
|
|
|
|
|
elif options.preprocess_deskew != 0 and options.preprocess_clean != 0:
|
|
|
|
|
input_file = infiles[-2]
|
|
|
|
|
elif options.preprocess_deskew != 0 and options.preprocess_clean == 0:
|
|
|
|
|
input_file = infiles[-1]
|
|
|
|
|
else:
|
|
|
|
|
input_file = infiles[0]
|
2015-02-11 10:23:45 -08:00
|
|
|
|
|
|
|
|
|
if all(image['enc'] == 'jpeg' for image in pageinfo['images']):
|
|
|
|
|
# If all images were JPEGs originally, produce a JPEG as output
|
|
|
|
|
check_call(['convert', input_file, 'jpg:' + output_file])
|
|
|
|
|
else:
|
2015-07-22 02:57:13 -07:00
|
|
|
|
re_symlink(input_file, output_file)
|
2014-11-14 00:23:22 -08:00
|
|
|
|
|
|
|
|
|
|
2015-03-10 14:28:38 -07:00
|
|
|
|
@active_if(ocr_required and not options.exact_image)
|
2014-11-14 00:23:22 -08:00
|
|
|
|
@merge([ocr_tesseract, select_image_for_pdf],
|
2015-02-08 17:26:16 -08:00
|
|
|
|
os.path.join(options.tmp_fld, '%04i.rendered.pdf' % pageno))
|
2014-10-08 03:54:06 -07:00
|
|
|
|
def render_page(infiles, output_file):
|
2015-02-20 17:20:48 -08:00
|
|
|
|
hocr, image = infiles[0], infiles[1]
|
2015-02-08 19:51:41 -08:00
|
|
|
|
|
2015-02-20 17:20:48 -08:00
|
|
|
|
dpi = round(max(pageinfo['xres'], pageinfo['yres']))
|
2015-02-08 19:51:41 -08:00
|
|
|
|
|
2015-02-20 17:20:48 -08:00
|
|
|
|
hocrtransform = HocrTransform(hocr, dpi)
|
|
|
|
|
hocrtransform.to_pdf(output_file, imageFileName=image,
|
2015-02-21 14:19:27 -08:00
|
|
|
|
showBoundingboxes=False, invisibleText=True)
|
2015-02-08 19:51:41 -08:00
|
|
|
|
|
|
|
|
|
|
|
|
|
|
@active_if(ocr_required and options.pdf_noimg)
|
|
|
|
|
@transform(ocr_tesseract, suffix(".hocr"), ".ocred.todebug.pdf")
|
|
|
|
|
def render_text_output_page(input_file, output_file):
|
2015-02-20 17:20:48 -08:00
|
|
|
|
dpi = round(max(pageinfo['xres'], pageinfo['yres']))
|
2014-10-08 03:54:06 -07:00
|
|
|
|
|
2015-02-20 17:20:48 -08:00
|
|
|
|
hocrtransform = HocrTransform(input_file, dpi)
|
|
|
|
|
hocrtransform.to_pdf(output_file, imageFileName=None,
|
2015-02-21 14:19:27 -08:00
|
|
|
|
showBoundingboxes=True, invisibleText=False)
|
2014-10-10 01:07:46 -07:00
|
|
|
|
|
2014-10-08 03:54:06 -07:00
|
|
|
|
|
2015-03-10 14:28:38 -07:00
|
|
|
|
@active_if(ocr_required and options.exact_image)
|
|
|
|
|
@transform(ocr_tesseract, suffix(".hocr"), ".hocr.pdf")
|
|
|
|
|
def render_hocr_blank_page(input_file, output_file):
|
|
|
|
|
dpi = round(max(pageinfo['xres'], pageinfo['yres']))
|
|
|
|
|
|
|
|
|
|
hocrtransform = HocrTransform(input_file, dpi)
|
|
|
|
|
hocrtransform.to_pdf(output_file, imageFileName=None,
|
|
|
|
|
showBoundingboxes=False, invisibleText=True)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
@active_if(ocr_required and options.exact_image)
|
|
|
|
|
@merge([render_hocr_blank_page, extract_single_page],
|
|
|
|
|
os.path.join(options.tmp_fld, "%04i.merged.pdf") % pageno)
|
|
|
|
|
def merge_hocr_with_original_page(infiles, output_file):
|
|
|
|
|
with open(infiles[0], 'rb') as hocr_input, \
|
|
|
|
|
open(infiles[1], 'rb') as page_input, \
|
|
|
|
|
open(output_file, 'wb') as output:
|
|
|
|
|
hocr_reader = pypdf.PdfFileReader(hocr_input)
|
|
|
|
|
page_reader = pypdf.PdfFileReader(page_input)
|
|
|
|
|
writer = pypdf.PdfFileWriter()
|
|
|
|
|
|
|
|
|
|
the_page = hocr_reader.getPage(0)
|
|
|
|
|
the_page.mergePage(page_reader.getPage(0))
|
|
|
|
|
writer.addPage(the_page)
|
|
|
|
|
writer.write(output)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
@merge([render_page, merge_hocr_with_original_page, extract_single_page],
|
2015-02-08 17:26:16 -08:00
|
|
|
|
os.path.join(options.tmp_fld, '%04i.ocred.pdf' % pageno))
|
|
|
|
|
def select_final_page(infiles, output_file):
|
2015-07-22 02:57:13 -07:00
|
|
|
|
re_symlink(infiles[-1], output_file)
|
2015-02-08 17:26:16 -08:00
|
|
|
|
|
|
|
|
|
|
2015-03-24 22:46:33 -07:00
|
|
|
|
if __name__ == '__main__':
|
|
|
|
|
cmdline.run(options)
|
2014-09-26 04:19:41 -07:00
|
|
|
|
|
|
|
|
|
|