mirror of
https://github.com/Unstructured-IO/unstructured.git
synced 2025-09-03 05:39:47 +00:00

This PR aims to add backward compatibility for the deprecated `pdf_infer_table_structure` parameter. A missing part of turning table extraction for PDFs and Images off by default in https://github.com/Unstructured-IO/unstructured/pull/3035, which was turned on in https://github.com/Unstructured-IO/unstructured/pull/2588. --------- Co-authored-by: ryannikolaidis <1208590+ryannikolaidis@users.noreply.github.com> Co-authored-by: christinestraub <christinestraub@users.noreply.github.com>
595 lines
23 KiB
Python
595 lines
23 KiB
Python
"""Provides partitioning with automatic file-type detection."""
|
|
|
|
from __future__ import annotations
|
|
|
|
import io
|
|
from typing import IO, Any, Callable, Literal, Optional
|
|
|
|
import requests
|
|
|
|
from unstructured.documents.elements import DataSourceMetadata, Element
|
|
from unstructured.file_utils.filetype import (
|
|
FILETYPE_TO_MIMETYPE,
|
|
STR_TO_FILETYPE,
|
|
FileType,
|
|
detect_filetype,
|
|
is_json_processable,
|
|
)
|
|
from unstructured.logger import logger
|
|
from unstructured.partition.common import exactly_one
|
|
from unstructured.partition.email import partition_email
|
|
from unstructured.partition.html import partition_html
|
|
from unstructured.partition.json import partition_json
|
|
from unstructured.partition.lang import check_language_args
|
|
from unstructured.partition.text import partition_text
|
|
from unstructured.partition.utils.constants import PartitionStrategy
|
|
from unstructured.partition.xml import partition_xml
|
|
from unstructured.utils import dependency_exists
|
|
|
|
PARTITION_WITH_EXTRAS_MAP: dict[str, Callable[..., list[Element]]] = {}
|
|
|
|
if dependency_exists("pandas"):
|
|
from unstructured.partition.csv import partition_csv
|
|
from unstructured.partition.tsv import partition_tsv
|
|
|
|
PARTITION_WITH_EXTRAS_MAP["csv"] = partition_csv
|
|
PARTITION_WITH_EXTRAS_MAP["tsv"] = partition_tsv
|
|
|
|
|
|
if dependency_exists("docx"):
|
|
from unstructured.partition.doc import partition_doc
|
|
from unstructured.partition.docx import partition_docx
|
|
|
|
PARTITION_WITH_EXTRAS_MAP["doc"] = partition_doc
|
|
PARTITION_WITH_EXTRAS_MAP["docx"] = partition_docx
|
|
|
|
|
|
if dependency_exists("docx") and dependency_exists("pypandoc"):
|
|
from unstructured.partition.odt import partition_odt
|
|
|
|
PARTITION_WITH_EXTRAS_MAP["odt"] = partition_odt
|
|
|
|
|
|
if dependency_exists("pypandoc"):
|
|
from unstructured.partition.epub import partition_epub
|
|
|
|
PARTITION_WITH_EXTRAS_MAP["epub"] = partition_epub
|
|
|
|
|
|
if dependency_exists("pypandoc"):
|
|
from unstructured.partition.org import partition_org
|
|
from unstructured.partition.rst import partition_rst
|
|
from unstructured.partition.rtf import partition_rtf
|
|
|
|
PARTITION_WITH_EXTRAS_MAP["org"] = partition_org
|
|
PARTITION_WITH_EXTRAS_MAP["rst"] = partition_rst
|
|
PARTITION_WITH_EXTRAS_MAP["rtf"] = partition_rtf
|
|
|
|
|
|
if dependency_exists("markdown"):
|
|
from unstructured.partition.md import partition_md
|
|
|
|
PARTITION_WITH_EXTRAS_MAP["md"] = partition_md
|
|
|
|
|
|
if dependency_exists("msg_parser"):
|
|
from unstructured.partition.msg import partition_msg
|
|
|
|
PARTITION_WITH_EXTRAS_MAP["msg"] = partition_msg
|
|
|
|
|
|
pdf_imports = ["pdf2image", "pdfminer", "PIL"]
|
|
if all(dependency_exists(dep) for dep in pdf_imports):
|
|
from unstructured.partition.pdf import partition_pdf
|
|
|
|
PARTITION_WITH_EXTRAS_MAP["pdf"] = partition_pdf
|
|
|
|
|
|
if dependency_exists("unstructured_inference"):
|
|
from unstructured.partition.image import partition_image
|
|
|
|
PARTITION_WITH_EXTRAS_MAP["image"] = partition_image
|
|
|
|
|
|
if dependency_exists("pptx"):
|
|
from unstructured.partition.ppt import partition_ppt
|
|
from unstructured.partition.pptx import partition_pptx
|
|
|
|
PARTITION_WITH_EXTRAS_MAP["ppt"] = partition_ppt
|
|
PARTITION_WITH_EXTRAS_MAP["pptx"] = partition_pptx
|
|
|
|
|
|
if dependency_exists("pandas") and dependency_exists("openpyxl"):
|
|
from unstructured.partition.xlsx import partition_xlsx
|
|
|
|
PARTITION_WITH_EXTRAS_MAP["xlsx"] = partition_xlsx
|
|
|
|
|
|
IMAGE_FILETYPES = [
|
|
FileType.HEIC,
|
|
FileType.PNG,
|
|
FileType.JPG,
|
|
FileType.TIFF,
|
|
FileType.BMP,
|
|
]
|
|
|
|
|
|
def _get_partition_with_extras(
|
|
doc_type: str,
|
|
partition_with_extras_map: Optional[dict[str, Callable[..., list[Element]]]] = None,
|
|
):
|
|
if partition_with_extras_map is None:
|
|
partition_with_extras_map = PARTITION_WITH_EXTRAS_MAP
|
|
_partition_func = partition_with_extras_map.get(doc_type)
|
|
if _partition_func is None:
|
|
raise ImportError(
|
|
f"partition_{doc_type} is not available. "
|
|
f"Install the {doc_type} dependencies with "
|
|
f'pip install "unstructured[{doc_type}]"',
|
|
)
|
|
return _partition_func
|
|
|
|
|
|
def partition(
|
|
filename: Optional[str] = None,
|
|
content_type: Optional[str] = None,
|
|
file: Optional[IO[bytes]] = None,
|
|
file_filename: Optional[str] = None,
|
|
url: Optional[str] = None,
|
|
include_page_breaks: bool = False,
|
|
strategy: str = PartitionStrategy.AUTO,
|
|
encoding: Optional[str] = None,
|
|
paragraph_grouper: Optional[Callable[[str], str]] | Literal[False] = None,
|
|
headers: dict[str, str] = {},
|
|
skip_infer_table_types: list[str] = ["pdf", "jpg", "png", "heic"],
|
|
ssl_verify: bool = True,
|
|
ocr_languages: Optional[str] = None, # changing to optional for deprecation
|
|
languages: Optional[list[str]] = None,
|
|
detect_language_per_element: bool = False,
|
|
pdf_infer_table_structure: bool = False,
|
|
extract_images_in_pdf: bool = False,
|
|
extract_image_block_types: Optional[list[str]] = None,
|
|
extract_image_block_output_dir: Optional[str] = None,
|
|
extract_image_block_to_payload: bool = False,
|
|
xml_keep_tags: bool = False,
|
|
data_source_metadata: Optional[DataSourceMetadata] = None,
|
|
metadata_filename: Optional[str] = None,
|
|
request_timeout: Optional[int] = None,
|
|
hi_res_model_name: Optional[str] = None,
|
|
model_name: Optional[str] = None, # to be deprecated
|
|
date_from_file_object: bool = False,
|
|
starting_page_number: int = 1,
|
|
**kwargs: Any,
|
|
):
|
|
"""Partitions a document into its constituent elements. Will use libmagic to determine
|
|
the file's type and route it to the appropriate partitioning function. Applies the default
|
|
parameters for each partitioning function. Use the document-type specific partitioning
|
|
functions if you need access to additional kwarg options.
|
|
|
|
Parameters
|
|
----------
|
|
filename
|
|
A string defining the target filename path.
|
|
content_type
|
|
A string defining the file content in MIME type
|
|
file
|
|
A file-like object using "rb" mode --> open(filename, "rb").
|
|
metadata_filename
|
|
When file is not None, the filename (string) to store in element metadata. E.g. "foo.txt"
|
|
url
|
|
The url for a remote document. Pass in content_type if you want partition to treat
|
|
the document as a specific content_type.
|
|
include_page_breaks
|
|
If True, the output will include page breaks if the filetype supports it
|
|
strategy
|
|
The strategy to use for partitioning PDF/image. Uses a layout detection model if set
|
|
to 'hi_res', otherwise partition simply extracts the text from the document
|
|
and processes it.
|
|
encoding
|
|
The encoding method used to decode the text input. If None, utf-8 will be used.
|
|
headers
|
|
The headers to be used in conjunction with the HTTP request if URL is set.
|
|
skip_infer_table_types
|
|
The document types that you want to skip table extraction with.
|
|
ssl_verify
|
|
If the URL parameter is set, determines whether or not partition uses SSL verification
|
|
in the HTTP request.
|
|
languages
|
|
The languages present in the document, for use in partitioning and/or OCR. For partitioning
|
|
image or pdf documents with Tesseract, you'll first need to install the appropriate
|
|
Tesseract language pack. For other partitions, language is detected using naive Bayesian
|
|
filter via `langdetect`. Multiple languages indicates text could be in either language.
|
|
Additional Parameters:
|
|
detect_language_per_element
|
|
Detect language per element instead of at the document level.
|
|
pdf_infer_table_structure
|
|
Deprecated! Use `skip_infer_table_types` to opt out of table extraction for any document
|
|
type.
|
|
If True and strategy=hi_res, any Table Elements extracted from a PDF will include an
|
|
additional metadata field, "text_as_html," where the value (string) is a just a
|
|
transformation of the data into an HTML <table>.
|
|
The "text" field for a partitioned Table Element is always present, whether True or False.
|
|
extract_images_in_pdf
|
|
Only applicable if `strategy=hi_res`.
|
|
If True, any detected images will be saved in the path specified by
|
|
'extract_image_block_output_dir' or stored as base64 encoded data within metadata fields.
|
|
Deprecation Note: This parameter is marked for deprecation. Future versions will use
|
|
'extract_image_block_types' for broader extraction capabilities.
|
|
extract_image_block_types
|
|
Only applicable if `strategy=hi_res`.
|
|
Images of the element type(s) specified in this list (e.g., ["Image", "Table"]) will be
|
|
saved in the path specified by 'extract_image_block_output_dir' or stored as base64
|
|
encoded data within metadata fields.
|
|
extract_image_block_to_payload
|
|
Only applicable if `strategy=hi_res`.
|
|
If True, images of the element type(s) defined in 'extract_image_block_types' will be
|
|
encoded as base64 data and stored in two metadata fields: 'image_base64' and
|
|
'image_mime_type'.
|
|
This parameter facilitates the inclusion of element data directly within the payload,
|
|
especially for web-based applications or APIs.
|
|
extract_image_block_output_dir
|
|
Only applicable if `strategy=hi_res` and `extract_image_block_to_payload=False`.
|
|
The filesystem path for saving images of the element type(s)
|
|
specified in 'extract_image_block_types'.
|
|
xml_keep_tags
|
|
If True, will retain the XML tags in the output. Otherwise it will simply extract
|
|
the text from within the tags. Only applies to partition_xml.
|
|
request_timeout
|
|
The timeout for the HTTP request if URL is set. Defaults to None meaning no timeout and
|
|
requests will block indefinitely.
|
|
hi_res_model_name
|
|
The layout detection model used when partitioning strategy is set to `hi_res`.
|
|
model_name
|
|
The layout detection model used when partitioning strategy is set to `hi_res`. To be
|
|
deprecated in favor of `hi_res_model_name`.
|
|
date_from_file_object
|
|
Applies only when providing file via `file` parameter. If this option is True and inference
|
|
from message header failed, attempt to infer last_modified metadata from bytes,
|
|
otherwise set it to None.
|
|
starting_page_number
|
|
Indicates what page number should be assigned to the first page in the document.
|
|
This information will be reflected in elements' metadata and can be be especially
|
|
useful when partitioning a document that is part of a larger document.
|
|
"""
|
|
exactly_one(file=file, filename=filename, url=url)
|
|
|
|
if metadata_filename and file_filename:
|
|
raise ValueError(
|
|
"Only one of metadata_filename and file_filename is specified. "
|
|
"metadata_filename is preferred. file_filename is marked for deprecation.",
|
|
)
|
|
|
|
if file_filename is not None:
|
|
metadata_filename = file_filename
|
|
logger.warn(
|
|
"The file_filename kwarg will be deprecated in a future version of unstructured. "
|
|
"Please use metadata_filename instead.",
|
|
)
|
|
kwargs.setdefault("metadata_filename", metadata_filename)
|
|
kwargs.setdefault("date_from_file_object", date_from_file_object)
|
|
|
|
if pdf_infer_table_structure:
|
|
logger.warning(
|
|
"The pdf_infer_table_structure kwarg is deprecated. Please use skip_infer_table_types "
|
|
"instead."
|
|
)
|
|
|
|
languages = check_language_args(languages or [], ocr_languages)
|
|
|
|
if url is not None:
|
|
file, filetype = file_and_type_from_url(
|
|
url=url,
|
|
content_type=content_type,
|
|
headers=headers,
|
|
ssl_verify=ssl_verify,
|
|
request_timeout=request_timeout,
|
|
)
|
|
else:
|
|
if headers != {}:
|
|
logger.warning(
|
|
"The headers kwarg is set but the url kwarg is not. "
|
|
"The headers kwarg will be ignored.",
|
|
)
|
|
filetype = detect_filetype(
|
|
filename=filename,
|
|
file=file,
|
|
file_filename=metadata_filename,
|
|
content_type=content_type,
|
|
encoding=encoding,
|
|
)
|
|
|
|
if file is not None:
|
|
file.seek(0)
|
|
|
|
infer_table_structure = decide_table_extraction(
|
|
filetype,
|
|
skip_infer_table_types,
|
|
pdf_infer_table_structure,
|
|
)
|
|
|
|
if filetype == FileType.DOC:
|
|
_partition_doc = _get_partition_with_extras("doc")
|
|
elements = _partition_doc(
|
|
filename=filename,
|
|
file=file,
|
|
infer_table_structure=infer_table_structure,
|
|
languages=languages,
|
|
detect_language_per_element=detect_language_per_element,
|
|
starting_page_number=starting_page_number,
|
|
**kwargs,
|
|
)
|
|
elif filetype == FileType.DOCX:
|
|
_partition_docx = _get_partition_with_extras("docx")
|
|
elements = _partition_docx(
|
|
filename=filename,
|
|
file=file,
|
|
infer_table_structure=infer_table_structure,
|
|
languages=languages,
|
|
detect_language_per_element=detect_language_per_element,
|
|
starting_page_number=starting_page_number,
|
|
**kwargs,
|
|
)
|
|
elif filetype == FileType.ODT:
|
|
_partition_odt = _get_partition_with_extras("odt")
|
|
elements = _partition_odt(
|
|
filename=filename,
|
|
file=file,
|
|
infer_table_structure=infer_table_structure,
|
|
languages=languages,
|
|
detect_language_per_element=detect_language_per_element,
|
|
starting_page_number=starting_page_number,
|
|
**kwargs,
|
|
)
|
|
elif filetype == FileType.EML:
|
|
elements = partition_email(
|
|
filename=filename,
|
|
file=file,
|
|
encoding=encoding,
|
|
languages=languages,
|
|
detect_language_per_element=detect_language_per_element,
|
|
**kwargs,
|
|
)
|
|
elif filetype == FileType.MSG:
|
|
_partition_msg = _get_partition_with_extras("msg")
|
|
elements = _partition_msg(
|
|
filename=filename,
|
|
file=file,
|
|
languages=languages,
|
|
detect_language_per_element=detect_language_per_element,
|
|
**kwargs,
|
|
)
|
|
elif filetype == FileType.HTML:
|
|
elements = partition_html(
|
|
filename=filename,
|
|
file=file,
|
|
include_page_breaks=include_page_breaks,
|
|
encoding=encoding,
|
|
languages=languages,
|
|
detect_language_per_element=detect_language_per_element,
|
|
**kwargs,
|
|
)
|
|
elif filetype == FileType.XML:
|
|
elements = partition_xml(
|
|
filename=filename,
|
|
file=file,
|
|
encoding=encoding,
|
|
xml_keep_tags=xml_keep_tags,
|
|
languages=languages,
|
|
detect_language_per_element=detect_language_per_element,
|
|
**kwargs,
|
|
)
|
|
elif filetype == FileType.EPUB:
|
|
_partition_epub = _get_partition_with_extras("epub")
|
|
elements = _partition_epub(
|
|
filename=filename,
|
|
file=file,
|
|
include_page_breaks=include_page_breaks,
|
|
infer_table_structure=infer_table_structure,
|
|
languages=languages,
|
|
detect_language_per_element=detect_language_per_element,
|
|
**kwargs,
|
|
)
|
|
elif filetype == FileType.ORG:
|
|
_partition_org = _get_partition_with_extras("org")
|
|
elements = _partition_org(
|
|
filename=filename,
|
|
file=file,
|
|
include_page_breaks=include_page_breaks,
|
|
languages=languages,
|
|
detect_language_per_element=detect_language_per_element,
|
|
**kwargs,
|
|
)
|
|
elif filetype == FileType.RST:
|
|
_partition_rst = _get_partition_with_extras("rst")
|
|
elements = _partition_rst(
|
|
filename=filename,
|
|
file=file,
|
|
include_page_breaks=include_page_breaks,
|
|
infer_table_structure=infer_table_structure,
|
|
languages=languages,
|
|
detect_language_per_element=detect_language_per_element,
|
|
**kwargs,
|
|
)
|
|
elif filetype == FileType.MD:
|
|
_partition_md = _get_partition_with_extras("md")
|
|
elements = _partition_md(
|
|
filename=filename,
|
|
file=file,
|
|
include_page_breaks=include_page_breaks,
|
|
infer_table_structure=infer_table_structure,
|
|
languages=languages,
|
|
detect_language_per_element=detect_language_per_element,
|
|
**kwargs,
|
|
)
|
|
elif filetype == FileType.PDF:
|
|
_partition_pdf = _get_partition_with_extras("pdf")
|
|
elements = _partition_pdf(
|
|
filename=filename,
|
|
file=file,
|
|
url=None,
|
|
include_page_breaks=include_page_breaks,
|
|
infer_table_structure=infer_table_structure,
|
|
strategy=strategy,
|
|
languages=languages,
|
|
hi_res_model_name=hi_res_model_name or model_name,
|
|
extract_images_in_pdf=extract_images_in_pdf,
|
|
extract_image_block_types=extract_image_block_types,
|
|
extract_image_block_output_dir=extract_image_block_output_dir,
|
|
extract_image_block_to_payload=extract_image_block_to_payload,
|
|
starting_page_number=starting_page_number,
|
|
**kwargs,
|
|
)
|
|
elif filetype in IMAGE_FILETYPES:
|
|
_partition_image = _get_partition_with_extras("image")
|
|
elements = _partition_image(
|
|
filename=filename,
|
|
file=file,
|
|
url=None,
|
|
include_page_breaks=include_page_breaks,
|
|
infer_table_structure=infer_table_structure,
|
|
strategy=strategy,
|
|
languages=languages,
|
|
hi_res_model_name=hi_res_model_name or model_name,
|
|
extract_images_in_pdf=extract_images_in_pdf,
|
|
extract_image_block_types=extract_image_block_types,
|
|
extract_image_block_output_dir=extract_image_block_output_dir,
|
|
extract_image_block_to_payload=extract_image_block_to_payload,
|
|
starting_page_number=starting_page_number,
|
|
**kwargs,
|
|
)
|
|
elif filetype == FileType.TXT:
|
|
elements = partition_text(
|
|
filename=filename,
|
|
file=file,
|
|
encoding=encoding,
|
|
paragraph_grouper=paragraph_grouper,
|
|
languages=languages,
|
|
detect_language_per_element=detect_language_per_element,
|
|
**kwargs,
|
|
)
|
|
elif filetype == FileType.RTF:
|
|
_partition_rtf = _get_partition_with_extras("rtf")
|
|
elements = _partition_rtf(
|
|
filename=filename,
|
|
file=file,
|
|
include_page_breaks=include_page_breaks,
|
|
infer_table_structure=infer_table_structure,
|
|
languages=languages,
|
|
detect_language_per_element=detect_language_per_element,
|
|
**kwargs,
|
|
)
|
|
elif filetype == FileType.PPT:
|
|
_partition_ppt = _get_partition_with_extras("ppt")
|
|
elements = _partition_ppt(
|
|
filename=filename,
|
|
file=file,
|
|
include_page_breaks=include_page_breaks,
|
|
infer_table_structure=infer_table_structure,
|
|
languages=languages,
|
|
detect_language_per_element=detect_language_per_element,
|
|
**kwargs,
|
|
)
|
|
elif filetype == FileType.PPTX:
|
|
_partition_pptx = _get_partition_with_extras("pptx")
|
|
elements = _partition_pptx(
|
|
filename=filename,
|
|
file=file,
|
|
include_page_breaks=include_page_breaks,
|
|
infer_table_structure=infer_table_structure,
|
|
languages=languages,
|
|
detect_language_per_element=detect_language_per_element,
|
|
starting_page_number=starting_page_number,
|
|
**kwargs,
|
|
)
|
|
elif filetype == FileType.JSON:
|
|
if not is_json_processable(filename=filename, file=file):
|
|
raise ValueError(
|
|
"Detected a JSON file that does not conform to the Unstructured schema. "
|
|
"partition_json currently only processes serialized Unstructured output.",
|
|
)
|
|
elements = partition_json(filename=filename, file=file, **kwargs)
|
|
elif (filetype == FileType.XLSX) or (filetype == FileType.XLS):
|
|
_partition_xlsx = _get_partition_with_extras("xlsx")
|
|
elements = _partition_xlsx(
|
|
filename=filename,
|
|
file=file,
|
|
infer_table_structure=infer_table_structure,
|
|
languages=languages,
|
|
detect_language_per_element=detect_language_per_element,
|
|
starting_page_number=starting_page_number,
|
|
**kwargs,
|
|
)
|
|
elif filetype == FileType.CSV:
|
|
_partition_csv = _get_partition_with_extras("csv")
|
|
elements = _partition_csv(
|
|
filename=filename,
|
|
file=file,
|
|
infer_table_structure=infer_table_structure,
|
|
languages=languages,
|
|
detect_language_per_element=detect_language_per_element,
|
|
**kwargs,
|
|
)
|
|
elif filetype == FileType.TSV:
|
|
_partition_tsv = _get_partition_with_extras("tsv")
|
|
elements = _partition_tsv(
|
|
filename=filename,
|
|
file=file,
|
|
languages=languages,
|
|
detect_language_per_element=detect_language_per_element,
|
|
**kwargs,
|
|
)
|
|
elif filetype == FileType.EMPTY:
|
|
elements = []
|
|
else:
|
|
msg = "Invalid file" if not filename else f"Invalid file {filename}"
|
|
raise ValueError(f"{msg}. The {filetype} file type is not supported in partition.")
|
|
|
|
for element in elements:
|
|
element.metadata.url = url
|
|
element.metadata.data_source = data_source_metadata
|
|
if content_type is not None:
|
|
out_filetype = STR_TO_FILETYPE.get(content_type)
|
|
element.metadata.filetype = (
|
|
FILETYPE_TO_MIMETYPE[out_filetype] if out_filetype is not None else None
|
|
)
|
|
else:
|
|
element.metadata.filetype = FILETYPE_TO_MIMETYPE[filetype]
|
|
|
|
return elements
|
|
|
|
|
|
def file_and_type_from_url(
|
|
url: str,
|
|
content_type: Optional[str] = None,
|
|
headers: dict[str, str] = {},
|
|
ssl_verify: bool = True,
|
|
request_timeout: Optional[int] = None,
|
|
) -> tuple[io.BytesIO, Optional[FileType]]:
|
|
response = requests.get(url, headers=headers, verify=ssl_verify, timeout=request_timeout)
|
|
file = io.BytesIO(response.content)
|
|
|
|
content_type = (
|
|
content_type or response.headers.get("Content-Type", "").split(";")[0].strip().lower()
|
|
)
|
|
encoding = response.headers.get("Content-Encoding", "utf-8")
|
|
|
|
filetype = detect_filetype(file=file, content_type=content_type, encoding=encoding)
|
|
return file, filetype
|
|
|
|
|
|
def decide_table_extraction(
|
|
filetype: Optional[FileType],
|
|
skip_infer_table_types: list[str],
|
|
pdf_infer_table_structure: bool,
|
|
) -> bool:
|
|
doc_type = filetype.name.lower() if filetype else None
|
|
|
|
if doc_type == "pdf":
|
|
# For backwards compatibility. Ultimately we want to remove pdf_infer_table_structure
|
|
# completely and rely exclusively on `skip_infer_table_types` for all file types.
|
|
# Until then for pdf files we first check pdf_infer_table_structure and then update
|
|
# based on skip_infer_tables.
|
|
return pdf_infer_table_structure or doc_type not in skip_infer_table_types
|
|
|
|
return doc_type not in skip_infer_table_types
|