|
|
""" |
|
|
A PDF Matplotlib backend. |
|
|
|
|
|
Author: Jouni K Seppänen <jks@iki.fi> and others. |
|
|
""" |
|
|
|
|
|
import codecs |
|
|
from datetime import timezone |
|
|
from datetime import datetime |
|
|
from enum import Enum |
|
|
from functools import total_ordering |
|
|
from io import BytesIO |
|
|
import itertools |
|
|
import logging |
|
|
import math |
|
|
import os |
|
|
import string |
|
|
import struct |
|
|
import sys |
|
|
import time |
|
|
import types |
|
|
import warnings |
|
|
import zlib |
|
|
|
|
|
import numpy as np |
|
|
from PIL import Image |
|
|
|
|
|
import matplotlib as mpl |
|
|
from matplotlib import _api, _text_helpers, _type1font, cbook, dviread |
|
|
from matplotlib._pylab_helpers import Gcf |
|
|
from matplotlib.backend_bases import ( |
|
|
_Backend, FigureCanvasBase, FigureManagerBase, GraphicsContextBase, |
|
|
RendererBase) |
|
|
from matplotlib.backends.backend_mixed import MixedModeRenderer |
|
|
from matplotlib.figure import Figure |
|
|
from matplotlib.font_manager import get_font, fontManager as _fontManager |
|
|
from matplotlib._afm import AFM |
|
|
from matplotlib.ft2font import (FIXED_WIDTH, ITALIC, LOAD_NO_SCALE, |
|
|
LOAD_NO_HINTING, KERNING_UNFITTED, FT2Font) |
|
|
from matplotlib.transforms import Affine2D, BboxBase |
|
|
from matplotlib.path import Path |
|
|
from matplotlib.dates import UTC |
|
|
from matplotlib import _path |
|
|
from . import _backend_pdf_ps |
|
|
|
|
|
_log = logging.getLogger(__name__) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
@_api.deprecated("3.6", alternative="a vendored copy of _fill") |
|
|
def fill(strings, linelen=75): |
|
|
return _fill(strings, linelen=linelen) |
|
|
|
|
|
|
|
|
def _fill(strings, linelen=75): |
|
|
""" |
|
|
Make one string from sequence of strings, with whitespace in between. |
|
|
|
|
|
The whitespace is chosen to form lines of at most *linelen* characters, |
|
|
if possible. |
|
|
""" |
|
|
currpos = 0 |
|
|
lasti = 0 |
|
|
result = [] |
|
|
for i, s in enumerate(strings): |
|
|
length = len(s) |
|
|
if currpos + length < linelen: |
|
|
currpos += length + 1 |
|
|
else: |
|
|
result.append(b' '.join(strings[lasti:i])) |
|
|
lasti = i |
|
|
currpos = length |
|
|
result.append(b' '.join(strings[lasti:])) |
|
|
return b'\n'.join(result) |
|
|
|
|
|
|
|
|
def _create_pdf_info_dict(backend, metadata): |
|
|
""" |
|
|
Create a PDF infoDict based on user-supplied metadata. |
|
|
|
|
|
A default ``Creator``, ``Producer``, and ``CreationDate`` are added, though |
|
|
the user metadata may override it. The date may be the current time, or a |
|
|
time set by the ``SOURCE_DATE_EPOCH`` environment variable. |
|
|
|
|
|
Metadata is verified to have the correct keys and their expected types. Any |
|
|
unknown keys/types will raise a warning. |
|
|
|
|
|
Parameters |
|
|
---------- |
|
|
backend : str |
|
|
The name of the backend to use in the Producer value. |
|
|
|
|
|
metadata : dict[str, Union[str, datetime, Name]] |
|
|
A dictionary of metadata supplied by the user with information |
|
|
following the PDF specification, also defined in |
|
|
`~.backend_pdf.PdfPages` below. |
|
|
|
|
|
If any value is *None*, then the key will be removed. This can be used |
|
|
to remove any pre-defined values. |
|
|
|
|
|
Returns |
|
|
------- |
|
|
dict[str, Union[str, datetime, Name]] |
|
|
A validated dictionary of metadata. |
|
|
""" |
|
|
|
|
|
|
|
|
|
|
|
source_date_epoch = os.getenv("SOURCE_DATE_EPOCH") |
|
|
if source_date_epoch: |
|
|
source_date = datetime.fromtimestamp(int(source_date_epoch), timezone.utc) |
|
|
source_date = source_date.replace(tzinfo=UTC) |
|
|
else: |
|
|
source_date = datetime.today() |
|
|
|
|
|
info = { |
|
|
'Creator': f'Matplotlib v{mpl.__version__}, https://matplotlib.org', |
|
|
'Producer': f'Matplotlib {backend} backend v{mpl.__version__}', |
|
|
'CreationDate': source_date, |
|
|
**metadata |
|
|
} |
|
|
info = {k: v for (k, v) in info.items() if v is not None} |
|
|
|
|
|
def is_string_like(x): |
|
|
return isinstance(x, str) |
|
|
is_string_like.text_for_warning = "an instance of str" |
|
|
|
|
|
def is_date(x): |
|
|
return isinstance(x, datetime) |
|
|
is_date.text_for_warning = "an instance of datetime.datetime" |
|
|
|
|
|
def check_trapped(x): |
|
|
if isinstance(x, Name): |
|
|
return x.name in (b'True', b'False', b'Unknown') |
|
|
else: |
|
|
return x in ('True', 'False', 'Unknown') |
|
|
check_trapped.text_for_warning = 'one of {"True", "False", "Unknown"}' |
|
|
|
|
|
keywords = { |
|
|
'Title': is_string_like, |
|
|
'Author': is_string_like, |
|
|
'Subject': is_string_like, |
|
|
'Keywords': is_string_like, |
|
|
'Creator': is_string_like, |
|
|
'Producer': is_string_like, |
|
|
'CreationDate': is_date, |
|
|
'ModDate': is_date, |
|
|
'Trapped': check_trapped, |
|
|
} |
|
|
for k in info: |
|
|
if k not in keywords: |
|
|
_api.warn_external(f'Unknown infodict keyword: {k!r}. ' |
|
|
f'Must be one of {set(keywords)!r}.') |
|
|
elif not keywords[k](info[k]): |
|
|
_api.warn_external(f'Bad value for infodict keyword {k}. ' |
|
|
f'Got {info[k]!r} which is not ' |
|
|
f'{keywords[k].text_for_warning}.') |
|
|
if 'Trapped' in info: |
|
|
info['Trapped'] = Name(info['Trapped']) |
|
|
|
|
|
return info |
|
|
|
|
|
|
|
|
def _datetime_to_pdf(d): |
|
|
""" |
|
|
Convert a datetime to a PDF string representing it. |
|
|
|
|
|
Used for PDF and PGF. |
|
|
""" |
|
|
r = d.strftime('D:%Y%m%d%H%M%S') |
|
|
z = d.utcoffset() |
|
|
if z is not None: |
|
|
z = z.seconds |
|
|
else: |
|
|
if time.daylight: |
|
|
z = time.altzone |
|
|
else: |
|
|
z = time.timezone |
|
|
if z == 0: |
|
|
r += 'Z' |
|
|
elif z < 0: |
|
|
r += "+%02d'%02d'" % ((-z) // 3600, (-z) % 3600) |
|
|
else: |
|
|
r += "-%02d'%02d'" % (z // 3600, z % 3600) |
|
|
return r |
|
|
|
|
|
|
|
|
def _calculate_quad_point_coordinates(x, y, width, height, angle=0): |
|
|
""" |
|
|
Calculate the coordinates of rectangle when rotated by angle around x, y |
|
|
""" |
|
|
|
|
|
angle = math.radians(-angle) |
|
|
sin_angle = math.sin(angle) |
|
|
cos_angle = math.cos(angle) |
|
|
a = x + height * sin_angle |
|
|
b = y + height * cos_angle |
|
|
c = x + width * cos_angle + height * sin_angle |
|
|
d = y - width * sin_angle + height * cos_angle |
|
|
e = x + width * cos_angle |
|
|
f = y - width * sin_angle |
|
|
return ((x, y), (e, f), (c, d), (a, b)) |
|
|
|
|
|
|
|
|
def _get_coordinates_of_block(x, y, width, height, angle=0): |
|
|
""" |
|
|
Get the coordinates of rotated rectangle and rectangle that covers the |
|
|
rotated rectangle. |
|
|
""" |
|
|
|
|
|
vertices = _calculate_quad_point_coordinates(x, y, width, |
|
|
height, angle) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
pad = 0.00001 if angle % 90 else 0 |
|
|
min_x = min(v[0] for v in vertices) - pad |
|
|
min_y = min(v[1] for v in vertices) - pad |
|
|
max_x = max(v[0] for v in vertices) + pad |
|
|
max_y = max(v[1] for v in vertices) + pad |
|
|
return (tuple(itertools.chain.from_iterable(vertices)), |
|
|
(min_x, min_y, max_x, max_y)) |
|
|
|
|
|
|
|
|
def _get_link_annotation(gc, x, y, width, height, angle=0): |
|
|
""" |
|
|
Create a link annotation object for embedding URLs. |
|
|
""" |
|
|
quadpoints, rect = _get_coordinates_of_block(x, y, width, height, angle) |
|
|
link_annotation = { |
|
|
'Type': Name('Annot'), |
|
|
'Subtype': Name('Link'), |
|
|
'Rect': rect, |
|
|
'Border': [0, 0, 0], |
|
|
'A': { |
|
|
'S': Name('URI'), |
|
|
'URI': gc.get_url(), |
|
|
}, |
|
|
} |
|
|
if angle % 90: |
|
|
|
|
|
link_annotation['QuadPoints'] = quadpoints |
|
|
return link_annotation |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
_str_escapes = str.maketrans({ |
|
|
'\\': '\\\\', '(': '\\(', ')': '\\)', '\n': '\\n', '\r': '\\r'}) |
|
|
|
|
|
|
|
|
def pdfRepr(obj): |
|
|
"""Map Python objects to PDF syntax.""" |
|
|
|
|
|
|
|
|
if hasattr(obj, 'pdfRepr'): |
|
|
return obj.pdfRepr() |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
elif isinstance(obj, (float, np.floating)): |
|
|
if not np.isfinite(obj): |
|
|
raise ValueError("Can only output finite numbers in PDF") |
|
|
r = b"%.10f" % obj |
|
|
return r.rstrip(b'0').rstrip(b'.') |
|
|
|
|
|
|
|
|
|
|
|
elif isinstance(obj, bool): |
|
|
return [b'false', b'true'][obj] |
|
|
|
|
|
|
|
|
elif isinstance(obj, (int, np.integer)): |
|
|
return b"%d" % obj |
|
|
|
|
|
|
|
|
elif isinstance(obj, str): |
|
|
return pdfRepr(obj.encode('ascii') if obj.isascii() |
|
|
else codecs.BOM_UTF16_BE + obj.encode('UTF-16BE')) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
elif isinstance(obj, bytes): |
|
|
return ( |
|
|
b'(' + |
|
|
obj.decode('latin-1').translate(_str_escapes).encode('latin-1') |
|
|
+ b')') |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
elif isinstance(obj, dict): |
|
|
return _fill([ |
|
|
b"<<", |
|
|
*[Name(k).pdfRepr() + b" " + pdfRepr(v) for k, v in obj.items()], |
|
|
b">>", |
|
|
]) |
|
|
|
|
|
|
|
|
elif isinstance(obj, (list, tuple)): |
|
|
return _fill([b"[", *[pdfRepr(val) for val in obj], b"]"]) |
|
|
|
|
|
|
|
|
elif obj is None: |
|
|
return b'null' |
|
|
|
|
|
|
|
|
elif isinstance(obj, datetime): |
|
|
return pdfRepr(_datetime_to_pdf(obj)) |
|
|
|
|
|
|
|
|
elif isinstance(obj, BboxBase): |
|
|
return _fill([pdfRepr(val) for val in obj.bounds]) |
|
|
|
|
|
else: |
|
|
raise TypeError("Don't know a PDF representation for {} objects" |
|
|
.format(type(obj))) |
|
|
|
|
|
|
|
|
def _font_supports_glyph(fonttype, glyph): |
|
|
""" |
|
|
Returns True if the font is able to provide codepoint *glyph* in a PDF. |
|
|
|
|
|
For a Type 3 font, this method returns True only for single-byte |
|
|
characters. For Type 42 fonts this method return True if the character is |
|
|
from the Basic Multilingual Plane. |
|
|
""" |
|
|
if fonttype == 3: |
|
|
return glyph <= 255 |
|
|
if fonttype == 42: |
|
|
return glyph <= 65535 |
|
|
raise NotImplementedError() |
|
|
|
|
|
|
|
|
class Reference: |
|
|
""" |
|
|
PDF reference object. |
|
|
|
|
|
Use PdfFile.reserveObject() to create References. |
|
|
""" |
|
|
|
|
|
def __init__(self, id): |
|
|
self.id = id |
|
|
|
|
|
def __repr__(self): |
|
|
return "<Reference %d>" % self.id |
|
|
|
|
|
def pdfRepr(self): |
|
|
return b"%d 0 R" % self.id |
|
|
|
|
|
def write(self, contents, file): |
|
|
write = file.write |
|
|
write(b"%d 0 obj\n" % self.id) |
|
|
write(pdfRepr(contents)) |
|
|
write(b"\nendobj\n") |
|
|
|
|
|
|
|
|
@total_ordering |
|
|
class Name: |
|
|
"""PDF name object.""" |
|
|
__slots__ = ('name',) |
|
|
_hexify = {c: '#%02x' % c |
|
|
for c in {*range(256)} - {*range(ord('!'), ord('~') + 1)}} |
|
|
|
|
|
def __init__(self, name): |
|
|
if isinstance(name, Name): |
|
|
self.name = name.name |
|
|
else: |
|
|
if isinstance(name, bytes): |
|
|
name = name.decode('ascii') |
|
|
self.name = name.translate(self._hexify).encode('ascii') |
|
|
|
|
|
def __repr__(self): |
|
|
return "<Name %s>" % self.name |
|
|
|
|
|
def __str__(self): |
|
|
return '/' + self.name.decode('ascii') |
|
|
|
|
|
def __eq__(self, other): |
|
|
return isinstance(other, Name) and self.name == other.name |
|
|
|
|
|
def __lt__(self, other): |
|
|
return isinstance(other, Name) and self.name < other.name |
|
|
|
|
|
def __hash__(self): |
|
|
return hash(self.name) |
|
|
|
|
|
@staticmethod |
|
|
@_api.deprecated("3.6") |
|
|
def hexify(match): |
|
|
return '#%02x' % ord(match.group()) |
|
|
|
|
|
def pdfRepr(self): |
|
|
return b'/' + self.name |
|
|
|
|
|
|
|
|
@_api.deprecated("3.6") |
|
|
class Operator: |
|
|
__slots__ = ('op',) |
|
|
|
|
|
def __init__(self, op): |
|
|
self.op = op |
|
|
|
|
|
def __repr__(self): |
|
|
return '<Operator %s>' % self.op |
|
|
|
|
|
def pdfRepr(self): |
|
|
return self.op |
|
|
|
|
|
|
|
|
class Verbatim: |
|
|
"""Store verbatim PDF command content for later inclusion in the stream.""" |
|
|
def __init__(self, x): |
|
|
self._x = x |
|
|
|
|
|
def pdfRepr(self): |
|
|
return self._x |
|
|
|
|
|
|
|
|
class Op(Enum): |
|
|
"""PDF operators (not an exhaustive list).""" |
|
|
|
|
|
close_fill_stroke = b'b' |
|
|
fill_stroke = b'B' |
|
|
fill = b'f' |
|
|
closepath = b'h' |
|
|
close_stroke = b's' |
|
|
stroke = b'S' |
|
|
endpath = b'n' |
|
|
begin_text = b'BT' |
|
|
end_text = b'ET' |
|
|
curveto = b'c' |
|
|
rectangle = b're' |
|
|
lineto = b'l' |
|
|
moveto = b'm' |
|
|
concat_matrix = b'cm' |
|
|
use_xobject = b'Do' |
|
|
setgray_stroke = b'G' |
|
|
setgray_nonstroke = b'g' |
|
|
setrgb_stroke = b'RG' |
|
|
setrgb_nonstroke = b'rg' |
|
|
setcolorspace_stroke = b'CS' |
|
|
setcolorspace_nonstroke = b'cs' |
|
|
setcolor_stroke = b'SCN' |
|
|
setcolor_nonstroke = b'scn' |
|
|
setdash = b'd' |
|
|
setlinejoin = b'j' |
|
|
setlinecap = b'J' |
|
|
setgstate = b'gs' |
|
|
gsave = b'q' |
|
|
grestore = b'Q' |
|
|
textpos = b'Td' |
|
|
selectfont = b'Tf' |
|
|
textmatrix = b'Tm' |
|
|
show = b'Tj' |
|
|
showkern = b'TJ' |
|
|
setlinewidth = b'w' |
|
|
clip = b'W' |
|
|
shading = b'sh' |
|
|
|
|
|
op = _api.deprecated('3.6')(property(lambda self: self.value)) |
|
|
|
|
|
def pdfRepr(self): |
|
|
return self.value |
|
|
|
|
|
@classmethod |
|
|
def paint_path(cls, fill, stroke): |
|
|
""" |
|
|
Return the PDF operator to paint a path. |
|
|
|
|
|
Parameters |
|
|
---------- |
|
|
fill : bool |
|
|
Fill the path with the fill color. |
|
|
stroke : bool |
|
|
Stroke the outline of the path with the line color. |
|
|
""" |
|
|
if stroke: |
|
|
if fill: |
|
|
return cls.fill_stroke |
|
|
else: |
|
|
return cls.stroke |
|
|
else: |
|
|
if fill: |
|
|
return cls.fill |
|
|
else: |
|
|
return cls.endpath |
|
|
|
|
|
|
|
|
class Stream: |
|
|
""" |
|
|
PDF stream object. |
|
|
|
|
|
This has no pdfRepr method. Instead, call begin(), then output the |
|
|
contents of the stream by calling write(), and finally call end(). |
|
|
""" |
|
|
__slots__ = ('id', 'len', 'pdfFile', 'file', 'compressobj', 'extra', 'pos') |
|
|
|
|
|
def __init__(self, id, len, file, extra=None, png=None): |
|
|
""" |
|
|
Parameters |
|
|
---------- |
|
|
id : int |
|
|
Object id of the stream. |
|
|
len : Reference or None |
|
|
An unused Reference object for the length of the stream; |
|
|
None means to use a memory buffer so the length can be inlined. |
|
|
file : PdfFile |
|
|
The underlying object to write the stream to. |
|
|
extra : dict from Name to anything, or None |
|
|
Extra key-value pairs to include in the stream header. |
|
|
png : dict or None |
|
|
If the data is already png encoded, the decode parameters. |
|
|
""" |
|
|
self.id = id |
|
|
self.len = len |
|
|
self.pdfFile = file |
|
|
self.file = file.fh |
|
|
self.compressobj = None |
|
|
if extra is None: |
|
|
self.extra = dict() |
|
|
else: |
|
|
self.extra = extra.copy() |
|
|
if png is not None: |
|
|
self.extra.update({'Filter': Name('FlateDecode'), |
|
|
'DecodeParms': png}) |
|
|
|
|
|
self.pdfFile.recordXref(self.id) |
|
|
if mpl.rcParams['pdf.compression'] and not png: |
|
|
self.compressobj = zlib.compressobj( |
|
|
mpl.rcParams['pdf.compression']) |
|
|
if self.len is None: |
|
|
self.file = BytesIO() |
|
|
else: |
|
|
self._writeHeader() |
|
|
self.pos = self.file.tell() |
|
|
|
|
|
def _writeHeader(self): |
|
|
write = self.file.write |
|
|
write(b"%d 0 obj\n" % self.id) |
|
|
dict = self.extra |
|
|
dict['Length'] = self.len |
|
|
if mpl.rcParams['pdf.compression']: |
|
|
dict['Filter'] = Name('FlateDecode') |
|
|
|
|
|
write(pdfRepr(dict)) |
|
|
write(b"\nstream\n") |
|
|
|
|
|
def end(self): |
|
|
"""Finalize stream.""" |
|
|
|
|
|
self._flush() |
|
|
if self.len is None: |
|
|
contents = self.file.getvalue() |
|
|
self.len = len(contents) |
|
|
self.file = self.pdfFile.fh |
|
|
self._writeHeader() |
|
|
self.file.write(contents) |
|
|
self.file.write(b"\nendstream\nendobj\n") |
|
|
else: |
|
|
length = self.file.tell() - self.pos |
|
|
self.file.write(b"\nendstream\nendobj\n") |
|
|
self.pdfFile.writeObject(self.len, length) |
|
|
|
|
|
def write(self, data): |
|
|
"""Write some data on the stream.""" |
|
|
|
|
|
if self.compressobj is None: |
|
|
self.file.write(data) |
|
|
else: |
|
|
compressed = self.compressobj.compress(data) |
|
|
self.file.write(compressed) |
|
|
|
|
|
def _flush(self): |
|
|
"""Flush the compression object.""" |
|
|
|
|
|
if self.compressobj is not None: |
|
|
compressed = self.compressobj.flush() |
|
|
self.file.write(compressed) |
|
|
self.compressobj = None |
|
|
|
|
|
|
|
|
def _get_pdf_charprocs(font_path, glyph_ids): |
|
|
font = get_font(font_path, hinting_factor=1) |
|
|
conv = 1000 / font.units_per_EM |
|
|
procs = {} |
|
|
for glyph_id in glyph_ids: |
|
|
g = font.load_glyph(glyph_id, LOAD_NO_SCALE) |
|
|
|
|
|
|
|
|
|
|
|
d1 = (np.array([g.horiAdvance, 0, *g.bbox]) * conv + .5).astype(int) |
|
|
v, c = font.get_path() |
|
|
v = (v * 64).astype(int) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
quads, = np.nonzero(c == 3) |
|
|
quads_on = quads[1::2] |
|
|
quads_mid_on = np.array( |
|
|
sorted({*quads_on} & {*(quads - 1)} & {*(quads + 1)}), int) |
|
|
implicit = quads_mid_on[ |
|
|
(v[quads_mid_on] |
|
|
== ((v[quads_mid_on - 1] + v[quads_mid_on + 1]) / 2).astype(int)) |
|
|
.all(axis=1)] |
|
|
if (font.postscript_name, glyph_id) in [ |
|
|
("DejaVuSerif-Italic", 77), |
|
|
("DejaVuSerif-Italic", 135), |
|
|
]: |
|
|
v[:, 0] -= 1 |
|
|
v = (v * conv + .5).astype(int) |
|
|
v[implicit] = (( |
|
|
(v[implicit - 1] + v[implicit + 1]) / 2).astype(int)) |
|
|
procs[font.get_glyph_name(glyph_id)] = ( |
|
|
" ".join(map(str, d1)).encode("ascii") + b" d1\n" |
|
|
+ _path.convert_to_string( |
|
|
Path(v, c), None, None, False, None, -1, |
|
|
|
|
|
[b"m", b"l", b"", b"c", b"h"], True) |
|
|
+ b"f") |
|
|
return procs |
|
|
|
|
|
|
|
|
class PdfFile: |
|
|
"""PDF file object.""" |
|
|
|
|
|
def __init__(self, filename, metadata=None): |
|
|
""" |
|
|
Parameters |
|
|
---------- |
|
|
filename : str or path-like or file-like |
|
|
Output target; if a string, a file will be opened for writing. |
|
|
|
|
|
metadata : dict from strings to strings and dates |
|
|
Information dictionary object (see PDF reference section 10.2.1 |
|
|
'Document Information Dictionary'), e.g.: |
|
|
``{'Creator': 'My software', 'Author': 'Me', 'Title': 'Awesome'}``. |
|
|
|
|
|
The standard keys are 'Title', 'Author', 'Subject', 'Keywords', |
|
|
'Creator', 'Producer', 'CreationDate', 'ModDate', and |
|
|
'Trapped'. Values have been predefined for 'Creator', 'Producer' |
|
|
and 'CreationDate'. They can be removed by setting them to `None`. |
|
|
""" |
|
|
super().__init__() |
|
|
|
|
|
self._object_seq = itertools.count(1) |
|
|
self.xrefTable = [[0, 65535, 'the zero object']] |
|
|
self.passed_in_file_object = False |
|
|
self.original_file_like = None |
|
|
self.tell_base = 0 |
|
|
fh, opened = cbook.to_filehandle(filename, "wb", return_opened=True) |
|
|
if not opened: |
|
|
try: |
|
|
self.tell_base = filename.tell() |
|
|
except IOError: |
|
|
fh = BytesIO() |
|
|
self.original_file_like = filename |
|
|
else: |
|
|
fh = filename |
|
|
self.passed_in_file_object = True |
|
|
|
|
|
self.fh = fh |
|
|
self.currentstream = None |
|
|
fh.write(b"%PDF-1.4\n") |
|
|
|
|
|
|
|
|
|
|
|
fh.write(b"%\254\334 \253\272\n") |
|
|
|
|
|
self.rootObject = self.reserveObject('root') |
|
|
self.pagesObject = self.reserveObject('pages') |
|
|
self.pageList = [] |
|
|
self.fontObject = self.reserveObject('fonts') |
|
|
self._extGStateObject = self.reserveObject('extended graphics states') |
|
|
self.hatchObject = self.reserveObject('tiling patterns') |
|
|
self.gouraudObject = self.reserveObject('Gouraud triangles') |
|
|
self.XObjectObject = self.reserveObject('external objects') |
|
|
self.resourceObject = self.reserveObject('resources') |
|
|
|
|
|
root = {'Type': Name('Catalog'), |
|
|
'Pages': self.pagesObject} |
|
|
self.writeObject(self.rootObject, root) |
|
|
|
|
|
self.infoDict = _create_pdf_info_dict('pdf', metadata or {}) |
|
|
|
|
|
self.fontNames = {} |
|
|
self._internal_font_seq = (Name(f'F{i}') for i in itertools.count(1)) |
|
|
self.dviFontInfo = {} |
|
|
|
|
|
self.type1Descriptors = {} |
|
|
self._character_tracker = _backend_pdf_ps.CharacterTracker() |
|
|
|
|
|
self.alphaStates = {} |
|
|
self._alpha_state_seq = (Name(f'A{i}') for i in itertools.count(1)) |
|
|
self._soft_mask_states = {} |
|
|
self._soft_mask_seq = (Name(f'SM{i}') for i in itertools.count(1)) |
|
|
self._soft_mask_groups = [] |
|
|
self.hatchPatterns = {} |
|
|
self._hatch_pattern_seq = (Name(f'H{i}') for i in itertools.count(1)) |
|
|
self.gouraudTriangles = [] |
|
|
|
|
|
self._images = {} |
|
|
self._image_seq = (Name(f'I{i}') for i in itertools.count(1)) |
|
|
|
|
|
self.markers = {} |
|
|
self.multi_byte_charprocs = {} |
|
|
|
|
|
self.paths = [] |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
self._annotations = [] |
|
|
|
|
|
|
|
|
self.pageAnnotations = [] |
|
|
|
|
|
|
|
|
procsets = [Name(x) for x in "PDF Text ImageB ImageC ImageI".split()] |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
resources = {'Font': self.fontObject, |
|
|
'XObject': self.XObjectObject, |
|
|
'ExtGState': self._extGStateObject, |
|
|
'Pattern': self.hatchObject, |
|
|
'Shading': self.gouraudObject, |
|
|
'ProcSet': procsets} |
|
|
self.writeObject(self.resourceObject, resources) |
|
|
|
|
|
def newPage(self, width, height): |
|
|
self.endStream() |
|
|
|
|
|
self.width, self.height = width, height |
|
|
contentObject = self.reserveObject('page contents') |
|
|
annotsObject = self.reserveObject('annotations') |
|
|
thePage = {'Type': Name('Page'), |
|
|
'Parent': self.pagesObject, |
|
|
'Resources': self.resourceObject, |
|
|
'MediaBox': [0, 0, 72 * width, 72 * height], |
|
|
'Contents': contentObject, |
|
|
'Annots': annotsObject, |
|
|
} |
|
|
pageObject = self.reserveObject('page') |
|
|
self.writeObject(pageObject, thePage) |
|
|
self.pageList.append(pageObject) |
|
|
self._annotations.append((annotsObject, self.pageAnnotations)) |
|
|
|
|
|
self.beginStream(contentObject.id, |
|
|
self.reserveObject('length of content stream')) |
|
|
|
|
|
|
|
|
self.output(Name('DeviceRGB'), Op.setcolorspace_stroke) |
|
|
self.output(Name('DeviceRGB'), Op.setcolorspace_nonstroke) |
|
|
self.output(GraphicsContextPdf.joinstyles['round'], Op.setlinejoin) |
|
|
|
|
|
|
|
|
self.pageAnnotations = [] |
|
|
|
|
|
def newTextnote(self, text, positionRect=[-100, -100, 0, 0]): |
|
|
|
|
|
theNote = {'Type': Name('Annot'), |
|
|
'Subtype': Name('Text'), |
|
|
'Contents': text, |
|
|
'Rect': positionRect, |
|
|
} |
|
|
self.pageAnnotations.append(theNote) |
|
|
|
|
|
def _get_subsetted_psname(self, ps_name, charmap): |
|
|
def toStr(n, base): |
|
|
if n < base: |
|
|
return string.ascii_uppercase[n] |
|
|
else: |
|
|
return ( |
|
|
toStr(n // base, base) + string.ascii_uppercase[n % base] |
|
|
) |
|
|
|
|
|
|
|
|
hashed = hash(frozenset(charmap.keys())) % ((sys.maxsize + 1) * 2) |
|
|
prefix = toStr(hashed, 26) |
|
|
|
|
|
|
|
|
return prefix[:6] + "+" + ps_name |
|
|
|
|
|
def finalize(self): |
|
|
"""Write out the various deferred objects and the pdf end matter.""" |
|
|
|
|
|
self.endStream() |
|
|
self._write_annotations() |
|
|
self.writeFonts() |
|
|
self.writeExtGSTates() |
|
|
self._write_soft_mask_groups() |
|
|
self.writeHatches() |
|
|
self.writeGouraudTriangles() |
|
|
xobjects = { |
|
|
name: ob for image, name, ob in self._images.values()} |
|
|
for tup in self.markers.values(): |
|
|
xobjects[tup[0]] = tup[1] |
|
|
for name, value in self.multi_byte_charprocs.items(): |
|
|
xobjects[name] = value |
|
|
for name, path, trans, ob, join, cap, padding, filled, stroked \ |
|
|
in self.paths: |
|
|
xobjects[name] = ob |
|
|
self.writeObject(self.XObjectObject, xobjects) |
|
|
self.writeImages() |
|
|
self.writeMarkers() |
|
|
self.writePathCollectionTemplates() |
|
|
self.writeObject(self.pagesObject, |
|
|
{'Type': Name('Pages'), |
|
|
'Kids': self.pageList, |
|
|
'Count': len(self.pageList)}) |
|
|
self.writeInfoDict() |
|
|
|
|
|
|
|
|
self.writeXref() |
|
|
self.writeTrailer() |
|
|
|
|
|
def close(self): |
|
|
"""Flush all buffers and free all resources.""" |
|
|
|
|
|
self.endStream() |
|
|
if self.passed_in_file_object: |
|
|
self.fh.flush() |
|
|
else: |
|
|
if self.original_file_like is not None: |
|
|
self.original_file_like.write(self.fh.getvalue()) |
|
|
self.fh.close() |
|
|
|
|
|
def write(self, data): |
|
|
if self.currentstream is None: |
|
|
self.fh.write(data) |
|
|
else: |
|
|
self.currentstream.write(data) |
|
|
|
|
|
def output(self, *data): |
|
|
self.write(_fill([pdfRepr(x) for x in data])) |
|
|
self.write(b'\n') |
|
|
|
|
|
def beginStream(self, id, len, extra=None, png=None): |
|
|
assert self.currentstream is None |
|
|
self.currentstream = Stream(id, len, self, extra, png) |
|
|
|
|
|
def endStream(self): |
|
|
if self.currentstream is not None: |
|
|
self.currentstream.end() |
|
|
self.currentstream = None |
|
|
|
|
|
def outputStream(self, ref, data, *, extra=None): |
|
|
self.beginStream(ref.id, None, extra) |
|
|
self.currentstream.write(data) |
|
|
self.endStream() |
|
|
|
|
|
def _write_annotations(self): |
|
|
for annotsObject, annotations in self._annotations: |
|
|
self.writeObject(annotsObject, annotations) |
|
|
|
|
|
def fontName(self, fontprop): |
|
|
""" |
|
|
Select a font based on fontprop and return a name suitable for |
|
|
Op.selectfont. If fontprop is a string, it will be interpreted |
|
|
as the filename of the font. |
|
|
""" |
|
|
|
|
|
if isinstance(fontprop, str): |
|
|
filenames = [fontprop] |
|
|
elif mpl.rcParams['pdf.use14corefonts']: |
|
|
filenames = _fontManager._find_fonts_by_props( |
|
|
fontprop, fontext='afm', directory=RendererPdf._afm_font_dir |
|
|
) |
|
|
else: |
|
|
filenames = _fontManager._find_fonts_by_props(fontprop) |
|
|
first_Fx = None |
|
|
for fname in filenames: |
|
|
Fx = self.fontNames.get(fname) |
|
|
if not first_Fx: |
|
|
first_Fx = Fx |
|
|
if Fx is None: |
|
|
Fx = next(self._internal_font_seq) |
|
|
self.fontNames[fname] = Fx |
|
|
_log.debug('Assigning font %s = %r', Fx, fname) |
|
|
if not first_Fx: |
|
|
first_Fx = Fx |
|
|
|
|
|
|
|
|
|
|
|
return first_Fx |
|
|
|
|
|
def dviFontName(self, dvifont): |
|
|
""" |
|
|
Given a dvi font object, return a name suitable for Op.selectfont. |
|
|
This registers the font information in ``self.dviFontInfo`` if not yet |
|
|
registered. |
|
|
""" |
|
|
|
|
|
dvi_info = self.dviFontInfo.get(dvifont.texname) |
|
|
if dvi_info is not None: |
|
|
return dvi_info.pdfname |
|
|
|
|
|
tex_font_map = dviread.PsfontsMap(dviread._find_tex_file('pdftex.map')) |
|
|
psfont = tex_font_map[dvifont.texname] |
|
|
if psfont.filename is None: |
|
|
raise ValueError( |
|
|
"No usable font file found for {} (TeX: {}); " |
|
|
"the font may lack a Type-1 version" |
|
|
.format(psfont.psname, dvifont.texname)) |
|
|
|
|
|
pdfname = next(self._internal_font_seq) |
|
|
_log.debug('Assigning font %s = %s (dvi)', pdfname, dvifont.texname) |
|
|
self.dviFontInfo[dvifont.texname] = types.SimpleNamespace( |
|
|
dvifont=dvifont, |
|
|
pdfname=pdfname, |
|
|
fontfile=psfont.filename, |
|
|
basefont=psfont.psname, |
|
|
encodingfile=psfont.encoding, |
|
|
effects=psfont.effects) |
|
|
return pdfname |
|
|
|
|
|
def writeFonts(self): |
|
|
fonts = {} |
|
|
for dviname, info in sorted(self.dviFontInfo.items()): |
|
|
Fx = info.pdfname |
|
|
_log.debug('Embedding Type-1 font %s from dvi.', dviname) |
|
|
fonts[Fx] = self._embedTeXFont(info) |
|
|
for filename in sorted(self.fontNames): |
|
|
Fx = self.fontNames[filename] |
|
|
_log.debug('Embedding font %s.', filename) |
|
|
if filename.endswith('.afm'): |
|
|
|
|
|
_log.debug('Writing AFM font.') |
|
|
fonts[Fx] = self._write_afm_font(filename) |
|
|
else: |
|
|
|
|
|
_log.debug('Writing TrueType font.') |
|
|
chars = self._character_tracker.used.get(filename) |
|
|
if chars: |
|
|
fonts[Fx] = self.embedTTF(filename, chars) |
|
|
self.writeObject(self.fontObject, fonts) |
|
|
|
|
|
def _write_afm_font(self, filename): |
|
|
with open(filename, 'rb') as fh: |
|
|
font = AFM(fh) |
|
|
fontname = font.get_fontname() |
|
|
fontdict = {'Type': Name('Font'), |
|
|
'Subtype': Name('Type1'), |
|
|
'BaseFont': Name(fontname), |
|
|
'Encoding': Name('WinAnsiEncoding')} |
|
|
fontdictObject = self.reserveObject('font dictionary') |
|
|
self.writeObject(fontdictObject, fontdict) |
|
|
return fontdictObject |
|
|
|
|
|
def _embedTeXFont(self, fontinfo): |
|
|
_log.debug('Embedding TeX font %s - fontinfo=%s', |
|
|
fontinfo.dvifont.texname, fontinfo.__dict__) |
|
|
|
|
|
|
|
|
widthsObject = self.reserveObject('font widths') |
|
|
self.writeObject(widthsObject, fontinfo.dvifont.widths) |
|
|
|
|
|
|
|
|
fontdictObject = self.reserveObject('font dictionary') |
|
|
fontdict = { |
|
|
'Type': Name('Font'), |
|
|
'Subtype': Name('Type1'), |
|
|
'FirstChar': 0, |
|
|
'LastChar': len(fontinfo.dvifont.widths) - 1, |
|
|
'Widths': widthsObject, |
|
|
} |
|
|
|
|
|
|
|
|
if fontinfo.encodingfile is not None: |
|
|
fontdict['Encoding'] = { |
|
|
'Type': Name('Encoding'), |
|
|
'Differences': [ |
|
|
0, *map(Name, dviread._parse_enc(fontinfo.encodingfile))], |
|
|
} |
|
|
|
|
|
|
|
|
if fontinfo.fontfile is None: |
|
|
_log.warning( |
|
|
"Because of TeX configuration (pdftex.map, see updmap option " |
|
|
"pdftexDownloadBase14) the font %s is not embedded. This is " |
|
|
"deprecated as of PDF 1.5 and it may cause the consumer " |
|
|
"application to show something that was not intended.", |
|
|
fontinfo.basefont) |
|
|
fontdict['BaseFont'] = Name(fontinfo.basefont) |
|
|
self.writeObject(fontdictObject, fontdict) |
|
|
return fontdictObject |
|
|
|
|
|
|
|
|
t1font = _type1font.Type1Font(fontinfo.fontfile) |
|
|
if fontinfo.effects: |
|
|
t1font = t1font.transform(fontinfo.effects) |
|
|
fontdict['BaseFont'] = Name(t1font.prop['FontName']) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
effects = (fontinfo.effects.get('slant', 0.0), |
|
|
fontinfo.effects.get('extend', 1.0)) |
|
|
fontdesc = self.type1Descriptors.get((fontinfo.fontfile, effects)) |
|
|
if fontdesc is None: |
|
|
fontdesc = self.createType1Descriptor(t1font, fontinfo.fontfile) |
|
|
self.type1Descriptors[(fontinfo.fontfile, effects)] = fontdesc |
|
|
fontdict['FontDescriptor'] = fontdesc |
|
|
|
|
|
self.writeObject(fontdictObject, fontdict) |
|
|
return fontdictObject |
|
|
|
|
|
def createType1Descriptor(self, t1font, fontfile): |
|
|
|
|
|
|
|
|
fontdescObject = self.reserveObject('font descriptor') |
|
|
fontfileObject = self.reserveObject('font file') |
|
|
|
|
|
italic_angle = t1font.prop['ItalicAngle'] |
|
|
fixed_pitch = t1font.prop['isFixedPitch'] |
|
|
|
|
|
flags = 0 |
|
|
|
|
|
if fixed_pitch: |
|
|
flags |= 1 << 0 |
|
|
|
|
|
if 0: |
|
|
flags |= 1 << 1 |
|
|
|
|
|
if 1: |
|
|
flags |= 1 << 2 |
|
|
|
|
|
else: |
|
|
flags |= 1 << 5 |
|
|
|
|
|
if italic_angle: |
|
|
flags |= 1 << 6 |
|
|
|
|
|
if 0: |
|
|
flags |= 1 << 16 |
|
|
|
|
|
if 0: |
|
|
flags |= 1 << 17 |
|
|
|
|
|
if 0: |
|
|
flags |= 1 << 18 |
|
|
|
|
|
ft2font = get_font(fontfile) |
|
|
|
|
|
descriptor = { |
|
|
'Type': Name('FontDescriptor'), |
|
|
'FontName': Name(t1font.prop['FontName']), |
|
|
'Flags': flags, |
|
|
'FontBBox': ft2font.bbox, |
|
|
'ItalicAngle': italic_angle, |
|
|
'Ascent': ft2font.ascender, |
|
|
'Descent': ft2font.descender, |
|
|
'CapHeight': 1000, |
|
|
'XHeight': 500, |
|
|
'FontFile': fontfileObject, |
|
|
'FontFamily': t1font.prop['FamilyName'], |
|
|
'StemV': 50, |
|
|
|
|
|
|
|
|
} |
|
|
|
|
|
self.writeObject(fontdescObject, descriptor) |
|
|
|
|
|
self.outputStream(fontfileObject, b"".join(t1font.parts[:2]), |
|
|
extra={'Length1': len(t1font.parts[0]), |
|
|
'Length2': len(t1font.parts[1]), |
|
|
'Length3': 0}) |
|
|
|
|
|
return fontdescObject |
|
|
|
|
|
def _get_xobject_glyph_name(self, filename, glyph_name): |
|
|
Fx = self.fontName(filename) |
|
|
return "-".join([ |
|
|
Fx.name.decode(), |
|
|
os.path.splitext(os.path.basename(filename))[0], |
|
|
glyph_name]) |
|
|
|
|
|
_identityToUnicodeCMap = b"""/CIDInit /ProcSet findresource begin |
|
|
12 dict begin |
|
|
begincmap |
|
|
/CIDSystemInfo |
|
|
<< /Registry (Adobe) |
|
|
/Ordering (UCS) |
|
|
/Supplement 0 |
|
|
>> def |
|
|
/CMapName /Adobe-Identity-UCS def |
|
|
/CMapType 2 def |
|
|
1 begincodespacerange |
|
|
<0000> <ffff> |
|
|
endcodespacerange |
|
|
%d beginbfrange |
|
|
%s |
|
|
endbfrange |
|
|
endcmap |
|
|
CMapName currentdict /CMap defineresource pop |
|
|
end |
|
|
end""" |
|
|
|
|
|
def embedTTF(self, filename, characters): |
|
|
"""Embed the TTF font from the named file into the document.""" |
|
|
|
|
|
font = get_font(filename) |
|
|
fonttype = mpl.rcParams['pdf.fonttype'] |
|
|
|
|
|
def cvt(length, upe=font.units_per_EM, nearest=True): |
|
|
"""Convert font coordinates to PDF glyph coordinates.""" |
|
|
value = length / upe * 1000 |
|
|
if nearest: |
|
|
return round(value) |
|
|
|
|
|
if value < 0: |
|
|
return math.floor(value) |
|
|
else: |
|
|
return math.ceil(value) |
|
|
|
|
|
def embedTTFType3(font, characters, descriptor): |
|
|
"""The Type 3-specific part of embedding a Truetype font""" |
|
|
widthsObject = self.reserveObject('font widths') |
|
|
fontdescObject = self.reserveObject('font descriptor') |
|
|
fontdictObject = self.reserveObject('font dictionary') |
|
|
charprocsObject = self.reserveObject('character procs') |
|
|
differencesArray = [] |
|
|
firstchar, lastchar = 0, 255 |
|
|
bbox = [cvt(x, nearest=False) for x in font.bbox] |
|
|
|
|
|
fontdict = { |
|
|
'Type': Name('Font'), |
|
|
'BaseFont': ps_name, |
|
|
'FirstChar': firstchar, |
|
|
'LastChar': lastchar, |
|
|
'FontDescriptor': fontdescObject, |
|
|
'Subtype': Name('Type3'), |
|
|
'Name': descriptor['FontName'], |
|
|
'FontBBox': bbox, |
|
|
'FontMatrix': [.001, 0, 0, .001, 0, 0], |
|
|
'CharProcs': charprocsObject, |
|
|
'Encoding': { |
|
|
'Type': Name('Encoding'), |
|
|
'Differences': differencesArray}, |
|
|
'Widths': widthsObject |
|
|
} |
|
|
|
|
|
from encodings import cp1252 |
|
|
|
|
|
|
|
|
def get_char_width(charcode): |
|
|
s = ord(cp1252.decoding_table[charcode]) |
|
|
width = font.load_char( |
|
|
s, flags=LOAD_NO_SCALE | LOAD_NO_HINTING).horiAdvance |
|
|
return cvt(width) |
|
|
with warnings.catch_warnings(): |
|
|
|
|
|
|
|
|
|
|
|
warnings.filterwarnings("ignore") |
|
|
widths = [get_char_width(charcode) |
|
|
for charcode in range(firstchar, lastchar+1)] |
|
|
descriptor['MaxWidth'] = max(widths) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
glyph_ids = [] |
|
|
differences = [] |
|
|
multi_byte_chars = set() |
|
|
for c in characters: |
|
|
ccode = c |
|
|
gind = font.get_char_index(ccode) |
|
|
glyph_ids.append(gind) |
|
|
glyph_name = font.get_glyph_name(gind) |
|
|
if ccode <= 255: |
|
|
differences.append((ccode, glyph_name)) |
|
|
else: |
|
|
multi_byte_chars.add(glyph_name) |
|
|
differences.sort() |
|
|
|
|
|
last_c = -2 |
|
|
for c, name in differences: |
|
|
if c != last_c + 1: |
|
|
differencesArray.append(c) |
|
|
differencesArray.append(Name(name)) |
|
|
last_c = c |
|
|
|
|
|
|
|
|
rawcharprocs = _get_pdf_charprocs(filename, glyph_ids) |
|
|
charprocs = {} |
|
|
for charname in sorted(rawcharprocs): |
|
|
stream = rawcharprocs[charname] |
|
|
charprocDict = {} |
|
|
|
|
|
|
|
|
if charname in multi_byte_chars: |
|
|
charprocDict = {'Type': Name('XObject'), |
|
|
'Subtype': Name('Form'), |
|
|
'BBox': bbox} |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
stream = stream[stream.find(b"d1") + 2:] |
|
|
charprocObject = self.reserveObject('charProc') |
|
|
self.outputStream(charprocObject, stream, extra=charprocDict) |
|
|
|
|
|
|
|
|
|
|
|
if charname in multi_byte_chars: |
|
|
name = self._get_xobject_glyph_name(filename, charname) |
|
|
self.multi_byte_charprocs[name] = charprocObject |
|
|
else: |
|
|
charprocs[charname] = charprocObject |
|
|
|
|
|
|
|
|
self.writeObject(fontdictObject, fontdict) |
|
|
self.writeObject(fontdescObject, descriptor) |
|
|
self.writeObject(widthsObject, widths) |
|
|
self.writeObject(charprocsObject, charprocs) |
|
|
|
|
|
return fontdictObject |
|
|
|
|
|
def embedTTFType42(font, characters, descriptor): |
|
|
"""The Type 42-specific part of embedding a Truetype font""" |
|
|
fontdescObject = self.reserveObject('font descriptor') |
|
|
cidFontDictObject = self.reserveObject('CID font dictionary') |
|
|
type0FontDictObject = self.reserveObject('Type 0 font dictionary') |
|
|
cidToGidMapObject = self.reserveObject('CIDToGIDMap stream') |
|
|
fontfileObject = self.reserveObject('font file stream') |
|
|
wObject = self.reserveObject('Type 0 widths') |
|
|
toUnicodeMapObject = self.reserveObject('ToUnicode map') |
|
|
|
|
|
subset_str = "".join(chr(c) for c in characters) |
|
|
_log.debug("SUBSET %s characters: %s", filename, subset_str) |
|
|
fontdata = _backend_pdf_ps.get_glyphs_subset(filename, subset_str) |
|
|
_log.debug( |
|
|
"SUBSET %s %d -> %d", filename, |
|
|
os.stat(filename).st_size, fontdata.getbuffer().nbytes |
|
|
) |
|
|
|
|
|
|
|
|
full_font = font |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
font = FT2Font(fontdata) |
|
|
|
|
|
cidFontDict = { |
|
|
'Type': Name('Font'), |
|
|
'Subtype': Name('CIDFontType2'), |
|
|
'BaseFont': ps_name, |
|
|
'CIDSystemInfo': { |
|
|
'Registry': 'Adobe', |
|
|
'Ordering': 'Identity', |
|
|
'Supplement': 0}, |
|
|
'FontDescriptor': fontdescObject, |
|
|
'W': wObject, |
|
|
'CIDToGIDMap': cidToGidMapObject |
|
|
} |
|
|
|
|
|
type0FontDict = { |
|
|
'Type': Name('Font'), |
|
|
'Subtype': Name('Type0'), |
|
|
'BaseFont': ps_name, |
|
|
'Encoding': Name('Identity-H'), |
|
|
'DescendantFonts': [cidFontDictObject], |
|
|
'ToUnicode': toUnicodeMapObject |
|
|
} |
|
|
|
|
|
|
|
|
descriptor['FontFile2'] = fontfileObject |
|
|
self.outputStream( |
|
|
fontfileObject, fontdata.getvalue(), |
|
|
extra={'Length1': fontdata.getbuffer().nbytes}) |
|
|
|
|
|
|
|
|
|
|
|
cid_to_gid_map = ['\0'] * 65536 |
|
|
widths = [] |
|
|
max_ccode = 0 |
|
|
for c in characters: |
|
|
ccode = c |
|
|
gind = font.get_char_index(ccode) |
|
|
glyph = font.load_char(ccode, |
|
|
flags=LOAD_NO_SCALE | LOAD_NO_HINTING) |
|
|
widths.append((ccode, cvt(glyph.horiAdvance))) |
|
|
if ccode < 65536: |
|
|
cid_to_gid_map[ccode] = chr(gind) |
|
|
max_ccode = max(ccode, max_ccode) |
|
|
widths.sort() |
|
|
cid_to_gid_map = cid_to_gid_map[:max_ccode + 1] |
|
|
|
|
|
last_ccode = -2 |
|
|
w = [] |
|
|
max_width = 0 |
|
|
unicode_groups = [] |
|
|
for ccode, width in widths: |
|
|
if ccode != last_ccode + 1: |
|
|
w.append(ccode) |
|
|
w.append([width]) |
|
|
unicode_groups.append([ccode, ccode]) |
|
|
else: |
|
|
w[-1].append(width) |
|
|
unicode_groups[-1][1] = ccode |
|
|
max_width = max(max_width, width) |
|
|
last_ccode = ccode |
|
|
|
|
|
unicode_bfrange = [] |
|
|
for start, end in unicode_groups: |
|
|
|
|
|
if start > 65535: |
|
|
continue |
|
|
end = min(65535, end) |
|
|
|
|
|
unicode_bfrange.append( |
|
|
b"<%04x> <%04x> [%s]" % |
|
|
(start, end, |
|
|
b" ".join(b"<%04x>" % x for x in range(start, end+1)))) |
|
|
unicode_cmap = (self._identityToUnicodeCMap % |
|
|
(len(unicode_groups), b"\n".join(unicode_bfrange))) |
|
|
|
|
|
|
|
|
glyph_ids = [] |
|
|
for ccode in characters: |
|
|
if not _font_supports_glyph(fonttype, ccode): |
|
|
gind = full_font.get_char_index(ccode) |
|
|
glyph_ids.append(gind) |
|
|
|
|
|
bbox = [cvt(x, nearest=False) for x in full_font.bbox] |
|
|
rawcharprocs = _get_pdf_charprocs(filename, glyph_ids) |
|
|
for charname in sorted(rawcharprocs): |
|
|
stream = rawcharprocs[charname] |
|
|
charprocDict = {'Type': Name('XObject'), |
|
|
'Subtype': Name('Form'), |
|
|
'BBox': bbox} |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
stream = stream[stream.find(b"d1") + 2:] |
|
|
charprocObject = self.reserveObject('charProc') |
|
|
self.outputStream(charprocObject, stream, extra=charprocDict) |
|
|
|
|
|
name = self._get_xobject_glyph_name(filename, charname) |
|
|
self.multi_byte_charprocs[name] = charprocObject |
|
|
|
|
|
|
|
|
cid_to_gid_map = "".join(cid_to_gid_map).encode("utf-16be") |
|
|
self.outputStream(cidToGidMapObject, cid_to_gid_map) |
|
|
|
|
|
|
|
|
self.outputStream(toUnicodeMapObject, unicode_cmap) |
|
|
|
|
|
descriptor['MaxWidth'] = max_width |
|
|
|
|
|
|
|
|
self.writeObject(cidFontDictObject, cidFontDict) |
|
|
self.writeObject(type0FontDictObject, type0FontDict) |
|
|
self.writeObject(fontdescObject, descriptor) |
|
|
self.writeObject(wObject, w) |
|
|
|
|
|
return type0FontDictObject |
|
|
|
|
|
|
|
|
|
|
|
ps_name = self._get_subsetted_psname( |
|
|
font.postscript_name, |
|
|
font.get_charmap() |
|
|
) |
|
|
ps_name = ps_name.encode('ascii', 'replace') |
|
|
ps_name = Name(ps_name) |
|
|
pclt = font.get_sfnt_table('pclt') or {'capHeight': 0, 'xHeight': 0} |
|
|
post = font.get_sfnt_table('post') or {'italicAngle': (0, 0)} |
|
|
ff = font.face_flags |
|
|
sf = font.style_flags |
|
|
|
|
|
flags = 0 |
|
|
symbolic = False |
|
|
if ff & FIXED_WIDTH: |
|
|
flags |= 1 << 0 |
|
|
if 0: |
|
|
flags |= 1 << 1 |
|
|
if symbolic: |
|
|
flags |= 1 << 2 |
|
|
else: |
|
|
flags |= 1 << 5 |
|
|
if sf & ITALIC: |
|
|
flags |= 1 << 6 |
|
|
if 0: |
|
|
flags |= 1 << 16 |
|
|
if 0: |
|
|
flags |= 1 << 17 |
|
|
if 0: |
|
|
flags |= 1 << 18 |
|
|
|
|
|
descriptor = { |
|
|
'Type': Name('FontDescriptor'), |
|
|
'FontName': ps_name, |
|
|
'Flags': flags, |
|
|
'FontBBox': [cvt(x, nearest=False) for x in font.bbox], |
|
|
'Ascent': cvt(font.ascender, nearest=False), |
|
|
'Descent': cvt(font.descender, nearest=False), |
|
|
'CapHeight': cvt(pclt['capHeight'], nearest=False), |
|
|
'XHeight': cvt(pclt['xHeight']), |
|
|
'ItalicAngle': post['italicAngle'][1], |
|
|
'StemV': 0 |
|
|
} |
|
|
|
|
|
if fonttype == 3: |
|
|
return embedTTFType3(font, characters, descriptor) |
|
|
elif fonttype == 42: |
|
|
return embedTTFType42(font, characters, descriptor) |
|
|
|
|
|
def alphaState(self, alpha): |
|
|
"""Return name of an ExtGState that sets alpha to the given value.""" |
|
|
|
|
|
state = self.alphaStates.get(alpha, None) |
|
|
if state is not None: |
|
|
return state[0] |
|
|
|
|
|
name = next(self._alpha_state_seq) |
|
|
self.alphaStates[alpha] = \ |
|
|
(name, {'Type': Name('ExtGState'), |
|
|
'CA': alpha[0], 'ca': alpha[1]}) |
|
|
return name |
|
|
|
|
|
def _soft_mask_state(self, smask): |
|
|
""" |
|
|
Return an ExtGState that sets the soft mask to the given shading. |
|
|
|
|
|
Parameters |
|
|
---------- |
|
|
smask : Reference |
|
|
Reference to a shading in DeviceGray color space, whose luminosity |
|
|
is to be used as the alpha channel. |
|
|
|
|
|
Returns |
|
|
------- |
|
|
Name |
|
|
""" |
|
|
|
|
|
state = self._soft_mask_states.get(smask, None) |
|
|
if state is not None: |
|
|
return state[0] |
|
|
|
|
|
name = next(self._soft_mask_seq) |
|
|
groupOb = self.reserveObject('transparency group for soft mask') |
|
|
self._soft_mask_states[smask] = ( |
|
|
name, |
|
|
{ |
|
|
'Type': Name('ExtGState'), |
|
|
'AIS': False, |
|
|
'SMask': { |
|
|
'Type': Name('Mask'), |
|
|
'S': Name('Luminosity'), |
|
|
'BC': [1], |
|
|
'G': groupOb |
|
|
} |
|
|
} |
|
|
) |
|
|
self._soft_mask_groups.append(( |
|
|
groupOb, |
|
|
{ |
|
|
'Type': Name('XObject'), |
|
|
'Subtype': Name('Form'), |
|
|
'FormType': 1, |
|
|
'Group': { |
|
|
'S': Name('Transparency'), |
|
|
'CS': Name('DeviceGray') |
|
|
}, |
|
|
'Matrix': [1, 0, 0, 1, 0, 0], |
|
|
'Resources': {'Shading': {'S': smask}}, |
|
|
'BBox': [0, 0, 1, 1] |
|
|
}, |
|
|
[Name('S'), Op.shading] |
|
|
)) |
|
|
return name |
|
|
|
|
|
def writeExtGSTates(self): |
|
|
self.writeObject( |
|
|
self._extGStateObject, |
|
|
dict([ |
|
|
*self.alphaStates.values(), |
|
|
*self._soft_mask_states.values() |
|
|
]) |
|
|
) |
|
|
|
|
|
def _write_soft_mask_groups(self): |
|
|
for ob, attributes, content in self._soft_mask_groups: |
|
|
self.beginStream(ob.id, None, attributes) |
|
|
self.output(*content) |
|
|
self.endStream() |
|
|
|
|
|
def hatchPattern(self, hatch_style): |
|
|
|
|
|
if hatch_style is not None: |
|
|
edge, face, hatch = hatch_style |
|
|
if edge is not None: |
|
|
edge = tuple(edge) |
|
|
if face is not None: |
|
|
face = tuple(face) |
|
|
hatch_style = (edge, face, hatch) |
|
|
|
|
|
pattern = self.hatchPatterns.get(hatch_style, None) |
|
|
if pattern is not None: |
|
|
return pattern |
|
|
|
|
|
name = next(self._hatch_pattern_seq) |
|
|
self.hatchPatterns[hatch_style] = name |
|
|
return name |
|
|
|
|
|
def writeHatches(self): |
|
|
hatchDict = dict() |
|
|
sidelen = 72.0 |
|
|
for hatch_style, name in self.hatchPatterns.items(): |
|
|
ob = self.reserveObject('hatch pattern') |
|
|
hatchDict[name] = ob |
|
|
res = {'Procsets': |
|
|
[Name(x) for x in "PDF Text ImageB ImageC ImageI".split()]} |
|
|
self.beginStream( |
|
|
ob.id, None, |
|
|
{'Type': Name('Pattern'), |
|
|
'PatternType': 1, 'PaintType': 1, 'TilingType': 1, |
|
|
'BBox': [0, 0, sidelen, sidelen], |
|
|
'XStep': sidelen, 'YStep': sidelen, |
|
|
'Resources': res, |
|
|
|
|
|
'Matrix': [1, 0, 0, 1, 0, self.height * 72]}) |
|
|
|
|
|
stroke_rgb, fill_rgb, hatch = hatch_style |
|
|
self.output(stroke_rgb[0], stroke_rgb[1], stroke_rgb[2], |
|
|
Op.setrgb_stroke) |
|
|
if fill_rgb is not None: |
|
|
self.output(fill_rgb[0], fill_rgb[1], fill_rgb[2], |
|
|
Op.setrgb_nonstroke, |
|
|
0, 0, sidelen, sidelen, Op.rectangle, |
|
|
Op.fill) |
|
|
|
|
|
self.output(mpl.rcParams['hatch.linewidth'], Op.setlinewidth) |
|
|
|
|
|
self.output(*self.pathOperations( |
|
|
Path.hatch(hatch), |
|
|
Affine2D().scale(sidelen), |
|
|
simplify=False)) |
|
|
self.output(Op.fill_stroke) |
|
|
|
|
|
self.endStream() |
|
|
self.writeObject(self.hatchObject, hatchDict) |
|
|
|
|
|
def addGouraudTriangles(self, points, colors): |
|
|
""" |
|
|
Add a Gouraud triangle shading. |
|
|
|
|
|
Parameters |
|
|
---------- |
|
|
points : np.ndarray |
|
|
Triangle vertices, shape (n, 3, 2) |
|
|
where n = number of triangles, 3 = vertices, 2 = x, y. |
|
|
colors : np.ndarray |
|
|
Vertex colors, shape (n, 3, 1) or (n, 3, 4) |
|
|
as with points, but last dimension is either (gray,) |
|
|
or (r, g, b, alpha). |
|
|
|
|
|
Returns |
|
|
------- |
|
|
Name, Reference |
|
|
""" |
|
|
name = Name('GT%d' % len(self.gouraudTriangles)) |
|
|
ob = self.reserveObject(f'Gouraud triangle {name}') |
|
|
self.gouraudTriangles.append((name, ob, points, colors)) |
|
|
return name, ob |
|
|
|
|
|
def writeGouraudTriangles(self): |
|
|
gouraudDict = dict() |
|
|
for name, ob, points, colors in self.gouraudTriangles: |
|
|
gouraudDict[name] = ob |
|
|
shape = points.shape |
|
|
flat_points = points.reshape((shape[0] * shape[1], 2)) |
|
|
colordim = colors.shape[2] |
|
|
assert colordim in (1, 4) |
|
|
flat_colors = colors.reshape((shape[0] * shape[1], colordim)) |
|
|
if colordim == 4: |
|
|
|
|
|
colordim = 3 |
|
|
points_min = np.min(flat_points, axis=0) - (1 << 8) |
|
|
points_max = np.max(flat_points, axis=0) + (1 << 8) |
|
|
factor = 0xffffffff / (points_max - points_min) |
|
|
|
|
|
self.beginStream( |
|
|
ob.id, None, |
|
|
{'ShadingType': 4, |
|
|
'BitsPerCoordinate': 32, |
|
|
'BitsPerComponent': 8, |
|
|
'BitsPerFlag': 8, |
|
|
'ColorSpace': Name( |
|
|
'DeviceRGB' if colordim == 3 else 'DeviceGray' |
|
|
), |
|
|
'AntiAlias': False, |
|
|
'Decode': ([points_min[0], points_max[0], |
|
|
points_min[1], points_max[1]] |
|
|
+ [0, 1] * colordim), |
|
|
}) |
|
|
|
|
|
streamarr = np.empty( |
|
|
(shape[0] * shape[1],), |
|
|
dtype=[('flags', 'u1'), |
|
|
('points', '>u4', (2,)), |
|
|
('colors', 'u1', (colordim,))]) |
|
|
streamarr['flags'] = 0 |
|
|
streamarr['points'] = (flat_points - points_min) * factor |
|
|
streamarr['colors'] = flat_colors[:, :colordim] * 255.0 |
|
|
|
|
|
self.write(streamarr.tobytes()) |
|
|
self.endStream() |
|
|
self.writeObject(self.gouraudObject, gouraudDict) |
|
|
|
|
|
def imageObject(self, image): |
|
|
"""Return name of an image XObject representing the given image.""" |
|
|
|
|
|
entry = self._images.get(id(image), None) |
|
|
if entry is not None: |
|
|
return entry[1] |
|
|
|
|
|
name = next(self._image_seq) |
|
|
ob = self.reserveObject(f'image {name}') |
|
|
self._images[id(image)] = (image, name, ob) |
|
|
return name |
|
|
|
|
|
def _unpack(self, im): |
|
|
""" |
|
|
Unpack image array *im* into ``(data, alpha)``, which have shape |
|
|
``(height, width, 3)`` (RGB) or ``(height, width, 1)`` (grayscale or |
|
|
alpha), except that alpha is None if the image is fully opaque. |
|
|
""" |
|
|
im = im[::-1] |
|
|
if im.ndim == 2: |
|
|
return im, None |
|
|
else: |
|
|
rgb = im[:, :, :3] |
|
|
rgb = np.array(rgb, order='C') |
|
|
|
|
|
if im.shape[2] == 4: |
|
|
alpha = im[:, :, 3][..., None] |
|
|
if np.all(alpha == 255): |
|
|
alpha = None |
|
|
else: |
|
|
alpha = np.array(alpha, order='C') |
|
|
else: |
|
|
alpha = None |
|
|
return rgb, alpha |
|
|
|
|
|
def _writePng(self, img): |
|
|
""" |
|
|
Write the image *img* into the pdf file using png |
|
|
predictors with Flate compression. |
|
|
""" |
|
|
buffer = BytesIO() |
|
|
img.save(buffer, format="png") |
|
|
buffer.seek(8) |
|
|
png_data = b'' |
|
|
bit_depth = palette = None |
|
|
while True: |
|
|
length, type = struct.unpack(b'!L4s', buffer.read(8)) |
|
|
if type in [b'IHDR', b'PLTE', b'IDAT']: |
|
|
data = buffer.read(length) |
|
|
if len(data) != length: |
|
|
raise RuntimeError("truncated data") |
|
|
if type == b'IHDR': |
|
|
bit_depth = int(data[8]) |
|
|
elif type == b'PLTE': |
|
|
palette = data |
|
|
elif type == b'IDAT': |
|
|
png_data += data |
|
|
elif type == b'IEND': |
|
|
break |
|
|
else: |
|
|
buffer.seek(length, 1) |
|
|
buffer.seek(4, 1) |
|
|
return png_data, bit_depth, palette |
|
|
|
|
|
def _writeImg(self, data, id, smask=None): |
|
|
""" |
|
|
Write the image *data*, of shape ``(height, width, 1)`` (grayscale) or |
|
|
``(height, width, 3)`` (RGB), as pdf object *id* and with the soft mask |
|
|
(alpha channel) *smask*, which should be either None or a ``(height, |
|
|
width, 1)`` array. |
|
|
""" |
|
|
height, width, color_channels = data.shape |
|
|
obj = {'Type': Name('XObject'), |
|
|
'Subtype': Name('Image'), |
|
|
'Width': width, |
|
|
'Height': height, |
|
|
'ColorSpace': Name({1: 'DeviceGray', 3: 'DeviceRGB'}[color_channels]), |
|
|
'BitsPerComponent': 8} |
|
|
if smask: |
|
|
obj['SMask'] = smask |
|
|
if mpl.rcParams['pdf.compression']: |
|
|
if data.shape[-1] == 1: |
|
|
data = data.squeeze(axis=-1) |
|
|
png = {'Predictor': 10, 'Colors': color_channels, 'Columns': width} |
|
|
img = Image.fromarray(data) |
|
|
img_colors = img.getcolors(maxcolors=256) |
|
|
if color_channels == 3 and img_colors is not None: |
|
|
|
|
|
|
|
|
num_colors = len(img_colors) |
|
|
palette = np.array([comp for _, color in img_colors for comp in color], |
|
|
dtype=np.uint8) |
|
|
palette24 = ((palette[0::3].astype(np.uint32) << 16) | |
|
|
(palette[1::3].astype(np.uint32) << 8) | |
|
|
palette[2::3]) |
|
|
rgb24 = ((data[:, :, 0].astype(np.uint32) << 16) | |
|
|
(data[:, :, 1].astype(np.uint32) << 8) | |
|
|
data[:, :, 2]) |
|
|
indices = np.argsort(palette24).astype(np.uint8) |
|
|
rgb8 = indices[np.searchsorted(palette24, rgb24, sorter=indices)] |
|
|
img = Image.fromarray(rgb8, mode='P') |
|
|
img.putpalette(palette) |
|
|
png_data, bit_depth, palette = self._writePng(img) |
|
|
if bit_depth is None or palette is None: |
|
|
raise RuntimeError("invalid PNG header") |
|
|
palette = palette[:num_colors * 3] |
|
|
obj['ColorSpace'] = [Name('Indexed'), Name('DeviceRGB'), |
|
|
num_colors - 1, palette] |
|
|
obj['BitsPerComponent'] = bit_depth |
|
|
png['Colors'] = 1 |
|
|
png['BitsPerComponent'] = bit_depth |
|
|
else: |
|
|
png_data, _, _ = self._writePng(img) |
|
|
else: |
|
|
png = None |
|
|
self.beginStream( |
|
|
id, |
|
|
self.reserveObject('length of image stream'), |
|
|
obj, |
|
|
png=png |
|
|
) |
|
|
if png: |
|
|
self.currentstream.write(png_data) |
|
|
else: |
|
|
self.currentstream.write(data.tobytes()) |
|
|
self.endStream() |
|
|
|
|
|
def writeImages(self): |
|
|
for img, name, ob in self._images.values(): |
|
|
data, adata = self._unpack(img) |
|
|
if adata is not None: |
|
|
smaskObject = self.reserveObject("smask") |
|
|
self._writeImg(adata, smaskObject.id) |
|
|
else: |
|
|
smaskObject = None |
|
|
self._writeImg(data, ob.id, smaskObject) |
|
|
|
|
|
def markerObject(self, path, trans, fill, stroke, lw, joinstyle, |
|
|
capstyle): |
|
|
"""Return name of a marker XObject representing the given path.""" |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
pathops = self.pathOperations(path, trans, simplify=False) |
|
|
key = (tuple(pathops), bool(fill), bool(stroke), joinstyle, capstyle) |
|
|
result = self.markers.get(key) |
|
|
if result is None: |
|
|
name = Name('M%d' % len(self.markers)) |
|
|
ob = self.reserveObject('marker %d' % len(self.markers)) |
|
|
bbox = path.get_extents(trans) |
|
|
self.markers[key] = [name, ob, bbox, lw] |
|
|
else: |
|
|
if result[-1] < lw: |
|
|
result[-1] = lw |
|
|
name = result[0] |
|
|
return name |
|
|
|
|
|
def writeMarkers(self): |
|
|
for ((pathops, fill, stroke, joinstyle, capstyle), |
|
|
(name, ob, bbox, lw)) in self.markers.items(): |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
bbox = bbox.padded(lw * 5) |
|
|
self.beginStream( |
|
|
ob.id, None, |
|
|
{'Type': Name('XObject'), 'Subtype': Name('Form'), |
|
|
'BBox': list(bbox.extents)}) |
|
|
self.output(GraphicsContextPdf.joinstyles[joinstyle], |
|
|
Op.setlinejoin) |
|
|
self.output(GraphicsContextPdf.capstyles[capstyle], Op.setlinecap) |
|
|
self.output(*pathops) |
|
|
self.output(Op.paint_path(fill, stroke)) |
|
|
self.endStream() |
|
|
|
|
|
def pathCollectionObject(self, gc, path, trans, padding, filled, stroked): |
|
|
name = Name('P%d' % len(self.paths)) |
|
|
ob = self.reserveObject('path %d' % len(self.paths)) |
|
|
self.paths.append( |
|
|
(name, path, trans, ob, gc.get_joinstyle(), gc.get_capstyle(), |
|
|
padding, filled, stroked)) |
|
|
return name |
|
|
|
|
|
def writePathCollectionTemplates(self): |
|
|
for (name, path, trans, ob, joinstyle, capstyle, padding, filled, |
|
|
stroked) in self.paths: |
|
|
pathops = self.pathOperations(path, trans, simplify=False) |
|
|
bbox = path.get_extents(trans) |
|
|
if not np.all(np.isfinite(bbox.extents)): |
|
|
extents = [0, 0, 0, 0] |
|
|
else: |
|
|
bbox = bbox.padded(padding) |
|
|
extents = list(bbox.extents) |
|
|
self.beginStream( |
|
|
ob.id, None, |
|
|
{'Type': Name('XObject'), 'Subtype': Name('Form'), |
|
|
'BBox': extents}) |
|
|
self.output(GraphicsContextPdf.joinstyles[joinstyle], |
|
|
Op.setlinejoin) |
|
|
self.output(GraphicsContextPdf.capstyles[capstyle], Op.setlinecap) |
|
|
self.output(*pathops) |
|
|
self.output(Op.paint_path(filled, stroked)) |
|
|
self.endStream() |
|
|
|
|
|
@staticmethod |
|
|
def pathOperations(path, transform, clip=None, simplify=None, sketch=None): |
|
|
return [Verbatim(_path.convert_to_string( |
|
|
path, transform, clip, simplify, sketch, |
|
|
6, |
|
|
[Op.moveto.value, Op.lineto.value, b'', Op.curveto.value, |
|
|
Op.closepath.value], |
|
|
True))] |
|
|
|
|
|
def writePath(self, path, transform, clip=False, sketch=None): |
|
|
if clip: |
|
|
clip = (0.0, 0.0, self.width * 72, self.height * 72) |
|
|
simplify = path.should_simplify |
|
|
else: |
|
|
clip = None |
|
|
simplify = False |
|
|
cmds = self.pathOperations(path, transform, clip, simplify=simplify, |
|
|
sketch=sketch) |
|
|
self.output(*cmds) |
|
|
|
|
|
def reserveObject(self, name=''): |
|
|
""" |
|
|
Reserve an ID for an indirect object. |
|
|
|
|
|
The name is used for debugging in case we forget to print out |
|
|
the object with writeObject. |
|
|
""" |
|
|
id = next(self._object_seq) |
|
|
self.xrefTable.append([None, 0, name]) |
|
|
return Reference(id) |
|
|
|
|
|
def recordXref(self, id): |
|
|
self.xrefTable[id][0] = self.fh.tell() - self.tell_base |
|
|
|
|
|
def writeObject(self, object, contents): |
|
|
self.recordXref(object.id) |
|
|
object.write(contents, self) |
|
|
|
|
|
def writeXref(self): |
|
|
"""Write out the xref table.""" |
|
|
self.startxref = self.fh.tell() - self.tell_base |
|
|
self.write(b"xref\n0 %d\n" % len(self.xrefTable)) |
|
|
for i, (offset, generation, name) in enumerate(self.xrefTable): |
|
|
if offset is None: |
|
|
raise AssertionError( |
|
|
'No offset for object %d (%s)' % (i, name)) |
|
|
else: |
|
|
key = b"f" if name == 'the zero object' else b"n" |
|
|
text = b"%010d %05d %b \n" % (offset, generation, key) |
|
|
self.write(text) |
|
|
|
|
|
def writeInfoDict(self): |
|
|
"""Write out the info dictionary, checking it for good form""" |
|
|
|
|
|
self.infoObject = self.reserveObject('info') |
|
|
self.writeObject(self.infoObject, self.infoDict) |
|
|
|
|
|
def writeTrailer(self): |
|
|
"""Write out the PDF trailer.""" |
|
|
|
|
|
self.write(b"trailer\n") |
|
|
self.write(pdfRepr( |
|
|
{'Size': len(self.xrefTable), |
|
|
'Root': self.rootObject, |
|
|
'Info': self.infoObject})) |
|
|
|
|
|
self.write(b"\nstartxref\n%d\n%%%%EOF\n" % self.startxref) |
|
|
|
|
|
|
|
|
class RendererPdf(_backend_pdf_ps.RendererPDFPSBase): |
|
|
|
|
|
_afm_font_dir = cbook._get_data_path("fonts/pdfcorefonts") |
|
|
_use_afm_rc_name = "pdf.use14corefonts" |
|
|
|
|
|
def __init__(self, file, image_dpi, height, width): |
|
|
super().__init__(width, height) |
|
|
self.file = file |
|
|
self.gc = self.new_gc() |
|
|
self.image_dpi = image_dpi |
|
|
|
|
|
def finalize(self): |
|
|
self.file.output(*self.gc.finalize()) |
|
|
|
|
|
def check_gc(self, gc, fillcolor=None): |
|
|
orig_fill = getattr(gc, '_fillcolor', (0., 0., 0.)) |
|
|
gc._fillcolor = fillcolor |
|
|
|
|
|
orig_alphas = getattr(gc, '_effective_alphas', (1.0, 1.0)) |
|
|
|
|
|
if gc.get_rgb() is None: |
|
|
|
|
|
|
|
|
|
|
|
gc.set_foreground((0, 0, 0, 0), isRGBA=True) |
|
|
|
|
|
if gc._forced_alpha: |
|
|
gc._effective_alphas = (gc._alpha, gc._alpha) |
|
|
elif fillcolor is None or len(fillcolor) < 4: |
|
|
gc._effective_alphas = (gc._rgb[3], 1.0) |
|
|
else: |
|
|
gc._effective_alphas = (gc._rgb[3], fillcolor[3]) |
|
|
|
|
|
delta = self.gc.delta(gc) |
|
|
if delta: |
|
|
self.file.output(*delta) |
|
|
|
|
|
|
|
|
gc._fillcolor = orig_fill |
|
|
gc._effective_alphas = orig_alphas |
|
|
|
|
|
def get_image_magnification(self): |
|
|
return self.image_dpi/72.0 |
|
|
|
|
|
def draw_image(self, gc, x, y, im, transform=None): |
|
|
|
|
|
|
|
|
h, w = im.shape[:2] |
|
|
if w == 0 or h == 0: |
|
|
return |
|
|
|
|
|
if transform is None: |
|
|
|
|
|
gc.set_alpha(1.0) |
|
|
|
|
|
self.check_gc(gc) |
|
|
|
|
|
w = 72.0 * w / self.image_dpi |
|
|
h = 72.0 * h / self.image_dpi |
|
|
|
|
|
imob = self.file.imageObject(im) |
|
|
|
|
|
if transform is None: |
|
|
self.file.output(Op.gsave, |
|
|
w, 0, 0, h, x, y, Op.concat_matrix, |
|
|
imob, Op.use_xobject, Op.grestore) |
|
|
else: |
|
|
tr1, tr2, tr3, tr4, tr5, tr6 = transform.frozen().to_values() |
|
|
|
|
|
self.file.output(Op.gsave, |
|
|
1, 0, 0, 1, x, y, Op.concat_matrix, |
|
|
tr1, tr2, tr3, tr4, tr5, tr6, Op.concat_matrix, |
|
|
imob, Op.use_xobject, Op.grestore) |
|
|
|
|
|
def draw_path(self, gc, path, transform, rgbFace=None): |
|
|
|
|
|
self.check_gc(gc, rgbFace) |
|
|
self.file.writePath( |
|
|
path, transform, |
|
|
rgbFace is None and gc.get_hatch_path() is None, |
|
|
gc.get_sketch_params()) |
|
|
self.file.output(self.gc.paint()) |
|
|
|
|
|
def draw_path_collection(self, gc, master_transform, paths, all_transforms, |
|
|
offsets, offset_trans, facecolors, edgecolors, |
|
|
linewidths, linestyles, antialiaseds, urls, |
|
|
offset_position): |
|
|
|
|
|
|
|
|
|
|
|
can_do_optimization = True |
|
|
facecolors = np.asarray(facecolors) |
|
|
edgecolors = np.asarray(edgecolors) |
|
|
|
|
|
if not len(facecolors): |
|
|
filled = False |
|
|
can_do_optimization = not gc.get_hatch() |
|
|
else: |
|
|
if np.all(facecolors[:, 3] == facecolors[0, 3]): |
|
|
filled = facecolors[0, 3] != 0.0 |
|
|
else: |
|
|
can_do_optimization = False |
|
|
|
|
|
if not len(edgecolors): |
|
|
stroked = False |
|
|
else: |
|
|
if np.all(np.asarray(linewidths) == 0.0): |
|
|
stroked = False |
|
|
elif np.all(edgecolors[:, 3] == edgecolors[0, 3]): |
|
|
stroked = edgecolors[0, 3] != 0.0 |
|
|
else: |
|
|
can_do_optimization = False |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
len_path = len(paths[0].vertices) if len(paths) > 0 else 0 |
|
|
uses_per_path = self._iter_collection_uses_per_path( |
|
|
paths, all_transforms, offsets, facecolors, edgecolors) |
|
|
should_do_optimization = \ |
|
|
len_path + uses_per_path + 5 < len_path * uses_per_path |
|
|
|
|
|
if (not can_do_optimization) or (not should_do_optimization): |
|
|
return RendererBase.draw_path_collection( |
|
|
self, gc, master_transform, paths, all_transforms, |
|
|
offsets, offset_trans, facecolors, edgecolors, |
|
|
linewidths, linestyles, antialiaseds, urls, |
|
|
offset_position) |
|
|
|
|
|
padding = np.max(linewidths) |
|
|
path_codes = [] |
|
|
for i, (path, transform) in enumerate(self._iter_collection_raw_paths( |
|
|
master_transform, paths, all_transforms)): |
|
|
name = self.file.pathCollectionObject( |
|
|
gc, path, transform, padding, filled, stroked) |
|
|
path_codes.append(name) |
|
|
|
|
|
output = self.file.output |
|
|
output(*self.gc.push()) |
|
|
lastx, lasty = 0, 0 |
|
|
for xo, yo, path_id, gc0, rgbFace in self._iter_collection( |
|
|
gc, path_codes, offsets, offset_trans, |
|
|
facecolors, edgecolors, linewidths, linestyles, |
|
|
antialiaseds, urls, offset_position): |
|
|
|
|
|
self.check_gc(gc0, rgbFace) |
|
|
dx, dy = xo - lastx, yo - lasty |
|
|
output(1, 0, 0, 1, dx, dy, Op.concat_matrix, path_id, |
|
|
Op.use_xobject) |
|
|
lastx, lasty = xo, yo |
|
|
output(*self.gc.pop()) |
|
|
|
|
|
def draw_markers(self, gc, marker_path, marker_trans, path, trans, |
|
|
rgbFace=None): |
|
|
|
|
|
|
|
|
|
|
|
len_marker_path = len(marker_path) |
|
|
uses = len(path) |
|
|
if len_marker_path * uses < len_marker_path + uses + 5: |
|
|
RendererBase.draw_markers(self, gc, marker_path, marker_trans, |
|
|
path, trans, rgbFace) |
|
|
return |
|
|
|
|
|
self.check_gc(gc, rgbFace) |
|
|
fill = gc.fill(rgbFace) |
|
|
stroke = gc.stroke() |
|
|
|
|
|
output = self.file.output |
|
|
marker = self.file.markerObject( |
|
|
marker_path, marker_trans, fill, stroke, self.gc._linewidth, |
|
|
gc.get_joinstyle(), gc.get_capstyle()) |
|
|
|
|
|
output(Op.gsave) |
|
|
lastx, lasty = 0, 0 |
|
|
for vertices, code in path.iter_segments( |
|
|
trans, |
|
|
clip=(0, 0, self.file.width*72, self.file.height*72), |
|
|
simplify=False): |
|
|
if len(vertices): |
|
|
x, y = vertices[-2:] |
|
|
if not (0 <= x <= self.file.width * 72 |
|
|
and 0 <= y <= self.file.height * 72): |
|
|
continue |
|
|
dx, dy = x - lastx, y - lasty |
|
|
output(1, 0, 0, 1, dx, dy, Op.concat_matrix, |
|
|
marker, Op.use_xobject) |
|
|
lastx, lasty = x, y |
|
|
output(Op.grestore) |
|
|
|
|
|
def draw_gouraud_triangle(self, gc, points, colors, trans): |
|
|
self.draw_gouraud_triangles(gc, points.reshape((1, 3, 2)), |
|
|
colors.reshape((1, 3, 4)), trans) |
|
|
|
|
|
def draw_gouraud_triangles(self, gc, points, colors, trans): |
|
|
assert len(points) == len(colors) |
|
|
if len(points) == 0: |
|
|
return |
|
|
assert points.ndim == 3 |
|
|
assert points.shape[1] == 3 |
|
|
assert points.shape[2] == 2 |
|
|
assert colors.ndim == 3 |
|
|
assert colors.shape[1] == 3 |
|
|
assert colors.shape[2] in (1, 4) |
|
|
|
|
|
shape = points.shape |
|
|
points = points.reshape((shape[0] * shape[1], 2)) |
|
|
tpoints = trans.transform(points) |
|
|
tpoints = tpoints.reshape(shape) |
|
|
name, _ = self.file.addGouraudTriangles(tpoints, colors) |
|
|
output = self.file.output |
|
|
|
|
|
if colors.shape[2] == 1: |
|
|
|
|
|
gc.set_alpha(1.0) |
|
|
self.check_gc(gc) |
|
|
output(name, Op.shading) |
|
|
return |
|
|
|
|
|
alpha = colors[0, 0, 3] |
|
|
if np.allclose(alpha, colors[:, :, 3]): |
|
|
|
|
|
gc.set_alpha(alpha) |
|
|
self.check_gc(gc) |
|
|
output(name, Op.shading) |
|
|
else: |
|
|
|
|
|
alpha = colors[:, :, 3][:, :, None] |
|
|
_, smask_ob = self.file.addGouraudTriangles(tpoints, alpha) |
|
|
gstate = self.file._soft_mask_state(smask_ob) |
|
|
output(Op.gsave, gstate, Op.setgstate, |
|
|
name, Op.shading, |
|
|
Op.grestore) |
|
|
|
|
|
def _setup_textpos(self, x, y, angle, oldx=0, oldy=0, oldangle=0): |
|
|
if angle == oldangle == 0: |
|
|
self.file.output(x - oldx, y - oldy, Op.textpos) |
|
|
else: |
|
|
angle = math.radians(angle) |
|
|
self.file.output(math.cos(angle), math.sin(angle), |
|
|
-math.sin(angle), math.cos(angle), |
|
|
x, y, Op.textmatrix) |
|
|
self.file.output(0, 0, Op.textpos) |
|
|
|
|
|
def draw_mathtext(self, gc, x, y, s, prop, angle): |
|
|
|
|
|
width, height, descent, glyphs, rects = \ |
|
|
self._text2path.mathtext_parser.parse(s, 72, prop) |
|
|
|
|
|
if gc.get_url() is not None: |
|
|
self.file._annotations[-1][1].append(_get_link_annotation( |
|
|
gc, x, y, width, height, angle)) |
|
|
|
|
|
fonttype = mpl.rcParams['pdf.fonttype'] |
|
|
|
|
|
|
|
|
a = math.radians(angle) |
|
|
self.file.output(Op.gsave) |
|
|
self.file.output(math.cos(a), math.sin(a), |
|
|
-math.sin(a), math.cos(a), |
|
|
x, y, Op.concat_matrix) |
|
|
|
|
|
self.check_gc(gc, gc._rgb) |
|
|
prev_font = None, None |
|
|
oldx, oldy = 0, 0 |
|
|
unsupported_chars = [] |
|
|
|
|
|
self.file.output(Op.begin_text) |
|
|
for font, fontsize, num, ox, oy in glyphs: |
|
|
self.file._character_tracker.track_glyph(font, num) |
|
|
fontname = font.fname |
|
|
if not _font_supports_glyph(fonttype, num): |
|
|
|
|
|
|
|
|
unsupported_chars.append((font, fontsize, ox, oy, num)) |
|
|
else: |
|
|
self._setup_textpos(ox, oy, 0, oldx, oldy) |
|
|
oldx, oldy = ox, oy |
|
|
if (fontname, fontsize) != prev_font: |
|
|
self.file.output(self.file.fontName(fontname), fontsize, |
|
|
Op.selectfont) |
|
|
prev_font = fontname, fontsize |
|
|
self.file.output(self.encode_string(chr(num), fonttype), |
|
|
Op.show) |
|
|
self.file.output(Op.end_text) |
|
|
|
|
|
for font, fontsize, ox, oy, num in unsupported_chars: |
|
|
self._draw_xobject_glyph( |
|
|
font, fontsize, font.get_char_index(num), ox, oy) |
|
|
|
|
|
|
|
|
for ox, oy, width, height in rects: |
|
|
self.file.output(Op.gsave, ox, oy, width, height, |
|
|
Op.rectangle, Op.fill, Op.grestore) |
|
|
|
|
|
|
|
|
self.file.output(Op.grestore) |
|
|
|
|
|
def draw_tex(self, gc, x, y, s, prop, angle, *, mtext=None): |
|
|
|
|
|
texmanager = self.get_texmanager() |
|
|
fontsize = prop.get_size_in_points() |
|
|
dvifile = texmanager.make_dvi(s, fontsize) |
|
|
with dviread.Dvi(dvifile, 72) as dvi: |
|
|
page, = dvi |
|
|
|
|
|
if gc.get_url() is not None: |
|
|
self.file._annotations[-1][1].append(_get_link_annotation( |
|
|
gc, x, y, page.width, page.height, angle)) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
oldfont, seq = None, [] |
|
|
for x1, y1, dvifont, glyph, width in page.text: |
|
|
if dvifont != oldfont: |
|
|
pdfname = self.file.dviFontName(dvifont) |
|
|
seq += [['font', pdfname, dvifont.size]] |
|
|
oldfont = dvifont |
|
|
seq += [['text', x1, y1, [bytes([glyph])], x1+width]] |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
i, curx, fontsize = 0, 0, None |
|
|
while i < len(seq)-1: |
|
|
elt, nxt = seq[i:i+2] |
|
|
if elt[0] == 'font': |
|
|
fontsize = elt[2] |
|
|
elif elt[0] == nxt[0] == 'text' and elt[2] == nxt[2]: |
|
|
offset = elt[4] - nxt[1] |
|
|
if abs(offset) < 0.1: |
|
|
elt[3][-1] += nxt[3][0] |
|
|
elt[4] += nxt[4]-nxt[1] |
|
|
else: |
|
|
elt[3] += [offset*1000.0/fontsize, nxt[3][0]] |
|
|
elt[4] = nxt[4] |
|
|
del seq[i+1] |
|
|
continue |
|
|
i += 1 |
|
|
|
|
|
|
|
|
mytrans = Affine2D().rotate_deg(angle).translate(x, y) |
|
|
|
|
|
|
|
|
self.check_gc(gc, gc._rgb) |
|
|
self.file.output(Op.begin_text) |
|
|
curx, cury, oldx, oldy = 0, 0, 0, 0 |
|
|
for elt in seq: |
|
|
if elt[0] == 'font': |
|
|
self.file.output(elt[1], elt[2], Op.selectfont) |
|
|
elif elt[0] == 'text': |
|
|
curx, cury = mytrans.transform((elt[1], elt[2])) |
|
|
self._setup_textpos(curx, cury, angle, oldx, oldy) |
|
|
oldx, oldy = curx, cury |
|
|
if len(elt[3]) == 1: |
|
|
self.file.output(elt[3][0], Op.show) |
|
|
else: |
|
|
self.file.output(elt[3], Op.showkern) |
|
|
else: |
|
|
assert False |
|
|
self.file.output(Op.end_text) |
|
|
|
|
|
|
|
|
|
|
|
boxgc = self.new_gc() |
|
|
boxgc.copy_properties(gc) |
|
|
boxgc.set_linewidth(0) |
|
|
pathops = [Path.MOVETO, Path.LINETO, Path.LINETO, Path.LINETO, |
|
|
Path.CLOSEPOLY] |
|
|
for x1, y1, h, w in page.boxes: |
|
|
path = Path([[x1, y1], [x1+w, y1], [x1+w, y1+h], [x1, y1+h], |
|
|
[0, 0]], pathops) |
|
|
self.draw_path(boxgc, path, mytrans, gc._rgb) |
|
|
|
|
|
def encode_string(self, s, fonttype): |
|
|
if fonttype in (1, 3): |
|
|
return s.encode('cp1252', 'replace') |
|
|
return s.encode('utf-16be', 'replace') |
|
|
|
|
|
def draw_text(self, gc, x, y, s, prop, angle, ismath=False, mtext=None): |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
self.check_gc(gc, gc._rgb) |
|
|
if ismath: |
|
|
return self.draw_mathtext(gc, x, y, s, prop, angle) |
|
|
|
|
|
fontsize = prop.get_size_in_points() |
|
|
|
|
|
if mpl.rcParams['pdf.use14corefonts']: |
|
|
font = self._get_font_afm(prop) |
|
|
fonttype = 1 |
|
|
else: |
|
|
font = self._get_font_ttf(prop) |
|
|
self.file._character_tracker.track(font, s) |
|
|
fonttype = mpl.rcParams['pdf.fonttype'] |
|
|
|
|
|
if gc.get_url() is not None: |
|
|
font.set_text(s) |
|
|
width, height = font.get_width_height() |
|
|
self.file._annotations[-1][1].append(_get_link_annotation( |
|
|
gc, x, y, width / 64, height / 64, angle)) |
|
|
|
|
|
|
|
|
|
|
|
if fonttype not in [3, 42]: |
|
|
self.file.output(Op.begin_text, |
|
|
self.file.fontName(prop), fontsize, Op.selectfont) |
|
|
self._setup_textpos(x, y, angle) |
|
|
self.file.output(self.encode_string(s, fonttype), |
|
|
Op.show, Op.end_text) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
else: |
|
|
|
|
|
|
|
|
singlebyte_chunks = [] |
|
|
|
|
|
multibyte_glyphs = [] |
|
|
prev_was_multibyte = True |
|
|
prev_font = font |
|
|
for item in _text_helpers.layout( |
|
|
s, font, kern_mode=KERNING_UNFITTED): |
|
|
if _font_supports_glyph(fonttype, ord(item.char)): |
|
|
if prev_was_multibyte or item.ft_object != prev_font: |
|
|
singlebyte_chunks.append((item.ft_object, item.x, [])) |
|
|
prev_font = item.ft_object |
|
|
if item.prev_kern: |
|
|
singlebyte_chunks[-1][2].append(item.prev_kern) |
|
|
singlebyte_chunks[-1][2].append(item.char) |
|
|
prev_was_multibyte = False |
|
|
else: |
|
|
multibyte_glyphs.append( |
|
|
(item.ft_object, item.x, item.glyph_idx) |
|
|
) |
|
|
prev_was_multibyte = True |
|
|
|
|
|
|
|
|
self.file.output(Op.gsave) |
|
|
a = math.radians(angle) |
|
|
self.file.output(math.cos(a), math.sin(a), |
|
|
-math.sin(a), math.cos(a), |
|
|
x, y, Op.concat_matrix) |
|
|
|
|
|
|
|
|
self.file.output(Op.begin_text) |
|
|
prev_start_x = 0 |
|
|
for ft_object, start_x, kerns_or_chars in singlebyte_chunks: |
|
|
ft_name = self.file.fontName(ft_object.fname) |
|
|
self.file.output(ft_name, fontsize, Op.selectfont) |
|
|
self._setup_textpos(start_x, 0, 0, prev_start_x, 0, 0) |
|
|
self.file.output( |
|
|
|
|
|
|
|
|
[-1000 * next(group) / fontsize if tp == float |
|
|
else self.encode_string("".join(group), fonttype) |
|
|
for tp, group in itertools.groupby(kerns_or_chars, type)], |
|
|
Op.showkern) |
|
|
prev_start_x = start_x |
|
|
self.file.output(Op.end_text) |
|
|
|
|
|
for ft_object, start_x, glyph_idx in multibyte_glyphs: |
|
|
self._draw_xobject_glyph( |
|
|
ft_object, fontsize, glyph_idx, start_x, 0 |
|
|
) |
|
|
self.file.output(Op.grestore) |
|
|
|
|
|
def _draw_xobject_glyph(self, font, fontsize, glyph_idx, x, y): |
|
|
"""Draw a multibyte character from a Type 3 font as an XObject.""" |
|
|
glyph_name = font.get_glyph_name(glyph_idx) |
|
|
name = self.file._get_xobject_glyph_name(font.fname, glyph_name) |
|
|
self.file.output( |
|
|
Op.gsave, |
|
|
0.001 * fontsize, 0, 0, 0.001 * fontsize, x, y, Op.concat_matrix, |
|
|
Name(name), Op.use_xobject, |
|
|
Op.grestore, |
|
|
) |
|
|
|
|
|
def new_gc(self): |
|
|
|
|
|
return GraphicsContextPdf(self.file) |
|
|
|
|
|
|
|
|
class GraphicsContextPdf(GraphicsContextBase): |
|
|
|
|
|
def __init__(self, file): |
|
|
super().__init__() |
|
|
self._fillcolor = (0.0, 0.0, 0.0) |
|
|
self._effective_alphas = (1.0, 1.0) |
|
|
self.file = file |
|
|
self.parent = None |
|
|
|
|
|
def __repr__(self): |
|
|
d = dict(self.__dict__) |
|
|
del d['file'] |
|
|
del d['parent'] |
|
|
return repr(d) |
|
|
|
|
|
def stroke(self): |
|
|
""" |
|
|
Predicate: does the path need to be stroked (its outline drawn)? |
|
|
This tests for the various conditions that disable stroking |
|
|
the path, in which case it would presumably be filled. |
|
|
""" |
|
|
|
|
|
|
|
|
return (self._linewidth > 0 and self._alpha > 0 and |
|
|
(len(self._rgb) <= 3 or self._rgb[3] != 0.0)) |
|
|
|
|
|
def fill(self, *args): |
|
|
""" |
|
|
Predicate: does the path need to be filled? |
|
|
|
|
|
An optional argument can be used to specify an alternative |
|
|
_fillcolor, as needed by RendererPdf.draw_markers. |
|
|
""" |
|
|
if len(args): |
|
|
_fillcolor = args[0] |
|
|
else: |
|
|
_fillcolor = self._fillcolor |
|
|
return (self._hatch or |
|
|
(_fillcolor is not None and |
|
|
(len(_fillcolor) <= 3 or _fillcolor[3] != 0.0))) |
|
|
|
|
|
def paint(self): |
|
|
""" |
|
|
Return the appropriate pdf operator to cause the path to be |
|
|
stroked, filled, or both. |
|
|
""" |
|
|
return Op.paint_path(self.fill(), self.stroke()) |
|
|
|
|
|
capstyles = {'butt': 0, 'round': 1, 'projecting': 2} |
|
|
joinstyles = {'miter': 0, 'round': 1, 'bevel': 2} |
|
|
|
|
|
def capstyle_cmd(self, style): |
|
|
return [self.capstyles[style], Op.setlinecap] |
|
|
|
|
|
def joinstyle_cmd(self, style): |
|
|
return [self.joinstyles[style], Op.setlinejoin] |
|
|
|
|
|
def linewidth_cmd(self, width): |
|
|
return [width, Op.setlinewidth] |
|
|
|
|
|
def dash_cmd(self, dashes): |
|
|
offset, dash = dashes |
|
|
if dash is None: |
|
|
dash = [] |
|
|
offset = 0 |
|
|
return [list(dash), offset, Op.setdash] |
|
|
|
|
|
def alpha_cmd(self, alpha, forced, effective_alphas): |
|
|
name = self.file.alphaState(effective_alphas) |
|
|
return [name, Op.setgstate] |
|
|
|
|
|
def hatch_cmd(self, hatch, hatch_color): |
|
|
if not hatch: |
|
|
if self._fillcolor is not None: |
|
|
return self.fillcolor_cmd(self._fillcolor) |
|
|
else: |
|
|
return [Name('DeviceRGB'), Op.setcolorspace_nonstroke] |
|
|
else: |
|
|
hatch_style = (hatch_color, self._fillcolor, hatch) |
|
|
name = self.file.hatchPattern(hatch_style) |
|
|
return [Name('Pattern'), Op.setcolorspace_nonstroke, |
|
|
name, Op.setcolor_nonstroke] |
|
|
|
|
|
def rgb_cmd(self, rgb): |
|
|
if mpl.rcParams['pdf.inheritcolor']: |
|
|
return [] |
|
|
if rgb[0] == rgb[1] == rgb[2]: |
|
|
return [rgb[0], Op.setgray_stroke] |
|
|
else: |
|
|
return [*rgb[:3], Op.setrgb_stroke] |
|
|
|
|
|
def fillcolor_cmd(self, rgb): |
|
|
if rgb is None or mpl.rcParams['pdf.inheritcolor']: |
|
|
return [] |
|
|
elif rgb[0] == rgb[1] == rgb[2]: |
|
|
return [rgb[0], Op.setgray_nonstroke] |
|
|
else: |
|
|
return [*rgb[:3], Op.setrgb_nonstroke] |
|
|
|
|
|
def push(self): |
|
|
parent = GraphicsContextPdf(self.file) |
|
|
parent.copy_properties(self) |
|
|
parent.parent = self.parent |
|
|
self.parent = parent |
|
|
return [Op.gsave] |
|
|
|
|
|
def pop(self): |
|
|
assert self.parent is not None |
|
|
self.copy_properties(self.parent) |
|
|
self.parent = self.parent.parent |
|
|
return [Op.grestore] |
|
|
|
|
|
def clip_cmd(self, cliprect, clippath): |
|
|
"""Set clip rectangle. Calls `.pop()` and `.push()`.""" |
|
|
cmds = [] |
|
|
|
|
|
while ((self._cliprect, self._clippath) != (cliprect, clippath) |
|
|
and self.parent is not None): |
|
|
cmds.extend(self.pop()) |
|
|
|
|
|
if ((self._cliprect, self._clippath) != (cliprect, clippath) or |
|
|
self.parent is None): |
|
|
cmds.extend(self.push()) |
|
|
if self._cliprect != cliprect: |
|
|
cmds.extend([cliprect, Op.rectangle, Op.clip, Op.endpath]) |
|
|
if self._clippath != clippath: |
|
|
path, affine = clippath.get_transformed_path_and_affine() |
|
|
cmds.extend( |
|
|
PdfFile.pathOperations(path, affine, simplify=False) + |
|
|
[Op.clip, Op.endpath]) |
|
|
return cmds |
|
|
|
|
|
commands = ( |
|
|
|
|
|
(('_cliprect', '_clippath'), clip_cmd), |
|
|
(('_alpha', '_forced_alpha', '_effective_alphas'), alpha_cmd), |
|
|
(('_capstyle',), capstyle_cmd), |
|
|
(('_fillcolor',), fillcolor_cmd), |
|
|
(('_joinstyle',), joinstyle_cmd), |
|
|
(('_linewidth',), linewidth_cmd), |
|
|
(('_dashes',), dash_cmd), |
|
|
(('_rgb',), rgb_cmd), |
|
|
|
|
|
(('_hatch', '_hatch_color'), hatch_cmd), |
|
|
) |
|
|
|
|
|
def delta(self, other): |
|
|
""" |
|
|
Copy properties of other into self and return PDF commands |
|
|
needed to transform *self* into *other*. |
|
|
""" |
|
|
cmds = [] |
|
|
fill_performed = False |
|
|
for params, cmd in self.commands: |
|
|
different = False |
|
|
for p in params: |
|
|
ours = getattr(self, p) |
|
|
theirs = getattr(other, p) |
|
|
try: |
|
|
if ours is None or theirs is None: |
|
|
different = ours is not theirs |
|
|
else: |
|
|
different = bool(ours != theirs) |
|
|
except ValueError: |
|
|
ours = np.asarray(ours) |
|
|
theirs = np.asarray(theirs) |
|
|
different = (ours.shape != theirs.shape or |
|
|
np.any(ours != theirs)) |
|
|
if different: |
|
|
break |
|
|
|
|
|
|
|
|
if params == ('_hatch', '_hatch_color') and fill_performed: |
|
|
different = True |
|
|
|
|
|
if different: |
|
|
if params == ('_fillcolor',): |
|
|
fill_performed = True |
|
|
theirs = [getattr(other, p) for p in params] |
|
|
cmds.extend(cmd(self, *theirs)) |
|
|
for p in params: |
|
|
setattr(self, p, getattr(other, p)) |
|
|
return cmds |
|
|
|
|
|
def copy_properties(self, other): |
|
|
""" |
|
|
Copy properties of other into self. |
|
|
""" |
|
|
super().copy_properties(other) |
|
|
fillcolor = getattr(other, '_fillcolor', self._fillcolor) |
|
|
effective_alphas = getattr(other, '_effective_alphas', |
|
|
self._effective_alphas) |
|
|
self._fillcolor = fillcolor |
|
|
self._effective_alphas = effective_alphas |
|
|
|
|
|
def finalize(self): |
|
|
""" |
|
|
Make sure every pushed graphics state is popped. |
|
|
""" |
|
|
cmds = [] |
|
|
while self.parent is not None: |
|
|
cmds.extend(self.pop()) |
|
|
return cmds |
|
|
|
|
|
|
|
|
class PdfPages: |
|
|
""" |
|
|
A multi-page PDF file. |
|
|
|
|
|
Examples |
|
|
-------- |
|
|
>>> import matplotlib.pyplot as plt |
|
|
>>> # Initialize: |
|
|
>>> with PdfPages('foo.pdf') as pdf: |
|
|
... # As many times as you like, create a figure fig and save it: |
|
|
... fig = plt.figure() |
|
|
... pdf.savefig(fig) |
|
|
... # When no figure is specified the current figure is saved |
|
|
... pdf.savefig() |
|
|
|
|
|
Notes |
|
|
----- |
|
|
In reality `PdfPages` is a thin wrapper around `PdfFile`, in order to avoid |
|
|
confusion when using `~.pyplot.savefig` and forgetting the format argument. |
|
|
""" |
|
|
__slots__ = ('_file', 'keep_empty') |
|
|
|
|
|
def __init__(self, filename, keep_empty=True, metadata=None): |
|
|
""" |
|
|
Create a new PdfPages object. |
|
|
|
|
|
Parameters |
|
|
---------- |
|
|
filename : str or path-like or file-like |
|
|
Plots using `PdfPages.savefig` will be written to a file at this |
|
|
location. The file is opened at once and any older file with the |
|
|
same name is overwritten. |
|
|
|
|
|
keep_empty : bool, optional |
|
|
If set to False, then empty pdf files will be deleted automatically |
|
|
when closed. |
|
|
|
|
|
metadata : dict, optional |
|
|
Information dictionary object (see PDF reference section 10.2.1 |
|
|
'Document Information Dictionary'), e.g.: |
|
|
``{'Creator': 'My software', 'Author': 'Me', 'Title': 'Awesome'}``. |
|
|
|
|
|
The standard keys are 'Title', 'Author', 'Subject', 'Keywords', |
|
|
'Creator', 'Producer', 'CreationDate', 'ModDate', and |
|
|
'Trapped'. Values have been predefined for 'Creator', 'Producer' |
|
|
and 'CreationDate'. They can be removed by setting them to `None`. |
|
|
""" |
|
|
self._file = PdfFile(filename, metadata=metadata) |
|
|
self.keep_empty = keep_empty |
|
|
|
|
|
def __enter__(self): |
|
|
return self |
|
|
|
|
|
def __exit__(self, exc_type, exc_val, exc_tb): |
|
|
self.close() |
|
|
|
|
|
def close(self): |
|
|
""" |
|
|
Finalize this object, making the underlying file a complete |
|
|
PDF file. |
|
|
""" |
|
|
self._file.finalize() |
|
|
self._file.close() |
|
|
if (self.get_pagecount() == 0 and not self.keep_empty and |
|
|
not self._file.passed_in_file_object): |
|
|
os.remove(self._file.fh.name) |
|
|
self._file = None |
|
|
|
|
|
def infodict(self): |
|
|
""" |
|
|
Return a modifiable information dictionary object |
|
|
(see PDF reference section 10.2.1 'Document Information |
|
|
Dictionary'). |
|
|
""" |
|
|
return self._file.infoDict |
|
|
|
|
|
def savefig(self, figure=None, **kwargs): |
|
|
""" |
|
|
Save a `.Figure` to this file as a new page. |
|
|
|
|
|
Any other keyword arguments are passed to `~.Figure.savefig`. |
|
|
|
|
|
Parameters |
|
|
---------- |
|
|
figure : `.Figure` or int, default: the active figure |
|
|
The figure, or index of the figure, that is saved to the file. |
|
|
""" |
|
|
if not isinstance(figure, Figure): |
|
|
if figure is None: |
|
|
manager = Gcf.get_active() |
|
|
else: |
|
|
manager = Gcf.get_fig_manager(figure) |
|
|
if manager is None: |
|
|
raise ValueError("No figure {}".format(figure)) |
|
|
figure = manager.canvas.figure |
|
|
|
|
|
try: |
|
|
orig_canvas = figure.canvas |
|
|
figure.canvas = FigureCanvasPdf(figure) |
|
|
figure.savefig(self, format="pdf", **kwargs) |
|
|
finally: |
|
|
figure.canvas = orig_canvas |
|
|
|
|
|
def get_pagecount(self): |
|
|
"""Return the current number of pages in the multipage pdf file.""" |
|
|
return len(self._file.pageList) |
|
|
|
|
|
def attach_note(self, text, positionRect=[-100, -100, 0, 0]): |
|
|
""" |
|
|
Add a new text note to the page to be saved next. The optional |
|
|
positionRect specifies the position of the new note on the |
|
|
page. It is outside the page per default to make sure it is |
|
|
invisible on printouts. |
|
|
""" |
|
|
self._file.newTextnote(text, positionRect) |
|
|
|
|
|
|
|
|
class FigureCanvasPdf(FigureCanvasBase): |
|
|
|
|
|
|
|
|
fixed_dpi = 72 |
|
|
filetypes = {'pdf': 'Portable Document Format'} |
|
|
|
|
|
def get_default_filetype(self): |
|
|
return 'pdf' |
|
|
|
|
|
def print_pdf(self, filename, *, |
|
|
bbox_inches_restore=None, metadata=None): |
|
|
|
|
|
dpi = self.figure.dpi |
|
|
self.figure.dpi = 72 |
|
|
width, height = self.figure.get_size_inches() |
|
|
if isinstance(filename, PdfPages): |
|
|
file = filename._file |
|
|
else: |
|
|
file = PdfFile(filename, metadata=metadata) |
|
|
try: |
|
|
file.newPage(width, height) |
|
|
renderer = MixedModeRenderer( |
|
|
self.figure, width, height, dpi, |
|
|
RendererPdf(file, dpi, height, width), |
|
|
bbox_inches_restore=bbox_inches_restore) |
|
|
self.figure.draw(renderer) |
|
|
renderer.finalize() |
|
|
if not isinstance(filename, PdfPages): |
|
|
file.finalize() |
|
|
finally: |
|
|
if isinstance(filename, PdfPages): |
|
|
file.endStream() |
|
|
else: |
|
|
file.close() |
|
|
|
|
|
def draw(self): |
|
|
self.figure.draw_without_rendering() |
|
|
return super().draw() |
|
|
|
|
|
|
|
|
FigureManagerPdf = FigureManagerBase |
|
|
|
|
|
|
|
|
@_Backend.export |
|
|
class _BackendPdf(_Backend): |
|
|
FigureCanvas = FigureCanvasPdf |
|
|
|