Last active
January 28, 2025 12:53
-
-
Save Alistair1231/8c759bb171800bd87d3c0725f1ed5d2a to your computer and use it in GitHub Desktop.
latexrun customizations
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
# for new image tags check: https://gitlab.com/islandoftex/images/texlive/container_registry/573747 | |
FROM registry.gitlab.com/islandoftex/images/texlive:TL2024-2025-01-26-full-doc | |
# create a non-root user | |
ARG USERNAME=vscode | |
ARG USER_UID=1000 | |
ARG USER_GID=$USER_UID | |
RUN apt-get update && \ | |
apt-get install -y man sudo openssh-client neovim inkscape zsh tmux && \ | |
ln -s /usr/bin/nvim /usr/bin/vim && \ | |
rm -rf /var/lib/apt/lists/* | |
RUN groupadd --gid $USER_GID $USERNAME && \ | |
useradd --uid $USER_UID --gid $USER_GID -m $USERNAME && \ | |
echo $USERNAME ALL=\(root\) NOPASSWD:ALL > /etc/sudoers.d/$USERNAME && \ | |
chmod 0440 /etc/sudoers.d/$USERNAME | |
USER $USERNAME | |
## install brew (watchman is only available on brew) | |
# first its dependencies | |
RUN sudo apt-get update && sudo apt-get install -y build-essential procps curl file git && sudo rm -rf /var/lib/apt/lists/* | |
# then brew | |
RUN NONINTERACTIVE=1 /bin/bash -c "$(curl -fsSL https://raw.githubusercontent.com/Homebrew/install/HEAD/install.sh)" | |
# then set the path and disable auto update | |
ENV HOMEBREW_NO_AUTO_UPDATE=1 | |
ENV PATH=/home/linuxbrew/.linuxbrew/bin:$PATH | |
# then install the packages | |
RUN brew install uv watchman pygments tex-fmt | |
# set zsh as default shell | |
RUN sudo chsh -s $(which zsh) $USERNAME | |
# Download latexrun | |
RUN sudo curl -Lo /usr/bin/latexrun https://gist.github.com/Alistair1231/8c759bb171800bd87d3c0725f1ed5d2a/raw/latexrun && \ | |
sudo chmod +x /usr/bin/latexrun |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
#!/usr/bin/env -S uv run --script | |
# /// script | |
# requires-python = ">=3.13" | |
# dependencies = [] | |
# /// | |
# Copyright (c) 2013, 2014 Austin Clements | |
# Copyright (c) 2025 Alistair1231 | |
# Copyright (c) 2025 Ferdi265 | |
# Changes: | |
# - Time the execution of the latex command | |
# - Set Default Latex Command to lualatex | |
# - Set Default Bibtex Command to biber | |
# - add shell-escape and synctex=1 to the default latex command | |
# - set verbose_cmds to True by default | |
# - Fix by @Ferdi265 for "SyntaxWarning: invalid escape sequence" on Python >=3.12 | |
# https://github.com/aclements/latexrun/pull/76 | |
# Permission is hereby granted, free of charge, to any person obtaining a copy | |
# of this software and associated documentation files (the "Software"), to deal | |
# in the Software without restriction, including without limitation the rights | |
# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | |
# copies of the Software, and to permit persons to whom the Software is | |
# furnished to do so, subject to the following conditions: | |
# The above copyright notice and this permission notice shall be included in | |
# all copies or substantial portions of the Software. | |
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | |
# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | |
# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE | |
# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | |
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | |
# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN | |
# THE SOFTWARE. | |
import sys | |
import os | |
import errno | |
import argparse | |
import shlex | |
import json | |
import subprocess | |
import re | |
import collections | |
import hashlib | |
import shutil | |
import curses | |
import filecmp | |
import io | |
import traceback | |
import time | |
try: | |
import fcntl | |
except ImportError: | |
# Non-UNIX platform | |
fcntl = None | |
def debug(string, *args): | |
if debug.enabled: | |
print(string.format(*args), file=sys.stderr) | |
debug.enabled = False | |
def debug_exc(): | |
if debug.enabled: | |
traceback.print_exc() | |
def main(): | |
# Parse command-line | |
arg_parser = argparse.ArgumentParser( | |
description="""A 21st century LaTeX wrapper, | |
%(prog)s runs latex (and bibtex) the right number of times so you | |
don't have to, | |
strips the log spew to make errors visible, | |
and plays well with standard build tools.""" | |
) | |
arg_parser.add_argument( | |
"-o", | |
metavar="FILE", | |
dest="output", | |
default=None, | |
help="Output file name (default: derived from input file)", | |
) | |
arg_parser.add_argument( | |
"--latex-cmd", | |
metavar="CMD", | |
default="lualatex", | |
help="Latex command (default: %(default)s)", | |
) | |
arg_parser.add_argument( | |
"--latex-args", | |
metavar="ARGS", | |
type=arg_parser_shlex, | |
help="Additional command-line arguments for latex." | |
" This will be parsed and split using POSIX shell rules.", | |
) | |
arg_parser.add_argument( | |
"--bibtex-cmd", | |
metavar="CMD", | |
default="biber", | |
help="Bibtex command (default: %(default)s)", | |
) | |
arg_parser.add_argument( | |
"--bibtex-args", | |
metavar="ARGS", | |
type=arg_parser_shlex, | |
help="Additional command-line arguments for bibtex", | |
) | |
arg_parser.add_argument( | |
"--max-iterations", | |
metavar="N", | |
type=int, | |
default=10, | |
help="Max number of times to run latex before giving up (default: %(default)s)", | |
) | |
arg_parser.add_argument( | |
"-W", | |
metavar="(no-)CLASS", | |
action=ArgParserWarnAction, | |
dest="nowarns", | |
default=set(["underfull"]), | |
help="Enable/disable warning from CLASS, which can be any package name, " | |
"LaTeX warning class (e.g., font), bad box type " | |
'(underfull, overfull, loose, tight), or "all"', | |
) | |
arg_parser.add_argument( | |
"-O", | |
metavar="DIR", | |
dest="obj_dir", | |
default="latex.out", | |
help="Directory for intermediate files and control database " | |
"(default: %(default)s)", | |
) | |
arg_parser.add_argument( | |
"--color", | |
choices=("auto", "always", "never"), | |
default="auto", | |
help="When to colorize messages", | |
) | |
arg_parser.add_argument( | |
"--verbose-cmds", | |
action="store_true", | |
default=True, | |
help="Print commands as they are executed", | |
) | |
arg_parser.add_argument( | |
"--debug", action="store_true", help="Enable detailed debug output" | |
) | |
actions = arg_parser.add_argument_group("actions") | |
actions.add_argument("--clean-all", action="store_true", help="Delete output files") | |
actions.add_argument("file", nargs="?", help=".tex file to compile") | |
args = arg_parser.parse_args() | |
if not any([args.clean_all, args.file]): | |
arg_parser.error("at least one action is required") | |
args.latex_args = args.latex_args or [] | |
args.bibtex_args = args.bibtex_args or [] | |
verbose_cmd.enabled = args.verbose_cmds | |
debug.enabled = args.debug | |
# A note about encodings: POSIX encoding is a mess; TeX encoding | |
# is a disaster. Our goal is to make things no worse, so we want | |
# byte-accurate round-tripping of TeX messages. Since TeX | |
# messages are *basically* text, we use strings and | |
# surrogateescape'ing for both input and output. I'm not fond of | |
# setting surrogateescape globally, but it's far easier than | |
# dealing with every place we pass TeX output through. | |
# Conveniently, JSON can round-trip surrogateescape'd strings, so | |
# our control database doesn't need special handling. | |
sys.stdout = io.TextIOWrapper( | |
sys.stdout.buffer, | |
encoding=sys.stdout.encoding, | |
errors="surrogateescape", | |
line_buffering=sys.stdout.line_buffering, | |
) | |
sys.stderr = io.TextIOWrapper( | |
sys.stderr.buffer, | |
encoding=sys.stderr.encoding, | |
errors="surrogateescape", | |
line_buffering=sys.stderr.line_buffering, | |
) | |
Message.setup_color(args.color) | |
# Open control database. | |
dbpath = os.path.join(args.obj_dir, ".latexrun.db") | |
if not os.path.exists(dbpath) and os.path.exists(".latexrun.db"): | |
# The control database used to live in the source directory. | |
# Support this for backwards compatibility. | |
dbpath = ".latexrun.db" | |
try: | |
db = DB(dbpath) | |
except (ValueError, OSError) as e: | |
print( | |
"error opening {}: {}".format( | |
e.filename if hasattr(e, "filename") else dbpath, e | |
), | |
file=sys.stderr, | |
) | |
debug_exc() | |
sys.exit(1) | |
# Clean | |
if args.clean_all: | |
try: | |
db.do_clean(args.obj_dir) | |
except OSError as e: | |
print(e, file=sys.stderr) | |
debug_exc() | |
sys.exit(1) | |
# Build | |
if not args.file: | |
return | |
task_commit = None | |
try: | |
task_latex = LaTeX( | |
db, args.file, args.latex_cmd, args.latex_args, args.obj_dir, args.nowarns | |
) | |
task_commit = LaTeXCommit(db, task_latex, args.output) | |
task_bibtex = BibTeX( | |
db, | |
task_latex, | |
args.bibtex_cmd, | |
args.bibtex_args, | |
args.nowarns, | |
args.obj_dir, | |
) | |
tasks = [task_latex, task_commit, task_bibtex] | |
stable = run_tasks(tasks, args.max_iterations) | |
# Print final task output and gather exit status | |
status = 0 | |
for task in tasks: | |
status = max(task.report(), status) | |
if not stable: | |
print( | |
"error: files are still changing after {} iterations; giving up".format( | |
args.max_iterations | |
), | |
file=sys.stderr, | |
) | |
status = max(status, 1) | |
except TaskError as e: | |
print(str(e), file=sys.stderr) | |
debug_exc() | |
status = 1 | |
# Report final status, if interesting | |
fstatus = "There were errors" if task_commit is None else task_commit.status | |
if fstatus: | |
output = args.output | |
if output is None: | |
if task_latex.get_outname() is not None: | |
output = os.path.basename(task_latex.get_outname()) | |
else: | |
output = "output" | |
if Message._color: | |
terminfo.send("bold", ("setaf", 1)) | |
print("{}; {} not updated".format(fstatus, output)) | |
if Message._color: | |
terminfo.send("sgr0") | |
sys.exit(status) | |
def arg_parser_shlex(string): | |
"""Argument parser for shell token lists.""" | |
try: | |
return shlex.split(string) | |
except ValueError as e: | |
raise argparse.ArgumentTypeError(str(e)) from None | |
class ArgParserWarnAction(argparse.Action): | |
def __call__(self, parser, namespace, value, option_string=None): | |
nowarn = getattr(namespace, self.dest) | |
if value == "all": | |
nowarn.clear() | |
elif value.startswith("no-"): | |
nowarn.add(value[3:]) | |
else: | |
nowarn.discard(value) | |
setattr(namespace, self.dest, nowarn) | |
def verbose_cmd(args, cwd=None, env=None): | |
if verbose_cmd.enabled: | |
cmd = " ".join(map(shlex.quote, args)) | |
if cwd is not None: | |
cmd = "(cd {} && {})".format(shlex.quote(cwd), cmd) | |
if env is not None: | |
for k, v in env.items(): | |
if os.environ.get(k) != v: | |
cmd = "{}={} {}".format(k, shlex.quote(v), cmd) | |
print(cmd, file=sys.stderr) | |
verbose_cmd.enabled = False | |
def mkdir_p(path): | |
try: | |
os.makedirs(path) | |
except OSError as exc: | |
if exc.errno == errno.EEXIST and os.path.isdir(path): | |
pass | |
else: | |
raise | |
class DB: | |
"""A latexrun control database.""" | |
_VERSION = "latexrun-db-v2" | |
def __init__(self, filename): | |
self.__filename = filename | |
# Make sure database directory exists | |
if os.path.dirname(self.__filename): | |
os.makedirs(os.path.dirname(self.__filename), exist_ok=True) | |
# Lock the database if possible. We don't release this lock | |
# until the process exits. | |
lockpath = self.__filename + ".lock" | |
if fcntl is not None: | |
lockfd = os.open(lockpath, os.O_CREAT | os.O_WRONLY | os.O_CLOEXEC, 0o666) | |
# Note that this is actually an fcntl lock, not a lockf | |
# lock. Don't be fooled. | |
fcntl.lockf(lockfd, fcntl.LOCK_EX, 1) | |
try: | |
fp = open(filename, "r") | |
except FileNotFoundError: | |
debug("creating new database") | |
self.__val = {"version": DB._VERSION} | |
else: | |
debug("loading database") | |
self.__val = json.load(fp) | |
if "version" not in self.__val: | |
raise ValueError( | |
"file exists, but does not appear to be a latexrun database".format( | |
filename | |
) | |
) | |
if self.__val["version"] != DB._VERSION: | |
raise ValueError( | |
"unknown database version {!r}".format(self.__val["version"]) | |
) | |
def commit(self): | |
debug("committing database") | |
# Atomically commit database | |
tmp_filename = self.__filename + ".tmp" | |
with open(tmp_filename, "w") as fp: | |
json.dump(self.__val, fp, indent=2, separators=(",", ": ")) | |
fp.flush() | |
os.fsync(fp.fileno()) | |
os.rename(tmp_filename, self.__filename) | |
def get_summary(self, task_id): | |
"""Return the recorded summary for the given task or None.""" | |
return self.__val.get("tasks", {}).get(task_id) | |
def set_summary(self, task_id, summary): | |
"""Set the summary for the given task.""" | |
self.__val.setdefault("tasks", {})[task_id] = summary | |
def add_clean(self, filename): | |
"""Add an output file to be cleaned. | |
Unlike the output files recorded in the task summaries, | |
cleanable files strictly accumulate until a clean is | |
performed. | |
""" | |
self.__val.setdefault("clean", {})[filename] = hash_cache.get(filename) | |
def do_clean(self, obj_dir=None): | |
"""Remove output files and delete database. | |
If obj_dir is not None and it is empty after all files are | |
removed, it will also be removed. | |
""" | |
for f, want_hash in self.__val.get("clean", {}).items(): | |
have_hash = hash_cache.get(f) | |
if have_hash is not None: | |
if want_hash == have_hash: | |
debug("unlinking {}", f) | |
hash_cache.invalidate(f) | |
os.unlink(f) | |
else: | |
print( | |
"warning: {} has changed; not removing".format(f), | |
file=sys.stderr, | |
) | |
self.__val = {"version": DB._VERSION} | |
try: | |
os.unlink(self.__filename) | |
except FileNotFoundError: | |
pass | |
if obj_dir is not None: | |
try: | |
os.rmdir(obj_dir) | |
except OSError: | |
pass | |
class HashCache: | |
"""Cache of file hashes. | |
As latexrun reaches fixed-point, it hashes the same files over and | |
over, many of which never change. Since hashing is somewhat | |
expensive, we keep a simple cache of these hashes. | |
""" | |
def __init__(self): | |
self.__cache = {} | |
def get(self, filename): | |
"""Return the hash of filename, or * if it was clobbered.""" | |
try: | |
with open(filename, "rb") as fp: | |
st = os.fstat(fp.fileno()) | |
key = (st.st_dev, st.st_ino) | |
if key in self.__cache: | |
return self.__cache[key] | |
debug("hashing {}", filename) | |
h = hashlib.sha256() | |
while True: | |
block = fp.read(256 * 1024) | |
if not len(block): | |
break | |
h.update(block) | |
self.__cache[key] = h.hexdigest() | |
return self.__cache[key] | |
except (FileNotFoundError, IsADirectoryError): | |
return None | |
def clobber(self, filename): | |
"""If filename's hash is not known, record an invalid hash. | |
This can be used when filename was overwritten before we were | |
necessarily able to obtain its hash. filename must exist. | |
""" | |
st = os.stat(filename) | |
key = (st.st_dev, st.st_ino) | |
if key not in self.__cache: | |
self.__cache[key] = "*" | |
def invalidate(self, filename): | |
try: | |
st = os.stat(filename) | |
except OSError as e: | |
# Pessimistically wipe the whole cache | |
debug("wiping hash cache ({})", e) | |
self.__cache.clear() | |
else: | |
key = (st.st_dev, st.st_ino) | |
if key in self.__cache: | |
del self.__cache[key] | |
hash_cache = HashCache() | |
class _Terminfo: | |
def __init__(self): | |
self.__tty = os.isatty(sys.stdout.fileno()) | |
if self.__tty: | |
curses.setupterm() | |
self.__ti = {} | |
def __ensure(self, cap): | |
if cap not in self.__ti: | |
if not self.__tty: | |
string = None | |
else: | |
string = curses.tigetstr(cap) | |
if string is None or b"$<" in string: | |
# Don't have this capability or it has a pause | |
string = None | |
self.__ti[cap] = string | |
return self.__ti[cap] | |
def has(self, *caps): | |
return all(self.__ensure(cap) is not None for cap in caps) | |
def send(self, *caps): | |
# Flush TextIOWrapper to the binary IO buffer | |
sys.stdout.flush() | |
for cap in caps: | |
# We should use curses.putp here, but it's broken in | |
# Python3 because it writes directly to C's buffered | |
# stdout and there's no way to flush that. | |
if isinstance(cap, tuple): | |
s = curses.tparm(self.__ensure(cap[0]), *cap[1:]) | |
else: | |
s = self.__ensure(cap) | |
sys.stdout.buffer.write(s) | |
terminfo = _Terminfo() | |
class Progress: | |
_enabled = None | |
def __init__(self, prefix): | |
self.__prefix = prefix | |
if Progress._enabled is None: | |
Progress._enabled = (not debug.enabled) and terminfo.has( | |
"cr", "el", "rmam", "smam" | |
) | |
def __enter__(self): | |
self.last = "" | |
self.update("") | |
return self | |
def __exit__(self, typ, value, traceback): | |
if Progress._enabled: | |
# Beginning of line and clear | |
terminfo.send("cr", "el") | |
sys.stdout.flush() | |
def update(self, msg): | |
if not Progress._enabled: | |
return | |
out = "[" + self.__prefix + "]" | |
if msg: | |
out += " " + msg | |
if out != self.last: | |
# Beginning of line, clear line, disable wrap | |
terminfo.send("cr", "el", "rmam") | |
sys.stdout.write(out) | |
# Enable wrap | |
terminfo.send("smam") | |
self.last = out | |
sys.stdout.flush() | |
class Message(collections.namedtuple("Message", "typ filename lineno msg")): | |
def emit(self): | |
if self.filename: | |
if self.filename.startswith("./"): | |
finfo = self.filename[2:] | |
else: | |
finfo = self.filename | |
else: | |
finfo = "<no file>" | |
if self.lineno is not None: | |
finfo += ":" + str(self.lineno) | |
finfo += ": " | |
if self._color: | |
terminfo.send("bold") | |
sys.stdout.write(finfo) | |
if self.typ != "info": | |
if self._color: | |
terminfo.send(("setaf", 5 if self.typ == "warning" else 1)) | |
sys.stdout.write(self.typ + ": ") | |
if self._color: | |
terminfo.send("sgr0") | |
sys.stdout.write(self.msg + "\n") | |
@classmethod | |
def setup_color(cls, state): | |
if state == "never": | |
cls._color = False | |
elif state == "always": | |
cls._color = True | |
elif state == "auto": | |
cls._color = terminfo.has("setaf", "bold", "sgr0") | |
else: | |
raise ValueError("Illegal color state {:r}".format(state)) | |
################################################################## | |
# Task framework | |
# | |
terminate_task_loop = False | |
start_time = time.time() | |
def run_tasks(tasks, max_iterations): | |
"""Execute tasks in round-robin order until all are stable. | |
This will also exit if terminate_task_loop is true. Tasks may use | |
this to terminate after a fatal error (even if that fatal error | |
doesn't necessarily indicate stability; as long as re-running the | |
task will never eliminate the fatal error). | |
Return True if fixed-point is reached or terminate_task_loop is | |
set within max_iterations iterations. | |
""" | |
global terminate_task_loop | |
terminate_task_loop = False | |
nstable = 0 | |
for iteration in range(max_iterations): | |
for task in tasks: | |
if task.stable(): | |
nstable += 1 | |
if nstable == len(tasks): | |
debug("fixed-point reached") | |
return True | |
else: | |
task.run() | |
nstable = 0 | |
if terminate_task_loop: | |
debug("terminate_task_loop set") | |
return True | |
debug("fixed-point not reached") | |
return False | |
class TaskError(Exception): | |
pass | |
class Task: | |
"""A deterministic computation whose inputs and outputs can be captured.""" | |
def __init__(self, db, task_id): | |
self.__db = db | |
self.__task_id = task_id | |
def __debug(self, string, *args): | |
if debug.enabled: | |
debug("task {}: {}", self.__task_id, string.format(*args)) | |
def stable(self): | |
"""Return True if running this task will not affect system state. | |
Functionally, let f be the task, and s be the system state. | |
Then s' = f(s). If it must be that s' == s (that is, f has | |
reached a fixed point), then this function must return True. | |
""" | |
last_summary = self.__db.get_summary(self.__task_id) | |
if last_summary is None: | |
# Task has never run, so running it will modify system | |
# state | |
changed = "never run" | |
else: | |
# If any of the inputs have changed since the last run of | |
# this task, the result may change, so re-run the task. | |
# Also, it's possible something else changed an output | |
# file, in which case we also want to re-run the task, so | |
# check the outputs, too. | |
changed = self.__summary_changed(last_summary) | |
if changed: | |
self.__debug("unstable (changed: {})", changed) | |
return False | |
else: | |
self.__debug("stable") | |
return True | |
def __summary_changed(self, summary): | |
"""Test if any inputs changed from summary. | |
Returns a string describing the changed input, or None. | |
""" | |
for dep in summary["deps"]: | |
fn, args, val = dep | |
method = getattr(self, "_input_" + fn, None) | |
if method is None: | |
return "unknown dependency method {}".format(fn) | |
if method == self._input_unstable or method(*args) != val: | |
return "{}{}".format(fn, tuple(args)) | |
return None | |
def _input(self, name, *args): | |
"""Register an input for this run. | |
This calls self._input_<name>(*args) to get the value of this | |
input. This function should run quickly and return some | |
projection of system state that affects the result of this | |
computation. | |
Both args and the return value must be JSON serializable. | |
""" | |
method = getattr(self, "_input_" + name) | |
val = method(*args) | |
if [name, args, val] not in self.__deps: | |
self.__deps.append([name, args, val]) | |
return val | |
def run(self): | |
# Before we run the task, pre-hash any files that were output | |
# files in the last run. These may be input by this run and | |
# then clobbered, at which point it will be too late to get an | |
# input hash. Ideally we would only hash files that were | |
# *both* input and output files, but latex doesn't tell us | |
# about input files that didn't exist, so if we start from a | |
# clean slate, we often require an extra run because we don't | |
# know a file is input/output until after the second run. | |
last_summary = self.__db.get_summary(self.__task_id) | |
if last_summary is not None: | |
for io_filename in last_summary["output_files"]: | |
self.__debug("pre-hashing {}", io_filename) | |
hash_cache.get(io_filename) | |
# Run the task | |
self.__debug("running") | |
self.__deps = [] | |
result = self._execute() | |
# Clear cached output file hashes | |
for filename in result.output_filenames: | |
hash_cache.invalidate(filename) | |
# If the output files change, then the computation needs to be | |
# re-run, so record them as inputs | |
for filename in result.output_filenames: | |
self._input("file", filename) | |
# Update task summary in database | |
self.__db.set_summary(self.__task_id, self.__make_summary(self.__deps, result)) | |
del self.__deps | |
# Add output files to be cleaned | |
for f in result.output_filenames: | |
self.__db.add_clean(f) | |
try: | |
self.__db.commit() | |
except OSError as e: | |
raise TaskError( | |
"error committing control database {}: {}".format( | |
getattr(e, "filename", "<unknown path>"), e | |
) | |
) from e | |
def __make_summary(self, deps, run_result): | |
"""Construct a new task summary.""" | |
return { | |
"deps": deps, | |
"output_files": {f: hash_cache.get(f) for f in run_result.output_filenames}, | |
"extra": run_result.extra, | |
} | |
def _execute(self): | |
"""Abstract: Execute this task. | |
Subclasses should implement this method to execute this task. | |
This method must return a RunResult giving the inputs that | |
were used by the task and the outputs it produced. | |
""" | |
raise NotImplementedError("Task._execute is abstract") | |
def _get_result_extra(self): | |
"""Return the 'extra' result from the previous run, or None.""" | |
summary = self.__db.get_summary(self.__task_id) | |
if summary is None: | |
return None | |
return summary["extra"] | |
def report(self): | |
"""Report the task's results to stdout and return exit status. | |
This may be called when the task has never executed. | |
Subclasses should override this. The default implementation | |
reports nothing and returns 0. | |
""" | |
return 0 | |
# Standard input functions | |
def _input_env(self, var): | |
return os.environ.get(var) | |
def _input_file(self, path): | |
return hash_cache.get(path) | |
def _input_unstable(self): | |
"""Mark this run as unstable, regardless of other inputs.""" | |
return None | |
def _input_unknown_input(self): | |
"""An unknown input that may change after latexrun exits. | |
This conservatively marks some unknown input that definitely | |
won't change while latexrun is running, but may change before | |
the user next runs latexrun. This allows the task to | |
stabilize during this invocation, but will cause the task to | |
re-run on the next invocation. | |
""" | |
return start_time | |
class RunResult(collections.namedtuple("RunResult", "output_filenames extra")): | |
"""The result of a single task execution. | |
This captures all files written by the task, and task-specific | |
results that need to be persisted between runs (for example, to | |
enable reporting of a task's results). | |
""" | |
pass | |
################################################################## | |
# LaTeX task | |
# | |
def normalize_input_path(path): | |
# Resolve the directory of the input path, but leave the file | |
# component alone because it affects TeX's behavior. | |
head, tail = os.path.split(path) | |
npath = os.path.join(os.path.realpath(head), tail) | |
return os.path.relpath(path) | |
class LaTeX(Task): | |
def __init__(self, db, tex_filename, cmd, cmd_args, obj_dir, nowarns): | |
super().__init__(db, "latex::" + normalize_input_path(tex_filename)) | |
self.__tex_filename = tex_filename | |
self.__cmd = cmd | |
self.__cmd_args = cmd_args | |
self.__obj_dir = obj_dir | |
self.__nowarns = nowarns | |
self.__pass = 0 | |
def _input_args(self): | |
# If filename starts with a character the tex command-line | |
# treats specially, then tweak it so it doesn't. | |
filename = self.__tex_filename | |
if filename.startswith(("-", "&", "\\")): | |
filename = "./" + filename | |
# XXX Put these at the beginning in case the provided | |
# arguments are malformed. Might want to do a best-effort | |
# check for incompatible user-provided arguments (note: | |
# arguments can be given with one or two dashes and those with | |
# values can use an equals or a space). | |
return ( | |
[self.__cmd] | |
+ self.__cmd_args | |
+ [ | |
"-interaction", | |
"nonstopmode", | |
"-recorder", | |
"-output-directory", | |
self.__obj_dir, | |
"-synctex", | |
"1", | |
"-shell-escape", | |
filename, | |
] | |
) | |
def _execute(self): | |
# Run latex | |
self.__pass += 1 | |
args = self._input("args") | |
debug("running {}", args) | |
try: | |
os.makedirs(self.__obj_dir, exist_ok=True) | |
except OSError as e: | |
raise TaskError("failed to create %s: " % self.__obj_dir + str(e)) from e | |
try: | |
verbose_cmd(args) | |
start_time = time.time() | |
p = subprocess.Popen( | |
args, | |
stdin=subprocess.DEVNULL, | |
stdout=subprocess.PIPE, | |
stderr=subprocess.STDOUT, | |
) | |
stdout, has_errors, missing_includes = self.__feed_terminal(p.stdout) | |
status = p.wait() | |
duration = time.time() - start_time | |
print(f"took {duration:.2f} seconds", file=sys.stderr) | |
except OSError as e: | |
raise TaskError("failed to execute latex task: " + str(e)) from e | |
# Register environment variable inputs | |
for env_var in [ | |
"TEXMFOUTPUT", | |
"TEXINPUTS", | |
"TEXFORMATS", | |
"TEXPOOL", | |
"TFMFONTS", | |
"PATH", | |
]: | |
self._input("env", env_var) | |
jobname, outname = self.__parse_jobname(stdout) | |
inputs, outputs = self.__parse_recorder(jobname) | |
# LaTeX overwrites its own inputs. Mark its output files as | |
# clobbered before we hash its input files. | |
for path in outputs: | |
# In some abort cases (e.g., >=100 errors), LaTeX claims | |
# output files that don't actually exist. | |
if os.path.exists(path): | |
hash_cache.clobber(path) | |
# Depend on input files. Task.run pre-hashed outputs from the | |
# previous run, so if this isn't the first run and as long as | |
# the set of outputs didn't change, we'll be able to get the | |
# input hashes, even if they were clobbered. | |
for path in inputs: | |
self._input("file", path) | |
if missing_includes: | |
# Missing \includes are tricky. Ideally we'd depend on | |
# the absence of some file, but in fact we'd have to | |
# depend on the failure of a whole kpathsea lookup. | |
# Rather than try to be clever, just mark this as an | |
# unknown input so we'll run at least once on the next | |
# invocation. | |
self._input("unknown_input") | |
if not self.__create_outdirs(stdout) and has_errors: | |
# LaTeX reported unrecoverable errors (other than output | |
# directory errors, which we just fixed). We could | |
# continue to stabilize the document, which may change | |
# some of the other problems reported (but not the | |
# unrecoverable errors), or we can just abort now and get | |
# back to the user quickly with the major errors. We opt | |
# for the latter. | |
global terminate_task_loop | |
terminate_task_loop = True | |
# This error could depend on something we failed to track. | |
# It would be really confusing if we continued to report | |
# the error after the user fixed it, so be conservative | |
# and force a re-run next time. | |
self._input("unknown_input") | |
return RunResult( | |
outputs, {"jobname": jobname, "outname": outname, "status": status} | |
) | |
def __feed_terminal(self, stdout): | |
prefix = "latex" | |
if self.__pass > 1: | |
prefix += " ({})".format(self.__pass) | |
with Progress(prefix) as progress: | |
buf = [] | |
filt = LaTeXFilter() | |
while True: | |
# Use os.read to read only what's available on the pipe, | |
# without waiting to fill a buffer | |
data = os.read(stdout.fileno(), 4096) | |
if not data: | |
break | |
# See "A note about encoding" above | |
data = data.decode("ascii", errors="surrogateescape") | |
buf.append(data) | |
filt.feed(data) | |
file_stack = filt.get_file_stack() | |
if file_stack: | |
tos = file_stack[-1] | |
if tos.startswith("./"): | |
tos = tos[2:] | |
progress.update(">" * len(file_stack) + " " + tos) | |
else: | |
progress.update("") | |
# Were there unrecoverable errors? | |
has_errors = any(msg.typ == "error" for msg in filt.get_messages()) | |
return "".join(buf), has_errors, filt.has_missing_includes() | |
def __parse_jobname(self, stdout): | |
"""Extract the job name and output name from latex's output. | |
We get these from latex because they depend on complicated | |
file name parsing rules, are affected by arguments like | |
-output-directory, and may be just "texput" if things fail | |
really early. The output name may be None if there were no | |
pages of output. | |
""" | |
jobname = outname = None | |
for m in re.finditer( | |
r'^Transcript written on "?(.*)\.log"?\.$', stdout, re.MULTILINE | re.DOTALL | |
): | |
jobname = m.group(1).replace("\n", "") | |
if jobname is None: | |
print(stdout, file=sys.stderr) | |
raise TaskError("failed to extract job name from latex log") | |
for m in re.finditer( | |
r'^Output written on "?(.*\.[^ ."]+)"? \([0-9]+ page', | |
stdout, | |
re.MULTILINE | re.DOTALL, | |
): | |
outname = m.group(1).replace("\n", "") | |
if outname is None and not re.search( | |
r"^No pages of output\.$|^! Emergency stop\.$" | |
r"|^! ==> Fatal error occurred, no output PDF file produced!$", | |
stdout, | |
re.MULTILINE, | |
): | |
print(stdout, file=sys.stderr) | |
raise TaskError("failed to extract output name from latex log") | |
# LuaTeX (0.76.0) doesn't include the output directory in the | |
# logged transcript or output file name. | |
if os.path.basename(jobname) == jobname and os.path.exists( | |
os.path.join(self.__obj_dir, jobname + ".log") | |
): | |
jobname = os.path.join(self.__obj_dir, jobname) | |
if outname is not None: | |
outname = os.path.join(self.__obj_dir, outname) | |
return jobname, outname | |
def __parse_recorder(self, jobname): | |
"""Parse file recorder output.""" | |
# XXX If latex fails because a file isn't found, that doesn't | |
# go into the .fls file, but creating that file will affect | |
# the computation, so it should be included as an input. | |
# Though it's generally true that files can be added earlier | |
# in search paths and will affect the output without us knowing. | |
# | |
# XXX This is a serious problem for bibtex, since the first | |
# run won't depend on the .bbl file! But maybe the .aux file | |
# will always cause a re-run, at which point the .bbl will | |
# exist? | |
filename = jobname + ".fls" | |
try: | |
recorder = open(filename) | |
except OSError as e: | |
raise TaskError("failed to open file recorder output: " + str(e)) from e | |
pwd, inputs, outputs = "", set(), set() | |
for linenum, line in enumerate(recorder): | |
parts = line.rstrip("\n").split(" ", 1) | |
if parts[0] == "PWD": | |
pwd = parts[1] | |
elif parts[0] in ("INPUT", "OUTPUT"): | |
if parts[1].startswith("/"): | |
path = parts[1] | |
else: | |
# Try to make "nice" paths, especially for clean | |
path = os.path.relpath(os.path.join(pwd, parts[1])) | |
if parts[0] == "INPUT": | |
inputs.add(path) | |
else: | |
outputs.add(path) | |
else: | |
raise TaskError( | |
"syntax error on line {} of {}".format(linenum, filename) | |
) | |
# Ironically, latex omits the .fls file itself | |
outputs.add(filename) | |
return inputs, outputs | |
def __create_outdirs(self, stdout): | |
# In some cases, such as \include'ing a file from a | |
# subdirectory, TeX will attempt to create files in | |
# subdirectories of the output directory that don't exist. | |
# Detect this, create the output directory, and re-run. | |
m = re.search("^! I can't write on file `(.*)'\\.$", stdout, re.M) | |
if m and m.group(1).find("/") > 0 and "../" not in m.group(1): | |
debug("considering creating output sub-directory for {}".format(m.group(1))) | |
subdir = os.path.dirname(m.group(1)) | |
newdir = os.path.join(self.__obj_dir, subdir) | |
if os.path.isdir(subdir) and not os.path.isdir(newdir): | |
debug("creating output subdirectory {}".format(newdir)) | |
try: | |
mkdir_p(newdir) | |
except OSError as e: | |
raise TaskError( | |
"failed to create output subdirectory: " + str(e) | |
) from e | |
self._input("unstable") | |
return True | |
def report(self): | |
extra = self._get_result_extra() | |
if extra is None: | |
return 0 | |
# Parse the log | |
logfile = open(extra["jobname"] + ".log", "rt", errors="surrogateescape") | |
for msg in self.__clean_messages( | |
LaTeXFilter(self.__nowarns).feed(logfile.read(), True).get_messages() | |
): | |
msg.emit() | |
# Return LaTeX's exit status | |
return extra["status"] | |
def __clean_messages(self, msgs): | |
"""Make some standard log messages more user-friendly.""" | |
have_undefined_reference = False | |
for msg in msgs: | |
if msg.msg == "==> Fatal error occurred, no output PDF file produced!": | |
msg = msg._replace( | |
typ="info", msg="Fatal error (no output file produced)" | |
) | |
if msg.msg.startswith("[LaTeX] "): | |
# Strip unnecessary package name | |
msg = msg._replace(msg=msg.msg.split(" ", 1)[1]) | |
if re.match(r"Reference .* undefined", msg.msg): | |
have_undefined_reference = True | |
if have_undefined_reference and re.match( | |
r"There were undefined references", msg.msg | |
): | |
# LaTeX prints this at the end so the user knows it's | |
# worthwhile looking back at the log. Since latexrun | |
# makes the earlier messages obvious, this is | |
# redundant. | |
continue | |
yield msg | |
def get_tex_filename(self): | |
return self.__tex_filename | |
def get_jobname(self): | |
extra = self._get_result_extra() | |
if extra is None: | |
return None | |
return extra["jobname"] | |
def get_outname(self): | |
extra = self._get_result_extra() | |
if extra is None: | |
return None | |
return extra["outname"] | |
def get_status(self): | |
extra = self._get_result_extra() | |
if extra is None: | |
return None | |
return extra["status"] | |
class LaTeXCommit(Task): | |
def __init__(self, db, latex_task, output_path): | |
super().__init__( | |
db, "latex_commit::" + normalize_input_path(latex_task.get_tex_filename()) | |
) | |
self.__latex_task = latex_task | |
self.__output_path = output_path | |
self.status = "There were errors" | |
def _input_latex(self): | |
return self.__latex_task.get_status(), self.__latex_task.get_outname() | |
def _execute(self): | |
self.status = "There were errors" | |
# If latex succeeded with output, atomically commit the output | |
status, outname = self._input("latex") | |
if status != 0 or outname is None: | |
debug("not committing (status {}, outname {})", status, outname) | |
if outname is None: | |
self.status = "No pages of output" | |
return RunResult([], None) | |
commit = self.__output_path or os.path.basename(outname) | |
if os.path.abspath(commit) == os.path.abspath(outname): | |
debug("skipping commit (outname is commit name)") | |
self.status = None | |
return RunResult([], None) | |
try: | |
if os.path.exists(commit) and filecmp.cmp(outname, commit): | |
debug("skipping commit ({} and {} are identical)", outname, commit) | |
# To avoid confusion, touch the output file | |
open(outname, "r+b").close() | |
else: | |
debug("commiting {} to {}", outname, commit) | |
shutil.copy(outname, outname + "~") | |
os.rename(outname + "~", commit) | |
except OSError as e: | |
raise TaskError("error committing latex output: {}".format(e)) from e | |
self._input("file", outname) | |
self.status = None | |
return RunResult([commit], None) | |
class LaTeXFilter: | |
TRACE = False # Set to enable detailed parse tracing | |
def __init__(self, nowarns=[]): | |
self.__data = "" | |
self.__restart_pos = 0 | |
self.__restart_file_stack = [] | |
self.__restart_messages_len = 0 | |
self.__messages = [] | |
self.__first_file = None | |
self.__fatal_error = False | |
self.__missing_includes = False | |
self.__pageno = 1 | |
self.__restart_pageno = 1 | |
self.__suppress = {cls: 0 for cls in nowarns} | |
def feed(self, data, eof=False): | |
"""Feed LaTeX log data to the parser. | |
The log data can be from LaTeX's standard output, or from the | |
log file. If there will be no more data, set eof to True. | |
""" | |
self.__data += data | |
self.__data_complete = eof | |
# Reset to last known-good restart point | |
self.__pos = self.__restart_pos | |
self.__file_stack = self.__restart_file_stack.copy() | |
self.__messages = self.__messages[: self.__restart_messages_len] | |
self.__lstart = self.__lend = -1 | |
self.__pageno = self.__restart_pageno | |
# Parse forward | |
while self.__pos < len(self.__data): | |
self.__noise() | |
# Handle suppressed warnings | |
if eof: | |
msgs = [ | |
"%d %s warning%s" % (count, cls, "s" if count > 1 else "") | |
for cls, count in self.__suppress.items() | |
if count | |
] | |
if msgs: | |
self.__message( | |
"info", | |
None, | |
"%s not shown (use -Wall to show them)" % ", ".join(msgs), | |
filename=self.__first_file, | |
) | |
if eof and len(self.__file_stack) and not self.__fatal_error: | |
# Fatal errors generally cause TeX to "succumb" without | |
# closing the file stack, so don't complain in that case. | |
self.__message( | |
"warning", None, "unbalanced `(' in log; file names may be wrong" | |
) | |
return self | |
def get_messages(self): | |
"""Return a list of warning and error Messages.""" | |
return self.__messages | |
def get_file_stack(self): | |
"""Return the file stack for the data that has been parsed. | |
This results a list from outermost file to innermost file. | |
The list may be empty. | |
""" | |
return self.__file_stack | |
def has_missing_includes(self): | |
"""Return True if the log reported missing \\include files.""" | |
return self.__missing_includes | |
def __save_restart_point(self): | |
"""Save the current state as a known-good restart point. | |
On the next call to feed, the parser will reset to this point. | |
""" | |
self.__restart_pos = self.__pos | |
self.__restart_file_stack = self.__file_stack.copy() | |
self.__restart_messages_len = len(self.__messages) | |
self.__restart_pageno = self.__pageno | |
def __message(self, typ, lineno, msg, cls=None, filename=None): | |
if cls is not None and cls in self.__suppress: | |
self.__suppress[cls] += 1 | |
return | |
filename = filename or ( | |
self.__file_stack[-1] if self.__file_stack else self.__first_file | |
) | |
self.__messages.append(Message(typ, filename, lineno, msg)) | |
def __ensure_line(self): | |
"""Update lstart and lend.""" | |
if self.__lstart <= self.__pos < self.__lend: | |
return | |
self.__lstart = self.__data.rfind("\n", 0, self.__pos) + 1 | |
self.__lend = self.__data.find("\n", self.__pos) + 1 | |
if self.__lend == 0: | |
self.__lend = len(self.__data) | |
@property | |
def __col(self): | |
"""The 0-based column number of __pos.""" | |
self.__ensure_line() | |
return self.__pos - self.__lstart | |
@property | |
def __avail(self): | |
return self.__pos < len(self.__data) | |
def __lookingat(self, needle): | |
return self.__data.startswith(needle, self.__pos) | |
def __lookingatre(self, regexp, flags=0): | |
return re.compile(regexp, flags=flags).match(self.__data, self.__pos) | |
def __skip_line(self): | |
self.__ensure_line() | |
self.__pos = self.__lend | |
def __consume_line(self, unwrap=False): | |
self.__ensure_line() | |
data = self.__data[self.__pos : self.__lend] | |
self.__pos = self.__lend | |
if unwrap: | |
# TeX helpfully wraps all terminal output at 79 columns | |
# (max_print_line). If requested, unwrap it. There's | |
# simply no way to do this perfectly, since there could be | |
# a line that happens to be 79 columns. | |
# | |
# We check for >=80 because a bug in LuaTeX causes it to | |
# wrap at 80 columns instead of 79 (LuaTeX #900). | |
while self.__lend - self.__lstart >= 80: | |
if self.TRACE: | |
print("<{}> wrapping".format(self.__pos)) | |
self.__ensure_line() | |
data = data[:-1] + self.__data[self.__pos : self.__lend] | |
self.__pos = self.__lend | |
return data | |
# Parser productions | |
def __noise(self): | |
# Most of TeX's output is line noise that combines error | |
# messages, warnings, file names, user errors and warnings, | |
# and echos of token lists and other input. This attempts to | |
# tease these apart, paying particular attention to all of the | |
# places where TeX echos input so that parens in the input do | |
# not confuse the file name scanner. There are three | |
# functions in TeX that echo input: show_token_list (used by | |
# runaway and show_context, which is used by print_err), | |
# short_display (used by overfull/etc h/vbox), and show_print | |
# (used in issue_message and the same places as | |
# show_token_list). | |
lookingat, lookingatre = self.__lookingat, self.__lookingatre | |
if self.__col == 0: | |
# The following messages are always preceded by a newline | |
if lookingat("! "): | |
return self.__errmessage() | |
if lookingat("!pdfTeX error: "): | |
return self.__pdftex_fail() | |
if lookingat("Runaway "): | |
return self.__runaway() | |
if lookingatre(r"(Overfull|Underfull|Loose|Tight) \\[hv]box \("): | |
return self.__bad_box() | |
if lookingatre(r"(Package |Class |LaTeX |pdfTeX )?(\w+ )?warning: ", re.I): | |
return self.__generic_warning() | |
if lookingatre("No file .*\\.tex\\.$", re.M): | |
# This happens with \includes of missing files. For | |
# whatever reason, LaTeX doesn't consider this even | |
# worth a warning, but I do! | |
self.__message( | |
"warning", | |
None, | |
self.__simplify_message(self.__consume_line(unwrap=True).strip()), | |
) | |
self.__missing_includes = True | |
return | |
# Other things that are common and irrelevant | |
if lookingatre(r"(Package|Class|LaTeX) (\w+ )?info: ", re.I): | |
return self.__generic_info() | |
if lookingatre(r"(Document Class|File|Package): "): | |
# Output from "\ProvidesX" | |
return self.__consume_line(unwrap=True) | |
if lookingatre(r"\\\w+=\\[a-z]+\d+\n"): | |
# Output from "\new{count,dimen,skip,...}" | |
return self.__consume_line(unwrap=True) | |
# print(self.__data[self.__lstart:self.__lend].rstrip()) | |
# self.__pos = self.__lend | |
# return | |
# Now that we've substantially reduced the spew and hopefully | |
# eliminated all input echoing, we're left with the file name | |
# stack, page outs, and random other messages from both TeX | |
# and various packages. We'll assume at this point that all | |
# parentheses belong to the file name stack or, if they're in | |
# random other messages, they're at least balanced and nothing | |
# interesting happens between them. For page outs, ship_out | |
# prints a space if not at the beginning of a line, then a | |
# "[", then the page number being shipped out (this is | |
# usually, but not always, followed by "]"). | |
m = re.compile(r"[(){}\n]|(?<=[\n ])\[\d+", re.M).search( | |
self.__data, self.__pos | |
) | |
if m is None: | |
self.__pos = len(self.__data) | |
return | |
self.__pos = m.start() + 1 | |
ch = self.__data[m.start()] | |
if ch == "\n": | |
# Save this as a known-good restart point for incremental | |
# parsing, since we definitely didn't match any of the | |
# known message types above. | |
self.__save_restart_point() | |
elif ch == "[": | |
# This is printed at the end of a page, so we're beginning | |
# page n+1. | |
self.__pageno = int(self.__lookingatre(r"\d+").group(0)) + 1 | |
elif ( | |
self.__data.startswith("`", m.start() - 1) | |
or self.__data.startswith("`\\", m.start() - 2) | |
) and self.__data.startswith("'", m.start() + 1): | |
# (, ), {, and } sometimes appear in TeX's error | |
# descriptions, but they're always in `'s (and sometimes | |
# backslashed) | |
return | |
elif ch == "(": | |
# XXX Check that the stack doesn't drop to empty and then re-grow | |
first = self.__first_file is None and self.__col == 1 | |
filename = self.__filename() | |
self.__file_stack.append(filename) | |
if first: | |
self.__first_file = filename | |
if self.TRACE: | |
print( | |
"<{}>{}enter {}".format( | |
m.start(), " " * len(self.__file_stack), filename | |
) | |
) | |
elif ch == ")": | |
if len(self.__file_stack): | |
if self.TRACE: | |
print( | |
"<{}>{}exit {}".format( | |
m.start(), | |
" " * len(self.__file_stack), | |
self.__file_stack[-1], | |
) | |
) | |
self.__file_stack.pop() | |
else: | |
self.__message( | |
"warning", None, "extra `)' in log; file names may be wrong " | |
) | |
elif ch == "{": | |
# TeX uses this for various things we want to ignore, like | |
# file names and print_mark. Consume up to the '}' | |
epos = self.__data.find("}", self.__pos) | |
if epos != -1: | |
self.__pos = epos + 1 | |
else: | |
self.__message( | |
"warning", None, "unbalanced `{' in log; file names may be wrong" | |
) | |
elif ch == "}": | |
self.__message("warning", None, "extra `}' in log; file names may be wrong") | |
def __filename(self): | |
initcol = self.__col | |
first = True | |
name = "" | |
# File names may wrap, but if they do, TeX will always print a | |
# newline before the open paren | |
while first or (initcol == 1 and self.__lookingat("\n") and self.__col >= 79): | |
if not first: | |
self.__pos += 1 | |
m = self.__lookingatre(r"[^(){} \n]*") | |
name += m.group() | |
self.__pos = m.end() | |
first = False | |
return name | |
def __simplify_message(self, msg): | |
msg = re.sub( | |
r"^(?:Package |Class |LaTeX |pdfTeX )?([^ ]+) (?:Error|Warning): ", | |
r"[\1] ", | |
msg, | |
flags=re.I, | |
) | |
msg = re.sub(r"\.$", "", msg) | |
msg = re.sub(r"has occurred (while \\output is active)", r"\1", msg) | |
return msg | |
def __errmessage(self): | |
# Procedure print_err (including \errmessage, itself used by | |
# LaTeX's \GenericError and all of its callers), as well as | |
# fatal_error. Prints "\n! " followed by error text | |
# ("Emergency stop" in the case of fatal_error). print_err is | |
# always followed by a call to error, which prints a period, | |
# and a newline... | |
msg = self.__consume_line(unwrap=True)[1:].strip() | |
is_fatal_error = msg == "Emergency stop." | |
msg = self.__simplify_message(msg) | |
# ... and then calls show_context, which prints the input | |
# stack as pairs of lines giving the context. These context | |
# lines are truncated so they never wrap. Each pair of lines | |
# will start with either "<something> " if the context is a | |
# token list, "<*> " for terminal input (or command line), | |
# "<read ...>" for stream reads, something like "\macroname | |
# #1->" for macros (though everything after \macroname is | |
# subject to being elided as "..."), or "l.[0-9]+ " if it's a | |
# file. This is followed by the errant input with a line | |
# break where the error occurred. | |
lineno = None | |
found_context = False | |
stack = [] | |
while self.__avail: | |
m1 = self.__lookingatre(r"<([a-z ]+|\*|read [^ >]*)> |\\.*(->|...)") | |
m2 = self.__lookingatre(r"l\.[0-9]+ ") | |
if m1: | |
found_context = True | |
pre = self.__consume_line().rstrip("\n") | |
stack.append(pre) | |
elif m2: | |
found_context = True | |
pre = self.__consume_line().rstrip("\n") | |
info, rest = pre.split(" ", 1) | |
lineno = int(info[2:]) | |
stack.append(rest) | |
elif found_context: | |
# Done with context | |
break | |
if found_context: | |
# Consume the second context line | |
post = self.__consume_line().rstrip("\n") | |
# Clean up goofy trailing ^^M TeX sometimes includes | |
post = re.sub(r"\^\^M$", "", post) | |
if post[: len(pre)].isspace() and not post.isspace(): | |
stack.append(len(stack[-1])) | |
stack[-2] += post[len(pre) :] | |
else: | |
# If we haven't found the context, skip the line. | |
self.__skip_line() | |
stack_msg = "" | |
for i, trace in enumerate(stack): | |
stack_msg += ( | |
"\n " + (" " * trace) + "^" | |
if isinstance(trace, int) | |
else "\n at " + trace.rstrip() | |
if i == 0 | |
else "\n from " + trace.rstrip() | |
) | |
if is_fatal_error: | |
# fatal_error always prints one additional line of message | |
info = self.__consume_line().strip() | |
if info.startswith("*** "): | |
info = info[4:] | |
msg += ": " + info.lstrip("(").rstrip(")") | |
self.__message("error", lineno, msg + stack_msg) | |
self.__fatal_error = True | |
def __pdftex_fail(self): | |
# Procedure pdftex_fail. Prints "\n!pdfTeX error: ", the | |
# message, and a newline. Unlike print_err, there's never | |
# context. | |
msg = self.__consume_line(unwrap=True)[1:].strip() | |
msg = self.__simplify_message(msg) | |
self.__message("error", None, msg) | |
def __runaway(self): | |
# Procedure runaway. Prints "\nRunaway ...\n" possibly | |
# followed by token list (user text). Always followed by a | |
# call to print_err, so skip lines until we see the print_err. | |
self.__skip_line() # Skip "Runaway ...\n" | |
if not self.__lookingat("! ") and self.__avail: | |
# Skip token list, which is limited to one line | |
self.__skip_line() | |
def __bad_box(self): | |
# Function hpack and vpack. hpack prints a warning, a | |
# newline, then a short_display of the offending text. | |
# Unfortunately, there's nothing indicating the end of the | |
# offending text, but it should be on one (possible wrapped) | |
# line. vpack prints a warning and then, *unless output is | |
# active*, a newline. The missing newline is probably a bug, | |
# but it sure makes our lives harder. | |
origpos = self.__pos | |
msg = self.__consume_line() | |
m = re.search( | |
r" in (?:paragraph|alignment) at lines ([0-9]+)--([0-9]+)", msg | |
) or re.search(r" detected at line ([0-9]+)", msg) | |
if m: | |
# Sometimes TeX prints crazy line ranges like "at lines | |
# 8500--250". The lower number seems roughly sane, so use | |
# that. I'm not sure what causes this, but it may be | |
# related to shipout routines messing up line registers. | |
lineno = min(int(m.group(1)), int(m.groups()[-1])) | |
msg = msg[: m.start()] | |
else: | |
m = re.search(r" while \\output is active", msg) | |
if m: | |
lineno = None | |
msg = msg[: m.end()] | |
else: | |
self.__message("warning", None, "malformed bad box message in log") | |
return | |
# Back up to the end of the known message text | |
self.__pos = origpos + m.end() | |
if self.__lookingat("\n"): | |
# We have a newline, so consume it and look for the | |
# offending text. | |
self.__pos += 1 | |
# If there is offending text, it will start with a font | |
# name, which will start with a \. | |
if "hbox" in msg and self.__lookingat("\\"): | |
self.__consume_line(unwrap=True) | |
msg = self.__simplify_message(msg) + " (page {})".format(self.__pageno) | |
cls = msg.split(None, 1)[0].lower() | |
self.__message("warning", lineno, msg, cls=cls) | |
def __generic_warning(self): | |
# Warnings produced by LaTeX's \GenericWarning (which is | |
# called by \{Package,Class}Warning and \@latex@warning), | |
# warnings produced by pdftex_warn, and other random warnings. | |
msg, cls = self.__generic_info() | |
# Most warnings include an input line emitted by \on@line | |
m = re.search(" on input line ([0-9]+)", msg) | |
if m: | |
lineno = int(m.group(1)) | |
msg = msg[: m.start()] | |
else: | |
lineno = None | |
msg = self.__simplify_message(msg) | |
self.__message("warning", lineno, msg, cls=cls) | |
def __generic_info(self): | |
# Messages produced by LaTeX's \Generic{Error,Warning,Info} | |
# and things that look like them | |
msg = self.__consume_line(unwrap=True).strip() | |
# Package and class messages are continued with lines | |
# containing '(package name) ' | |
pkg_name = msg.split(" ", 2)[1] | |
prefix = "(" + pkg_name + ") " | |
while self.__lookingat(prefix): | |
# Collect extra lines. It's important that we keep these | |
# because they may contain context information like line | |
# numbers. | |
extra = self.__consume_line(unwrap=True) | |
msg += " " + extra[len(prefix) :].strip() | |
return msg, pkg_name.lower() | |
################################################################## | |
# BibTeX task | |
# | |
class BibTeX(Task): | |
def __init__(self, db, latex_task, cmd, cmd_args, nowarns, obj_dir): | |
super().__init__( | |
db, "bibtex::" + normalize_input_path(latex_task.get_tex_filename()) | |
) | |
self.__latex_task = latex_task | |
self.__cmd = cmd | |
self.__cmd_args = cmd_args | |
self.__obj_dir = obj_dir | |
def stable(self): | |
# If bibtex doesn't have its inputs, then it's stable because | |
# it has no effect on system state. | |
jobname = self.__latex_task.get_jobname() | |
if jobname is None: | |
# We don't know where the .aux file is until latex has run | |
return True | |
if not os.path.exists(jobname + ".aux"): | |
# Input isn't ready, so bibtex will simply fail without | |
# affecting system state. Hence, this task is trivially | |
# stable. | |
return True | |
if not self.__find_bib_cmds(os.path.dirname(jobname), jobname + ".aux"): | |
# The tex file doesn't refer to any bibliographic data, so | |
# don't run bibtex. | |
return True | |
return super().stable() | |
def __find_bib_cmds(self, basedir, auxname, stack=()): | |
debug("scanning for bib commands in {}".format(auxname)) | |
if auxname in stack: | |
raise TaskError(".aux file loop") | |
stack = stack + (auxname,) | |
try: | |
aux_data = open(auxname, errors="surrogateescape").read() | |
except FileNotFoundError: | |
# The aux file may not exist if latex aborted | |
return False | |
if re.search(r"^\\bibstyle\{", aux_data, flags=re.M) or re.search( | |
r"^\\bibdata\{", aux_data, flags=re.M | |
): | |
return True | |
if re.search(r"^\\abx@aux@cite\{", aux_data, flags=re.M): | |
# biber citation | |
return True | |
# Recurse into included aux files (see aux_input_command), in | |
# case \bibliography appears in an \included file. | |
for m in re.finditer(r"^\\@input\{([^}]*)\}", aux_data, flags=re.M): | |
if self.__find_bib_cmds(basedir, os.path.join(basedir, m.group(1)), stack): | |
return True | |
return False | |
def _input_args(self): | |
if self.__is_biber(): | |
aux_name = os.path.basename(self.__latex_task.get_jobname()) | |
else: | |
aux_name = os.path.basename(self.__latex_task.get_jobname()) + ".aux" | |
return [self.__cmd] + self.__cmd_args + [aux_name] | |
def _input_cwd(self): | |
return os.path.dirname(self.__latex_task.get_jobname()) | |
def _input_auxfile(self, auxname): | |
# We don't consider the .aux files regular inputs. | |
# Instead, we extract just the bit that BibTeX cares about | |
# and depend on that. See get_aux_command_and_process in | |
# bibtex.web. | |
debug("hashing filtered aux file {}", auxname) | |
try: | |
with open(auxname, "rb") as aux: | |
h = hashlib.sha256() | |
for line in aux: | |
if line.startswith( | |
( | |
b"\\citation{", | |
b"\\bibdata{", | |
b"\\bibstyle{", | |
b"\\@input{", | |
b"\\abx@aux@cite{", | |
) | |
): | |
h.update(line) | |
return h.hexdigest() | |
except FileNotFoundError: | |
debug("{} does not exist", auxname) | |
return None | |
def __path_join(self, first, rest): | |
if rest is None: | |
# Append ':' to keep the default search path | |
return first + ":" | |
return first + ":" + rest | |
def __is_biber(self): | |
return "biber" in self.__cmd | |
def _execute(self): | |
# This gets complicated when \include is involved. \include | |
# switches to a different aux file and records its path in the | |
# main aux file. However, BibTeX does not consider this path | |
# to be relative to the location of the main aux file, so we | |
# have to run BibTeX *in the output directory* for it to | |
# follow these includes (there's no way to tell BibTeX other | |
# locations to search). Unfortunately, this means BibTeX will | |
# no longer be able to find local bib or bst files, but so we | |
# tell it where to look by setting BIBINPUTS and BSTINPUTS | |
# (luckily we can control this search). We have to pass this | |
# same environment down to Kpathsea when we resolve the paths | |
# in BibTeX's log. | |
args, cwd = self._input("args"), self._input("cwd") | |
debug("running {} in {}", args, cwd) | |
env = os.environ.copy() | |
env["BIBINPUTS"] = self.__path_join(os.getcwd(), env.get("BIBINPUTS")) | |
env["BSTINPUTS"] = self.__path_join(os.getcwd(), env.get("BSTINPUTS")) | |
try: | |
verbose_cmd(args, cwd, env) | |
p = subprocess.Popen( | |
args, | |
cwd=cwd, | |
env=env, | |
stdin=subprocess.DEVNULL, | |
stdout=subprocess.PIPE, | |
stderr=subprocess.STDOUT, | |
) | |
stdout = self.__feed_terminal(p.stdout) | |
status = p.wait() | |
except OSError as e: | |
raise TaskError("failed to execute bibtex task: " + str(e)) from e | |
inputs, auxnames, outbase = self.__parse_inputs(stdout, cwd, env) | |
if not inputs and not auxnames: | |
# BibTeX failed catastrophically. | |
print(stdout, file=sys.stderr) | |
raise TaskError("failed to execute bibtex task") | |
# Register environment variable inputs | |
for env_var in ["TEXMFOUTPUT", "BSTINPUTS", "BIBINPUTS", "PATH"]: | |
self._input("env", env_var) | |
# Register file inputs | |
for path in auxnames: | |
self._input("auxfile", path) | |
for path in inputs: | |
self._input("file", path) | |
if self.__is_biber(): | |
outbase = os.path.join(cwd, outbase) | |
outputs = [outbase + ".bbl", outbase + ".blg"] | |
return RunResult( | |
outputs, {"outbase": outbase, "status": status, "inputs": inputs} | |
) | |
def __feed_terminal(self, stdout): | |
with Progress("bibtex") as progress: | |
buf, linebuf = [], "" | |
while True: | |
data = os.read(stdout.fileno(), 4096) | |
if not data: | |
break | |
# See "A note about encoding" above | |
data = data.decode("ascii", errors="surrogateescape") | |
buf.append(data) | |
linebuf += data | |
while "\n" in linebuf: | |
line, _, linebuf = linebuf.partition("\n") | |
if line.startswith("Database file"): | |
progress.update(line.split(": ", 1)[1]) | |
return "".join(buf) | |
def __parse_inputs(self, log, cwd, env): | |
# BibTeX conveniently logs every file that it opens, and its | |
# log is actually sensible (see calls to a_open_in in | |
# bibtex.web.) The only trick is that these file names are | |
# pre-kpathsea lookup and may be relative to the directory we | |
# ran BibTeX in. | |
# | |
# Because BibTeX actually depends on very little in the .aux | |
# file (and it's likely other things will change in the .aux | |
# file), we don't count the whole .aux file as an input, but | |
# instead depend only on the lines that matter to BibTeX. | |
kpathsea = Kpathsea("bibtex") | |
inputs = [] | |
auxnames = [] | |
outbase = None | |
for line in log.splitlines(): | |
m = re.match( | |
"(?:The top-level auxiliary file:|A level-[0-9]+ auxiliary file:) (.*)", | |
line, | |
) | |
if m: | |
auxnames.append(os.path.join(cwd, m.group(1))) | |
continue | |
m = re.match("(?:(The style file:)|(Database file #[0-9]+:)) (.*)", line) | |
if m: | |
filename = m.group(3) | |
if m.group(1): | |
filename = kpathsea.find_file(filename, "bst", cwd, env) | |
elif m.group(2): | |
filename = kpathsea.find_file(filename, "bib", cwd, env) | |
# If this path is relative to the source directory, | |
# clean it up for error reporting and portability of | |
# the dependency DB | |
if filename.startswith("/"): | |
relname = os.path.relpath(filename) | |
if "../" not in relname: | |
filename = relname | |
inputs.append(filename) | |
# biber output | |
m = re.search("Found BibTeX data source '(.*?)'", line) | |
if m: | |
filename = m.group(1) | |
inputs.append(filename) | |
m = re.search("Logfile is '(.*?)'", line) | |
if m: | |
outbase = m.group(1)[:-4] | |
if outbase is None: | |
outbase = auxnames[0][:-4] | |
return inputs, auxnames, outbase | |
def report(self): | |
extra = self._get_result_extra() | |
if extra is None: | |
return 0 | |
# Parse and pretty-print the log | |
log = open(extra["outbase"] + ".blg", "rt").read() | |
inputs = extra["inputs"] | |
for msg in BibTeXFilter(log, inputs).get_messages(): | |
msg.emit() | |
# BibTeX exits with 1 if there are warnings, 2 if there are | |
# errors, and 3 if there are fatal errors (sysdep.h). | |
# Translate to a normal UNIX exit status. | |
if extra["status"] >= 2: | |
return 1 | |
return 0 | |
class BibTeXFilter: | |
def __init__(self, data, inputs): | |
self.__inputs = inputs | |
self.__key_locs = None | |
self.__messages = [] | |
prev_line = "" | |
for line in data.splitlines(): | |
msg = self.__process_line(prev_line, line) | |
if msg is not None: | |
self.__messages.append(Message(*msg)) | |
prev_line = line | |
def get_messages(self): | |
"""Return a list of warning and error Messages.""" | |
# BibTeX reports most errors in no particular order. Sort by | |
# file and line. | |
return sorted( | |
self.__messages, key=lambda msg: (msg.filename or "", msg.lineno or 0) | |
) | |
def __process_line(self, prev_line, line): | |
m = None | |
def match(regexp): | |
nonlocal m | |
m = re.match(regexp, line) | |
return m | |
# BibTeX has many error paths, but luckily the set is closed, | |
# so we can find all of them. This first case is the | |
# workhorse format. | |
# | |
# AUX errors: aux_err/aux_err_return/aux_err_print | |
# | |
# BST errors: bst_ln_num_print/bst_err/ | |
# bst_err_print_and_look_for_blank_line_return/ | |
# bst_warn_print/bst_warn/ | |
# skip_token/skip_token_print/ | |
# bst_ext_warn/bst_ext_warn_print/ | |
# bst_ex_warn/bst_ex_warn_print/ | |
# bst_mild_ex_warn/bst_mild_ex_warn_print/ | |
# bst_string_size_exceeded | |
# | |
# BIB errors: bib_ln_num_print/ | |
# bib_err_print/bib_err/ | |
# bib_warn_print/bib_warn/ | |
# bib_one_of_two_expected_err/macro_name_warning/ | |
if match("(.*?)---?line ([0-9]+) of file (.*)"): | |
# Sometimes the real error is printed on the previous line | |
if m.group(1) == "while executing": | |
# bst_ex_warn. The real message is on the previous line | |
text = prev_line | |
else: | |
text = m.group(1) or prev_line | |
typ, msg = self.__canonicalize(text) | |
return (typ, m.group(3), int(m.group(2)), msg) | |
# overflow/print_overflow | |
if match("Sorry---you've exceeded BibTeX's (.*)"): | |
return ("error", None, None, "capacity exceeded: " + m.group(1)) | |
# confusion/print_confusion | |
if match("(.*)---this can't happen$"): | |
return ("error", None, None, "internal error: " + m.group(1)) | |
# aux_end_err | |
if match("I found (no .*)---while reading file (.*)"): | |
return ("error", m.group(2), None, m.group(1)) | |
# bad_cross_reference_print/ | |
# nonexistent_cross_reference_error/ | |
# @<Complain about a nested cross reference@> | |
# | |
# This is split across two lines. Match the second. | |
if match('^refers to entry "'): | |
typ, msg = self.__canonicalize(prev_line + " " + line) | |
msg = re.sub("^a (bad cross reference)", "\\1", msg) | |
# Try to give this key a location | |
filename = lineno = None | |
m2 = re.search(r'--entry "[^"]"', prev_line) | |
if m2: | |
filename, lineno = self.__find_key(m2.group(1)) | |
return (typ, filename, lineno, msg) | |
# print_missing_entry | |
if match('Warning--I didn\'t find a database entry for (".*")'): | |
return ("warning", None, None, "no database entry for " + m.group(1)) | |
# x_warning | |
if match("Warning--(.*)"): | |
# Most formats give warnings about "something in <key>". | |
# Try to match it up. | |
filename = lineno = None | |
for m2 in reversed(list(re.finditer(r" in ([^, \t\n]+)\b", line))): | |
if m2: | |
filename, lineno = self.__find_key(m2.group(1)) | |
if filename: | |
break | |
return ("warning", filename, lineno, m.group(1)) | |
# @<Clean up and leave@> | |
if match("Aborted at line ([0-9]+) of file (.*)"): | |
return ("info", m.group(2), int(m.group(1)), "aborted") | |
# biber type errors | |
if match("^.*> WARN - (.*)$"): | |
print("warning", None, None, m.group(1)) | |
m2 = re.match("(.*) in file '(.*?)', skipping ...", m.group(1)) | |
if m2: | |
return ("warning", m2.group(2), "0", m2.group(1)) | |
return ("warning", None, None, m.group(1)) | |
if match("^.*> ERROR - (.*)$"): | |
m2 = re.match(r"BibTeX subsystem: (.*?), line (\d+), (.*)$", m.group(1)) | |
if m2: | |
return ("error", m2.group(1), m2.group(2), m2.group(3)) | |
return ("error", None, None, m.group(1)) | |
def __canonicalize(self, msg): | |
if msg.startswith("Warning"): | |
msg = re.sub("^Warning-*", "", msg) | |
typ = "warning" | |
else: | |
typ = "error" | |
msg = re.sub("^I('m| was)? ", "", msg) | |
msg = msg[:1].lower() + msg[1:] | |
return typ, msg | |
def __find_key(self, key): | |
if self.__key_locs is None: | |
p = BibTeXKeyParser() | |
self.__key_locs = {} | |
for filename in self.__inputs: | |
data = open(filename, "rt", errors="surrogateescape").read() | |
for pkey, lineno in p.parse(data): | |
self.__key_locs.setdefault(pkey, (filename, lineno)) | |
return self.__key_locs.get(key, (None, None)) | |
class BibTeXKeyParser: | |
"""Just enough of a BibTeX parser to find keys.""" | |
def parse(self, data): | |
IDENT_RE = "(?![0-9])([^\x00-\x20\x80-\xff \t\"#%'(),={}]+)" | |
self.__pos, self.__data = 0, data | |
# Find the next entry | |
while self.__consume("[^@]*@[ \t\n]*"): | |
# What type of entry? | |
if not self.__consume(IDENT_RE + "[ \t\n]*"): | |
continue | |
typ = self.__m.group(1) | |
if typ == "comment": | |
continue | |
start = self.__pos | |
if not self.__consume("([{(])[ \t\n]*"): | |
continue | |
closing, key_re = {"{": ("}", "([^, \t\n}]*)"), "(": (")", "([^, \t\n]*)")}[ | |
self.__m.group(1) | |
] | |
if typ not in ("preamble", "string"): | |
# Regular entry; get key | |
if self.__consume(key_re): | |
yield self.__m.group(1), self.__lineno() | |
# Consume body of entry | |
self.__pos = start | |
self.__balanced(closing) | |
def __consume(self, regexp): | |
self.__m = re.compile(regexp).match(self.__data, self.__pos) | |
if self.__m: | |
self.__pos = self.__m.end() | |
return self.__m | |
def __lineno(self): | |
return self.__data.count("\n", 0, self.__pos) + 1 | |
def __balanced(self, closing): | |
self.__pos += 1 | |
level = 0 | |
skip = re.compile("[{}" + closing + "]") | |
while True: | |
m = skip.search(self.__data, self.__pos) | |
if not m: | |
break | |
self.__pos = m.end() | |
ch = m.group(0) | |
if level == 0 and ch == closing: | |
break | |
elif ch == "{": | |
level += 1 | |
elif ch == "}": | |
level -= 1 | |
class Kpathsea: | |
def __init__(self, program_name): | |
self.__progname = program_name | |
def find_file(self, name, format, cwd=None, env=None): | |
"""Return the resolved path of 'name' or None.""" | |
args = ["kpsewhich", "-progname", self.__progname, "-format", format, name] | |
try: | |
verbose_cmd(args, cwd, env) | |
path = subprocess.check_output( | |
args, cwd=cwd, env=env, universal_newlines=True | |
).strip() | |
except subprocess.CalledProcessError as e: | |
if e.returncode != 1: | |
raise | |
return None | |
if cwd is None: | |
return path | |
return os.path.join(cwd, path) | |
if __name__ == "__main__": | |
main() |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
# https://gist.github.com/alexjohnj/35fa1617ed6a21fe11f126fb906d6001 | |
# This is a generic Makefile that can be used to compile somewhat complicated | |
# LaTeX documents. It assumes a directory structure like this | |
# . | |
# ├── figures/*.{pdf,tex} | |
# ├── MAIN.tex | |
# ├── Makefile | |
# ├── frontmatter.tex | |
# ├── sections/*.tex | |
# └── latex.out/ | |
# and that you want to use a LaTeX runner such as latexrun or latexmk. The | |
# Makefile's configured for latexrun but it should be easy to switch to | |
# something else by customising the variables $(LATEXRUN) and $(LATEXRUN_OPTS). | |
# | |
# The watch task uses [watchman](https://github.com/facebook/watchman) to watch | |
# all the files defined in the variables $(MAIN), $(BIB_FILE), $(FIGURES), | |
# $(PREAMBLE), $(SECTIONS) and $(OTHER_FILES) running the all task if any of | |
# them changes. | |
MAIN = 20250113-KubernetesClusterConcept | |
PREAMBLE = frontmatter.tex | |
# No references? Just assign this to nothing. | |
BIB_FILE = sources.bib | |
TMP_DIR = latex.out | |
FIGURE_DIR = figures | |
SECTIONS_DIR = sections | |
FIGURES = $(wildcard $(FIGURE_DIR)/*.tex $(FIGURE_DIR)/*.pdf) | |
SECTIONS = $(wildcard $(SECTIONS_DIR)/*.tex) | |
# Define any other files (e.g., BibLaTeX configuration files) here. | |
OTHER_FILES = $(wildcard macros/*.sty) glossary.tex | |
LATEXRUN = latexrun | |
# latexrun --latex-cmd lualatex --bibtex-cmd biber --latex-args="-shell-escape -synctex=1" --verbose-cmds main.tex | |
.PHONY: FORCE | |
$(MAIN).pdf: FORCE $(MAIN).tex $(FIGURES) $(SECTIONS) $(BIB_FILE) $(OTHER_FILES) | |
clear | |
$(LATEXRUN) $(MAIN).tex | |
.PHONY: clean | |
clean: | |
$(LATEXRUN) --clean-all | |
.PHONY: all | |
all: $(MAIN).pdf # tex-fmt | |
.PHONY: watch | |
watch: | |
watchman-make -p $(MAIN).tex $(PREAMBLE) $(FIGURES) $(SECTIONS) $(BIB_FILE) $(OTHER_FILES) -t MAIN=$(MAIN) all | |
# make for subfiles; | |
# use like this `make sub WHAT=diagram WHERE=./figures/diagram` | |
# for `./figures/diagram/diagram.tex -> ./figures/diagram/diagram.pdf` | |
.PHONY: sub | |
sub: | |
@if [ -z "$(MAIN)" ] || [ -z "$(WHERE)" ]; then \ | |
echo "Error: MAIN and WHERE must be specified. Usage: make sub MAIN=<diagram> WHERE=<directory>"; \ | |
exit 1; \ | |
fi | |
$(LATEXRUN) "$(WHERE)/$(MAIN)" -O "$(WHERE)/latex.out" -o "$(WHERE)/$(MAIN).pdf" |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment