# taken and modified from https://github.com/harvardnlp/im2markup # tokenize latex formulas import sys import os import re import argparse import subprocess import shutil from threading import Timer from datetime import datetime def run_cmd(cmd, timeout_sec=30): proc = subprocess.Popen(cmd, shell=True) kill_proc = lambda p: p.kill() timer = Timer(timeout_sec, kill_proc, [proc]) try: timer.start() stdout,stderr = proc.communicate() finally: timer.cancel() def tokenize_latex(latex_code, latex_type="", middle_file=""): if not latex_code: return False, latex_code if not latex_type: latex_type = "tabular" if "tabular" in latex_code else "formula" if not middle_file: middle_file = "out-" + datetime.now().strftime('%Y_%m_%d_%H_%M_%S') + ".txt" temp_file = middle_file + '.tmp' if latex_type == "formula": with open(temp_file, 'w') as f: prepre = latex_code # replace split, align with aligned prepre = re.sub(r'\\begin{(split|align|alignedat|alignat|eqnarray)\*?}(.+?)\\end{\1\*?}', r'\\begin{aligned}\2\\end{aligned}', prepre, flags=re.S) prepre = re.sub(r'\\begin{(smallmatrix)\*?}(.+?)\\end{\1\*?}', r'\\begin{matrix}\2\\end{matrix}', prepre, flags=re.S) f.write(prepre) cmd = r"cat %s | node %s %s > %s " % (temp_file, os.path.join(os.path.dirname(__file__), 'preprocess_formula.js'), 'normalize', middle_file) ret = subprocess.call(cmd, shell=True) os.remove(temp_file) if ret != 0: return False, latex_code operators = '\s?'.join('|'.join(['arccos', 'arcsin', 'arctan', 'arg', 'cos', 'cosh', 'cot', 'coth', 'csc', 'deg', 'det', 'dim', 'exp', 'gcd', 'hom', 'inf', 'injlim', 'ker', 'lg', 'lim', 'liminf', 'limsup', 'ln', 'log', 'max', 'min', 'Pr', 'projlim', 'sec', 'sin', 'sinh', 'sup', 'tan', 'tanh'])) ops = re.compile(r'\\operatorname {(%s)}' % operators) with open(middle_file, 'r') as fin: for line in fin: tokens = line.strip().split() tokens_out = [] for token in tokens: tokens_out.append(token) post = ' '.join(tokens_out) # use \sin instead of \operatorname{sin} names = ['\\'+x.replace(' ', '') for x in re.findall(ops, post)] post = re.sub(ops, lambda match: str(names.pop(0)), post).replace(r'\\ \end{array}', r'\end{array}') os.remove(middle_file) return True, post elif latex_type == "tabular": latex_code = latex_code.replace("\\\\%", "\\\\ %") latex_code = latex_code.replace("\%", "") latex_code = latex_code.split('%')[0] latex_code = latex_code.replace("", "\%") if not "\\end{tabular}" in latex_code: latex_code += "\\end{tabular}" with open(middle_file, 'w') as f: f.write(latex_code.replace('\r', ' ').replace('\n', ' ')) cmd = "perl -pe 's|hskip(.*?)(cm\\|in\\|pt\\|mm\\|em)|hspace{\\1\\2}|g' %s > %s"%(middle_file, temp_file) ret = subprocess.call(cmd, shell=True) if ret != 0: return False, latex_code os.remove(middle_file) cmd = r"cat %s | node %s %s > %s " % (temp_file, os.path.join(os.path.dirname(__file__), 'preprocess_tabular.js'), 'tokenize', middle_file) ret = subprocess.call(cmd, shell=True) os.remove(temp_file) if ret != 0: return False, latex_code with open(middle_file, 'r') as fin: for line in fin: tokens = line.strip().split() tokens_out = [] for token in tokens: tokens_out.append(token) post = ' '.join(tokens_out) os.remove(middle_file) return True, post else: print(f"latex type{latex_type} unrecognized.") return False, latex_code if __name__ == '__main__': latex_code = open("2.txt", 'r').read().replace('\r', ' ') print("=>", latex_code) new_code = tokenize_latex(latex_code) print("=>", new_code)