# This file is part of lyx2lyx # Copyright (C) 2016 The LyX team # # This program is free software; you can redistribute it and/or # modify it under the terms of the GNU General Public License # as published by the Free Software Foundation; either version 2 # of the License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. """Convert files to the file format generated by lyx 2.3""" import re from lyx2lyx_tools import ( add_to_preamble, insert_document_option, insert_to_preamble, is_document_option, latex_length, put_cmd_in_ert, remove_document_option, revert_font_attrs, revert_language, ) # Uncomment only what you need to import, please. from parser_tools import ( del_complete_lines, del_token, del_value, find_complete_lines, find_end_of_inset, find_end_of_layout, find_re, find_substring, find_token, find_token_backwards, get_bool_value, get_containing_inset, get_containing_layout, get_quoted_value, get_value, is_in_inset, set_bool_value, ) #################################################################### # Private helper functions ############################################################################### ### ### Conversion and reversion routines ### ############################################################################### def convert_microtype(document): "Add microtype settings." i = find_token(document.header, "\\font_tt_scale") j = find_token(document.preamble, "\\usepackage{microtype}") if j == -1: document.header.insert(i + 1, "\\use_microtype false") else: document.header.insert(i + 1, "\\use_microtype true") del document.preamble[j] if j and document.preamble[j - 1] == "% Added by lyx2lyx": del document.preamble[j - 1] def revert_microtype(document): "Remove microtype settings." use_microtype = get_bool_value(document.header, "\\use_microtype", delete=True) if use_microtype: add_to_preamble(document, ["\\usepackage{microtype}"]) def convert_dateinset(document): "Convert date external inset to ERT" i = 0 while True: i = find_token(document.body, "\\begin_inset External", i + 1) if i == -1: return j = find_end_of_inset(document.body, i) if j == -1: document.warning( "Malformed lyx document: Missing '\\end_inset' in convert_dateinset." ) continue if get_value(document.body, "template", i, j) == "Date": document.body[i : j + 1] = put_cmd_in_ert("\\today ") i = j # skip inset def convert_inputenc(document): """Replace no longer supported input encoding setting.""" i = find_token(document.header, "\\inputencoding pt254") if i != -1: document.header[i] = "\\inputencoding pt154" def convert_ibranches(document): 'Add "inverted 0" to branch insets' i = 0 while True: i = find_token(document.body, "\\begin_inset Branch", i + 1) if i == -1: return document.body.insert(i + 1, "inverted 0") def revert_ibranches(document): "Convert inverted branches to explicit anti-branches" # Get list of branches ourbranches = {} i = 0 while True: i = find_token(document.header, "\\branch", i + 1) if i == -1: break branch = document.header[i][8:].strip() selected = get_bool_value(document.header, "\\selected", i + 1, i + 2) if selected is None: document.warning( "Malformed LyX document: No selection indicator " "for branch %s." % branch ) selected = True # the value tells us whether the branch is selected ourbranches[branch] = selected # Find branch insets, remove "inverted" tag and # convert inverted insets to "Anti-OldBranch" insets antibranches = {} i = 0 while True: i = find_token(document.body, "\\begin_inset Branch", i + 1) if i == -1: break inverted = get_bool_value(document.body, "inverted", i + 1, i + 2, delete=True) if inverted is None: document.warning("Malformed LyX document: Missing 'inverted' tag in branch inset.") continue if inverted: branch = document.body[i][20:].strip() if branch not in antibranches: antibranch = "Anti-" + branch while antibranch in antibranches: antibranch = "x" + antibranch antibranches[branch] = antibranch else: antibranch = antibranches[branch] document.body[i] = "\\begin_inset Branch " + antibranch # now we need to add the new branches to the header for old, new in antibranches.items(): i = find_token(document.header, "\\branch " + old, 0) if i == -1: document.warning("Can't find branch %s even though we found it before!" % (old)) continue j = find_token(document.header, "\\end_branch", i) if j == -1: document.warning("Malformed LyX document! Can't find end of branch " + old) continue lines = ["\\branch " + new, "\\selected %d" % (not ourbranches[old])] # these are the old lines telling us color, etc. lines += document.header[i + 2 : j + 1] document.header[i:i] = lines beamer_article_styles = [ "### Inserted by lyx2lyx (more [scr]article styles) ###", "Input article.layout", "Input beamer.layout", "Provides geometry 0", "Provides hyperref 0", "DefaultFont", " Family Roman", " Series Medium", " Shape Up", " Size Normal", " Color None", "EndFont", "Preamble", " \\usepackage{beamerarticle,pgf}", " % this default might be overridden by plain title style", " \\newcommand\\makebeamertitle{\\frame{\\maketitle}}%", " \\AtBeginDocument{", " \\let\\origtableofcontents=\\tableofcontents", " \\def\\tableofcontents{\\@ifnextchar[{\\origtableofcontents}{\\gobbletableofcontents}}", " \\def\\gobbletableofcontents#1{\\origtableofcontents}", " }", "EndPreamble", "### End of insertion by lyx2lyx (more [scr]article styles) ###", ] def revert_beamer_article_styles(document): "Include (scr)article styles in beamer article" beamer_articles = ["article-beamer", "scrarticle-beamer"] if document.textclass not in beamer_articles: return if document.textclass == "scrarticle-beamer": beamer_article_styles[1] = "Input scrartcl.layout" document.append_local_layout(beamer_article_styles) def convert_beamer_article_styles(document): "Remove included (scr)article styles in beamer article" beamer_articles = ["article-beamer", "scrarticle-beamer"] if document.textclass not in beamer_articles: return if document.textclass == "scrarticle-beamer": beamer_article_styles[1] = "Input scrartcl.layout" document.del_local_layout(beamer_article_styles) def revert_new_babel_languages(document): """Revert "bosnian", "friulan", "macedonian", "piedmontese", "romansh". Set the document language to English but use correct babel setting. """ nblanguages = ["bosnian", "friulan", "macedonian", "piedmontese", "romansh"] for lang in nblanguages: if lang == "bosnian" or lang == "macedonian": # These are only supported by babel revert_language(document, lang, lang, "") else: # These are supported by babel and polyglossia revert_language(document, lang, lang, lang) # TODO: # def convert_new_babel_languages(document) # set to native support if get_value(document.header, "\\options") in # ["bosnian", "friulan", "macedonian", "piedmontese", "romansh"] # and Babel is used. def revert_amharic(document): "Set the document language to English but assure Amharic output" revert_language(document, "amharic", "", "amharic") def revert_asturian(document): "Set the document language to English but assure Asturian output" revert_language(document, "asturian", "", "asturian") def revert_kannada(document): "Set the document language to English but assure Kannada output" revert_language(document, "kannada", "", "kannada") def revert_khmer(document): "Set the document language to English but assure Khmer output" revert_language(document, "khmer", "", "khmer") def revert_urdu(document): "Set the document language to English but assure Urdu output" revert_language(document, "urdu", "", "urdu") def revert_syriac(document): "Set the document language to English but assure Syriac output" revert_language(document, "syriac", "", "syriac") def revert_quotes(document): "Revert Quote Insets in verbatim or Hebrew context to plain quotes" # First handle verbatim insets i = 0 j = 0 while i < len(document.body): words = document.body[i].split() if ( len(words) > 1 and words[0] == "\\begin_inset" and ( words[1] in ["ERT", "listings"] or (len(words) > 2 and words[2] in ["URL", "Chunk", "Sweave", "S/R"]) ) ): j = find_end_of_inset(document.body, i) if j == -1: document.warning( "Malformed LyX document: Can't find end of " + words[1] + " inset at line " + str(i) ) i += 1 continue while True: k = find_token(document.body, "\\begin_inset Quotes", i, j) if k == -1: i += 1 break l = find_end_of_inset(document.body, k) if l == -1: document.warning( "Malformed LyX document: Can't find end of Quote inset at line " + str(k) ) i = k continue replace = '"' if document.body[k].endswith("s"): replace = "'" document.body[k : l + 2] = [replace] else: i += 1 continue # Now verbatim layouts i = 0 j = 0 while i < len(document.body): words = document.body[i].split() if ( len(words) > 1 and words[0] == "\\begin_layout" and words[1] in ["Verbatim", "Verbatim*", "Code", "Author_Email", "Author_URL"] ): j = find_end_of_layout(document.body, i) if j == -1: document.warning( "Malformed LyX document: Can't find end of " + words[1] + " layout at line " + str(i) ) i += 1 continue while True: k = find_token(document.body, "\\begin_inset Quotes", i, j) if k == -1: i += 1 break l = find_end_of_inset(document.body, k) if l == -1: document.warning( "Malformed LyX document: Can't find end of Quote inset at line " + str(k) ) i = k continue replace = '"' if document.body[k].endswith("s"): replace = "'" document.body[k : l + 2] = [replace] else: i += 1 continue # Now handle Hebrew if ( not document.language == "hebrew" and find_token(document.body, "\\lang hebrew", 0) == -1 ): return i = 0 j = 0 while True: k = find_token(document.body, "\\begin_inset Quotes", i) if k == -1: return l = find_end_of_inset(document.body, k) if l == -1: document.warning( "Malformed LyX document: Can't find end of Quote inset at line " + str(k) ) i = k continue hebrew = False parent = get_containing_layout(document.body, k) ql = find_token_backwards(document.body, "\\lang", k) if ql == -1 or ql < parent[1]: hebrew = document.language == "hebrew" elif document.body[ql] == "\\lang hebrew": hebrew = True if hebrew: replace = '"' if document.body[k].endswith("s"): replace = "'" document.body[k : l + 2] = [replace] i = l iopart_local_layout = [ "### Inserted by lyx2lyx (stdlayouts) ###", "Input stdlayouts.inc", "### End of insertion by lyx2lyx (stdlayouts) ###" "", ] def revert_iopart(document): "Input new styles via local layout" if document.textclass != "iopart": return document.append_local_layout(iopart_local_layout) def convert_iopart(document): "Remove local layout we added, if it is there" if document.textclass != "iopart": return document.del_local_layout(iopart_local_layout) def convert_quotestyle(document): "Convert \\quotes_language to \\quotes_style" i = find_token(document.header, "\\quotes_language", 0) if i == -1: document.warning("Malformed LyX document! Can't find \\quotes_language!") return val = get_value(document.header, "\\quotes_language", i) document.header[i] = "\\quotes_style " + val def revert_quotestyle(document): "Revert \\quotes_style to \\quotes_language" i = find_token(document.header, "\\quotes_style", 0) if i == -1: document.warning("Malformed LyX document! Can't find \\quotes_style!") return val = get_value(document.header, "\\quotes_style", i) document.header[i] = "\\quotes_language " + val def revert_plainquote(document): "Revert plain quote insets" # First, revert style setting i = find_token(document.header, "\\quotes_style plain", 0) if i != -1: document.header[i] = "\\quotes_style english" # now the insets i = 0 j = 0 while True: k = find_token(document.body, "\\begin_inset Quotes q", i) if k == -1: return l = find_end_of_inset(document.body, k) if l == -1: document.warning( "Malformed LyX document: Can't find end of Quote inset at line " + str(k) ) i = k continue replace = '"' if document.body[k].endswith("s"): replace = "'" document.body[k : l + 2] = [replace] i = l def convert_frenchquotes(document): "Convert french quote insets to swiss" # First, revert style setting i = find_token(document.header, "\\quotes_style french", 0) if i != -1: document.header[i] = "\\quotes_style swiss" # now the insets i = 0 while True: i = find_token(document.body, "\\begin_inset Quotes f", i) if i == -1: return val = get_value(document.body, "\\begin_inset Quotes", i)[7:] newval = val.replace("f", "c", 1) document.body[i] = document.body[i].replace(val, newval) i += 1 def revert_swissquotes(document): "Revert swiss quote insets to french" # First, revert style setting i = find_token(document.header, "\\quotes_style swiss", 0) if i != -1: document.header[i] = "\\quotes_style french" # now the insets i = 0 while True: i = find_token(document.body, "\\begin_inset Quotes c", i) if i == -1: return val = get_value(document.body, "\\begin_inset Quotes", i)[7:] newval = val.replace("c", "f", 1) document.body[i] = document.body[i].replace(val, newval) i += 1 def revert_britishquotes(document): "Revert british quote insets to english" # First, revert style setting i = find_token(document.header, "\\quotes_style british", 0) if i != -1: document.header[i] = "\\quotes_style english" # now the insets i = 0 while True: i = find_token(document.body, "\\begin_inset Quotes b", i) if i == -1: return val = get_value(document.body, "\\begin_inset Quotes", i)[7:] newval = val.replace("b", "e", 1) if val[2] == "d": # opening mark newval = newval.replace("d", "s") else: # closing mark newval = newval.replace("s", "d") document.body[i] = document.body[i].replace(val, newval) i += 1 def revert_swedishgquotes(document): "Revert swedish quote insets" # First, revert style setting i = find_token(document.header, "\\quotes_style swedishg", 0) if i != -1: document.header[i] = "\\quotes_style danish" # now the insets i = 0 while True: i = find_token(document.body, "\\begin_inset Quotes w", i) if i == -1: return val = get_value(document.body, "\\begin_inset Quotes", i)[7:] if val[2] == "d": # outer marks newval = val.replace("w", "a", 1).replace("r", "l") else: # inner marks newval = val.replace("w", "s", 1) document.body[i] = document.body[i].replace(val, newval) i += 1 def revert_frenchquotes(document): "Revert french inner quote insets" i = 0 while True: i = find_token(document.body, "\\begin_inset Quotes f", i) if i == -1: return val = get_value(document.body, "\\begin_inset Quotes", i)[7:] if val[2] == "s": # inner marks newval = val.replace("f", "e", 1).replace("s", "d") document.body[i] = document.body[i].replace(val, newval) i += 1 def revert_frenchinquotes(document): "Revert inner frenchin quote insets" # First, revert style setting i = find_token(document.header, "\\quotes_style frenchin", 0) if i != -1: document.header[i] = "\\quotes_style french" # now the insets i = 0 while True: i = find_token(document.body, "\\begin_inset Quotes i", i) if i == -1: return val = get_value(document.body, "\\begin_inset Quotes", i)[7:] newval = val.replace("i", "f", 1) if val[2] == "s": # inner marks newval = newval.replace("s", "d") document.body[i] = document.body[i].replace(val, newval) i += 1 def revert_russianquotes(document): "Revert russian quote insets" # First, revert style setting i = find_token(document.header, "\\quotes_style russian", 0) if i != -1: document.header[i] = "\\quotes_style french" # now the insets i = 0 while True: i = find_token(document.body, "\\begin_inset Quotes r", i) if i == -1: return val = get_value(document.body, "\\begin_inset Quotes", i)[7:] newval = val if val[2] == "s": # inner marks newval = val.replace("r", "g", 1).replace("s", "d") else: # outer marks newval = val.replace("r", "f", 1) document.body[i] = document.body[i].replace(val, newval) i += 1 def revert_dynamicquotes(document): "Revert dynamic quote insets" # First, revert header i = find_token(document.header, "\\dynamic_quotes", 0) if i != -1: del document.header[i] # Get global style style = "english" i = find_token(document.header, "\\quotes_style", 0) if i == -1: document.warning("Malformed document! Missing \\quotes_style") else: style = get_value(document.header, "\\quotes_style", i) s = "e" if style == "english": s = "e" elif style == "swedish": s = "s" elif style == "german": s = "g" elif style == "polish": s = "p" elif style == "swiss": s = "c" elif style == "danish": s = "a" elif style == "plain": s = "q" elif style == "british": s = "b" elif style == "swedishg": s = "w" elif style == "french": s = "f" elif style == "frenchin": s = "i" elif style == "russian": s = "r" # now transform the insets i = 0 while True: i = find_token(document.body, "\\begin_inset Quotes x", i) if i == -1: return document.body[i] = document.body[i].replace("x", s) i += 1 def revert_cjkquotes(document): "Revert cjk quote insets" # Get global style style = "english" i = find_token(document.header, "\\quotes_style", 0) if i == -1: document.warning("Malformed document! Missing \\quotes_style") else: style = get_value(document.header, "\\quotes_style", i) global_cjk = style.find("cjk") != -1 if global_cjk: document.header[i] = "\\quotes_style english" # transform dynamic insets s = "j" if style == "cjkangle": s = "k" i = 0 while True: i = find_token(document.body, "\\begin_inset Quotes x", i) if i == -1: break document.body[i] = document.body[i].replace("x", s) i += 1 cjk_langs = [ "chinese-simplified", "chinese-traditional", "japanese", "japanese-cjk", "korean", ] i = 0 j = 0 while True: k = find_token(document.body, "\\begin_inset Quotes j", i) if k == -1: break l = find_end_of_inset(document.body, k) if l == -1: document.warning( "Malformed LyX document: Can't find end of Quote inset at line " + str(k) ) i = k continue cjk = False parent = get_containing_layout(document.body, k) ql = find_token_backwards(document.body, "\\lang", k) if ql == -1 or ql < parent[1]: cjk = document.language in cjk_langs elif document.body[ql].split()[1] in cjk_langs: cjk = True val = get_value(document.body, "\\begin_inset Quotes", i)[7:] replace = [] if val[2] == "s": # inner marks if val[1] == "l": # inner opening mark if cjk: replace = ["\u300e"] else: replace = ["\\begin_inset Formula $\\llceil$", "\\end_inset"] else: # inner closing mark if cjk: replace = ["\u300f"] else: replace = ["\\begin_inset Formula $\\rrfloor$", "\\end_inset"] else: # outer marks if val[1] == "l": # outer opening mark if cjk: replace = ["\u300c"] else: replace = ["\\begin_inset Formula $\\lceil$", "\\end_inset"] else: # outer closing mark if cjk: replace = ["\u300d"] else: replace = ["\\begin_inset Formula $\\rfloor$", "\\end_inset"] document.body[k : l + 1] = replace i = l i = 0 j = 0 while True: k = find_token(document.body, "\\begin_inset Quotes k", i) if k == -1: return l = find_end_of_inset(document.body, k) if l == -1: document.warning( "Malformed LyX document: Can't find end of Quote inset at line " + str(k) ) i = k continue cjk = False parent = get_containing_layout(document.body, k) ql = find_token_backwards(document.body, "\\lang", k) if ql == -1 or ql < parent[1]: cjk = document.language in cjk_langs elif document.body[ql].split()[1] in cjk_langs: cjk = True val = get_value(document.body, "\\begin_inset Quotes", i)[7:] replace = [] if val[2] == "s": # inner marks if val[1] == "l": # inner opening mark if cjk: replace = ["\u3008"] else: replace = ["\\begin_inset Formula $\\langle$", "\\end_inset"] else: # inner closing mark if cjk: replace = ["\u3009"] else: replace = ["\\begin_inset Formula $\\rangle$", "\\end_inset"] else: # outer marks if val[1] == "l": # outer opening mark if cjk: replace = ["\u300a"] else: replace = [ "\\begin_inset Formula $\\langle\\kern -2.5pt\\langle$", "\\end_inset", ] else: # outer closing mark if cjk: replace = ["\u300b"] else: replace = [ "\\begin_inset Formula $\\rangle\\kern -2.5pt\\rangle$", "\\end_inset", ] document.body[k : l + 1] = replace i = l def convert_crimson(document): """Transform preamble code to native font setting.""" # Quick-check: i = find_substring(document.preamble, "{cochineal}") if i == -1: return # Find and delete user-preamble code: if document.preamble[i] == "\\usepackage[proportional,osf]{cochineal}": osf = True elif document.preamble[i] == "\\usepackage{cochineal}": osf = False else: return del document.preamble[i] if i and document.preamble[i - 1] == "% Added by lyx2lyx": del document.preamble[i - 1] # Convert to native font setting: j = find_token(document.header, "\\font_roman") if j == -1: romanfont = ["\font_roman", '"cochineal"', '"default"'] else: romanfont = document.header[j].split() romanfont[1] = '"cochineal"' document.header[j] = " ".join(romanfont) try: set_bool_value(document.header, "\\font_osf", osf) except ValueError: # no \\font_osf setting in document.header if osf: document.header.insert(-1, "\\font_osf true") def revert_crimson(document): "Revert native Cochineal/Crimson font definition to LaTeX" i = find_token(document.header, '\\font_roman "cochineal"') if i == -1: return # replace unsupported font setting document.header[i] = document.header[i].replace("cochineal", "default") # no need for preamble code with system fonts if get_bool_value(document.header, "\\use_non_tex_fonts"): return # transfer old style figures setting to package options j = find_token(document.header, "\\font_osf true") if j != -1: options = "[proportional,osf]" document.header[j] = "\\font_osf false" else: options = "" add_to_preamble(document, ["\\usepackage%s{cochineal}" % options]) def revert_cochinealmath(document): "Revert cochineal newtxmath definitions to LaTeX" if find_token(document.header, "\\use_non_tex_fonts false", 0) != -1: i = find_token(document.header, '\\font_math "cochineal-ntxm"', 0) if i != -1: add_to_preamble(document, "\\usepackage[cochineal]{newtxmath}") document.header[i] = document.header[i].replace("cochineal-ntxm", "auto") def revert_labelonly(document): "Revert labelonly tag for InsetRef" i = 0 while True: i = find_token(document.body, "\\begin_inset CommandInset ref", i) if i == -1: return j = find_end_of_inset(document.body, i) if j == -1: document.warning("Can't find end of reference inset at line %d!!" % (i)) i += 1 continue k = find_token(document.body, "LatexCommand labelonly", i, j) if k == -1: i = j continue label = get_quoted_value(document.body, "reference", i, j) if not label: document.warning("Can't find label for reference at line %d!" % (i)) i = j + 1 continue document.body[i : j + 1] = put_cmd_in_ert([label]) i += 1 def revert_plural_refs(document): "Revert plural and capitalized references" i = find_token(document.header, "\\use_refstyle 1", 0) use_refstyle = i != 0 i = 0 while True: i = find_token(document.body, "\\begin_inset CommandInset ref", i) if i == -1: return j = find_end_of_inset(document.body, i) if j == -1: document.warning("Can't find end of reference inset at line %d!!" % (i)) i += 1 continue plural = caps = suffix = False k = find_token(document.body, "LaTeXCommand formatted", i, j) if k != -1 and use_refstyle: plural = get_bool_value(document.body, "plural", i, j, False) caps = get_bool_value(document.body, "caps", i, j, False) label = get_quoted_value(document.body, "reference", i, j) if label: try: (prefix, suffix) = label.split(":", 1) except: document.warning( "No `:' separator in formatted reference at line %d!" % (i) ) else: document.warning("Can't find label for reference at line %d!" % (i)) # this effectively tests also for use_refstyle and a formatted reference # we do this complicated test because we would otherwise do this erasure # over and over and over if not ((plural or caps) and suffix): del_token(document.body, "plural", i, j) del_token(document.body, "caps", i, j - 1) # since we deleted a line i = j - 1 continue if caps: prefix = prefix[0].title() + prefix[1:] cmd = "\\" + prefix + "ref" if plural: cmd += "[s]" cmd += "{" + suffix + "}" document.body[i : j + 1] = put_cmd_in_ert([cmd]) i += 1 def revert_noprefix(document): "Revert labelonly tags with 'noprefix' set" i = 0 while True: i = find_token(document.body, "\\begin_inset CommandInset ref", i) if i == -1: return j = find_end_of_inset(document.body, i) if j == -1: document.warning("Can't find end of reference inset at line %d!!" % (i)) i += 1 continue k = find_token(document.body, "LatexCommand labelonly", i, j) noprefix = False if k != -1: noprefix = get_bool_value(document.body, "noprefix", i, j) if not noprefix: # either it was not a labelonly command, or else noprefix was not set. # in that case, we just delete the option. del_token(document.body, "noprefix", i, j) i = j continue label = get_quoted_value(document.body, "reference", i, j) if not label: document.warning("Can't find label for reference at line %d!" % (i)) i = j + 1 continue try: (prefix, suffix) = label.split(":", 1) except: document.warning("No `:' separator in formatted reference at line %d!" % (i)) # we'll leave this as an ordinary labelonly reference del_token(document.body, "noprefix", i, j) i = j continue document.body[i : j + 1] = put_cmd_in_ert([suffix]) i += 1 def revert_biblatex(document): "Revert biblatex support" # # Header # # 1. Get cite engine engine = "basic" i = find_token(document.header, "\\cite_engine", 0) if i == -1: document.warning("Malformed document! Missing \\cite_engine") else: engine = get_value(document.header, "\\cite_engine", i) # 2. Store biblatex state and revert to natbib biblatex = False if engine in ["biblatex", "biblatex-natbib"]: biblatex = True document.header[i] = "\\cite_engine natbib" # 3. Store and remove new document headers bibstyle = "" i = find_token(document.header, "\\biblatex_bibstyle", 0) if i != -1: bibstyle = get_value(document.header, "\\biblatex_bibstyle", i) del document.header[i] citestyle = "" i = find_token(document.header, "\\biblatex_citestyle", 0) if i != -1: citestyle = get_value(document.header, "\\biblatex_citestyle", i) del document.header[i] biblio_options = "" i = find_token(document.header, "\\biblio_options", 0) if i != -1: biblio_options = get_value(document.header, "\\biblio_options", i) del document.header[i] if biblatex: bbxopts = "[natbib=true" if bibstyle != "": bbxopts += ",bibstyle=" + bibstyle if citestyle != "": bbxopts += ",citestyle=" + citestyle if biblio_options != "": bbxopts += "," + biblio_options bbxopts += "]" add_to_preamble(document, "\\usepackage" + bbxopts + "{biblatex}") # # Body # # 1. Bibtex insets i = 0 bibresources = [] while True: i = find_token(document.body, "\\begin_inset CommandInset bibtex", i) if i == -1: break j = find_end_of_inset(document.body, i) if j == -1: document.warning("Can't find end of bibtex inset at line %d!!" % (i)) i += 1 continue bibs = get_quoted_value(document.body, "bibfiles", i, j) opts = get_quoted_value(document.body, "biblatexopts", i, j) # store resources if bibs: bibresources += bibs.split(",") else: document.warning("Can't find bibfiles for bibtex inset at line %d!" % (i)) # remove biblatexopts line k = find_token(document.body, "biblatexopts", i, j) if k != -1: del document.body[k] # Re-find inset end line j = find_end_of_inset(document.body, i) # Insert ERT \\printbibliography and wrap bibtex inset to a Note if biblatex: pcmd = "printbibliography" if opts: pcmd += "[" + opts + "]" repl = [ "\\begin_inset ERT", "status open", "", "\\begin_layout Plain Layout", "", "", "\\backslash", pcmd, "\\end_layout", "", "\\end_inset", "", "", "\\end_layout", "", "\\begin_layout Standard", "\\begin_inset Note Note", "status open", "", "\\begin_layout Plain Layout", ] repl += document.body[i : j + 1] repl += ["", "\\end_layout", "", "\\end_inset", "", ""] document.body[i : j + 1] = repl j += 27 i = j + 1 if biblatex: for b in bibresources: add_to_preamble(document, "\\addbibresource{" + b + ".bib}") # 2. Citation insets # Specific citation insets used in biblatex that need to be reverted to ERT new_citations = { "Cite": "Cite", "citebyear": "citeyear", "citeyear": "cite*", "Footcite": "Smartcite", "footcite": "smartcite", "Autocite": "Autocite", "autocite": "autocite", "citetitle": "citetitle", "citetitle*": "citetitle*", "fullcite": "fullcite", "footfullcite": "footfullcite", "supercite": "supercite", "citeauthor": "citeauthor", "citeauthor*": "citeauthor*", "Citeauthor": "Citeauthor", "Citeauthor*": "Citeauthor*", } # All commands accepted by LyX < 2.3. Everything else throws an error. old_citations = [ "cite", "nocite", "citet", "citep", "citealt", "citealp", "citeauthor", "citeyear", "citeyearpar", "citet*", "citep*", "citealt*", "citealp*", "citeauthor*", "Citet", "Citep", "Citealt", "Citealp", "Citeauthor", "Citet*", "Citep*", "Citealt*", "Citealp*", "Citeauthor*", "fullcite", "footcite", "footcitet", "footcitep", "footcitealt", "footcitealp", "footciteauthor", "footciteyear", "footciteyearpar", "citefield", "citetitle", "cite*", ] i = 0 while True: i = find_token(document.body, "\\begin_inset CommandInset citation", i) if i == -1: break j = find_end_of_inset(document.body, i) if j == -1: document.warning("Can't find end of citation inset at line %d!!" % (i)) i += 1 continue k = find_token(document.body, "LatexCommand", i, j) if k == -1: document.warning("Can't find LatexCommand for citation inset at line %d!" % (i)) i = j + 1 continue cmd = get_value(document.body, "LatexCommand", k) if biblatex and cmd in list(new_citations.keys()): pre = get_quoted_value(document.body, "before", i, j) post = get_quoted_value(document.body, "after", i, j) key = get_quoted_value(document.body, "key", i, j) if not key: document.warning("Citation inset at line %d does not have a key!" % (i)) key = "???" # Replace known new commands with ERT res = "\\" + new_citations[cmd] if pre: res += "[" + pre + "]" if post: res += "[" + post + "]" elif pre: res += "[]" res += "{" + key + "}" document.body[i : j + 1] = put_cmd_in_ert([res]) elif cmd not in old_citations: # Reset unknown commands to cite. This is what LyX does as well # (but LyX 2.2 would break on unknown commands) document.body[k] = "LatexCommand cite" document.warning("Reset unknown cite command '%s' with cite" % cmd) i = j + 1 # Emulate the old biblatex-workaround (pretend natbib in order to use the styles) if biblatex: biblatex_emulation = [ "### Inserted by lyx2lyx (biblatex emulation) ###", "Provides natbib 1", "### End of insertion by lyx2lyx (biblatex emulation) ###", ] document.append_local_layout(biblatex_emulation) def revert_citekeyonly(document): "Revert keyonly cite command to ERT" i = 0 while True: i = find_token(document.body, "\\begin_inset CommandInset citation", i) if i == -1: break j = find_end_of_inset(document.body, i) if j == -1: document.warning("Can't find end of citation inset at line %d!!" % (i)) i += 1 continue k = find_token(document.body, "LatexCommand", i, j) if k == -1: document.warning("Can't find LatexCommand for citation inset at line %d!" % (i)) i = j + 1 continue cmd = get_value(document.body, "LatexCommand", k) if cmd != "keyonly": i = j + 1 continue key = get_quoted_value(document.body, "key", i, j) if not key: document.warning("Citation inset at line %d does not have a key!" % (i)) # Replace known new commands with ERT document.body[i : j + 1] = put_cmd_in_ert([key]) i = j + 1 def revert_bibpackopts(document): "Revert support for natbib/jurabib package options" engine = "basic" i = find_token(document.header, "\\cite_engine", 0) if i == -1: document.warning("Malformed document! Missing \\cite_engine") else: engine = get_value(document.header, "\\cite_engine", i) biblatex = False if engine not in ["natbib", "jurabib"]: return i = find_token(document.header, "\\biblio_options", 0) if i == -1: # Nothing to do if we have no options return biblio_options = get_value(document.header, "\\biblio_options", i) del document.header[i] if not biblio_options: # Nothing to do for empty options return bibliography_package_options = [ "### Inserted by lyx2lyx (bibliography package options) ###", "PackageOptions " + engine + " " + biblio_options, "### End of insertion by lyx2lyx (bibliography package options) ###", ] document.append_local_layout(bibliography_package_options) def revert_qualicites(document): "Revert qualified citation list commands to ERT" # Citation insets that support qualified lists, with their LaTeX code ql_citations = { "cite": "cites", "Cite": "Cites", "citet": "textcites", "Citet": "Textcites", "citep": "parencites", "Citep": "Parencites", "Footcite": "Smartcites", "footcite": "smartcites", "Autocite": "Autocites", "autocite": "autocites", } # Get cite engine engine = "basic" i = find_token(document.header, "\\cite_engine", 0) if i == -1: document.warning("Malformed document! Missing \\cite_engine") else: engine = get_value(document.header, "\\cite_engine", i) biblatex = engine in ["biblatex", "biblatex-natbib"] i = 0 while True: i = find_token(document.body, "\\begin_inset CommandInset citation", i) if i == -1: break j = find_end_of_inset(document.body, i) if j == -1: document.warning("Can't find end of citation inset at line %d!!" % (i)) i += 1 continue pres = find_token(document.body, "pretextlist", i, j) posts = find_token(document.body, "posttextlist", i, j) if pres == -1 and posts == -1: # nothing to do. i = j + 1 continue pretexts = get_quoted_value(document.body, "pretextlist", pres) posttexts = get_quoted_value(document.body, "posttextlist", posts) k = find_token(document.body, "LatexCommand", i, j) if k == -1: document.warning("Can't find LatexCommand for citation inset at line %d!" % (i)) i = j + 1 continue cmd = get_value(document.body, "LatexCommand", k) if biblatex and cmd in list(ql_citations.keys()): pre = get_quoted_value(document.body, "before", i, j) post = get_quoted_value(document.body, "after", i, j) key = get_quoted_value(document.body, "key", i, j) if not key: document.warning("Citation inset at line %d does not have a key!" % (i)) key = "???" keys = key.split(",") prelist = pretexts.split("\t") premap = dict() for pp in prelist: ppp = pp.split(" ", 1) premap[ppp[0]] = ppp[1] postlist = posttexts.split("\t") postmap = dict() for pp in postlist: ppp = pp.split(" ", 1) postmap[ppp[0]] = ppp[1] # Replace known new commands with ERT if "(" in pre or ")" in pre: pre = "{" + pre + "}" if "(" in post or ")" in post: post = "{" + post + "}" res = "\\" + ql_citations[cmd] if pre: res += "(" + pre + ")" if post: res += "(" + post + ")" elif pre: res += "()" for kk in keys: if premap.get(kk, "") != "": res += "[" + premap[kk] + "]" if postmap.get(kk, "") != "": res += "[" + postmap[kk] + "]" elif premap.get(kk, "") != "": res += "[]" res += "{" + kk + "}" document.body[i : j + 1] = put_cmd_in_ert([res]) else: # just remove the params del document.body[posttexts] del document.body[pretexts] i += 1 command_insets = ["bibitem", "citation", "href", "index_print", "nomenclature"] def convert_literalparam(document): "Add param literal" pos = len("\\begin_inset CommandInset ") i = 0 while True: i = find_token(document.body, "\\begin_inset CommandInset", i) if i == -1: break inset = document.body[i][pos:].strip() if inset not in command_insets: i += 1 continue j = find_end_of_inset(document.body, i) if j == -1: document.warning( "Malformed LyX document: Can't find end of %s inset at line %d" % (inset, i) ) i += 1 continue while i < j and document.body[i].strip() != "": i += 1 # href is already fully latexified. Here we can switch off literal. if inset == "href": document.body.insert(i, 'literal "false"') else: document.body.insert(i, 'literal "true"') i = j + 1 def revert_literalparam(document): "Remove param literal" for inset in command_insets: i = 0 while True: i = find_token(document.body, "\\begin_inset CommandInset %s" % inset, i + 1) if i == -1: break j = find_end_of_inset(document.body, i) if j == -1: document.warning( "Malformed LyX document: Can't find end of %s inset at line %d" % (inset, i) ) continue del_token(document.body, "literal", i, j) def revert_multibib(document): "Revert multibib support" # 1. Get cite engine engine = "basic" i = find_token(document.header, "\\cite_engine", 0) if i == -1: document.warning("Malformed document! Missing \\cite_engine") else: engine = get_value(document.header, "\\cite_engine", i) # 2. Do we use biblatex? biblatex = False if engine in ["biblatex", "biblatex-natbib"]: biblatex = True # 3. Store and remove multibib document header multibib = "" i = find_token(document.header, "\\multibib", 0) if i != -1: multibib = get_value(document.header, "\\multibib", i) del document.header[i] if not multibib: return # 4. The easy part: Biblatex if biblatex: i = find_token(document.header, "\\biblio_options", 0) if i == -1: k = find_token(document.header, "\\use_bibtopic", 0) if k == -1: # this should not happen document.warning("Malformed LyX document! No \\use_bibtopic header found!") return document.header[k - 1 : k - 1] = ["\\biblio_options " + "refsection=" + multibib] else: biblio_options = get_value(document.header, "\\biblio_options", i) if biblio_options: biblio_options += "," biblio_options += "refsection=" + multibib document.header[i] = "\\biblio_options " + biblio_options # Bibtex insets i = 0 while True: i = find_token(document.body, "\\begin_inset CommandInset bibtex", i) if i == -1: break j = find_end_of_inset(document.body, i) if j == -1: document.warning("Can't find end of bibtex inset at line %d!!" % (i)) i += 1 continue btprint = get_quoted_value(document.body, "btprint", i, j) if btprint != "bibbysection": i += 1 continue opts = get_quoted_value(document.body, "biblatexopts", i, j) # change btprint line k = find_token(document.body, "btprint", i, j) if k != -1: document.body[k] = 'btprint "btPrintCited"' # Insert ERT \\bibbysection and wrap bibtex inset to a Note pcmd = "bibbysection" if opts: pcmd += "[" + opts + "]" repl = [ "\\begin_inset ERT", "status open", "", "\\begin_layout Plain Layout", "", "", "\\backslash", pcmd, "\\end_layout", "", "\\end_inset", "", "", "\\end_layout", "", "\\begin_layout Standard", "\\begin_inset Note Note", "status open", "", "\\begin_layout Plain Layout", ] repl += document.body[i : j + 1] repl += ["", "\\end_layout", "", "\\end_inset", "", ""] document.body[i : j + 1] = repl j += 27 i = j + 1 return # 5. More tricky: Bibtex/Bibtopic k = find_token(document.header, "\\use_bibtopic", 0) if k == -1: # this should not happen document.warning("Malformed LyX document! No \\use_bibtopic header found!") return document.header[k] = "\\use_bibtopic true" # Possible units. This assumes that the LyX name follows the std, # which might not always be the case. But it's as good as we can get. units = { "part": "Part", "chapter": "Chapter", "section": "Section", "subsection": "Subsection", } if multibib not in units.keys(): document.warning("Unknown multibib value `%s'!" % multibib) return unit = units[multibib] btunit = False i = 0 while True: i = find_token(document.body, "\\begin_layout " + unit, i) if i == -1: break if btunit: document.body[i - 1 : i - 1] = [ "\\begin_layout Standard", "\\begin_inset ERT", "status open", "", "\\begin_layout Plain Layout", "", "", "\\backslash", "end{btUnit}", "\\end_layout", "\\begin_layout Plain Layout", "", "\\backslash", "begin{btUnit}" "\\end_layout", "", "\\end_inset", "", "", "\\end_layout", "", ] i += 21 else: document.body[i - 1 : i - 1] = [ "\\begin_layout Standard", "\\begin_inset ERT", "status open", "", "\\begin_layout Plain Layout", "", "", "\\backslash", "begin{btUnit}" "\\end_layout", "", "\\end_inset", "", "", "\\end_layout", "", ] i += 16 btunit = True i += 1 if btunit: i = find_token(document.body, "\\end_body", i) document.body[i - 1 : i - 1] = [ "\\begin_layout Standard", "\\begin_inset ERT", "status open", "", "\\begin_layout Plain Layout", "", "", "\\backslash", "end{btUnit}" "\\end_layout", "", "\\end_inset", "", "", "\\end_layout", "", ] def revert_chapterbib(document): "Revert chapterbib support" # 1. Get cite engine engine = "basic" i = find_token(document.header, "\\cite_engine", 0) if i == -1: document.warning("Malformed document! Missing \\cite_engine") else: engine = get_value(document.header, "\\cite_engine", i) # 2. Do we use biblatex? biblatex = False if engine in ["biblatex", "biblatex-natbib"]: biblatex = True # 3. Store multibib document header value multibib = "" i = find_token(document.header, "\\multibib", 0) if i != -1: multibib = get_value(document.header, "\\multibib", i) if not multibib or multibib != "child": # nothing to do return # 4. remove multibib header del document.header[i] # 5. Biblatex if biblatex: # find include insets i = 0 while True: i = find_token(document.body, "\\begin_inset CommandInset include", i) if i == -1: break j = find_end_of_inset(document.body, i) if j == -1: document.warning("Can't find end of bibtex inset at line %d!!" % (i)) i += 1 continue parent = get_containing_layout(document.body, i) parbeg = parent[1] # Insert ERT \\newrefsection before inset beg = [ "\\begin_layout Standard", "\\begin_inset ERT", "status open", "", "\\begin_layout Plain Layout", "", "", "\\backslash", "newrefsection" "\\end_layout", "", "\\end_inset", "", "", "\\end_layout", "", ] document.body[parbeg - 1 : parbeg - 1] = beg j += len(beg) i = j + 1 return # 6. Bibtex/Bibtopic i = find_token(document.header, "\\use_bibtopic", 0) if i == -1: # this should not happen document.warning("Malformed LyX document! No \\use_bibtopic header found!") return if get_value(document.header, "\\use_bibtopic", i) == "true": # find include insets i = 0 while True: i = find_token(document.body, "\\begin_inset CommandInset include", i) if i == -1: break j = find_end_of_inset(document.body, i) if j == -1: document.warning("Can't find end of bibtex inset at line %d!!" % (i)) i += 1 continue parent = get_containing_layout(document.body, i) parbeg = parent[1] parend = parent[2] # Insert wrap inset into \\begin{btUnit}...\\end{btUnit} beg = [ "\\begin_layout Standard", "\\begin_inset ERT", "status open", "", "\\begin_layout Plain Layout", "", "", "\\backslash", "begin{btUnit}" "\\end_layout", "", "\\end_inset", "", "", "\\end_layout", "", ] end = [ "\\begin_layout Standard", "\\begin_inset ERT", "status open", "", "\\begin_layout Plain Layout", "", "", "\\backslash", "end{btUnit}" "\\end_layout", "", "\\end_inset", "", "", "\\end_layout", "", ] document.body[parend + 1 : parend + 1] = end document.body[parbeg - 1 : parbeg - 1] = beg j += len(beg) + len(end) i = j + 1 return # 7. Chapterbib proper add_to_preamble(document, ["\\usepackage{chapterbib}"]) def convert_dashligatures(document): """Set 'use_dash_ligatures' according to content.""" # Look for and remove dashligatures workaround from 2.3->2.2 reversion, # set use_dash_ligatures to True if found, to None else. use_dash_ligatures = ( del_complete_lines( document.preamble, [ "% Added by lyx2lyx", r"\renewcommand{\textendash}{--}", r"\renewcommand{\textemdash}{---}", ], ) or None ) if use_dash_ligatures is None: # Look for dashes (Documents by LyX 2.1 or older have "\twohyphens\n" # or "\threehyphens\n" as interim representation for -- an ---.) lines = document.body has_literal_dashes = has_ligature_dashes = False dash_pattern = re.compile(".*[\u2013\u2014]|\\twohyphens|\\threehyphens") i = j = 0 while True: # skip lines without dashes: i = find_re(lines, dash_pattern, i + 1) if i == -1: break line = lines[i] # skip label width string (see bug 10243): if line.startswith("\\labelwidthstring"): continue # do not touch hyphens in some insets (cf. lyx_2_2.convert_dashes): try: inset_type, start, end = get_containing_inset(lines, i) except TypeError: # no containing inset inset_type, start, end = "no inset", -1, -1 if ( inset_type.split()[0] in [ "CommandInset", "ERT", "External", "Formula", "FormulaMacro", "Graphics", "IPA", "listings", ] or inset_type == "Flex Code" ): i = end continue try: layoutname, start, end, j = get_containing_layout(lines, i) except TypeError: # no (or malformed) containing layout document.warning("Malformed LyX document: " "Can't find layout at line %d" % i) continue if not layoutname: document.warning( "Malformed LyX document: " "Missing layout name on line %d" % start ) if layoutname == "LyX-Code": i = end continue # literal dash followed by a non-white-character or no-break space: if re.search("[\u2013\u2014]([\\S\u00a0\u202f\u2060]|$)", line, flags=re.UNICODE): has_literal_dashes = True # ligature dash followed by non-white-char or no-break space on next line: if re.search(r"(\\twohyphens|\\threehyphens)", line) and re.match( "[\\S\u00a0\u202f\u2060]", lines[i + 1], flags=re.UNICODE ): has_ligature_dashes = True if has_literal_dashes and has_ligature_dashes: # TODO: insert a warning note in the document? document.warning( "This document contained both literal and " '"ligature" dashes.\n Line breaks may have changed. ' "See UserGuide chapter 3.9.1 for details." ) break if has_literal_dashes and not has_ligature_dashes: use_dash_ligatures = False elif has_ligature_dashes and not has_literal_dashes: use_dash_ligatures = True # insert the setting if there is a preferred value if use_dash_ligatures is True: document.header.insert(-1, "\\use_dash_ligatures true") elif use_dash_ligatures is False: document.header.insert(-1, "\\use_dash_ligatures false") def revert_dashligatures(document): """Remove font ligature settings for en- and em-dashes. Revert conversion of \twodashes or \threedashes to literal dashes. """ use_dash_ligatures = del_value(document.header, "\\use_dash_ligatures") if use_dash_ligatures != "true" or document.backend != "latex": return i = 0 dash_pattern = re.compile(".*[\u2013\u2014]") while True: # skip lines without dashes: i = find_re(document.body, dash_pattern, i + 1) if i == -1: break line = document.body[i] # skip label width string (see bug 10243): if line.startswith("\\labelwidthstring"): continue # do not touch hyphens in some insets (cf. lyx_2_2.convert_dashes): try: inset_type, start, end = get_containing_inset(document.body, i) except TypeError: # no containing inset inset_type, start, end = "no inset", -1, -1 if ( inset_type.split()[0] in [ "CommandInset", "ERT", "External", "Formula", "FormulaMacro", "Graphics", "IPA", "listings", ] or inset_type == "Flex Code" ): i = end continue try: layoutname, start, end, j = get_containing_layout(document.body, i) except TypeError: # no (or malformed) containing layout document.warning("Malformed LyX document: " "Can't find layout at body line %d" % i) continue if layoutname == "LyX-Code": i = end continue # TODO: skip replacement in typewriter fonts line = line.replace("\u2013", "\\twohyphens\n") line = line.replace("\u2014", "\\threehyphens\n") document.body[i : i + 1] = line.split("\n") # redefine the dash LICRs to use ligature dashes: add_to_preamble( document, [r"\renewcommand{\textendash}{--}", r"\renewcommand{\textemdash}{---}"], ) def revert_noto(document): "Revert Noto font definitions to LaTeX" if find_token(document.header, "\\use_non_tex_fonts false", 0) != -1: preamble = "" i = find_token(document.header, '\\font_roman "NotoSerif-TLF"', 0) if i != -1: add_to_preamble(document, ["\\renewcommand{\\rmdefault}{NotoSerif-TLF}"]) document.header[i] = document.header[i].replace("NotoSerif-TLF", "default") i = find_token(document.header, '\\font_sans "NotoSans-TLF"', 0) if i != -1: add_to_preamble(document, ["\\renewcommand{\\sfdefault}{NotoSans-TLF}"]) document.header[i] = document.header[i].replace("NotoSans-TLF", "default") i = find_token(document.header, '\\font_typewriter "NotoMono-TLF"', 0) if i != -1: add_to_preamble(document, ["\\renewcommand{\\ttdefault}{NotoMono-TLF}"]) document.header[i] = document.header[i].replace("NotoMono-TLF", "default") def revert_xout(document): "Reverts \\xout font attribute" changed = revert_font_attrs(document.body, "\\xout", "\\xout") if changed == True: insert_to_preamble( document, [ "% for proper cross-out", "\\PassOptionsToPackage{normalem}{ulem}", "\\usepackage{ulem}", ], ) def convert_mathindent(document): """Add the \\is_math_indent tag.""" k = find_token(document.header, "\\quotes_style") # where to insert # check if the document uses the class option "fleqn" options = get_value(document.header, "\\options") if "fleqn" in options: document.header.insert(k, "\\is_math_indent 1") # delete the fleqn option i = find_token(document.header, "\\options") options = [option for option in options.split(",") if option.strip() != "fleqn"] if options: document.header[i] = "\\options " + ",".join(options) else: del document.header[i] else: document.header.insert(k, "\\is_math_indent 0") def revert_mathindent(document): "Define mathindent if set in the document" # emulate and delete \math_indentation value = get_value(document.header, "\\math_indentation", default="default", delete=True) if value != "default": add_to_preamble(document, [r"\setlength{\mathindent}{%s}" % value]) # delete \is_math_indent and emulate via document class option if not get_bool_value(document.header, "\\is_math_indent", delete=True): return i = find_token(document.header, "\\options") if i != -1: document.header[i] = document.header[i].replace("\\options ", "\\options fleqn,") else: l = find_token(document.header, "\\use_default_options") document.header.insert(l, "\\options fleqn") def revert_baselineskip(document): "Revert baselineskips to TeX code" i = 0 while True: i = find_substring(document.body, "baselineskip%", i + 1) if i == -1: return if document.body[i].startswith("\\begin_inset VSpace"): # output VSpace inset as TeX code end = find_end_of_inset(document.body, i) if end == -1: document.warning( "Malformed LyX document: " "Can't find end of VSpace inset at line %d." % i ) continue # read out the value baselineskip = document.body[i].split()[-1] # check if it is the starred version star = "*" if "*" in document.body[i] else "" # now output TeX code cmd = f"\\vspace{star}{{{latex_length(baselineskip)[1]}}}" document.body[i : end + 1] = put_cmd_in_ert(cmd) i += 8 continue begin, end = is_in_inset(document.body, i, "\\begin_inset space \\hspace") if begin != -1: # output space inset as TeX code baselineskip = document.body[i].split()[-1] star = "*" if "*" in document.body[i - 1] else "" cmd = f"\\hspace{star}{{{latex_length(baselineskip)[1]}}}" document.body[begin : end + 1] = put_cmd_in_ert(cmd) def revert_rotfloat(document): "Revert placement options for rotated floats" i = 0 j = 0 k = 0 while True: i = find_token(document.body, "sideways true", i) if i == -1: return if not document.body[i - 2].startswith("placement "): i = i + 1 continue # we found a sideways float with placement options # at first store the placement beg = document.body[i - 2].rfind(" ") placement = document.body[i - 2][beg + 1 :] # check if the option'H' is used if placement.find("H") != -1: add_to_preamble(document, ["\\usepackage{float}"]) # now check if it is a starred type if document.body[i - 1].find("wide true") != -1: star = "*" else: star = "" # store the float type beg = document.body[i - 3].rfind(" ") fType = document.body[i - 3][beg + 1 :] # now output TeX code endInset = find_end_of_inset(document.body, i - 3) if endInset == -1: document.warning("Malformed LyX document: Missing '\\end_inset' of Float inset.") return else: document.body[endInset - 2 : endInset + 1] = put_cmd_in_ert( "\\end{sideways" + fType + star + "}" ) document.body[i - 3 : i + 2] = put_cmd_in_ert( "\\begin{sideways" + fType + star + "}[" + placement + "]" ) add_to_preamble(document, ["\\usepackage{rotfloat}"]) i = i + 1 allowbreak_emulation = [ r"\begin_inset space \hspace{}", r"\length 0dd", r"\end_inset", r"", ] def convert_allowbreak(document): r"Zero widths Space-inset -> \SpecialChar allowbreak." lines = document.body i = find_complete_lines(lines, allowbreak_emulation, 2) while i != -1: lines[i - 1 : i + 4] = [lines[i - 1] + r"\SpecialChar allowbreak"] i = find_complete_lines(lines, allowbreak_emulation, i + 3) def revert_allowbreak(document): r"\SpecialChar allowbreak -> Zero widths Space-inset." i = 1 lines = document.body while i < len(lines): if lines[i].endswith(r"\SpecialChar allowbreak"): lines[i : i + 1] = [ lines[i].replace(r"\SpecialChar allowbreak", "") ] + allowbreak_emulation i += 5 else: i += 1 def convert_mathnumberpos(document): "add the \\math_number_before tag" i = find_token(document.header, "\\quotes_style") # check if the document uses the class option "leqno" if is_document_option(document, "leqno"): remove_document_option(document, "leqno") document.header.insert(i, "\\math_number_before 1") else: document.header.insert(i, "\\math_number_before 0") def revert_mathnumberpos(document): """Remove \\math_number_before tag, add the document class option leqno if required. """ math_number_before = get_bool_value(document.header, "\\math_number_before", delete=True) if math_number_before: insert_document_option(document, "leqno") def convert_mathnumberingname(document): "rename the \\math_number_before tag to \\math_numbering_side" i = find_token(document.header, "\\math_number_before") math_number_before = get_bool_value(document.header, "\\math_number_before", i) if math_number_before: document.header[i] = "\\math_numbering_side left" return # check if the document uses the class option "reqno" k = find_token(document.header, "\\options") if "reqno" in document.header[k]: document.header[i] = "\\math_numbering_side right" # delete the found option document.header[k] = document.header[k].replace(",reqno", "") document.header[k] = document.header[k].replace(", reqno", "") document.header[k] = document.header[k].replace("reqno,", "") if "reqno" in document.header[k]: # then we have reqno as the only option del document.header[k] else: document.header[i] = "\\math_numbering_side default" def revert_mathnumberingname(document): "rename the \\math_numbering_side tag back to \\math_number_before" i = find_token(document.header, "\\math_numbering_side") math_numbering_side = get_value(document.header, "\\math_numbering_side", i) # rename tag and set boolean value: if math_numbering_side == "left": document.header[i] = "\\math_number_before 1" elif math_numbering_side == "right": # also add the option reqno: document.header[i] = "\\math_number_before 0" k = find_token(document.header, "\\options") if k != -1 and "reqno" not in document.header[k]: document.header[k] = document.header[k].replace("\\options", "\\options reqno,") else: l = find_token(document.header, "\\use_default_options", 0) document.header.insert(l, "\\options reqno") else: document.header[i] = "\\math_number_before 0" def convert_minted(document): "add the \\use_minted tag" i = find_token(document.header, "\\index ") document.header.insert(i, "\\use_minted 0") def revert_minted(document): "remove the \\use_minted tag" del_token(document.header, "\\use_minted") def revert_longtable_lscape(document): "revert the longtable landcape mode to ERT" i = 0 regexp = re.compile(r"^