mirror of
https://git.lyx.org/repos/lyx.git
synced 2024-11-27 03:36:39 +00:00
Copy lyx2lyx files from 2.3.x.
Because of extensive changes in the support files, it seems to me that we actually do need to copy over everything.
This commit is contained in:
parent
4607b84bd0
commit
c2ac6440f7
@ -28,15 +28,21 @@ import locale
|
||||
import sys
|
||||
import re
|
||||
import time
|
||||
import io
|
||||
import codecs
|
||||
|
||||
try:
|
||||
import lyx2lyx_version
|
||||
version__ = lyx2lyx_version.version
|
||||
except: # we are running from build directory so assume the last version
|
||||
version__ = '2.2'
|
||||
version__ = '2.3'
|
||||
|
||||
default_debug__ = 2
|
||||
|
||||
# Provide support for both python 2 and 3
|
||||
PY2 = sys.version_info[0] == 2
|
||||
# End of code to support for both python 2 and 3
|
||||
|
||||
####################################################################
|
||||
# Private helper functions
|
||||
|
||||
@ -86,7 +92,8 @@ format_relation = [("0_06", [200], minor_versions("0.6" , 4)),
|
||||
("1_6", list(range(277,346)), minor_versions("1.6" , 10)),
|
||||
("2_0", list(range(346,414)), minor_versions("2.0" , 8)),
|
||||
("2_1", list(range(414,475)), minor_versions("2.1" , 5)),
|
||||
("2_2", list(range(475,509)), minor_versions("2.2" , 0))
|
||||
("2_2", list(range(475,509)), minor_versions("2.2" , 0)),
|
||||
("2_3", (), minor_versions("2.3" , 0))
|
||||
]
|
||||
|
||||
####################################################################
|
||||
@ -157,6 +164,17 @@ def trim_eol(line):
|
||||
return line[:-1]
|
||||
|
||||
|
||||
def trim_eol_binary(line):
|
||||
" Remove end of line char(s)."
|
||||
if line[-1] != 10 and line[-1] != 13:
|
||||
# May happen for the last line of a document
|
||||
return line
|
||||
if line[-2:-1] == 13:
|
||||
return line[:-2]
|
||||
else:
|
||||
return line[:-1]
|
||||
|
||||
|
||||
def get_encoding(language, inputencoding, format, cjk_encoding):
|
||||
" Returns enconding of the lyx file"
|
||||
if format > 248:
|
||||
@ -187,10 +205,10 @@ def get_encoding(language, inputencoding, format, cjk_encoding):
|
||||
class LyX_base:
|
||||
"""This class carries all the information of the LyX file."""
|
||||
|
||||
def __init__(self, end_format = 0, input = "", output = "", error = "",
|
||||
debug = default_debug__, try_hard = 0, cjk_encoding = '',
|
||||
final_version = "", systemlyxdir = '', language = "english",
|
||||
encoding = "auto"):
|
||||
def __init__(self, end_format = 0, input = u'', output = u'', error = u'',
|
||||
debug = default_debug__, try_hard = 0, cjk_encoding = u'',
|
||||
final_version = u'', systemlyxdir = u'', language = u'english',
|
||||
encoding = u'auto'):
|
||||
|
||||
"""Arguments:
|
||||
end_format: final format that the file should be converted. (integer)
|
||||
@ -199,7 +217,8 @@ class LyX_base:
|
||||
error: the name of the error file, if empty use the standard error.
|
||||
debug: debug level, O means no debug, as its value increases be more verbose.
|
||||
"""
|
||||
self.choose_io(input, output)
|
||||
self.choose_input(input)
|
||||
self.output = output
|
||||
|
||||
if error:
|
||||
self.err = open(error, "w")
|
||||
@ -279,23 +298,50 @@ class LyX_base:
|
||||
"""Reads a file into the self.header and
|
||||
self.body parts, from self.input."""
|
||||
|
||||
# First pass: Read header to determine file encoding
|
||||
# If we are running under python3 then all strings are binary in this
|
||||
# pass. In some cases we need to convert binary to unicode in order to
|
||||
# use our parser tools. Since we do not know the true encoding yet we
|
||||
# use latin1. This works since a) the parts we are interested in are
|
||||
# pure ASCII (subset of latin1) and b) in contrast to pure ascii or
|
||||
# utf8, one can decode any 8byte string using latin1.
|
||||
first_line = True
|
||||
while True:
|
||||
line = self.input.readline()
|
||||
if not line:
|
||||
self.error("Invalid LyX file.")
|
||||
# eof found before end of header
|
||||
self.error("Invalid LyX file: Missing body.")
|
||||
|
||||
line = trim_eol(line)
|
||||
if check_token(line, '\\begin_preamble'):
|
||||
while 1:
|
||||
if first_line:
|
||||
# Remove UTF8 BOM marker if present
|
||||
if line.startswith(codecs.BOM_UTF8):
|
||||
line = line[len(codecs.BOM_UTF8):]
|
||||
|
||||
first_line = False
|
||||
|
||||
if PY2:
|
||||
line = trim_eol(line)
|
||||
decoded = line
|
||||
else:
|
||||
line = trim_eol_binary(line)
|
||||
decoded = line.decode('latin1')
|
||||
if check_token(decoded, '\\begin_preamble'):
|
||||
while True:
|
||||
line = self.input.readline()
|
||||
if not line:
|
||||
self.error("Invalid LyX file.")
|
||||
# eof found before end of header
|
||||
self.error("Invalid LyX file: Missing body.")
|
||||
|
||||
line = trim_eol(line)
|
||||
if check_token(line, '\\end_preamble'):
|
||||
if PY2:
|
||||
line = trim_eol(line)
|
||||
decoded = line
|
||||
else:
|
||||
line = trim_eol_binary(line)
|
||||
decoded = line.decode('latin1')
|
||||
if check_token(decoded, '\\end_preamble'):
|
||||
break
|
||||
|
||||
if line.split()[:0] in ("\\layout",
|
||||
if decoded.split()[:0] in ("\\layout",
|
||||
"\\begin_layout", "\\begin_body"):
|
||||
|
||||
self.warning("Malformed LyX file:"
|
||||
@ -305,33 +351,49 @@ class LyX_base:
|
||||
|
||||
self.preamble.append(line)
|
||||
|
||||
if check_token(line, '\\end_preamble'):
|
||||
if check_token(decoded, '\\end_preamble'):
|
||||
continue
|
||||
|
||||
line = line.rstrip()
|
||||
if not line:
|
||||
continue
|
||||
|
||||
if line.split()[0] in ("\\layout", "\\begin_layout",
|
||||
if decoded.split()[0] in ("\\layout", "\\begin_layout",
|
||||
"\\begin_body", "\\begin_deeper"):
|
||||
self.body.append(line)
|
||||
break
|
||||
|
||||
self.header.append(line)
|
||||
|
||||
i = find_token(self.header, '\\textclass', 0)
|
||||
if PY2:
|
||||
i = find_token(self.header, '\\textclass', 0)
|
||||
else:
|
||||
i = find_token(self.header, b'\\textclass', 0)
|
||||
if i == -1:
|
||||
self.warning("Malformed LyX file: Missing '\\textclass'.")
|
||||
i = find_token(self.header, '\\lyxformat', 0) + 1
|
||||
self.header[i:i] = ['\\textclass article']
|
||||
if PY2:
|
||||
i = find_token(self.header, '\\lyxformat', 0) + 1
|
||||
self.header[i:i] = ['\\textclass article']
|
||||
else:
|
||||
i = find_token(self.header, b'\\lyxformat', 0) + 1
|
||||
self.header[i:i] = [b'\\textclass article']
|
||||
|
||||
self.textclass = get_value(self.header, "\\textclass", 0)
|
||||
self.backend = get_backend(self.textclass)
|
||||
self.format = self.read_format()
|
||||
self.language = get_value(self.header, "\\language", 0,
|
||||
default = "english")
|
||||
self.inputencoding = get_value(self.header, "\\inputencoding",
|
||||
0, default = "auto")
|
||||
if PY2:
|
||||
self.textclass = get_value(self.header, "\\textclass", 0,
|
||||
default = "")
|
||||
self.language = get_value(self.header, "\\language", 0,
|
||||
default = "english")
|
||||
self.inputencoding = get_value(self.header, "\\inputencoding", 0,
|
||||
default = "auto")
|
||||
else:
|
||||
self.textclass = get_value(self.header, b"\\textclass", 0,
|
||||
default = b"")
|
||||
self.language = get_value(self.header, b"\\language", 0,
|
||||
default = b"english").decode('ascii')
|
||||
self.inputencoding = get_value(self.header, b"\\inputencoding", 0,
|
||||
default = b"auto").decode('ascii')
|
||||
self.format = self.read_format()
|
||||
self.initial_format = self.format
|
||||
self.encoding = get_encoding(self.language,
|
||||
self.inputencoding, self.format,
|
||||
self.cjk_encoding)
|
||||
@ -340,13 +402,16 @@ class LyX_base:
|
||||
# Second pass over header and preamble, now we know the file encoding
|
||||
# Do not forget the textclass (Debian bug #700828)
|
||||
self.textclass = self.textclass.decode(self.encoding)
|
||||
self.backend = get_backend(self.textclass)
|
||||
for i in range(len(self.header)):
|
||||
self.header[i] = self.header[i].decode(self.encoding)
|
||||
for i in range(len(self.preamble)):
|
||||
self.preamble[i] = self.preamble[i].decode(self.encoding)
|
||||
for i in range(len(self.body)):
|
||||
self.body[i] = self.body[i].decode(self.encoding)
|
||||
|
||||
# Read document body
|
||||
while 1:
|
||||
while True:
|
||||
line = self.input.readline().decode(self.encoding)
|
||||
if not line:
|
||||
break
|
||||
@ -355,6 +420,7 @@ class LyX_base:
|
||||
|
||||
def write(self):
|
||||
" Writes the LyX file to self.output."
|
||||
self.choose_output(self.output)
|
||||
self.set_version()
|
||||
self.set_format()
|
||||
self.set_textclass()
|
||||
@ -369,29 +435,53 @@ class LyX_base:
|
||||
header = self.header
|
||||
|
||||
for line in header + [''] + self.body:
|
||||
self.output.write(line.encode(self.encoding)+"\n")
|
||||
self.output.write(line+u"\n")
|
||||
|
||||
|
||||
def choose_io(self, input, output):
|
||||
"""Choose input and output streams, dealing transparently with
|
||||
def choose_output(self, output):
|
||||
"""Choose output streams dealing transparently with
|
||||
compressed files."""
|
||||
|
||||
if output:
|
||||
self.output = open(output, "wb")
|
||||
# This is a bit complicated, because we need to be compatible both with
|
||||
# python 2 and python 3. Therefore we handle the encoding here and not
|
||||
# when writing individual lines and may need up to 3 layered file like
|
||||
# interfaces.
|
||||
if self.compressed:
|
||||
if output:
|
||||
outputfileobj = open(output, 'wb')
|
||||
else:
|
||||
# We cannot not use stdout directly since it needs text, not bytes in python 3
|
||||
outputfileobj = os.fdopen(sys.stdout.fileno(), 'wb')
|
||||
# We cannot not use gzip.open() since it is not supported by python 2
|
||||
zipbuffer = gzip.GzipFile(mode='wb', fileobj=outputfileobj)
|
||||
# We do not want to use different newlines on different OSes inside zipped files
|
||||
self.output = io.TextIOWrapper(zipbuffer, encoding=self.encoding, newline='\n')
|
||||
else:
|
||||
self.output = sys.stdout
|
||||
if output:
|
||||
self.output = io.open(output, 'w', encoding=self.encoding)
|
||||
else:
|
||||
self.output = io.open(sys.stdout.fileno(), 'w', encoding=self.encoding)
|
||||
|
||||
if input and input != '-':
|
||||
|
||||
def choose_input(self, input):
|
||||
"""Choose input stream, dealing transparently with
|
||||
compressed files."""
|
||||
|
||||
# Since we do not know the encoding yet we need to read the input as
|
||||
# bytes in binary mode, and convert later to unicode.
|
||||
if input and input != u'-':
|
||||
self.dir = os.path.dirname(os.path.abspath(input))
|
||||
try:
|
||||
gzip.open(input).readline()
|
||||
self.input = gzip.open(input)
|
||||
self.output = gzip.GzipFile(mode="wb", fileobj=self.output)
|
||||
self.compressed = True
|
||||
except:
|
||||
self.input = open(input)
|
||||
self.input = open(input, 'rb')
|
||||
self.compressed = False
|
||||
else:
|
||||
self.dir = ''
|
||||
self.input = sys.stdin
|
||||
self.dir = u''
|
||||
self.input = os.fdopen(sys.stdin.fileno(), 'rb')
|
||||
self.compressed = False
|
||||
|
||||
|
||||
def lyxformat(self, format):
|
||||
@ -464,11 +554,14 @@ class LyX_base:
|
||||
def read_format(self):
|
||||
" Read from the header the fileformat of the present LyX file."
|
||||
for line in self.header:
|
||||
result = fileformat.match(line)
|
||||
if PY2:
|
||||
result = fileformat.match(line)
|
||||
else:
|
||||
result = fileformat.match(line.decode('ascii'))
|
||||
if result:
|
||||
return self.lyxformat(result.group(1))
|
||||
else:
|
||||
self.error("Invalid LyX File.")
|
||||
self.error("Invalid LyX File: Missing format.")
|
||||
return None
|
||||
|
||||
|
||||
@ -606,7 +699,6 @@ class LyX_base:
|
||||
conversion are taken. It returns a list of modules needed to
|
||||
convert the LyX file from self.format to self.end_format"""
|
||||
|
||||
self.start = self.format
|
||||
format = self.format
|
||||
correct_version = 0
|
||||
|
||||
@ -640,7 +732,7 @@ class LyX_base:
|
||||
|
||||
# Convertion mode, back or forth
|
||||
steps = []
|
||||
if (initial_step, self.start) < (final_step, self.end_format):
|
||||
if (initial_step, self.initial_format) < (final_step, self.end_format):
|
||||
mode = "convert"
|
||||
full_steps = []
|
||||
for step in format_relation:
|
||||
@ -684,7 +776,7 @@ class LyX_base:
|
||||
|
||||
# toc_par = []
|
||||
# i = 0
|
||||
# while 1:
|
||||
# while True:
|
||||
# i = find_tokens(self.body, sections, i)
|
||||
# if i == -1:
|
||||
# break
|
||||
@ -734,9 +826,9 @@ class LyX_base:
|
||||
class File(LyX_base):
|
||||
" This class reads existing LyX files."
|
||||
|
||||
def __init__(self, end_format = 0, input = "", output = "", error = "",
|
||||
debug = default_debug__, try_hard = 0, cjk_encoding = '',
|
||||
final_version = '', systemlyxdir = ''):
|
||||
def __init__(self, end_format = 0, input = u'', output = u'', error = u'',
|
||||
debug = default_debug__, try_hard = 0, cjk_encoding = u'',
|
||||
final_version = u'', systemlyxdir = u''):
|
||||
LyX_base.__init__(self, end_format, input, output, error,
|
||||
debug, try_hard, cjk_encoding, final_version,
|
||||
systemlyxdir)
|
||||
|
@ -33,6 +33,7 @@ dist_lyx2lyx_PYTHON = \
|
||||
lyx_2_0.py \
|
||||
lyx_2_1.py \
|
||||
lyx_2_2.py \
|
||||
lyx_2_3.py \
|
||||
profiling.py \
|
||||
test_parser_tools.py
|
||||
|
||||
|
@ -19,58 +19,66 @@
|
||||
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
||||
|
||||
" Program used to convert between different versions of the lyx file format."
|
||||
import optparse
|
||||
import argparse
|
||||
import sys
|
||||
import LyX
|
||||
|
||||
# Provide support for both python 2 and 3
|
||||
PY2 = sys.version_info[0] == 2
|
||||
if PY2:
|
||||
# argparse returns strings in the commandline encoding, we need to convert.
|
||||
# sys.getdefaultencoding() would not always be correct, see
|
||||
# http://legacy.python.org/dev/peps/pep-0383/
|
||||
def cmd_arg(arg):
|
||||
return arg.decode(sys.getfilesystemencoding())
|
||||
else:
|
||||
cmd_arg = str
|
||||
# End of code to support for both python 2 and 3
|
||||
|
||||
def main():
|
||||
args = {}
|
||||
args["usage"] = "usage: %prog [options] [file]"
|
||||
|
||||
args["version"] = """lyx2lyx, version %s
|
||||
Copyright (C) 2011 The LyX Team, José Matos and Dekel Tsur""" % LyX.version__
|
||||
args["usage"] = "%(prog)s [options] [file]"
|
||||
|
||||
args["description"] = """Convert old lyx file <file> to newer format,
|
||||
files can be compressed with gzip. If there no file is specified then
|
||||
the standard input is assumed, in this case gziped files are not
|
||||
handled."""
|
||||
|
||||
parser = optparse.OptionParser(**args)
|
||||
parser = argparse.ArgumentParser(**args)
|
||||
|
||||
parser.set_defaults(debug=LyX.default_debug__, cjk_encoding = '')
|
||||
parser.add_option("-d", "--debug", type="int",
|
||||
parser.add_argument("-d", "--debug", type=int, dest="debug",
|
||||
help="level=0..2 (O_ quiet, 10_verbose) default: 2")
|
||||
parser.add_option("-q", "--quiet",
|
||||
parser.add_argument("-q", "--quiet",
|
||||
action="store_const", const=0, dest="debug")
|
||||
parser.add_option("-v", "--verbose",
|
||||
parser.add_argument("-v", "--verbose",
|
||||
action="store_const", const=1, dest="debug")
|
||||
parser.add_option("--noisy",
|
||||
parser.add_argument("--noisy",
|
||||
action="store_const", const=10, dest="debug")
|
||||
parser.add_option("-c", "--encoding", dest="cjk_encoding",
|
||||
parser.add_argument("-c", "--encoding", type=cmd_arg, dest="cjk_encoding",
|
||||
help="files in format 413 and lower are read and"
|
||||
" written in the format of CJK-LyX."
|
||||
"If encoding is not given or 'auto' the encoding"
|
||||
"is determined from the locale.")
|
||||
parser.add_option("-e", "--err", dest="error",
|
||||
parser.add_argument("-e", "--err", type=cmd_arg, dest="error",
|
||||
help= "file name of the error file else goes to stderr")
|
||||
parser.add_option("-o", "--output",
|
||||
parser.add_argument("-o", "--output", type=cmd_arg, dest="output",
|
||||
help= "name of the output file else goes to stdout")
|
||||
parser.add_option("-t", "--to", dest= "end_format",
|
||||
parser.add_argument("-t", "--to", type=cmd_arg, dest= "end_format",
|
||||
help= "destination file format, default (latest)")
|
||||
parser.add_option("-V", "--final_version", dest= "final_version",
|
||||
parser.add_argument("-V", "--final_version", type=cmd_arg, dest= "final_version",
|
||||
help= "destination version, default (latest)")
|
||||
parser.add_option("-l", "--list", action="store_true",
|
||||
parser.add_argument("-l", "--list", action="store_true",
|
||||
help = "list all available formats and supported versions")
|
||||
parser.add_option("-n", "--try-hard", action="store_true",
|
||||
parser.add_argument("-n", "--try-hard", action="store_true",
|
||||
help = "try hard (ignore any convertion errors)")
|
||||
parser.add_option("-s", "--systemlyxdir", dest= "systemlyxdir",
|
||||
parser.add_argument("-s", "--systemlyxdir", type=cmd_arg, dest= "systemlyxdir",
|
||||
help= "LyX system directory for conversion from version 489 or older")
|
||||
parser.add_argument('--version', action='version', version="""lyx2lyx, version %s
|
||||
Copyright (C) 2011 The LyX Team, José Matos and Dekel Tsur""" % LyX.version__)
|
||||
parser.add_argument("input", nargs='?', type=cmd_arg, default=None)
|
||||
|
||||
(options, args) = parser.parse_args()
|
||||
if args:
|
||||
options.input = args[0]
|
||||
else:
|
||||
options.input = None
|
||||
options = parser.parse_args()
|
||||
|
||||
if options.list:
|
||||
sys.stderr.write(LyX.format_info())
|
||||
|
@ -26,7 +26,7 @@ def space_before_layout(document):
|
||||
" Remove empty line before \\layout. "
|
||||
lines = document.body
|
||||
i = 2 # skip first layout
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(lines, '\\layout', i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -41,7 +41,7 @@ def formula_inset_space_eat(document):
|
||||
" Remove space after inset formula."
|
||||
lines = document.body
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(lines, "\\begin_inset Formula", i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -56,7 +56,7 @@ def update_tabular(document):
|
||||
lines = document.body
|
||||
lyxtable_re = re.compile(r".*\\LyXTable$")
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_re(lines, lyxtable_re, i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -108,7 +108,7 @@ def update_inset_label(document):
|
||||
" Update inset Label."
|
||||
lines = document.body
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(lines, '\\begin_inset Label', i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -120,7 +120,7 @@ def update_latexdel(document):
|
||||
" Update inset LatexDel."
|
||||
lines = document.body
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(lines, '\\begin_inset LatexDel', i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -163,7 +163,7 @@ def remove_cursor(document):
|
||||
lines = document.body
|
||||
i = 0
|
||||
cursor_re = re.compile(r'.*(\\cursor \d*)')
|
||||
while 1:
|
||||
while True:
|
||||
i = find_re(lines, cursor_re, i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -176,7 +176,7 @@ def remove_empty_insets(document):
|
||||
" Remove empty insets."
|
||||
lines = document.body
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(lines, '\\begin_inset ', i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -190,7 +190,7 @@ def remove_formula_latex(document):
|
||||
" Remove formula latex."
|
||||
lines = document.body
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(lines, '\\latex formula_latex ', i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -268,7 +268,7 @@ def update_latexaccents(document):
|
||||
" Update latex accent insets."
|
||||
body = document.body
|
||||
i = 1
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(body, '\\i ', i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -297,7 +297,7 @@ def obsolete_latex_title(document):
|
||||
" Replace layout Latex_Title with Title."
|
||||
body = document.body
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(body, '\\layout', i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -313,7 +313,7 @@ def remove_inset_latex(document):
|
||||
body = document.body
|
||||
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(body, '\\begin_inset Latex', i)
|
||||
if i == -1:
|
||||
return
|
||||
|
@ -26,7 +26,7 @@ def obsolete_latex_title(document):
|
||||
|
||||
body = document.body
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(body, '\\layout', i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -43,7 +43,7 @@ def update_tabular(document):
|
||||
lines = document.body
|
||||
lyxtable_re = re.compile(r".*\\LyXTable$")
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_re(lines, lyxtable_re, i)
|
||||
if i == -1:
|
||||
break
|
||||
|
@ -40,7 +40,7 @@ def replace_protected_separator(document):
|
||||
" Replace protected separator. "
|
||||
lines = document.body
|
||||
i=0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(lines, "\\protected_separator", i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -65,7 +65,7 @@ def merge_formula_inset(document):
|
||||
" Merge formula insets. "
|
||||
lines = document.body
|
||||
i=0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(lines, "\\begin_inset Formula", i)
|
||||
if i == -1: break
|
||||
if lines[i+1] in math_env:
|
||||
@ -79,7 +79,7 @@ def update_tabular(document):
|
||||
lines = document.body
|
||||
lyxtable_re = re.compile(r".*\\LyXTable$")
|
||||
i=0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_re(lines, lyxtable_re, i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -111,7 +111,7 @@ def update_toc(document):
|
||||
" Update table of contents. "
|
||||
lines = document.body
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(lines,
|
||||
'\\begin_inset LatexCommand \\tableofcontents', i)
|
||||
if i == -1:
|
||||
@ -158,7 +158,7 @@ def remove_space_in_units(document):
|
||||
|
||||
for margin in margins:
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(lines, margin, i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -174,7 +174,7 @@ def latexdel_getargs(document, i):
|
||||
lines = document.body
|
||||
|
||||
# play safe, clean empty lines
|
||||
while 1:
|
||||
while True:
|
||||
if lines[i]:
|
||||
break
|
||||
del lines[i]
|
||||
@ -191,7 +191,7 @@ def latexdel_getargs(document, i):
|
||||
del lines[i:j + 1]
|
||||
|
||||
# play safe, clean empty lines
|
||||
while 1:
|
||||
while True:
|
||||
if lines[i]:
|
||||
break
|
||||
del lines[i]
|
||||
@ -212,7 +212,7 @@ def update_ref(document):
|
||||
" Update reference inset. "
|
||||
lines = document.body
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(lines, '\\begin_inset LatexCommand', i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -230,7 +230,7 @@ def update_latexdel(document):
|
||||
lines = document.body
|
||||
i = 0
|
||||
latexdel_re = re.compile(r".*\\begin_inset LatexDel")
|
||||
while 1:
|
||||
while True:
|
||||
i = find_re(lines, latexdel_re, i)
|
||||
if i == -1:
|
||||
return
|
||||
|
@ -26,7 +26,7 @@ def update_tabular(document):
|
||||
" Update tabular to version 1 (xml like syntax). "
|
||||
lines = document.body
|
||||
i=0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_re(lines, lyxtable_re, i)
|
||||
if i == -1:
|
||||
break
|
||||
|
@ -39,7 +39,7 @@ def update_tabular(document):
|
||||
regexp = re.compile(r'^\\begin_inset\s+Tabular')
|
||||
lines = document.body
|
||||
i=0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_re(lines, regexp, i)
|
||||
if i == -1:
|
||||
break
|
||||
|
@ -146,7 +146,7 @@ def remove_oldfloat(document):
|
||||
" Change \begin_float .. \end_float into \begin_inset Float .. \end_inset"
|
||||
lines = document.body
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(lines, "\\begin_float", i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -215,7 +215,7 @@ def remove_pextra(document):
|
||||
lines = document.body
|
||||
i = 0
|
||||
flag = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_re(lines, pextra_type2_rexp, i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -258,7 +258,7 @@ def remove_pextra(document):
|
||||
j = get_next_paragraph(lines, i, document.format + 1)
|
||||
|
||||
count = 0
|
||||
while 1:
|
||||
while True:
|
||||
# collect more paragraphs to the minipage
|
||||
count = count+1
|
||||
if j == -1 or not check_token(lines[j], "\\layout"):
|
||||
@ -301,12 +301,12 @@ def remove_oldert(document):
|
||||
""]
|
||||
lines = document.body
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_tokens(lines, ["\\latex latex", "\\layout LaTeX"], i)
|
||||
if i == -1:
|
||||
break
|
||||
j = i+1
|
||||
while 1:
|
||||
while True:
|
||||
# \end_inset is for ert inside a tabular cell. The other tokens
|
||||
# are obvious.
|
||||
j = find_tokens(lines, ["\\latex default", "\\layout", "\\begin_inset", "\\end_inset", "\\end_float", "\\the_end"],
|
||||
@ -327,7 +327,7 @@ def remove_oldert(document):
|
||||
new = ['\layout %s' % document.default_layout, "", ""]
|
||||
|
||||
k = i+1
|
||||
while 1:
|
||||
while True:
|
||||
k2 = find_re(lines, ert_rexp, k, j)
|
||||
inset = hfill = specialchar = 0
|
||||
if k2 == -1:
|
||||
@ -399,7 +399,7 @@ def remove_oldert(document):
|
||||
|
||||
# Delete remaining "\latex xxx" tokens
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(lines, "\\latex ", i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -410,7 +410,7 @@ def remove_oldertinset(document):
|
||||
" ERT insert are hidden feature of lyx 1.1.6. This might be removed in the future."
|
||||
lines = document.body
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(lines, "\\begin_inset ERT", i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -445,7 +445,7 @@ def combine_ert(document):
|
||||
" Combine ERT paragraphs."
|
||||
lines = document.body
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(lines, "\\begin_inset ERT", i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -490,7 +490,7 @@ def remove_figinset(document):
|
||||
" Remove figinset."
|
||||
lines = document.body
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(lines, "\\begin_inset Figure", i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -564,7 +564,7 @@ def update_tabular(document):
|
||||
regexp = re.compile(r'^\\begin_inset\s+Tabular')
|
||||
lines = document.body
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_re(lines, regexp, i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -697,7 +697,7 @@ def update_longtables(document):
|
||||
regexp = re.compile(r'^\\begin_inset\s+Tabular')
|
||||
body = document.body
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_re(body, regexp, i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -766,7 +766,7 @@ def fix_oldfloatinset(document):
|
||||
" Figure insert are hidden feature of lyx 1.1.6. This might be removed in the future."
|
||||
lines = document.body
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(lines, "\\begin_inset Float ", i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -780,7 +780,7 @@ def change_listof(document):
|
||||
" Change listof insets."
|
||||
lines = document.body
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(lines, "\\begin_inset LatexCommand \\listof", i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -793,7 +793,7 @@ def change_infoinset(document):
|
||||
" Change info inset."
|
||||
lines = document.body
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(lines, "\\begin_inset Info", i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -824,7 +824,7 @@ def change_header(document):
|
||||
if i == -1:
|
||||
return
|
||||
lines[i+1:i+1] = ["\\use_natbib 0",
|
||||
"\use_numerical_citations 0"]
|
||||
"\\use_numerical_citations 0"]
|
||||
|
||||
|
||||
supported_versions = ["1.2.%d" % i for i in range(5)] + ["1.2"]
|
||||
|
@ -54,7 +54,7 @@ def change_insetgraphics(document):
|
||||
" Change inset Graphics."
|
||||
lines = document.body
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(lines, "\\begin_inset Graphics", i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -113,7 +113,7 @@ def change_tabular(document):
|
||||
" Change tabular."
|
||||
lines = document.body
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(lines, "<column", i)
|
||||
if i == -1:
|
||||
break
|
||||
|
@ -105,7 +105,7 @@ def del_token(lines, token, start, end):
|
||||
def remove_color_default(document):
|
||||
" Remove \color default"
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\color default", i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -179,7 +179,7 @@ def revert_spaces(document):
|
||||
" \InsetSpace ~ -> \SpecialChar ~"
|
||||
regexp = re.compile(r'(.*)(\\InsetSpace\s+)(\S+)')
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_re(document.body, regexp, i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -236,7 +236,7 @@ def revert_eqref(document):
|
||||
"\\begin_inset LatexCommand \\eqref -> ERT"
|
||||
regexp = re.compile(r'^\\begin_inset\s+LatexCommand\s+\\eqref')
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_re(document.body, regexp, i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -264,7 +264,7 @@ def revert_bibtex(document):
|
||||
def remove_insetparent(document):
|
||||
" Remove \lyxparent"
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_inset LatexCommand \\lyxparent", i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -276,7 +276,7 @@ def convert_external(document):
|
||||
external_rexp = re.compile(r'\\begin_inset External ([^,]*),"([^"]*)",')
|
||||
external_header = "\\begin_inset External"
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, external_header, i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -311,7 +311,7 @@ def revert_external_1(document):
|
||||
" Revert inset External."
|
||||
external_header = "\\begin_inset External"
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, external_header, i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -336,7 +336,7 @@ def revert_external_2(document):
|
||||
" Revert inset External. (part II)"
|
||||
draft_token = '\tdraft'
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, '\\begin_inset External', i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -354,7 +354,7 @@ def convert_comment(document):
|
||||
" Convert \\layout comment"
|
||||
i = 0
|
||||
comment = "\\layout Comment"
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, comment, i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -365,7 +365,7 @@ def convert_comment(document):
|
||||
'\\layout %s' % document.default_layout]
|
||||
i = i + 7
|
||||
|
||||
while 1:
|
||||
while True:
|
||||
old_i = i
|
||||
i = find_token(document.body, "\\layout", i)
|
||||
if i == -1:
|
||||
@ -418,7 +418,7 @@ def convert_comment(document):
|
||||
def revert_comment(document):
|
||||
" Revert comments"
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_tokens(document.body, ["\\begin_inset Comment", "\\begin_inset Greyedout"], i)
|
||||
|
||||
if i == -1:
|
||||
@ -437,7 +437,7 @@ def add_end_layout(document):
|
||||
i = i + 1
|
||||
struct_stack = ["\\layout"]
|
||||
|
||||
while 1:
|
||||
while True:
|
||||
i = find_tokens(document.body, ["\\begin_inset", "\\end_inset", "\\layout",
|
||||
"\\begin_deeper", "\\end_deeper", "\\the_end"], i)
|
||||
|
||||
@ -504,7 +504,7 @@ def add_end_layout(document):
|
||||
def rm_end_layout(document):
|
||||
" Remove \end_layout"
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, '\\end_layout', i)
|
||||
|
||||
if i == -1:
|
||||
@ -535,7 +535,7 @@ def rm_tracking_changes(document):
|
||||
def rm_body_changes(document):
|
||||
" Remove body changes."
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\change_", i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -546,7 +546,7 @@ def rm_body_changes(document):
|
||||
def layout2begin_layout(document):
|
||||
" \layout -> \begin_layout "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, '\\layout', i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -558,7 +558,7 @@ def layout2begin_layout(document):
|
||||
def begin_layout2layout(document):
|
||||
" \begin_layout -> \layout "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, '\\begin_layout', i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -578,7 +578,7 @@ def convert_table_valignment_middle(document):
|
||||
" Convert table valignment, center -> middle"
|
||||
regexp = re.compile(r'^\\begin_inset\s+Tabular')
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_re(document.body, regexp, i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -602,7 +602,7 @@ def revert_valignment_middle(document):
|
||||
" Convert table valignment, middle -> center"
|
||||
regexp = re.compile(r'^\\begin_inset\s+Tabular')
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_re(document.body, regexp, i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -677,7 +677,7 @@ parskip}
|
||||
attribute_values = ['default', 'default', 'default', 'default',
|
||||
'default', 'default', 'default', 'none', document.language]
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_layout", i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -835,7 +835,7 @@ parskip}
|
||||
def convert_note(document):
|
||||
" Convert Notes. "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_tokens(document.body, ["\\begin_inset Note",
|
||||
"\\begin_inset Comment",
|
||||
"\\begin_inset Greyedout"], i)
|
||||
@ -850,7 +850,7 @@ def revert_note(document):
|
||||
" Revert Notes. "
|
||||
note_header = "\\begin_inset Note "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, note_header, i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -862,7 +862,7 @@ def revert_note(document):
|
||||
def convert_box(document):
|
||||
" Convert Boxes. "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_tokens(document.body, ["\\begin_inset Boxed",
|
||||
"\\begin_inset Doublebox",
|
||||
"\\begin_inset Frameless",
|
||||
@ -880,7 +880,7 @@ def revert_box(document):
|
||||
" Revert Boxes."
|
||||
box_header = "\\begin_inset Box "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, box_header, i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -889,10 +889,10 @@ def revert_box(document):
|
||||
i = i + 1
|
||||
|
||||
|
||||
def convert_collapsable(document):
|
||||
def convert_collapsible(document):
|
||||
" Convert collapsed insets. "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_tokens_exact(document.body, ["\\begin_inset Box",
|
||||
"\\begin_inset Branch",
|
||||
"\\begin_inset CharStyle",
|
||||
@ -909,7 +909,7 @@ def convert_collapsable(document):
|
||||
# If, however, we find a line starting '\begin_layout'
|
||||
# (_always_ present) then break with a warning message
|
||||
i = i + 1
|
||||
while 1:
|
||||
while True:
|
||||
if (document.body[i] == "collapsed false"):
|
||||
document.body[i] = "status open"
|
||||
break
|
||||
@ -924,10 +924,10 @@ def convert_collapsable(document):
|
||||
i = i + 1
|
||||
|
||||
|
||||
def revert_collapsable(document):
|
||||
def revert_collapsible(document):
|
||||
" Revert collapsed insets. "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_tokens_exact(document.body, ["\\begin_inset Box",
|
||||
"\\begin_inset Branch",
|
||||
"\\begin_inset CharStyle",
|
||||
@ -944,7 +944,7 @@ def revert_collapsable(document):
|
||||
# If, however, we find a line starting '\begin_layout'
|
||||
# (_always_ present) then break with a warning message
|
||||
i = i + 1
|
||||
while 1:
|
||||
while True:
|
||||
if (document.body[i] == "status open"):
|
||||
document.body[i] = "collapsed false"
|
||||
break
|
||||
@ -963,7 +963,7 @@ def revert_collapsable(document):
|
||||
def convert_ert(document):
|
||||
" Convert ERT. "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_inset ERT", i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -972,7 +972,7 @@ def convert_ert(document):
|
||||
# If, however, we find a line starting '\begin_layout'
|
||||
# (_always_ present) then break with a warning message
|
||||
i = i + 1
|
||||
while 1:
|
||||
while True:
|
||||
if (document.body[i] == "status Open"):
|
||||
document.body[i] = "status open"
|
||||
break
|
||||
@ -993,7 +993,7 @@ def convert_ert(document):
|
||||
def revert_ert(document):
|
||||
" Revert ERT. "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_inset ERT", i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -1002,7 +1002,7 @@ def revert_ert(document):
|
||||
# If, however, we find a line starting '\begin_layout'
|
||||
# (_always_ present) then break with a warning message
|
||||
i = i + 1
|
||||
while 1:
|
||||
while True:
|
||||
if (document.body[i] == "status open"):
|
||||
document.body[i] = "status Open"
|
||||
break
|
||||
@ -1028,7 +1028,7 @@ def convert_minipage(document):
|
||||
inner_pos = ["c","t","b","s"]
|
||||
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_inset Minipage", i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -1199,7 +1199,7 @@ def revert_breaks(document):
|
||||
|
||||
# Convert the insets
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_tokens(document.body, tokens, i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -1554,7 +1554,7 @@ def convert_frameless_box(document):
|
||||
pos = ['t', 'c', 'b']
|
||||
inner_pos = ['c', 't', 'b', 's']
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, '\\begin_inset Frameless', i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -1733,7 +1733,7 @@ def convert_frameless_box(document):
|
||||
def remove_branches(document):
|
||||
" Remove branches. "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.header, "\\branch", i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -1745,7 +1745,7 @@ def remove_branches(document):
|
||||
del document.header[i:j+1]
|
||||
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_inset Branch", i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -1760,7 +1760,7 @@ def remove_branches(document):
|
||||
# If, however, we find a line starting '\layout'
|
||||
# (_always_ present) then break with a warning message
|
||||
collapsed_found = 0
|
||||
while 1:
|
||||
while True:
|
||||
if (document.body[i][:9] == "collapsed"):
|
||||
del document.body[i]
|
||||
collapsed_found = 1
|
||||
@ -1821,7 +1821,7 @@ def revert_bibtopic(document):
|
||||
def convert_float(document):
|
||||
" Convert sideway floats. "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token_exact(document.body, '\\begin_inset Float', i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -1829,7 +1829,7 @@ def convert_float(document):
|
||||
# If, however, we find a line starting '\begin_layout'
|
||||
# (_always_ present) then break with a warning message
|
||||
i = i + 1
|
||||
while 1:
|
||||
while True:
|
||||
if (document.body[i][:4] == "wide"):
|
||||
document.body.insert(i + 1, 'sideways false')
|
||||
break
|
||||
@ -1843,7 +1843,7 @@ def convert_float(document):
|
||||
def revert_float(document):
|
||||
" Revert sideways floats. "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token_exact(document.body, '\\begin_inset Float', i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -1883,7 +1883,7 @@ def convert_graphics(document):
|
||||
""" Add extension to documentnames of insetgraphics if necessary.
|
||||
"""
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_inset Graphics", i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -1893,8 +1893,7 @@ def convert_graphics(document):
|
||||
return
|
||||
i = i + 1
|
||||
filename = document.body[j].split()[1]
|
||||
absname = os.path.normpath(os.path.join(document.dir, filename))
|
||||
if document.input == stdin and not os.path.isabs(filename):
|
||||
if document.dir == u'' and not os.path.isabs(filename):
|
||||
# We don't know the directory and cannot check the document.
|
||||
# We could use a heuristic and take the current directory,
|
||||
# and we could try to find out if documentname has an extension,
|
||||
@ -1905,6 +1904,7 @@ def convert_graphics(document):
|
||||
You may need to correct the document manually or run
|
||||
lyx2lyx again with the .lyx document as commandline argument.""" % filename)
|
||||
continue
|
||||
absname = os.path.normpath(os.path.join(document.dir, filename))
|
||||
# This needs to be the same algorithm as in pre 233 insetgraphics
|
||||
if access(absname, F_OK):
|
||||
continue
|
||||
@ -1924,7 +1924,7 @@ def convert_names(document):
|
||||
|
||||
i = 0
|
||||
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_layout Author", i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -2092,7 +2092,7 @@ def revert_paperpackage(document):
|
||||
def convert_bullets(document):
|
||||
" Convert bullets. "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.header, "\\bullet", i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -2110,7 +2110,7 @@ def convert_bullets(document):
|
||||
def revert_bullets(document):
|
||||
" Revert bullets. "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.header, "\\bullet", i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -2235,13 +2235,13 @@ def normalize_paragraph_params(document):
|
||||
"\\leftindent"
|
||||
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, '\\begin_layout', i)
|
||||
if i == -1:
|
||||
return
|
||||
|
||||
i = i + 1
|
||||
while 1:
|
||||
while True:
|
||||
if body[i].strip() and body[i].split()[0] not in allowed_parameters:
|
||||
break
|
||||
|
||||
@ -2282,7 +2282,7 @@ def convert_ert_paragraphs(document):
|
||||
'\\emph', '\\numeric', '\\bar', '\\noun',
|
||||
'\\color', '\\lang']
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, '\\begin_inset ERT', i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -2294,7 +2294,7 @@ def convert_ert_paragraphs(document):
|
||||
|
||||
# convert non-standard paragraphs to standard
|
||||
k = i
|
||||
while 1:
|
||||
while True:
|
||||
k = find_token(document.body, "\\begin_layout", k, j)
|
||||
if k == -1:
|
||||
break
|
||||
@ -2314,7 +2314,7 @@ def convert_ert_paragraphs(document):
|
||||
# insert an empty paragraph before each paragraph but the first
|
||||
k = i
|
||||
first_pagraph = 1
|
||||
while 1:
|
||||
while True:
|
||||
k = find_token(document.body, "\\begin_layout", k, j)
|
||||
if k == -1:
|
||||
break
|
||||
@ -2329,7 +2329,7 @@ def convert_ert_paragraphs(document):
|
||||
|
||||
# convert \\newline to new paragraph
|
||||
k = i
|
||||
while 1:
|
||||
while True:
|
||||
k = find_token(document.body, "\\newline", k, j)
|
||||
if k == -1:
|
||||
break
|
||||
@ -2347,7 +2347,7 @@ def convert_ert_paragraphs(document):
|
||||
def revert_ert_paragraphs(document):
|
||||
" Remove double paragraph breaks. "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, '\\begin_inset ERT', i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -2359,7 +2359,7 @@ def revert_ert_paragraphs(document):
|
||||
|
||||
# replace paragraph breaks with \newline
|
||||
k = i
|
||||
while 1:
|
||||
while True:
|
||||
k = find_token(document.body, "\\end_layout", k, j)
|
||||
l = find_token(document.body, "\\begin_layout", k, j)
|
||||
if k == -1 or l == -1:
|
||||
@ -2370,7 +2370,7 @@ def revert_ert_paragraphs(document):
|
||||
|
||||
# replace double \newlines with paragraph breaks
|
||||
k = i
|
||||
while 1:
|
||||
while True:
|
||||
k = find_token(document.body, "\\newline", k, j)
|
||||
if k == -1:
|
||||
break
|
||||
@ -2402,7 +2402,7 @@ def convert_french(document):
|
||||
# Change language in the document body
|
||||
regexp = re.compile(r'^\\lang\s+frenchb')
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_re(document.body, regexp, i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -2448,7 +2448,7 @@ def convert_sgml_paragraphs(document):
|
||||
return
|
||||
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_layout SGML", i)
|
||||
|
||||
if i == -1:
|
||||
@ -2474,7 +2474,7 @@ convert = [[222, [insert_tracking_changes, add_end_header, convert_amsmath]],
|
||||
convert_table_valignment_middle, convert_breaks]],
|
||||
[226, [convert_note]],
|
||||
[227, [convert_box]],
|
||||
[228, [convert_collapsable, convert_ert]],
|
||||
[228, [convert_collapsible, convert_ert]],
|
||||
[229, [convert_minipage]],
|
||||
[230, [convert_jurabib]],
|
||||
[231, [convert_float]],
|
||||
@ -2512,7 +2512,7 @@ revert = [[244, []],
|
||||
[230, [revert_float]],
|
||||
[229, [revert_jurabib]],
|
||||
[228, []],
|
||||
[227, [revert_collapsable, revert_ert]],
|
||||
[227, [revert_collapsible, revert_ert]],
|
||||
[226, [revert_box, revert_external_2]],
|
||||
[225, [revert_note]],
|
||||
[224, [rm_end_layout, begin_layout2layout, revert_end_document,
|
||||
|
@ -33,7 +33,6 @@ if not PY2:
|
||||
unichr = chr
|
||||
else:
|
||||
text_type = unicode
|
||||
unichr = unichr
|
||||
# End of code to support for both python 2 and 3
|
||||
|
||||
####################################################################
|
||||
@ -62,7 +61,7 @@ def find_beginning_of_layout(lines, i):
|
||||
def revert_framed(document):
|
||||
"Revert framed notes. "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_tokens(document.body, ["\\begin_inset Note Framed", "\\begin_inset Note Shaded"], i)
|
||||
|
||||
if i == -1:
|
||||
@ -217,7 +216,7 @@ def revert_booktabs(document):
|
||||
re_bspace = re.compile(r'\s+bottomspace="[^"]+"')
|
||||
re_ispace = re.compile(r'\s+interlinespace="[^"]+"')
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_inset Tabular", i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -471,14 +470,14 @@ implemented.'''
|
||||
def revert_cs_label(document):
|
||||
" Remove status flag of charstyle label. "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_inset CharStyle", i)
|
||||
if i == -1:
|
||||
return
|
||||
# Seach for a line starting 'show_label'
|
||||
# If it is not there, break with a warning message
|
||||
i = i + 1
|
||||
while 1:
|
||||
while True:
|
||||
if (document.body[i][:10] == "show_label"):
|
||||
del document.body[i]
|
||||
break
|
||||
@ -505,7 +504,7 @@ key "argument"
|
||||
This must be called after convert_commandparams.
|
||||
"""
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\bibitem", i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -593,7 +592,7 @@ def convert_commandparams(document):
|
||||
# convert_bibitem()), but could be read in, so we convert it here, too.
|
||||
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_inset LatexCommand", i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -681,7 +680,7 @@ def convert_commandparams(document):
|
||||
def revert_commandparams(document):
|
||||
regex = re.compile(r'(\S+)\s+(.+)')
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_inset LatexCommand", i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -739,7 +738,7 @@ def revert_nomenclature(document):
|
||||
regex = re.compile(r'(\S+)\s+(.+)')
|
||||
i = 0
|
||||
use_nomencl = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_inset LatexCommand nomenclature", i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -790,7 +789,7 @@ def revert_printnomenclature(document):
|
||||
regex = re.compile(r'(\S+)\s+(.+)')
|
||||
i = 0
|
||||
use_nomencl = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_inset LatexCommand printnomenclature", i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -856,7 +855,7 @@ def revert_esint(document):
|
||||
def revert_clearpage(document):
|
||||
" clearpage -> ERT "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\clearpage", i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -877,7 +876,7 @@ def revert_clearpage(document):
|
||||
def revert_cleardoublepage(document):
|
||||
" cleardoublepage -> ERT "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\cleardoublepage", i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -935,7 +934,7 @@ def revert_encodings(document):
|
||||
def convert_caption(document):
|
||||
" Convert caption layouts to caption insets. "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_layout Caption", i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -955,7 +954,7 @@ def revert_caption(document):
|
||||
" Convert caption insets to caption layouts. "
|
||||
" This assumes that the text class has a caption style. "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_inset Caption", i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -1120,7 +1119,7 @@ def convert_accent(document):
|
||||
re_contents = re.compile(r'^([^\s{]+)(.*)$')
|
||||
re_accentedcontents = re.compile(r'^\s*{?([^{}]*)}?\s*$')
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_re(document.body, re_wholeinset, i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -1453,13 +1452,13 @@ def revert_utf8plain(document):
|
||||
def revert_beamer_alert(document):
|
||||
" Revert beamer's \\alert inset back to ERT. "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_inset CharStyle Alert", i)
|
||||
if i == -1:
|
||||
return
|
||||
document.body[i] = "\\begin_inset ERT"
|
||||
i = i + 1
|
||||
while 1:
|
||||
while True:
|
||||
if (document.body[i][:13] == "\\begin_layout"):
|
||||
# Insert the \alert command
|
||||
document.body[i + 1] = "\\alert{" + document.body[i + 1] + '}'
|
||||
@ -1472,13 +1471,13 @@ def revert_beamer_alert(document):
|
||||
def revert_beamer_structure(document):
|
||||
" Revert beamer's \\structure inset back to ERT. "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_inset CharStyle Structure", i)
|
||||
if i == -1:
|
||||
return
|
||||
document.body[i] = "\\begin_inset ERT"
|
||||
i = i + 1
|
||||
while 1:
|
||||
while True:
|
||||
if (document.body[i][:13] == "\\begin_layout"):
|
||||
document.body[i + 1] = "\\structure{" + document.body[i + 1] + '}'
|
||||
break
|
||||
@ -1551,7 +1550,7 @@ def revert_cv_textclass(document):
|
||||
def convert_graphics_rotation(document):
|
||||
" add scaleBeforeRotation graphics parameter. "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_inset Graphics", i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -1573,7 +1572,7 @@ def convert_graphics_rotation(document):
|
||||
def revert_graphics_rotation(document):
|
||||
" remove scaleBeforeRotation graphics parameter. "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_inset Graphics", i)
|
||||
if i == -1:
|
||||
return
|
||||
|
@ -23,6 +23,7 @@ import unicodedata
|
||||
import sys, os
|
||||
|
||||
from parser_tools import find_token, find_end_of, find_tokens, get_value
|
||||
from unicode_symbols import read_unicodesymbols
|
||||
|
||||
####################################################################
|
||||
# Private helper functions
|
||||
@ -145,9 +146,18 @@ def set_option(document, m, option, value):
|
||||
return l
|
||||
|
||||
|
||||
# FIXME: Use the version in unicode_symbols.py which has some bug fixes
|
||||
def read_unicodesymbols():
|
||||
# FIXME: Remove this function if the version imported from unicode_symbols works.
|
||||
# This function was the predecessor from that function, that in the meanwhile got
|
||||
# new fixes.
|
||||
def read_unicodesymbols2():
|
||||
" Read the unicodesymbols list of unicode characters and corresponding commands."
|
||||
|
||||
# Provide support for both python 2 and 3
|
||||
PY2 = sys.version_info[0] == 2
|
||||
if not PY2:
|
||||
unichr = chr
|
||||
# End of code to support for both python 2 and 3
|
||||
|
||||
pathname = os.path.abspath(os.path.dirname(sys.argv[0]))
|
||||
fp = open(os.path.join(pathname.strip('lyx2lyx'), 'unicodesymbols'))
|
||||
spec_chars = []
|
||||
@ -424,7 +434,7 @@ def lyx2latex(document, lines):
|
||||
continue
|
||||
inert = ert_end >= curline
|
||||
content += lyxline2latex(document, lines[curline], inert)
|
||||
|
||||
|
||||
return content
|
||||
|
||||
|
||||
@ -966,7 +976,7 @@ def revert_pdf_options(document):
|
||||
def remove_inzip_options(document):
|
||||
"Remove inzipName and embed options from the Graphics inset"
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_inset Graphics", i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -994,7 +1004,7 @@ def convert_inset_command(document):
|
||||
LatexCommand cmd
|
||||
"""
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_inset LatexCommand", i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -1033,7 +1043,7 @@ def revert_inset_command(document):
|
||||
will not be able to recognize. Not sure what to do about that.
|
||||
"""
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_inset CommandInset", i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -1103,7 +1113,7 @@ def revert_wrapfig_options(document):
|
||||
|
||||
|
||||
def convert_latexcommand_index(document):
|
||||
"Convert from LatexCommand form to collapsable form."
|
||||
"Convert from LatexCommand form to collapsible form."
|
||||
i = 0
|
||||
r1 = re.compile('name "(.*)"')
|
||||
while True:
|
||||
@ -1135,7 +1145,7 @@ def convert_latexcommand_index(document):
|
||||
|
||||
|
||||
def revert_latexcommand_index(document):
|
||||
"Revert from collapsable form to LatexCommand form."
|
||||
"Revert from collapsible form to LatexCommand form."
|
||||
i = 0
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_inset Index", i)
|
||||
@ -1256,7 +1266,7 @@ def revert_japanese_encoding(document):
|
||||
def revert_inset_info(document):
|
||||
'Replace info inset with its content'
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, '\\begin_inset Info', i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -1737,7 +1747,7 @@ def convert_serbocroatian(document):
|
||||
def convert_framed_notes(document):
|
||||
"Convert framed notes to boxes. "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_tokens(document.body, ["\\begin_inset Note Framed", "\\begin_inset Note Shaded"], i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -1814,7 +1824,7 @@ def revert_colsep(document):
|
||||
def revert_framed_notes(document):
|
||||
"Revert framed boxes to notes. "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_tokens(document.body, ["\\begin_inset Box Framed", "\\begin_inset Box Shaded"], i)
|
||||
|
||||
if i == -1:
|
||||
@ -1932,7 +1942,7 @@ def revert_nocite_key(body, start, end):
|
||||
def revert_nocite(document):
|
||||
"Revert LatexCommand nocite to ERT"
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_inset CommandInset citation", i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -2050,7 +2060,7 @@ def revert_serbianlatin(document):
|
||||
def revert_rotfloat(document):
|
||||
" Revert sideways custom floats. "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
# whitespace intended (exclude \\begin_inset FloatList)
|
||||
i = find_token(document.body, "\\begin_inset Float ", i)
|
||||
if i == -1:
|
||||
@ -2113,7 +2123,7 @@ def revert_rotfloat(document):
|
||||
def revert_widesideways(document):
|
||||
" Revert wide sideways floats. "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
# whitespace intended (exclude \\begin_inset FloatList)
|
||||
i = find_token(document.body, '\\begin_inset Float ', i)
|
||||
if i == -1:
|
||||
@ -2165,7 +2175,7 @@ def revert_widesideways(document):
|
||||
def revert_inset_embedding(document, type):
|
||||
' Remove embed tag from certain type of insets'
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_inset %s" % type, i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -2190,7 +2200,7 @@ def revert_external_embedding(document):
|
||||
def convert_subfig(document):
|
||||
" Convert subfigures to subfloats. "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
addedLines = 0
|
||||
i = find_token(document.body, '\\begin_inset Graphics', i)
|
||||
if i == -1:
|
||||
@ -2230,7 +2240,7 @@ def convert_subfig(document):
|
||||
def revert_subfig(document):
|
||||
" Revert subfloats. "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
# whitespace intended (exclude \\begin_inset FloatList)
|
||||
i = find_tokens(document.body, ['\\begin_inset Float ', '\\begin_inset Wrap'], i)
|
||||
if i == -1:
|
||||
@ -2734,7 +2744,7 @@ def convert_japanese_plain(document):
|
||||
def revert_pdfpages(document):
|
||||
' Revert pdfpages external inset to ERT '
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_inset External", i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -2836,7 +2846,7 @@ def revert_master(document):
|
||||
def revert_graphics_group(document):
|
||||
' Revert group information from graphics insets '
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_inset Graphics", i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -2866,7 +2876,7 @@ def update_apa_styles(document):
|
||||
"Paragraph*": "Paragraph",
|
||||
"Subparagraph*": "Subparagraph"}
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_layout", i)
|
||||
if i == -1:
|
||||
return
|
||||
|
@ -1,6 +1,5 @@
|
||||
# -*- coding: utf-8 -*-
|
||||
# This file is part of lyx2lyx
|
||||
# -*- coding: utf-8 -*-
|
||||
# Copyright (C) 2011 The LyX team
|
||||
#
|
||||
# This program is free software; you can redistribute it and/or
|
||||
@ -27,7 +26,7 @@ from parser_tools import find_token, find_end_of, find_tokens, \
|
||||
find_token_exact, find_end_of_inset, find_end_of_layout, \
|
||||
find_token_backwards, is_in_inset, get_value, get_quoted_value, \
|
||||
del_token, check_token, get_option_value
|
||||
|
||||
|
||||
from lyx2lyx_tools import add_to_preamble, insert_to_preamble, \
|
||||
put_cmd_in_ert, lyx2latex, latex_length, revert_flex_inset, \
|
||||
revert_font_attrs, hex2ratio, str2bool
|
||||
@ -156,7 +155,7 @@ def revert_phantom_types(document, ptype, cmd):
|
||||
i = end
|
||||
continue
|
||||
substi = ["\\begin_inset ERT", "status collapsed", "",
|
||||
"\\begin_layout Plain Layout", "", "", "\\backslash",
|
||||
"\\begin_layout Plain Layout", "", "", "\\backslash",
|
||||
cmd + "{", "\\end_layout", "", "\\end_inset"]
|
||||
substj = ["\\size default", "", "\\begin_inset ERT", "status collapsed", "",
|
||||
"\\begin_layout Plain Layout", "", "}", "\\end_layout", "", "\\end_inset"]
|
||||
@ -168,7 +167,7 @@ def revert_phantom_types(document, ptype, cmd):
|
||||
|
||||
def revert_phantom(document):
|
||||
revert_phantom_types(document, "Phantom", "phantom")
|
||||
|
||||
|
||||
def revert_hphantom(document):
|
||||
revert_phantom_types(document, "HPhantom", "hphantom")
|
||||
|
||||
@ -200,7 +199,7 @@ def revert_xetex(document):
|
||||
roman = sans = typew = "default"
|
||||
osf = False
|
||||
sf_scale = tt_scale = 100.0
|
||||
|
||||
|
||||
i = find_token(document.header, "\\font_roman", 0)
|
||||
if i == -1:
|
||||
document.warning("Malformed LyX document: Missing \\font_roman.")
|
||||
@ -214,7 +213,7 @@ def revert_xetex(document):
|
||||
else:
|
||||
sans = get_value(document.header, "\\font_sans", i)
|
||||
document.header[i] = "\\font_sans default"
|
||||
|
||||
|
||||
i = find_token(document.header, "\\font_typewriter", 0)
|
||||
if i == -1:
|
||||
document.warning("Malformed LyX document: Missing \\font_typewriter.")
|
||||
@ -235,7 +234,7 @@ def revert_xetex(document):
|
||||
else:
|
||||
# we do not need this value.
|
||||
document.header[i] = "\\font_sc false"
|
||||
|
||||
|
||||
i = find_token(document.header, "\\font_sf_scale", 0)
|
||||
if i == -1:
|
||||
document.warning("Malformed LyX document: Missing \\font_sf_scale.")
|
||||
@ -279,14 +278,14 @@ def revert_xetex(document):
|
||||
pretext.append(tw)
|
||||
if osf:
|
||||
pretext.append('\\defaultfontfeatures{Numbers=OldStyle}')
|
||||
pretext.append('\usepackage{xunicode}')
|
||||
pretext.append('\usepackage{xltxtra}')
|
||||
pretext.append('\\usepackage{xunicode}')
|
||||
pretext.append('\\usepackage{xltxtra}')
|
||||
insert_to_preamble(document, pretext)
|
||||
|
||||
|
||||
def revert_outputformat(document):
|
||||
" Remove default output format param "
|
||||
|
||||
|
||||
if not del_token(document.header, '\\default_output_format', 0):
|
||||
document.warning("Malformed LyX document: Missing \\default_output_format.")
|
||||
|
||||
@ -344,7 +343,7 @@ def revert_splitindex(document):
|
||||
preamble = []
|
||||
if useindices:
|
||||
preamble.append("\\usepackage{splitidx})")
|
||||
|
||||
|
||||
# deal with index declarations in the preamble
|
||||
i = 0
|
||||
while True:
|
||||
@ -355,7 +354,7 @@ def revert_splitindex(document):
|
||||
if k == -1:
|
||||
document.warning("Malformed LyX document: Missing \\end_index.")
|
||||
return
|
||||
if useindices:
|
||||
if useindices:
|
||||
line = document.header[i]
|
||||
l = re.compile(r'\\index (.*)$')
|
||||
m = l.match(line)
|
||||
@ -366,7 +365,7 @@ def revert_splitindex(document):
|
||||
del document.header[i:k + 1]
|
||||
if preamble:
|
||||
insert_to_preamble(document, preamble)
|
||||
|
||||
|
||||
# deal with index insets
|
||||
# these need to have the argument removed
|
||||
i = 0
|
||||
@ -392,7 +391,7 @@ def revert_splitindex(document):
|
||||
subst = put_cmd_in_ert("\\sindex[" + itype + "]{" + content + "}")
|
||||
document.body[i:k + 1] = subst
|
||||
i = i + 1
|
||||
|
||||
|
||||
# deal with index_print insets
|
||||
i = 0
|
||||
while True:
|
||||
@ -431,7 +430,7 @@ def convert_splitindex(document):
|
||||
if document.body[i + 1].find('LatexCommand printindex') == -1:
|
||||
document.warning("Malformed LyX document: Incomplete printindex inset.")
|
||||
return
|
||||
subst = ["LatexCommand printindex",
|
||||
subst = ["LatexCommand printindex",
|
||||
"type \"idx\""]
|
||||
document.body[i + 1:i + 2] = subst
|
||||
i = i + 1
|
||||
@ -515,10 +514,10 @@ def revert_ulinelatex(document):
|
||||
|
||||
def revert_custom_processors(document):
|
||||
" Remove bibtex_command and index_command params "
|
||||
|
||||
|
||||
if not del_token(document.header, '\\bibtex_command', 0):
|
||||
document.warning("Malformed LyX document: Missing \\bibtex_command.")
|
||||
|
||||
|
||||
if not del_token(document.header, '\\index_command', 0):
|
||||
document.warning("Malformed LyX document: Missing \\index_command.")
|
||||
|
||||
@ -596,10 +595,10 @@ def revert_longtable_align(document):
|
||||
if j == -1:
|
||||
i += 1
|
||||
continue
|
||||
# FIXME Is this correct? It wipes out everything after the
|
||||
# FIXME Is this correct? It wipes out everything after the
|
||||
# one we found.
|
||||
document.body[fline] = document.body[fline][:j - 1] + '>'
|
||||
# since there could be a tabular inside this one, we
|
||||
# since there could be a tabular inside this one, we
|
||||
# cannot jump to end.
|
||||
i += 1
|
||||
|
||||
@ -743,7 +742,7 @@ def convert_author_id(document):
|
||||
i = 0
|
||||
anum = 1
|
||||
re_author = re.compile(r'(\\author) (\".*\")\s*(.*)$')
|
||||
|
||||
|
||||
while True:
|
||||
i = find_token(document.header, "\\author", i)
|
||||
if i == -1:
|
||||
@ -755,7 +754,7 @@ def convert_author_id(document):
|
||||
document.header[i] = "\\author %i %s %s" % (anum, name, email)
|
||||
anum += 1
|
||||
i += 1
|
||||
|
||||
|
||||
i = 0
|
||||
while True:
|
||||
i = find_token(document.body, "\\change_", i)
|
||||
@ -821,7 +820,7 @@ def revert_suppress_date(document):
|
||||
|
||||
def convert_mhchem(document):
|
||||
"Set mhchem to off for versions older than 1.6.x"
|
||||
if document.start < 277:
|
||||
if document.initial_format < 277:
|
||||
# LyX 1.5.x and older did never load mhchem.
|
||||
# Therefore we must switch it off: Documents that use mhchem have
|
||||
# a manual \usepackage anyway, and documents not using mhchem but
|
||||
@ -859,7 +858,7 @@ def revert_mhchem(document):
|
||||
|
||||
if mhchem == "off":
|
||||
# don't load case
|
||||
return
|
||||
return
|
||||
|
||||
if mhchem == "auto":
|
||||
i = 0
|
||||
@ -874,9 +873,9 @@ def revert_mhchem(document):
|
||||
i += 1
|
||||
|
||||
if mhchem == "on":
|
||||
pre = ["\\PassOptionsToPackage{version=3}{mhchem}",
|
||||
pre = ["\\PassOptionsToPackage{version=3}{mhchem}",
|
||||
"\\usepackage{mhchem}"]
|
||||
insert_to_preamble(document, pre)
|
||||
insert_to_preamble(document, pre)
|
||||
|
||||
|
||||
def revert_fontenc(document):
|
||||
@ -916,7 +915,7 @@ def merge_gbrief(document):
|
||||
"Verteiler": "cc",
|
||||
"Gruss": "Closing"}
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_layout", i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -926,7 +925,7 @@ def merge_gbrief(document):
|
||||
document.body[i] = "\\begin_layout " + obsoletedby[layout]
|
||||
|
||||
i += 1
|
||||
|
||||
|
||||
document.textclass = "g-brief"
|
||||
document.set_textclass()
|
||||
|
||||
@ -989,7 +988,7 @@ def revert_multirow(document):
|
||||
if i == -1:
|
||||
begin_table = end_table
|
||||
continue
|
||||
|
||||
|
||||
# store the number of rows and columns
|
||||
numrows = get_option_value(document.body[begin_table], "rows")
|
||||
numcols = get_option_value(document.body[begin_table], "columns")
|
||||
@ -1016,13 +1015,13 @@ def revert_multirow(document):
|
||||
break
|
||||
begin_cell = begin_row
|
||||
multirows.append([])
|
||||
for column in range(numcols):
|
||||
for column in range(numcols):
|
||||
begin_cell = find_token(document.body, '<cell ', begin_cell, end_row)
|
||||
if begin_cell == -1:
|
||||
document.warning("Can't find column " + str(column + 1) + \
|
||||
"in row " + str(row + 1))
|
||||
break
|
||||
# NOTE
|
||||
# NOTE
|
||||
# this will fail if someone puts "</cell>" in a cell, but
|
||||
# that seems fairly unlikely.
|
||||
end_cell = find_end_of(document.body, begin_cell, '<cell', '</cell>')
|
||||
@ -1122,7 +1121,7 @@ def revert_math_output(document):
|
||||
else:
|
||||
document.warning("Unable to match " + document.header[i])
|
||||
document.header[i] = "\\html_use_mathml " + newval
|
||||
|
||||
|
||||
|
||||
|
||||
def revert_inset_preview(document):
|
||||
@ -1137,7 +1136,7 @@ def revert_inset_preview(document):
|
||||
document.warning("Malformed LyX document: Could not find end of Preview inset.")
|
||||
i += 1
|
||||
continue
|
||||
|
||||
|
||||
# This has several issues.
|
||||
# We need to do something about the layouts inside InsetPreview.
|
||||
# If we just leave the first one, then we have something like:
|
||||
@ -1146,16 +1145,16 @@ def revert_inset_preview(document):
|
||||
# \begin_layout Standard
|
||||
# and we get a "no \end_layout" error. So something has to be done.
|
||||
# Ideally, we would check if it is the same as the layout we are in.
|
||||
# If so, we just remove it; if not, we end the active one. But it is
|
||||
# If so, we just remove it; if not, we end the active one. But it is
|
||||
# not easy to know what layout we are in, due to depth changes, etc,
|
||||
# and it is not clear to me how much work it is worth doing. In most
|
||||
# cases, the layout will probably be the same.
|
||||
#
|
||||
#
|
||||
# For the same reason, we have to remove the \end_layout tag at the
|
||||
# end of the last layout in the inset. Again, that will sometimes be
|
||||
# wrong, but it will usually be right. To know what to do, we would
|
||||
# again have to know what layout the inset is in.
|
||||
|
||||
|
||||
blay = find_token(document.body, "\\begin_layout", i, iend)
|
||||
if blay == -1:
|
||||
document.warning("Can't find layout for preview inset!")
|
||||
@ -1167,13 +1166,13 @@ def revert_inset_preview(document):
|
||||
|
||||
# This is where we would check what layout we are in.
|
||||
# The check for Standard is definitely wrong.
|
||||
#
|
||||
#
|
||||
# lay = document.body[blay].split(None, 1)[1]
|
||||
# if lay != oldlayout:
|
||||
# # record a boolean to tell us what to do later....
|
||||
# # better to do it later, since (a) it won't mess up
|
||||
# # the numbering and (b) we only modify at the end.
|
||||
|
||||
|
||||
# we want to delete the last \\end_layout in this inset, too.
|
||||
# note that this may not be the \\end_layout that goes with blay!!
|
||||
bend = find_end_of_layout(document.body, blay)
|
||||
@ -1193,7 +1192,7 @@ def revert_inset_preview(document):
|
||||
del document.body[bend]
|
||||
del document.body[i:blay + 1]
|
||||
# we do not need to reset i
|
||||
|
||||
|
||||
|
||||
def revert_equalspacing_xymatrix(document):
|
||||
" Revert a Formula with xymatrix@! to an ERT inset "
|
||||
@ -1210,12 +1209,12 @@ def revert_equalspacing_xymatrix(document):
|
||||
document.warning("Malformed LyX document: Could not find end of Formula inset.")
|
||||
i += 1
|
||||
continue
|
||||
|
||||
|
||||
for curline in range(i,j):
|
||||
found = document.body[curline].find("\\xymatrix@!")
|
||||
if found != -1:
|
||||
break
|
||||
|
||||
|
||||
if found != -1:
|
||||
has_equal_spacing = True
|
||||
content = [document.body[i][21:]]
|
||||
@ -1230,7 +1229,7 @@ def revert_equalspacing_xymatrix(document):
|
||||
has_preamble = True;
|
||||
break;
|
||||
i = j + 1
|
||||
|
||||
|
||||
if has_equal_spacing and not has_preamble:
|
||||
add_to_preamble(document, ['\\usepackage[all]{xy}'])
|
||||
|
||||
@ -1247,7 +1246,7 @@ def revert_notefontcolor(document):
|
||||
|
||||
# are there any grey notes?
|
||||
if find_token(document.body, "\\begin_inset Note Greyedout", 0) == -1:
|
||||
# no need to do anything else, and \renewcommand will throw
|
||||
# no need to do anything else, and \renewcommand will throw
|
||||
# an error since lyxgreyedout will not exist.
|
||||
return
|
||||
|
||||
@ -1265,21 +1264,21 @@ def revert_notefontcolor(document):
|
||||
|
||||
|
||||
def revert_turkmen(document):
|
||||
"Set language Turkmen to English"
|
||||
"Set language Turkmen to English"
|
||||
|
||||
if document.language == "turkmen":
|
||||
document.language = "english"
|
||||
i = find_token(document.header, "\\language", 0)
|
||||
if i != -1:
|
||||
document.header[i] = "\\language english"
|
||||
if document.language == "turkmen":
|
||||
document.language = "english"
|
||||
i = find_token(document.header, "\\language", 0)
|
||||
if i != -1:
|
||||
document.header[i] = "\\language english"
|
||||
|
||||
j = 0
|
||||
while True:
|
||||
j = find_token(document.body, "\\lang turkmen", j)
|
||||
if j == -1:
|
||||
return
|
||||
document.body[j] = document.body[j].replace("\\lang turkmen", "\\lang english")
|
||||
j += 1
|
||||
j = 0
|
||||
while True:
|
||||
j = find_token(document.body, "\\lang turkmen", j)
|
||||
if j == -1:
|
||||
return
|
||||
document.body[j] = document.body[j].replace("\\lang turkmen", "\\lang english")
|
||||
j += 1
|
||||
|
||||
|
||||
def revert_fontcolor(document):
|
||||
@ -1332,7 +1331,7 @@ def revert_lyx_version(document):
|
||||
pass
|
||||
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, '\\begin_inset Info', i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -1406,7 +1405,7 @@ def convert_html_quotes(document):
|
||||
m = l.match(line)
|
||||
if m:
|
||||
document.header[i] = "\\html_latex_start " + m.group(1)
|
||||
|
||||
|
||||
i = find_token(document.header, '\\html_latex_end', 0)
|
||||
if i != -1:
|
||||
line = document.header[i]
|
||||
@ -1414,11 +1413,11 @@ def convert_html_quotes(document):
|
||||
m = l.match(line)
|
||||
if m:
|
||||
document.header[i] = "\\html_latex_end " + m.group(1)
|
||||
|
||||
|
||||
|
||||
def revert_html_quotes(document):
|
||||
" Remove quotes around html_latex_start and html_latex_end "
|
||||
|
||||
|
||||
i = find_token(document.header, '\\html_latex_start', 0)
|
||||
if i != -1:
|
||||
line = document.header[i]
|
||||
@ -1429,7 +1428,7 @@ def revert_html_quotes(document):
|
||||
del document.header[i]
|
||||
else:
|
||||
document.header[i] = "\\html_latex_start \"" + m.group(1) + "\""
|
||||
|
||||
|
||||
i = find_token(document.header, '\\html_latex_end', 0)
|
||||
if i != -1:
|
||||
line = document.header[i]
|
||||
@ -1481,7 +1480,7 @@ def revert_align_decimal(document):
|
||||
def convert_optarg(document):
|
||||
" Convert \\begin_inset OptArg to \\begin_inset Argument "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, '\\begin_inset OptArg', i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -1492,7 +1491,7 @@ def convert_optarg(document):
|
||||
def revert_argument(document):
|
||||
" Convert \\begin_inset Argument to \\begin_inset OptArg "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, '\\begin_inset Argument', i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -1621,8 +1620,8 @@ def convert_prettyref(document):
|
||||
document.body[k] = "LatexCommand formatted"
|
||||
i = j + 1
|
||||
document.header.insert(-1, "\\use_refstyle 0")
|
||||
|
||||
|
||||
|
||||
|
||||
def revert_refstyle(document):
|
||||
" Reverts neutral formatted refs to prettyref "
|
||||
re_ref = re.compile("^reference\s+\"(\w+):(\S+)\"")
|
||||
@ -1645,7 +1644,7 @@ def revert_refstyle(document):
|
||||
i = find_token(document.header, "\\use_refstyle", 0)
|
||||
if i != -1:
|
||||
document.header.pop(i)
|
||||
|
||||
|
||||
|
||||
def revert_nameref(document):
|
||||
" Convert namerefs to regular references "
|
||||
@ -1655,7 +1654,7 @@ def revert_nameref(document):
|
||||
for cmd in cmds:
|
||||
i = 0
|
||||
oldcmd = "LatexCommand " + cmd
|
||||
while 1:
|
||||
while True:
|
||||
# It seems better to look for this, as most of the reference
|
||||
# insets won't be ones we care about.
|
||||
i = find_token(document.body, oldcmd, i)
|
||||
@ -1686,13 +1685,13 @@ def revert_nameref(document):
|
||||
document.body[stins:endins + 1] = newcontent
|
||||
|
||||
if foundone:
|
||||
add_to_preamble(document, ["\usepackage{nameref}"])
|
||||
add_to_preamble(document, ["\\usepackage{nameref}"])
|
||||
|
||||
|
||||
def remove_Nameref(document):
|
||||
" Convert Nameref commands to nameref commands "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
# It seems better to look for this, as most of the reference
|
||||
# insets won't be ones we care about.
|
||||
i = find_token(document.body, "LatexCommand Nameref" , i)
|
||||
@ -1700,7 +1699,7 @@ def remove_Nameref(document):
|
||||
break
|
||||
cmdloc = i
|
||||
i += 1
|
||||
|
||||
|
||||
# Make sure it is actually in an inset!
|
||||
val = is_in_inset(document.body, cmdloc, \
|
||||
"\\begin_inset CommandInset ref")
|
||||
@ -1720,7 +1719,7 @@ def revert_mathrsfs(document):
|
||||
|
||||
def convert_flexnames(document):
|
||||
"Convert \\begin_inset Flex Custom:Style to \\begin_inset Flex Style and similarly for CharStyle and Element."
|
||||
|
||||
|
||||
i = 0
|
||||
rx = re.compile(r'^\\begin_inset Flex (?:Custom|CharStyle|Element):(.+)$')
|
||||
while True:
|
||||
@ -1800,7 +1799,7 @@ def revert_flexnames(document):
|
||||
flexlist = flex_insets
|
||||
else:
|
||||
flexlist = flex_elements
|
||||
|
||||
|
||||
rx = re.compile(r'^\\begin_inset Flex\s+(.+)$')
|
||||
i = 0
|
||||
while True:
|
||||
@ -1857,7 +1856,7 @@ def revert_mathdots(document):
|
||||
# force load case
|
||||
add_to_preamble(document, ["\\usepackage{mathdots}"])
|
||||
return
|
||||
|
||||
|
||||
# so we are in the auto case. we want to load mathdots if \iddots is used.
|
||||
i = 0
|
||||
while True:
|
||||
@ -1879,7 +1878,7 @@ def revert_mathdots(document):
|
||||
def convert_rule(document):
|
||||
" Convert \\lyxline to CommandInset line. "
|
||||
i = 0
|
||||
|
||||
|
||||
inset = ['\\begin_inset CommandInset line',
|
||||
'LatexCommand rule',
|
||||
'offset "0.5ex"',
|
||||
@ -1934,7 +1933,7 @@ def convert_rule(document):
|
||||
def revert_rule(document):
|
||||
" Revert line insets to Tex code "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_inset CommandInset line" , i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -1969,7 +1968,7 @@ def revert_diagram(document):
|
||||
j = find_end_of_inset(document.body, i)
|
||||
if j == -1:
|
||||
document.warning("Malformed LyX document: Can't find end of Formula inset.")
|
||||
return
|
||||
return
|
||||
lines = "\n".join(document.body[i:j])
|
||||
if lines.find("\\Diagram") == -1:
|
||||
i = j
|
||||
@ -1978,9 +1977,9 @@ def revert_diagram(document):
|
||||
# only need to do it once!
|
||||
return
|
||||
|
||||
chapters = ("amsbook", "book", "docbook-book", "elsart", "extbook", "extreport",
|
||||
"jbook", "jreport", "jsbook", "literate-book", "literate-report", "memoir",
|
||||
"mwbk", "mwrep", "recipebook", "report", "scrbook", "scrreprt", "svmono",
|
||||
chapters = ("amsbook", "book", "docbook-book", "elsart", "extbook", "extreport",
|
||||
"jbook", "jreport", "jsbook", "literate-book", "literate-report", "memoir",
|
||||
"mwbk", "mwrep", "recipebook", "report", "scrbook", "scrreprt", "svmono",
|
||||
"svmult", "tbook", "treport", "tufte-book")
|
||||
|
||||
def convert_bibtex_clearpage(document):
|
||||
@ -2024,13 +2023,13 @@ def convert_bibtex_clearpage(document):
|
||||
document.warning("Can't find options for bibliography inset at line " + str(j))
|
||||
j = k
|
||||
continue
|
||||
|
||||
|
||||
if val.find("bibtotoc") == -1:
|
||||
j = k
|
||||
continue
|
||||
|
||||
|
||||
# so we want to insert a new page right before the paragraph that
|
||||
# this bibliography thing is in.
|
||||
# this bibliography thing is in.
|
||||
lay = find_token_backwards(document.body, "\\begin_layout", j)
|
||||
if lay == -1:
|
||||
document.warning("Can't find layout containing bibliography inset at line " + str(j))
|
||||
@ -2065,7 +2064,7 @@ def convert_passthru(document):
|
||||
" http://www.mail-archive.com/lyx-devel@lists.lyx.org/msg161298.html "
|
||||
if not check_passthru:
|
||||
return
|
||||
|
||||
|
||||
rx = re.compile("\\\\begin_layout \s*(\w+)")
|
||||
beg = 0
|
||||
for lay in ["Chunk", "Scrap"]:
|
||||
@ -2124,7 +2123,7 @@ def convert_passthru(document):
|
||||
beg = end + 1
|
||||
if didit:
|
||||
beg += 4 # for the extra layout
|
||||
|
||||
|
||||
|
||||
def revert_passthru(document):
|
||||
" http://www.mail-archive.com/lyx-devel@lists.lyx.org/msg161298.html "
|
||||
@ -2142,7 +2141,7 @@ def revert_passthru(document):
|
||||
document.warning("Can't find end of layout at line " + str(beg))
|
||||
beg += 1
|
||||
continue
|
||||
|
||||
|
||||
# we now want to find out if the next layout is the
|
||||
# same as this one. but we will need to do this over and
|
||||
# over again.
|
||||
@ -2160,7 +2159,6 @@ def revert_passthru(document):
|
||||
# but first let's check and make sure there is no content between the
|
||||
# two layouts. i'm not sure if that can happen or not.
|
||||
for l in range(end + 1, next):
|
||||
document.warning("c'" + document.body[l] + "'")
|
||||
if document.body[l] != "":
|
||||
document.warning("Found content between adjacent " + lay + " layouts!")
|
||||
break
|
||||
@ -2170,7 +2168,6 @@ def revert_passthru(document):
|
||||
break
|
||||
empty = True
|
||||
for l in range(next + 1, nextend):
|
||||
document.warning("e'" + document.body[l] + "'")
|
||||
if document.body[l] != "":
|
||||
empty = False
|
||||
break
|
||||
@ -2225,7 +2222,7 @@ def revert_multirowOffset(document):
|
||||
if i == -1:
|
||||
begin_table = end_table
|
||||
continue
|
||||
|
||||
|
||||
# store the number of rows and columns
|
||||
numrows = get_option_value(document.body[begin_table], "rows")
|
||||
numcols = get_option_value(document.body[begin_table], "columns")
|
||||
@ -2252,13 +2249,13 @@ def revert_multirowOffset(document):
|
||||
break
|
||||
begin_cell = begin_row
|
||||
multirows.append([])
|
||||
for column in range(numcols):
|
||||
for column in range(numcols):
|
||||
begin_cell = find_token(document.body, '<cell ', begin_cell, end_row)
|
||||
if begin_cell == -1:
|
||||
document.warning("Can't find column " + str(column + 1) + \
|
||||
"in row " + str(row + 1))
|
||||
break
|
||||
# NOTE
|
||||
# NOTE
|
||||
# this will fail if someone puts "</cell>" in a cell, but
|
||||
# that seems fairly unlikely.
|
||||
end_cell = find_end_of(document.body, begin_cell, '<cell', '</cell>')
|
||||
@ -2310,7 +2307,7 @@ def revert_multirowOffset(document):
|
||||
replace('valignment="middle"', 'valignment="top"')
|
||||
# remove mroffset option
|
||||
document.body[bcell] = rgx.sub('', document.body[bcell])
|
||||
|
||||
|
||||
blay = find_token(document.body, "\\begin_layout", bcell, ecell)
|
||||
if blay == -1:
|
||||
document.warning("Can't find layout for cell!")
|
||||
@ -2336,7 +2333,7 @@ def revert_script(document):
|
||||
" Convert subscript/superscript inset to TeX code "
|
||||
i = 0
|
||||
foundsubscript = False
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, '\\begin_inset script', i)
|
||||
if i == -1:
|
||||
break
|
||||
|
@ -1,6 +1,5 @@
|
||||
# -*- coding: utf-8 -*-
|
||||
# This file is part of lyx2lyx
|
||||
# -*- coding: utf-8 -*-
|
||||
# Copyright (C) 2011 The LyX team
|
||||
#
|
||||
# This program is free software; you can redistribute it and/or
|
||||
@ -117,7 +116,7 @@ def convert_TeX_brace_to_Argument(document, line, n, nmax, inset, environment, o
|
||||
isInset must be true, if braces inside an InsetLayout needs to be converted
|
||||
isEnvironment must be true, if the layout is for a LaTeX environment
|
||||
isOpt must be true, if the argument is an optional one
|
||||
|
||||
|
||||
Todo: this routine can currently handle only one mandatory argument of environments
|
||||
'''
|
||||
|
||||
@ -168,7 +167,7 @@ def convert_TeX_brace_to_Argument(document, line, n, nmax, inset, environment, o
|
||||
n += 1
|
||||
endn = end
|
||||
loop += 1
|
||||
else:
|
||||
else:
|
||||
# no brace pair found
|
||||
# now check the case that we have "}" + "{" in two ERTs
|
||||
if opt:
|
||||
@ -449,16 +448,16 @@ def revert_justification(document):
|
||||
|
||||
|
||||
def revert_australian(document):
|
||||
"Set English language variants Australian and Newzealand to English"
|
||||
"Set English language variants Australian and Newzealand to English"
|
||||
|
||||
if document.language == "australian" or document.language == "newzealand":
|
||||
if document.language == "australian" or document.language == "newzealand":
|
||||
document.language = "english"
|
||||
i = find_token(document.header, "\\language", 0)
|
||||
if i != -1:
|
||||
document.header[i] = "\\language english"
|
||||
j = 0
|
||||
while True:
|
||||
j = find_token(document.body, "\\lang australian", j)
|
||||
i = find_token(document.header, "\\language", 0)
|
||||
if i != -1:
|
||||
document.header[i] = "\\language english"
|
||||
j = 0
|
||||
while True:
|
||||
j = find_token(document.body, "\\lang australian", j)
|
||||
if j == -1:
|
||||
j = find_token(document.body, "\\lang newzealand", 0)
|
||||
if j == -1:
|
||||
@ -466,7 +465,7 @@ def revert_australian(document):
|
||||
else:
|
||||
document.body[j] = document.body[j].replace("\\lang newzealand", "\\lang english")
|
||||
else:
|
||||
document.body[j] = document.body[j].replace("\\lang australian", "\\lang english")
|
||||
document.body[j] = document.body[j].replace("\\lang australian", "\\lang english")
|
||||
j += 1
|
||||
|
||||
|
||||
@ -565,7 +564,7 @@ def handle_longtable_captions(document, forward):
|
||||
if get_option_value(document.body[begin_row], 'endlastfoot') == 'true':
|
||||
document.body[begin_row] = set_option_value(document.body[begin_row], 'endlastfoot', 'false')
|
||||
begin_row = end_row
|
||||
# since there could be a tabular inside this one, we
|
||||
# since there could be a tabular inside this one, we
|
||||
# cannot jump to end.
|
||||
begin_table += 1
|
||||
|
||||
@ -848,7 +847,7 @@ def revert_verbatim(document, starred = False):
|
||||
'begin{%s}' % (latex_name),
|
||||
'\\end_layout', '', '\\begin_layout Plain Layout', '']
|
||||
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_layout %s" % (layout_name), i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -860,7 +859,7 @@ def revert_verbatim(document, starred = False):
|
||||
continue
|
||||
# delete all line breaks insets (there are no other insets)
|
||||
l = i
|
||||
while 1:
|
||||
while True:
|
||||
n = find_token(document.body, "\\begin_inset Newline newline", l, j)
|
||||
if n == -1:
|
||||
n = find_token(document.body, "\\begin_inset Newline linebreak", l, j)
|
||||
@ -908,7 +907,7 @@ def revert_verbatim(document, starred = False):
|
||||
def revert_tipa(document):
|
||||
" Revert native TIPA insets to mathed or ERT. "
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_inset IPA", i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -975,12 +974,12 @@ def revert_cell_rotation(document):
|
||||
put_cmd_in_ert("\\end{turn}")
|
||||
document.body[i + 4 : i + 4] = \
|
||||
put_cmd_in_ert("\\begin{turn}{" + value + "}")
|
||||
|
||||
|
||||
i += 1
|
||||
|
||||
|
||||
finally:
|
||||
if load_rotating:
|
||||
add_to_preamble(document, ["\\@ifundefined{turnbox}{\usepackage{rotating}}{}"])
|
||||
add_to_preamble(document, ["\\@ifundefined{turnbox}{\\usepackage{rotating}}{}"])
|
||||
|
||||
|
||||
def convert_cell_rotation(document):
|
||||
@ -997,7 +996,7 @@ def convert_cell_rotation(document):
|
||||
rgx = re.compile(r'rotate="[^"]+?"')
|
||||
# convert "true" to "90"
|
||||
document.body[i] = rgx.sub('rotate="90"', document.body[i])
|
||||
|
||||
|
||||
i += 1
|
||||
|
||||
|
||||
@ -1034,12 +1033,12 @@ def revert_table_rotation(document):
|
||||
put_cmd_in_ert("\\end{turn}")
|
||||
document.body[i - 2 : i - 2] = \
|
||||
put_cmd_in_ert("\\begin{turn}{" + value + "}")
|
||||
|
||||
|
||||
i += 1
|
||||
|
||||
|
||||
finally:
|
||||
if load_rotating:
|
||||
add_to_preamble(document, ["\\@ifundefined{turnbox}{\usepackage{rotating}}{}"])
|
||||
add_to_preamble(document, ["\\@ifundefined{turnbox}{\\usepackage{rotating}}{}"])
|
||||
|
||||
|
||||
def convert_table_rotation(document):
|
||||
@ -1056,7 +1055,7 @@ def convert_table_rotation(document):
|
||||
rgx = re.compile(r'rotate="[^"]+?"')
|
||||
# convert "true" to "90"
|
||||
document.body[i] = rgx.sub('rotate="90"', document.body[i])
|
||||
|
||||
|
||||
i += 1
|
||||
|
||||
|
||||
@ -1150,25 +1149,25 @@ def revert_use_cancel(document):
|
||||
|
||||
|
||||
def revert_ancientgreek(document):
|
||||
"Set the document language for ancientgreek to greek"
|
||||
"Set the document language for ancientgreek to greek"
|
||||
|
||||
if document.language == "ancientgreek":
|
||||
if document.language == "ancientgreek":
|
||||
document.language = "greek"
|
||||
i = find_token(document.header, "\\language", 0)
|
||||
if i != -1:
|
||||
document.header[i] = "\\language greek"
|
||||
j = 0
|
||||
while True:
|
||||
j = find_token(document.body, "\\lang ancientgreek", j)
|
||||
i = find_token(document.header, "\\language", 0)
|
||||
if i != -1:
|
||||
document.header[i] = "\\language greek"
|
||||
j = 0
|
||||
while True:
|
||||
j = find_token(document.body, "\\lang ancientgreek", j)
|
||||
if j == -1:
|
||||
return
|
||||
else:
|
||||
document.body[j] = document.body[j].replace("\\lang ancientgreek", "\\lang greek")
|
||||
document.body[j] = document.body[j].replace("\\lang ancientgreek", "\\lang greek")
|
||||
j += 1
|
||||
|
||||
|
||||
def revert_languages(document):
|
||||
"Set the document language for new supported languages to English"
|
||||
"Set the document language for new supported languages to English"
|
||||
|
||||
languages = [
|
||||
"coptic", "divehi", "hindi", "kurmanji", "lao", "marathi", "occitan", "sanskrit",
|
||||
@ -1177,11 +1176,11 @@ def revert_languages(document):
|
||||
for n in range(len(languages)):
|
||||
if document.language == languages[n]:
|
||||
document.language = "english"
|
||||
i = find_token(document.header, "\\language", 0)
|
||||
if i != -1:
|
||||
document.header[i] = "\\language english"
|
||||
i = find_token(document.header, "\\language", 0)
|
||||
if i != -1:
|
||||
document.header[i] = "\\language english"
|
||||
j = 0
|
||||
while j < len(document.body):
|
||||
while j < len(document.body):
|
||||
j = find_token(document.body, "\\lang " + languages[n], j)
|
||||
if j != -1:
|
||||
document.body[j] = document.body[j].replace("\\lang " + languages[n], "\\lang english")
|
||||
@ -1191,27 +1190,27 @@ def revert_languages(document):
|
||||
|
||||
|
||||
def convert_armenian(document):
|
||||
"Use polyglossia and thus non-TeX fonts for Armenian"
|
||||
"Use polyglossia and thus non-TeX fonts for Armenian"
|
||||
|
||||
if document.language == "armenian":
|
||||
i = find_token(document.header, "\\use_non_tex_fonts", 0)
|
||||
if i != -1:
|
||||
document.header[i] = "\\use_non_tex_fonts true"
|
||||
if document.language == "armenian":
|
||||
i = find_token(document.header, "\\use_non_tex_fonts", 0)
|
||||
if i != -1:
|
||||
document.header[i] = "\\use_non_tex_fonts true"
|
||||
|
||||
|
||||
def revert_armenian(document):
|
||||
"Use ArmTeX and thus TeX fonts for Armenian"
|
||||
"Use ArmTeX and thus TeX fonts for Armenian"
|
||||
|
||||
if document.language == "armenian":
|
||||
i = find_token(document.header, "\\use_non_tex_fonts", 0)
|
||||
if i != -1:
|
||||
document.header[i] = "\\use_non_tex_fonts false"
|
||||
if document.language == "armenian":
|
||||
i = find_token(document.header, "\\use_non_tex_fonts", 0)
|
||||
if i != -1:
|
||||
document.header[i] = "\\use_non_tex_fonts false"
|
||||
|
||||
|
||||
def revert_libertine(document):
|
||||
" Revert native libertine font definition to LaTeX "
|
||||
" Revert native libertine font definition to LaTeX "
|
||||
|
||||
if find_token(document.header, "\\use_non_tex_fonts false", 0) != -1:
|
||||
if find_token(document.header, "\\use_non_tex_fonts false", 0) != -1:
|
||||
i = find_token(document.header, "\\font_roman libertine", 0)
|
||||
if i != -1:
|
||||
osf = False
|
||||
@ -1230,9 +1229,9 @@ def revert_libertine(document):
|
||||
|
||||
|
||||
def revert_txtt(document):
|
||||
" Revert native txtt font definition to LaTeX "
|
||||
" Revert native txtt font definition to LaTeX "
|
||||
|
||||
if find_token(document.header, "\\use_non_tex_fonts false", 0) != -1:
|
||||
if find_token(document.header, "\\use_non_tex_fonts false", 0) != -1:
|
||||
i = find_token(document.header, "\\font_typewriter txtt", 0)
|
||||
if i != -1:
|
||||
preamble = "\\renewcommand{\\ttdefault}{txtt}"
|
||||
@ -1241,9 +1240,9 @@ def revert_txtt(document):
|
||||
|
||||
|
||||
def revert_mathdesign(document):
|
||||
" Revert native mathdesign font definition to LaTeX "
|
||||
" Revert native mathdesign font definition to LaTeX "
|
||||
|
||||
if find_token(document.header, "\\use_non_tex_fonts false", 0) != -1:
|
||||
if find_token(document.header, "\\use_non_tex_fonts false", 0) != -1:
|
||||
mathdesign_dict = {
|
||||
"mdbch": "charter",
|
||||
"mdput": "utopia",
|
||||
@ -1272,9 +1271,9 @@ def revert_mathdesign(document):
|
||||
|
||||
|
||||
def revert_texgyre(document):
|
||||
" Revert native TeXGyre font definition to LaTeX "
|
||||
" Revert native TeXGyre font definition to LaTeX "
|
||||
|
||||
if find_token(document.header, "\\use_non_tex_fonts false", 0) != -1:
|
||||
if find_token(document.header, "\\use_non_tex_fonts false", 0) != -1:
|
||||
texgyre_fonts = ["tgadventor", "tgbonum", "tgchorus", "tgcursor", \
|
||||
"tgheros", "tgpagella", "tgschola", "tgtermes"]
|
||||
i = find_token(document.header, "\\font_roman", 0)
|
||||
@ -1331,7 +1330,7 @@ def revert_ipadeco(document):
|
||||
i = end
|
||||
continue
|
||||
substi = ["\\begin_inset ERT", "status collapsed", "",
|
||||
"\\begin_layout Plain Layout", "", "", "\\backslash",
|
||||
"\\begin_layout Plain Layout", "", "", "\\backslash",
|
||||
decotype + "{", "\\end_layout", "", "\\end_inset"]
|
||||
substj = ["\\size default", "", "\\begin_inset ERT", "status collapsed", "",
|
||||
"\\begin_layout Plain Layout", "", "}", "\\end_layout", "", "\\end_inset"]
|
||||
@ -1371,9 +1370,9 @@ def revert_ipachar(document):
|
||||
|
||||
|
||||
def revert_minionpro(document):
|
||||
" Revert native MinionPro font definition to LaTeX "
|
||||
" Revert native MinionPro font definition to LaTeX "
|
||||
|
||||
if find_token(document.header, "\\use_non_tex_fonts false", 0) != -1:
|
||||
if find_token(document.header, "\\use_non_tex_fonts false", 0) != -1:
|
||||
i = find_token(document.header, "\\font_roman minionpro", 0)
|
||||
if i != -1:
|
||||
osf = False
|
||||
@ -1391,12 +1390,12 @@ def revert_minionpro(document):
|
||||
|
||||
|
||||
def revert_mathfonts(document):
|
||||
" Revert native math font definitions to LaTeX "
|
||||
" Revert native math font definitions to LaTeX "
|
||||
|
||||
i = find_token(document.header, "\\font_math", 0)
|
||||
if i == -1:
|
||||
return
|
||||
if find_token(document.header, "\\use_non_tex_fonts false", 0) != -1:
|
||||
if find_token(document.header, "\\use_non_tex_fonts false", 0) != -1:
|
||||
val = get_value(document.header, "\\font_math", i)
|
||||
if val == "eulervm":
|
||||
add_to_preamble(document, "\\usepackage{eulervm}")
|
||||
@ -1426,9 +1425,9 @@ def revert_mathfonts(document):
|
||||
|
||||
|
||||
def revert_mdnomath(document):
|
||||
" Revert mathdesign and fourier without math "
|
||||
" Revert mathdesign and fourier without math "
|
||||
|
||||
if find_token(document.header, "\\use_non_tex_fonts false", 0) != -1:
|
||||
if find_token(document.header, "\\use_non_tex_fonts false", 0) != -1:
|
||||
mathdesign_dict = {
|
||||
"md-charter": "mdbch",
|
||||
"md-utopia": "mdput",
|
||||
@ -1455,9 +1454,9 @@ def convert_mathfonts(document):
|
||||
|
||||
|
||||
def convert_mdnomath(document):
|
||||
" Change mathdesign font name "
|
||||
" Change mathdesign font name "
|
||||
|
||||
if find_token(document.header, "\\use_non_tex_fonts false", 0) != -1:
|
||||
if find_token(document.header, "\\use_non_tex_fonts false", 0) != -1:
|
||||
mathdesign_dict = {
|
||||
"mdbch": "md-charter",
|
||||
"mdput": "md-utopia",
|
||||
@ -1472,12 +1471,12 @@ def convert_mdnomath(document):
|
||||
|
||||
|
||||
def revert_newtxmath(document):
|
||||
" Revert native newtxmath definitions to LaTeX "
|
||||
" Revert native newtxmath definitions to LaTeX "
|
||||
|
||||
i = find_token(document.header, "\\font_math", 0)
|
||||
if i == -1:
|
||||
return
|
||||
if find_token(document.header, "\\use_non_tex_fonts false", 0) != -1:
|
||||
if find_token(document.header, "\\use_non_tex_fonts false", 0) != -1:
|
||||
val = get_value(document.header, "\\font_math", i)
|
||||
mathfont_dict = {
|
||||
"libertine-ntxm": "\\usepackage[libertine]{newtxmath}",
|
||||
@ -1490,9 +1489,9 @@ def revert_newtxmath(document):
|
||||
|
||||
|
||||
def revert_biolinum(document):
|
||||
" Revert native biolinum font definition to LaTeX "
|
||||
" Revert native biolinum font definition to LaTeX "
|
||||
|
||||
if find_token(document.header, "\\use_non_tex_fonts false", 0) != -1:
|
||||
if find_token(document.header, "\\use_non_tex_fonts false", 0) != -1:
|
||||
i = find_token(document.header, "\\font_sans biolinum", 0)
|
||||
if i != -1:
|
||||
osf = False
|
||||
@ -1595,7 +1594,7 @@ def convert_latexargs(document):
|
||||
document.body[i] = "\\begin_inset Argument 999"
|
||||
i += 1
|
||||
continue
|
||||
|
||||
|
||||
# Find containing paragraph layout
|
||||
parent = get_containing_layout(document.body, i)
|
||||
if parent == False:
|
||||
@ -2174,7 +2173,7 @@ def revert_literate(document):
|
||||
|
||||
def convert_literate(document):
|
||||
" Convert Literate document to new format"
|
||||
i = find_token(document.header, "\\textclass", 0)
|
||||
i = find_token(document.header, "\\textclass", 0)
|
||||
if (i != -1) and "literate-" in document.header[i]:
|
||||
document.textclass = document.header[i].replace("\\textclass literate-", "")
|
||||
j = find_token(document.header, "\\begin_modules", 0)
|
||||
@ -2218,12 +2217,12 @@ def revert_itemargs(document):
|
||||
|
||||
|
||||
def revert_garamondx_newtxmath(document):
|
||||
" Revert native garamond newtxmath definition to LaTeX "
|
||||
" Revert native garamond newtxmath definition to LaTeX "
|
||||
|
||||
i = find_token(document.header, "\\font_math", 0)
|
||||
if i == -1:
|
||||
return
|
||||
if find_token(document.header, "\\use_non_tex_fonts false", 0) != -1:
|
||||
if find_token(document.header, "\\use_non_tex_fonts false", 0) != -1:
|
||||
val = get_value(document.header, "\\font_math", i)
|
||||
if val == "garamondx-ntxm":
|
||||
add_to_preamble(document, "\\usepackage[garamondx]{newtxmath}")
|
||||
@ -2231,9 +2230,9 @@ def revert_garamondx_newtxmath(document):
|
||||
|
||||
|
||||
def revert_garamondx(document):
|
||||
" Revert native garamond font definition to LaTeX "
|
||||
" Revert native garamond font definition to LaTeX "
|
||||
|
||||
if find_token(document.header, "\\use_non_tex_fonts false", 0) != -1:
|
||||
if find_token(document.header, "\\use_non_tex_fonts false", 0) != -1:
|
||||
i = find_token(document.header, "\\font_roman garamondx", 0)
|
||||
if i != -1:
|
||||
osf = False
|
||||
@ -2250,7 +2249,7 @@ def revert_garamondx(document):
|
||||
|
||||
def convert_beamerargs(document):
|
||||
" Converts beamer arguments to new layout "
|
||||
|
||||
|
||||
beamer_classes = ["beamer", "article-beamer", "scrarticle-beamer"]
|
||||
if document.textclass not in beamer_classes:
|
||||
return
|
||||
@ -2421,7 +2420,7 @@ def convert_againframe_args(document):
|
||||
beamer_classes = ["beamer", "article-beamer", "scrarticle-beamer"]
|
||||
if document.textclass not in beamer_classes:
|
||||
return
|
||||
|
||||
|
||||
i = 0
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_layout AgainFrame", i)
|
||||
@ -2443,11 +2442,11 @@ def convert_againframe_args(document):
|
||||
|
||||
def convert_corollary_args(document):
|
||||
" Converts beamer corrolary-style ERT arguments native InsetArgs "
|
||||
|
||||
|
||||
beamer_classes = ["beamer", "article-beamer", "scrarticle-beamer"]
|
||||
if document.textclass not in beamer_classes:
|
||||
return
|
||||
|
||||
|
||||
corollary_layouts = ["Corollary", "Definition", "Definitions", "Example", "Examples", "Fact", "Proof", "Theorem"]
|
||||
for lay in corollary_layouts:
|
||||
i = 0
|
||||
@ -2533,11 +2532,11 @@ def convert_corollary_args(document):
|
||||
|
||||
def convert_quote_args(document):
|
||||
" Converts beamer quote style ERT args to native InsetArgs "
|
||||
|
||||
|
||||
beamer_classes = ["beamer", "article-beamer", "scrarticle-beamer"]
|
||||
if document.textclass not in beamer_classes:
|
||||
return
|
||||
|
||||
|
||||
quote_layouts = ["Uncover", "Only", "Quotation", "Quote", "Verse"]
|
||||
for lay in quote_layouts:
|
||||
i = 0
|
||||
@ -2602,7 +2601,7 @@ def cleanup_beamerargs(document):
|
||||
|
||||
def revert_beamerargs(document):
|
||||
" Reverts beamer arguments to old layout "
|
||||
|
||||
|
||||
beamer_classes = ["beamer", "article-beamer", "scrarticle-beamer"]
|
||||
if document.textclass not in beamer_classes:
|
||||
return
|
||||
@ -2793,13 +2792,13 @@ def revert_beamerargs(document):
|
||||
del document.body[p : endInset + 1]
|
||||
subst = put_cmd_in_ert("[") + content + put_cmd_in_ert("]")
|
||||
document.body[realparbeg : realparbeg] = subst
|
||||
|
||||
|
||||
i = realparend
|
||||
|
||||
|
||||
def revert_beamerargs2(document):
|
||||
" Reverts beamer arguments to old layout, step 2 "
|
||||
|
||||
|
||||
beamer_classes = ["beamer", "article-beamer", "scrarticle-beamer"]
|
||||
if document.textclass not in beamer_classes:
|
||||
return
|
||||
@ -2833,7 +2832,7 @@ def revert_beamerargs2(document):
|
||||
if m:
|
||||
argnr = m.group(1)
|
||||
if argnr == "2":
|
||||
document.body[p] = "\\begin_inset Argument 1"
|
||||
document.body[p] = "\\begin_inset Argument 1"
|
||||
if layoutname in corollary_layouts:
|
||||
m = rx.match(document.body[p])
|
||||
if m:
|
||||
@ -2884,7 +2883,7 @@ def revert_beamerargs2(document):
|
||||
|
||||
def revert_beamerargs3(document):
|
||||
" Reverts beamer arguments to old layout, step 3 "
|
||||
|
||||
|
||||
beamer_classes = ["beamer", "article-beamer", "scrarticle-beamer"]
|
||||
if document.textclass not in beamer_classes:
|
||||
return
|
||||
@ -2930,7 +2929,7 @@ def revert_beamerargs3(document):
|
||||
|
||||
def revert_beamerflex(document):
|
||||
" Reverts beamer Flex insets "
|
||||
|
||||
|
||||
beamer_classes = ["beamer", "article-beamer", "scrarticle-beamer"]
|
||||
if document.textclass not in beamer_classes:
|
||||
return
|
||||
@ -3030,13 +3029,13 @@ def revert_beamerflex(document):
|
||||
document.body[i : beginPlain + 1] = pre
|
||||
post = put_cmd_in_ert("}")
|
||||
document.body[z - 2 : z + 1] = post
|
||||
|
||||
|
||||
i += 1
|
||||
|
||||
|
||||
def revert_beamerblocks(document):
|
||||
" Reverts beamer block arguments to ERT "
|
||||
|
||||
|
||||
beamer_classes = ["beamer", "article-beamer", "scrarticle-beamer"]
|
||||
if document.textclass not in beamer_classes:
|
||||
return
|
||||
@ -3096,11 +3095,11 @@ def revert_beamerblocks(document):
|
||||
|
||||
def convert_beamerblocks(document):
|
||||
" Converts beamer block ERT args to native InsetArgs "
|
||||
|
||||
|
||||
beamer_classes = ["beamer", "article-beamer", "scrarticle-beamer"]
|
||||
if document.textclass not in beamer_classes:
|
||||
return
|
||||
|
||||
|
||||
blocks = ["Block", "ExampleBlock", "AlertBlock"]
|
||||
for lay in blocks:
|
||||
i = 0
|
||||
@ -3211,7 +3210,7 @@ def convert_beamerblocks(document):
|
||||
document.body[ertcontlastline : ertcontlastline + 1] = [
|
||||
document.body[ertcontlastline], '\\end_layout', '', '\\end_inset']
|
||||
document.body[ertcontdivline : ertcontdivlinetwo + 1] = [document.body[ertcontdivline][:tok],
|
||||
'\\end_layout', '', '\\end_inset', '', '\\end_layout', '',
|
||||
'\\end_layout', '', '\\end_inset', '', '\\end_layout', '',
|
||||
'\\end_inset', '', '', '\\begin_inset Argument 2',
|
||||
'status collapsed', '', '\\begin_layout Plain Layout',
|
||||
'\\begin_inset ERT', '', 'status open' '', '\\begin_layout Plain Layout',
|
||||
@ -3300,7 +3299,7 @@ def convert_beamerblocks(document):
|
||||
document.body[ertcontlastline : ertcontlastline + 1] = [
|
||||
document.body[ertcontlastline], '\\end_layout', '', '\\end_inset']
|
||||
document.body[ertcontdivline : ertcontdivlinetwo + 1] = [document.body[ertcontdivline][:tok],
|
||||
'\\end_layout', '', '\\end_inset', '', '\\end_layout', '',
|
||||
'\\end_layout', '', '\\end_inset', '', '\\end_layout', '',
|
||||
'\\end_inset', '', '', '\\begin_inset Argument 1',
|
||||
'status collapsed', '', '\\begin_layout Plain Layout',
|
||||
'\\begin_inset ERT', '', 'status open' '', '\\begin_layout Plain Layout',
|
||||
@ -3346,7 +3345,7 @@ def convert_beamerblocks(document):
|
||||
|
||||
def convert_overprint(document):
|
||||
" Convert old beamer overprint layouts to ERT "
|
||||
|
||||
|
||||
beamer_classes = ["beamer", "article-beamer", "scrarticle-beamer"]
|
||||
if document.textclass not in beamer_classes:
|
||||
return
|
||||
@ -3386,11 +3385,11 @@ def convert_overprint(document):
|
||||
# Remove arg inset
|
||||
del document.body[argbeg : argend + 1]
|
||||
subst += put_cmd_in_ert("[") + content + put_cmd_in_ert("]")
|
||||
|
||||
|
||||
endseq = endseq - len(document.body[i : i])
|
||||
document.body[i : i] = subst + ["\\end_layout"]
|
||||
endseq += len(subst)
|
||||
|
||||
|
||||
for p in range(i, endseq):
|
||||
if document.body[p] == "\\begin_layout Overprint":
|
||||
document.body[p] = "\\begin_layout Standard"
|
||||
@ -3400,7 +3399,7 @@ def convert_overprint(document):
|
||||
|
||||
def revert_overprint(document):
|
||||
" Revert old beamer overprint layouts to ERT "
|
||||
|
||||
|
||||
beamer_classes = ["beamer", "article-beamer", "scrarticle-beamer"]
|
||||
if document.textclass not in beamer_classes:
|
||||
return
|
||||
@ -3452,11 +3451,11 @@ def revert_overprint(document):
|
||||
# Remove arg inset
|
||||
del document.body[argbeg : argend + 1]
|
||||
subst += put_cmd_in_ert("[") + content + put_cmd_in_ert("]")
|
||||
|
||||
|
||||
endseq = endseq - len(document.body[i : i])
|
||||
document.body[i : i] = subst + ["\\end_layout"]
|
||||
endseq += len(subst)
|
||||
|
||||
|
||||
p = i
|
||||
while True:
|
||||
if p >= endseq:
|
||||
@ -3492,7 +3491,7 @@ def revert_overprint(document):
|
||||
|
||||
def revert_frametitle(document):
|
||||
" Reverts beamer frametitle layout to ERT "
|
||||
|
||||
|
||||
beamer_classes = ["beamer", "article-beamer", "scrarticle-beamer"]
|
||||
if document.textclass not in beamer_classes:
|
||||
return
|
||||
@ -3538,7 +3537,7 @@ def revert_frametitle(document):
|
||||
# Remove arg inset
|
||||
del document.body[p : endInset + 1]
|
||||
subst += put_cmd_in_ert("[") + content + put_cmd_in_ert("]")
|
||||
|
||||
|
||||
subst += put_cmd_in_ert("{")
|
||||
document.body[i : i + 1] = subst
|
||||
i = endlay
|
||||
@ -3546,7 +3545,7 @@ def revert_frametitle(document):
|
||||
|
||||
def convert_epigraph(document):
|
||||
" Converts memoir epigraph to new syntax "
|
||||
|
||||
|
||||
if document.textclass != "memoir":
|
||||
return
|
||||
|
||||
@ -3581,13 +3580,13 @@ def convert_epigraph(document):
|
||||
endlay += len(begsubst) + len(endsubst)
|
||||
endlay = endlay - len(document.body[ert : endInset + 1])
|
||||
del document.body[ert : endInset + 1]
|
||||
|
||||
|
||||
i = endlay
|
||||
|
||||
|
||||
def revert_epigraph(document):
|
||||
" Reverts memoir epigraph argument to ERT "
|
||||
|
||||
|
||||
if document.textclass != "memoir":
|
||||
return
|
||||
|
||||
@ -3616,14 +3615,14 @@ def revert_epigraph(document):
|
||||
subst += put_cmd_in_ert("}{") + content
|
||||
else:
|
||||
subst += put_cmd_in_ert("}{")
|
||||
|
||||
|
||||
document.body[j : j] = subst + document.body[j : j]
|
||||
i = endlay
|
||||
|
||||
|
||||
def convert_captioninsets(document):
|
||||
" Converts caption insets to new syntax "
|
||||
|
||||
|
||||
i = 0
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_inset Caption", i)
|
||||
@ -3635,7 +3634,7 @@ def convert_captioninsets(document):
|
||||
|
||||
def revert_captioninsets(document):
|
||||
" Reverts caption insets to old syntax "
|
||||
|
||||
|
||||
i = 0
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_inset Caption Standard", i)
|
||||
@ -3678,7 +3677,7 @@ def convert_captionlayouts(document):
|
||||
|
||||
def revert_captionlayouts(document):
|
||||
" Revert caption insets to caption layouts. "
|
||||
|
||||
|
||||
caption_dict = {
|
||||
"Above" : "Captionabove",
|
||||
"Below" : "Captionbelow",
|
||||
@ -3687,7 +3686,7 @@ def revert_captionlayouts(document):
|
||||
"Centered" : "CenteredCaption",
|
||||
"Bicaption" : "Bicaption",
|
||||
}
|
||||
|
||||
|
||||
i = 0
|
||||
rx = re.compile(r'^\\begin_inset Caption (\S+)$')
|
||||
while True:
|
||||
@ -3702,7 +3701,7 @@ def revert_captionlayouts(document):
|
||||
if val not in list(caption_dict.keys()):
|
||||
i += 1
|
||||
continue
|
||||
|
||||
|
||||
# We either need to delete the previous \begin_layout line, or we
|
||||
# need to end the previous layout if this inset is not in the first
|
||||
# position of the paragraph.
|
||||
@ -3770,7 +3769,7 @@ def revert_captionlayouts(document):
|
||||
|
||||
def revert_fragileframe(document):
|
||||
" Reverts beamer FragileFrame layout to ERT "
|
||||
|
||||
|
||||
beamer_classes = ["beamer", "article-beamer", "scrarticle-beamer"]
|
||||
if document.textclass not in beamer_classes:
|
||||
return
|
||||
@ -3852,14 +3851,14 @@ def revert_fragileframe(document):
|
||||
subst += put_cmd_in_ert("{") + content + put_cmd_in_ert("}")
|
||||
elif p == 3:
|
||||
subst += put_cmd_in_ert("[fragile]")
|
||||
|
||||
|
||||
document.body[i : i + 1] = subst
|
||||
i = j
|
||||
|
||||
|
||||
def revert_newframes(document):
|
||||
" Reverts beamer Frame and PlainFrame layouts to old forms "
|
||||
|
||||
|
||||
beamer_classes = ["beamer", "article-beamer", "scrarticle-beamer"]
|
||||
if document.textclass not in beamer_classes:
|
||||
return
|
||||
@ -3954,7 +3953,7 @@ def revert_newframes(document):
|
||||
# Remove arg inset
|
||||
del document.body[arg : endInset + 1]
|
||||
subst += content
|
||||
|
||||
|
||||
document.body[i : i + 1] = subst
|
||||
i = j
|
||||
|
||||
@ -4079,18 +4078,18 @@ def revert_IEEEtran_3(document):
|
||||
|
||||
def revert_kurier_fonts(document):
|
||||
" Revert kurier font definition to LaTeX "
|
||||
|
||||
|
||||
i = find_token(document.header, "\\font_math", 0)
|
||||
if i != -1:
|
||||
if find_token(document.header, "\\use_non_tex_fonts false", 0) != -1:
|
||||
if find_token(document.header, "\\use_non_tex_fonts false", 0) != -1:
|
||||
val = get_value(document.header, "\\font_math", i)
|
||||
if val == "kurier-math":
|
||||
add_to_preamble(document, "\\let\\Myrmdefault\\rmdefault\n" \
|
||||
"\\usepackage[math]{kurier}\n" \
|
||||
"\\renewcommand{\\rmdefault}{\\Myrmdefault}")
|
||||
document.header[i] = "\\font_math auto"
|
||||
|
||||
if find_token(document.header, "\\use_non_tex_fonts false", 0) != -1:
|
||||
|
||||
if find_token(document.header, "\\use_non_tex_fonts false", 0) != -1:
|
||||
kurier_fonts = ["kurier", "kurierc", "kurierl", "kurierlc"]
|
||||
k = find_token(document.header, "\\font_sans kurier", 0)
|
||||
if k != -1:
|
||||
@ -4101,18 +4100,18 @@ def revert_kurier_fonts(document):
|
||||
|
||||
def revert_iwona_fonts(document):
|
||||
" Revert iwona font definition to LaTeX "
|
||||
|
||||
|
||||
i = find_token(document.header, "\\font_math", 0)
|
||||
if i != -1:
|
||||
if find_token(document.header, "\\use_non_tex_fonts false", 0) != -1:
|
||||
if find_token(document.header, "\\use_non_tex_fonts false", 0) != -1:
|
||||
val = get_value(document.header, "\\font_math", i)
|
||||
if val == "iwona-math":
|
||||
add_to_preamble(document, "\\let\\Myrmdefault\\rmdefault\n" \
|
||||
"\\usepackage[math]{iwona}\n" \
|
||||
"\\renewcommand{\\rmdefault}{\\Myrmdefault}")
|
||||
document.header[i] = "\\font_math auto"
|
||||
|
||||
if find_token(document.header, "\\use_non_tex_fonts false", 0) != -1:
|
||||
|
||||
if find_token(document.header, "\\use_non_tex_fonts false", 0) != -1:
|
||||
iwona_fonts = ["iwona", "iwonac", "iwonal", "iwonalc"]
|
||||
k = find_token(document.header, "\\font_sans iwona", 0)
|
||||
if k != -1:
|
||||
@ -4124,7 +4123,7 @@ def revert_iwona_fonts(document):
|
||||
|
||||
def revert_new_libertines(document):
|
||||
" Revert new libertine font definition to LaTeX "
|
||||
|
||||
|
||||
if find_token(document.header, "\\use_non_tex_fonts true", 0) != -1:
|
||||
return
|
||||
|
||||
@ -4140,7 +4139,7 @@ def revert_new_libertines(document):
|
||||
preamble += "{libertineMono-type1}"
|
||||
add_to_preamble(document, [preamble])
|
||||
document.header[i] = "\\font_typewriter default"
|
||||
|
||||
|
||||
k = find_token(document.header, "\\font_sans biolinum", 0)
|
||||
if k != -1:
|
||||
preamble = "\\usepackage"
|
||||
@ -4165,11 +4164,11 @@ def revert_new_libertines(document):
|
||||
|
||||
def convert_lyxframes(document):
|
||||
" Converts old beamer frames to new style "
|
||||
|
||||
|
||||
beamer_classes = ["beamer", "article-beamer", "scrarticle-beamer"]
|
||||
if document.textclass not in beamer_classes:
|
||||
return
|
||||
|
||||
|
||||
framebeg = ["BeginFrame", "BeginPlainFrame"]
|
||||
frameend = ["Frame", "PlainFrame", "EndFrame", "BeginFrame", "BeginPlainFrame", "AgainFrame",
|
||||
"Section", "Section*", "Subsection", "Subsection*", "Subsubsection", "Subsubsection*"]
|
||||
@ -4238,11 +4237,11 @@ def convert_lyxframes(document):
|
||||
|
||||
def remove_endframes(document):
|
||||
" Remove deprecated beamer endframes "
|
||||
|
||||
|
||||
beamer_classes = ["beamer", "article-beamer", "scrarticle-beamer"]
|
||||
if document.textclass not in beamer_classes:
|
||||
return
|
||||
|
||||
|
||||
i = 0
|
||||
while True:
|
||||
i = find_token_exact(document.body, "\\begin_layout EndFrame", i)
|
||||
@ -4258,7 +4257,7 @@ def remove_endframes(document):
|
||||
|
||||
def revert_powerdot_flexes(document):
|
||||
" Reverts powerdot flex insets "
|
||||
|
||||
|
||||
if document.textclass != "powerdot":
|
||||
return
|
||||
|
||||
@ -4306,13 +4305,13 @@ def revert_powerdot_flexes(document):
|
||||
z += len(pre)
|
||||
document.body[i : beginPlain + 1] = pre
|
||||
post = put_cmd_in_ert("}")
|
||||
document.body[z - 2 : z + 1] = post
|
||||
document.body[z - 2 : z + 1] = post
|
||||
i += 1
|
||||
|
||||
|
||||
def revert_powerdot_pause(document):
|
||||
" Reverts powerdot pause layout to ERT "
|
||||
|
||||
|
||||
if document.textclass != "powerdot":
|
||||
return
|
||||
|
||||
@ -4342,14 +4341,14 @@ def revert_powerdot_pause(document):
|
||||
# Remove arg inset
|
||||
del document.body[p : endInset + 1]
|
||||
subst += put_cmd_in_ert("[") + content + put_cmd_in_ert("]")
|
||||
|
||||
|
||||
document.body[i : i + 1] = subst
|
||||
i = endlay
|
||||
|
||||
|
||||
def revert_powerdot_itemargs(document):
|
||||
" Reverts powerdot item arguments to ERT "
|
||||
|
||||
|
||||
if document.textclass != "powerdot":
|
||||
return
|
||||
|
||||
@ -4398,7 +4397,7 @@ def revert_powerdot_itemargs(document):
|
||||
del document.body[i:j+1]
|
||||
subst = put_cmd_in_ert("<") + content + put_cmd_in_ert(">")
|
||||
document.body[realparbeg : realparbeg] = subst
|
||||
|
||||
|
||||
i = realparend
|
||||
|
||||
|
||||
@ -4448,7 +4447,7 @@ def revert_powerdot_columns(document):
|
||||
# Remove arg inset
|
||||
del document.body[p : endInset + 1]
|
||||
subst += put_cmd_in_ert("{") + content + put_cmd_in_ert("}")
|
||||
|
||||
|
||||
subst += put_cmd_in_ert("{")
|
||||
document.body[i : i + 1] = subst
|
||||
i = endlay
|
||||
@ -4485,7 +4484,7 @@ def revert_mbox_fbox(document):
|
||||
|
||||
def revert_starred_caption(document):
|
||||
" Reverts unnumbered longtable caption insets "
|
||||
|
||||
|
||||
i = 0
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_inset Caption LongTableNoNumber", i)
|
||||
@ -4560,15 +4559,15 @@ def revert_aa2(document):
|
||||
|
||||
|
||||
def revert_tibetan(document):
|
||||
"Set the document language for Tibetan to English"
|
||||
"Set the document language for Tibetan to English"
|
||||
|
||||
if document.language == "tibetan":
|
||||
document.language = "english"
|
||||
i = find_token(document.header, "\\language", 0)
|
||||
if i != -1:
|
||||
document.header[i] = "\\language english"
|
||||
i = find_token(document.header, "\\language", 0)
|
||||
if i != -1:
|
||||
document.header[i] = "\\language english"
|
||||
j = 0
|
||||
while j < len(document.body):
|
||||
while j < len(document.body):
|
||||
j = find_token(document.body, "\\lang tibetan", j)
|
||||
if j != -1:
|
||||
document.body[j] = document.body[j].replace("\\lang tibetan", "\\lang english")
|
||||
@ -4612,14 +4611,14 @@ def convert_chunks(document):
|
||||
# there is no point continuing, as we will run into the same error again.
|
||||
return
|
||||
this_chunk = "".join(document.body[i + 1:j])
|
||||
|
||||
|
||||
# there may be empty lines between chunks
|
||||
# we just skip them.
|
||||
if not chunk_started:
|
||||
if this_chunk != "":
|
||||
# new chunk starts
|
||||
chunk_started = True
|
||||
|
||||
|
||||
if chunk_started:
|
||||
contents.append(document.body[i + 1:j])
|
||||
|
||||
@ -4666,7 +4665,7 @@ def convert_chunks(document):
|
||||
# the first par (separated from the options by a newline).
|
||||
# We collect such stuff to re-insert it later.
|
||||
postoptstuff = []
|
||||
|
||||
|
||||
match = first_re.search(optarg)
|
||||
if match:
|
||||
optarg = match.groups()[0]
|
||||
@ -4806,7 +4805,7 @@ def revert_chunks(document):
|
||||
# replace old content with new content
|
||||
document.body[lstart : lend + 1] = newlines
|
||||
i = lstart + len(newlines)
|
||||
|
||||
|
||||
|
||||
##
|
||||
# Conversion hub
|
||||
|
@ -1,6 +1,5 @@
|
||||
# -*- coding: utf-8 -*-
|
||||
# This file is part of lyx2lyx
|
||||
# -*- coding: utf-8 -*-
|
||||
# Copyright (C) 2015 The LyX team
|
||||
#
|
||||
# This program is free software; you can redistribute it and/or
|
||||
@ -78,7 +77,7 @@ def revert_Argument_to_TeX_brace(document, line, endline, n, nmax, environment,
|
||||
if nolastopt == False:
|
||||
document.body[endInset - 2 : endInset + 1] = put_cmd_in_ert("}{")
|
||||
else:
|
||||
document.body[endInset - 2 : endInset + 1] = put_cmd_in_ert("}")
|
||||
document.body[endInset - 2 : endInset + 1] = put_cmd_in_ert("}")
|
||||
del(document.body[lineArg : beginPlain + 1])
|
||||
wasOpt = False
|
||||
else:
|
||||
@ -153,7 +152,7 @@ def convert_separator(document):
|
||||
}
|
||||
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_deeper", i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -175,7 +174,7 @@ def convert_separator(document):
|
||||
i = i + 1
|
||||
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\align", i)
|
||||
if i == -1:
|
||||
break
|
||||
@ -187,6 +186,13 @@ def convert_separator(document):
|
||||
|
||||
j = find_token_backwards(document.body, "\\end_layout", i-1)
|
||||
if j != -1:
|
||||
# Very old LyX files do not have Plain Layout in insets (but Standard).
|
||||
# So we additionally check here if there is no inset boundary
|
||||
# between the previous layout and this one.
|
||||
n = find_token(document.body, "\\end_inset", j, lay[1])
|
||||
if n != -1:
|
||||
i = i + 1
|
||||
continue
|
||||
lay = get_containing_layout(document.body, j-1)
|
||||
if lay != False and lay[0] == "Standard" \
|
||||
and find_token(document.body, "\\align", lay[1], lay[2]) == -1 \
|
||||
@ -208,7 +214,7 @@ def convert_separator(document):
|
||||
regexp = re.compile(r'^\\begin_layout (?:(-*)|(\s*))(Separator|EndOfSlide)(?:(-*)|(\s*))$', re.IGNORECASE)
|
||||
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_re(document.body, regexp, i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -249,7 +255,7 @@ def revert_separator(document):
|
||||
"", "\\end_inset", ""]
|
||||
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_inset Separator", i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -341,7 +347,7 @@ def convert_parbreak(document):
|
||||
"""
|
||||
parbreakinset = "\\begin_inset Separator parbreak"
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, parbreakinset, i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -366,7 +372,7 @@ def revert_parbreak(document):
|
||||
Revert latexpar separators to parbreak separators.
|
||||
"""
|
||||
i = 0
|
||||
while 1:
|
||||
while True:
|
||||
i = find_token(document.body, "\\begin_inset Separator latexpar", i)
|
||||
if i == -1:
|
||||
return
|
||||
@ -619,12 +625,15 @@ def convert_dashes(document):
|
||||
i = 0
|
||||
while i < len(document.body):
|
||||
words = document.body[i].split()
|
||||
if len(words) > 1 and words[0] == "\\begin_inset" and \
|
||||
words[1] in ["CommandInset", "ERT", "External", "Formula", "FormulaMacro", "Graphics", "IPA", "listings"]:
|
||||
if (len(words) > 1 and words[0] == "\\begin_inset"
|
||||
and (words[1] in ["CommandInset", "ERT", "External", "Formula",
|
||||
"FormulaMacro", "Graphics", "IPA", "listings"]
|
||||
or ' '.join(words[1:]) == "Flex Code")):
|
||||
# must not replace anything in insets that store LaTeX contents in .lyx files
|
||||
# (math and command insets withut overridden read() and write() methods
|
||||
# (math and command insets without overridden read() and write() methods
|
||||
# filtering out IPA makes Text::readParToken() more simple
|
||||
# skip ERT as well since it is not needed there
|
||||
# Flex Code is logical markup, typically rendered as typewriter
|
||||
j = find_end_of_inset(document.body, i)
|
||||
if j == -1:
|
||||
document.warning("Malformed LyX document: Can't find end of " + words[1] + " inset at line " + str(i))
|
||||
@ -632,6 +641,16 @@ def convert_dashes(document):
|
||||
else:
|
||||
i = j
|
||||
continue
|
||||
if document.body[i] == "\\begin_layout LyX-Code":
|
||||
j = find_end_of_layout(document.body, i)
|
||||
if j == -1:
|
||||
document.warning("Malformed LyX document: "
|
||||
"Can't find end of %s layout at line %d" % (words[1],i))
|
||||
i += 1
|
||||
else:
|
||||
i = j
|
||||
continue
|
||||
|
||||
if len(words) > 0 and words[0] in ["\\leftindent", "\\paragraph_spacing", "\\align", "\\labelwidthstring"]:
|
||||
# skip paragraph parameters (bug 10243)
|
||||
i += 1
|
||||
@ -660,6 +679,12 @@ def convert_dashes(document):
|
||||
def revert_dashes(document):
|
||||
"convert \\twohyphens and \\threehyphens to -- and ---"
|
||||
|
||||
# eventually remove preamble code from 2.3->2.2 conversion:
|
||||
for i, line in enumerate(document.preamble):
|
||||
if i > 1 and line == r'\renewcommand{\textemdash}{---}':
|
||||
if (document.preamble[i-1] == r'\renewcommand{\textendash}{--}'
|
||||
and document.preamble[i-2] == '% Added by lyx2lyx'):
|
||||
del document.preamble[i-2:i+1]
|
||||
i = 0
|
||||
while i < len(document.body):
|
||||
words = document.body[i].split()
|
||||
@ -800,7 +825,7 @@ def convert_specialchar_internal(document, forward):
|
||||
else:
|
||||
i = j
|
||||
continue
|
||||
for key, value in specialchars.iteritems():
|
||||
for key, value in specialchars.items():
|
||||
if forward:
|
||||
document.body[i] = document.body[i].replace("\\SpecialChar " + key, "\\SpecialChar " + value)
|
||||
document.body[i] = document.body[i].replace("\\SpecialCharNoPassThru " + key, "\\SpecialCharNoPassThru " + value)
|
||||
@ -1151,11 +1176,11 @@ def convert_origin(document):
|
||||
if i == -1:
|
||||
document.warning("Malformed LyX document: No \\textclass!!")
|
||||
return
|
||||
if document.dir == "":
|
||||
origin = "stdin"
|
||||
if document.dir == u'':
|
||||
origin = u'stdin'
|
||||
else:
|
||||
relpath = ''
|
||||
if document.systemlyxdir and document.systemlyxdir != '':
|
||||
relpath = u''
|
||||
if document.systemlyxdir and document.systemlyxdir != u'':
|
||||
try:
|
||||
if os.path.isabs(document.dir):
|
||||
absdir = os.path.normpath(document.dir)
|
||||
@ -1166,15 +1191,14 @@ def convert_origin(document):
|
||||
else:
|
||||
abssys = os.path.normpath(os.path.abspath(document.systemlyxdir))
|
||||
relpath = os.path.relpath(absdir, abssys)
|
||||
if relpath.find('..') == 0:
|
||||
relpath = ''
|
||||
if relpath.find(u'..') == 0:
|
||||
relpath = u''
|
||||
except:
|
||||
relpath = ''
|
||||
if relpath == '':
|
||||
origin = document.dir.replace('\\', '/') + '/'
|
||||
relpath = u''
|
||||
if relpath == u'':
|
||||
origin = document.dir.replace(u'\\', u'/') + u'/'
|
||||
else:
|
||||
origin = os.path.join("/systemlyxdir", relpath).replace('\\', '/') + '/'
|
||||
origin = unicode(origin, sys.getfilesystemencoding())
|
||||
origin = os.path.join(u"/systemlyxdir", relpath).replace(u'\\', u'/') + u'/'
|
||||
document.header[i:i] = ["\\origin " + origin]
|
||||
|
||||
|
||||
@ -1213,7 +1237,7 @@ def revert_textcolor(document):
|
||||
j = find_token(document.body, "\\color", i + 1)
|
||||
k = find_token(document.body, "\\end_layout", i + 1)
|
||||
if j == -1 and k != -1:
|
||||
j = k +1
|
||||
j = k +1
|
||||
# output TeX code
|
||||
# first output the closing brace
|
||||
if k < j:
|
||||
@ -1239,7 +1263,7 @@ def convert_colorbox(document):
|
||||
|
||||
def revert_colorbox(document):
|
||||
" outputs color settings for boxes as TeX code "
|
||||
|
||||
|
||||
i = 0
|
||||
defaultframecolor = "black"
|
||||
defaultbackcolor = "none"
|
||||
@ -1883,7 +1907,7 @@ def revert_tcolorbox_8(document):
|
||||
|
||||
def revert_moderncv_1(document):
|
||||
" Reverts the new inset of moderncv to TeX-code in preamble "
|
||||
|
||||
|
||||
if document.textclass != "moderncv":
|
||||
return
|
||||
i = 0
|
||||
@ -1951,7 +1975,7 @@ def revert_moderncv_1(document):
|
||||
|
||||
def revert_moderncv_2(document):
|
||||
" Reverts the phone inset of moderncv to the obsoleted mobile or fax "
|
||||
|
||||
|
||||
if document.textclass != "moderncv":
|
||||
return
|
||||
i = 0
|
||||
@ -2093,7 +2117,7 @@ def convert_moderncv_name(document):
|
||||
|
||||
def revert_achemso(document):
|
||||
" Reverts the flex inset Latin to TeX code "
|
||||
|
||||
|
||||
if document.textclass != "achemso":
|
||||
return
|
||||
i = 0
|
||||
|
2344
lib/lyx2lyx/lyx_2_3.py
Normal file
2344
lib/lyx2lyx/lyx_2_3.py
Normal file
File diff suppressed because it is too large
Load Diff
@ -64,7 +64,7 @@ get_value(lines, token, start[, end[, default]):
|
||||
and is what is returned if we do not find anything. So you
|
||||
can use that to set a default.
|
||||
|
||||
get_quoted_value(lines, token, start[, end[, default]):
|
||||
get_quoted_value(lines, token, start[, end[, default]]):
|
||||
Similar to get_value, but it will strip quotes off the
|
||||
value, if they are present. So use this one for cases
|
||||
where the value is normally quoted.
|
||||
@ -74,6 +74,9 @@ get_option_value(line, option):
|
||||
option="value"
|
||||
and returns value. Returns "" if not found.
|
||||
|
||||
get_bool_value(lines, token, start[, end[, default]]):
|
||||
Like get_value, but returns a boolean.
|
||||
|
||||
del_token(lines, token, start[, end]):
|
||||
Like find_token, but deletes the line if it finds one.
|
||||
Returns True if a line got deleted, otherwise False.
|
||||
@ -315,6 +318,25 @@ def get_quoted_value(lines, token, start, end = 0, default = ""):
|
||||
return val.strip('"')
|
||||
|
||||
|
||||
def get_bool_value(lines, token, start, end = 0, default = None):
|
||||
""" get_value(lines, token, start[[, end], default]) -> string
|
||||
|
||||
Find the next line that looks like:
|
||||
token bool_value
|
||||
|
||||
Returns True if bool_value is 1 or true and
|
||||
False if bool_value is 0 or false
|
||||
"""
|
||||
|
||||
val = get_quoted_value(lines, token, start, end, "")
|
||||
|
||||
if val == "1" or val == "true":
|
||||
return True
|
||||
if val == "0" or val == "false":
|
||||
return False
|
||||
return default
|
||||
|
||||
|
||||
def get_option_value(line, option):
|
||||
rx = option + '\s*=\s*"([^"]+)"'
|
||||
rx = re.compile(rx)
|
||||
|
@ -65,28 +65,28 @@ class TestParserTools(unittest.TestCase):
|
||||
def test_check_token(self):
|
||||
line = "\\begin_layout Standard"
|
||||
|
||||
self.assertEquals(check_token(line, '\\begin_layout'), True)
|
||||
self.assertEquals(check_token(line, 'Standard'), False)
|
||||
self.assertEqual(check_token(line, '\\begin_layout'), True)
|
||||
self.assertEqual(check_token(line, 'Standard'), False)
|
||||
|
||||
|
||||
def test_is_nonempty_line(self):
|
||||
self.assertEquals(is_nonempty_line(lines[0]), False)
|
||||
self.assertEquals(is_nonempty_line(lines[1]), True)
|
||||
self.assertEquals(is_nonempty_line(" "*5), False)
|
||||
self.assertEqual(is_nonempty_line(lines[0]), False)
|
||||
self.assertEqual(is_nonempty_line(lines[1]), True)
|
||||
self.assertEqual(is_nonempty_line(" "*5), False)
|
||||
|
||||
|
||||
def test_find_token(self):
|
||||
self.assertEquals(find_token(lines, '\\emph', 0), 7)
|
||||
self.assertEquals(find_token(lines, '\\emph', 0, 5), -1)
|
||||
self.assertEquals(find_token(lines, '\\emp', 0, 0, True), -1)
|
||||
self.assertEquals(find_token(lines, '\\emp', 0, 0, False), 7)
|
||||
self.assertEquals(find_token(lines, 'emph', 0), -1)
|
||||
self.assertEqual(find_token(lines, '\\emph', 0), 7)
|
||||
self.assertEqual(find_token(lines, '\\emph', 0, 5), -1)
|
||||
self.assertEqual(find_token(lines, '\\emp', 0, 0, True), -1)
|
||||
self.assertEqual(find_token(lines, '\\emp', 0, 0, False), 7)
|
||||
self.assertEqual(find_token(lines, 'emph', 0), -1)
|
||||
|
||||
|
||||
def test_find_tokens(self):
|
||||
tokens = ['\\emph', '\\end_inset']
|
||||
self.assertEquals(find_tokens(lines, tokens, 0), 4)
|
||||
self.assertEquals(find_tokens(lines, tokens, 0, 4), -1)
|
||||
self.assertEqual(find_tokens(lines, tokens, 0), 4)
|
||||
self.assertEqual(find_tokens(lines, tokens, 0, 4), -1)
|
||||
|
||||
|
||||
if __name__ == '__main__':
|
||||
|
@ -29,7 +29,12 @@ if not PY2:
|
||||
def read_unicodesymbols():
|
||||
" Read the unicodesymbols list of unicode characters and corresponding commands."
|
||||
pathname = os.path.abspath(os.path.dirname(sys.argv[0]))
|
||||
fp = open(os.path.join(pathname.strip('lyx2lyx'), 'unicodesymbols'))
|
||||
filename = os.path.join(pathname.strip('lyx2lyx'), 'unicodesymbols')
|
||||
|
||||
# For python 3+ we have to specify the encoding for those systems
|
||||
# where the default is not UTF-8
|
||||
fp = open(filename, encoding="utf8") if (not PY2) else open(filename)
|
||||
|
||||
spec_chars = []
|
||||
# A backslash, followed by some non-word character, and then a character
|
||||
# in brackets. The idea is to check for constructs like: \"{u}, which is how
|
||||
|
Loading…
Reference in New Issue
Block a user