2007-08-12 13:25:36 +00:00
|
|
|
# This file is part of lyx2lyx
|
|
|
|
# -*- coding: utf-8 -*-
|
2008-02-03 11:56:44 +00:00
|
|
|
# Copyright (C) 2007-2008 The LyX Team <lyx-devel@lists.lyx.org>
|
2007-08-12 13:25:36 +00:00
|
|
|
#
|
|
|
|
# This program is free software; you can redistribute it and/or
|
|
|
|
# modify it under the terms of the GNU General Public License
|
|
|
|
# as published by the Free Software Foundation; either version 2
|
|
|
|
# of the License, or (at your option) any later version.
|
|
|
|
#
|
|
|
|
# This program is distributed in the hope that it will be useful,
|
|
|
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
# GNU General Public License for more details.
|
|
|
|
#
|
|
|
|
# You should have received a copy of the GNU General Public License
|
|
|
|
# along with this program; if not, write to the Free Software
|
|
|
|
# Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA.
|
|
|
|
|
|
|
|
""" Convert files to the file format generated by lyx 1.6"""
|
|
|
|
|
|
|
|
import re
|
|
|
|
import unicodedata
|
|
|
|
import sys, os
|
|
|
|
|
2008-04-12 13:48:17 +00:00
|
|
|
from parser_tools import find_token, find_end_of, find_tokens, get_value, get_value_string
|
2007-08-12 13:25:36 +00:00
|
|
|
|
|
|
|
####################################################################
|
|
|
|
# Private helper functions
|
|
|
|
|
|
|
|
def find_end_of_inset(lines, i):
|
|
|
|
" Find end of inset, where lines[i] is included."
|
|
|
|
return find_end_of(lines, i, "\\begin_inset", "\\end_inset")
|
|
|
|
|
2008-06-02 19:42:53 +00:00
|
|
|
# WARNING!
|
|
|
|
# DO NOT do this:
|
|
|
|
# document.body[i] = wrap_insert_ert(...)
|
|
|
|
# wrap_into_ert may returns a multiline string, which should NOT appear
|
|
|
|
# in document.body. Insetad, do something like this:
|
|
|
|
# subst = wrap_inset_ert(...)
|
|
|
|
# subst = subst.split('\n')
|
|
|
|
# document.body[i:i+1] = subst
|
|
|
|
# i+= len(subst) - 1
|
|
|
|
# where the last statement resets the counter to accord with the added
|
|
|
|
# lines.
|
2007-11-21 18:18:24 +00:00
|
|
|
def wrap_into_ert(string, src, dst):
|
2008-07-08 20:25:53 +00:00
|
|
|
'''Within string, replace occurrences of src with dst, wrapped into ERT
|
|
|
|
E.g.: wrap_into_ert('sch\"on', "\\", "\\backslash") is:
|
|
|
|
sch<ERT>\\backslash</ERT>"on'''
|
2008-03-19 09:40:24 +00:00
|
|
|
return string.replace(src, '\n\\begin_inset ERT\nstatus collapsed\n\\begin_layout Standard\n'
|
2007-11-21 18:18:24 +00:00
|
|
|
+ dst + '\n\\end_layout\n\\end_inset\n')
|
|
|
|
|
2008-07-08 20:25:53 +00:00
|
|
|
def put_cmd_in_ert(string):
|
2008-12-02 07:45:34 +00:00
|
|
|
for rep in unicode_reps:
|
|
|
|
string = string.replace(rep[1], rep[0].replace('\\\\', '\\'))
|
2008-07-08 20:25:53 +00:00
|
|
|
string = string.replace('\\', "\\backslash\n")
|
|
|
|
string = "\\begin_inset ERT\nstatus collapsed\n\\begin_layout Standard\n" \
|
|
|
|
+ string + "\n\\end_layout\n\\end_inset"
|
|
|
|
return string
|
|
|
|
|
2008-01-11 18:56:53 +00:00
|
|
|
def add_to_preamble(document, text):
|
|
|
|
""" Add text to the preamble if it is not already there.
|
|
|
|
Only the first line is checked!"""
|
|
|
|
|
|
|
|
if find_token(document.preamble, text[0], 0) != -1:
|
|
|
|
return
|
|
|
|
|
|
|
|
document.preamble.extend(text)
|
|
|
|
|
2008-08-05 22:37:28 +00:00
|
|
|
def insert_to_preamble(index, document, text):
|
|
|
|
""" Insert text to the preamble at a given line"""
|
|
|
|
|
|
|
|
document.preamble.insert(index, text)
|
|
|
|
|
2008-04-11 14:07:25 +00:00
|
|
|
# Convert a LyX length into a LaTeX length
|
|
|
|
def convert_len(len):
|
|
|
|
units = {"text%":"\\backslash\ntextwidth", "col%":"\\backslash\ncolumnwidth",
|
|
|
|
"page%":"\\backslash\npagewidth", "line%":"\\backslash\nlinewidth",
|
|
|
|
"theight%":"\\backslash\ntextheight", "pheight%":"\\backslash\npageheight"}
|
|
|
|
|
|
|
|
# Convert LyX units to LaTeX units
|
|
|
|
for unit in units.keys():
|
|
|
|
if len.find(unit) != -1:
|
|
|
|
len = '%f' % (len2value(len) / 100)
|
|
|
|
len = len.strip('0') + units[unit]
|
|
|
|
break
|
|
|
|
|
|
|
|
return len
|
|
|
|
|
|
|
|
# Return the value of len without the unit in numerical form.
|
|
|
|
def len2value(len):
|
|
|
|
result = re.search('([+-]?[0-9.]+)', len)
|
|
|
|
if result:
|
|
|
|
return float(result.group(1))
|
|
|
|
# No number means 1.0
|
|
|
|
return 1.0
|
|
|
|
|
2008-04-15 17:59:01 +00:00
|
|
|
# Unfortunately, this doesn't really work, since Standard isn't always default.
|
|
|
|
# But it's as good as we can do right now.
|
2008-04-19 17:03:21 +00:00
|
|
|
def find_default_layout(document, start, end):
|
2008-04-15 17:59:01 +00:00
|
|
|
l = find_token(document.body, "\\begin_layout Standard", start, end)
|
|
|
|
if l == -1:
|
|
|
|
l = find_token(document.body, "\\begin_layout PlainLayout", start, end)
|
|
|
|
if l == -1:
|
|
|
|
l = find_token(document.body, "\\begin_layout Plain Layout", start, end)
|
|
|
|
return l
|
|
|
|
|
2008-03-18 18:08:32 +00:00
|
|
|
def get_option(document, m, option, default):
|
|
|
|
l = document.body[m].find(option)
|
|
|
|
val = default
|
|
|
|
if l != -1:
|
|
|
|
val = document.body[m][l:].split('"')[1]
|
|
|
|
return val
|
|
|
|
|
|
|
|
def remove_option(document, m, option):
|
|
|
|
l = document.body[m].find(option)
|
|
|
|
if l != -1:
|
|
|
|
val = document.body[m][l:].split('"')[1]
|
|
|
|
document.body[m] = document.body[m][:l-1] + document.body[m][l+len(option + '="' + val + '"'):]
|
|
|
|
return l
|
|
|
|
|
|
|
|
def set_option(document, m, option, value):
|
|
|
|
l = document.body[m].find(option)
|
|
|
|
if l != -1:
|
|
|
|
oldval = document.body[m][l:].split('"')[1]
|
|
|
|
l = l + len(option + '="')
|
|
|
|
document.body[m] = document.body[m][:l] + value + document.body[m][l+len(oldval):]
|
|
|
|
else:
|
|
|
|
document.body[m] = document.body[m][:-1] + ' ' + option + '="' + value + '">'
|
|
|
|
return l
|
|
|
|
|
2008-05-04 07:51:50 +00:00
|
|
|
|
2008-07-07 20:31:06 +00:00
|
|
|
def read_unicodesymbols():
|
|
|
|
" Read the unicodesymbols list of unicode characters and corresponding commands."
|
|
|
|
pathname = os.path.abspath(os.path.dirname(sys.argv[0]))
|
|
|
|
fp = open(os.path.join(pathname.strip('lyx2lyx'), 'unicodesymbols'))
|
|
|
|
spec_chars = []
|
|
|
|
# Two backslashes, followed by some non-word character, and then a character
|
|
|
|
# in brackets. The idea is to check for constructs like: \"{u}, which is how
|
|
|
|
# they are written in the unicodesymbols file; but they can also be written
|
2008-12-02 07:45:34 +00:00
|
|
|
# as: \"u or even \" u.
|
2008-07-07 20:31:06 +00:00
|
|
|
r = re.compile(r'\\\\(\W)\{(\w)\}')
|
|
|
|
for line in fp.readlines():
|
2008-07-08 20:25:53 +00:00
|
|
|
if line[0] != '#' and line.strip() != "":
|
2008-07-07 20:31:06 +00:00
|
|
|
line=line.replace(' "',' ') # remove all quotation marks with spaces before
|
|
|
|
line=line.replace('" ',' ') # remove all quotation marks with spaces after
|
|
|
|
line=line.replace(r'\"','"') # replace \" by " (for characters with diaeresis)
|
|
|
|
try:
|
|
|
|
[ucs4,command,dead] = line.split(None,2)
|
2008-07-08 20:25:53 +00:00
|
|
|
if command[0:1] != "\\":
|
|
|
|
continue
|
2008-07-07 20:31:06 +00:00
|
|
|
spec_chars.append([command, unichr(eval(ucs4))])
|
|
|
|
except:
|
|
|
|
continue
|
|
|
|
m = r.match(command)
|
|
|
|
if m != None:
|
|
|
|
command = "\\\\"
|
|
|
|
# If the character is a double-quote, then we need to escape it, too,
|
|
|
|
# since it is done that way in the LyX file.
|
|
|
|
if m.group(1) == "\"":
|
|
|
|
command += "\\"
|
2008-12-02 07:45:34 +00:00
|
|
|
commandbl = command
|
2008-07-07 20:31:06 +00:00
|
|
|
command += m.group(1) + m.group(2)
|
2008-12-02 07:45:34 +00:00
|
|
|
commandbl += m.group(1) + ' ' + m.group(2)
|
2008-07-07 20:31:06 +00:00
|
|
|
spec_chars.append([command, unichr(eval(ucs4))])
|
2008-12-02 07:45:34 +00:00
|
|
|
spec_chars.append([commandbl, unichr(eval(ucs4))])
|
2008-07-07 20:31:06 +00:00
|
|
|
fp.close()
|
|
|
|
return spec_chars
|
|
|
|
|
2008-07-07 19:13:01 +00:00
|
|
|
|
2008-07-08 20:25:53 +00:00
|
|
|
def extract_argument(line):
|
|
|
|
'Extracts a LaTeX argument from the start of line. Returns (arg, rest).'
|
|
|
|
|
|
|
|
if not line:
|
|
|
|
return (None, "")
|
|
|
|
|
|
|
|
bracere = re.compile("(\s*)(.*)")
|
|
|
|
n = bracere.match(line)
|
|
|
|
whitespace = n.group(1)
|
|
|
|
stuff = n.group(2)
|
|
|
|
brace = stuff[:1]
|
|
|
|
if brace != "[" and brace != "{":
|
|
|
|
return (None, line)
|
|
|
|
|
|
|
|
# find closing brace
|
|
|
|
remain = stuff[1:]
|
|
|
|
pos = 0
|
|
|
|
num = 1
|
|
|
|
term = "}"
|
|
|
|
if brace == "[":
|
|
|
|
term = "]"
|
|
|
|
skip = False
|
|
|
|
for c in remain:
|
|
|
|
if skip:
|
|
|
|
skip = False
|
|
|
|
elif c == "\\":
|
|
|
|
skip = True
|
|
|
|
elif c == brace:
|
|
|
|
num += 1
|
|
|
|
elif c == term:
|
|
|
|
num -= 1
|
|
|
|
if c == 0:
|
|
|
|
break
|
|
|
|
pos += 1
|
|
|
|
if num != 0:
|
|
|
|
# We never found the matching brace
|
|
|
|
# So, to be on the safe side, let's just return everything
|
|
|
|
# which will then get wrapped as ERT
|
|
|
|
return (line, "")
|
|
|
|
return (line[:pos + 1], line[pos + 1:])
|
|
|
|
|
|
|
|
|
2009-09-10 12:10:18 +00:00
|
|
|
def latex2ert(line, isindex):
|
2008-07-08 20:25:53 +00:00
|
|
|
'''Converts LaTeX commands into ERT. line may well be a multi-line
|
|
|
|
string when it is returned.'''
|
2008-07-07 21:20:30 +00:00
|
|
|
if not line:
|
|
|
|
return line
|
2008-07-08 20:25:53 +00:00
|
|
|
|
|
|
|
retval = ""
|
|
|
|
## FIXME Escaped \ ??
|
2008-07-13 19:24:09 +00:00
|
|
|
# This regex looks for a LaTeX command---i.e., something of the form
|
|
|
|
# "\alPhaStuFF", or "\X", where X is any character---where the command
|
|
|
|
# may also be preceded by an additional backslash, which is how it would
|
|
|
|
# appear (e.g.) in an InsetIndex.
|
|
|
|
labelre = re.compile(r'(.*?)\\?(\\(?:[a-zA-Z]+|.))(.*)')
|
2008-07-08 20:25:53 +00:00
|
|
|
|
|
|
|
m = labelre.match(line)
|
|
|
|
while m != None:
|
|
|
|
retval += m.group(1)
|
|
|
|
cmd = m.group(2)
|
|
|
|
end = m.group(3)
|
|
|
|
|
|
|
|
while True:
|
|
|
|
(arg, rest) = extract_argument(end)
|
|
|
|
if arg == None:
|
|
|
|
break
|
|
|
|
cmd += arg
|
|
|
|
end = rest
|
2008-07-13 19:24:09 +00:00
|
|
|
# If we wanted to put labels into an InsetLabel, for example, then we
|
|
|
|
# would just need to test here for cmd == "label" and then take some
|
2008-08-01 17:57:01 +00:00
|
|
|
# appropriate action, i.e., to use arg to get the content and then
|
2008-07-13 19:24:09 +00:00
|
|
|
# wrap it appropriately.
|
2008-07-08 20:25:53 +00:00
|
|
|
cmd = put_cmd_in_ert(cmd)
|
|
|
|
retval += "\n" + cmd + "\n"
|
|
|
|
line = end
|
|
|
|
m = labelre.match(line)
|
2008-12-02 07:45:34 +00:00
|
|
|
# put all remaining braces in ERT
|
|
|
|
line = wrap_into_ert(line, '}', '}')
|
|
|
|
line = wrap_into_ert(line, '{', '{')
|
2009-09-10 12:10:18 +00:00
|
|
|
if isindex:
|
|
|
|
# active character that is not available in all font encodings
|
|
|
|
line = wrap_into_ert(line, '|', '|')
|
2008-07-08 20:25:53 +00:00
|
|
|
retval += line
|
|
|
|
return retval
|
2008-07-07 21:20:30 +00:00
|
|
|
|
|
|
|
|
2008-10-28 16:00:42 +00:00
|
|
|
unicode_reps = read_unicodesymbols()
|
|
|
|
|
2008-08-06 12:59:54 +00:00
|
|
|
#Bug 5022....
|
|
|
|
#Might should do latex2ert first, then deal with stuff that DOESN'T
|
|
|
|
#end up inside ERT. That routine could be modified so that it returned
|
|
|
|
#a list of lines, and we could then skip ERT bits and only deal with
|
|
|
|
#the other bits.
|
2009-09-10 12:10:18 +00:00
|
|
|
def latex2lyx(data, isindex):
|
2008-08-01 17:57:01 +00:00
|
|
|
'''Takes a string, possibly multi-line, and returns the result of
|
2008-07-07 19:13:01 +00:00
|
|
|
converting LaTeX constructs into LyX constructs. Returns a list of
|
2009-09-10 12:10:18 +00:00
|
|
|
lines, suitable for insertion into document.body.
|
|
|
|
The bool isindex specifies whether we are in an index macro (which
|
|
|
|
has some specific active characters that need to be ERTed).'''
|
2008-07-07 20:31:06 +00:00
|
|
|
|
2008-10-23 18:08:26 +00:00
|
|
|
if not data:
|
2008-10-25 13:43:07 +00:00
|
|
|
return [""]
|
2008-07-07 19:13:01 +00:00
|
|
|
retval = []
|
|
|
|
|
|
|
|
# Convert LaTeX to Unicode
|
2008-07-08 20:25:53 +00:00
|
|
|
# Commands of this sort need to be checked to make sure they are
|
|
|
|
# followed by a non-alpha character, lest we replace too much.
|
|
|
|
hardone = re.compile(r'^\\\\[a-zA-Z]+$')
|
2008-07-13 19:24:09 +00:00
|
|
|
|
2008-10-28 16:00:42 +00:00
|
|
|
for rep in unicode_reps:
|
2008-07-08 20:25:53 +00:00
|
|
|
if hardone.match(rep[0]):
|
|
|
|
pos = 0
|
|
|
|
while True:
|
|
|
|
pos = data.find(rep[0], pos)
|
|
|
|
if pos == -1:
|
|
|
|
break
|
|
|
|
nextpos = pos + len(rep[0])
|
2008-07-10 16:24:19 +00:00
|
|
|
if nextpos < len(data) and data[nextpos].isalpha():
|
2008-07-08 20:25:53 +00:00
|
|
|
# not the end of that command
|
|
|
|
pos = nextpos
|
|
|
|
continue
|
|
|
|
data = data[:pos] + rep[1] + data[nextpos:]
|
|
|
|
pos = nextpos
|
|
|
|
else:
|
2008-07-07 20:31:06 +00:00
|
|
|
data = data.replace(rep[0], rep[1])
|
2008-07-08 20:25:53 +00:00
|
|
|
|
2008-12-02 07:45:34 +00:00
|
|
|
# Generic
|
|
|
|
# \" -> ":
|
2008-07-07 19:13:01 +00:00
|
|
|
data = wrap_into_ert(data, r'\"', '"')
|
2008-12-02 07:45:34 +00:00
|
|
|
# \\ -> \:
|
|
|
|
data = data.replace('\\\\', '\\')
|
2008-07-08 20:25:53 +00:00
|
|
|
|
2008-07-07 19:13:01 +00:00
|
|
|
# Math:
|
2008-07-08 20:25:53 +00:00
|
|
|
mathre = re.compile('^(.*?)(\$.*?\$)(.*)')
|
2008-07-07 19:13:01 +00:00
|
|
|
lines = data.split('\n')
|
|
|
|
for line in lines:
|
|
|
|
#document.warning("LINE: " + line)
|
|
|
|
#document.warning(str(i) + ":" + document.body[i])
|
|
|
|
#document.warning("LAST: " + document.body[-1])
|
|
|
|
g = line
|
2008-07-07 21:13:02 +00:00
|
|
|
m = mathre.match(g)
|
|
|
|
while m != None:
|
2008-07-07 19:13:01 +00:00
|
|
|
s = m.group(1)
|
|
|
|
f = m.group(2).replace('\\\\', '\\')
|
|
|
|
g = m.group(3)
|
|
|
|
if s:
|
2008-07-07 21:13:02 +00:00
|
|
|
# this is non-math!
|
2009-09-10 12:10:18 +00:00
|
|
|
s = latex2ert(s, isindex)
|
2008-07-07 21:13:02 +00:00
|
|
|
subst = s.split('\n')
|
|
|
|
retval += subst
|
2008-07-07 19:13:01 +00:00
|
|
|
retval.append("\\begin_inset Formula " + f)
|
|
|
|
retval.append("\\end_inset")
|
2008-07-07 21:13:02 +00:00
|
|
|
m = mathre.match(g)
|
2008-07-07 21:20:30 +00:00
|
|
|
# Handle whatever is left, which is just text
|
2009-09-10 12:10:18 +00:00
|
|
|
g = latex2ert(g, isindex)
|
2008-07-07 19:13:01 +00:00
|
|
|
subst = g.split('\n')
|
|
|
|
retval += subst
|
|
|
|
return retval
|
|
|
|
|
|
|
|
|
2009-04-27 09:51:38 +00:00
|
|
|
def lyxline2latex(document, line, inert):
|
|
|
|
'Convert some LyX stuff into corresponding LaTeX stuff line-wise, as best we can.'
|
|
|
|
if line.startswith("\\begin_inset Formula"):
|
|
|
|
line = line[20:]
|
|
|
|
elif line.startswith("\\begin_inset Quotes"):
|
|
|
|
# For now, we do a very basic reversion. Someone who understands
|
|
|
|
# quotes is welcome to fix it up.
|
|
|
|
qtype = line[20:].strip()
|
|
|
|
# lang = qtype[0]
|
|
|
|
side = qtype[1]
|
|
|
|
dbls = qtype[2]
|
|
|
|
if side == "l":
|
|
|
|
if dbls == "d":
|
|
|
|
line = "``"
|
|
|
|
else:
|
|
|
|
line = "`"
|
|
|
|
else:
|
|
|
|
if dbls == "d":
|
|
|
|
line = "''"
|
|
|
|
else:
|
|
|
|
line = "'"
|
|
|
|
elif line.isspace() or \
|
|
|
|
line.startswith("\\begin_layout") or \
|
|
|
|
line.startswith("\\end_layout") or \
|
|
|
|
line.startswith("\\begin_inset") or \
|
|
|
|
line.startswith("\\end_inset") or \
|
|
|
|
line.startswith("\\lang") or \
|
|
|
|
line.strip() == "status collapsed" or \
|
|
|
|
line.strip() == "status open":
|
|
|
|
#skip all that stuff
|
|
|
|
return ""
|
|
|
|
|
|
|
|
# this needs to be added to the preamble because of cases like
|
|
|
|
# \textmu, \textbackslash, etc.
|
|
|
|
add_to_preamble(document, ['% added by lyx2lyx for converted entries',
|
|
|
|
'\\@ifundefined{textmu}',
|
|
|
|
' {\\usepackage{textcomp}}{}'])
|
|
|
|
# a lossless reversion is not possible
|
|
|
|
# try at least to handle some common insets and settings
|
|
|
|
if inert:
|
|
|
|
line = line.replace(r'\backslash', '\\')
|
|
|
|
else:
|
|
|
|
line = line.replace('&', '\\&{}')
|
|
|
|
line = line.replace('#', '\\#{}')
|
|
|
|
line = line.replace('^', '\\^{}')
|
|
|
|
line = line.replace('%', '\\%{}')
|
|
|
|
line = line.replace('_', '\\_{}')
|
|
|
|
line = line.replace('$', '\\${}')
|
|
|
|
|
|
|
|
# Do the LyX text --> LaTeX conversion
|
|
|
|
for rep in unicode_reps:
|
|
|
|
line = line.replace(rep[1], rep[0].replace('\\\\', '\\') + "{}")
|
|
|
|
line = line.replace(r'\backslash', r'\textbackslash{}')
|
|
|
|
line = line.replace(r'\series bold', r'\bfseries{}').replace(r'\series default', r'\mdseries{}')
|
|
|
|
line = line.replace(r'\shape italic', r'\itshape{}').replace(r'\shape smallcaps', r'\scshape{}')
|
|
|
|
line = line.replace(r'\shape slanted', r'\slshape{}').replace(r'\shape default', r'\upshape{}')
|
|
|
|
line = line.replace(r'\emph on', r'\em{}').replace(r'\emph default', r'\em{}')
|
|
|
|
line = line.replace(r'\noun on', r'\scshape{}').replace(r'\noun default', r'\upshape{}')
|
|
|
|
line = line.replace(r'\bar under', r'\underbar{').replace(r'\bar default', r'}')
|
|
|
|
line = line.replace(r'\family sans', r'\sffamily{}').replace(r'\family default', r'\normalfont{}')
|
|
|
|
line = line.replace(r'\family typewriter', r'\ttfamily{}').replace(r'\family roman', r'\rmfamily{}')
|
|
|
|
line = line.replace(r'\InsetSpace ', r'').replace(r'\SpecialChar ', r'')
|
|
|
|
return line
|
|
|
|
|
|
|
|
|
2008-08-07 22:37:27 +00:00
|
|
|
def lyx2latex(document, lines):
|
2008-07-10 22:01:57 +00:00
|
|
|
'Convert some LyX stuff into corresponding LaTeX stuff, as best we can.'
|
|
|
|
# clean up multiline stuff
|
|
|
|
content = ""
|
|
|
|
ert_end = 0
|
2008-08-01 17:57:01 +00:00
|
|
|
|
2008-07-10 22:01:57 +00:00
|
|
|
for curline in range(len(lines)):
|
2009-04-27 09:51:38 +00:00
|
|
|
line = lines[curline]
|
|
|
|
if line.startswith("\\begin_inset ERT"):
|
|
|
|
# We don't want to replace things inside ERT, so figure out
|
|
|
|
# where the end of the inset is.
|
|
|
|
ert_end = find_end_of_inset(lines, curline + 1)
|
|
|
|
continue
|
|
|
|
inert = ert_end >= curline
|
|
|
|
content += lyxline2latex(document, lines[curline], inert)
|
|
|
|
|
2008-07-10 22:01:57 +00:00
|
|
|
return content
|
|
|
|
|
|
|
|
|
2008-05-04 07:51:50 +00:00
|
|
|
####################################################################
|
|
|
|
|
|
|
|
def convert_ltcaption(document):
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, "\\begin_inset Tabular", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
j = find_end_of_inset(document.body, i + 1)
|
|
|
|
if j == -1:
|
|
|
|
document.warning("Malformed LyX document: Could not find end of tabular.")
|
|
|
|
continue
|
|
|
|
|
|
|
|
nrows = int(document.body[i+1].split('"')[3])
|
|
|
|
ncols = int(document.body[i+1].split('"')[5])
|
|
|
|
|
|
|
|
m = i + 1
|
|
|
|
for k in range(nrows):
|
|
|
|
m = find_token(document.body, "<row", m)
|
|
|
|
r = m
|
|
|
|
caption = 'false'
|
|
|
|
for k in range(ncols):
|
|
|
|
m = find_token(document.body, "<cell", m)
|
|
|
|
if (k == 0):
|
|
|
|
mend = find_token(document.body, "</cell>", m + 1)
|
|
|
|
# first look for caption insets
|
|
|
|
mcap = find_token(document.body, "\\begin_inset Caption", m + 1, mend)
|
|
|
|
# then look for ERT captions
|
|
|
|
if mcap == -1:
|
|
|
|
mcap = find_token(document.body, "caption", m + 1, mend)
|
|
|
|
if mcap > -1:
|
|
|
|
mcap = find_token(document.body, "\\backslash", mcap - 1, mcap)
|
|
|
|
if mcap > -1:
|
|
|
|
caption = 'true'
|
|
|
|
if caption == 'true':
|
|
|
|
if (k == 0):
|
|
|
|
set_option(document, r, 'caption', 'true')
|
|
|
|
set_option(document, m, 'multicolumn', '1')
|
|
|
|
set_option(document, m, 'bottomline', 'false')
|
|
|
|
set_option(document, m, 'topline', 'false')
|
|
|
|
set_option(document, m, 'rightline', 'false')
|
|
|
|
set_option(document, m, 'leftline', 'false')
|
|
|
|
#j = find_end_of_inset(document.body, j + 1)
|
|
|
|
else:
|
|
|
|
set_option(document, m, 'multicolumn', '2')
|
|
|
|
m = m + 1
|
|
|
|
m = m + 1
|
|
|
|
|
|
|
|
i = j + 1
|
|
|
|
|
2008-06-02 19:42:53 +00:00
|
|
|
|
|
|
|
#FIXME Use of wrap_into_ert can confuse lyx2lyx
|
2008-05-04 07:51:50 +00:00
|
|
|
def revert_ltcaption(document):
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, "\\begin_inset Tabular", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
j = find_end_of_inset(document.body, i + 1)
|
|
|
|
if j == -1:
|
|
|
|
document.warning("Malformed LyX document: Could not find end of tabular.")
|
|
|
|
continue
|
|
|
|
|
|
|
|
m = i + 1
|
|
|
|
nrows = int(document.body[i+1].split('"')[3])
|
|
|
|
ncols = int(document.body[i+1].split('"')[5])
|
|
|
|
|
|
|
|
for k in range(nrows):
|
|
|
|
m = find_token(document.body, "<row", m)
|
|
|
|
caption = get_option(document, m, 'caption', 'false')
|
|
|
|
if caption == 'true':
|
|
|
|
remove_option(document, m, 'caption')
|
|
|
|
for k in range(ncols):
|
|
|
|
m = find_token(document.body, "<cell", m)
|
|
|
|
remove_option(document, m, 'multicolumn')
|
|
|
|
if k == 0:
|
|
|
|
m = find_token(document.body, "\\begin_inset Caption", m)
|
|
|
|
if m == -1:
|
|
|
|
return
|
|
|
|
m = find_end_of_inset(document.body, m + 1)
|
|
|
|
document.body[m] += wrap_into_ert("","","\\backslash\n\\backslash\n%")
|
|
|
|
m = m + 1
|
|
|
|
m = m + 1
|
|
|
|
i = j + 1
|
|
|
|
|
|
|
|
|
2008-03-18 18:08:32 +00:00
|
|
|
def convert_tablines(document):
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, "\\begin_inset Tabular", i)
|
|
|
|
if i == -1:
|
2008-03-19 09:40:24 +00:00
|
|
|
# LyX 1.3 inserted an extra space between \begin_inset
|
|
|
|
# and Tabular so let us try if this is the case and fix it.
|
2008-03-19 08:41:44 +00:00
|
|
|
i = find_token(document.body, "\\begin_inset Tabular", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
2008-03-19 09:40:24 +00:00
|
|
|
else:
|
|
|
|
document.body[i] = "\\begin_inset Tabular"
|
2008-03-18 18:08:32 +00:00
|
|
|
j = find_end_of_inset(document.body, i + 1)
|
|
|
|
if j == -1:
|
|
|
|
document.warning("Malformed LyX document: Could not find end of tabular.")
|
|
|
|
continue
|
|
|
|
|
|
|
|
m = i + 1
|
|
|
|
nrows = int(document.body[i+1].split('"')[3])
|
|
|
|
ncols = int(document.body[i+1].split('"')[5])
|
|
|
|
|
|
|
|
col_info = []
|
|
|
|
for k in range(ncols):
|
|
|
|
m = find_token(document.body, "<column", m)
|
|
|
|
left = get_option(document, m, 'leftline', 'false')
|
|
|
|
right = get_option(document, m, 'rightline', 'false')
|
|
|
|
col_info.append([left, right])
|
|
|
|
remove_option(document, m, 'leftline')
|
|
|
|
remove_option(document, m, 'rightline')
|
|
|
|
m = m + 1
|
|
|
|
|
|
|
|
row_info = []
|
|
|
|
for k in range(nrows):
|
|
|
|
m = find_token(document.body, "<row", m)
|
|
|
|
top = get_option(document, m, 'topline', 'false')
|
|
|
|
bottom = get_option(document, m, 'bottomline', 'false')
|
|
|
|
row_info.append([top, bottom])
|
|
|
|
remove_option(document, m, 'topline')
|
|
|
|
remove_option(document, m, 'bottomline')
|
|
|
|
m = m + 1
|
|
|
|
|
|
|
|
m = i + 1
|
|
|
|
mc_info = []
|
|
|
|
for k in range(nrows*ncols):
|
|
|
|
m = find_token(document.body, "<cell", m)
|
|
|
|
mc_info.append(get_option(document, m, 'multicolumn', '0'))
|
|
|
|
m = m + 1
|
|
|
|
m = i + 1
|
|
|
|
for l in range(nrows):
|
|
|
|
for k in range(ncols):
|
|
|
|
m = find_token(document.body, '<cell', m)
|
|
|
|
if mc_info[l*ncols + k] == '0':
|
|
|
|
r = set_option(document, m, 'topline', row_info[l][0])
|
|
|
|
r = set_option(document, m, 'bottomline', row_info[l][1])
|
|
|
|
r = set_option(document, m, 'leftline', col_info[k][0])
|
|
|
|
r = set_option(document, m, 'rightline', col_info[k][1])
|
|
|
|
elif mc_info[l*ncols + k] == '1':
|
|
|
|
s = k + 1
|
|
|
|
while s < ncols and mc_info[l*ncols + s] == '2':
|
|
|
|
s = s + 1
|
|
|
|
if s < ncols and mc_info[l*ncols + s] != '1':
|
|
|
|
r = set_option(document, m, 'rightline', col_info[k][1])
|
|
|
|
if k > 0 and mc_info[l*ncols + k - 1] == '0':
|
|
|
|
r = set_option(document, m, 'leftline', col_info[k][0])
|
|
|
|
m = m + 1
|
|
|
|
i = j + 1
|
|
|
|
|
|
|
|
|
|
|
|
def revert_tablines(document):
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, "\\begin_inset Tabular", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
j = find_end_of_inset(document.body, i + 1)
|
|
|
|
if j == -1:
|
|
|
|
document.warning("Malformed LyX document: Could not find end of tabular.")
|
|
|
|
continue
|
|
|
|
|
|
|
|
m = i + 1
|
|
|
|
nrows = int(document.body[i+1].split('"')[3])
|
|
|
|
ncols = int(document.body[i+1].split('"')[5])
|
|
|
|
|
|
|
|
lines = []
|
|
|
|
for k in range(nrows*ncols):
|
|
|
|
m = find_token(document.body, "<cell", m)
|
|
|
|
top = get_option(document, m, 'topline', 'false')
|
|
|
|
bottom = get_option(document, m, 'bottomline', 'false')
|
|
|
|
left = get_option(document, m, 'leftline', 'false')
|
|
|
|
right = get_option(document, m, 'rightline', 'false')
|
|
|
|
lines.append([top, bottom, left, right])
|
|
|
|
m = m + 1
|
|
|
|
|
2008-05-04 07:51:50 +00:00
|
|
|
# we will want to ignore longtable captions
|
|
|
|
m = i + 1
|
|
|
|
caption_info = []
|
|
|
|
for k in range(nrows):
|
|
|
|
m = find_token(document.body, "<row", m)
|
|
|
|
caption = get_option(document, m, 'caption', 'false')
|
|
|
|
caption_info.append([caption])
|
|
|
|
m = m + 1
|
|
|
|
|
2008-03-18 18:08:32 +00:00
|
|
|
m = i + 1
|
|
|
|
col_info = []
|
|
|
|
for k in range(ncols):
|
|
|
|
m = find_token(document.body, "<column", m)
|
|
|
|
left = 'true'
|
|
|
|
for l in range(nrows):
|
2008-04-17 21:12:24 +00:00
|
|
|
left = lines[l*ncols + k][2]
|
2008-05-04 07:51:50 +00:00
|
|
|
if left == 'false' and caption_info[l] == 'false':
|
2008-03-18 18:08:32 +00:00
|
|
|
break
|
|
|
|
set_option(document, m, 'leftline', left)
|
|
|
|
right = 'true'
|
|
|
|
for l in range(nrows):
|
2008-04-17 21:12:24 +00:00
|
|
|
right = lines[l*ncols + k][3]
|
2008-05-04 07:51:50 +00:00
|
|
|
if right == 'false' and caption_info[l] == 'false':
|
2008-03-18 18:08:32 +00:00
|
|
|
break
|
|
|
|
set_option(document, m, 'rightline', right)
|
|
|
|
m = m + 1
|
|
|
|
|
|
|
|
row_info = []
|
|
|
|
for k in range(nrows):
|
|
|
|
m = find_token(document.body, "<row", m)
|
|
|
|
top = 'true'
|
|
|
|
for l in range(ncols):
|
|
|
|
top = lines[k*ncols + l][0]
|
|
|
|
if top == 'false':
|
|
|
|
break
|
2008-05-04 07:51:50 +00:00
|
|
|
if caption_info[k] == 'false':
|
|
|
|
top = 'false'
|
2008-03-18 18:08:32 +00:00
|
|
|
set_option(document, m, 'topline', top)
|
|
|
|
bottom = 'true'
|
|
|
|
for l in range(ncols):
|
|
|
|
bottom = lines[k*ncols + l][1]
|
|
|
|
if bottom == 'false':
|
|
|
|
break
|
2008-05-04 07:51:50 +00:00
|
|
|
if caption_info[k] == 'false':
|
|
|
|
bottom = 'false'
|
2008-03-18 18:08:32 +00:00
|
|
|
set_option(document, m, 'bottomline', bottom)
|
|
|
|
m = m + 1
|
|
|
|
|
|
|
|
i = j + 1
|
|
|
|
|
|
|
|
|
2007-08-12 13:25:36 +00:00
|
|
|
def fix_wrong_tables(document):
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, "\\begin_inset Tabular", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
j = find_end_of_inset(document.body, i + 1)
|
|
|
|
if j == -1:
|
|
|
|
document.warning("Malformed LyX document: Could not find end of tabular.")
|
|
|
|
continue
|
|
|
|
|
|
|
|
m = i + 1
|
|
|
|
nrows = int(document.body[i+1].split('"')[3])
|
|
|
|
ncols = int(document.body[i+1].split('"')[5])
|
|
|
|
|
|
|
|
for l in range(nrows):
|
|
|
|
prev_multicolumn = 0
|
|
|
|
for k in range(ncols):
|
|
|
|
m = find_token(document.body, '<cell', m)
|
|
|
|
|
|
|
|
if document.body[m].find('multicolumn') != -1:
|
|
|
|
multicol_cont = int(document.body[m].split('"')[1])
|
|
|
|
|
|
|
|
if multicol_cont == 2 and (k == 0 or prev_multicolumn == 0):
|
|
|
|
document.body[m] = document.body[m][:5] + document.body[m][21:]
|
|
|
|
prev_multicolumn = 0
|
|
|
|
else:
|
|
|
|
prev_multicolumn = multicol_cont
|
|
|
|
else:
|
|
|
|
prev_multicolumn = 0
|
|
|
|
|
|
|
|
i = j + 1
|
|
|
|
|
|
|
|
|
2007-08-12 14:56:49 +00:00
|
|
|
def close_begin_deeper(document):
|
|
|
|
i = 0
|
|
|
|
depth = 0
|
|
|
|
while True:
|
|
|
|
i = find_tokens(document.body, ["\\begin_deeper", "\\end_deeper"], i)
|
|
|
|
|
|
|
|
if i == -1:
|
|
|
|
break
|
|
|
|
|
|
|
|
if document.body[i][:13] == "\\begin_deeper":
|
|
|
|
depth += 1
|
|
|
|
else:
|
|
|
|
depth -= 1
|
|
|
|
|
|
|
|
i += 1
|
|
|
|
|
|
|
|
document.body[-2:-2] = ['\\end_deeper' for i in range(depth)]
|
2007-08-17 15:48:41 +00:00
|
|
|
|
|
|
|
|
|
|
|
def long_charstyle_names(document):
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, "\\begin_inset CharStyle", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
document.body[i] = document.body[i].replace("CharStyle ", "CharStyle CharStyle:")
|
|
|
|
i += 1
|
|
|
|
|
|
|
|
def revert_long_charstyle_names(document):
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, "\\begin_inset CharStyle", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
2008-08-05 14:33:42 +00:00
|
|
|
document.body[i] = document.body[i].replace("CharStyle CharStyle:", "CharStyle ")
|
2007-08-17 15:48:41 +00:00
|
|
|
i += 1
|
2007-08-12 14:56:49 +00:00
|
|
|
|
|
|
|
|
2007-08-18 23:26:07 +00:00
|
|
|
def axe_show_label(document):
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, "\\begin_inset CharStyle", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
if document.body[i + 1].find("show_label") != -1:
|
|
|
|
if document.body[i + 1].find("true") != -1:
|
|
|
|
document.body[i + 1] = "status open"
|
|
|
|
del document.body[ i + 2]
|
|
|
|
else:
|
|
|
|
if document.body[i + 1].find("false") != -1:
|
|
|
|
document.body[i + 1] = "status collapsed"
|
|
|
|
del document.body[ i + 2]
|
|
|
|
else:
|
|
|
|
document.warning("Malformed LyX document: show_label neither false nor true.")
|
|
|
|
else:
|
|
|
|
document.warning("Malformed LyX document: show_label missing in CharStyle.")
|
2008-03-19 09:40:24 +00:00
|
|
|
|
2007-08-19 14:03:51 +00:00
|
|
|
i += 1
|
2007-08-18 23:26:07 +00:00
|
|
|
|
|
|
|
|
|
|
|
def revert_show_label(document):
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, "\\begin_inset CharStyle", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
if document.body[i + 1].find("status open") != -1:
|
|
|
|
document.body.insert(i + 1, "show_label true")
|
|
|
|
else:
|
|
|
|
if document.body[i + 1].find("status collapsed") != -1:
|
|
|
|
document.body.insert(i + 1, "show_label false")
|
|
|
|
else:
|
|
|
|
document.warning("Malformed LyX document: no legal status line in CharStyle.")
|
|
|
|
i += 1
|
|
|
|
|
This is one of a series of patches that will merge the layout modules development in personal/branches/rgheck back into the tree.
Design goal: Allow the use of layout "modules", which are to LaTeX packages as layout files are to LaTeX document classes. Thus, one could have a module that defined certain character styles, environments, commands, or what have you, and include it in various documents, each of which uses a different document class, without having to modify the layout files themselves. For example, a theorems.module could be used with article.layout to provide support for theorem-type environments, without having to modify article.layout itself, and the same module could be used with book.layout, etc.
This patch adds the backend. The ModuleList class holds a list of the available modules, which are retrieved from lyxmodules.lst, itself generated by configure.py. There are two LFUNs available: modules-clear and module-add, which do the obvious thing; you can test by typing these into the minibuffer, along with the name of one of the available modules: URL (a CharStyle), Endnote (a Custom Inset), and---with the spaces---End To Foot (View>LaTeX and look at the user preamble), which are themselves in lib/layouts. There are some others, too, that allow theorems to be added to classes like article and book.
The GUI will come next.
Issues: (i) The configure.py script could be improved. It'd be nice, for example, if it tested for the presence of the LaTeX packages a particular module needs. But this would mean re-working the LaTeX script, and I don't know how to do that. Note that at present, the packages are ignored. This will change shortly. (ii) I've used std::string in LyXModule, following what seemed to be a precedent in TextClass. If some of these should be docstrings, please let me know, and I'll change them. (iii) There is at present no distinction between LaTeX and DocBook modules. Should there be? That is: Should there be modules that are available when the document class is a LaTeX class and others that are available only when it is DocBook? Or should there just be one set of modules? Each module can of course indicate for what it is suitable in its description.
git-svn-id: svn://svn.lyx.org/lyx/lyx-devel/trunk@19893 a592a061-630c-0410-9148-cb99ea01b6c8
2007-08-29 17:59:49 +00:00
|
|
|
def revert_begin_modules(document):
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.header, "\\begin_modules", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
j = find_end_of(document.header, i, "\\begin_modules", "\\end_modules")
|
|
|
|
if j == -1:
|
|
|
|
# this should not happen
|
|
|
|
break
|
|
|
|
document.header[i : j + 1] = []
|
2007-08-18 23:26:07 +00:00
|
|
|
|
2007-09-08 13:01:29 +00:00
|
|
|
def convert_flex(document):
|
|
|
|
"Convert CharStyle to Flex"
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, "\\begin_inset CharStyle", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
document.body[i] = document.body[i].replace('\\begin_inset CharStyle', '\\begin_inset Flex')
|
|
|
|
|
|
|
|
def revert_flex(document):
|
|
|
|
"Convert Flex to CharStyle"
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, "\\begin_inset Flex", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
document.body[i] = document.body[i].replace('\\begin_inset Flex', '\\begin_inset CharStyle')
|
|
|
|
|
|
|
|
|
2007-09-24 14:37:10 +00:00
|
|
|
# Discard PDF options for hyperref
|
2007-09-20 22:32:20 +00:00
|
|
|
def revert_pdf_options(document):
|
2007-09-24 14:37:10 +00:00
|
|
|
"Revert PDF options for hyperref."
|
2008-04-19 21:29:44 +00:00
|
|
|
# store the PDF options and delete the entries from the Lyx file
|
2007-09-23 12:39:31 +00:00
|
|
|
i = 0
|
2008-04-12 13:48:17 +00:00
|
|
|
hyperref = False
|
|
|
|
title = ""
|
|
|
|
author = ""
|
|
|
|
subject = ""
|
|
|
|
keywords = ""
|
|
|
|
bookmarks = ""
|
|
|
|
bookmarksnumbered = ""
|
|
|
|
bookmarksopen = ""
|
|
|
|
bookmarksopenlevel = ""
|
|
|
|
breaklinks = ""
|
|
|
|
pdfborder = ""
|
|
|
|
colorlinks = ""
|
|
|
|
backref = ""
|
|
|
|
pagebackref = ""
|
|
|
|
pagemode = ""
|
|
|
|
otheroptions = ""
|
2008-04-15 14:02:42 +00:00
|
|
|
i = find_token(document.header, "\\use_hyperref", i)
|
2007-09-23 12:39:31 +00:00
|
|
|
if i != -1:
|
2008-04-15 14:02:42 +00:00
|
|
|
hyperref = get_value(document.header, "\\use_hyperref", i) == 'true'
|
2007-09-23 12:39:31 +00:00
|
|
|
del document.header[i]
|
|
|
|
i = find_token(document.header, "\\pdf_store_options", i)
|
|
|
|
if i != -1:
|
|
|
|
del document.header[i]
|
|
|
|
i = find_token(document.header, "\\pdf_title", 0)
|
|
|
|
if i != -1:
|
2008-04-12 13:48:17 +00:00
|
|
|
title = get_value_string(document.header, '\\pdf_title', 0, 0, True)
|
|
|
|
title = ' pdftitle={' + title + '}'
|
2007-09-23 12:39:31 +00:00
|
|
|
del document.header[i]
|
|
|
|
i = find_token(document.header, "\\pdf_author", 0)
|
|
|
|
if i != -1:
|
2008-04-12 13:48:17 +00:00
|
|
|
author = get_value_string(document.header, '\\pdf_author', 0, 0, True)
|
|
|
|
if title == "":
|
|
|
|
author = ' pdfauthor={' + author + '}'
|
|
|
|
else:
|
|
|
|
author = ',\n pdfauthor={' + author + '}'
|
2007-09-23 12:39:31 +00:00
|
|
|
del document.header[i]
|
|
|
|
i = find_token(document.header, "\\pdf_subject", 0)
|
|
|
|
if i != -1:
|
2008-04-12 13:48:17 +00:00
|
|
|
subject = get_value_string(document.header, '\\pdf_subject', 0, 0, True)
|
|
|
|
if title == "" and author == "":
|
|
|
|
subject = ' pdfsubject={' + subject + '}'
|
|
|
|
else:
|
|
|
|
subject = ',\n pdfsubject={' + subject + '}'
|
2007-09-23 12:39:31 +00:00
|
|
|
del document.header[i]
|
|
|
|
i = find_token(document.header, "\\pdf_keywords", 0)
|
|
|
|
if i != -1:
|
2008-04-12 13:48:17 +00:00
|
|
|
keywords = get_value_string(document.header, '\\pdf_keywords', 0, 0, True)
|
|
|
|
if title == "" and author == "" and subject == "":
|
|
|
|
keywords = ' pdfkeywords={' + keywords + '}'
|
|
|
|
else:
|
|
|
|
keywords = ',\n pdfkeywords={' + keywords + '}'
|
2007-09-23 12:39:31 +00:00
|
|
|
del document.header[i]
|
|
|
|
i = find_token(document.header, "\\pdf_bookmarks", 0)
|
|
|
|
if i != -1:
|
2008-04-12 13:48:17 +00:00
|
|
|
bookmarks = get_value_string(document.header, '\\pdf_bookmarks', 0)
|
|
|
|
bookmarks = ',\n bookmarks=' + bookmarks
|
2007-09-23 12:39:31 +00:00
|
|
|
del document.header[i]
|
|
|
|
i = find_token(document.header, "\\pdf_bookmarksnumbered", i)
|
|
|
|
if i != -1:
|
2008-04-12 13:48:17 +00:00
|
|
|
bookmarksnumbered = get_value_string(document.header, '\\pdf_bookmarksnumbered', 0)
|
|
|
|
bookmarksnumbered = ',\n bookmarksnumbered=' + bookmarksnumbered
|
2007-09-23 12:39:31 +00:00
|
|
|
del document.header[i]
|
|
|
|
i = find_token(document.header, "\\pdf_bookmarksopen", i)
|
|
|
|
if i != -1:
|
2008-04-12 13:48:17 +00:00
|
|
|
bookmarksopen = get_value_string(document.header, '\\pdf_bookmarksopen', 0)
|
|
|
|
bookmarksopen = ',\n bookmarksopen=' + bookmarksopen
|
2007-09-23 12:39:31 +00:00
|
|
|
del document.header[i]
|
|
|
|
i = find_token(document.header, "\\pdf_bookmarksopenlevel", i)
|
|
|
|
if i != -1:
|
2008-04-12 13:48:17 +00:00
|
|
|
bookmarksopenlevel = get_value_string(document.header, '\\pdf_bookmarksopenlevel', 0, 0, True)
|
|
|
|
bookmarksopenlevel = ',\n bookmarksopenlevel=' + bookmarksopenlevel
|
2007-09-23 12:39:31 +00:00
|
|
|
del document.header[i]
|
|
|
|
i = find_token(document.header, "\\pdf_breaklinks", i)
|
|
|
|
if i != -1:
|
2008-04-12 13:48:17 +00:00
|
|
|
breaklinks = get_value_string(document.header, '\\pdf_breaklinks', 0)
|
|
|
|
breaklinks = ',\n breaklinks=' + breaklinks
|
2007-09-23 12:39:31 +00:00
|
|
|
del document.header[i]
|
|
|
|
i = find_token(document.header, "\\pdf_pdfborder", i)
|
|
|
|
if i != -1:
|
2008-04-12 13:48:17 +00:00
|
|
|
pdfborder = get_value_string(document.header, '\\pdf_pdfborder', 0)
|
|
|
|
if pdfborder == 'true':
|
|
|
|
pdfborder = ',\n pdfborder={0 0 0}'
|
|
|
|
else:
|
|
|
|
pdfborder = ',\n pdfborder={0 0 1}'
|
2007-09-23 12:39:31 +00:00
|
|
|
del document.header[i]
|
|
|
|
i = find_token(document.header, "\\pdf_colorlinks", i)
|
|
|
|
if i != -1:
|
2008-04-12 13:48:17 +00:00
|
|
|
colorlinks = get_value_string(document.header, '\\pdf_colorlinks', 0)
|
|
|
|
colorlinks = ',\n colorlinks=' + colorlinks
|
2007-09-23 12:39:31 +00:00
|
|
|
del document.header[i]
|
|
|
|
i = find_token(document.header, "\\pdf_backref", i)
|
|
|
|
if i != -1:
|
2008-04-12 13:48:17 +00:00
|
|
|
backref = get_value_string(document.header, '\\pdf_backref', 0)
|
|
|
|
backref = ',\n backref=' + backref
|
2007-09-23 12:39:31 +00:00
|
|
|
del document.header[i]
|
|
|
|
i = find_token(document.header, "\\pdf_pagebackref", i)
|
|
|
|
if i != -1:
|
2008-04-12 13:48:17 +00:00
|
|
|
pagebackref = get_value_string(document.header, '\\pdf_pagebackref', 0)
|
|
|
|
pagebackref = ',\n pagebackref=' + pagebackref
|
2007-09-23 12:39:31 +00:00
|
|
|
del document.header[i]
|
|
|
|
i = find_token(document.header, "\\pdf_pagemode", 0)
|
|
|
|
if i != -1:
|
2008-04-12 13:48:17 +00:00
|
|
|
pagemode = get_value_string(document.header, '\\pdf_pagemode', 0)
|
|
|
|
pagemode = ',\n pdfpagemode=' + pagemode
|
2007-09-23 12:39:31 +00:00
|
|
|
del document.header[i]
|
|
|
|
i = find_token(document.header, "\\pdf_quoted_options", 0)
|
|
|
|
if i != -1:
|
2008-04-12 13:48:17 +00:00
|
|
|
otheroptions = get_value_string(document.header, '\\pdf_quoted_options', 0, 0, True)
|
|
|
|
if title == "" and author == "" and subject == "" and keywords == "":
|
|
|
|
otheroptions = ' ' + otheroptions
|
|
|
|
else:
|
|
|
|
otheroptions = ',\n ' + otheroptions
|
2007-09-23 12:39:31 +00:00
|
|
|
del document.header[i]
|
2007-09-20 22:32:20 +00:00
|
|
|
|
2008-04-12 13:48:17 +00:00
|
|
|
# write to the preamble when hyperref was used
|
|
|
|
if hyperref == True:
|
2008-04-12 13:54:30 +00:00
|
|
|
# preamble write preparations
|
|
|
|
# bookmark numbers are only output when they are turned on
|
2008-04-12 13:48:17 +00:00
|
|
|
if bookmarksopen == ',\n bookmarksopen=true':
|
|
|
|
bookmarksopen = bookmarksopen + bookmarksopenlevel
|
|
|
|
if bookmarks == ',\n bookmarks=true':
|
|
|
|
bookmarks = bookmarks + bookmarksnumbered + bookmarksopen
|
|
|
|
else:
|
|
|
|
bookmarks = bookmarks
|
2008-04-12 13:54:30 +00:00
|
|
|
# hypersetup is only output when there are things to be set up
|
2008-04-12 13:48:17 +00:00
|
|
|
setupstart = '\\hypersetup{%\n'
|
|
|
|
setupend = ' }\n'
|
|
|
|
if otheroptions == "" and title == "" and author == ""\
|
|
|
|
and subject == "" and keywords == "":
|
|
|
|
setupstart = ""
|
|
|
|
setupend = ""
|
2008-04-12 13:54:30 +00:00
|
|
|
# write the preamble
|
2008-08-05 22:37:28 +00:00
|
|
|
# babel must be loaded before hyperref and hyperref the first part
|
|
|
|
# of the preamble, like in LyX 1.6
|
|
|
|
insert_to_preamble(0, document,
|
|
|
|
'% Commands inserted by lyx2lyx for PDF properties\n'
|
|
|
|
+ '\\usepackage{babel}\n'
|
|
|
|
+ '\\usepackage[unicode=true'
|
2008-04-12 13:48:17 +00:00
|
|
|
+ bookmarks
|
|
|
|
+ breaklinks
|
|
|
|
+ pdfborder
|
|
|
|
+ backref
|
|
|
|
+ pagebackref
|
|
|
|
+ colorlinks
|
|
|
|
+ pagemode
|
|
|
|
+ ']\n'
|
2008-08-05 22:37:28 +00:00
|
|
|
+ ' {hyperref}\n'
|
2008-04-12 13:48:17 +00:00
|
|
|
+ setupstart
|
|
|
|
+ title
|
|
|
|
+ author
|
|
|
|
+ subject
|
|
|
|
+ keywords
|
|
|
|
+ otheroptions
|
2008-08-05 22:37:28 +00:00
|
|
|
+ setupend)
|
2008-04-12 13:48:17 +00:00
|
|
|
|
2007-08-18 23:26:07 +00:00
|
|
|
|
2007-09-15 18:31:32 +00:00
|
|
|
def remove_inzip_options(document):
|
|
|
|
"Remove inzipName and embed options from the Graphics inset"
|
|
|
|
i = 0
|
|
|
|
while 1:
|
|
|
|
i = find_token(document.body, "\\begin_inset Graphics", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
j = find_end_of_inset(document.body, i + 1)
|
|
|
|
if j == -1:
|
|
|
|
# should not happen
|
|
|
|
document.warning("Malformed LyX document: Could not find end of graphics inset.")
|
|
|
|
# If there's a inzip param, just remove that
|
|
|
|
k = find_token(document.body, "\tinzipName", i + 1, j)
|
|
|
|
if k != -1:
|
|
|
|
del document.body[k]
|
|
|
|
# embed option must follow the inzipName option
|
|
|
|
del document.body[k+1]
|
|
|
|
i = i + 1
|
|
|
|
|
|
|
|
|
2007-09-27 18:24:18 +00:00
|
|
|
def convert_inset_command(document):
|
2007-10-07 14:26:16 +00:00
|
|
|
"""
|
|
|
|
Convert:
|
2008-03-19 09:40:24 +00:00
|
|
|
\begin_inset LatexCommand cmd
|
|
|
|
to
|
2007-10-07 14:26:16 +00:00
|
|
|
\begin_inset CommandInset InsetType
|
|
|
|
LatexCommand cmd
|
|
|
|
"""
|
|
|
|
i = 0
|
|
|
|
while 1:
|
|
|
|
i = find_token(document.body, "\\begin_inset LatexCommand", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
line = document.body[i]
|
|
|
|
r = re.compile(r'\\begin_inset LatexCommand (.*)$')
|
|
|
|
m = r.match(line)
|
|
|
|
cmdName = m.group(1)
|
|
|
|
insetName = ""
|
|
|
|
#this is adapted from factory.cpp
|
|
|
|
if cmdName[0:4].lower() == "cite":
|
|
|
|
insetName = "citation"
|
|
|
|
elif cmdName == "url" or cmdName == "htmlurl":
|
|
|
|
insetName = "url"
|
|
|
|
elif cmdName[-3:] == "ref":
|
|
|
|
insetName = "ref"
|
|
|
|
elif cmdName == "tableofcontents":
|
|
|
|
insetName = "toc"
|
|
|
|
elif cmdName == "printnomenclature":
|
|
|
|
insetName = "nomencl_print"
|
|
|
|
elif cmdName == "printindex":
|
|
|
|
insetName = "index_print"
|
|
|
|
else:
|
|
|
|
insetName = cmdName
|
|
|
|
insertion = ["\\begin_inset CommandInset " + insetName, "LatexCommand " + cmdName]
|
|
|
|
document.body[i : i+1] = insertion
|
2007-09-27 18:24:18 +00:00
|
|
|
|
|
|
|
|
|
|
|
def revert_inset_command(document):
|
2007-10-07 14:26:16 +00:00
|
|
|
"""
|
|
|
|
Convert:
|
|
|
|
\begin_inset CommandInset InsetType
|
|
|
|
LatexCommand cmd
|
2008-03-19 09:40:24 +00:00
|
|
|
to
|
|
|
|
\begin_inset LatexCommand cmd
|
2007-10-07 14:26:16 +00:00
|
|
|
Some insets may end up being converted to insets earlier versions of LyX
|
|
|
|
will not be able to recognize. Not sure what to do about that.
|
|
|
|
"""
|
|
|
|
i = 0
|
|
|
|
while 1:
|
|
|
|
i = find_token(document.body, "\\begin_inset CommandInset", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
nextline = document.body[i+1]
|
|
|
|
r = re.compile(r'LatexCommand\s+(.*)$')
|
|
|
|
m = r.match(nextline)
|
|
|
|
if not m:
|
|
|
|
document.warning("Malformed LyX document: Missing LatexCommand in " + document.body[i] + ".")
|
2008-11-11 17:44:55 +00:00
|
|
|
i += 1
|
2007-10-07 14:26:16 +00:00
|
|
|
continue
|
|
|
|
cmdName = m.group(1)
|
|
|
|
insertion = ["\\begin_inset LatexCommand " + cmdName]
|
|
|
|
document.body[i : i+2] = insertion
|
2007-09-27 18:24:18 +00:00
|
|
|
|
|
|
|
|
2007-09-24 14:37:10 +00:00
|
|
|
def convert_wrapfig_options(document):
|
|
|
|
"Convert optional options for wrap floats (wrapfig)."
|
|
|
|
# adds the tokens "lines", "placement", and "overhang"
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, "\\begin_inset Wrap figure", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
document.body.insert(i + 1, "lines 0")
|
|
|
|
j = find_token(document.body, "placement", i)
|
|
|
|
# placement can be already set or not; if not, set it
|
|
|
|
if j == i+2:
|
|
|
|
document.body.insert(i + 3, "overhang 0col%")
|
|
|
|
else:
|
|
|
|
document.body.insert(i + 2, "placement o")
|
|
|
|
document.body.insert(i + 3, "overhang 0col%")
|
|
|
|
i = i + 1
|
|
|
|
|
|
|
|
|
2007-09-24 13:43:58 +00:00
|
|
|
def revert_wrapfig_options(document):
|
2007-09-24 14:37:10 +00:00
|
|
|
"Revert optional options for wrap floats (wrapfig)."
|
2007-09-24 13:43:58 +00:00
|
|
|
i = 0
|
|
|
|
while True:
|
2008-07-07 18:09:29 +00:00
|
|
|
i = find_token(document.body, "\\begin_inset Wrap figure", i)
|
2007-09-24 13:43:58 +00:00
|
|
|
if i == -1:
|
|
|
|
return
|
2008-07-07 18:09:29 +00:00
|
|
|
j = find_end_of_inset(document.body, i)
|
2007-09-24 13:43:58 +00:00
|
|
|
if j == -1:
|
2008-07-07 18:09:29 +00:00
|
|
|
document.warning("Can't find end of Wrap inset at line " + str(i))
|
|
|
|
i += 1
|
|
|
|
continue
|
|
|
|
k = find_default_layout(document, i, j)
|
|
|
|
if k == -1:
|
|
|
|
document.warning("Can't find default layout for Wrap figure!")
|
|
|
|
i = j
|
|
|
|
continue
|
|
|
|
# Options should be between i and k now
|
|
|
|
l = find_token(document.body, "lines", i, k)
|
|
|
|
if l == -1:
|
|
|
|
document.warning("Can't find lines option for Wrap figure!")
|
|
|
|
i = k
|
|
|
|
continue
|
|
|
|
m = find_token(document.body, "overhang", i + 1, k)
|
|
|
|
if m == -1:
|
|
|
|
document.warning("Malformed LyX document: Couldn't find overhang parameter of wrap float!")
|
|
|
|
i = k
|
|
|
|
continue
|
|
|
|
# Do these in reverse order
|
|
|
|
del document.body[m]
|
|
|
|
del document.body[l]
|
|
|
|
i = k
|
2007-09-24 13:43:58 +00:00
|
|
|
|
|
|
|
|
2007-10-03 13:38:19 +00:00
|
|
|
def convert_latexcommand_index(document):
|
|
|
|
"Convert from LatexCommand form to collapsable form."
|
2008-03-19 09:40:24 +00:00
|
|
|
i = 0
|
2008-07-07 18:50:50 +00:00
|
|
|
r1 = re.compile('name "(.*)"')
|
2007-10-03 13:38:19 +00:00
|
|
|
while True:
|
|
|
|
i = find_token(document.body, "\\begin_inset CommandInset index", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
if document.body[i + 1] != "LatexCommand index": # Might also be index_print
|
2009-01-25 10:59:52 +00:00
|
|
|
i += 1
|
|
|
|
continue
|
2009-03-13 12:17:21 +00:00
|
|
|
j = find_end_of_inset(document.body, i + 1)
|
2008-08-05 15:07:37 +00:00
|
|
|
if j == -1:
|
2009-03-13 12:17:21 +00:00
|
|
|
document.warning("Unable to find end of index inset at line " + str(i) + "!")
|
2008-08-05 15:07:37 +00:00
|
|
|
i += 2
|
|
|
|
continue
|
2008-07-07 18:50:50 +00:00
|
|
|
m = r1.match(document.body[i + 2])
|
|
|
|
if m == None:
|
|
|
|
document.warning("Unable to match: " + document.body[i+2])
|
2008-10-12 08:58:42 +00:00
|
|
|
# this can happen with empty index insets!
|
|
|
|
linelist = [""]
|
|
|
|
else:
|
|
|
|
fullcontent = m.group(1)
|
2009-09-10 12:10:18 +00:00
|
|
|
linelist = latex2lyx(fullcontent, True)
|
2008-08-05 15:07:37 +00:00
|
|
|
#document.warning(fullcontent)
|
2008-07-07 19:13:01 +00:00
|
|
|
|
2008-08-05 15:07:37 +00:00
|
|
|
linelist = ["\\begin_inset Index", "status collapsed", "\\begin_layout Standard", ""] + \
|
|
|
|
linelist + ["\\end_layout"]
|
|
|
|
document.body[i : j] = linelist
|
|
|
|
i += len(linelist) - (j - i)
|
2007-10-03 13:38:19 +00:00
|
|
|
|
|
|
|
|
|
|
|
def revert_latexcommand_index(document):
|
2007-12-06 11:04:56 +00:00
|
|
|
"Revert from collapsable form to LatexCommand form."
|
2007-10-03 13:38:19 +00:00
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, "\\begin_inset Index", i)
|
|
|
|
if i == -1:
|
2007-11-21 18:18:24 +00:00
|
|
|
return
|
|
|
|
j = find_end_of_inset(document.body, i + 1)
|
|
|
|
if j == -1:
|
|
|
|
return
|
2008-07-10 16:24:19 +00:00
|
|
|
|
2008-08-07 22:37:27 +00:00
|
|
|
content = lyx2latex(document, document.body[i:j])
|
2008-07-10 16:24:19 +00:00
|
|
|
# escape quotes
|
|
|
|
content = content.replace('"', r'\"')
|
|
|
|
document.body[i:j] = ["\\begin_inset CommandInset index", "LatexCommand index",
|
|
|
|
"name " + '"' + content + '"', ""]
|
|
|
|
i += 5
|
2007-10-03 13:38:19 +00:00
|
|
|
|
|
|
|
|
2007-10-03 21:07:01 +00:00
|
|
|
def revert_wraptable(document):
|
|
|
|
"Revert wrap table to wrap figure."
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, "\\begin_inset Wrap table", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
document.body[i] = document.body[i].replace('\\begin_inset Wrap table', '\\begin_inset Wrap figure')
|
|
|
|
i = i + 1
|
|
|
|
|
|
|
|
|
2007-10-04 23:20:39 +00:00
|
|
|
def revert_vietnamese(document):
|
|
|
|
"Set language Vietnamese to English"
|
|
|
|
# Set document language from Vietnamese to English
|
|
|
|
i = 0
|
|
|
|
if document.language == "vietnamese":
|
|
|
|
document.language = "english"
|
|
|
|
i = find_token(document.header, "\\language", 0)
|
|
|
|
if i != -1:
|
|
|
|
document.header[i] = "\\language english"
|
|
|
|
j = 0
|
|
|
|
while True:
|
|
|
|
j = find_token(document.body, "\\lang vietnamese", j)
|
|
|
|
if j == -1:
|
|
|
|
return
|
|
|
|
document.body[j] = document.body[j].replace("\\lang vietnamese", "\\lang english")
|
|
|
|
j = j + 1
|
|
|
|
|
|
|
|
|
2008-07-09 07:27:31 +00:00
|
|
|
def convert_japanese_cjk(document):
|
|
|
|
"Set language japanese to japanese-cjk"
|
|
|
|
# Set document language from japanese-plain to japanese
|
|
|
|
i = 0
|
|
|
|
if document.language == "japanese":
|
|
|
|
document.language = "japanese-cjk"
|
|
|
|
i = find_token(document.header, "\\language", 0)
|
|
|
|
if i != -1:
|
|
|
|
document.header[i] = "\\language japanese-cjk"
|
|
|
|
j = 0
|
|
|
|
while True:
|
|
|
|
j = find_token(document.body, "\\lang japanese", j)
|
|
|
|
if j == -1:
|
|
|
|
return
|
|
|
|
document.body[j] = document.body[j].replace("\\lang japanese", "\\lang japanese-cjk")
|
|
|
|
j = j + 1
|
|
|
|
|
|
|
|
|
2007-10-08 23:40:23 +00:00
|
|
|
def revert_japanese(document):
|
|
|
|
"Set language japanese-plain to japanese"
|
|
|
|
# Set document language from japanese-plain to japanese
|
|
|
|
i = 0
|
|
|
|
if document.language == "japanese-plain":
|
|
|
|
document.language = "japanese"
|
|
|
|
i = find_token(document.header, "\\language", 0)
|
|
|
|
if i != -1:
|
|
|
|
document.header[i] = "\\language japanese"
|
|
|
|
j = 0
|
|
|
|
while True:
|
|
|
|
j = find_token(document.body, "\\lang japanese-plain", j)
|
|
|
|
if j == -1:
|
|
|
|
return
|
|
|
|
document.body[j] = document.body[j].replace("\\lang japanese-plain", "\\lang japanese")
|
|
|
|
j = j + 1
|
|
|
|
|
|
|
|
|
2008-07-09 07:27:31 +00:00
|
|
|
def revert_japanese_cjk(document):
|
|
|
|
"Set language japanese-cjk to japanese"
|
|
|
|
# Set document language from japanese-plain to japanese
|
|
|
|
i = 0
|
|
|
|
if document.language == "japanese-cjk":
|
|
|
|
document.language = "japanese"
|
|
|
|
i = find_token(document.header, "\\language", 0)
|
|
|
|
if i != -1:
|
|
|
|
document.header[i] = "\\language japanese"
|
|
|
|
j = 0
|
|
|
|
while True:
|
|
|
|
j = find_token(document.body, "\\lang japanese-cjk", j)
|
|
|
|
if j == -1:
|
|
|
|
return
|
|
|
|
document.body[j] = document.body[j].replace("\\lang japanese-cjk", "\\lang japanese")
|
|
|
|
j = j + 1
|
|
|
|
|
|
|
|
|
2007-10-09 00:12:48 +00:00
|
|
|
def revert_japanese_encoding(document):
|
|
|
|
"Set input encoding form EUC-JP-plain to EUC-JP etc."
|
|
|
|
# Set input encoding form EUC-JP-plain to EUC-JP etc.
|
|
|
|
i = 0
|
|
|
|
i = find_token(document.header, "\\inputencoding EUC-JP-plain", 0)
|
|
|
|
if i != -1:
|
|
|
|
document.header[i] = "\\inputencoding EUC-JP"
|
|
|
|
j = 0
|
|
|
|
j = find_token(document.header, "\\inputencoding JIS-plain", 0)
|
|
|
|
if j != -1:
|
|
|
|
document.header[j] = "\\inputencoding JIS"
|
|
|
|
k = 0
|
|
|
|
k = find_token(document.header, "\\inputencoding SJIS-plain", 0)
|
2008-03-19 09:40:24 +00:00
|
|
|
if k != -1: # convert to UTF8 since there is currently no SJIS encoding
|
2007-10-09 00:12:48 +00:00
|
|
|
document.header[k] = "\\inputencoding UTF8"
|
|
|
|
|
|
|
|
|
2007-10-11 14:56:34 +00:00
|
|
|
def revert_inset_info(document):
|
|
|
|
'Replace info inset with its content'
|
|
|
|
i = 0
|
|
|
|
while 1:
|
|
|
|
i = find_token(document.body, '\\begin_inset Info', i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
j = find_end_of_inset(document.body, i + 1)
|
|
|
|
if j == -1:
|
|
|
|
# should not happen
|
|
|
|
document.warning("Malformed LyX document: Could not find end of Info inset.")
|
|
|
|
type = 'unknown'
|
|
|
|
arg = ''
|
|
|
|
for k in range(i, j+1):
|
|
|
|
if document.body[k].startswith("arg"):
|
2008-11-11 17:24:07 +00:00
|
|
|
arg = document.body[k][3:].strip()
|
|
|
|
# remove embracing quotation marks
|
|
|
|
if arg[0] == '"':
|
|
|
|
arg = arg[1:]
|
|
|
|
if arg[len(arg) - 1] == '"':
|
|
|
|
arg = arg[:len(arg) - 1]
|
|
|
|
# \" to straight quote
|
2009-04-27 09:51:38 +00:00
|
|
|
arg = arg.replace(r'\"', '"')
|
|
|
|
# \ to \backslash
|
|
|
|
arg = arg.replace(r'\\', "\\backslash\n")
|
2007-10-11 14:56:34 +00:00
|
|
|
if document.body[k].startswith("type"):
|
|
|
|
type = document.body[k][4:].strip().strip('"')
|
|
|
|
# I think there is a newline after \\end_inset, which should be removed.
|
|
|
|
if document.body[j + 1].strip() == "":
|
|
|
|
document.body[i : (j + 2)] = [type + ':' + arg]
|
|
|
|
else:
|
|
|
|
document.body[i : (j + 1)] = [type + ':' + arg]
|
|
|
|
|
|
|
|
|
2007-10-12 23:51:56 +00:00
|
|
|
def convert_pdf_options(document):
|
|
|
|
# Set the pdfusetitle tag, delete the pdf_store_options,
|
|
|
|
# set quotes for bookmarksopenlevel"
|
|
|
|
has_hr = get_value(document.header, "\\use_hyperref", 0, default = "0")
|
|
|
|
if has_hr == "1":
|
|
|
|
k = find_token(document.header, "\\use_hyperref", 0)
|
|
|
|
document.header.insert(k + 1, "\\pdf_pdfusetitle true")
|
|
|
|
k = find_token(document.header, "\\pdf_store_options", 0)
|
|
|
|
if k != -1:
|
|
|
|
del document.header[k]
|
|
|
|
i = find_token(document.header, "\\pdf_bookmarksopenlevel", k)
|
|
|
|
if i == -1: return
|
|
|
|
document.header[i] = document.header[i].replace('"', '')
|
|
|
|
|
|
|
|
|
2007-10-13 03:18:51 +00:00
|
|
|
def revert_pdf_options_2(document):
|
2007-10-12 23:51:56 +00:00
|
|
|
# reset the pdfusetitle tag, set quotes for bookmarksopenlevel"
|
|
|
|
k = find_token(document.header, "\\use_hyperref", 0)
|
|
|
|
i = find_token(document.header, "\\pdf_pdfusetitle", k)
|
|
|
|
if i != -1:
|
|
|
|
del document.header[i]
|
|
|
|
i = find_token(document.header, "\\pdf_bookmarksopenlevel", k)
|
|
|
|
if i == -1: return
|
|
|
|
values = document.header[i].split()
|
|
|
|
values[1] = ' "' + values[1] + '"'
|
|
|
|
document.header[i] = ''.join(values)
|
|
|
|
|
|
|
|
|
2007-10-13 19:06:09 +00:00
|
|
|
def convert_htmlurl(document):
|
2007-10-15 16:40:47 +00:00
|
|
|
'Convert "htmlurl" to "href" insets for docbook'
|
|
|
|
if document.backend != "docbook":
|
|
|
|
return
|
2007-10-13 19:06:09 +00:00
|
|
|
i = 0
|
|
|
|
while True:
|
2007-10-15 16:40:47 +00:00
|
|
|
i = find_token(document.body, "\\begin_inset CommandInset url", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
document.body[i] = "\\begin_inset CommandInset href"
|
|
|
|
document.body[i + 1] = "LatexCommand href"
|
|
|
|
i = i + 1
|
|
|
|
|
2007-10-29 22:38:36 +00:00
|
|
|
|
2007-10-15 16:40:47 +00:00
|
|
|
def convert_url(document):
|
|
|
|
'Convert url insets to url charstyles'
|
|
|
|
if document.backend == "docbook":
|
|
|
|
return
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, "\\begin_inset CommandInset url", i)
|
|
|
|
if i == -1:
|
|
|
|
break
|
2007-11-01 22:17:54 +00:00
|
|
|
n = find_token(document.body, "name", i)
|
|
|
|
if n == i + 2:
|
|
|
|
# place the URL name in typewriter before the new URL insert
|
|
|
|
# grab the name 'bla' from the e.g. the line 'name "bla"',
|
|
|
|
# therefore start with the 6th character
|
|
|
|
name = document.body[n][6:-1]
|
2007-11-01 22:40:31 +00:00
|
|
|
newname = [name + " "]
|
|
|
|
document.body[i:i] = newname
|
2007-11-01 22:17:54 +00:00
|
|
|
i = i + 1
|
2007-10-15 16:40:47 +00:00
|
|
|
j = find_token(document.body, "target", i)
|
|
|
|
if j == -1:
|
|
|
|
document.warning("Malformed LyX document: Can't find target for url inset")
|
2009-08-06 08:04:59 +00:00
|
|
|
i += 1
|
2007-10-15 16:40:47 +00:00
|
|
|
continue
|
2007-11-01 22:17:54 +00:00
|
|
|
target = document.body[j][8:-1]
|
2007-10-15 16:40:47 +00:00
|
|
|
k = find_token(document.body, "\\end_inset", j)
|
|
|
|
if k == -1:
|
|
|
|
document.warning("Malformed LyX document: Can't find end of url inset")
|
2009-08-06 08:04:59 +00:00
|
|
|
i = j
|
2007-10-15 16:40:47 +00:00
|
|
|
continue
|
|
|
|
newstuff = ["\\begin_inset Flex URL",
|
2008-03-19 09:40:24 +00:00
|
|
|
"status collapsed", "",
|
2007-10-15 16:40:47 +00:00
|
|
|
"\\begin_layout Standard",
|
2007-11-01 22:17:54 +00:00
|
|
|
"",
|
2007-10-15 16:40:47 +00:00
|
|
|
target,
|
|
|
|
"\\end_layout",
|
|
|
|
""]
|
|
|
|
document.body[i:k] = newstuff
|
2009-08-06 08:04:59 +00:00
|
|
|
i = i + len(newstuff)
|
2007-10-15 16:40:47 +00:00
|
|
|
|
2008-01-10 07:22:43 +00:00
|
|
|
def convert_ams_classes(document):
|
|
|
|
tc = document.textclass
|
|
|
|
if (tc != "amsart" and tc != "amsart-plain" and
|
2008-01-11 08:03:48 +00:00
|
|
|
tc != "amsart-seq" and tc != "amsbook"):
|
2008-01-10 07:22:43 +00:00
|
|
|
return
|
|
|
|
if tc == "amsart-plain":
|
|
|
|
document.textclass = "amsart"
|
|
|
|
document.set_textclass()
|
|
|
|
document.add_module("Theorems (Starred)")
|
|
|
|
return
|
|
|
|
if tc == "amsart-seq":
|
|
|
|
document.textclass = "amsart"
|
|
|
|
document.set_textclass()
|
|
|
|
document.add_module("Theorems (AMS)")
|
|
|
|
|
|
|
|
#Now we want to see if any of the environments in the extended theorems
|
|
|
|
#module were used in this document. If so, we'll add that module, too.
|
|
|
|
layouts = ["Criterion", "Algorithm", "Axiom", "Condition", "Note", \
|
|
|
|
"Notation", "Summary", "Acknowledgement", "Conclusion", "Fact", \
|
|
|
|
"Assumption"]
|
|
|
|
|
|
|
|
r = re.compile(r'^\\begin_layout (.*?)\*?\s*$')
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, "\\begin_layout", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
m = r.match(document.body[i])
|
|
|
|
if m == None:
|
2008-08-01 17:57:01 +00:00
|
|
|
# This is an empty layout
|
|
|
|
# document.warning("Weirdly formed \\begin_layout at line %d of body!" % i)
|
2008-01-10 07:22:43 +00:00
|
|
|
i += 1
|
|
|
|
continue
|
|
|
|
m = m.group(1)
|
|
|
|
if layouts.count(m) != 0:
|
|
|
|
document.add_module("Theorems (AMS-Extended)")
|
|
|
|
return
|
|
|
|
i += 1
|
2007-10-29 22:38:36 +00:00
|
|
|
|
2007-10-13 19:06:09 +00:00
|
|
|
def revert_href(document):
|
|
|
|
'Reverts hyperlink insets (href) to url insets (url)'
|
|
|
|
i = 0
|
|
|
|
while True:
|
2007-10-18 05:34:00 +00:00
|
|
|
i = find_token(document.body, "\\begin_inset CommandInset href", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
2007-10-18 16:49:22 +00:00
|
|
|
document.body[i : i + 2] = \
|
2007-10-18 05:34:00 +00:00
|
|
|
["\\begin_inset CommandInset url", "LatexCommand url"]
|
|
|
|
i = i + 2
|
2007-10-13 19:06:09 +00:00
|
|
|
|
2008-07-07 16:12:32 +00:00
|
|
|
def revert_url(document):
|
2008-07-07 16:22:55 +00:00
|
|
|
'Reverts Flex URL insets to old-style URL insets'
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, "\\begin_inset Flex URL", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
j = find_end_of_inset(document.body, i)
|
|
|
|
if j == -1:
|
|
|
|
document.warning("Can't find end of inset in revert_url!")
|
|
|
|
return
|
|
|
|
k = find_default_layout(document, i, j)
|
|
|
|
if k == -1:
|
|
|
|
document.warning("Can't find default layout in revert_url!")
|
|
|
|
i = j
|
|
|
|
continue
|
|
|
|
l = find_end_of(document.body, k, "\\begin_layout", "\\end_layout")
|
|
|
|
if l == -1 or l >= j:
|
|
|
|
document.warning("Can't find end of default layout in revert_url!")
|
|
|
|
i = j
|
|
|
|
continue
|
|
|
|
# OK, so the inset's data is between lines k and l.
|
|
|
|
data = " ".join(document.body[k+1:l])
|
|
|
|
data = data.strip()
|
|
|
|
newinset = ["\\begin_inset LatexCommand url", "target \"" + data + "\"",\
|
|
|
|
"", "\\end_inset"]
|
|
|
|
document.body[i:j+1] = newinset
|
|
|
|
i = i + len(newinset)
|
|
|
|
|
2007-10-29 22:38:36 +00:00
|
|
|
|
2007-10-23 15:02:15 +00:00
|
|
|
def convert_include(document):
|
|
|
|
'Converts include insets to new format.'
|
|
|
|
i = 0
|
|
|
|
r = re.compile(r'\\begin_inset Include\s+\\([^{]+){([^}]*)}(?:\[(.*)\])?')
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, "\\begin_inset Include", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
line = document.body[i]
|
|
|
|
previewline = document.body[i + 1]
|
|
|
|
m = r.match(line)
|
|
|
|
if m == None:
|
|
|
|
document.warning("Unable to match line " + str(i) + " of body!")
|
|
|
|
i += 1
|
|
|
|
continue
|
|
|
|
cmd = m.group(1)
|
|
|
|
fn = m.group(2)
|
|
|
|
opt = m.group(3)
|
2008-03-19 09:40:24 +00:00
|
|
|
insertion = ["\\begin_inset CommandInset include",
|
2007-10-23 15:02:15 +00:00
|
|
|
"LatexCommand " + cmd, previewline,
|
|
|
|
"filename \"" + fn + "\""]
|
|
|
|
newlines = 2
|
|
|
|
if opt:
|
|
|
|
insertion.append("lstparams " + '"' + opt + '"')
|
|
|
|
newlines += 1
|
|
|
|
document.body[i : i + 2] = insertion
|
|
|
|
i += newlines
|
|
|
|
|
2007-10-29 22:38:36 +00:00
|
|
|
|
2007-10-23 15:02:15 +00:00
|
|
|
def revert_include(document):
|
|
|
|
'Reverts include insets to old format.'
|
|
|
|
i = 0
|
2008-07-07 17:41:56 +00:00
|
|
|
r0 = re.compile('preview.*')
|
2007-10-23 15:02:15 +00:00
|
|
|
r1 = re.compile('LatexCommand (.+)')
|
2008-07-07 18:50:50 +00:00
|
|
|
r2 = re.compile('filename "(.+)"')
|
2008-07-07 18:17:16 +00:00
|
|
|
r3 = re.compile('lstparams "(.*)"')
|
2007-10-23 15:02:15 +00:00
|
|
|
while True:
|
|
|
|
i = find_token(document.body, "\\begin_inset CommandInset include", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
2008-07-07 17:41:56 +00:00
|
|
|
nextline = i + 1
|
|
|
|
m = r1.match(document.body[nextline])
|
2007-10-23 15:02:15 +00:00
|
|
|
if m == None:
|
|
|
|
document.warning("Malformed LyX document: No LatexCommand line for `" +
|
|
|
|
document.body[i] + "' on line " + str(i) + ".")
|
|
|
|
i += 1
|
|
|
|
continue
|
|
|
|
cmd = m.group(1)
|
2008-07-07 17:41:56 +00:00
|
|
|
nextline += 1
|
2008-10-09 12:51:37 +00:00
|
|
|
if r0.match(document.body[nextline]):
|
|
|
|
previewline = document.body[nextline]
|
|
|
|
nextline += 1
|
|
|
|
else:
|
|
|
|
previewline = ""
|
2008-07-07 17:41:56 +00:00
|
|
|
m = r2.match(document.body[nextline])
|
2007-10-23 15:02:15 +00:00
|
|
|
if m == None:
|
|
|
|
document.warning("Malformed LyX document: No filename line for `" + \
|
|
|
|
document.body[i] + "' on line " + str(i) + ".")
|
|
|
|
i += 2
|
|
|
|
continue
|
|
|
|
fn = m.group(1)
|
2008-07-07 17:41:56 +00:00
|
|
|
nextline += 1
|
2007-10-23 15:02:15 +00:00
|
|
|
options = ""
|
|
|
|
if (cmd == "lstinputlisting"):
|
2008-07-07 17:41:56 +00:00
|
|
|
m = r3.match(document.body[nextline])
|
2007-10-23 15:02:15 +00:00
|
|
|
if m != None:
|
|
|
|
options = m.group(1)
|
|
|
|
numlines = 5
|
2008-07-07 17:41:56 +00:00
|
|
|
nextline += 1
|
2007-10-23 15:02:15 +00:00
|
|
|
newline = "\\begin_inset Include \\" + cmd + "{" + fn + "}"
|
|
|
|
if options:
|
|
|
|
newline += ("[" + options + "]")
|
2008-07-07 17:41:56 +00:00
|
|
|
insertion = [newline]
|
|
|
|
if previewline != "":
|
|
|
|
insertion.append(previewline)
|
|
|
|
document.body[i : nextline] = insertion
|
2007-10-23 15:02:15 +00:00
|
|
|
i += 2
|
2007-10-29 22:38:36 +00:00
|
|
|
|
|
|
|
|
|
|
|
def revert_albanian(document):
|
|
|
|
"Set language Albanian to English"
|
|
|
|
i = 0
|
|
|
|
if document.language == "albanian":
|
|
|
|
document.language = "english"
|
|
|
|
i = find_token(document.header, "\\language", 0)
|
|
|
|
if i != -1:
|
|
|
|
document.header[i] = "\\language english"
|
|
|
|
j = 0
|
|
|
|
while True:
|
|
|
|
j = find_token(document.body, "\\lang albanian", j)
|
|
|
|
if j == -1:
|
|
|
|
return
|
|
|
|
document.body[j] = document.body[j].replace("\\lang albanian", "\\lang english")
|
|
|
|
j = j + 1
|
|
|
|
|
|
|
|
|
|
|
|
def revert_lowersorbian(document):
|
|
|
|
"Set language lower Sorbian to English"
|
|
|
|
i = 0
|
|
|
|
if document.language == "lowersorbian":
|
|
|
|
document.language = "english"
|
|
|
|
i = find_token(document.header, "\\language", 0)
|
|
|
|
if i != -1:
|
|
|
|
document.header[i] = "\\language english"
|
|
|
|
j = 0
|
|
|
|
while True:
|
|
|
|
j = find_token(document.body, "\\lang lowersorbian", j)
|
|
|
|
if j == -1:
|
|
|
|
return
|
|
|
|
document.body[j] = document.body[j].replace("\\lang lowersorbian", "\\lang english")
|
|
|
|
j = j + 1
|
|
|
|
|
2007-10-13 19:06:09 +00:00
|
|
|
|
2007-10-30 22:07:16 +00:00
|
|
|
def revert_uppersorbian(document):
|
|
|
|
"Set language uppersorbian to usorbian as this was used in LyX 1.5"
|
|
|
|
i = 0
|
|
|
|
if document.language == "uppersorbian":
|
|
|
|
document.language = "usorbian"
|
|
|
|
i = find_token(document.header, "\\language", 0)
|
|
|
|
if i != -1:
|
|
|
|
document.header[i] = "\\language usorbian"
|
|
|
|
j = 0
|
|
|
|
while True:
|
|
|
|
j = find_token(document.body, "\\lang uppersorbian", j)
|
|
|
|
if j == -1:
|
|
|
|
return
|
|
|
|
document.body[j] = document.body[j].replace("\\lang uppersorbian", "\\lang usorbian")
|
|
|
|
j = j + 1
|
|
|
|
|
|
|
|
|
|
|
|
def convert_usorbian(document):
|
2007-11-25 22:39:04 +00:00
|
|
|
"Set language usorbian to uppersorbian"
|
2007-10-30 22:07:16 +00:00
|
|
|
i = 0
|
|
|
|
if document.language == "usorbian":
|
|
|
|
document.language = "uppersorbian"
|
|
|
|
i = find_token(document.header, "\\language", 0)
|
|
|
|
if i != -1:
|
|
|
|
document.header[i] = "\\language uppersorbian"
|
|
|
|
j = 0
|
|
|
|
while True:
|
|
|
|
j = find_token(document.body, "\\lang usorbian", j)
|
|
|
|
if j == -1:
|
|
|
|
return
|
|
|
|
document.body[j] = document.body[j].replace("\\lang usorbian", "\\lang uppersorbian")
|
|
|
|
j = j + 1
|
|
|
|
|
|
|
|
|
2008-10-19 10:50:32 +00:00
|
|
|
def convert_macro_global(document):
|
|
|
|
"Remove TeX code command \global when it is in front of a macro"
|
|
|
|
# math macros are nowadays already defined \global, so that an additional
|
|
|
|
# \global would make the document uncompilable, see
|
|
|
|
# http://bugzilla.lyx.org/show_bug.cgi?id=5371
|
2008-10-23 18:08:26 +00:00
|
|
|
# We're looking for something like this:
|
|
|
|
# \begin_inset ERT
|
|
|
|
# status collapsed
|
|
|
|
#
|
|
|
|
# \begin_layout Plain Layout
|
2008-10-27 16:51:51 +00:00
|
|
|
#
|
|
|
|
#
|
2008-10-23 18:08:26 +00:00
|
|
|
# \backslash
|
|
|
|
# global
|
|
|
|
# \end_layout
|
2008-10-27 16:51:51 +00:00
|
|
|
#
|
2008-10-23 18:08:26 +00:00
|
|
|
# \end_inset
|
2008-10-27 16:51:51 +00:00
|
|
|
#
|
|
|
|
#
|
2008-10-23 18:08:26 +00:00
|
|
|
# \begin_inset FormulaMacro
|
|
|
|
# \renewcommand{\foo}{123}
|
|
|
|
# \end_inset
|
2008-10-19 10:50:32 +00:00
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, "\\begin_inset FormulaMacro", i)
|
2008-10-23 19:18:21 +00:00
|
|
|
if i == -1:
|
2008-10-19 10:50:32 +00:00
|
|
|
return
|
2008-10-23 18:08:26 +00:00
|
|
|
# if i <= 13, then there isn't enough room for the ERT
|
|
|
|
if i <= 12:
|
|
|
|
i += 1
|
|
|
|
continue
|
|
|
|
if document.body[i-6] == "global":
|
|
|
|
del document.body[i-13 : i]
|
|
|
|
i = i - 12
|
|
|
|
else:
|
|
|
|
i += 1
|
2008-10-19 10:50:32 +00:00
|
|
|
|
|
|
|
|
2007-11-01 17:37:43 +00:00
|
|
|
def revert_macro_optional_params(document):
|
|
|
|
"Convert macro definitions with optional parameters into ERTs"
|
|
|
|
# Stub to convert macro definitions with one or more optional parameters
|
|
|
|
# into uninterpreted ERT insets
|
|
|
|
|
|
|
|
|
|
|
|
def revert_hyperlinktype(document):
|
|
|
|
'Reverts hyperlink type'
|
|
|
|
i = 0
|
|
|
|
j = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, "target", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
j = find_token(document.body, "type", i)
|
|
|
|
if j == -1:
|
|
|
|
return
|
|
|
|
if j == i + 1:
|
|
|
|
del document.body[j]
|
|
|
|
i = i + 1
|
|
|
|
|
|
|
|
|
2007-11-23 02:10:00 +00:00
|
|
|
def revert_pagebreak(document):
|
2007-11-26 22:36:49 +00:00
|
|
|
'Reverts pagebreak to ERT'
|
2007-11-23 02:10:00 +00:00
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, "\\pagebreak", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
2007-11-26 22:36:49 +00:00
|
|
|
document.body[i] = '\\begin_inset ERT\nstatus collapsed\n\n' \
|
|
|
|
'\\begin_layout Standard\n\n\n\\backslash\n' \
|
|
|
|
'pagebreak{}\n\\end_layout\n\n\\end_inset\n\n'
|
2007-11-23 02:10:00 +00:00
|
|
|
i = i + 1
|
|
|
|
|
|
|
|
|
2007-11-25 18:34:37 +00:00
|
|
|
def revert_linebreak(document):
|
2007-11-26 22:36:49 +00:00
|
|
|
'Reverts linebreak to ERT'
|
2007-11-25 18:34:37 +00:00
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, "\\linebreak", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
2007-11-26 22:36:49 +00:00
|
|
|
document.body[i] = '\\begin_inset ERT\nstatus collapsed\n\n' \
|
|
|
|
'\\begin_layout Standard\n\n\n\\backslash\n' \
|
|
|
|
'linebreak{}\n\\end_layout\n\n\\end_inset\n\n'
|
2007-11-25 18:34:37 +00:00
|
|
|
i = i + 1
|
|
|
|
|
|
|
|
|
2007-11-25 21:09:01 +00:00
|
|
|
def revert_latin(document):
|
|
|
|
"Set language Latin to English"
|
|
|
|
i = 0
|
|
|
|
if document.language == "latin":
|
|
|
|
document.language = "english"
|
|
|
|
i = find_token(document.header, "\\language", 0)
|
|
|
|
if i != -1:
|
|
|
|
document.header[i] = "\\language english"
|
|
|
|
j = 0
|
|
|
|
while True:
|
|
|
|
j = find_token(document.body, "\\lang latin", j)
|
|
|
|
if j == -1:
|
|
|
|
return
|
|
|
|
document.body[j] = document.body[j].replace("\\lang latin", "\\lang english")
|
|
|
|
j = j + 1
|
|
|
|
|
|
|
|
|
|
|
|
def revert_samin(document):
|
|
|
|
"Set language North Sami to English"
|
|
|
|
i = 0
|
|
|
|
if document.language == "samin":
|
|
|
|
document.language = "english"
|
|
|
|
i = find_token(document.header, "\\language", 0)
|
|
|
|
if i != -1:
|
|
|
|
document.header[i] = "\\language english"
|
|
|
|
j = 0
|
|
|
|
while True:
|
|
|
|
j = find_token(document.body, "\\lang samin", j)
|
|
|
|
if j == -1:
|
|
|
|
return
|
|
|
|
document.body[j] = document.body[j].replace("\\lang samin", "\\lang english")
|
|
|
|
j = j + 1
|
|
|
|
|
|
|
|
|
2007-11-25 22:39:04 +00:00
|
|
|
def convert_serbocroatian(document):
|
|
|
|
"Set language Serbocroatian to Croatian as this was really Croatian in LyX 1.5"
|
|
|
|
i = 0
|
|
|
|
if document.language == "serbocroatian":
|
|
|
|
document.language = "croatian"
|
|
|
|
i = find_token(document.header, "\\language", 0)
|
|
|
|
if i != -1:
|
|
|
|
document.header[i] = "\\language croatian"
|
|
|
|
j = 0
|
|
|
|
while True:
|
|
|
|
j = find_token(document.body, "\\lang serbocroatian", j)
|
|
|
|
if j == -1:
|
|
|
|
return
|
|
|
|
document.body[j] = document.body[j].replace("\\lang serbocroatian", "\\lang croatian")
|
|
|
|
j = j + 1
|
|
|
|
|
|
|
|
|
2007-12-04 09:25:50 +00:00
|
|
|
def convert_framed_notes(document):
|
|
|
|
"Convert framed notes to boxes. "
|
|
|
|
i = 0
|
|
|
|
while 1:
|
|
|
|
i = find_tokens(document.body, ["\\begin_inset Note Framed", "\\begin_inset Note Shaded"], i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
2008-06-01 01:45:19 +00:00
|
|
|
subst = [document.body[i].replace("\\begin_inset Note", "\\begin_inset Box"),
|
|
|
|
'position "t"',
|
|
|
|
'hor_pos "c"',
|
2008-06-02 23:24:04 +00:00
|
|
|
'has_inner_box 0',
|
2008-08-01 17:57:01 +00:00
|
|
|
'inner_pos "t"',
|
2008-06-01 01:45:19 +00:00
|
|
|
'use_parbox 0',
|
|
|
|
'width "100col%"',
|
|
|
|
'special "none"',
|
|
|
|
'height "1in"',
|
|
|
|
'height_special "totalheight"']
|
|
|
|
document.body[i:i+1] = subst
|
|
|
|
i = i + 9
|
2007-12-04 09:25:50 +00:00
|
|
|
|
|
|
|
|
This commit changes the way individual LyXModule's are represented, both internally and in the .lyx files. The earlier version represented them by their `descriptive name', e.g., "Endnote" or "Theorems (AMS)", these being the same names used in the UI. This was a mistake, as becomes readily apparent when one starts to think about translating these strings. The modules ought to be represented by their filename, without the extension, just as TextClass's are.
The changes that accomplish this part are in ModuleList.{h,cpp}, configure.py, and the *.module files themselves. This is a format change, and the lyx2lyx is in those files.
By itself, that change would not be major, except for the fact that we do not want the module to be represented in the UI by its filename---e.g., theorems-std---but rather by a descriptive name, such as "Theorems". But that change turns out to be wholly non-trivial. The mechanism for choosing modules was the same as---indeed, was borrowed from---that in GuiCitation: You get a list of modules, and choosing them involves moving strings from one QListView to another. The models underlying these views are just QStringListModels, which means that, when you want to know what modules have been selected, you see what strings are in the "selected" QListView. But these are just the descriptive names, and we can't look up a module by its descriptive name if it's been translated. That, indeed, was the whole point of the change to the new representation.
So, we need a more complicated model underlying the QListView, one that will pair an identifying string---the filename minus the extension, in this case---with each item. This turns out not to be terribly difficult, though it took rather a while for me to understand why it's not difficult. There are two parts:
(i) GuiSelectionManger gets re-written to use any QAbstractListModel, not just a QStringListModel. This actually seems to improve the code, independently.
(ii) We then subclass QAbstractListModel to get the associated ID string, using the Qt::UserRole slot associated with each item to store its ID. This would be almost completely trivial if QAbstractListItem::itemData() included the QVariant associated with this role, but it doesn't, so there are some additional hoops through which to jump.
The new model, a GuiIdListModel, is defined in the files by that name. The changes in GuiSelectionManger.{h,cpp} make it more abstract; the changes in GuiDocument.{h,cpp} adapt it to the new framework.
I've also updated the module documenation to accord with this change.
git-svn-id: svn://svn.lyx.org/lyx/lyx-devel/trunk@22501 a592a061-630c-0410-9148-cb99ea01b6c8
2008-01-12 04:28:12 +00:00
|
|
|
def convert_module_names(document):
|
|
|
|
modulemap = { 'Braille' : 'braille', 'Endnote' : 'endnotes', 'Foot to End' : 'foottoend',\
|
|
|
|
'Hanging' : 'hanging', 'Linguistics' : 'linguistics', 'Logical Markup' : 'logicalmkup', \
|
|
|
|
'Theorems (AMS-Extended)' : 'theorems-ams-extended', 'Theorems (AMS)' : 'theorems-ams', \
|
|
|
|
'Theorems (Order By Chapter)' : 'theorems-chap', 'Theorems (Order By Section)' : 'theorems-sec', \
|
|
|
|
'Theorems (Starred)' : 'theorems-starred', 'Theorems' : 'theorems-std' }
|
|
|
|
modlist = document.get_module_list()
|
2008-01-12 06:43:44 +00:00
|
|
|
if len(modlist) == 0:
|
|
|
|
return
|
This commit changes the way individual LyXModule's are represented, both internally and in the .lyx files. The earlier version represented them by their `descriptive name', e.g., "Endnote" or "Theorems (AMS)", these being the same names used in the UI. This was a mistake, as becomes readily apparent when one starts to think about translating these strings. The modules ought to be represented by their filename, without the extension, just as TextClass's are.
The changes that accomplish this part are in ModuleList.{h,cpp}, configure.py, and the *.module files themselves. This is a format change, and the lyx2lyx is in those files.
By itself, that change would not be major, except for the fact that we do not want the module to be represented in the UI by its filename---e.g., theorems-std---but rather by a descriptive name, such as "Theorems". But that change turns out to be wholly non-trivial. The mechanism for choosing modules was the same as---indeed, was borrowed from---that in GuiCitation: You get a list of modules, and choosing them involves moving strings from one QListView to another. The models underlying these views are just QStringListModels, which means that, when you want to know what modules have been selected, you see what strings are in the "selected" QListView. But these are just the descriptive names, and we can't look up a module by its descriptive name if it's been translated. That, indeed, was the whole point of the change to the new representation.
So, we need a more complicated model underlying the QListView, one that will pair an identifying string---the filename minus the extension, in this case---with each item. This turns out not to be terribly difficult, though it took rather a while for me to understand why it's not difficult. There are two parts:
(i) GuiSelectionManger gets re-written to use any QAbstractListModel, not just a QStringListModel. This actually seems to improve the code, independently.
(ii) We then subclass QAbstractListModel to get the associated ID string, using the Qt::UserRole slot associated with each item to store its ID. This would be almost completely trivial if QAbstractListItem::itemData() included the QVariant associated with this role, but it doesn't, so there are some additional hoops through which to jump.
The new model, a GuiIdListModel, is defined in the files by that name. The changes in GuiSelectionManger.{h,cpp} make it more abstract; the changes in GuiDocument.{h,cpp} adapt it to the new framework.
I've also updated the module documenation to accord with this change.
git-svn-id: svn://svn.lyx.org/lyx/lyx-devel/trunk@22501 a592a061-630c-0410-9148-cb99ea01b6c8
2008-01-12 04:28:12 +00:00
|
|
|
newmodlist = []
|
|
|
|
for mod in modlist:
|
|
|
|
if modulemap.has_key(mod):
|
|
|
|
newmodlist.append(modulemap[mod])
|
|
|
|
else:
|
|
|
|
document.warning("Can't find module %s in the module map!" % mod)
|
|
|
|
newmodlist.append(mod)
|
|
|
|
document.set_module_list(newmodlist)
|
|
|
|
|
|
|
|
|
|
|
|
def revert_module_names(document):
|
|
|
|
modulemap = { 'braille' : 'Braille', 'endnotes' : 'Endnote', 'foottoend' : 'Foot to End',\
|
|
|
|
'hanging' : 'Hanging', 'linguistics' : 'Linguistics', 'logicalmkup' : 'Logical Markup', \
|
|
|
|
'theorems-ams-extended' : 'Theorems (AMS-Extended)', 'theorems-ams' : 'Theorems (AMS)', \
|
|
|
|
'theorems-chap' : 'Theorems (Order By Chapter)', 'theorems-sec' : 'Theorems (Order By Section)', \
|
|
|
|
'theorems-starred' : 'Theorems (Starred)', 'theorems-std' : 'Theorems'}
|
|
|
|
modlist = document.get_module_list()
|
2008-01-12 06:43:44 +00:00
|
|
|
if len(modlist) == 0:
|
|
|
|
return
|
This commit changes the way individual LyXModule's are represented, both internally and in the .lyx files. The earlier version represented them by their `descriptive name', e.g., "Endnote" or "Theorems (AMS)", these being the same names used in the UI. This was a mistake, as becomes readily apparent when one starts to think about translating these strings. The modules ought to be represented by their filename, without the extension, just as TextClass's are.
The changes that accomplish this part are in ModuleList.{h,cpp}, configure.py, and the *.module files themselves. This is a format change, and the lyx2lyx is in those files.
By itself, that change would not be major, except for the fact that we do not want the module to be represented in the UI by its filename---e.g., theorems-std---but rather by a descriptive name, such as "Theorems". But that change turns out to be wholly non-trivial. The mechanism for choosing modules was the same as---indeed, was borrowed from---that in GuiCitation: You get a list of modules, and choosing them involves moving strings from one QListView to another. The models underlying these views are just QStringListModels, which means that, when you want to know what modules have been selected, you see what strings are in the "selected" QListView. But these are just the descriptive names, and we can't look up a module by its descriptive name if it's been translated. That, indeed, was the whole point of the change to the new representation.
So, we need a more complicated model underlying the QListView, one that will pair an identifying string---the filename minus the extension, in this case---with each item. This turns out not to be terribly difficult, though it took rather a while for me to understand why it's not difficult. There are two parts:
(i) GuiSelectionManger gets re-written to use any QAbstractListModel, not just a QStringListModel. This actually seems to improve the code, independently.
(ii) We then subclass QAbstractListModel to get the associated ID string, using the Qt::UserRole slot associated with each item to store its ID. This would be almost completely trivial if QAbstractListItem::itemData() included the QVariant associated with this role, but it doesn't, so there are some additional hoops through which to jump.
The new model, a GuiIdListModel, is defined in the files by that name. The changes in GuiSelectionManger.{h,cpp} make it more abstract; the changes in GuiDocument.{h,cpp} adapt it to the new framework.
I've also updated the module documenation to accord with this change.
git-svn-id: svn://svn.lyx.org/lyx/lyx-devel/trunk@22501 a592a061-630c-0410-9148-cb99ea01b6c8
2008-01-12 04:28:12 +00:00
|
|
|
newmodlist = []
|
|
|
|
for mod in modlist:
|
|
|
|
if modulemap.has_key(mod):
|
|
|
|
newmodlist.append(modulemap[mod])
|
|
|
|
else:
|
|
|
|
document.warning("Can't find module %s in the module map!" % mod)
|
|
|
|
newmodlist.append(mod)
|
|
|
|
document.set_module_list(newmodlist)
|
|
|
|
|
|
|
|
|
2008-02-19 05:24:48 +00:00
|
|
|
def revert_colsep(document):
|
|
|
|
i = find_token(document.header, "\\columnsep", 0)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
colsepline = document.header[i]
|
|
|
|
r = re.compile(r'\\columnsep (.*)')
|
|
|
|
m = r.match(colsepline)
|
|
|
|
if not m:
|
|
|
|
document.warning("Malformed column separation line!")
|
|
|
|
return
|
|
|
|
colsep = m.group(1)
|
|
|
|
del document.header[i]
|
|
|
|
#it seems to be safe to add the package even if it is already used
|
|
|
|
pretext = ["\\usepackage{geometry}", "\\geometry{columnsep=" + colsep + "}"]
|
|
|
|
|
|
|
|
add_to_preamble(document, pretext)
|
|
|
|
|
|
|
|
|
2007-12-04 09:25:50 +00:00
|
|
|
def revert_framed_notes(document):
|
|
|
|
"Revert framed boxes to notes. "
|
|
|
|
i = 0
|
|
|
|
while 1:
|
|
|
|
i = find_tokens(document.body, ["\\begin_inset Box Framed", "\\begin_inset Box Shaded"], i)
|
|
|
|
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
j = find_end_of_inset(document.body, i + 1)
|
|
|
|
if j == -1:
|
|
|
|
# should not happen
|
|
|
|
document.warning("Malformed LyX document: Could not find end of Box inset.")
|
|
|
|
k = find_token(document.body, "status", i + 1, j)
|
|
|
|
if k == -1:
|
|
|
|
document.warning("Malformed LyX document: Missing `status' tag in Box inset.")
|
|
|
|
return
|
|
|
|
status = document.body[k]
|
2008-04-15 17:59:01 +00:00
|
|
|
l = find_default_layout(document, i + 1, j)
|
2008-04-15 15:52:01 +00:00
|
|
|
if l == -1:
|
|
|
|
document.warning("Malformed LyX document: Missing `\\begin_layout' in Box inset.")
|
2007-12-04 09:25:50 +00:00
|
|
|
return
|
|
|
|
m = find_token(document.body, "\\end_layout", i + 1, j)
|
|
|
|
if m == -1:
|
|
|
|
document.warning("Malformed LyX document: Missing `\\end_layout' in Box inset.")
|
|
|
|
return
|
|
|
|
ibox = find_token(document.body, "has_inner_box 1", i + 1, k)
|
|
|
|
pbox = find_token(document.body, "use_parbox 1", i + 1, k)
|
|
|
|
if ibox == -1 and pbox == -1:
|
|
|
|
document.body[i] = document.body[i].replace("\\begin_inset Box", "\\begin_inset Note")
|
|
|
|
del document.body[i+1:k]
|
|
|
|
else:
|
|
|
|
document.body[i] = document.body[i].replace("\\begin_inset Box Shaded", "\\begin_inset Box Frameless")
|
2008-06-01 01:50:39 +00:00
|
|
|
subst1 = [document.body[l],
|
|
|
|
"\\begin_inset Note Shaded",
|
|
|
|
status,
|
|
|
|
'\\begin_layout Standard']
|
|
|
|
document.body[l:l + 1] = subst1
|
|
|
|
subst2 = [document.body[m], "\\end_layout", "\\end_inset"]
|
|
|
|
document.body[m:m + 1] = subst2
|
2007-12-04 09:25:50 +00:00
|
|
|
i = i + 1
|
|
|
|
|
|
|
|
|
2007-12-06 11:04:56 +00:00
|
|
|
def revert_slash(document):
|
|
|
|
'Revert \\SpecialChar \\slash{} to ERT'
|
2008-06-01 14:27:43 +00:00
|
|
|
i = 0
|
|
|
|
while i < len(document.body):
|
2008-11-12 09:42:37 +00:00
|
|
|
m = re.match(r'(.*)\\SpecialChar \\slash{}(.*)', document.body[i])
|
2008-06-01 02:04:33 +00:00
|
|
|
if m:
|
2008-11-12 09:42:37 +00:00
|
|
|
before = m.group(1)
|
|
|
|
after = m.group(2)
|
|
|
|
subst = [before,
|
|
|
|
'\\begin_inset ERT',
|
|
|
|
'status collapsed', '',
|
|
|
|
'\\begin_layout Standard',
|
|
|
|
'', '', '\\backslash',
|
|
|
|
'slash{}',
|
|
|
|
'\\end_layout', '',
|
|
|
|
'\\end_inset', '',
|
|
|
|
after]
|
|
|
|
document.body[i: i+1] = subst
|
|
|
|
i = i + len(subst)
|
2008-06-01 14:27:43 +00:00
|
|
|
else:
|
2008-11-12 09:42:37 +00:00
|
|
|
i = i + 1
|
2007-12-26 20:15:31 +00:00
|
|
|
|
2007-12-06 11:04:56 +00:00
|
|
|
|
|
|
|
def revert_nobreakdash(document):
|
|
|
|
'Revert \\SpecialChar \\nobreakdash- to ERT'
|
2008-06-01 14:55:47 +00:00
|
|
|
i = 0
|
|
|
|
while i < len(document.body):
|
2008-11-12 09:42:37 +00:00
|
|
|
m = re.match(r'(.*)\\SpecialChar \\nobreakdash-(.*)', document.body[i])
|
2007-12-06 11:04:56 +00:00
|
|
|
if m:
|
2008-11-12 09:42:37 +00:00
|
|
|
before = m.group(1)
|
|
|
|
after = m.group(2)
|
|
|
|
subst = [before,
|
|
|
|
'\\begin_inset ERT',
|
2008-06-01 02:18:51 +00:00
|
|
|
'status collapsed', '',
|
|
|
|
'\\begin_layout Standard', '', '',
|
|
|
|
'\\backslash',
|
|
|
|
'nobreakdash-',
|
|
|
|
'\\end_layout', '',
|
2008-11-12 09:42:37 +00:00
|
|
|
'\\end_inset', '',
|
|
|
|
after]
|
|
|
|
document.body[i: i+1] = subst
|
2008-06-01 14:55:47 +00:00
|
|
|
i = i + len(subst)
|
2008-06-01 02:18:51 +00:00
|
|
|
j = find_token(document.header, "\\use_amsmath", 0)
|
|
|
|
if j == -1:
|
|
|
|
document.warning("Malformed LyX document: Missing '\\use_amsmath'.")
|
|
|
|
return
|
|
|
|
document.header[j] = "\\use_amsmath 2"
|
2008-06-01 14:55:47 +00:00
|
|
|
else:
|
|
|
|
i = i + 1
|
2007-12-06 11:04:56 +00:00
|
|
|
|
|
|
|
|
2008-06-01 19:59:15 +00:00
|
|
|
#Returns number of lines added/removed
|
2007-12-20 15:46:14 +00:00
|
|
|
def revert_nocite_key(body, start, end):
|
2008-08-01 17:57:01 +00:00
|
|
|
'key "..." -> \nocite{...}'
|
2008-06-01 02:26:56 +00:00
|
|
|
r = re.compile(r'^key "(.*)"')
|
2008-06-01 14:55:47 +00:00
|
|
|
i = start
|
|
|
|
j = end
|
2008-06-01 19:59:15 +00:00
|
|
|
while i < j:
|
2008-06-01 02:26:56 +00:00
|
|
|
m = r.match(body[i])
|
|
|
|
if m:
|
|
|
|
body[i:i+1] = ["\\backslash", "nocite{" + m.group(1) + "}"]
|
2008-06-01 19:59:15 +00:00
|
|
|
j += 1 # because we added a line
|
|
|
|
i += 2 # skip that line
|
2007-12-20 15:46:14 +00:00
|
|
|
else:
|
2008-06-01 19:59:15 +00:00
|
|
|
del body[i]
|
|
|
|
j -= 1 # because we deleted a line
|
|
|
|
# no need to change i, since it now points to the next line
|
|
|
|
return j - end
|
2007-12-20 15:46:14 +00:00
|
|
|
|
|
|
|
|
|
|
|
def revert_nocite(document):
|
|
|
|
"Revert LatexCommand nocite to ERT"
|
|
|
|
i = 0
|
|
|
|
while 1:
|
|
|
|
i = find_token(document.body, "\\begin_inset CommandInset citation", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
2008-06-01 19:59:15 +00:00
|
|
|
if (document.body[i+1] != "LatexCommand nocite"):
|
2008-06-01 14:55:47 +00:00
|
|
|
# note that we already incremented i
|
2008-06-01 19:59:15 +00:00
|
|
|
i = i + 1
|
2008-06-01 14:55:47 +00:00
|
|
|
continue
|
2008-06-01 19:59:15 +00:00
|
|
|
insetEnd = find_end_of_inset(document.body, i)
|
|
|
|
if insetEnd == -1:
|
2008-06-01 14:55:47 +00:00
|
|
|
#this should not happen
|
|
|
|
document.warning("End of CommandInset citation not found in revert_nocite!")
|
|
|
|
return
|
2008-06-01 19:59:15 +00:00
|
|
|
|
|
|
|
paramLocation = i + 2 #start of the inset's parameters
|
|
|
|
addedLines = 0
|
|
|
|
document.body[i:i+2] = \
|
|
|
|
["\\begin_inset ERT", "status collapsed", "", "\\begin_layout Standard"]
|
|
|
|
# that added two lines
|
|
|
|
paramLocation += 2
|
|
|
|
insetEnd += 2
|
|
|
|
#print insetEnd, document.body[i: insetEnd + 1]
|
|
|
|
insetEnd += revert_nocite_key(document.body, paramLocation, insetEnd)
|
|
|
|
#print insetEnd, document.body[i: insetEnd + 1]
|
|
|
|
document.body.insert(insetEnd, "\\end_layout")
|
|
|
|
document.body.insert(insetEnd + 1, "")
|
|
|
|
i = insetEnd + 1
|
2007-12-20 15:46:14 +00:00
|
|
|
|
|
|
|
|
2007-12-28 16:56:57 +00:00
|
|
|
def revert_btprintall(document):
|
|
|
|
"Revert (non-bibtopic) btPrintAll option to ERT \nocite{*}"
|
|
|
|
i = find_token(document.header, '\\use_bibtopic', 0)
|
|
|
|
if i == -1:
|
|
|
|
document.warning("Malformed lyx document: Missing '\\use_bibtopic'.")
|
|
|
|
return
|
|
|
|
if get_value(document.header, '\\use_bibtopic', 0) == "false":
|
|
|
|
i = 0
|
|
|
|
while i < len(document.body):
|
|
|
|
i = find_token(document.body, "\\begin_inset CommandInset bibtex", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
j = find_end_of_inset(document.body, i + 1)
|
|
|
|
if j == -1:
|
|
|
|
#this should not happen
|
|
|
|
document.warning("End of CommandInset bibtex not found in revert_btprintall!")
|
|
|
|
j = len(document.body)
|
2008-06-01 14:57:33 +00:00
|
|
|
# this range isn't really right, but it should be OK, since we shouldn't
|
|
|
|
# see more than one matching line in each inset
|
2008-06-01 15:01:28 +00:00
|
|
|
addedlines = 0
|
2007-12-28 16:56:57 +00:00
|
|
|
for k in range(i, j):
|
|
|
|
if (document.body[k] == 'btprint "btPrintAll"'):
|
|
|
|
del document.body[k]
|
2008-06-01 02:21:29 +00:00
|
|
|
subst = ["\\begin_inset ERT",
|
|
|
|
"status collapsed", "",
|
|
|
|
"\\begin_layout Standard", "",
|
|
|
|
"\\backslash",
|
|
|
|
"nocite{*}",
|
|
|
|
"\\end_layout",
|
|
|
|
"\\end_inset"]
|
|
|
|
document.body[i:i] = subst
|
2008-06-01 15:01:28 +00:00
|
|
|
addlines = addedlines + len(subst) - 1
|
|
|
|
i = j + addedlines
|
2007-12-28 16:56:57 +00:00
|
|
|
|
|
|
|
|
2007-12-05 21:42:57 +00:00
|
|
|
def revert_bahasam(document):
|
|
|
|
"Set language Bahasa Malaysia to Bahasa Indonesia"
|
|
|
|
i = 0
|
|
|
|
if document.language == "bahasam":
|
|
|
|
document.language = "bahasa"
|
|
|
|
i = find_token(document.header, "\\language", 0)
|
|
|
|
if i != -1:
|
|
|
|
document.header[i] = "\\language bahasa"
|
|
|
|
j = 0
|
|
|
|
while True:
|
|
|
|
j = find_token(document.body, "\\lang bahasam", j)
|
|
|
|
if j == -1:
|
|
|
|
return
|
|
|
|
document.body[j] = document.body[j].replace("\\lang bahasam", "\\lang bahasa")
|
|
|
|
j = j + 1
|
|
|
|
|
|
|
|
|
2007-12-05 22:04:33 +00:00
|
|
|
def revert_interlingua(document):
|
|
|
|
"Set language Interlingua to English"
|
|
|
|
i = 0
|
|
|
|
if document.language == "interlingua":
|
|
|
|
document.language = "english"
|
|
|
|
i = find_token(document.header, "\\language", 0)
|
|
|
|
if i != -1:
|
|
|
|
document.header[i] = "\\language english"
|
|
|
|
j = 0
|
|
|
|
while True:
|
|
|
|
j = find_token(document.body, "\\lang interlingua", j)
|
|
|
|
if j == -1:
|
|
|
|
return
|
|
|
|
document.body[j] = document.body[j].replace("\\lang interlingua", "\\lang english")
|
|
|
|
j = j + 1
|
|
|
|
|
|
|
|
|
2007-12-13 23:29:56 +00:00
|
|
|
def revert_serbianlatin(document):
|
|
|
|
"Set language Serbian-Latin to Croatian"
|
|
|
|
i = 0
|
|
|
|
if document.language == "serbian-latin":
|
|
|
|
document.language = "croatian"
|
|
|
|
i = find_token(document.header, "\\language", 0)
|
|
|
|
if i != -1:
|
|
|
|
document.header[i] = "\\language croatian"
|
|
|
|
j = 0
|
|
|
|
while True:
|
|
|
|
j = find_token(document.body, "\\lang serbian-latin", j)
|
|
|
|
if j == -1:
|
|
|
|
return
|
|
|
|
document.body[j] = document.body[j].replace("\\lang serbian-latin", "\\lang croatian")
|
|
|
|
j = j + 1
|
|
|
|
|
|
|
|
|
2008-01-11 18:56:53 +00:00
|
|
|
def revert_rotfloat(document):
|
2008-01-12 09:59:37 +00:00
|
|
|
" Revert sideways custom floats. "
|
2008-01-11 18:56:53 +00:00
|
|
|
i = 0
|
|
|
|
while 1:
|
2008-05-05 06:37:25 +00:00
|
|
|
# whitespace intended (exclude \\begin_inset FloatList)
|
|
|
|
i = find_token(document.body, "\\begin_inset Float ", i)
|
2008-01-11 18:56:53 +00:00
|
|
|
if i == -1:
|
|
|
|
return
|
2008-01-12 09:59:37 +00:00
|
|
|
line = document.body[i]
|
|
|
|
r = re.compile(r'\\begin_inset Float (.*)$')
|
|
|
|
m = r.match(line)
|
2008-05-05 06:37:25 +00:00
|
|
|
if m == None:
|
|
|
|
document.warning("Unable to match line " + str(i) + " of body!")
|
|
|
|
i += 1
|
|
|
|
continue
|
2008-01-12 09:59:37 +00:00
|
|
|
floattype = m.group(1)
|
|
|
|
if floattype == "figure" or floattype == "table":
|
2008-06-02 19:42:53 +00:00
|
|
|
i += 1
|
2008-01-12 09:59:37 +00:00
|
|
|
continue
|
2008-01-11 18:56:53 +00:00
|
|
|
j = find_end_of_inset(document.body, i)
|
|
|
|
if j == -1:
|
2008-06-05 05:00:53 +00:00
|
|
|
document.warning("Malformed lyx document: Missing '\\end_inset' in revert_rotfloat.")
|
2008-06-02 19:42:53 +00:00
|
|
|
i += 1
|
2008-01-11 18:56:53 +00:00
|
|
|
continue
|
2008-06-02 19:42:53 +00:00
|
|
|
addedLines = 0
|
|
|
|
if get_value(document.body, 'sideways', i, j) == "false":
|
|
|
|
i += 1
|
2008-01-11 18:56:53 +00:00
|
|
|
continue
|
2008-06-02 19:42:53 +00:00
|
|
|
l = find_default_layout(document, i + 1, j)
|
|
|
|
if l == -1:
|
|
|
|
document.warning("Malformed LyX document: Missing `\\begin_layout' in Float inset.")
|
|
|
|
return
|
|
|
|
subst = ['\\begin_layout Standard',
|
|
|
|
'\\begin_inset ERT',
|
|
|
|
'status collapsed', '',
|
2008-08-01 17:57:01 +00:00
|
|
|
'\\begin_layout Standard', '', '',
|
2008-06-02 19:42:53 +00:00
|
|
|
'\\backslash', '',
|
|
|
|
'end{sideways' + floattype + '}',
|
|
|
|
'\\end_layout', '', '\\end_inset']
|
|
|
|
document.body[j : j+1] = subst
|
|
|
|
addedLines = len(subst) - 1
|
|
|
|
del document.body[i+1 : l]
|
2008-08-01 17:57:01 +00:00
|
|
|
addedLines -= (l-1) - (i+1)
|
2008-06-02 19:42:53 +00:00
|
|
|
subst = ['\\begin_inset ERT', 'status collapsed', '',
|
2008-08-01 17:57:01 +00:00
|
|
|
'\\begin_layout Standard', '', '', '\\backslash',
|
|
|
|
'begin{sideways' + floattype + '}',
|
2008-06-02 19:42:53 +00:00
|
|
|
'\\end_layout', '', '\\end_inset', '',
|
|
|
|
'\\end_layout', '']
|
|
|
|
document.body[i : i+1] = subst
|
|
|
|
addedLines += len(subst) - 1
|
|
|
|
if floattype == "algorithm":
|
|
|
|
add_to_preamble(document,
|
|
|
|
['% Commands inserted by lyx2lyx for sideways algorithm float',
|
|
|
|
'\\usepackage{rotfloat}',
|
|
|
|
'\\floatstyle{ruled}',
|
|
|
|
'\\newfloat{algorithm}{tbp}{loa}',
|
|
|
|
'\\floatname{algorithm}{Algorithm}'])
|
|
|
|
else:
|
|
|
|
document.warning("Cannot create preamble definition for custom float" + floattype + ".")
|
|
|
|
i += addedLines + 1
|
2008-01-11 18:56:53 +00:00
|
|
|
|
|
|
|
|
|
|
|
def revert_widesideways(document):
|
|
|
|
" Revert wide sideways floats. "
|
|
|
|
i = 0
|
|
|
|
while 1:
|
2008-05-05 06:37:25 +00:00
|
|
|
# whitespace intended (exclude \\begin_inset FloatList)
|
|
|
|
i = find_token(document.body, '\\begin_inset Float ', i)
|
2008-01-11 18:56:53 +00:00
|
|
|
if i == -1:
|
|
|
|
return
|
2008-01-12 09:59:37 +00:00
|
|
|
line = document.body[i]
|
|
|
|
r = re.compile(r'\\begin_inset Float (.*)$')
|
|
|
|
m = r.match(line)
|
2008-05-05 06:37:25 +00:00
|
|
|
if m == None:
|
|
|
|
document.warning("Unable to match line " + str(i) + " of body!")
|
|
|
|
i += 1
|
|
|
|
continue
|
2008-01-12 09:59:37 +00:00
|
|
|
floattype = m.group(1)
|
|
|
|
if floattype != "figure" and floattype != "table":
|
2008-06-02 19:42:53 +00:00
|
|
|
i += 1
|
2008-01-12 09:59:37 +00:00
|
|
|
continue
|
2008-01-11 18:56:53 +00:00
|
|
|
j = find_end_of_inset(document.body, i)
|
|
|
|
if j == -1:
|
2008-06-05 05:00:53 +00:00
|
|
|
document.warning("Malformed lyx document: Missing '\\end_inset' in revert_widesideways.")
|
2008-06-02 19:42:53 +00:00
|
|
|
i += 1
|
2008-01-11 18:56:53 +00:00
|
|
|
continue
|
2008-06-02 19:42:53 +00:00
|
|
|
if get_value(document.body, 'sideways', i, j) == "false" or \
|
|
|
|
get_value(document.body, 'wide', i, j) == "false":
|
|
|
|
i += 1
|
|
|
|
continue
|
|
|
|
l = find_default_layout(document, i + 1, j)
|
|
|
|
if l == -1:
|
|
|
|
document.warning("Malformed LyX document: Missing `\\begin_layout' in Float inset.")
|
|
|
|
return
|
2008-08-01 17:57:01 +00:00
|
|
|
subst = ['\\begin_layout Standard', '\\begin_inset ERT',
|
|
|
|
'status collapsed', '',
|
2008-06-02 19:42:53 +00:00
|
|
|
'\\begin_layout Standard', '', '', '\\backslash',
|
2008-08-01 17:57:01 +00:00
|
|
|
'end{sideways' + floattype + '*}',
|
2008-06-02 19:42:53 +00:00
|
|
|
'\\end_layout', '', '\\end_inset']
|
|
|
|
document.body[j : j+1] = subst
|
|
|
|
addedLines = len(subst) - 1
|
|
|
|
del document.body[i+1:l-1]
|
|
|
|
addedLines -= (l-1) - (i+1)
|
|
|
|
subst = ['\\begin_inset ERT', 'status collapsed', '',
|
|
|
|
'\\begin_layout Standard', '', '', '\\backslash',
|
|
|
|
'begin{sideways' + floattype + '*}', '\\end_layout', '',
|
|
|
|
'\\end_inset', '', '\\end_layout', '']
|
|
|
|
document.body[i : i+1] = subst
|
|
|
|
addedLines += len(subst) - 1
|
|
|
|
add_to_preamble(document, ['\\usepackage{rotfloat}\n'])
|
|
|
|
i += addedLines + 1
|
2008-01-11 18:56:53 +00:00
|
|
|
|
|
|
|
|
2008-04-18 17:06:03 +00:00
|
|
|
def revert_inset_embedding(document, type):
|
|
|
|
' Remove embed tag from certain type of insets'
|
2008-04-11 17:20:59 +00:00
|
|
|
i = 0
|
|
|
|
while 1:
|
2008-04-18 17:06:03 +00:00
|
|
|
i = find_token(document.body, "\\begin_inset %s" % type, i)
|
2008-04-11 17:20:59 +00:00
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
j = find_end_of_inset(document.body, i)
|
|
|
|
if j == -1:
|
2008-06-05 05:00:53 +00:00
|
|
|
document.warning("Malformed lyx document: Missing '\\end_inset' in revert_inset_embedding.")
|
2008-04-11 17:20:59 +00:00
|
|
|
i = i + 1
|
|
|
|
continue
|
|
|
|
k = find_token(document.body, "\tembed", i, j)
|
2008-04-18 17:06:03 +00:00
|
|
|
if k == -1:
|
|
|
|
k = find_token(document.body, "embed", i, j)
|
2008-04-11 17:20:59 +00:00
|
|
|
if k != -1:
|
|
|
|
del document.body[k]
|
|
|
|
i = i + 1
|
|
|
|
|
|
|
|
|
2008-04-18 17:06:03 +00:00
|
|
|
def revert_external_embedding(document):
|
|
|
|
' Remove embed tag from external inset '
|
|
|
|
revert_inset_embedding(document, 'External')
|
|
|
|
|
|
|
|
|
2008-03-02 11:30:50 +00:00
|
|
|
def convert_subfig(document):
|
|
|
|
" Convert subfigures to subfloats. "
|
|
|
|
i = 0
|
|
|
|
while 1:
|
2008-10-25 13:43:07 +00:00
|
|
|
addedLines = 0
|
2008-03-02 11:30:50 +00:00
|
|
|
i = find_token(document.body, '\\begin_inset Graphics', i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
2008-06-03 00:41:38 +00:00
|
|
|
endInset = find_end_of_inset(document.body, i)
|
|
|
|
if endInset == -1:
|
2008-06-05 05:00:53 +00:00
|
|
|
document.warning("Malformed lyx document: Missing '\\end_inset' in convert_subfig.")
|
2008-06-02 19:42:53 +00:00
|
|
|
i += 1
|
2008-03-02 11:30:50 +00:00
|
|
|
continue
|
2008-06-03 00:41:38 +00:00
|
|
|
k = find_token(document.body, '\tsubcaption', i, endInset)
|
2008-03-02 11:30:50 +00:00
|
|
|
if k == -1:
|
2008-07-07 18:21:53 +00:00
|
|
|
i = endInset
|
2008-03-02 11:30:50 +00:00
|
|
|
continue
|
2008-06-03 00:41:38 +00:00
|
|
|
l = find_token(document.body, '\tsubcaptionText', i, endInset)
|
2008-07-07 18:21:53 +00:00
|
|
|
if l == -1:
|
2008-10-25 13:43:07 +00:00
|
|
|
caption = ""
|
2008-10-27 16:51:51 +00:00
|
|
|
else:
|
2008-10-25 13:43:07 +00:00
|
|
|
caption = document.body[l][16:].strip('"')
|
|
|
|
del document.body[l]
|
|
|
|
addedLines -= 1
|
2008-03-17 14:12:42 +00:00
|
|
|
del document.body[k]
|
2008-10-25 13:43:07 +00:00
|
|
|
addedLines -= 1
|
2008-08-01 17:57:01 +00:00
|
|
|
subst = ['\\begin_inset Float figure', 'wide false', 'sideways false',
|
|
|
|
'status open', '', '\\begin_layout Plain Layout', '\\begin_inset Caption',
|
2009-09-10 12:10:18 +00:00
|
|
|
'', '\\begin_layout Plain Layout'] + latex2lyx(caption, False) + \
|
2008-08-01 17:57:01 +00:00
|
|
|
[ '\\end_layout', '', '\\end_inset', '',
|
2008-07-07 18:21:53 +00:00
|
|
|
'\\end_layout', '', '\\begin_layout Plain Layout']
|
|
|
|
document.body[i : i] = subst
|
|
|
|
addedLines += len(subst)
|
2008-06-03 00:41:38 +00:00
|
|
|
endInset += addedLines
|
2008-07-07 18:21:53 +00:00
|
|
|
subst = ['', '\\end_inset', '', '\\end_layout']
|
|
|
|
document.body[endInset : endInset] = subst
|
|
|
|
addedLines += len(subst)
|
2008-06-02 19:42:53 +00:00
|
|
|
i += addedLines + 1
|
2008-03-02 11:30:50 +00:00
|
|
|
|
|
|
|
|
|
|
|
def revert_subfig(document):
|
|
|
|
" Revert subfloats. "
|
|
|
|
i = 0
|
|
|
|
while 1:
|
2008-05-05 06:37:25 +00:00
|
|
|
# whitespace intended (exclude \\begin_inset FloatList)
|
2008-06-05 16:05:52 +00:00
|
|
|
i = find_tokens(document.body, ['\\begin_inset Float ', '\\begin_inset Wrap'], i)
|
2008-03-02 11:30:50 +00:00
|
|
|
if i == -1:
|
|
|
|
return
|
2008-06-05 05:00:53 +00:00
|
|
|
j = 0
|
|
|
|
addedLines = 0
|
|
|
|
while j != -1:
|
2008-03-02 11:30:50 +00:00
|
|
|
j = find_end_of_inset(document.body, i)
|
|
|
|
if j == -1:
|
2008-06-05 05:00:53 +00:00
|
|
|
document.warning("Malformed lyx document: Missing '\\end_inset' (float) at line " + str(i + len(document.header)) + ".\n\t" + document.body[i])
|
|
|
|
# document.warning(document.body[i-1] + "\n" + document.body[i+1])
|
|
|
|
i += 1
|
|
|
|
continue # this will get us back to the outer loop, since j == -1
|
2008-03-02 11:30:50 +00:00
|
|
|
# look for embedded float (= subfloat)
|
2008-05-05 06:37:25 +00:00
|
|
|
# whitespace intended (exclude \\begin_inset FloatList)
|
|
|
|
k = find_token(document.body, '\\begin_inset Float ', i + 1, j)
|
2008-03-02 11:30:50 +00:00
|
|
|
if k == -1:
|
|
|
|
break
|
2008-10-13 07:51:02 +00:00
|
|
|
# is the subfloat aligned?
|
2008-11-12 01:17:47 +00:00
|
|
|
al = find_token(document.body, '\\align ', k - 1, j)
|
2008-10-13 07:51:02 +00:00
|
|
|
alignment_beg = ""
|
|
|
|
alignment_end = ""
|
|
|
|
if al != -1:
|
|
|
|
if get_value(document.body, '\\align', al) == "center":
|
|
|
|
alignment_beg = "\\backslash\nbegin{centering}"
|
|
|
|
alignment_end = "\\backslash\npar\\backslash\nend{centering}"
|
|
|
|
elif get_value(document.body, '\\align', al) == "left":
|
|
|
|
alignment_beg = "\\backslash\nbegin{raggedright}"
|
|
|
|
alignment_end = "\\backslash\npar\\backslash\nend{raggedright}"
|
|
|
|
elif get_value(document.body, '\\align', al) == "right":
|
|
|
|
alignment_beg = "\\backslash\nbegin{raggedleft}"
|
|
|
|
alignment_end = "\\backslash\npar\\backslash\nend{raggedleft}"
|
2008-03-02 11:30:50 +00:00
|
|
|
l = find_end_of_inset(document.body, k)
|
|
|
|
if l == -1:
|
|
|
|
document.warning("Malformed lyx document: Missing '\\end_inset' (embedded float).")
|
2008-06-05 05:00:53 +00:00
|
|
|
i += 1
|
2009-04-27 09:51:38 +00:00
|
|
|
j = -1
|
2008-06-05 05:00:53 +00:00
|
|
|
continue # escape to the outer loop
|
2008-04-19 21:29:44 +00:00
|
|
|
m = find_default_layout(document, k + 1, l)
|
2008-03-02 11:30:50 +00:00
|
|
|
# caption?
|
|
|
|
cap = find_token(document.body, '\\begin_inset Caption', k + 1, l)
|
|
|
|
caption = ''
|
|
|
|
shortcap = ''
|
2008-06-05 05:00:53 +00:00
|
|
|
capend = cap
|
2008-03-02 11:30:50 +00:00
|
|
|
if cap != -1:
|
|
|
|
capend = find_end_of_inset(document.body, cap)
|
|
|
|
if capend == -1:
|
|
|
|
document.warning("Malformed lyx document: Missing '\\end_inset' (caption).")
|
|
|
|
return
|
|
|
|
# label?
|
|
|
|
label = ''
|
|
|
|
lbl = find_token(document.body, '\\begin_inset CommandInset label', cap, capend)
|
|
|
|
if lbl != -1:
|
|
|
|
lblend = find_end_of_inset(document.body, lbl + 1)
|
|
|
|
if lblend == -1:
|
|
|
|
document.warning("Malformed lyx document: Missing '\\end_inset' (label).")
|
|
|
|
return
|
|
|
|
for line in document.body[lbl:lblend + 1]:
|
|
|
|
if line.startswith('name '):
|
|
|
|
label = line.split()[1].strip('"')
|
|
|
|
break
|
|
|
|
else:
|
|
|
|
lbl = capend
|
|
|
|
lblend = capend
|
|
|
|
label = ''
|
|
|
|
# opt arg?
|
|
|
|
opt = find_token(document.body, '\\begin_inset OptArg', cap, capend)
|
|
|
|
if opt != -1:
|
|
|
|
optend = find_end_of_inset(document.body, opt)
|
|
|
|
if optend == -1:
|
|
|
|
document.warning("Malformed lyx document: Missing '\\end_inset' (OptArg).")
|
|
|
|
return
|
2008-04-19 21:29:44 +00:00
|
|
|
optc = find_default_layout(document, opt, optend)
|
2008-03-02 11:30:50 +00:00
|
|
|
if optc == -1:
|
2008-04-15 17:59:01 +00:00
|
|
|
document.warning("Malformed LyX document: Missing `\\begin_layout' in Float inset.")
|
2008-03-02 11:30:50 +00:00
|
|
|
return
|
|
|
|
optcend = find_end_of(document.body, optc, "\\begin_layout", "\\end_layout")
|
|
|
|
for line in document.body[optc:optcend]:
|
|
|
|
if not line.startswith('\\'):
|
|
|
|
shortcap += line.strip()
|
|
|
|
else:
|
|
|
|
opt = capend
|
|
|
|
optend = capend
|
|
|
|
for line in document.body[cap:capend]:
|
|
|
|
if line in document.body[lbl:lblend]:
|
|
|
|
continue
|
|
|
|
elif line in document.body[opt:optend]:
|
|
|
|
continue
|
2009-04-27 09:51:38 +00:00
|
|
|
else:
|
|
|
|
inert = True
|
|
|
|
caption += lyxline2latex(document, line, inert)
|
2008-03-02 11:30:50 +00:00
|
|
|
if len(label) > 0:
|
2009-04-27 09:51:38 +00:00
|
|
|
caption += "\n\\backslash\nlabel{" + label + "}"
|
|
|
|
subst = '\\begin_layout PlainLayout\n\\begin_inset ERT\nstatus collapsed\n\n' \
|
|
|
|
'\\begin_layout PlainLayout\n\n}' + alignment_end + \
|
2008-10-13 07:51:02 +00:00
|
|
|
'\n\\end_layout\n\n\\end_inset\n\n' \
|
2009-04-27 09:51:38 +00:00
|
|
|
'\\end_layout\n\n\\begin_layout PlainLayout\n'
|
2008-06-05 05:00:53 +00:00
|
|
|
subst = subst.split('\n')
|
|
|
|
document.body[l : l+1] = subst
|
|
|
|
addedLines = len(subst) - 1
|
|
|
|
# this is before l and so is unchanged by the multiline insertion
|
|
|
|
if cap != capend:
|
|
|
|
del document.body[cap:capend+1]
|
|
|
|
addedLines -= (capend + 1 - cap)
|
2008-03-02 11:30:50 +00:00
|
|
|
del document.body[k+1:m-1]
|
2008-06-05 05:00:53 +00:00
|
|
|
addedLines -= (m - 1 - (k + 1))
|
2008-03-02 11:30:50 +00:00
|
|
|
insertion = '\\begin_inset ERT\nstatus collapsed\n\n' \
|
2009-04-27 09:51:38 +00:00
|
|
|
'\\begin_layout PlainLayout\n\n' + alignment_beg + '\n\\backslash\n' \
|
2008-06-05 05:00:53 +00:00
|
|
|
'subfloat'
|
2008-03-02 11:30:50 +00:00
|
|
|
if len(shortcap) > 0:
|
|
|
|
insertion = insertion + "[" + shortcap + "]"
|
|
|
|
if len(caption) > 0:
|
|
|
|
insertion = insertion + "[" + caption + "]"
|
|
|
|
insertion = insertion + '{%\n\\end_layout\n\n\\end_inset\n\n\\end_layout\n'
|
2008-06-05 05:00:53 +00:00
|
|
|
insertion = insertion.split('\n')
|
|
|
|
document.body[k : k + 1] = insertion
|
|
|
|
addedLines += len(insertion) - 1
|
2009-04-27 09:51:38 +00:00
|
|
|
al = find_token(document.body, '\\align ', k - 1, j + addedLines)
|
2008-10-13 07:51:02 +00:00
|
|
|
if al != -1:
|
|
|
|
del document.body[al]
|
|
|
|
addedLines -= 1
|
2008-07-07 18:22:59 +00:00
|
|
|
add_to_preamble(document, ['\\usepackage{subfig}\n'])
|
2008-06-05 05:00:53 +00:00
|
|
|
i += addedLines + 1
|
2008-03-02 11:30:50 +00:00
|
|
|
|
|
|
|
|
2008-03-02 17:12:22 +00:00
|
|
|
def revert_wrapplacement(document):
|
*** File Format Change: UI and enhancement for InsetSpace ***
* src/insets/InsetSpace.{cpp,h}:
- merge in HFill inset, add support for dotfill, hrulefill,
hspace and hspace*
* src/insets/InsetHFill.{cpp,h}:
- remove
* src/frontends/qt4/GuiHSpace.{cpp,h}:
* src/frontends/qt4/ui/HSpaceUi.ui:
* src/frontends/qt4/GuiView:
- new GUI for Space insets.
* src/insets/Inset.{cpp,h}:
* src/insets/InsetCode.h:
* src/insets/InsetCollapsable.cpp:
* src/insets/InsetCommandParams.cpp:
- remove HFILL_CODE and LFUN_HFILL_INSERT, add SPACE_CODE where necessary,
new Inset member isStretchableSpace() to indicate HFill and friends.
* Buffer.cpp:
- increase format to 319
* lib/lyx2lyx/LyX.py:
* lib/lyx2lyx/lyx_1_6.py:
- conversion/reversion routines
* development/FORMAT:
- document file format change
* src/Makefile.am:
* src/frontends/qt4/Makefile.am:
* development/scons/scons_manifest.py:
- deal with UI changes.
* src/LyXAction.cpp:
- remove LFUN_HFILL_INSERT
* src/LyXFunc.cpp:
- handle space dialog.
* src/factory.cpp:
* src/Paragraph.cpp (isHFill):
* src/Text.cpp:
* src/Text3.cpp:
* src/TextMetrics.cpp:
- adapt to changes
* lib/ui/classic.ui:
* lib/ui/stdmenus.ui:
- add HSpace dialog, remove HFill.
git-svn-id: svn://svn.lyx.org/lyx/lyx-devel/trunk@23787 a592a061-630c-0410-9148-cb99ea01b6c8
2008-03-17 09:23:43 +00:00
|
|
|
" Revert placement options wrap floats (wrapfig). "
|
2008-03-02 17:12:22 +00:00
|
|
|
i = 0
|
|
|
|
while True:
|
2008-07-07 18:09:29 +00:00
|
|
|
i = find_token(document.body, "\\begin_inset Wrap figure", i)
|
2008-03-02 17:12:22 +00:00
|
|
|
if i == -1:
|
|
|
|
return
|
2008-07-07 18:09:29 +00:00
|
|
|
e = find_end_of_inset(document.body, i)
|
|
|
|
j = find_token(document.body, "placement", i + 1, e)
|
|
|
|
if j == -1:
|
2008-03-02 17:12:22 +00:00
|
|
|
document.warning("Malformed LyX document: Couldn't find placement parameter of wrap float.")
|
2008-07-07 18:09:29 +00:00
|
|
|
i += 1
|
|
|
|
continue
|
2008-12-15 09:46:48 +00:00
|
|
|
r = re.compile("placement (o|i|l|r|O|I|L|R)")
|
2008-07-07 18:09:29 +00:00
|
|
|
m = r.match(document.body[j])
|
|
|
|
if m == None:
|
|
|
|
document.warning("Malformed LyX document: Placement option isn't O|I|R|L!")
|
2008-12-15 09:46:48 +00:00
|
|
|
else:
|
|
|
|
document.body[j] = "placement " + m.group(1).lower()
|
2008-07-07 18:09:29 +00:00
|
|
|
i = j
|
2008-03-02 17:12:22 +00:00
|
|
|
|
|
|
|
|
2008-03-09 23:04:14 +00:00
|
|
|
def remove_extra_embedded_files(document):
|
*** File Format Change: UI and enhancement for InsetSpace ***
* src/insets/InsetSpace.{cpp,h}:
- merge in HFill inset, add support for dotfill, hrulefill,
hspace and hspace*
* src/insets/InsetHFill.{cpp,h}:
- remove
* src/frontends/qt4/GuiHSpace.{cpp,h}:
* src/frontends/qt4/ui/HSpaceUi.ui:
* src/frontends/qt4/GuiView:
- new GUI for Space insets.
* src/insets/Inset.{cpp,h}:
* src/insets/InsetCode.h:
* src/insets/InsetCollapsable.cpp:
* src/insets/InsetCommandParams.cpp:
- remove HFILL_CODE and LFUN_HFILL_INSERT, add SPACE_CODE where necessary,
new Inset member isStretchableSpace() to indicate HFill and friends.
* Buffer.cpp:
- increase format to 319
* lib/lyx2lyx/LyX.py:
* lib/lyx2lyx/lyx_1_6.py:
- conversion/reversion routines
* development/FORMAT:
- document file format change
* src/Makefile.am:
* src/frontends/qt4/Makefile.am:
* development/scons/scons_manifest.py:
- deal with UI changes.
* src/LyXAction.cpp:
- remove LFUN_HFILL_INSERT
* src/LyXFunc.cpp:
- handle space dialog.
* src/factory.cpp:
* src/Paragraph.cpp (isHFill):
* src/Text.cpp:
* src/Text3.cpp:
* src/TextMetrics.cpp:
- adapt to changes
* lib/ui/classic.ui:
* lib/ui/stdmenus.ui:
- add HSpace dialog, remove HFill.
git-svn-id: svn://svn.lyx.org/lyx/lyx-devel/trunk@23787 a592a061-630c-0410-9148-cb99ea01b6c8
2008-03-17 09:23:43 +00:00
|
|
|
" Remove \extra_embedded_files from buffer params "
|
2008-03-09 23:04:14 +00:00
|
|
|
i = find_token(document.header, '\\extra_embedded_files', 0)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
document.header.pop(i)
|
|
|
|
|
*** File Format Change: UI and enhancement for InsetSpace ***
* src/insets/InsetSpace.{cpp,h}:
- merge in HFill inset, add support for dotfill, hrulefill,
hspace and hspace*
* src/insets/InsetHFill.{cpp,h}:
- remove
* src/frontends/qt4/GuiHSpace.{cpp,h}:
* src/frontends/qt4/ui/HSpaceUi.ui:
* src/frontends/qt4/GuiView:
- new GUI for Space insets.
* src/insets/Inset.{cpp,h}:
* src/insets/InsetCode.h:
* src/insets/InsetCollapsable.cpp:
* src/insets/InsetCommandParams.cpp:
- remove HFILL_CODE and LFUN_HFILL_INSERT, add SPACE_CODE where necessary,
new Inset member isStretchableSpace() to indicate HFill and friends.
* Buffer.cpp:
- increase format to 319
* lib/lyx2lyx/LyX.py:
* lib/lyx2lyx/lyx_1_6.py:
- conversion/reversion routines
* development/FORMAT:
- document file format change
* src/Makefile.am:
* src/frontends/qt4/Makefile.am:
* development/scons/scons_manifest.py:
- deal with UI changes.
* src/LyXAction.cpp:
- remove LFUN_HFILL_INSERT
* src/LyXFunc.cpp:
- handle space dialog.
* src/factory.cpp:
* src/Paragraph.cpp (isHFill):
* src/Text.cpp:
* src/Text3.cpp:
* src/TextMetrics.cpp:
- adapt to changes
* lib/ui/classic.ui:
* lib/ui/stdmenus.ui:
- add HSpace dialog, remove HFill.
git-svn-id: svn://svn.lyx.org/lyx/lyx-devel/trunk@23787 a592a061-630c-0410-9148-cb99ea01b6c8
2008-03-17 09:23:43 +00:00
|
|
|
|
|
|
|
def convert_spaceinset(document):
|
2008-06-01 14:22:11 +00:00
|
|
|
" Convert '\\InsetSpace foo' to '\\begin_inset Space foo\n\\end_inset' "
|
|
|
|
i = 0
|
|
|
|
while i < len(document.body):
|
|
|
|
m = re.match(r'(.*)\\InsetSpace (.*)', document.body[i])
|
|
|
|
if m:
|
|
|
|
before = m.group(1)
|
|
|
|
after = m.group(2)
|
|
|
|
subst = [before, "\\begin_inset Space " + after, "\\end_inset"]
|
|
|
|
document.body[i: i+1] = subst
|
2008-06-01 14:27:43 +00:00
|
|
|
i = i + len(subst)
|
2008-06-01 14:22:11 +00:00
|
|
|
else:
|
|
|
|
i = i + 1
|
*** File Format Change: UI and enhancement for InsetSpace ***
* src/insets/InsetSpace.{cpp,h}:
- merge in HFill inset, add support for dotfill, hrulefill,
hspace and hspace*
* src/insets/InsetHFill.{cpp,h}:
- remove
* src/frontends/qt4/GuiHSpace.{cpp,h}:
* src/frontends/qt4/ui/HSpaceUi.ui:
* src/frontends/qt4/GuiView:
- new GUI for Space insets.
* src/insets/Inset.{cpp,h}:
* src/insets/InsetCode.h:
* src/insets/InsetCollapsable.cpp:
* src/insets/InsetCommandParams.cpp:
- remove HFILL_CODE and LFUN_HFILL_INSERT, add SPACE_CODE where necessary,
new Inset member isStretchableSpace() to indicate HFill and friends.
* Buffer.cpp:
- increase format to 319
* lib/lyx2lyx/LyX.py:
* lib/lyx2lyx/lyx_1_6.py:
- conversion/reversion routines
* development/FORMAT:
- document file format change
* src/Makefile.am:
* src/frontends/qt4/Makefile.am:
* development/scons/scons_manifest.py:
- deal with UI changes.
* src/LyXAction.cpp:
- remove LFUN_HFILL_INSERT
* src/LyXFunc.cpp:
- handle space dialog.
* src/factory.cpp:
* src/Paragraph.cpp (isHFill):
* src/Text.cpp:
* src/Text3.cpp:
* src/TextMetrics.cpp:
- adapt to changes
* lib/ui/classic.ui:
* lib/ui/stdmenus.ui:
- add HSpace dialog, remove HFill.
git-svn-id: svn://svn.lyx.org/lyx/lyx-devel/trunk@23787 a592a061-630c-0410-9148-cb99ea01b6c8
2008-03-17 09:23:43 +00:00
|
|
|
|
|
|
|
|
|
|
|
def revert_spaceinset(document):
|
|
|
|
" Revert '\\begin_inset Space foo\n\\end_inset' to '\\InsetSpace foo' "
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, "\\begin_inset Space", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
j = find_end_of_inset(document.body, i)
|
|
|
|
if j == -1:
|
|
|
|
document.warning("Malformed LyX document: Could not find end of space inset.")
|
|
|
|
continue
|
|
|
|
document.body[i] = document.body[i].replace('\\begin_inset Space', '\\InsetSpace')
|
|
|
|
del document.body[j]
|
|
|
|
|
|
|
|
|
|
|
|
def convert_hfill(document):
|
|
|
|
" Convert hfill to space inset "
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, "\\hfill", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
2008-06-02 19:42:53 +00:00
|
|
|
subst = document.body[i].replace('\\hfill', \
|
|
|
|
'\n\\begin_inset Space \\hfill{}\n\\end_inset')
|
|
|
|
subst = subst.split('\n')
|
|
|
|
document.body[i : i+1] = subst
|
|
|
|
i += len(subst)
|
*** File Format Change: UI and enhancement for InsetSpace ***
* src/insets/InsetSpace.{cpp,h}:
- merge in HFill inset, add support for dotfill, hrulefill,
hspace and hspace*
* src/insets/InsetHFill.{cpp,h}:
- remove
* src/frontends/qt4/GuiHSpace.{cpp,h}:
* src/frontends/qt4/ui/HSpaceUi.ui:
* src/frontends/qt4/GuiView:
- new GUI for Space insets.
* src/insets/Inset.{cpp,h}:
* src/insets/InsetCode.h:
* src/insets/InsetCollapsable.cpp:
* src/insets/InsetCommandParams.cpp:
- remove HFILL_CODE and LFUN_HFILL_INSERT, add SPACE_CODE where necessary,
new Inset member isStretchableSpace() to indicate HFill and friends.
* Buffer.cpp:
- increase format to 319
* lib/lyx2lyx/LyX.py:
* lib/lyx2lyx/lyx_1_6.py:
- conversion/reversion routines
* development/FORMAT:
- document file format change
* src/Makefile.am:
* src/frontends/qt4/Makefile.am:
* development/scons/scons_manifest.py:
- deal with UI changes.
* src/LyXAction.cpp:
- remove LFUN_HFILL_INSERT
* src/LyXFunc.cpp:
- handle space dialog.
* src/factory.cpp:
* src/Paragraph.cpp (isHFill):
* src/Text.cpp:
* src/Text3.cpp:
* src/TextMetrics.cpp:
- adapt to changes
* lib/ui/classic.ui:
* lib/ui/stdmenus.ui:
- add HSpace dialog, remove HFill.
git-svn-id: svn://svn.lyx.org/lyx/lyx-devel/trunk@23787 a592a061-630c-0410-9148-cb99ea01b6c8
2008-03-17 09:23:43 +00:00
|
|
|
|
|
|
|
|
|
|
|
def revert_hfills(document):
|
|
|
|
' Revert \\hfill commands '
|
2008-06-02 19:42:53 +00:00
|
|
|
hfill = re.compile(r'\\hfill')
|
|
|
|
dotfill = re.compile(r'\\dotfill')
|
|
|
|
hrulefill = re.compile(r'\\hrulefill')
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, "\\InsetSpace", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
if hfill.search(document.body[i]):
|
|
|
|
document.body[i] = \
|
|
|
|
document.body[i].replace('\\InsetSpace \\hfill{}', '\\hfill')
|
|
|
|
i += 1
|
|
|
|
continue
|
|
|
|
if dotfill.search(document.body[i]):
|
|
|
|
subst = document.body[i].replace('\\InsetSpace \\dotfill{}', \
|
|
|
|
'\\begin_inset ERT\nstatus collapsed\n\n' \
|
|
|
|
'\\begin_layout Standard\n\n\n\\backslash\n' \
|
|
|
|
'dotfill{}\n\\end_layout\n\n\\end_inset\n\n')
|
|
|
|
subst = subst.split('\n')
|
|
|
|
document.body[i : i+1] = subst
|
|
|
|
i += len(subst)
|
|
|
|
continue
|
|
|
|
if hrulefill.search(document.body[i]):
|
|
|
|
subst = document.body[i].replace('\\InsetSpace \\hrulefill{}', \
|
|
|
|
'\\begin_inset ERT\nstatus collapsed\n\n' \
|
|
|
|
'\\begin_layout Standard\n\n\n\\backslash\n' \
|
|
|
|
'hrulefill{}\n\\end_layout\n\n\\end_inset\n\n')
|
|
|
|
subst = subst.split('\n')
|
|
|
|
document.body[i : i+1] = subst
|
|
|
|
i += len(subst)
|
|
|
|
continue
|
|
|
|
i += 1
|
*** File Format Change: UI and enhancement for InsetSpace ***
* src/insets/InsetSpace.{cpp,h}:
- merge in HFill inset, add support for dotfill, hrulefill,
hspace and hspace*
* src/insets/InsetHFill.{cpp,h}:
- remove
* src/frontends/qt4/GuiHSpace.{cpp,h}:
* src/frontends/qt4/ui/HSpaceUi.ui:
* src/frontends/qt4/GuiView:
- new GUI for Space insets.
* src/insets/Inset.{cpp,h}:
* src/insets/InsetCode.h:
* src/insets/InsetCollapsable.cpp:
* src/insets/InsetCommandParams.cpp:
- remove HFILL_CODE and LFUN_HFILL_INSERT, add SPACE_CODE where necessary,
new Inset member isStretchableSpace() to indicate HFill and friends.
* Buffer.cpp:
- increase format to 319
* lib/lyx2lyx/LyX.py:
* lib/lyx2lyx/lyx_1_6.py:
- conversion/reversion routines
* development/FORMAT:
- document file format change
* src/Makefile.am:
* src/frontends/qt4/Makefile.am:
* development/scons/scons_manifest.py:
- deal with UI changes.
* src/LyXAction.cpp:
- remove LFUN_HFILL_INSERT
* src/LyXFunc.cpp:
- handle space dialog.
* src/factory.cpp:
* src/Paragraph.cpp (isHFill):
* src/Text.cpp:
* src/Text3.cpp:
* src/TextMetrics.cpp:
- adapt to changes
* lib/ui/classic.ui:
* lib/ui/stdmenus.ui:
- add HSpace dialog, remove HFill.
git-svn-id: svn://svn.lyx.org/lyx/lyx-devel/trunk@23787 a592a061-630c-0410-9148-cb99ea01b6c8
2008-03-17 09:23:43 +00:00
|
|
|
|
|
|
|
def revert_hspace(document):
|
|
|
|
' Revert \\InsetSpace \\hspace{} to ERT '
|
|
|
|
i = 0
|
2008-06-02 19:42:53 +00:00
|
|
|
hspace = re.compile(r'\\hspace{}')
|
|
|
|
hstar = re.compile(r'\\hspace\*{}')
|
*** File Format Change: UI and enhancement for InsetSpace ***
* src/insets/InsetSpace.{cpp,h}:
- merge in HFill inset, add support for dotfill, hrulefill,
hspace and hspace*
* src/insets/InsetHFill.{cpp,h}:
- remove
* src/frontends/qt4/GuiHSpace.{cpp,h}:
* src/frontends/qt4/ui/HSpaceUi.ui:
* src/frontends/qt4/GuiView:
- new GUI for Space insets.
* src/insets/Inset.{cpp,h}:
* src/insets/InsetCode.h:
* src/insets/InsetCollapsable.cpp:
* src/insets/InsetCommandParams.cpp:
- remove HFILL_CODE and LFUN_HFILL_INSERT, add SPACE_CODE where necessary,
new Inset member isStretchableSpace() to indicate HFill and friends.
* Buffer.cpp:
- increase format to 319
* lib/lyx2lyx/LyX.py:
* lib/lyx2lyx/lyx_1_6.py:
- conversion/reversion routines
* development/FORMAT:
- document file format change
* src/Makefile.am:
* src/frontends/qt4/Makefile.am:
* development/scons/scons_manifest.py:
- deal with UI changes.
* src/LyXAction.cpp:
- remove LFUN_HFILL_INSERT
* src/LyXFunc.cpp:
- handle space dialog.
* src/factory.cpp:
* src/Paragraph.cpp (isHFill):
* src/Text.cpp:
* src/Text3.cpp:
* src/TextMetrics.cpp:
- adapt to changes
* lib/ui/classic.ui:
* lib/ui/stdmenus.ui:
- add HSpace dialog, remove HFill.
git-svn-id: svn://svn.lyx.org/lyx/lyx-devel/trunk@23787 a592a061-630c-0410-9148-cb99ea01b6c8
2008-03-17 09:23:43 +00:00
|
|
|
while True:
|
|
|
|
i = find_token(document.body, "\\InsetSpace \\hspace", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
length = get_value(document.body, '\\length', i+1)
|
|
|
|
if length == '':
|
|
|
|
document.warning("Malformed lyx document: Missing '\\length' in Space inset.")
|
|
|
|
return
|
|
|
|
del document.body[i+1]
|
2008-06-02 19:42:53 +00:00
|
|
|
addedLines = -1
|
|
|
|
if hstar.search(document.body[i]):
|
|
|
|
subst = document.body[i].replace('\\InsetSpace \\hspace*{}', \
|
|
|
|
'\\begin_inset ERT\nstatus collapsed\n\n' \
|
|
|
|
'\\begin_layout Standard\n\n\n\\backslash\n' \
|
|
|
|
'hspace*{' + length + '}\n\\end_layout\n\n\\end_inset\n\n')
|
|
|
|
subst = subst.split('\n')
|
|
|
|
document.body[i : i+1] = subst
|
|
|
|
addedLines += len(subst) - 1
|
|
|
|
i += addedLines + 1
|
|
|
|
continue
|
|
|
|
if hspace.search(document.body[i]):
|
|
|
|
subst = document.body[i].replace('\\InsetSpace \\hspace{}', \
|
|
|
|
'\\begin_inset ERT\nstatus collapsed\n\n' \
|
|
|
|
'\\begin_layout Standard\n\n\n\\backslash\n' \
|
|
|
|
'hspace{' + length + '}\n\\end_layout\n\n\\end_inset\n\n')
|
|
|
|
subst = subst.split('\n')
|
|
|
|
document.body[i : i+1] = subst
|
|
|
|
addedLines += len(subst) - 1
|
|
|
|
i += addedLines + 1
|
|
|
|
continue
|
|
|
|
i += 1
|
*** File Format Change: UI and enhancement for InsetSpace ***
* src/insets/InsetSpace.{cpp,h}:
- merge in HFill inset, add support for dotfill, hrulefill,
hspace and hspace*
* src/insets/InsetHFill.{cpp,h}:
- remove
* src/frontends/qt4/GuiHSpace.{cpp,h}:
* src/frontends/qt4/ui/HSpaceUi.ui:
* src/frontends/qt4/GuiView:
- new GUI for Space insets.
* src/insets/Inset.{cpp,h}:
* src/insets/InsetCode.h:
* src/insets/InsetCollapsable.cpp:
* src/insets/InsetCommandParams.cpp:
- remove HFILL_CODE and LFUN_HFILL_INSERT, add SPACE_CODE where necessary,
new Inset member isStretchableSpace() to indicate HFill and friends.
* Buffer.cpp:
- increase format to 319
* lib/lyx2lyx/LyX.py:
* lib/lyx2lyx/lyx_1_6.py:
- conversion/reversion routines
* development/FORMAT:
- document file format change
* src/Makefile.am:
* src/frontends/qt4/Makefile.am:
* development/scons/scons_manifest.py:
- deal with UI changes.
* src/LyXAction.cpp:
- remove LFUN_HFILL_INSERT
* src/LyXFunc.cpp:
- handle space dialog.
* src/factory.cpp:
* src/Paragraph.cpp (isHFill):
* src/Text.cpp:
* src/Text3.cpp:
* src/TextMetrics.cpp:
- adapt to changes
* lib/ui/classic.ui:
* lib/ui/stdmenus.ui:
- add HSpace dialog, remove HFill.
git-svn-id: svn://svn.lyx.org/lyx/lyx-devel/trunk@23787 a592a061-630c-0410-9148-cb99ea01b6c8
2008-03-17 09:23:43 +00:00
|
|
|
|
|
|
|
|
2008-03-18 12:11:28 +00:00
|
|
|
def revert_protected_hfill(document):
|
|
|
|
' Revert \\begin_inset Space \\hspace*{\\fill} to ERT '
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, '\\begin_inset Space \\hspace*{\\fill}', i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
j = find_end_of_inset(document.body, i)
|
|
|
|
if j == -1:
|
|
|
|
document.warning("Malformed LyX document: Could not find end of space inset.")
|
|
|
|
continue
|
|
|
|
del document.body[j]
|
2008-06-02 19:42:53 +00:00
|
|
|
subst = document.body[i].replace('\\begin_inset Space \\hspace*{\\fill}', \
|
|
|
|
'\\begin_inset ERT\nstatus collapsed\n\n' \
|
|
|
|
'\\begin_layout Standard\n\n\n\\backslash\n' \
|
|
|
|
'hspace*{\n\\backslash\nfill}\n\\end_layout\n\n\\end_inset\n\n')
|
|
|
|
subst = subst.split('\n')
|
|
|
|
document.body[i : i+1] = subst
|
|
|
|
i += len(subst)
|
2008-03-18 12:11:28 +00:00
|
|
|
|
|
|
|
|
2008-04-30 20:11:57 +00:00
|
|
|
def revert_leftarrowfill(document):
|
|
|
|
' Revert \\begin_inset Space \\leftarrowfill{} to ERT '
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, '\\begin_inset Space \\leftarrowfill{}', i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
j = find_end_of_inset(document.body, i)
|
|
|
|
if j == -1:
|
|
|
|
document.warning("Malformed LyX document: Could not find end of space inset.")
|
|
|
|
continue
|
|
|
|
del document.body[j]
|
2008-06-02 19:42:53 +00:00
|
|
|
subst = document.body[i].replace('\\begin_inset Space \\leftarrowfill{}', \
|
|
|
|
'\\begin_inset ERT\nstatus collapsed\n\n' \
|
|
|
|
'\\begin_layout Standard\n\n\n\\backslash\n' \
|
|
|
|
'leftarrowfill{}\n\\end_layout\n\n\\end_inset\n\n')
|
|
|
|
subst = subst.split('\n')
|
|
|
|
document.body[i : i+1] = subst
|
|
|
|
i += len(subst)
|
2008-04-30 20:11:57 +00:00
|
|
|
|
|
|
|
|
|
|
|
def revert_rightarrowfill(document):
|
|
|
|
' Revert \\begin_inset Space \\rightarrowfill{} to ERT '
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, '\\begin_inset Space \\rightarrowfill{}', i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
j = find_end_of_inset(document.body, i)
|
|
|
|
if j == -1:
|
|
|
|
document.warning("Malformed LyX document: Could not find end of space inset.")
|
|
|
|
continue
|
|
|
|
del document.body[j]
|
2008-06-02 19:42:53 +00:00
|
|
|
subst = document.body[i].replace('\\begin_inset Space \\rightarrowfill{}', \
|
|
|
|
'\\begin_inset ERT\nstatus collapsed\n\n' \
|
|
|
|
'\\begin_layout Standard\n\n\n\\backslash\n' \
|
|
|
|
'rightarrowfill{}\n\\end_layout\n\n\\end_inset\n\n')
|
|
|
|
subst = subst.split('\n')
|
|
|
|
document.body[i : i+1] = subst
|
|
|
|
i += len(subst)
|
2008-04-30 20:11:57 +00:00
|
|
|
|
|
|
|
|
|
|
|
def revert_upbracefill(document):
|
|
|
|
' Revert \\begin_inset Space \\upbracefill{} to ERT '
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, '\\begin_inset Space \\upbracefill{}', i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
j = find_end_of_inset(document.body, i)
|
|
|
|
if j == -1:
|
|
|
|
document.warning("Malformed LyX document: Could not find end of space inset.")
|
|
|
|
continue
|
|
|
|
del document.body[j]
|
2008-06-02 19:42:53 +00:00
|
|
|
subst = document.body[i].replace('\\begin_inset Space \\upbracefill{}', \
|
|
|
|
'\\begin_inset ERT\nstatus collapsed\n\n' \
|
|
|
|
'\\begin_layout Standard\n\n\n\\backslash\n' \
|
|
|
|
'upbracefill{}\n\\end_layout\n\n\\end_inset\n\n')
|
|
|
|
subst = subst.split('\n')
|
|
|
|
document.body[i : i+1] = subst
|
|
|
|
i += len(subst)
|
2008-04-30 20:11:57 +00:00
|
|
|
|
|
|
|
|
|
|
|
def revert_downbracefill(document):
|
|
|
|
' Revert \\begin_inset Space \\downbracefill{} to ERT '
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, '\\begin_inset Space \\downbracefill{}', i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
j = find_end_of_inset(document.body, i)
|
|
|
|
if j == -1:
|
|
|
|
document.warning("Malformed LyX document: Could not find end of space inset.")
|
|
|
|
continue
|
|
|
|
del document.body[j]
|
2008-06-02 19:42:53 +00:00
|
|
|
subst = document.body[i].replace('\\begin_inset Space \\downbracefill{}', \
|
|
|
|
'\\begin_inset ERT\nstatus collapsed\n\n' \
|
|
|
|
'\\begin_layout Standard\n\n\n\\backslash\n' \
|
|
|
|
'downbracefill{}\n\\end_layout\n\n\\end_inset\n\n')
|
|
|
|
subst = subst.split('\n')
|
|
|
|
document.body[i : i+1] = subst
|
|
|
|
i += len(subst)
|
2008-04-30 20:11:57 +00:00
|
|
|
|
|
|
|
|
2008-03-24 12:31:07 +00:00
|
|
|
def revert_local_layout(document):
|
|
|
|
' Revert local layout headers.'
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.header, "\\begin_local_layout", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
j = find_end_of(document.header, i, "\\begin_local_layout", "\\end_local_layout")
|
|
|
|
if j == -1:
|
|
|
|
# this should not happen
|
|
|
|
break
|
|
|
|
document.header[i : j + 1] = []
|
|
|
|
|
|
|
|
|
2008-03-25 09:26:03 +00:00
|
|
|
def convert_pagebreaks(document):
|
|
|
|
' Convert inline Newpage insets to new format '
|
2008-03-25 19:10:04 +00:00
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, '\\newpage', i)
|
|
|
|
if i == -1:
|
|
|
|
break
|
|
|
|
document.body[i:i+1] = ['\\begin_inset Newpage newpage',
|
2008-06-02 19:42:53 +00:00
|
|
|
'\\end_inset']
|
2008-03-25 19:10:04 +00:00
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, '\\pagebreak', i)
|
|
|
|
if i == -1:
|
|
|
|
break
|
|
|
|
document.body[i:i+1] = ['\\begin_inset Newpage pagebreak',
|
2008-06-02 19:42:53 +00:00
|
|
|
'\\end_inset']
|
2008-03-25 19:10:04 +00:00
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, '\\clearpage', i)
|
|
|
|
if i == -1:
|
|
|
|
break
|
|
|
|
document.body[i:i+1] = ['\\begin_inset Newpage clearpage',
|
2008-06-02 19:42:53 +00:00
|
|
|
'\\end_inset']
|
2008-03-25 19:10:04 +00:00
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, '\\cleardoublepage', i)
|
|
|
|
if i == -1:
|
|
|
|
break
|
|
|
|
document.body[i:i+1] = ['\\begin_inset Newpage cleardoublepage',
|
2008-06-02 19:42:53 +00:00
|
|
|
'\\end_inset']
|
2008-03-25 09:26:03 +00:00
|
|
|
|
|
|
|
|
|
|
|
def revert_pagebreaks(document):
|
|
|
|
' Revert \\begin_inset Newpage to previous inline format '
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, '\\begin_inset Newpage', i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
j = find_end_of_inset(document.body, i)
|
|
|
|
if j == -1:
|
|
|
|
document.warning("Malformed LyX document: Could not find end of Newpage inset.")
|
|
|
|
continue
|
|
|
|
del document.body[j]
|
|
|
|
document.body[i] = document.body[i].replace('\\begin_inset Newpage newpage', '\\newpage')
|
|
|
|
document.body[i] = document.body[i].replace('\\begin_inset Newpage pagebreak', '\\pagebreak')
|
|
|
|
document.body[i] = document.body[i].replace('\\begin_inset Newpage clearpage', '\\clearpage')
|
|
|
|
document.body[i] = document.body[i].replace('\\begin_inset Newpage cleardoublepage', '\\cleardoublepage')
|
|
|
|
|
2008-03-24 12:31:07 +00:00
|
|
|
|
2008-03-26 08:10:01 +00:00
|
|
|
def convert_linebreaks(document):
|
|
|
|
' Convert inline Newline insets to new format '
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, '\\newline', i)
|
|
|
|
if i == -1:
|
|
|
|
break
|
|
|
|
document.body[i:i+1] = ['\\begin_inset Newline newline',
|
2008-06-02 19:42:53 +00:00
|
|
|
'\\end_inset']
|
2008-03-26 08:10:01 +00:00
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, '\\linebreak', i)
|
|
|
|
if i == -1:
|
|
|
|
break
|
|
|
|
document.body[i:i+1] = ['\\begin_inset Newline linebreak',
|
2008-06-02 19:42:53 +00:00
|
|
|
'\\end_inset']
|
2008-03-26 08:10:01 +00:00
|
|
|
|
|
|
|
|
|
|
|
def revert_linebreaks(document):
|
|
|
|
' Revert \\begin_inset Newline to previous inline format '
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, '\\begin_inset Newline', i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
j = find_end_of_inset(document.body, i)
|
|
|
|
if j == -1:
|
|
|
|
document.warning("Malformed LyX document: Could not find end of Newline inset.")
|
|
|
|
continue
|
|
|
|
del document.body[j]
|
|
|
|
document.body[i] = document.body[i].replace('\\begin_inset Newline newline', '\\newline')
|
|
|
|
document.body[i] = document.body[i].replace('\\begin_inset Newline linebreak', '\\linebreak')
|
|
|
|
|
|
|
|
|
2008-03-29 23:25:40 +00:00
|
|
|
def convert_japanese_plain(document):
|
2008-04-11 14:07:25 +00:00
|
|
|
' Set language japanese-plain to japanese '
|
2008-03-29 23:25:40 +00:00
|
|
|
i = 0
|
|
|
|
if document.language == "japanese-plain":
|
|
|
|
document.language = "japanese"
|
|
|
|
i = find_token(document.header, "\\language", 0)
|
|
|
|
if i != -1:
|
|
|
|
document.header[i] = "\\language japanese"
|
|
|
|
j = 0
|
|
|
|
while True:
|
|
|
|
j = find_token(document.body, "\\lang japanese-plain", j)
|
|
|
|
if j == -1:
|
|
|
|
return
|
|
|
|
document.body[j] = document.body[j].replace("\\lang japanese-plain", "\\lang japanese")
|
|
|
|
j = j + 1
|
|
|
|
|
|
|
|
|
2008-04-11 14:07:25 +00:00
|
|
|
def revert_pdfpages(document):
|
|
|
|
' Revert pdfpages external inset to ERT '
|
|
|
|
i = 0
|
|
|
|
while 1:
|
|
|
|
i = find_token(document.body, "\\begin_inset External", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
j = find_end_of_inset(document.body, i)
|
|
|
|
if j == -1:
|
2008-06-05 05:00:53 +00:00
|
|
|
document.warning("Malformed lyx document: Missing '\\end_inset' in revert_pdfpages.")
|
2008-04-11 14:07:25 +00:00
|
|
|
i = i + 1
|
|
|
|
continue
|
|
|
|
if get_value(document.body, 'template', i, j) == "PDFPages":
|
|
|
|
filename = get_value(document.body, 'filename', i, j)
|
|
|
|
extra = ''
|
|
|
|
r = re.compile(r'\textra PDFLaTeX \"(.*)\"$')
|
|
|
|
for k in range(i, j):
|
|
|
|
m = r.match(document.body[k])
|
|
|
|
if m:
|
|
|
|
extra = m.group(1)
|
|
|
|
angle = get_value(document.body, 'rotateAngle', i, j)
|
|
|
|
width = get_value(document.body, 'width', i, j)
|
|
|
|
height = get_value(document.body, 'height', i, j)
|
|
|
|
scale = get_value(document.body, 'scale', i, j)
|
|
|
|
keepAspectRatio = find_token(document.body, "\tkeepAspectRatio", i, j)
|
|
|
|
options = extra
|
|
|
|
if angle != '':
|
|
|
|
if options != '':
|
|
|
|
options += ",angle=" + angle
|
|
|
|
else:
|
|
|
|
options += "angle=" + angle
|
|
|
|
if width != '':
|
|
|
|
if options != '':
|
|
|
|
options += ",width=" + convert_len(width)
|
|
|
|
else:
|
|
|
|
options += "width=" + convert_len(width)
|
|
|
|
if height != '':
|
|
|
|
if options != '':
|
|
|
|
options += ",height=" + convert_len(height)
|
|
|
|
else:
|
|
|
|
options += "height=" + convert_len(height)
|
|
|
|
if scale != '':
|
|
|
|
if options != '':
|
|
|
|
options += ",scale=" + scale
|
|
|
|
else:
|
|
|
|
options += "scale=" + scale
|
|
|
|
if keepAspectRatio != '':
|
|
|
|
if options != '':
|
|
|
|
options += ",keepaspectratio"
|
|
|
|
else:
|
|
|
|
options += "keepaspectratio"
|
|
|
|
if options != '':
|
|
|
|
options = '[' + options + ']'
|
|
|
|
del document.body[i+1:j+1]
|
|
|
|
document.body[i:i+1] = ['\\begin_inset ERT',
|
|
|
|
'status collapsed',
|
|
|
|
'',
|
|
|
|
'\\begin_layout Standard',
|
|
|
|
'',
|
|
|
|
'\\backslash',
|
|
|
|
'includepdf' + options + '{' + filename + '}',
|
|
|
|
'\\end_layout',
|
|
|
|
'',
|
|
|
|
'\\end_inset']
|
|
|
|
add_to_preamble(document, ['\\usepackage{pdfpages}\n'])
|
|
|
|
i = i + 1
|
|
|
|
continue
|
|
|
|
i = i + 1
|
|
|
|
|
|
|
|
|
2008-04-16 00:25:08 +00:00
|
|
|
def revert_mexican(document):
|
2008-04-28 16:38:56 +00:00
|
|
|
' Set language Spanish(Mexico) to Spanish '
|
2008-04-16 00:25:08 +00:00
|
|
|
i = 0
|
|
|
|
if document.language == "spanish-mexico":
|
|
|
|
document.language = "spanish"
|
|
|
|
i = find_token(document.header, "\\language", 0)
|
|
|
|
if i != -1:
|
|
|
|
document.header[i] = "\\language spanish"
|
|
|
|
j = 0
|
|
|
|
while True:
|
|
|
|
j = find_token(document.body, "\\lang spanish-mexico", j)
|
|
|
|
if j == -1:
|
|
|
|
return
|
|
|
|
document.body[j] = document.body[j].replace("\\lang spanish-mexico", "\\lang spanish")
|
|
|
|
j = j + 1
|
|
|
|
|
|
|
|
|
2008-04-18 17:06:03 +00:00
|
|
|
def remove_embedding(document):
|
2008-04-28 16:38:56 +00:00
|
|
|
' Remove embed tag from all insets '
|
2008-04-18 17:06:03 +00:00
|
|
|
revert_inset_embedding(document, 'Graphics')
|
|
|
|
revert_inset_embedding(document, 'External')
|
|
|
|
revert_inset_embedding(document, 'CommandInset include')
|
|
|
|
revert_inset_embedding(document, 'CommandInset bibtex')
|
|
|
|
|
|
|
|
|
2008-04-28 16:38:56 +00:00
|
|
|
def revert_master(document):
|
|
|
|
' Remove master param '
|
|
|
|
i = find_token(document.header, "\\master", 0)
|
|
|
|
if i != -1:
|
|
|
|
del document.header[i]
|
|
|
|
|
|
|
|
|
2008-05-06 21:13:09 +00:00
|
|
|
def revert_graphics_group(document):
|
|
|
|
' Revert group information from graphics insets '
|
|
|
|
i = 0
|
|
|
|
while 1:
|
|
|
|
i = find_token(document.body, "\\begin_inset Graphics", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
j = find_end_of_inset(document.body, i)
|
|
|
|
if j == -1:
|
2008-06-05 05:00:53 +00:00
|
|
|
document.warning("Malformed lyx document: Missing '\\end_inset' in revert_graphics_group.")
|
2008-05-06 21:13:09 +00:00
|
|
|
i = i + 1
|
|
|
|
continue
|
|
|
|
k = find_token(document.body, " groupId", i, j)
|
|
|
|
if k == -1:
|
|
|
|
i = i + 1
|
|
|
|
continue
|
|
|
|
del document.body[k]
|
|
|
|
i = i + 1
|
|
|
|
|
|
|
|
|
2008-05-09 08:57:32 +00:00
|
|
|
def update_apa_styles(document):
|
2008-05-15 22:54:13 +00:00
|
|
|
' Replace obsolete styles '
|
2008-05-09 08:57:32 +00:00
|
|
|
|
|
|
|
if document.textclass != "apa":
|
|
|
|
return
|
|
|
|
|
|
|
|
obsoletedby = { "Acknowledgments": "Acknowledgements",
|
|
|
|
"Section*": "Section",
|
|
|
|
"Subsection*": "Subsection",
|
|
|
|
"Subsubsection*": "Subsubsection",
|
|
|
|
"Paragraph*": "Paragraph",
|
|
|
|
"Subparagraph*": "Subparagraph"}
|
|
|
|
i = 0
|
|
|
|
while 1:
|
|
|
|
i = find_token(document.body, "\\begin_layout", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
|
|
|
|
layout = document.body[i][14:]
|
|
|
|
if layout in obsoletedby:
|
|
|
|
document.body[i] = "\\begin_layout " + obsoletedby[layout]
|
|
|
|
|
|
|
|
i += 1
|
|
|
|
|
2008-05-15 22:54:13 +00:00
|
|
|
|
|
|
|
def convert_paper_sizes(document):
|
|
|
|
' exchange size options legalpaper and executivepaper to correct order '
|
|
|
|
# routine is needed to fix http://bugzilla.lyx.org/show_bug.cgi?id=4868
|
|
|
|
i = 0
|
|
|
|
j = 0
|
|
|
|
i = find_token(document.header, "\\papersize executivepaper", 0)
|
|
|
|
if i != -1:
|
|
|
|
document.header[i] = "\\papersize legalpaper"
|
|
|
|
return
|
|
|
|
j = find_token(document.header, "\\papersize legalpaper", 0)
|
|
|
|
if j != -1:
|
|
|
|
document.header[j] = "\\papersize executivepaper"
|
|
|
|
|
|
|
|
|
|
|
|
def revert_paper_sizes(document):
|
|
|
|
' exchange size options legalpaper and executivepaper to correct order '
|
|
|
|
i = 0
|
|
|
|
j = 0
|
|
|
|
i = find_token(document.header, "\\papersize executivepaper", 0)
|
|
|
|
if i != -1:
|
|
|
|
document.header[i] = "\\papersize legalpaper"
|
|
|
|
return
|
|
|
|
j = find_token(document.header, "\\papersize legalpaper", 0)
|
|
|
|
if j != -1:
|
|
|
|
document.header[j] = "\\papersize executivepaper"
|
|
|
|
|
|
|
|
|
2008-05-30 15:38:43 +00:00
|
|
|
def convert_InsetSpace(document):
|
|
|
|
" Convert '\\begin_inset Space foo' to '\\begin_inset space foo'"
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, "\\begin_inset Space", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
document.body[i] = document.body[i].replace('\\begin_inset Space', '\\begin_inset space')
|
|
|
|
|
|
|
|
|
|
|
|
def revert_InsetSpace(document):
|
|
|
|
" Revert '\\begin_inset space foo' to '\\begin_inset Space foo'"
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, "\\begin_inset space", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
document.body[i] = document.body[i].replace('\\begin_inset space', '\\begin_inset Space')
|
|
|
|
|
|
|
|
|
2008-06-13 07:34:55 +00:00
|
|
|
def convert_display_enum(document):
|
|
|
|
" Convert 'display foo' to 'display false/true'"
|
|
|
|
i = 0
|
|
|
|
while True:
|
2008-06-15 14:06:14 +00:00
|
|
|
i = find_token(document.body, "\tdisplay", i)
|
2008-06-13 07:34:55 +00:00
|
|
|
if i == -1:
|
|
|
|
return
|
2008-06-15 14:06:14 +00:00
|
|
|
val = get_value(document.body, 'display', i)
|
|
|
|
if val == "none":
|
2008-06-13 07:34:55 +00:00
|
|
|
document.body[i] = document.body[i].replace('none', 'false')
|
2008-06-15 14:06:14 +00:00
|
|
|
if val == "default":
|
2008-06-13 07:34:55 +00:00
|
|
|
document.body[i] = document.body[i].replace('default', 'true')
|
2008-06-15 14:06:14 +00:00
|
|
|
if val == "monochrome":
|
2008-06-13 07:34:55 +00:00
|
|
|
document.body[i] = document.body[i].replace('monochrome', 'true')
|
2008-06-15 14:06:14 +00:00
|
|
|
if val == "grayscale":
|
2008-06-13 07:34:55 +00:00
|
|
|
document.body[i] = document.body[i].replace('grayscale', 'true')
|
2008-06-15 14:06:14 +00:00
|
|
|
if val == "color":
|
2008-06-13 07:34:55 +00:00
|
|
|
document.body[i] = document.body[i].replace('color', 'true')
|
2008-06-15 14:06:14 +00:00
|
|
|
if val == "preview":
|
2008-06-13 07:34:55 +00:00
|
|
|
document.body[i] = document.body[i].replace('preview', 'true')
|
2008-06-15 14:06:14 +00:00
|
|
|
i += 1
|
2008-06-13 07:34:55 +00:00
|
|
|
|
|
|
|
|
|
|
|
def revert_display_enum(document):
|
|
|
|
" Revert 'display false/true' to 'display none/color'"
|
|
|
|
i = 0
|
|
|
|
while True:
|
2008-06-15 14:06:14 +00:00
|
|
|
i = find_token(document.body, "\tdisplay", i)
|
2008-06-13 07:34:55 +00:00
|
|
|
if i == -1:
|
|
|
|
return
|
2008-06-15 14:06:14 +00:00
|
|
|
val = get_value(document.body, 'display', i)
|
|
|
|
if val == "false":
|
2008-06-13 07:34:55 +00:00
|
|
|
document.body[i] = document.body[i].replace('false', 'none')
|
2008-06-15 14:06:14 +00:00
|
|
|
if val == "true":
|
2008-06-13 07:34:55 +00:00
|
|
|
document.body[i] = document.body[i].replace('true', 'default')
|
2008-06-15 14:06:14 +00:00
|
|
|
i += 1
|
2008-06-13 07:34:55 +00:00
|
|
|
|
|
|
|
|
2008-06-05 06:18:34 +00:00
|
|
|
def remove_fontsCJK(document):
|
|
|
|
' Remove font_cjk param '
|
|
|
|
i = find_token(document.header, "\\font_cjk", 0)
|
|
|
|
if i != -1:
|
|
|
|
del document.header[i]
|
|
|
|
|
|
|
|
|
2008-06-05 03:54:18 +00:00
|
|
|
def convert_plain_layout(document):
|
2008-08-01 17:57:01 +00:00
|
|
|
" Convert 'PlainLayout' to 'Plain Layout'"
|
2008-06-15 14:06:14 +00:00
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, '\\begin_layout PlainLayout', i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
document.body[i] = document.body[i].replace('\\begin_layout PlainLayout', \
|
|
|
|
'\\begin_layout Plain Layout')
|
|
|
|
i += 1
|
2008-06-05 03:54:18 +00:00
|
|
|
|
|
|
|
|
|
|
|
def revert_plain_layout(document):
|
2009-04-27 09:51:38 +00:00
|
|
|
" Revert 'Plain Layout' to 'PlainLayout'"
|
2008-06-15 14:06:14 +00:00
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, '\\begin_layout Plain Layout', i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
document.body[i] = document.body[i].replace('\\begin_layout Plain Layout', \
|
|
|
|
'\\begin_layout PlainLayout')
|
|
|
|
i += 1
|
|
|
|
|
2008-06-05 03:54:18 +00:00
|
|
|
|
|
|
|
def revert_plainlayout(document):
|
2009-04-27 09:51:38 +00:00
|
|
|
" Revert 'PlainLayout' to 'Standard'"
|
2008-06-15 14:06:14 +00:00
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, '\\begin_layout PlainLayout', i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
# This will be incorrect for some document classes, since Standard is not always
|
|
|
|
# the default. But (a) it is probably the best we can do and (b) it will actually
|
|
|
|
# work, in fact, since an unknown layout will be converted to default.
|
|
|
|
document.body[i] = document.body[i].replace('\\begin_layout PlainLayout', \
|
|
|
|
'\\begin_layout Standard')
|
|
|
|
i += 1
|
2008-06-05 03:54:18 +00:00
|
|
|
|
|
|
|
|
2008-06-22 23:21:46 +00:00
|
|
|
def revert_polytonicgreek(document):
|
|
|
|
"Set language polytonic Greek to Greek"
|
|
|
|
i = 0
|
|
|
|
if document.language == "polutonikogreek":
|
|
|
|
document.language = "greek"
|
|
|
|
i = find_token(document.header, "\\language", 0)
|
|
|
|
if i != -1:
|
|
|
|
document.header[i] = "\\language greek"
|
|
|
|
j = 0
|
|
|
|
while True:
|
|
|
|
j = find_token(document.body, "\\lang polutonikogreek", j)
|
|
|
|
if j == -1:
|
|
|
|
return
|
|
|
|
document.body[j] = document.body[j].replace("\\lang polutonikogreek", "\\lang greek")
|
|
|
|
j = j + 1
|
|
|
|
|
|
|
|
|
2008-07-25 20:43:32 +00:00
|
|
|
def revert_removed_modules(document):
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.header, "\\begin_remove_modules", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
j = find_end_of(document.header, i, "\\begin_remove_modules", "\\end_remove_modules")
|
|
|
|
if j == -1:
|
|
|
|
# this should not happen
|
|
|
|
break
|
|
|
|
document.header[i : j + 1] = []
|
|
|
|
|
|
|
|
|
2008-08-01 17:57:01 +00:00
|
|
|
def add_plain_layout(document):
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, "\\begin_layout", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
if len(document.body[i].split()) == 1:
|
|
|
|
document.body[i] = "\\begin_layout Plain Layout"
|
|
|
|
i += 1
|
|
|
|
|
2008-09-30 18:00:02 +00:00
|
|
|
|
|
|
|
def revert_tabulators(document):
|
|
|
|
"Revert tabulators to 4 spaces"
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, "\t", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
document.body[i] = document.body[i].replace("\t", " ")
|
|
|
|
i += 1
|
|
|
|
|
|
|
|
|
|
|
|
def revert_tabsize(document):
|
|
|
|
"Revert the tabsize parameter of listings"
|
|
|
|
i = 0
|
|
|
|
j = 0
|
|
|
|
while True:
|
|
|
|
# either it is the only parameter
|
|
|
|
i = find_token(document.body, 'lstparams "tabsize=4"', i)
|
|
|
|
if i != -1:
|
|
|
|
del document.body[i]
|
|
|
|
# or the last one
|
|
|
|
j = find_token(document.body, "lstparams", j)
|
|
|
|
if j == -1:
|
|
|
|
return
|
|
|
|
pos = document.body[j].find(",tabsize=")
|
|
|
|
document.body[j] = document.body[j][:pos] + '"'
|
|
|
|
i += 1
|
|
|
|
j += 1
|
|
|
|
|
|
|
|
|
2008-10-11 23:52:31 +00:00
|
|
|
def revert_mongolian(document):
|
|
|
|
"Set language Mongolian to English"
|
|
|
|
i = 0
|
|
|
|
if document.language == "mongolian":
|
|
|
|
document.language = "english"
|
|
|
|
i = find_token(document.header, "\\language", 0)
|
|
|
|
if i != -1:
|
|
|
|
document.header[i] = "\\language english"
|
|
|
|
j = 0
|
|
|
|
while True:
|
|
|
|
j = find_token(document.body, "\\lang mongolian", j)
|
|
|
|
if j == -1:
|
|
|
|
return
|
|
|
|
document.body[j] = document.body[j].replace("\\lang mongolian", "\\lang english")
|
|
|
|
j = j + 1
|
|
|
|
|
|
|
|
|
2008-10-12 09:36:00 +00:00
|
|
|
def revert_default_options(document):
|
|
|
|
' Remove param use_default_options '
|
|
|
|
i = find_token(document.header, "\\use_default_options", 0)
|
|
|
|
if i != -1:
|
|
|
|
del document.header[i]
|
|
|
|
|
|
|
|
|
|
|
|
def convert_default_options(document):
|
|
|
|
' Add param use_default_options and set it to false '
|
|
|
|
i = find_token(document.header, "\\textclass", 0)
|
|
|
|
if i == -1:
|
|
|
|
document.warning("Malformed LyX document: Missing `\\textclass'.")
|
|
|
|
return
|
|
|
|
document.header.insert(i, '\\use_default_options false')
|
|
|
|
|
|
|
|
|
2008-10-13 16:01:02 +00:00
|
|
|
def revert_backref_options(document):
|
2008-10-13 21:43:14 +00:00
|
|
|
' Revert option pdf_backref=page to pagebackref '
|
2008-10-13 16:01:02 +00:00
|
|
|
i = find_token(document.header, "\\pdf_backref page", 0)
|
|
|
|
if i != -1:
|
|
|
|
document.header[i] = "\\pdf_pagebackref true"
|
|
|
|
|
|
|
|
|
|
|
|
def convert_backref_options(document):
|
|
|
|
' We have changed the option pagebackref to backref=true '
|
|
|
|
i = find_token(document.header, "\\pdf_pagebackref true", 0)
|
|
|
|
if i != -1:
|
|
|
|
document.header[i] = "\\pdf_backref page"
|
|
|
|
j = find_token(document.header, "\\pdf_pagebackref false", 0)
|
|
|
|
if j != -1:
|
|
|
|
del document.header[j]
|
2008-10-13 21:43:14 +00:00
|
|
|
# backref=true was not a valid option, we meant backref=section
|
|
|
|
k = find_token(document.header, "\\pdf_backref true", 0)
|
2008-10-14 00:56:17 +00:00
|
|
|
if k != -1 and i != -1:
|
|
|
|
del document.header[k]
|
|
|
|
elif k != -1 and j != -1:
|
2008-10-13 21:43:14 +00:00
|
|
|
document.header[k] = "\\pdf_backref section"
|
2008-10-13 16:01:02 +00:00
|
|
|
|
2008-11-07 19:17:35 +00:00
|
|
|
|
|
|
|
def convert_charstyle_element(document):
|
|
|
|
"Convert CharStyle to Element for docbook backend"
|
|
|
|
if document.backend != "docbook":
|
|
|
|
return
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, "\\begin_inset Flex CharStyle:", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
document.body[i] = document.body[i].replace('\\begin_inset Flex CharStyle:',
|
|
|
|
'\\begin_inset Flex Element:')
|
|
|
|
|
|
|
|
def revert_charstyle_element(document):
|
|
|
|
"Convert Element to CharStyle for docbook backend"
|
|
|
|
if document.backend != "docbook":
|
|
|
|
return
|
|
|
|
i = 0
|
|
|
|
while True:
|
|
|
|
i = find_token(document.body, "\\begin_inset Flex Element:", i)
|
|
|
|
if i == -1:
|
|
|
|
return
|
|
|
|
document.body[i] = document.body[i].replace('\\begin_inset Flex Element:',
|
|
|
|
'\\begin_inset Flex CharStyle:')
|
|
|
|
|
2007-08-12 13:25:36 +00:00
|
|
|
##
|
|
|
|
# Conversion hub
|
|
|
|
#
|
|
|
|
|
|
|
|
supported_versions = ["1.6.0","1.6"]
|
2007-10-03 13:38:19 +00:00
|
|
|
convert = [[277, [fix_wrong_tables]],
|
2007-08-12 14:56:49 +00:00
|
|
|
[278, [close_begin_deeper]],
|
2007-08-18 23:26:07 +00:00
|
|
|
[279, [long_charstyle_names]],
|
This is one of a series of patches that will merge the layout modules development in personal/branches/rgheck back into the tree.
Design goal: Allow the use of layout "modules", which are to LaTeX packages as layout files are to LaTeX document classes. Thus, one could have a module that defined certain character styles, environments, commands, or what have you, and include it in various documents, each of which uses a different document class, without having to modify the layout files themselves. For example, a theorems.module could be used with article.layout to provide support for theorem-type environments, without having to modify article.layout itself, and the same module could be used with book.layout, etc.
This patch adds the backend. The ModuleList class holds a list of the available modules, which are retrieved from lyxmodules.lst, itself generated by configure.py. There are two LFUNs available: modules-clear and module-add, which do the obvious thing; you can test by typing these into the minibuffer, along with the name of one of the available modules: URL (a CharStyle), Endnote (a Custom Inset), and---with the spaces---End To Foot (View>LaTeX and look at the user preamble), which are themselves in lib/layouts. There are some others, too, that allow theorems to be added to classes like article and book.
The GUI will come next.
Issues: (i) The configure.py script could be improved. It'd be nice, for example, if it tested for the presence of the LaTeX packages a particular module needs. But this would mean re-working the LaTeX script, and I don't know how to do that. Note that at present, the packages are ignored. This will change shortly. (ii) I've used std::string in LyXModule, following what seemed to be a precedent in TextClass. If some of these should be docstrings, please let me know, and I'll change them. (iii) There is at present no distinction between LaTeX and DocBook modules. Should there be? That is: Should there be modules that are available when the document class is a LaTeX class and others that are available only when it is DocBook? Or should there just be one set of modules? Each module can of course indicate for what it is suitable in its description.
git-svn-id: svn://svn.lyx.org/lyx/lyx-devel/trunk@19893 a592a061-630c-0410-9148-cb99ea01b6c8
2007-08-29 17:59:49 +00:00
|
|
|
[280, [axe_show_label]],
|
2007-08-31 09:46:14 +00:00
|
|
|
[281, []],
|
2007-09-08 13:01:29 +00:00
|
|
|
[282, []],
|
2007-09-09 23:47:22 +00:00
|
|
|
[283, [convert_flex]],
|
2007-09-11 14:23:12 +00:00
|
|
|
[284, []],
|
2007-10-21 18:06:00 +00:00
|
|
|
[285, []],
|
2007-09-24 13:43:58 +00:00
|
|
|
[286, []],
|
2007-09-27 18:24:18 +00:00
|
|
|
[287, [convert_wrapfig_options]],
|
2007-10-03 13:38:19 +00:00
|
|
|
[288, [convert_inset_command]],
|
2007-10-03 22:00:45 +00:00
|
|
|
[289, [convert_latexcommand_index]],
|
2007-10-04 23:20:39 +00:00
|
|
|
[290, []],
|
2007-10-08 23:40:23 +00:00
|
|
|
[291, []],
|
2008-07-09 07:27:31 +00:00
|
|
|
[292, [convert_japanese_cjk]],
|
2007-10-12 23:51:56 +00:00
|
|
|
[293, []],
|
2007-10-13 19:06:09 +00:00
|
|
|
[294, [convert_pdf_options]],
|
2007-10-23 15:02:15 +00:00
|
|
|
[295, [convert_htmlurl, convert_url]],
|
2007-10-29 22:38:36 +00:00
|
|
|
[296, [convert_include]],
|
2007-11-01 11:03:51 +00:00
|
|
|
[297, [convert_usorbian]],
|
2008-10-19 10:50:32 +00:00
|
|
|
[298, [convert_macro_global]],
|
2007-11-23 02:10:00 +00:00
|
|
|
[299, []],
|
2007-11-25 18:34:37 +00:00
|
|
|
[300, []],
|
2007-11-25 21:09:01 +00:00
|
|
|
[301, []],
|
2007-11-25 22:39:04 +00:00
|
|
|
[302, []],
|
2007-12-04 09:25:50 +00:00
|
|
|
[303, [convert_serbocroatian]],
|
2007-12-05 21:42:57 +00:00
|
|
|
[304, [convert_framed_notes]],
|
2007-12-05 22:04:33 +00:00
|
|
|
[305, []],
|
2007-12-06 11:04:56 +00:00
|
|
|
[306, []],
|
2007-12-13 23:29:56 +00:00
|
|
|
[307, []],
|
2007-12-20 15:46:14 +00:00
|
|
|
[308, []],
|
2007-12-28 16:56:57 +00:00
|
|
|
[309, []],
|
2008-01-10 07:22:43 +00:00
|
|
|
[310, []],
|
2008-01-11 18:56:53 +00:00
|
|
|
[311, [convert_ams_classes]],
|
|
|
|
[312, []],
|
2008-02-03 11:56:44 +00:00
|
|
|
[313, [convert_module_names]],
|
2008-02-19 05:24:48 +00:00
|
|
|
[314, []],
|
2008-03-02 11:30:50 +00:00
|
|
|
[315, []],
|
2008-03-02 17:12:22 +00:00
|
|
|
[316, [convert_subfig]],
|
2008-03-09 23:04:14 +00:00
|
|
|
[317, []],
|
|
|
|
[318, []],
|
2008-03-18 12:11:28 +00:00
|
|
|
[319, [convert_spaceinset, convert_hfill]],
|
2008-03-18 18:08:32 +00:00
|
|
|
[320, []],
|
2008-03-24 12:31:07 +00:00
|
|
|
[321, [convert_tablines]],
|
2008-06-05 03:54:18 +00:00
|
|
|
[322, [convert_plain_layout]],
|
2008-03-26 08:10:01 +00:00
|
|
|
[323, [convert_pagebreaks]],
|
2008-03-29 23:25:40 +00:00
|
|
|
[324, [convert_linebreaks]],
|
|
|
|
[325, [convert_japanese_plain]],
|
2008-04-16 00:25:08 +00:00
|
|
|
[326, []],
|
2008-04-18 17:06:03 +00:00
|
|
|
[327, []],
|
|
|
|
[328, [remove_embedding, remove_extra_embedded_files, remove_inzip_options]],
|
2008-04-28 16:38:56 +00:00
|
|
|
[329, []],
|
2008-04-30 20:16:37 +00:00
|
|
|
[330, []],
|
2008-05-04 07:51:50 +00:00
|
|
|
[331, [convert_ltcaption]],
|
2008-05-06 21:13:09 +00:00
|
|
|
[332, []],
|
2008-05-09 08:57:32 +00:00
|
|
|
[333, [update_apa_styles]],
|
2008-05-15 22:54:13 +00:00
|
|
|
[334, [convert_paper_sizes]],
|
2008-05-30 15:38:43 +00:00
|
|
|
[335, [convert_InsetSpace]],
|
2008-06-05 06:18:34 +00:00
|
|
|
[336, []],
|
2008-06-13 07:34:55 +00:00
|
|
|
[337, [convert_display_enum]],
|
2008-06-22 23:21:46 +00:00
|
|
|
[338, []],
|
2008-07-25 20:43:32 +00:00
|
|
|
[339, []],
|
2008-09-30 18:00:02 +00:00
|
|
|
[340, [add_plain_layout]],
|
2008-10-11 23:52:31 +00:00
|
|
|
[341, []],
|
2008-10-12 09:36:00 +00:00
|
|
|
[342, []],
|
2008-10-13 16:01:02 +00:00
|
|
|
[343, [convert_default_options]],
|
2008-11-07 19:17:35 +00:00
|
|
|
[344, [convert_backref_options]],
|
|
|
|
[345, [convert_charstyle_element]]
|
2007-08-12 13:25:36 +00:00
|
|
|
]
|
|
|
|
|
2008-11-07 19:17:35 +00:00
|
|
|
revert = [[344, [revert_charstyle_element]],
|
|
|
|
[343, [revert_backref_options]],
|
2008-10-13 16:01:02 +00:00
|
|
|
[342, [revert_default_options]],
|
2008-10-12 09:36:00 +00:00
|
|
|
[341, [revert_mongolian]],
|
2008-10-11 23:52:31 +00:00
|
|
|
[340, [revert_tabulators, revert_tabsize]],
|
2008-09-30 18:00:02 +00:00
|
|
|
[339, []],
|
2008-08-01 17:57:01 +00:00
|
|
|
[338, [revert_removed_modules]],
|
2008-07-25 20:43:32 +00:00
|
|
|
[337, [revert_polytonicgreek]],
|
2008-06-22 23:21:46 +00:00
|
|
|
[336, [revert_display_enum]],
|
2008-06-13 07:34:55 +00:00
|
|
|
[335, [remove_fontsCJK]],
|
2008-06-05 06:18:34 +00:00
|
|
|
[334, [revert_InsetSpace]],
|
2008-05-30 15:38:43 +00:00
|
|
|
[333, [revert_paper_sizes]],
|
2008-05-15 22:54:13 +00:00
|
|
|
[332, []],
|
2008-05-09 08:57:32 +00:00
|
|
|
[331, [revert_graphics_group]],
|
2008-05-06 21:13:09 +00:00
|
|
|
[330, [revert_ltcaption]],
|
2008-05-04 07:51:50 +00:00
|
|
|
[329, [revert_leftarrowfill, revert_rightarrowfill, revert_upbracefill, revert_downbracefill]],
|
2008-04-30 20:16:37 +00:00
|
|
|
[328, [revert_master]],
|
2008-04-28 16:38:56 +00:00
|
|
|
[327, []],
|
2008-04-18 17:06:03 +00:00
|
|
|
[326, [revert_mexican]],
|
2008-04-16 00:25:08 +00:00
|
|
|
[325, [revert_pdfpages]],
|
2008-04-11 14:07:25 +00:00
|
|
|
[324, []],
|
2008-03-29 23:25:40 +00:00
|
|
|
[323, [revert_linebreaks]],
|
2008-03-26 08:10:01 +00:00
|
|
|
[322, [revert_pagebreaks]],
|
2008-06-05 03:54:18 +00:00
|
|
|
[321, [revert_local_layout, revert_plain_layout]],
|
2008-03-24 12:31:07 +00:00
|
|
|
[320, [revert_tablines]],
|
2008-03-18 18:08:32 +00:00
|
|
|
[319, [revert_protected_hfill]],
|
2008-03-18 12:11:28 +00:00
|
|
|
[318, [revert_spaceinset, revert_hfills, revert_hspace]],
|
*** File Format Change: UI and enhancement for InsetSpace ***
* src/insets/InsetSpace.{cpp,h}:
- merge in HFill inset, add support for dotfill, hrulefill,
hspace and hspace*
* src/insets/InsetHFill.{cpp,h}:
- remove
* src/frontends/qt4/GuiHSpace.{cpp,h}:
* src/frontends/qt4/ui/HSpaceUi.ui:
* src/frontends/qt4/GuiView:
- new GUI for Space insets.
* src/insets/Inset.{cpp,h}:
* src/insets/InsetCode.h:
* src/insets/InsetCollapsable.cpp:
* src/insets/InsetCommandParams.cpp:
- remove HFILL_CODE and LFUN_HFILL_INSERT, add SPACE_CODE where necessary,
new Inset member isStretchableSpace() to indicate HFill and friends.
* Buffer.cpp:
- increase format to 319
* lib/lyx2lyx/LyX.py:
* lib/lyx2lyx/lyx_1_6.py:
- conversion/reversion routines
* development/FORMAT:
- document file format change
* src/Makefile.am:
* src/frontends/qt4/Makefile.am:
* development/scons/scons_manifest.py:
- deal with UI changes.
* src/LyXAction.cpp:
- remove LFUN_HFILL_INSERT
* src/LyXFunc.cpp:
- handle space dialog.
* src/factory.cpp:
* src/Paragraph.cpp (isHFill):
* src/Text.cpp:
* src/Text3.cpp:
* src/TextMetrics.cpp:
- adapt to changes
* lib/ui/classic.ui:
* lib/ui/stdmenus.ui:
- add HSpace dialog, remove HFill.
git-svn-id: svn://svn.lyx.org/lyx/lyx-devel/trunk@23787 a592a061-630c-0410-9148-cb99ea01b6c8
2008-03-17 09:23:43 +00:00
|
|
|
[317, [remove_extra_embedded_files]],
|
2008-03-09 23:04:14 +00:00
|
|
|
[316, [revert_wrapplacement]],
|
2008-03-02 17:12:22 +00:00
|
|
|
[315, [revert_subfig]],
|
2008-06-05 03:54:18 +00:00
|
|
|
[314, [revert_colsep, revert_plainlayout]],
|
2008-02-19 05:24:48 +00:00
|
|
|
[313, []],
|
2008-02-03 11:56:44 +00:00
|
|
|
[312, [revert_module_names]],
|
This commit changes the way individual LyXModule's are represented, both internally and in the .lyx files. The earlier version represented them by their `descriptive name', e.g., "Endnote" or "Theorems (AMS)", these being the same names used in the UI. This was a mistake, as becomes readily apparent when one starts to think about translating these strings. The modules ought to be represented by their filename, without the extension, just as TextClass's are.
The changes that accomplish this part are in ModuleList.{h,cpp}, configure.py, and the *.module files themselves. This is a format change, and the lyx2lyx is in those files.
By itself, that change would not be major, except for the fact that we do not want the module to be represented in the UI by its filename---e.g., theorems-std---but rather by a descriptive name, such as "Theorems". But that change turns out to be wholly non-trivial. The mechanism for choosing modules was the same as---indeed, was borrowed from---that in GuiCitation: You get a list of modules, and choosing them involves moving strings from one QListView to another. The models underlying these views are just QStringListModels, which means that, when you want to know what modules have been selected, you see what strings are in the "selected" QListView. But these are just the descriptive names, and we can't look up a module by its descriptive name if it's been translated. That, indeed, was the whole point of the change to the new representation.
So, we need a more complicated model underlying the QListView, one that will pair an identifying string---the filename minus the extension, in this case---with each item. This turns out not to be terribly difficult, though it took rather a while for me to understand why it's not difficult. There are two parts:
(i) GuiSelectionManger gets re-written to use any QAbstractListModel, not just a QStringListModel. This actually seems to improve the code, independently.
(ii) We then subclass QAbstractListModel to get the associated ID string, using the Qt::UserRole slot associated with each item to store its ID. This would be almost completely trivial if QAbstractListItem::itemData() included the QVariant associated with this role, but it doesn't, so there are some additional hoops through which to jump.
The new model, a GuiIdListModel, is defined in the files by that name. The changes in GuiSelectionManger.{h,cpp} make it more abstract; the changes in GuiDocument.{h,cpp} adapt it to the new framework.
I've also updated the module documenation to accord with this change.
git-svn-id: svn://svn.lyx.org/lyx/lyx-devel/trunk@22501 a592a061-630c-0410-9148-cb99ea01b6c8
2008-01-12 04:28:12 +00:00
|
|
|
[311, [revert_rotfloat, revert_widesideways]],
|
2008-04-11 17:20:59 +00:00
|
|
|
[310, [revert_external_embedding]],
|
2008-01-10 07:22:43 +00:00
|
|
|
[309, [revert_btprintall]],
|
2007-12-28 16:56:57 +00:00
|
|
|
[308, [revert_nocite]],
|
2007-12-20 15:46:14 +00:00
|
|
|
[307, [revert_serbianlatin]],
|
2007-12-13 23:29:56 +00:00
|
|
|
[306, [revert_slash, revert_nobreakdash]],
|
2007-12-06 11:04:56 +00:00
|
|
|
[305, [revert_interlingua]],
|
2007-12-05 22:04:33 +00:00
|
|
|
[304, [revert_bahasam]],
|
2007-12-05 21:42:57 +00:00
|
|
|
[303, [revert_framed_notes]],
|
2007-12-04 09:25:50 +00:00
|
|
|
[302, []],
|
2007-11-25 22:39:04 +00:00
|
|
|
[301, [revert_latin, revert_samin]],
|
2007-11-25 21:09:01 +00:00
|
|
|
[300, [revert_linebreak]],
|
2007-11-25 18:34:37 +00:00
|
|
|
[299, [revert_pagebreak]],
|
2007-11-23 02:10:00 +00:00
|
|
|
[298, [revert_hyperlinktype]],
|
2007-11-01 17:37:43 +00:00
|
|
|
[297, [revert_macro_optional_params]],
|
2007-11-01 11:03:51 +00:00
|
|
|
[296, [revert_albanian, revert_lowersorbian, revert_uppersorbian]],
|
2007-10-29 22:42:58 +00:00
|
|
|
[295, [revert_include]],
|
2008-07-07 16:12:32 +00:00
|
|
|
[294, [revert_href, revert_url]],
|
2007-10-13 19:06:09 +00:00
|
|
|
[293, [revert_pdf_options_2]],
|
2007-10-12 23:51:56 +00:00
|
|
|
[292, [revert_inset_info]],
|
2008-07-09 07:27:31 +00:00
|
|
|
[291, [revert_japanese, revert_japanese_encoding, revert_japanese_cjk]],
|
2007-10-08 23:40:23 +00:00
|
|
|
[290, [revert_vietnamese]],
|
2007-10-04 23:20:39 +00:00
|
|
|
[289, [revert_wraptable]],
|
2007-10-03 13:38:19 +00:00
|
|
|
[288, [revert_latexcommand_index]],
|
2007-09-27 18:24:18 +00:00
|
|
|
[287, [revert_inset_command]],
|
2007-09-24 13:43:58 +00:00
|
|
|
[286, [revert_wrapfig_options]],
|
2007-09-20 22:57:23 +00:00
|
|
|
[285, [revert_pdf_options]],
|
2007-10-21 18:06:00 +00:00
|
|
|
[284, [remove_inzip_options]],
|
2007-09-09 23:47:22 +00:00
|
|
|
[283, []],
|
2007-09-08 13:01:29 +00:00
|
|
|
[282, [revert_flex]],
|
2007-08-31 09:46:14 +00:00
|
|
|
[281, []],
|
This is one of a series of patches that will merge the layout modules development in personal/branches/rgheck back into the tree.
Design goal: Allow the use of layout "modules", which are to LaTeX packages as layout files are to LaTeX document classes. Thus, one could have a module that defined certain character styles, environments, commands, or what have you, and include it in various documents, each of which uses a different document class, without having to modify the layout files themselves. For example, a theorems.module could be used with article.layout to provide support for theorem-type environments, without having to modify article.layout itself, and the same module could be used with book.layout, etc.
This patch adds the backend. The ModuleList class holds a list of the available modules, which are retrieved from lyxmodules.lst, itself generated by configure.py. There are two LFUNs available: modules-clear and module-add, which do the obvious thing; you can test by typing these into the minibuffer, along with the name of one of the available modules: URL (a CharStyle), Endnote (a Custom Inset), and---with the spaces---End To Foot (View>LaTeX and look at the user preamble), which are themselves in lib/layouts. There are some others, too, that allow theorems to be added to classes like article and book.
The GUI will come next.
Issues: (i) The configure.py script could be improved. It'd be nice, for example, if it tested for the presence of the LaTeX packages a particular module needs. But this would mean re-working the LaTeX script, and I don't know how to do that. Note that at present, the packages are ignored. This will change shortly. (ii) I've used std::string in LyXModule, following what seemed to be a precedent in TextClass. If some of these should be docstrings, please let me know, and I'll change them. (iii) There is at present no distinction between LaTeX and DocBook modules. Should there be? That is: Should there be modules that are available when the document class is a LaTeX class and others that are available only when it is DocBook? Or should there just be one set of modules? Each module can of course indicate for what it is suitable in its description.
git-svn-id: svn://svn.lyx.org/lyx/lyx-devel/trunk@19893 a592a061-630c-0410-9148-cb99ea01b6c8
2007-08-29 17:59:49 +00:00
|
|
|
[280, [revert_begin_modules]],
|
2007-08-18 23:26:07 +00:00
|
|
|
[279, [revert_show_label]],
|
2007-08-17 15:48:41 +00:00
|
|
|
[278, [revert_long_charstyle_names]],
|
2007-08-12 14:56:49 +00:00
|
|
|
[277, []],
|
2007-08-17 15:48:41 +00:00
|
|
|
[276, []]
|
2007-08-12 13:25:36 +00:00
|
|
|
]
|
|
|
|
|
|
|
|
|
|
|
|
if __name__ == "__main__":
|
|
|
|
pass
|