mirror of
https://git.lyx.org/repos/lyx.git
synced 2024-12-14 01:22:33 +00:00
6b2fe73e91
necessary to mark them in the proper language for typesetting. * src/Paragraph.cpp (Paragraph::Pimpl::writeScriptChars): new method. (Paragraph::Pimpl::simpleTeXSpecialChars): output proper latex code for greek and cyrillic chars when they cannot be encoded using the current latex encoding. * src/LaTeXFeatures.cpp: add textgreek and textcyr features. * src/Encoding.{cpp,h} (Encodings::isKnownScriptChar): new method. * lib/unicodesymbols: add greek and cyrillic alphabetic chars. git-svn-id: svn://svn.lyx.org/lyx/lyx-devel/branches/BRANCH_1_5_X@21184 a592a061-630c-0410-9148-cb99ea01b6c8
585 lines
15 KiB
C++
585 lines
15 KiB
C++
/**
|
|
* \file Encoding.cpp
|
|
* This file is part of LyX, the document processor.
|
|
* Licence details can be found in the file COPYING.
|
|
*
|
|
* \author Lars Gullik Bjønnes
|
|
* \author Jean-Marc Lasgouttes
|
|
* \author Dekel Tsur
|
|
*
|
|
* Full author contact details are available in file CREDITS.
|
|
*/
|
|
|
|
#include <config.h>
|
|
|
|
#include "Encoding.h"
|
|
|
|
#include "debug.h"
|
|
#include "LaTeXFeatures.h"
|
|
#include "Lexer.h"
|
|
#include "LyXRC.h"
|
|
|
|
#include "support/FileName.h"
|
|
#include "support/lstrings.h"
|
|
#include "support/unicode.h"
|
|
|
|
#include <sstream>
|
|
|
|
#ifndef CXX_GLOBAL_CSTD
|
|
using std::strtol;
|
|
#endif
|
|
using std::endl;
|
|
using std::string;
|
|
|
|
|
|
namespace lyx {
|
|
|
|
using support::FileName;
|
|
|
|
Encodings encodings;
|
|
|
|
namespace {
|
|
|
|
char_type arabic_table[172][4] = {
|
|
{0xfe80, 0xfe80, 0xfe80, 0xfe80}, // 0x0621 = hamza
|
|
{0xfe81, 0xfe82, 0xfe81, 0xfe82}, // 0x0622 = ligature madda on alef
|
|
{0xfe83, 0xfe84, 0xfe83, 0xfe84}, // 0x0623 = ligature hamza on alef
|
|
{0xfe85, 0xfe86, 0xfe85, 0xfe86}, // 0x0624 = ligature hamza on waw
|
|
{0xfe87, 0xfe88, 0xfe87, 0xfe88}, // 0x0625 = ligature hamza under alef
|
|
{0xfe89, 0xfe8a, 0xfe8b, 0xfe8c}, // 0x0626 = ligature hamza on ya
|
|
{0xfe8d, 0xfe8e, 0xfe8d, 0xfe8e}, // 0x0627 = alef
|
|
{0xfe8f, 0xfe90, 0xfe91, 0xfe92}, // 0x0628 = baa
|
|
{0xfe93, 0xfe94, 0xfe93, 0xfe94}, // 0x0629 = taa marbuta
|
|
{0xfe95, 0xfe96, 0xfe97, 0xfe98}, // 0x062a = taa
|
|
{0xfe99, 0xfe9a, 0xfe9b, 0xfe9c}, // 0x062b = thaa
|
|
{0xfe9d, 0xfe9e, 0xfe9f, 0xfea0}, // 0x062c = jeem
|
|
{0xfea1, 0xfea2, 0xfea3, 0xfea4}, // 0x062d = haa
|
|
{0xfea5, 0xfea6, 0xfea7, 0xfea8}, // 0x062e = khaa
|
|
{0xfea9, 0xfeaa, 0xfea9, 0xfeaa}, // 0x062f = dal
|
|
|
|
{0xfeab, 0xfeac, 0xfeab, 0xfeac}, // 0x0630 = thal
|
|
{0xfead, 0xfeae, 0xfead, 0xfeae}, // 0x0631 = ra
|
|
{0xfeaf, 0xfeb0, 0xfeaf, 0xfeb0}, // 0x0632 = zain
|
|
{0xfeb1, 0xfeb2, 0xfeb3, 0xfeb4}, // 0x0633 = seen
|
|
{0xfeb5, 0xfeb6, 0xfeb7, 0xfeb8}, // 0x0634 = sheen
|
|
{0xfeb9, 0xfeba, 0xfebb, 0xfebc}, // 0x0635 = sad
|
|
{0xfebd, 0xfebe, 0xfebf, 0xfec0}, // 0x0636 = dad
|
|
{0xfec1, 0xfec2, 0xfec3, 0xfec4}, // 0x0637 = tah
|
|
{0xfec5, 0xfec6, 0xfec7, 0xfec8}, // 0x0638 = zah
|
|
{0xfec9, 0xfeca, 0xfecb, 0xfecc}, // 0x0639 = ain
|
|
{0xfecd, 0xfece, 0xfecf, 0xfed0}, // 0x063a = ghain
|
|
{0, 0, 0, 0}, // 0x063b
|
|
{0, 0, 0, 0}, // 0x063c
|
|
{0, 0, 0, 0}, // 0x063d
|
|
{0, 0, 0, 0}, // 0x063e
|
|
{0, 0, 0, 0}, // 0x063f
|
|
|
|
{0, 0, 0, 0}, // 0x0640
|
|
{0xfed1, 0xfed2, 0xfed3, 0xfed4}, // 0x0641 = fa
|
|
{0xfed5, 0xfed6, 0xfed7, 0xfed8}, // 0x0642 = qaf
|
|
{0xfed9, 0xfeda, 0xfedb, 0xfedc}, // 0x0643 = kaf
|
|
{0xfedd, 0xfede, 0xfedf, 0xfee0}, // 0x0644 = lam
|
|
{0xfee1, 0xfee2, 0xfee3, 0xfee4}, // 0x0645 = meem
|
|
{0xfee5, 0xfee6, 0xfee7, 0xfee8}, // 0x0646 = noon
|
|
{0xfee9, 0xfeea, 0xfeeb, 0xfeec}, // 0x0647 = ha
|
|
{0xfeed, 0xfeee, 0xfeed, 0xfeee}, // 0x0648 = waw
|
|
{0xfeef, 0xfef0, 0xfeef, 0xfef0}, // 0x0649 = alef maksura
|
|
{0xfef1, 0xfef2, 0xfef3, 0xfef4}, // 0x064a = ya
|
|
{0x065b, 0x065b, 0x065b, 0x065b}, // 0x064b = fathatan
|
|
{0x065c, 0x065c, 0x065c, 0x065c}, // 0x064c = dammatan
|
|
{0x064d, 0x064d, 0x064d, 0x064d}, // 0x064d = kasratan
|
|
{0x064e, 0x064e, 0x064e, 0x064e}, // 0x064e = fatha
|
|
{0x064f, 0x064f, 0x064f, 0x064f}, // 0x064f = damma
|
|
|
|
{0x0650, 0x0650, 0x0650, 0x0650}, // 0x0650 = kasra
|
|
{0x0651, 0x0651, 0x0651, 0x0651}, // 0x0651 = shadda
|
|
{0x0652, 0x0652, 0x0652, 0x0652}, // 0x0652 = sukun
|
|
|
|
{0, 0, 0, 0}, // 0x0653
|
|
{0, 0, 0, 0}, // 0x0654
|
|
{0, 0, 0, 0}, // 0x0655
|
|
{0, 0, 0, 0}, // 0x0656
|
|
{0, 0, 0, 0}, // 0x0657
|
|
{0, 0, 0, 0}, // 0x0658
|
|
{0, 0, 0, 0}, // 0x0659
|
|
{0, 0, 0, 0}, // 0x065a
|
|
{0, 0, 0, 0}, // 0x065b
|
|
{0, 0, 0, 0}, // 0x065c
|
|
{0, 0, 0, 0}, // 0x065d
|
|
{0, 0, 0, 0}, // 0x065e
|
|
{0, 0, 0, 0}, // 0x065f
|
|
{0, 0, 0, 0}, // 0x0660
|
|
{0, 0, 0, 0}, // 0x0661
|
|
{0, 0, 0, 0}, // 0x0662
|
|
{0, 0, 0, 0}, // 0x0663
|
|
{0, 0, 0, 0}, // 0x0664
|
|
{0, 0, 0, 0}, // 0x0665
|
|
{0, 0, 0, 0}, // 0x0666
|
|
{0, 0, 0, 0}, // 0x0667
|
|
{0, 0, 0, 0}, // 0x0668
|
|
{0, 0, 0, 0}, // 0x0669
|
|
{0, 0, 0, 0}, // 0x066a
|
|
{0, 0, 0, 0}, // 0x066b
|
|
{0, 0, 0, 0}, // 0x066c
|
|
{0, 0, 0, 0}, // 0x066d
|
|
{0, 0, 0, 0}, // 0x066e
|
|
{0, 0, 0, 0}, // 0x066f
|
|
{0, 0, 0, 0}, // 0x0670
|
|
{0, 0, 0, 0}, // 0x0671
|
|
{0, 0, 0, 0}, // 0x0672
|
|
{0, 0, 0, 0}, // 0x0673
|
|
{0, 0, 0, 0}, // 0x0674
|
|
{0, 0, 0, 0}, // 0x0675
|
|
{0, 0, 0, 0}, // 0x0676
|
|
{0, 0, 0, 0}, // 0x0677
|
|
{0, 0, 0, 0}, // 0x0678
|
|
{0, 0, 0, 0}, // 0x0679
|
|
{0, 0, 0, 0}, // 0x067a
|
|
{0, 0, 0, 0}, // 0x067b
|
|
{0, 0, 0, 0}, // 0x067c
|
|
{0, 0, 0, 0}, // 0x067d
|
|
{0xfb56, 0xfb57, 0xfb58, 0xfb59}, // 0x067e = peh
|
|
{0, 0, 0, 0}, // 0x067f
|
|
{0, 0, 0, 0}, // 0x0680
|
|
{0, 0, 0, 0}, // 0x0681
|
|
{0, 0, 0, 0}, // 0x0682
|
|
{0, 0, 0, 0}, // 0x0683
|
|
{0, 0, 0, 0}, // 0x0684
|
|
{0, 0, 0, 0}, // 0x0685
|
|
{0xfb7a, 0xfb7b, 0xfb7c, 0xfb7d}, // 0x0686 = tcheh
|
|
{0, 0, 0, 0}, // 0x0687
|
|
{0, 0, 0, 0}, // 0x0688
|
|
{0, 0, 0, 0}, // 0x0689
|
|
{0, 0, 0, 0}, // 0x068a
|
|
{0, 0, 0, 0}, // 0x068b
|
|
{0, 0, 0, 0}, // 0x068c
|
|
{0, 0, 0, 0}, // 0x068d
|
|
{0, 0, 0, 0}, // 0x068e
|
|
{0, 0, 0, 0}, // 0x068f
|
|
{0, 0, 0, 0}, // 0x0690
|
|
{0, 0, 0, 0}, // 0x0691
|
|
{0, 0, 0, 0}, // 0x0692
|
|
{0, 0, 0, 0}, // 0x0693
|
|
{0, 0, 0, 0}, // 0x0694
|
|
{0, 0, 0, 0}, // 0x0695
|
|
{0, 0, 0, 0}, // 0x0696
|
|
{0, 0, 0, 0}, // 0x0697
|
|
{0xfb8a, 0xfb8b, 0xfb8a, 0xfb8b}, // 0x0698 = jeh
|
|
{0, 0, 0, 0}, // 0x0699
|
|
{0, 0, 0, 0}, // 0x069a
|
|
{0, 0, 0, 0}, // 0x069b
|
|
{0, 0, 0, 0}, // 0x069c
|
|
{0, 0, 0, 0}, // 0x069d
|
|
{0, 0, 0, 0}, // 0x069e
|
|
{0, 0, 0, 0}, // 0x069f
|
|
{0, 0, 0, 0}, // 0x06a0
|
|
{0, 0, 0, 0}, // 0x06a1
|
|
{0, 0, 0, 0}, // 0x06a2
|
|
{0, 0, 0, 0}, // 0x06a3
|
|
{0, 0, 0, 0}, // 0x06a4
|
|
{0, 0, 0, 0}, // 0x06a5
|
|
{0, 0, 0, 0}, // 0x06a6
|
|
{0, 0, 0, 0}, // 0x06a7
|
|
{0, 0, 0, 0}, // 0x06a8
|
|
{0xfb8e, 0xfb8f, 0xfb90, 0xfb91}, // 0x06a9 = farsi kaf
|
|
{0, 0, 0, 0}, // 0x06aa
|
|
{0, 0, 0, 0}, // 0x06ab
|
|
{0, 0, 0, 0}, // 0x06ac
|
|
{0, 0, 0, 0}, // 0x06ad
|
|
{0, 0, 0, 0}, // 0x06ae
|
|
{0xfb92, 0xfb93, 0xfb94, 0xfb95}, // 0x06af = gaf
|
|
{0, 0, 0, 0}, // 0x06b0
|
|
{0, 0, 0, 0}, // 0x06b1
|
|
{0, 0, 0, 0}, // 0x06b2
|
|
{0, 0, 0, 0}, // 0x06b3
|
|
{0, 0, 0, 0}, // 0x06b4
|
|
{0, 0, 0, 0}, // 0x06b5
|
|
{0, 0, 0, 0}, // 0x06b6
|
|
{0, 0, 0, 0}, // 0x06b7
|
|
{0, 0, 0, 0}, // 0x06b8
|
|
{0, 0, 0, 0}, // 0x06b9
|
|
{0, 0, 0, 0}, // 0x06ba
|
|
{0, 0, 0, 0}, // 0x06bb
|
|
{0, 0, 0, 0}, // 0x06bc
|
|
{0, 0, 0, 0}, // 0x06bd
|
|
{0, 0, 0, 0}, // 0x06be
|
|
{0, 0, 0, 0}, // 0x06bf
|
|
{0, 0, 0, 0}, // 0x06c0
|
|
{0, 0, 0, 0}, // 0x06c1
|
|
{0, 0, 0, 0}, // 0x06c2
|
|
{0, 0, 0, 0}, // 0x06c3
|
|
{0, 0, 0, 0}, // 0x06c4
|
|
{0, 0, 0, 0}, // 0x06c5
|
|
{0, 0, 0, 0}, // 0x06c6
|
|
{0, 0, 0, 0}, // 0x06c7
|
|
{0, 0, 0, 0}, // 0x06c8
|
|
{0, 0, 0, 0}, // 0x06c9
|
|
{0, 0, 0, 0}, // 0x06ca
|
|
{0, 0, 0, 0}, // 0x06cb
|
|
{0xfbfc, 0xfbfd, 0xfbfe, 0xfbff} // 0x06cc = farsi yeh
|
|
};
|
|
|
|
|
|
char_type const arabic_start = 0x0621;
|
|
char_type const arabic_end = 0x06cc;
|
|
|
|
|
|
/// Information about a single UCS4 character
|
|
struct CharInfo {
|
|
/// LaTeX command for this character
|
|
docstring command;
|
|
/// Needed LaTeX preamble (or feature)
|
|
string preamble;
|
|
/// Is this a combining character?
|
|
bool combining;
|
|
/// Is \c preamble a feature known by LaTeXFeatures, or a raw LaTeX
|
|
/// command?
|
|
bool feature;
|
|
/// Always force the LaTeX command, even if the encoding contains
|
|
/// this character?
|
|
bool force;
|
|
};
|
|
|
|
|
|
typedef std::map<char_type, CharInfo> CharInfoMap;
|
|
CharInfoMap unicodesymbols;
|
|
|
|
|
|
/// The highest code point in UCS4 encoding (1<<20 + 1<<16)
|
|
char_type const max_ucs4 = 0x110000;
|
|
|
|
} // namespace anon
|
|
|
|
|
|
Encoding::Encoding(string const & n, string const & l, string const & i,
|
|
bool f, Encoding::Package p)
|
|
: Name_(n), LatexName_(l), iconvName_(i), fixedwidth_(f), package_(p)
|
|
{
|
|
if (n == "ascii") {
|
|
// ASCII can encode 128 code points and nothing else
|
|
start_encodable_ = 128;
|
|
complete_ = true;
|
|
} else if (i == "UTF-8") {
|
|
// UTF8 can encode all UCS4 code points
|
|
start_encodable_ = max_ucs4;
|
|
complete_ = true;
|
|
} else {
|
|
complete_ = false;
|
|
}
|
|
}
|
|
|
|
|
|
void Encoding::init() const
|
|
{
|
|
start_encodable_ = 0;
|
|
// temporarily switch off lyxerr, since we will generate iconv errors
|
|
lyxerr.disable();
|
|
if (fixedwidth_) {
|
|
// We do not need to check all UCS4 code points, it is enough
|
|
// if we check all 256 code points of this encoding.
|
|
for (unsigned short j = 0; j < 256; ++j) {
|
|
char const c = j;
|
|
std::vector<char_type> const ucs4 = eightbit_to_ucs4(&c, 1, iconvName_);
|
|
if (ucs4.size() == 1) {
|
|
char_type const c = ucs4[0];
|
|
CharInfoMap::const_iterator const it = unicodesymbols.find(c);
|
|
if (it == unicodesymbols.end() || !it->second.force)
|
|
encodable_.insert(c);
|
|
}
|
|
}
|
|
} else {
|
|
// We do not know how many code points this encoding has, and
|
|
// they do not have a direct representation as a single byte,
|
|
// therefore we need to check all UCS4 code points.
|
|
// This is expensive!
|
|
for (char_type c = 0; c < max_ucs4; ++c) {
|
|
std::vector<char> const eightbit = ucs4_to_eightbit(&c, 1, iconvName_);
|
|
if (!eightbit.empty()) {
|
|
CharInfoMap::const_iterator const it = unicodesymbols.find(c);
|
|
if (it == unicodesymbols.end() || !it->second.force)
|
|
encodable_.insert(c);
|
|
}
|
|
}
|
|
}
|
|
lyxerr.enable();
|
|
CharSet::iterator it = encodable_.find(start_encodable_);
|
|
while (it != encodable_.end()) {
|
|
encodable_.erase(it);
|
|
++start_encodable_;
|
|
it = encodable_.find(start_encodable_);
|
|
}
|
|
complete_ = true;
|
|
}
|
|
|
|
|
|
docstring const Encoding::latexChar(char_type c) const
|
|
{
|
|
// assure the used encoding is properly initialized
|
|
if (!complete_)
|
|
init();
|
|
BOOST_ASSERT(complete_);
|
|
|
|
if (c < start_encodable_)
|
|
return docstring(1, c);
|
|
if (encodable_.find(c) == encodable_.end()) {
|
|
// c cannot be encoded in this encoding
|
|
CharInfoMap::const_iterator const it = unicodesymbols.find(c);
|
|
if (it == unicodesymbols.end())
|
|
lyxerr << "Could not find LaTeX command for character 0x"
|
|
<< std::hex << c << std::dec
|
|
<< ".\nLaTeX export will fail."
|
|
<< endl;
|
|
else
|
|
return it->second.command;
|
|
}
|
|
return docstring(1, c);
|
|
}
|
|
|
|
|
|
void Encodings::validate(char_type c, LaTeXFeatures & features)
|
|
{
|
|
CharInfoMap::const_iterator const it = unicodesymbols.find(c);
|
|
if (it != unicodesymbols.end() && !it->second.preamble.empty()) {
|
|
if (it->second.feature)
|
|
features.require(it->second.preamble);
|
|
else
|
|
features.addPreambleSnippet(it->second.preamble);
|
|
}
|
|
}
|
|
|
|
|
|
bool Encodings::isComposeChar_hebrew(char_type c)
|
|
{
|
|
return c <= 0x05c2 && c >= 0x05b0 &&
|
|
c != 0x05be && c != 0x05c0;
|
|
}
|
|
|
|
|
|
// Special Arabic letters are ones that do not get connected from left
|
|
// they are hamza, alef_madda, alef_hamza, waw_hamza, alef_hamza_under,
|
|
// alef, tah_marbota, dal, thal, rah, zai, wow, alef_maksoura
|
|
|
|
bool Encodings::is_arabic_special(char_type c)
|
|
{
|
|
return (c >= 0x0621 && c <= 0x0625) ||
|
|
c == 0x0627 || c == 0x0629 ||
|
|
c == 0x062f || c == 0x0648 ||
|
|
(c >= 0x0630 && c <= 0x0632) ||
|
|
c == 0x0649 || c == 0x0698;
|
|
}
|
|
|
|
|
|
bool Encodings::isComposeChar_arabic(char_type c)
|
|
{
|
|
return c >= 0x064b && c <= 0x0652;
|
|
}
|
|
|
|
|
|
bool Encodings::is_arabic(char_type c)
|
|
{
|
|
return c >= arabic_start && c <= arabic_end &&
|
|
arabic_table[c-arabic_start][0];
|
|
}
|
|
|
|
|
|
char_type Encodings::transformChar(char_type c,
|
|
Encodings::Letter_Form form)
|
|
{
|
|
if (!is_arabic(c))
|
|
return c;
|
|
|
|
return arabic_table[c-arabic_start][form];
|
|
}
|
|
|
|
|
|
bool Encodings::isCombiningChar(char_type c)
|
|
{
|
|
CharInfoMap::const_iterator const it = unicodesymbols.find(c);
|
|
if (it != unicodesymbols.end())
|
|
return it->second.combining;
|
|
return false;
|
|
}
|
|
|
|
|
|
bool Encodings::isKnownScriptChar(char_type const c, string & preamble)
|
|
{
|
|
CharInfoMap::const_iterator const it = unicodesymbols.find(c);
|
|
|
|
if (it == unicodesymbols.end())
|
|
return false;
|
|
|
|
if (it->second.preamble != "textgreek" &&
|
|
it->second.preamble != "textcyr")
|
|
return false;
|
|
|
|
if (preamble.empty()) {
|
|
preamble = it->second.preamble;
|
|
return true;
|
|
}
|
|
return it->second.preamble == preamble;
|
|
}
|
|
|
|
|
|
Encoding const * Encodings::getFromLyXName(string const & name) const
|
|
{
|
|
EncodingList::const_iterator it = encodinglist.find(name);
|
|
if (it != encodinglist.end())
|
|
return &it->second;
|
|
else
|
|
return 0;
|
|
}
|
|
|
|
|
|
Encoding const * Encodings::getFromLaTeXName(string const & name) const
|
|
{
|
|
// We don't use std::find_if because it makes copies of the pairs in
|
|
// the map.
|
|
// This linear search is OK since we don't have many encodings.
|
|
// Users could even optimize it by putting the encodings they use
|
|
// most at the top of lib/encodings.
|
|
EncodingList::const_iterator const end = encodinglist.end();
|
|
for (EncodingList::const_iterator it = encodinglist.begin(); it != end; ++it)
|
|
if (it->second.latexName() == name)
|
|
return &it->second;
|
|
return 0;
|
|
}
|
|
|
|
|
|
Encodings::Encodings()
|
|
{
|
|
}
|
|
|
|
|
|
void Encodings::read(FileName const & encfile, FileName const & symbolsfile)
|
|
{
|
|
// We must read the symbolsfile first, because the Encoding
|
|
// constructor depends on it.
|
|
Lexer symbolslex(0, 0);
|
|
symbolslex.setFile(symbolsfile);
|
|
while (symbolslex.isOK()) {
|
|
char_type symbol;
|
|
CharInfo info;
|
|
string flags;
|
|
|
|
if (symbolslex.next(true)) {
|
|
std::istringstream is(symbolslex.getString());
|
|
// reading symbol directly does not work if
|
|
// char_type == std::wchar_t.
|
|
boost::uint32_t tmp;
|
|
if(!(is >> std::hex >> tmp))
|
|
break;
|
|
symbol = tmp;
|
|
} else
|
|
break;
|
|
if (symbolslex.next(true))
|
|
info.command = symbolslex.getDocString();
|
|
else
|
|
break;
|
|
if (symbolslex.next(true))
|
|
info.preamble = symbolslex.getString();
|
|
else
|
|
break;
|
|
if (symbolslex.next(true))
|
|
flags = symbolslex.getString();
|
|
else
|
|
break;
|
|
|
|
info.combining = false;
|
|
info.feature = false;
|
|
info.force = false;
|
|
while (!flags.empty()) {
|
|
string flag;
|
|
flags = support::split(flags, flag, ',');
|
|
if (flag == "combining")
|
|
info.combining = true;
|
|
else if (flag == "force")
|
|
info.force = true;
|
|
else
|
|
lyxerr << "Ignoring unknown flag `" << flag
|
|
<< "' for symbol `0x"
|
|
<< std::hex << symbol << std::dec
|
|
<< "'." << endl;
|
|
}
|
|
|
|
if (!info.preamble.empty())
|
|
info.feature = info.preamble[0] != '\\';
|
|
|
|
LYXERR(Debug::INFO)
|
|
<< "Read unicode symbol " << symbol << " '"
|
|
<< to_utf8(info.command) << "' '" << info.preamble
|
|
<< "' " << info.combining << ' ' << info.feature
|
|
<< endl;
|
|
unicodesymbols[symbol] = info;
|
|
}
|
|
|
|
// Now read the encodings
|
|
enum Encodingtags {
|
|
et_encoding = 1,
|
|
et_end,
|
|
et_last
|
|
};
|
|
|
|
struct keyword_item encodingtags[et_last - 1] = {
|
|
{ "encoding", et_encoding },
|
|
{ "end", et_end }
|
|
};
|
|
|
|
Lexer lex(encodingtags, et_last - 1);
|
|
lex.setFile(encfile);
|
|
while (lex.isOK()) {
|
|
switch (lex.lex()) {
|
|
case et_encoding:
|
|
{
|
|
lex.next();
|
|
string const name = lex.getString();
|
|
lex.next();
|
|
string const latexname = lex.getString();
|
|
lex.next();
|
|
string const iconvname = lex.getString();
|
|
lex.next();
|
|
string const width = lex.getString();
|
|
bool fixedwidth = true;
|
|
if (width == "fixed")
|
|
fixedwidth = true;
|
|
else if (width == "variable")
|
|
fixedwidth = false;
|
|
else
|
|
lex.printError("Encodings::read: "
|
|
"Unknown width: `$$Token'");
|
|
lex.next();
|
|
string const p = lex.getString();
|
|
Encoding::Package package = Encoding::none;
|
|
if (p == "none")
|
|
package = Encoding::none;
|
|
else if (p == "inputenc")
|
|
package = Encoding::inputenc;
|
|
else if (p == "CJK")
|
|
package = Encoding::CJK;
|
|
else
|
|
lex.printError("Encodings::read: "
|
|
"Unknown package: `$$Token'");
|
|
LYXERR(Debug::INFO) << "Reading encoding " << name << endl;
|
|
encodinglist[name] = Encoding(name, latexname,
|
|
iconvname, fixedwidth,
|
|
package);
|
|
if (lex.lex() != et_end)
|
|
lex.printError("Encodings::read: "
|
|
"missing end");
|
|
break;
|
|
}
|
|
case et_end:
|
|
lex.printError("Encodings::read: Misplaced end");
|
|
break;
|
|
case Lexer::LEX_FEOF:
|
|
break;
|
|
default:
|
|
lex.printError("Encodings::read: "
|
|
"Unknown tag: `$$Token'");
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
} // namespace lyx
|