-/** The .tex to .lyx converter
- \author André Pönitz (2003)
+/**
+ * \file tex2lyx/text.C
+ * This file is part of LyX, the document processor.
+ * Licence details can be found in the file COPYING.
+ *
+ * \author André Pönitz
+ * \author Jean-Marc Lasgouttes
+ *
+ * Full author contact details are available in file CREDITS.
*/
// {[(
#include <config.h>
#include "tex2lyx.h"
+#include "context.h"
#include "FloatList.h"
-#include "lyxtextclass.h"
+#include "lengthcommon.h"
#include "support/lstrings.h"
-#include "support/tostr.h"
+#include "support/convert.h"
+#include "support/filetools.h"
+
+#include <boost/filesystem/operations.hpp>
+#include <boost/tuple/tuple.hpp>
#include <iostream>
#include <map>
#include <sstream>
#include <vector>
+
+namespace lyx {
+
+using support::addExtension;
+using support::changeExtension;
+using support::FileName;
+using support::makeAbsPath;
+using support::makeRelPath;
+using support::rtrim;
+using support::suffixIs;
+using support::contains;
+using support::subst;
+
using std::cerr;
using std::endl;
+
using std::map;
using std::ostream;
using std::ostringstream;
+using std::istringstream;
using std::string;
using std::vector;
-using lyx::support::rtrim;
-using lyx::support::suffixIs;
+namespace fs = boost::filesystem;
+
+
+void parse_text_in_inset(Parser & p, ostream & os, unsigned flags, bool outer,
+ Context const & context)
+{
+ Context newcontext(true, context.textclass);
+ newcontext.font = context.font;
+ parse_text(p, os, flags, outer, newcontext);
+ newcontext.check_end_layout(os);
+}
+
namespace {
-char const * known_latex_commands[] = { "ref", "cite", "label", "index",
-"printindex", "pageref", "url", 0 };
+/// parses a paragraph snippet, useful for example for \\emph{...}
+void parse_text_snippet(Parser & p, ostream & os, unsigned flags, bool outer,
+ Context & context)
+{
+ Context newcontext(context);
+ // Don't inherit the extra stuff
+ newcontext.extra_stuff.clear();
+ parse_text(p, os, flags, outer, newcontext);
+ // Make sure that we don't create invalid .lyx files
+ context.need_layout = newcontext.need_layout;
+ context.need_end_layout = newcontext.need_end_layout;
+}
+
+
+/*!
+ * Thin wrapper around parse_text_snippet() using a string.
+ *
+ * We completely ignore \c context.need_layout and \c context.need_end_layout,
+ * because our return value is not used directly (otherwise the stream version
+ * of parse_text_snippet() could be used). That means that the caller needs
+ * to do layout management manually.
+ * This is intended to parse text that does not create any layout changes.
+ */
+string parse_text_snippet(Parser & p, unsigned flags, const bool outer,
+ Context & context)
+{
+ Context newcontext(context);
+ newcontext.need_layout = false;
+ newcontext.need_end_layout = false;
+ newcontext.new_layout_allowed = false;
+ // Avoid warning by Context::~Context()
+ newcontext.extra_stuff.clear();
+ ostringstream os;
+ parse_text_snippet(p, os, flags, outer, newcontext);
+ return os.str();
+}
+
+
+char const * const known_latex_commands[] = { "ref", "cite", "label", "index",
+"printindex", "pageref", "url", "vref", "vpageref", "prettyref", "eqref", 0 };
-// LaTeX names for quotes
-char const * known_quotes[] = { "glqq", "grqq", "quotedblbase",
+/*!
+ * natbib commands.
+ * We can't put these into known_latex_commands because the argument order
+ * is reversed in lyx if there are 2 arguments.
+ * The starred forms are also known.
+ */
+char const * const known_natbib_commands[] = { "cite", "citet", "citep",
+"citealt", "citealp", "citeauthor", "citeyear", "citeyearpar",
+"citefullauthor", "Citet", "Citep", "Citealt", "Citealp", "Citeauthor", 0 };
+
+/*!
+ * jurabib commands.
+ * We can't put these into known_latex_commands because the argument order
+ * is reversed in lyx if there are 2 arguments.
+ * No starred form other than "cite*" known.
+ */
+char const * const known_jurabib_commands[] = { "cite", "citet", "citep",
+"citealt", "citealp", "citeauthor", "citeyear", "citeyearpar",
+// jurabib commands not (yet) supported by LyX:
+// "fullcite",
+// "footcite", "footcitet", "footcitep", "footcitealt", "footcitealp",
+// "footciteauthor", "footciteyear", "footciteyearpar",
+"citefield", "citetitle", "cite*", 0 };
+
+/// LaTeX names for quotes
+char const * const known_quotes[] = { "glqq", "grqq", "quotedblbase",
"textquotedblleft", "quotesinglbase", "guilsinglleft", "guilsinglright", 0};
-// the same as known_quotes with .lyx names
-char const * known_coded_quotes[] = { "gld", "grd", "gld",
-"grd", "gls", "fls", "frd", 0};
+/// the same as known_quotes with .lyx names
+char const * const known_coded_quotes[] = { "gld", "grd", "gld",
+"grd", "gls", "fls", "frs", 0};
-char const * known_sizes[] = { "tiny", "scriptsize", "footnotesize",
+/// LaTeX names for font sizes
+char const * const known_sizes[] = { "tiny", "scriptsize", "footnotesize",
"small", "normalsize", "large", "Large", "LARGE", "huge", "Huge", 0};
-char const * known_coded_sizes[] = { "tiny", "scriptsize", "footnotesize",
+/// the same as known_sizes with .lyx names
+char const * const known_coded_sizes[] = { "tiny", "scriptsize", "footnotesize",
"small", "normal", "large", "larger", "largest", "huge", "giant", 0};
+/// LaTeX 2.09 names for font families
+char const * const known_old_font_families[] = { "rm", "sf", "tt", 0};
-string cap(string s)
-{
- if (s.size())
- s[0] = toupper(s[0]);
- return s;
-}
+/// LaTeX names for font families
+char const * const known_font_families[] = { "rmfamily", "sffamily",
+"ttfamily", 0};
+
+/// the same as known_old_font_families and known_font_families with .lyx names
+char const * const known_coded_font_families[] = { "roman", "sans",
+"typewriter", 0};
+
+/// LaTeX 2.09 names for font series
+char const * const known_old_font_series[] = { "bf", 0};
+
+/// LaTeX names for font series
+char const * const known_font_series[] = { "bfseries", "mdseries", 0};
+/// the same as known_old_font_series and known_font_series with .lyx names
+char const * const known_coded_font_series[] = { "bold", "medium", 0};
-// splits "x=z, y=b" into a map
+/// LaTeX 2.09 names for font shapes
+char const * const known_old_font_shapes[] = { "it", "sl", "sc", 0};
+
+/// LaTeX names for font shapes
+char const * const known_font_shapes[] = { "itshape", "slshape", "scshape",
+"upshape", 0};
+
+/// the same as known_old_font_shapes and known_font_shapes with .lyx names
+char const * const known_coded_font_shapes[] = { "italic", "slanted",
+"smallcaps", "up", 0};
+
+/*!
+ * Graphics file extensions known by the dvips driver of the graphics package.
+ * These extensions are used to complete the filename of an included
+ * graphics file if it does not contain an extension.
+ * The order must be the same that latex uses to find a file, because we
+ * will use the first extension that matches.
+ * This is only an approximation for the common cases. If we would want to
+ * do it right in all cases, we would need to know which graphics driver is
+ * used and know the extensions of every driver of the graphics package.
+ */
+char const * const known_dvips_graphics_formats[] = {"eps", "ps", "eps.gz",
+"ps.gz", "eps.Z", "ps.Z", 0};
+
+/*!
+ * Graphics file extensions known by the pdftex driver of the graphics package.
+ * \sa known_dvips_graphics_formats
+ */
+char const * const known_pdftex_graphics_formats[] = {"png", "pdf", "jpg",
+"mps", "tif", 0};
+
+/*!
+ * Known file extensions for TeX files as used by \\include.
+ */
+char const * const known_tex_extensions[] = {"tex", 0};
+
+/// spaces known by InsetSpace
+char const * const known_spaces[] = { " ", "space", ",", "thinspace", "quad",
+"qquad", "enspace", "enskip", "negthinspace", 0};
+
+/// the same as known_spaces with .lyx names
+char const * const known_coded_spaces[] = { "space{}", "space{}",
+"thinspace{}", "thinspace{}", "quad{}", "qquad{}", "enspace{}", "enskip{}",
+"negthinspace{}", 0};
+
+
+/// splits "x=z, y=b" into a map
map<string, string> split_map(string const & s)
{
map<string, string> res;
}
+/*!
+ * Split a LaTeX length into value and unit.
+ * The latter can be a real unit like "pt", or a latex length variable
+ * like "\textwidth". The unit may contain additional stuff like glue
+ * lengths, but we don't care, because such lengths are ERT anyway.
+ * \returns true if \p value and \p unit are valid.
+ */
+bool splitLatexLength(string const & len, string & value, string & unit)
+{
+ if (len.empty())
+ return false;
+ const string::size_type i = len.find_first_not_of(" -+0123456789.,");
+ //'4,5' is a valid LaTeX length number. Change it to '4.5'
+ string const length = subst(len, ',', '.');
+ if (i == string::npos)
+ return false;
+ if (i == 0) {
+ if (len[0] == '\\') {
+ // We had something like \textwidth without a factor
+ value = "1.0";
+ } else {
+ return false;
+ }
+ } else {
+ value = trim(string(length, 0, i));
+ }
+ if (value == "-")
+ value = "-1.0";
+ // 'cM' is a valid LaTeX length unit. Change it to 'cm'
+ if (contains(len, '\\'))
+ unit = trim(string(len, i));
+ else
+ unit = lyx::support::lowercase(trim(string(len, i)));
+ return true;
+}
+
+
+/// A simple function to translate a latex length to something lyx can
+/// understand. Not perfect, but rather best-effort.
+bool translate_len(string const & length, string & valstring, string & unit)
+{
+ if (!splitLatexLength(length, valstring, unit))
+ return false;
+ // LyX uses percent values
+ double value;
+ istringstream iss(valstring);
+ iss >> value;
+ value *= 100;
+ ostringstream oss;
+ oss << value;
+ string const percentval = oss.str();
+ // a normal length
+ if (unit.empty() || unit[0] != '\\')
+ return true;
+ string::size_type const i = unit.find(' ');
+ string const endlen = (i == string::npos) ? string() : string(unit, i);
+ if (unit == "\\textwidth") {
+ valstring = percentval;
+ unit = "text%" + endlen;
+ } else if (unit == "\\columnwidth") {
+ valstring = percentval;
+ unit = "col%" + endlen;
+ } else if (unit == "\\paperwidth") {
+ valstring = percentval;
+ unit = "page%" + endlen;
+ } else if (unit == "\\linewidth") {
+ valstring = percentval;
+ unit = "line%" + endlen;
+ } else if (unit == "\\paperheight") {
+ valstring = percentval;
+ unit = "pheight%" + endlen;
+ } else if (unit == "\\textheight") {
+ valstring = percentval;
+ unit = "theight%" + endlen;
+ }
+ return true;
+}
+
+}
+
+
+string translate_len(string const & length)
+{
+ string unit;
+ string value;
+ if (translate_len(length, value, unit))
+ return value + unit;
+ // If the input is invalid, return what we have.
+ return length;
+}
+
+
+namespace {
+
+/*!
+ * Translates a LaTeX length into \p value, \p unit and
+ * \p special parts suitable for a box inset.
+ * The difference from translate_len() is that a box inset knows about
+ * some special "units" that are stored in \p special.
+ */
+void translate_box_len(string const & length, string & value, string & unit, string & special)
+{
+ if (translate_len(length, value, unit)) {
+ if (unit == "\\height" || unit == "\\depth" ||
+ unit == "\\totalheight" || unit == "\\width") {
+ special = unit.substr(1);
+ // The unit is not used, but LyX requires a dummy setting
+ unit = "in";
+ } else
+ special = "none";
+ } else {
+ value.clear();
+ unit = length;
+ special = "none";
+ }
+}
+
+
+/*!
+ * Find a file with basename \p name in path \p path and an extension
+ * in \p extensions.
+ */
+string find_file(string const & name, string const & path,
+ char const * const * extensions)
+{
+ // FIXME UNICODE encoding of name and path may be wrong (makeAbsPath
+ // expects utf8)
+ for (char const * const * what = extensions; *what; ++what) {
+ string const trial = addExtension(name, *what);
+ if (fs::exists(makeAbsPath(trial, path).toFilesystemEncoding()))
+ return trial;
+ }
+ return string();
+}
+
+
void begin_inset(ostream & os, string const & name)
{
os << "\n\\begin_inset " << name;
void end_inset(ostream & os)
{
- os << "\n\\end_inset \n\n";
+ os << "\n\\end_inset\n\n";
}
}
-void handle_ert(ostream & os, string const & s)
+void handle_ert(ostream & os, string const & s, Context & context)
{
+ // We must have a valid layout before outputting the ERT inset.
+ context.check_layout(os);
+ Context newcontext(true, context.textclass);
begin_inset(os, "ERT");
- os << "\nstatus Collapsed\n\n\\layout Standard\n\n";
+ os << "\nstatus collapsed\n";
+ newcontext.check_layout(os);
for (string::const_iterator it = s.begin(), et = s.end(); it != et; ++it) {
if (*it == '\\')
- os << "\n\\backslash \n";
- else
+ os << "\n\\backslash\n";
+ else if (*it == '\n') {
+ newcontext.new_paragraph(os);
+ newcontext.check_layout(os);
+ } else
os << *it;
}
+ newcontext.check_end_layout(os);
end_inset(os);
}
-void handle_par(ostream & os)
+void handle_comment(ostream & os, string const & s, Context & context)
{
- if (active_environments.empty())
- return;
- os << "\n\\layout ";
- string s = active_environment();
- if (s == "document" || s == "table")
- os << "Standard\n\n";
- else if (s == "center")
- os << "Standard\n\n\\align center\n";
- else if (s == "lyxcode")
- os << "LyX-Code\n\n";
- else if (s == "lyxlist")
- os << "List\n\n";
- else if (s == "thebibliography")
- os << "Bibliography\n\n";
- else
- os << cap(s) << "\n\n";
+ // TODO: Handle this better
+ Context newcontext(true, context.textclass);
+ begin_inset(os, "ERT");
+ os << "\nstatus collapsed\n";
+ newcontext.check_layout(os);
+ for (string::const_iterator it = s.begin(), et = s.end(); it != et; ++it) {
+ if (*it == '\\')
+ os << "\n\\backslash\n";
+ else
+ os << *it;
+ }
+ // make sure that our comment is the last thing on the line
+ newcontext.new_paragraph(os);
+ newcontext.check_layout(os);
+ newcontext.check_end_layout(os);
+ end_inset(os);
}
-struct isLayout {
+class isLayout : public std::unary_function<LyXLayout_ptr, bool> {
+public:
isLayout(string const name) : name_(name) {}
- bool operator()(LyXLayout_ptr const & ptr) {
- return ptr.get() && ptr->latexname() == name_;
+ bool operator()(LyXLayout_ptr const & ptr) const {
+ return ptr->latexname() == name_;
}
private:
string const name_;
LyXLayout_ptr findLayout(LyXTextClass const & textclass,
- string const & name)
+ string const & name)
{
- LyXTextClass::const_iterator it = textclass.begin();
+ LyXTextClass::const_iterator beg = textclass.begin();
LyXTextClass::const_iterator end = textclass.end();
- it = std::find_if(it, end, isLayout(name));
+
+ LyXTextClass::const_iterator
+ it = std::find_if(beg, end, isLayout(name));
+
return (it == end) ? LyXLayout_ptr() : *it;
}
-void output_layout(ostream & os, LyXLayout_ptr const & layout_ptr,
- Parser & p, bool outer, LyXTextClass const & textclass)
+void eat_whitespace(Parser &, ostream &, Context &, bool);
+
+
+void output_command_layout(ostream & os, Parser & p, bool outer,
+ Context & parent_context,
+ LyXLayout_ptr newlayout)
{
- string name = layout_ptr->name();
- os << "\n\n\\layout " << name << "\n\n";
- if (layout_ptr->optionalargs > 0) {
- string s;
+ parent_context.check_end_layout(os);
+ Context context(true, parent_context.textclass, newlayout,
+ parent_context.layout, parent_context.font);
+ if (parent_context.deeper_paragraph) {
+ // We are beginning a nested environment after a
+ // deeper paragraph inside the outer list environment.
+ // Therefore we don't need to output a "begin deeper".
+ context.need_end_deeper = true;
+ }
+ context.check_deeper(os);
+ context.check_layout(os);
+ if (context.layout->optionalargs > 0) {
+ eat_whitespace(p, os, context, false);
if (p.next_token().character() == '[') {
p.get_token(); // eat '['
begin_inset(os, "OptArg\n");
- os << "collapsed true\n\n\\layout Standard\n\n";
- parse_text(p, os, FLAG_BRACK_LAST, outer, textclass);
+ os << "status collapsed\n\n";
+ parse_text_in_inset(p, os, FLAG_BRACK_LAST, outer, context);
end_inset(os);
+ eat_whitespace(p, os, context, false);
+ }
+ }
+ parse_text(p, os, FLAG_ITEM, outer, context);
+ context.check_end_layout(os);
+ if (parent_context.deeper_paragraph) {
+ // We must suppress the "end deeper" because we
+ // suppressed the "begin deeper" above.
+ context.need_end_deeper = false;
+ }
+ context.check_end_deeper(os);
+ // We don't need really a new paragraph, but
+ // we must make sure that the next item gets a \begin_layout.
+ parent_context.new_paragraph(os);
+}
+
+
+/*!
+ * Output a space if necessary.
+ * This function gets called for every whitespace token.
+ *
+ * We have three cases here:
+ * 1. A space must be suppressed. Example: The lyxcode case below
+ * 2. A space may be suppressed. Example: Spaces before "\par"
+ * 3. A space must not be suppressed. Example: A space between two words
+ *
+ * We currently handle only 1. and 3 and from 2. only the case of
+ * spaces before newlines as a side effect.
+ *
+ * 2. could be used to suppress as many spaces as possible. This has two effects:
+ * - Reimporting LyX generated LaTeX files changes almost no whitespace
+ * - Superflous whitespace from non LyX generated LaTeX files is removed.
+ * The drawback is that the logic inside the function becomes
+ * complicated, and that is the reason why it is not implemented.
+ */
+void check_space(Parser const & p, ostream & os, Context & context)
+{
+ Token const next = p.next_token();
+ Token const curr = p.curr_token();
+ // A space before a single newline and vice versa must be ignored
+ // LyX emits a newline before \end{lyxcode}.
+ // This newline must be ignored,
+ // otherwise LyX will add an additional protected space.
+ if (next.cat() == catSpace ||
+ next.cat() == catNewline ||
+ (next.cs() == "end" && context.layout->free_spacing && curr.cat() == catNewline)) {
+ return;
+ }
+ context.check_layout(os);
+ os << ' ';
+}
+
+
+/*!
+ * Parse all arguments of \p command
+ */
+void parse_arguments(string const & command,
+ vector<ArgumentType> const & template_arguments,
+ Parser & p, ostream & os, bool outer, Context & context)
+{
+ string ert = command;
+ size_t no_arguments = template_arguments.size();
+ for (size_t i = 0; i < no_arguments; ++i) {
+ switch (template_arguments[i]) {
+ case required:
+ // This argument contains regular LaTeX
+ handle_ert(os, ert + '{', context);
+ eat_whitespace(p, os, context, false);
+ parse_text(p, os, FLAG_ITEM, outer, context);
+ ert = "}";
+ break;
+ case verbatim:
+ // This argument may contain special characters
+ ert += '{' + p.verbatim_item() + '}';
+ break;
+ case optional:
+ ert += p.getOpt();
+ break;
+ }
+ }
+ handle_ert(os, ert, context);
+}
+
+
+/*!
+ * Check whether \p command is a known command. If yes,
+ * handle the command with all arguments.
+ * \return true if the command was parsed, false otherwise.
+ */
+bool parse_command(string const & command, Parser & p, ostream & os,
+ bool outer, Context & context)
+{
+ if (known_commands.find(command) != known_commands.end()) {
+ parse_arguments(command, known_commands[command], p, os,
+ outer, context);
+ return true;
+ }
+ return false;
+}
+
+
+/// Parses a minipage or parbox
+void parse_box(Parser & p, ostream & os, unsigned flags, bool outer,
+ Context & parent_context, bool use_parbox)
+{
+ string position;
+ string inner_pos;
+ string height_value = "0";
+ string height_unit = "pt";
+ string height_special = "none";
+ string latex_height;
+ if (p.next_token().asInput() == "[") {
+ position = p.getArg('[', ']');
+ if (position != "t" && position != "c" && position != "b") {
+ position = "c";
+ cerr << "invalid position for minipage/parbox" << endl;
+ }
+ if (p.next_token().asInput() == "[") {
+ latex_height = p.getArg('[', ']');
+ translate_box_len(latex_height, height_value, height_unit, height_special);
+
+ if (p.next_token().asInput() == "[") {
+ inner_pos = p.getArg('[', ']');
+ if (inner_pos != "c" && inner_pos != "t" &&
+ inner_pos != "b" && inner_pos != "s") {
+ inner_pos = position;
+ cerr << "invalid inner_pos for minipage/parbox"
+ << endl;
+ }
+ }
+ }
+ }
+ string width_value;
+ string width_unit;
+ string const latex_width = p.verbatim_item();
+ translate_len(latex_width, width_value, width_unit);
+ if (contains(width_unit, '\\') || contains(height_unit, '\\')) {
+ // LyX can't handle length variables
+ ostringstream ss;
+ if (use_parbox)
+ ss << "\\parbox";
+ else
+ ss << "\\begin{minipage}";
+ if (!position.empty())
+ ss << '[' << position << ']';
+ if (!latex_height.empty())
+ ss << '[' << latex_height << ']';
+ if (!inner_pos.empty())
+ ss << '[' << inner_pos << ']';
+ ss << "{" << latex_width << "}";
+ if (use_parbox)
+ ss << '{';
+ handle_ert(os, ss.str(), parent_context);
+ parent_context.new_paragraph(os);
+ parse_text_in_inset(p, os, flags, outer, parent_context);
+ if (use_parbox)
+ handle_ert(os, "}", parent_context);
+ else
+ handle_ert(os, "\\end{minipage}", parent_context);
+ } else {
+ // LyX does not like empty positions, so we have
+ // to set them to the LaTeX default values here.
+ if (position.empty())
+ position = "c";
+ if (inner_pos.empty())
+ inner_pos = position;
+ parent_context.check_layout(os);
+ begin_inset(os, "Box Frameless\n");
+ os << "position \"" << position << "\"\n";
+ os << "hor_pos \"c\"\n";
+ os << "has_inner_box 1\n";
+ os << "inner_pos \"" << inner_pos << "\"\n";
+ os << "use_parbox " << use_parbox << "\n";
+ os << "width \"" << width_value << width_unit << "\"\n";
+ os << "special \"none\"\n";
+ os << "height \"" << height_value << height_unit << "\"\n";
+ os << "height_special \"" << height_special << "\"\n";
+ os << "status open\n\n";
+ parse_text_in_inset(p, os, flags, outer, parent_context);
+ end_inset(os);
+#ifdef PRESERVE_LAYOUT
+ // lyx puts a % after the end of the minipage
+ if (p.next_token().cat() == catNewline && p.next_token().cs().size() > 1) {
+ // new paragraph
+ //handle_comment(os, "%dummy", parent_context);
+ p.get_token();
+ p.skip_spaces();
+ parent_context.new_paragraph(os);
+ }
+ else if (p.next_token().cat() == catSpace || p.next_token().cat() == catNewline) {
+ //handle_comment(os, "%dummy", parent_context);
+ p.get_token();
+ p.skip_spaces();
+ // We add a protected space if something real follows
+ if (p.good() && p.next_token().cat() != catComment) {
+ os << "\\InsetSpace ~\n";
+ }
}
+#endif
}
- parse_text(p, os, FLAG_ITEM, outer, textclass);
- os << "\n\n\\layout Standard\n\n";
+}
+
+
+/// parse an unknown environment
+void parse_unknown_environment(Parser & p, string const & name, ostream & os,
+ unsigned flags, bool outer,
+ Context & parent_context)
+{
+ if (name == "tabbing")
+ // We need to remember that we have to handle '\=' specially
+ flags |= FLAG_TABBING;
+
+ // We need to translate font changes and paragraphs inside the
+ // environment to ERT if we have a non standard font.
+ // Otherwise things like
+ // \large\begin{foo}\huge bar\end{foo}
+ // will not work.
+ bool const specialfont =
+ (parent_context.font != parent_context.normalfont);
+ bool const new_layout_allowed = parent_context.new_layout_allowed;
+ if (specialfont)
+ parent_context.new_layout_allowed = false;
+ handle_ert(os, "\\begin{" + name + "}", parent_context);
+ parse_text_snippet(p, os, flags, outer, parent_context);
+ handle_ert(os, "\\end{" + name + "}", parent_context);
+ if (specialfont)
+ parent_context.new_layout_allowed = new_layout_allowed;
+}
+
+
+void parse_environment(Parser & p, ostream & os, bool outer,
+ Context & parent_context)
+{
+ LyXLayout_ptr newlayout;
+ string const name = p.getArg('{', '}');
+ const bool is_starred = suffixIs(name, '*');
+ string const unstarred_name = rtrim(name, "*");
+ active_environments.push_back(name);
+
+ if (is_math_env(name)) {
+ parent_context.check_layout(os);
+ begin_inset(os, "Formula ");
+ os << "\\begin{" << name << "}";
+ parse_math(p, os, FLAG_END, MATH_MODE);
+ os << "\\end{" << name << "}";
+ end_inset(os);
+ }
+
+ else if (name == "tabular" || name == "longtable") {
+ eat_whitespace(p, os, parent_context, false);
+ parent_context.check_layout(os);
+ begin_inset(os, "Tabular ");
+ handle_tabular(p, os, name == "longtable", parent_context);
+ end_inset(os);
+ p.skip_spaces();
+ }
+
+ else if (parent_context.textclass.floats().typeExist(unstarred_name)) {
+ eat_whitespace(p, os, parent_context, false);
+ parent_context.check_layout(os);
+ begin_inset(os, "Float " + unstarred_name + "\n");
+ if (p.next_token().asInput() == "[") {
+ os << "placement " << p.getArg('[', ']') << '\n';
+ }
+ os << "wide " << convert<string>(is_starred)
+ << "\nsideways false"
+ << "\nstatus open\n\n";
+ parse_text_in_inset(p, os, FLAG_END, outer, parent_context);
+ end_inset(os);
+ // We don't need really a new paragraph, but
+ // we must make sure that the next item gets a \begin_layout.
+ parent_context.new_paragraph(os);
+ p.skip_spaces();
+ }
+
+ else if (name == "minipage") {
+ eat_whitespace(p, os, parent_context, false);
+ parse_box(p, os, FLAG_END, outer, parent_context, false);
+ p.skip_spaces();
+ }
+
+ else if (name == "comment") {
+ eat_whitespace(p, os, parent_context, false);
+ parent_context.check_layout(os);
+ begin_inset(os, "Note Comment\n");
+ os << "status open\n";
+ parse_text_in_inset(p, os, FLAG_END, outer, parent_context);
+ end_inset(os);
+ p.skip_spaces();
+ }
+
+ else if (name == "lyxgreyedout") {
+ eat_whitespace(p, os, parent_context, false);
+ parent_context.check_layout(os);
+ begin_inset(os, "Note Greyedout\n");
+ os << "status open\n";
+ parse_text_in_inset(p, os, FLAG_END, outer, parent_context);
+ end_inset(os);
+ p.skip_spaces();
+ }
+
+ else if (!parent_context.new_layout_allowed)
+ parse_unknown_environment(p, name, os, FLAG_END, outer,
+ parent_context);
+
+ // Alignment settings
+ else if (name == "center" || name == "flushleft" || name == "flushright" ||
+ name == "centering" || name == "raggedright" || name == "raggedleft") {
+ eat_whitespace(p, os, parent_context, false);
+ // We must begin a new paragraph if not already done
+ if (! parent_context.atParagraphStart()) {
+ parent_context.check_end_layout(os);
+ parent_context.new_paragraph(os);
+ }
+ if (name == "flushleft" || name == "raggedright")
+ parent_context.add_extra_stuff("\\align left\n");
+ else if (name == "flushright" || name == "raggedleft")
+ parent_context.add_extra_stuff("\\align right\n");
+ else
+ parent_context.add_extra_stuff("\\align center\n");
+ parse_text(p, os, FLAG_END, outer, parent_context);
+ // Just in case the environment is empty ..
+ parent_context.extra_stuff.erase();
+ // We must begin a new paragraph to reset the alignment
+ parent_context.new_paragraph(os);
+ p.skip_spaces();
+ }
+
+ // The single '=' is meant here.
+ else if ((newlayout = findLayout(parent_context.textclass, name)).get() &&
+ newlayout->isEnvironment()) {
+ eat_whitespace(p, os, parent_context, false);
+ Context context(true, parent_context.textclass, newlayout,
+ parent_context.layout, parent_context.font);
+ if (parent_context.deeper_paragraph) {
+ // We are beginning a nested environment after a
+ // deeper paragraph inside the outer list environment.
+ // Therefore we don't need to output a "begin deeper".
+ context.need_end_deeper = true;
+ }
+ parent_context.check_end_layout(os);
+ switch (context.layout->latextype) {
+ case LATEX_LIST_ENVIRONMENT:
+ context.extra_stuff = "\\labelwidthstring "
+ + p.verbatim_item() + '\n';
+ p.skip_spaces();
+ break;
+ case LATEX_BIB_ENVIRONMENT:
+ p.verbatim_item(); // swallow next arg
+ p.skip_spaces();
+ break;
+ default:
+ break;
+ }
+ context.check_deeper(os);
+ parse_text(p, os, FLAG_END, outer, context);
+ context.check_end_layout(os);
+ if (parent_context.deeper_paragraph) {
+ // We must suppress the "end deeper" because we
+ // suppressed the "begin deeper" above.
+ context.need_end_deeper = false;
+ }
+ context.check_end_deeper(os);
+ parent_context.new_paragraph(os);
+ p.skip_spaces();
+ }
+
+ else if (name == "appendix") {
+ // This is no good latex style, but it works and is used in some documents...
+ eat_whitespace(p, os, parent_context, false);
+ parent_context.check_end_layout(os);
+ Context context(true, parent_context.textclass, parent_context.layout,
+ parent_context.layout, parent_context.font);
+ context.check_layout(os);
+ os << "\\start_of_appendix\n";
+ parse_text(p, os, FLAG_END, outer, context);
+ context.check_end_layout(os);
+ p.skip_spaces();
+ }
+
+ else if (known_environments.find(name) != known_environments.end()) {
+ vector<ArgumentType> arguments = known_environments[name];
+ // The last "argument" denotes wether we may translate the
+ // environment contents to LyX
+ // The default required if no argument is given makes us
+ // compatible with the reLyXre environment.
+ ArgumentType contents = arguments.empty() ?
+ required :
+ arguments.back();
+ if (!arguments.empty())
+ arguments.pop_back();
+ // See comment in parse_unknown_environment()
+ bool const specialfont =
+ (parent_context.font != parent_context.normalfont);
+ bool const new_layout_allowed =
+ parent_context.new_layout_allowed;
+ if (specialfont)
+ parent_context.new_layout_allowed = false;
+ parse_arguments("\\begin{" + name + "}", arguments, p, os,
+ outer, parent_context);
+ if (contents == verbatim)
+ handle_ert(os, p.verbatimEnvironment(name),
+ parent_context);
+ else
+ parse_text_snippet(p, os, FLAG_END, outer,
+ parent_context);
+ handle_ert(os, "\\end{" + name + "}", parent_context);
+ if (specialfont)
+ parent_context.new_layout_allowed = new_layout_allowed;
+ }
+
+ else
+ parse_unknown_environment(p, name, os, FLAG_END, outer,
+ parent_context);
+
+ active_environments.pop_back();
+}
+
+
+/// parses a comment and outputs it to \p os.
+void parse_comment(Parser & p, ostream & os, Token const & t, Context & context)
+{
+ BOOST_ASSERT(t.cat() == catComment);
+ if (!t.cs().empty()) {
+ context.check_layout(os);
+ handle_comment(os, '%' + t.cs(), context);
+ if (p.next_token().cat() == catNewline) {
+ // A newline after a comment line starts a new
+ // paragraph
+ if (context.new_layout_allowed) {
+ if(!context.atParagraphStart())
+ // Only start a new paragraph if not already
+ // done (we might get called recursively)
+ context.new_paragraph(os);
+ } else
+ handle_ert(os, "\n", context);
+ eat_whitespace(p, os, context, true);
+ }
+ } else {
+ // "%\n" combination
+ p.skip_spaces();
+ }
+}
+
+
+/*!
+ * Reads spaces and comments until the first non-space, non-comment token.
+ * New paragraphs (double newlines or \\par) are handled like simple spaces
+ * if \p eatParagraph is true.
+ * Spaces are skipped, but comments are written to \p os.
+ */
+void eat_whitespace(Parser & p, ostream & os, Context & context,
+ bool eatParagraph)
+{
+ while (p.good()) {
+ Token const & t = p.get_token();
+ if (t.cat() == catComment)
+ parse_comment(p, os, t, context);
+ else if ((! eatParagraph && p.isParagraph()) ||
+ (t.cat() != catSpace && t.cat() != catNewline)) {
+ p.putback();
+ return;
+ }
+ }
+}
+
+
+/*!
+ * Set a font attribute, parse text and reset the font attribute.
+ * \param attribute Attribute name (e.g. \\family, \\shape etc.)
+ * \param currentvalue Current value of the attribute. Is set to the new
+ * value during parsing.
+ * \param newvalue New value of the attribute
+ */
+void parse_text_attributes(Parser & p, ostream & os, unsigned flags, bool outer,
+ Context & context, string const & attribute,
+ string & currentvalue, string const & newvalue)
+{
+ context.check_layout(os);
+ string const oldvalue = currentvalue;
+ currentvalue = newvalue;
+ os << '\n' << attribute << ' ' << newvalue << "\n";
+ parse_text_snippet(p, os, flags, outer, context);
+ context.check_layout(os);
+ os << '\n' << attribute << ' ' << oldvalue << "\n";
+ currentvalue = oldvalue;
+}
+
+
+/// get the arguments of a natbib or jurabib citation command
+std::pair<string, string> getCiteArguments(Parser & p, bool natbibOrder)
+{
+ // We need to distinguish "" and "[]", so we can't use p.getOpt().
+
+ // text before the citation
+ string before;
+ // text after the citation
+ string after = p.getFullOpt();
+
+ if (!after.empty()) {
+ before = p.getFullOpt();
+ if (natbibOrder && !before.empty())
+ std::swap(before, after);
+ }
+ return std::make_pair(before, after);
+}
+
+
+/// Convert filenames with TeX macros and/or quotes to something LyX can understand
+string const normalize_filename(string const & name)
+{
+ Parser p(trim(name, "\""));
+ ostringstream os;
+ while (p.good()) {
+ Token const & t = p.get_token();
+ if (t.cat() != catEscape)
+ os << t.asInput();
+ else if (t.cs() == "lyxdot") {
+ // This is used by LyX for simple dots in relative
+ // names
+ os << '.';
+ p.skip_spaces();
+ } else if (t.cs() == "space") {
+ os << ' ';
+ p.skip_spaces();
+ } else
+ os << t.asInput();
+ }
+ return os.str();
+}
+
+
+/// Convert \p name from TeX convention (relative to master file) to LyX
+/// convention (relative to .lyx file) if it is relative
+void fix_relative_filename(string & name)
+{
+ if (lyx::support::absolutePath(name))
+ return;
+ // FIXME UNICODE encoding of name may be wrong (makeAbsPath expects
+ // utf8)
+ name = makeRelPath(makeAbsPath(name, getMasterFilePath()).absFilename(),
+ getParentFilePath());
+}
+
+
+/// Parse a NoWeb Scrap section. The initial "<<" is already parsed.
+void parse_noweb(Parser & p, ostream & os, Context & context)
+{
+ // assemble the rest of the keyword
+ string name("<<");
+ bool scrap = false;
+ while (p.good()) {
+ Token const & t = p.get_token();
+ if (t.asInput() == ">" && p.next_token().asInput() == ">") {
+ name += ">>";
+ p.get_token();
+ scrap = (p.good() && p.next_token().asInput() == "=");
+ if (scrap)
+ name += p.get_token().asInput();
+ break;
+ }
+ name += t.asInput();
+ }
+
+ if (!scrap || !context.new_layout_allowed ||
+ !context.textclass.hasLayout("Scrap")) {
+ cerr << "Warning: Could not interpret '" << name
+ << "'. Ignoring it." << endl;
+ return;
+ }
+
+ // We use new_paragraph instead of check_end_layout because the stuff
+ // following the noweb chunk needs to start with a \begin_layout.
+ // This may create a new paragraph even if there was none in the
+ // noweb file, but the alternative is an invalid LyX file. Since
+ // noweb code chunks are implemented with a layout style in LyX they
+ // always must be in an own paragraph.
+ context.new_paragraph(os);
+ Context newcontext(true, context.textclass, context.textclass["Scrap"]);
+ newcontext.check_layout(os);
+ os << name;
+ while (p.good()) {
+ Token const & t = p.get_token();
+ // We abuse the parser a bit, because this is no TeX syntax
+ // at all.
+ if (t.cat() == catEscape)
+ os << subst(t.asInput(), "\\", "\n\\backslash\n");
+ else
+ os << subst(t.asInput(), "\n", "\n\\newline\n");
+ // The scrap chunk is ended by an @ at the beginning of a line.
+ // After the @ the line may contain a comment and/or
+ // whitespace, but nothing else.
+ if (t.asInput() == "@" && p.prev_token().cat() == catNewline &&
+ (p.next_token().cat() == catSpace ||
+ p.next_token().cat() == catNewline ||
+ p.next_token().cat() == catComment)) {
+ while (p.good() && p.next_token().cat() == catSpace)
+ os << p.get_token().asInput();
+ if (p.next_token().cat() == catComment)
+ // The comment includes a final '\n'
+ os << p.get_token().asInput();
+ else {
+ if (p.next_token().cat() == catNewline)
+ p.get_token();
+ os << '\n';
+ }
+ break;
+ }
+ }
+ newcontext.check_end_layout(os);
}
} // anonymous namespace
void parse_text(Parser & p, ostream & os, unsigned flags, bool outer,
- LyXTextClass const & textclass)
+ Context & context)
{
+ LyXLayout_ptr newlayout;
+ // Store the latest bibliographystyle (needed for bibtex inset)
+ string bibliographystyle;
+ bool const use_natbib = used_packages.find("natbib") != used_packages.end();
+ bool const use_jurabib = used_packages.find("jurabib") != used_packages.end();
while (p.good()) {
- LyXLayout_ptr layout_ptr;
Token const & t = p.get_token();
#ifdef FILEDEBUG
flags |= FLAG_LEAVE;
}
- if (t.character() == ']' && (flags & FLAG_BRACK_LAST))
+ if (t.character() == ']' && (flags & FLAG_BRACK_LAST))
return;
//
//
if (t.cat() == catMath) {
// we are inside some text mode thingy, so opening new math is allowed
+ context.check_layout(os);
begin_inset(os, "Formula ");
Token const & n = p.get_token();
if (n.cat() == catMath && outer) {
else if (t.cat() == catSuper || t.cat() == catSub)
cerr << "catcode " << t << " illegal in text mode\n";
+ // Basic support for english quotes. This should be
+ // extended to other quotes, but is not so easy (a
+ // left english quote is the same as a right german
+ // quote...)
+ else if (t.asInput() == "`"
+ && p.next_token().asInput() == "`") {
+ context.check_layout(os);
+ begin_inset(os, "Quotes ");
+ os << "eld";
+ end_inset(os);
+ p.get_token();
+ skip_braces(p);
+ }
+ else if (t.asInput() == "'"
+ && p.next_token().asInput() == "'") {
+ context.check_layout(os);
+ begin_inset(os, "Quotes ");
+ os << "erd";
+ end_inset(os);
+ p.get_token();
+ skip_braces(p);
+ }
+
+ else if (t.asInput() == "<"
+ && p.next_token().asInput() == "<" && noweb_mode) {
+ p.get_token();
+ parse_noweb(p, os, context);
+ }
+
+ else if (t.cat() == catSpace || (t.cat() == catNewline && ! p.isParagraph()))
+ check_space(p, os, context);
+
+ else if (t.character() == '[' && noweb_mode &&
+ p.next_token().character() == '[') {
+ // These can contain underscores
+ p.putback();
+ string const s = p.getFullOpt() + ']';
+ if (p.next_token().character() == ']')
+ p.get_token();
+ else
+ cerr << "Warning: Inserting missing ']' in '"
+ << s << "'." << endl;
+ handle_ert(os, s, context);
+ }
+
else if (t.cat() == catLetter ||
- t.cat() == catSpace ||
t.cat() == catOther ||
t.cat() == catAlign ||
- t.cat() == catParameter)
+ t.cat() == catParameter) {
+ // This translates "&" to "\\&" which may be wrong...
+ context.check_layout(os);
os << t.character();
+ }
- else if (t.cat() == catNewline) {
- if (p.next_token().cat() == catNewline) {
- p.get_token();
- handle_par(os);
- } else {
- os << " "; // note the space
- }
+ else if (p.isParagraph()) {
+ if (context.new_layout_allowed)
+ context.new_paragraph(os);
+ else
+ handle_ert(os, "\\par ", context);
+ eat_whitespace(p, os, context, true);
}
else if (t.cat() == catActive) {
+ context.check_layout(os);
if (t.character() == '~') {
- if (active_environment() == "lyxcode")
+ if (context.layout->free_spacing)
os << ' ';
- else
+ else
os << "\\InsetSpace ~\n";
} else
os << t.character();
}
+ else if (t.cat() == catBegin &&
+ p.next_token().cat() == catEnd) {
+ // {}
+ Token const prev = p.prev_token();
+ p.get_token();
+ if (p.next_token().character() == '`' ||
+ (prev.character() == '-' &&
+ p.next_token().character() == '-'))
+ ; // ignore it in {}`` or -{}-
+ else
+ handle_ert(os, "{}", context);
+
+ }
+
else if (t.cat() == catBegin) {
- // special handling of size changes
- bool const is_size = is_known(p.next_token().cs(), known_sizes);
- string const s = parse_text(p, FLAG_BRACE_LAST, outer, textclass);
- if (s.empty() && p.next_token().character() == '`')
- ; // ignore it in {}``
- else if (is_size || s == "[" || s == "]" || s == "*")
- os << s;
- else {
- handle_ert(os, "{");
- os << s;
- handle_ert(os, "}");
+ context.check_layout(os);
+ // special handling of font attribute changes
+ Token const prev = p.prev_token();
+ Token const next = p.next_token();
+ Font const oldFont = context.font;
+ if (next.character() == '[' ||
+ next.character() == ']' ||
+ next.character() == '*') {
+ p.get_token();
+ if (p.next_token().cat() == catEnd) {
+ os << next.character();
+ p.get_token();
+ } else {
+ p.putback();
+ handle_ert(os, "{", context);
+ parse_text_snippet(p, os,
+ FLAG_BRACE_LAST,
+ outer, context);
+ handle_ert(os, "}", context);
+ }
+ } else if (! context.new_layout_allowed) {
+ handle_ert(os, "{", context);
+ parse_text_snippet(p, os, FLAG_BRACE_LAST,
+ outer, context);
+ handle_ert(os, "}", context);
+ } else if (is_known(next.cs(), known_sizes)) {
+ // next will change the size, so we must
+ // reset it here
+ parse_text_snippet(p, os, FLAG_BRACE_LAST,
+ outer, context);
+ if (!context.atParagraphStart())
+ os << "\n\\size "
+ << context.font.size << "\n";
+ } else if (is_known(next.cs(), known_font_families)) {
+ // next will change the font family, so we
+ // must reset it here
+ parse_text_snippet(p, os, FLAG_BRACE_LAST,
+ outer, context);
+ if (!context.atParagraphStart())
+ os << "\n\\family "
+ << context.font.family << "\n";
+ } else if (is_known(next.cs(), known_font_series)) {
+ // next will change the font series, so we
+ // must reset it here
+ parse_text_snippet(p, os, FLAG_BRACE_LAST,
+ outer, context);
+ if (!context.atParagraphStart())
+ os << "\n\\series "
+ << context.font.series << "\n";
+ } else if (is_known(next.cs(), known_font_shapes)) {
+ // next will change the font shape, so we
+ // must reset it here
+ parse_text_snippet(p, os, FLAG_BRACE_LAST,
+ outer, context);
+ if (!context.atParagraphStart())
+ os << "\n\\shape "
+ << context.font.shape << "\n";
+ } else if (is_known(next.cs(), known_old_font_families) ||
+ is_known(next.cs(), known_old_font_series) ||
+ is_known(next.cs(), known_old_font_shapes)) {
+ // next will change the font family, series
+ // and shape, so we must reset it here
+ parse_text_snippet(p, os, FLAG_BRACE_LAST,
+ outer, context);
+ if (!context.atParagraphStart())
+ os << "\n\\family "
+ << context.font.family
+ << "\n\\series "
+ << context.font.series
+ << "\n\\shape "
+ << context.font.shape << "\n";
+ } else {
+ handle_ert(os, "{", context);
+ parse_text_snippet(p, os, FLAG_BRACE_LAST,
+ outer, context);
+ handle_ert(os, "}", context);
}
}
else if (t.cat() == catEnd) {
- if (flags & FLAG_BRACE_LAST)
+ if (flags & FLAG_BRACE_LAST) {
return;
+ }
cerr << "stray '}' in text\n";
- handle_ert(os, "}");
+ handle_ert(os, "}", context);
}
- else if (t.cat() == catOther)
- os << string(1, t.character());
-
else if (t.cat() == catComment)
- handle_comment(p);
+ parse_comment(p, os, t, context);
//
// control sequences
//
else if (t.cs() == "(") {
+ context.check_layout(os);
begin_inset(os, "Formula");
os << " \\(";
parse_math(p, os, FLAG_SIMPLE2, MATH_MODE);
}
else if (t.cs() == "[") {
+ context.check_layout(os);
begin_inset(os, "Formula");
os << " \\[";
parse_math(p, os, FLAG_EQUATION, MATH_MODE);
end_inset(os);
}
- else if (t.cs() == "begin") {
- string const name = p.getArg('{', '}');
- const bool is_starred = suffixIs(name, '*');
- string const unstarred_name = rtrim(name, "*");
- active_environments.push_back(name);
- if (is_math_env(name)) {
- begin_inset(os, "Formula ");
- os << "\\begin{" << name << "}";
- parse_math(p, os, FLAG_END, MATH_MODE);
- os << "\\end{" << name << "}";
- end_inset(os);
- } else if (name == "tabular") {
- begin_inset(os, "Tabular ");
- handle_tabular(p, os, textclass);
- end_inset(os);
- } else if (textclass.floats().typeExist(unstarred_name)) {
- begin_inset(os, "Float " + unstarred_name + "\n");
- if (p.next_token().asInput() == "[") {
- os << "placement "
- << p.getArg('[', ']') << '\n';
- }
- os << "wide " << tostr(is_starred)
- << "\ncollapsed false\n\n"
- << "\\layout Standard\n";
- parse_text(p, os, FLAG_END, outer,
- textclass);
- end_inset(os);
- } else if (name == "center") {
- handle_par(os);
- parse_text(p, os, FLAG_END, outer,
- textclass);
- // The single '=' is meant here.
- } else if ((layout_ptr = findLayout(textclass, t.cs())).get() &&
- layout_ptr->isEnvironment()) {
- size_t const n = active_environments.size();
- string const s = active_environments[n - 2];
- bool const deeper = s == "enumerate" || s == "itemize"
- || s == "lyxlist";
- if (deeper)
- os << "\n\\begin_deeper";
- os << "\n\\layout " << layout_ptr->name()
- << "\n\n";
- switch (layout_ptr->latextype) {
- case LATEX_LIST_ENVIRONMENT:
- os << "\\labelwidthstring "
- << p.verbatim_item() << '\n';
- break;
- case LATEX_BIB_ENVIRONMENT:
- p.verbatim_item(); // swallow next arg
- break;
- default:
- break;
- }
- parse_text(p, os, FLAG_END, outer, textclass);
- if (deeper)
- os << "\n\\end_deeper\n";
- handle_par(os);
- } else {
- handle_par(os);
- parse_text(p, os, FLAG_END, outer, textclass);
- }
- }
+ else if (t.cs() == "begin")
+ parse_environment(p, os, outer, context);
else if (t.cs() == "end") {
if (flags & FLAG_END) {
if (name != active_environment())
cerr << "\\end{" + name + "} does not match \\begin{"
+ active_environment() + "}\n";
- active_environments.pop_back();
- handle_par(os);
return;
}
p.error("found 'end' unexpectedly");
}
else if (t.cs() == "item") {
- // should be done automatically by Parser::tokenize
- //p.skip_spaces();
- string s;
+ p.skip_spaces();
+ string s;
+ bool optarg = false;
if (p.next_token().character() == '[') {
p.get_token(); // eat '['
- s = parse_text(p, FLAG_BRACK_LAST, outer, textclass);
+ s = parse_text_snippet(p, FLAG_BRACK_LAST,
+ outer, context);
+ optarg = true;
+ }
+ context.set_item();
+ context.check_layout(os);
+ if (context.has_item) {
+ // An item in an unknown list-like environment
+ // FIXME: Do this in check_layout()!
+ context.has_item = false;
+ if (optarg)
+ handle_ert(os, "\\item", context);
+ else
+ handle_ert(os, "\\item ", context);
+ }
+ if (optarg) {
+ if (context.layout->labeltype != LABEL_MANUAL) {
+ // lyx does not support \item[\mybullet]
+ // in itemize environments
+ handle_ert(os, "[", context);
+ os << s;
+ handle_ert(os, "]", context);
+ } else if (!s.empty()) {
+ // The space is needed to separate the
+ // item from the rest of the sentence.
+ os << s << ' ';
+ eat_whitespace(p, os, context, false);
+ }
}
- handle_par(os);
- if (s.size())
- os << s << ' ';
+ }
+
+ else if (t.cs() == "bibitem") {
+ context.set_item();
+ context.check_layout(os);
+ os << "\\bibitem ";
+ os << p.getOpt();
+ os << '{' << p.verbatim_item() << '}' << "\n";
}
else if (t.cs() == "def") {
+ context.check_layout(os);
+ eat_whitespace(p, os, context, false);
string name = p.get_token().cs();
while (p.next_token().cat() != catBegin)
name += p.get_token().asString();
- handle_ert(os, "\\def\\" + name + '{' + p.verbatim_item() + '}');
+ handle_ert(os, "\\def\\" + name + '{' + p.verbatim_item() + '}', context);
}
- else if (t.cs() == "par") {
+ else if (t.cs() == "noindent") {
p.skip_spaces();
- if (p.next_token().cs() != "\\begin")
- handle_par(os);
- //cerr << "next token: '" << p.next_token().cs() << "'\n";
+ context.add_extra_stuff("\\noindent\n");
+ }
+
+ else if (t.cs() == "appendix") {
+ context.add_extra_stuff("\\start_of_appendix\n");
+ // We need to start a new paragraph. Otherwise the
+ // appendix in 'bla\appendix\chapter{' would start
+ // too late.
+ context.new_paragraph(os);
+ // We need to make sure that the paragraph is
+ // generated even if it is empty. Otherwise the
+ // appendix in '\par\appendix\par\chapter{' would
+ // start too late.
+ context.check_layout(os);
+ // FIXME: This is a hack to prevent paragraph
+ // deletion if it is empty. Handle this better!
+ handle_comment(os,
+ "%dummy comment inserted by tex2lyx to "
+ "ensure that this paragraph is not empty",
+ context);
+ // Both measures above may generate an additional
+ // empty paragraph, but that does not hurt, because
+ // whitespace does not matter here.
+ eat_whitespace(p, os, context, true);
}
// Must attempt to parse "Section*" before "Section".
else if ((p.next_token().asInput() == "*") &&
+ context.new_layout_allowed &&
// The single '=' is meant here.
- (layout_ptr = findLayout(textclass,
- t.cs() + '*')).get() &&
- layout_ptr->isCommand()) {
+ (newlayout = findLayout(context.textclass,
+ t.cs() + '*')).get() &&
+ newlayout->isCommand()) {
p.get_token();
- output_layout(os, layout_ptr, p, outer, textclass);
+ output_command_layout(os, p, outer, context, newlayout);
+ p.skip_spaces();
}
// The single '=' is meant here.
- else if ((layout_ptr = findLayout(textclass, t.cs())).get() &&
- layout_ptr->isCommand()) {
- output_layout(os, layout_ptr, p, outer, textclass);
+ else if (context.new_layout_allowed &&
+ (newlayout = findLayout(context.textclass, t.cs())).get() &&
+ newlayout->isCommand()) {
+ output_command_layout(os, p, outer, context, newlayout);
+ p.skip_spaces();
}
else if (t.cs() == "includegraphics") {
+ bool const clip = p.next_token().asInput() == "*";
+ if (clip)
+ p.get_token();
map<string, string> opts = split_map(p.getArg('[', ']'));
- string name = p.verbatim_item();
+ if (clip)
+ opts["clip"] = string();
+ string name = normalize_filename(p.verbatim_item());
+
+ string const path = getMasterFilePath();
+ // We want to preserve relative / absolute filenames,
+ // therefore path is only used for testing
+ // FIXME UNICODE encoding of name and path may be
+ // wrong (makeAbsPath expects utf8)
+ if (!fs::exists(makeAbsPath(name, path).toFilesystemEncoding())) {
+ // The file extension is probably missing.
+ // Now try to find it out.
+ string const dvips_name =
+ find_file(name, path,
+ known_dvips_graphics_formats);
+ string const pdftex_name =
+ find_file(name, path,
+ known_pdftex_graphics_formats);
+ if (!dvips_name.empty()) {
+ if (!pdftex_name.empty()) {
+ cerr << "This file contains the "
+ "latex snippet\n"
+ "\"\\includegraphics{"
+ << name << "}\".\n"
+ "However, files\n\""
+ << dvips_name << "\" and\n\""
+ << pdftex_name << "\"\n"
+ "both exist, so I had to make a "
+ "choice and took the first one.\n"
+ "Please move the unwanted one "
+ "someplace else and try again\n"
+ "if my choice was wrong."
+ << endl;
+ }
+ name = dvips_name;
+ } else if (!pdftex_name.empty())
+ name = pdftex_name;
+ }
+
+ // FIXME UNICODE encoding of name and path may be
+ // wrong (makeAbsPath expects utf8)
+ if (fs::exists(makeAbsPath(name, path).toFilesystemEncoding()))
+ fix_relative_filename(name);
+ else
+ cerr << "Warning: Could not find graphics file '"
+ << name << "'." << endl;
+
+ context.check_layout(os);
begin_inset(os, "Graphics ");
os << "\n\tfilename " << name << '\n';
if (opts.find("width") != opts.end())
- os << "\twidth " << opts["width"] << '\n';
+ os << "\twidth "
+ << translate_len(opts["width"]) << '\n';
if (opts.find("height") != opts.end())
- os << "\theight " << opts["height"] << '\n';
+ os << "\theight "
+ << translate_len(opts["height"]) << '\n';
+ if (opts.find("scale") != opts.end()) {
+ istringstream iss(opts["scale"]);
+ double val;
+ iss >> val;
+ val = val*100;
+ os << "\tscale " << val << '\n';
+ }
+ if (opts.find("angle") != opts.end())
+ os << "\trotateAngle "
+ << opts["angle"] << '\n';
+ if (opts.find("origin") != opts.end()) {
+ ostringstream ss;
+ string const opt = opts["origin"];
+ if (opt.find('l') != string::npos) ss << "left";
+ if (opt.find('r') != string::npos) ss << "right";
+ if (opt.find('c') != string::npos) ss << "center";
+ if (opt.find('t') != string::npos) ss << "Top";
+ if (opt.find('b') != string::npos) ss << "Bottom";
+ if (opt.find('B') != string::npos) ss << "Baseline";
+ if (!ss.str().empty())
+ os << "\trotateOrigin " << ss.str() << '\n';
+ else
+ cerr << "Warning: Ignoring unknown includegraphics origin argument '" << opt << "'\n";
+ }
+ if (opts.find("keepaspectratio") != opts.end())
+ os << "\tkeepAspectRatio\n";
+ if (opts.find("clip") != opts.end())
+ os << "\tclip\n";
+ if (opts.find("draft") != opts.end())
+ os << "\tdraft\n";
+ if (opts.find("bb") != opts.end())
+ os << "\tBoundingBox "
+ << opts["bb"] << '\n';
+ int numberOfbbOptions = 0;
+ if (opts.find("bbllx") != opts.end())
+ numberOfbbOptions++;
+ if (opts.find("bblly") != opts.end())
+ numberOfbbOptions++;
+ if (opts.find("bburx") != opts.end())
+ numberOfbbOptions++;
+ if (opts.find("bbury") != opts.end())
+ numberOfbbOptions++;
+ if (numberOfbbOptions == 4)
+ os << "\tBoundingBox "
+ << opts["bbllx"] << opts["bblly"]
+ << opts["bburx"] << opts["bbury"] << '\n';
+ else if (numberOfbbOptions > 0)
+ cerr << "Warning: Ignoring incomplete includegraphics boundingbox arguments.\n";
+ numberOfbbOptions = 0;
+ if (opts.find("natwidth") != opts.end())
+ numberOfbbOptions++;
+ if (opts.find("natheight") != opts.end())
+ numberOfbbOptions++;
+ if (numberOfbbOptions == 2)
+ os << "\tBoundingBox 0bp 0bp "
+ << opts["natwidth"] << opts["natheight"] << '\n';
+ else if (numberOfbbOptions > 0)
+ cerr << "Warning: Ignoring incomplete includegraphics boundingbox arguments.\n";
+ ostringstream special;
+ if (opts.find("hiresbb") != opts.end())
+ special << "hiresbb,";
+ if (opts.find("trim") != opts.end())
+ special << "trim,";
+ if (opts.find("viewport") != opts.end())
+ special << "viewport=" << opts["viewport"] << ',';
+ if (opts.find("totalheight") != opts.end())
+ special << "totalheight=" << opts["totalheight"] << ',';
+ if (opts.find("type") != opts.end())
+ special << "type=" << opts["type"] << ',';
+ if (opts.find("ext") != opts.end())
+ special << "ext=" << opts["ext"] << ',';
+ if (opts.find("read") != opts.end())
+ special << "read=" << opts["read"] << ',';
+ if (opts.find("command") != opts.end())
+ special << "command=" << opts["command"] << ',';
+ string s_special = special.str();
+ if (!s_special.empty()) {
+ // We had special arguments. Remove the trailing ','.
+ os << "\tspecial " << s_special.substr(0, s_special.size() - 1) << '\n';
+ }
+ // TODO: Handle the unknown settings better.
+ // Warn about invalid options.
+ // Check whether some option was given twice.
end_inset(os);
}
-
- else if (t.cs() == "footnote") {
+
+ else if (t.cs() == "footnote" ||
+ (t.cs() == "thanks" && context.layout->intitle)) {
+ p.skip_spaces();
+ context.check_layout(os);
begin_inset(os, "Foot\n");
- os << "collapsed true\n\n\\layout Standard\n\n";
- parse_text(p, os, FLAG_ITEM, false, textclass);
+ os << "status collapsed\n\n";
+ parse_text_in_inset(p, os, FLAG_ITEM, false, context);
+ end_inset(os);
+ }
+
+ else if (t.cs() == "marginpar") {
+ p.skip_spaces();
+ context.check_layout(os);
+ begin_inset(os, "Marginal\n");
+ os << "status collapsed\n\n";
+ parse_text_in_inset(p, os, FLAG_ITEM, false, context);
end_inset(os);
}
else if (t.cs() == "ensuremath") {
- string s = parse_text(p, FLAG_ITEM, false, textclass);
+ p.skip_spaces();
+ context.check_layout(os);
+ string const s = p.verbatim_item();
if (s == "±" || s == "³" || s == "²" || s == "µ")
os << s;
else
- handle_ert(os, "\\ensuremath{" + s + "}");
- }
-
- else if (t.cs() == "marginpar") {
- begin_inset(os, "Marginal\n");
- os << "collapsed true\n\n\\layout Standard\n\n";
- parse_text(p, os, FLAG_ITEM, false, textclass);
- end_inset(os);
+ handle_ert(os, "\\ensuremath{" + s + "}",
+ context);
}
else if (t.cs() == "hfill") {
+ context.check_layout(os);
os << "\n\\hfill\n";
skip_braces(p);
+ p.skip_spaces();
}
- else if (t.cs() == "makeindex" || t.cs() == "maketitle")
+ else if (t.cs() == "makeindex" || t.cs() == "maketitle") {
+ // FIXME: Somehow prevent title layouts if
+ // "maketitle" was not found
+ p.skip_spaces();
skip_braces(p); // swallow this
+ }
else if (t.cs() == "tableofcontents") {
- begin_inset(os, "LatexCommand ");
- os << '\\' << t.cs() << "{}\n";
+ p.skip_spaces();
+ context.check_layout(os);
+ begin_inset(os, "LatexCommand \\tableofcontents\n");
end_inset(os);
skip_braces(p); // swallow this
}
-
- else if (t.cs() == "textrm") {
- os << "\n\\family roman \n";
- parse_text(p, os, FLAG_ITEM, outer, textclass);
- os << "\n\\family default \n";
- }
-
- else if (t.cs() == "textsf") {
- os << "\n\\family sans \n";
- parse_text(p, os, FLAG_ITEM, outer, textclass);
- os << "\n\\family default \n";
- }
-
- else if (t.cs() == "texttt") {
- os << "\n\\family typewriter \n";
- parse_text(p, os, FLAG_ITEM, outer, textclass);
- os << "\n\\family default \n";
+ else if (t.cs() == "listoffigures") {
+ p.skip_spaces();
+ context.check_layout(os);
+ begin_inset(os, "FloatList figure\n");
+ end_inset(os);
+ skip_braces(p); // swallow this
}
- else if (t.cs() == "textit") {
- os << "\n\\shape italic \n";
- parse_text(p, os, FLAG_ITEM, outer, textclass);
- os << "\n\\shape default \n";
+ else if (t.cs() == "listoftables") {
+ p.skip_spaces();
+ context.check_layout(os);
+ begin_inset(os, "FloatList table\n");
+ end_inset(os);
+ skip_braces(p); // swallow this
}
- else if (t.cs() == "textsc") {
- os << "\n\\noun on \n";
- parse_text(p, os, FLAG_ITEM, outer, textclass);
- os << "\n\\noun default \n";
+ else if (t.cs() == "listof") {
+ p.skip_spaces(true);
+ string const name = p.get_token().asString();
+ if (context.textclass.floats().typeExist(name)) {
+ context.check_layout(os);
+ begin_inset(os, "FloatList ");
+ os << name << "\n";
+ end_inset(os);
+ p.get_token(); // swallow second arg
+ } else
+ handle_ert(os, "\\listof{" + name + "}", context);
}
- else if (t.cs() == "textbf") {
- os << "\n\\series bold \n";
- parse_text(p, os, FLAG_ITEM, outer, textclass);
- os << "\n\\series default \n";
+ else if (t.cs() == "textrm")
+ parse_text_attributes(p, os, FLAG_ITEM, outer,
+ context, "\\family",
+ context.font.family, "roman");
+
+ else if (t.cs() == "textsf")
+ parse_text_attributes(p, os, FLAG_ITEM, outer,
+ context, "\\family",
+ context.font.family, "sans");
+
+ else if (t.cs() == "texttt")
+ parse_text_attributes(p, os, FLAG_ITEM, outer,
+ context, "\\family",
+ context.font.family, "typewriter");
+
+ else if (t.cs() == "textmd")
+ parse_text_attributes(p, os, FLAG_ITEM, outer,
+ context, "\\series",
+ context.font.series, "medium");
+
+ else if (t.cs() == "textbf")
+ parse_text_attributes(p, os, FLAG_ITEM, outer,
+ context, "\\series",
+ context.font.series, "bold");
+
+ else if (t.cs() == "textup")
+ parse_text_attributes(p, os, FLAG_ITEM, outer,
+ context, "\\shape",
+ context.font.shape, "up");
+
+ else if (t.cs() == "textit")
+ parse_text_attributes(p, os, FLAG_ITEM, outer,
+ context, "\\shape",
+ context.font.shape, "italic");
+
+ else if (t.cs() == "textsl")
+ parse_text_attributes(p, os, FLAG_ITEM, outer,
+ context, "\\shape",
+ context.font.shape, "slanted");
+
+ else if (t.cs() == "textsc")
+ parse_text_attributes(p, os, FLAG_ITEM, outer,
+ context, "\\shape",
+ context.font.shape, "smallcaps");
+
+ else if (t.cs() == "textnormal" || t.cs() == "normalfont") {
+ context.check_layout(os);
+ Font oldFont = context.font;
+ context.font.init();
+ context.font.size = oldFont.size;
+ os << "\n\\family " << context.font.family << "\n";
+ os << "\n\\series " << context.font.series << "\n";
+ os << "\n\\shape " << context.font.shape << "\n";
+ if (t.cs() == "textnormal") {
+ parse_text_snippet(p, os, FLAG_ITEM, outer, context);
+ output_font_change(os, context.font, oldFont);
+ context.font = oldFont;
+ } else
+ eat_whitespace(p, os, context, false);
}
else if (t.cs() == "underbar") {
- os << "\n\\bar under \n";
- parse_text(p, os, FLAG_ITEM, outer, textclass);
- os << "\n\\bar default \n";
+ // Do NOT handle \underline.
+ // \underbar cuts through y, g, q, p etc.,
+ // \underline does not.
+ context.check_layout(os);
+ os << "\n\\bar under\n";
+ parse_text_snippet(p, os, FLAG_ITEM, outer, context);
+ context.check_layout(os);
+ os << "\n\\bar default\n";
}
else if (t.cs() == "emph" || t.cs() == "noun") {
- os << "\n\\" << t.cs() << " on \n";
- parse_text(p, os, FLAG_ITEM, outer, textclass);
- os << "\n\\" << t.cs() << " default \n";
+ context.check_layout(os);
+ os << "\n\\" << t.cs() << " on\n";
+ parse_text_snippet(p, os, FLAG_ITEM, outer, context);
+ context.check_layout(os);
+ os << "\n\\" << t.cs() << " default\n";
}
- else if (t.cs() == "bibitem") {
- os << "\n\\layout Bibliography\n\\bibitem ";
- os << p.getOpt();
- os << '{' << p.verbatim_item() << '}' << "\n";
+ else if (use_natbib &&
+ is_known(t.cs(), known_natbib_commands) &&
+ ((t.cs() != "citefullauthor" &&
+ t.cs() != "citeyear" &&
+ t.cs() != "citeyearpar") ||
+ p.next_token().asInput() != "*")) {
+ context.check_layout(os);
+ // tex lyx
+ // \citet[before][after]{a} \citet[after][before]{a}
+ // \citet[before][]{a} \citet[][before]{a}
+ // \citet[after]{a} \citet[after]{a}
+ // \citet{a} \citet{a}
+ string command = '\\' + t.cs();
+ if (p.next_token().asInput() == "*") {
+ command += '*';
+ p.get_token();
+ }
+ if (command == "\\citefullauthor")
+ // alternative name for "\\citeauthor*"
+ command = "\\citeauthor*";
+
+ // text before the citation
+ string before;
+ // text after the citation
+ string after;
+
+ boost::tie(before, after) = getCiteArguments(p, true);
+ if (command == "\\cite") {
+ // \cite without optional argument means
+ // \citet, \cite with at least one optional
+ // argument means \citep.
+ if (before.empty() && after.empty())
+ command = "\\citet";
+ else
+ command = "\\citep";
+ }
+ if (before.empty() && after == "[]")
+ // avoid \citet[]{a}
+ after.erase();
+ else if (before == "[]" && after == "[]") {
+ // avoid \citet[][]{a}
+ before.erase();
+ after.erase();
+ }
+ begin_inset(os, "LatexCommand ");
+ os << command << after << before
+ << '{' << p.verbatim_item() << "}\n";
+ end_inset(os);
+ }
+
+ else if (use_jurabib &&
+ is_known(t.cs(), known_jurabib_commands)) {
+ context.check_layout(os);
+ string const command = '\\' + t.cs();
+ char argumentOrder = '\0';
+ vector<string> const & options = used_packages["jurabib"];
+ if (std::find(options.begin(), options.end(),
+ "natbiborder") != options.end())
+ argumentOrder = 'n';
+ else if (std::find(options.begin(), options.end(),
+ "jurabiborder") != options.end())
+ argumentOrder = 'j';
+
+ // text before the citation
+ string before;
+ // text after the citation
+ string after;
+
+ boost::tie(before, after) =
+ getCiteArguments(p, argumentOrder != 'j');
+ string const citation = p.verbatim_item();
+ if (!before.empty() && argumentOrder == '\0') {
+ cerr << "Warning: Assuming argument order "
+ "of jurabib version 0.6 for\n'"
+ << command << before << after << '{'
+ << citation << "}'.\n"
+ "Add 'jurabiborder' to the jurabib "
+ "package options if you used an\n"
+ "earlier jurabib version." << endl;
+ }
+ begin_inset(os, "LatexCommand ");
+ os << command << after << before
+ << '{' << citation << "}\n";
+ end_inset(os);
}
else if (is_known(t.cs(), known_latex_commands)) {
+ // This needs to be after the check for natbib and
+ // jurabib commands, because "cite" has different
+ // arguments with natbib and jurabib.
+ context.check_layout(os);
begin_inset(os, "LatexCommand ");
os << '\\' << t.cs();
- os << p.getOpt();
- os << p.getOpt();
- os << '{' << p.verbatim_item() << "}\n";
+ // lyx cannot handle newlines in a latex command
+ // FIXME: Move the substitution into parser::getOpt()?
+ os << subst(p.getOpt(), "\n", " ");
+ os << subst(p.getOpt(), "\n", " ");
+ os << '{' << subst(p.verbatim_item(), "\n", " ") << "}\n";
end_inset(os);
}
else if (is_known(t.cs(), known_quotes)) {
- char const ** where = is_known(t.cs(), known_quotes);
+ char const * const * where = is_known(t.cs(), known_quotes);
+ context.check_layout(os);
begin_inset(os, "Quotes ");
os << known_coded_quotes[where - known_quotes];
end_inset(os);
+ // LyX adds {} after the quote, so we have to eat
+ // spaces here if there are any before a possible
+ // {} pair.
+ eat_whitespace(p, os, context, false);
skip_braces(p);
}
- else if (is_known(t.cs(), known_sizes)) {
- char const ** where = is_known(t.cs(), known_sizes);
- os << "\n\\size " << known_coded_sizes[where - known_sizes] << "\n";
+ else if (is_known(t.cs(), known_sizes) &&
+ context.new_layout_allowed) {
+ char const * const * where = is_known(t.cs(), known_sizes);
+ context.check_layout(os);
+ Font const oldFont = context.font;
+ context.font.size = known_coded_sizes[where - known_sizes];
+ output_font_change(os, oldFont, context.font);
+ eat_whitespace(p, os, context, false);
+ }
+
+ else if (is_known(t.cs(), known_font_families) &&
+ context.new_layout_allowed) {
+ char const * const * where =
+ is_known(t.cs(), known_font_families);
+ context.check_layout(os);
+ Font const oldFont = context.font;
+ context.font.family =
+ known_coded_font_families[where - known_font_families];
+ output_font_change(os, oldFont, context.font);
+ eat_whitespace(p, os, context, false);
+ }
+
+ else if (is_known(t.cs(), known_font_series) &&
+ context.new_layout_allowed) {
+ char const * const * where =
+ is_known(t.cs(), known_font_series);
+ context.check_layout(os);
+ Font const oldFont = context.font;
+ context.font.series =
+ known_coded_font_series[where - known_font_series];
+ output_font_change(os, oldFont, context.font);
+ eat_whitespace(p, os, context, false);
}
- else if (t.cs() == "LyX" || t.cs() == "TeX"
+ else if (is_known(t.cs(), known_font_shapes) &&
+ context.new_layout_allowed) {
+ char const * const * where =
+ is_known(t.cs(), known_font_shapes);
+ context.check_layout(os);
+ Font const oldFont = context.font;
+ context.font.shape =
+ known_coded_font_shapes[where - known_font_shapes];
+ output_font_change(os, oldFont, context.font);
+ eat_whitespace(p, os, context, false);
+ }
+ else if (is_known(t.cs(), known_old_font_families) &&
+ context.new_layout_allowed) {
+ char const * const * where =
+ is_known(t.cs(), known_old_font_families);
+ context.check_layout(os);
+ Font const oldFont = context.font;
+ context.font.init();
+ context.font.size = oldFont.size;
+ context.font.family =
+ known_coded_font_families[where - known_old_font_families];
+ output_font_change(os, oldFont, context.font);
+ eat_whitespace(p, os, context, false);
+ }
+
+ else if (is_known(t.cs(), known_old_font_series) &&
+ context.new_layout_allowed) {
+ char const * const * where =
+ is_known(t.cs(), known_old_font_series);
+ context.check_layout(os);
+ Font const oldFont = context.font;
+ context.font.init();
+ context.font.size = oldFont.size;
+ context.font.series =
+ known_coded_font_series[where - known_old_font_series];
+ output_font_change(os, oldFont, context.font);
+ eat_whitespace(p, os, context, false);
+ }
+
+ else if (is_known(t.cs(), known_old_font_shapes) &&
+ context.new_layout_allowed) {
+ char const * const * where =
+ is_known(t.cs(), known_old_font_shapes);
+ context.check_layout(os);
+ Font const oldFont = context.font;
+ context.font.init();
+ context.font.size = oldFont.size;
+ context.font.shape =
+ known_coded_font_shapes[where - known_old_font_shapes];
+ output_font_change(os, oldFont, context.font);
+ eat_whitespace(p, os, context, false);
+ }
+
+ else if (t.cs() == "LyX" || t.cs() == "TeX"
|| t.cs() == "LaTeX") {
+ context.check_layout(os);
os << t.cs();
skip_braces(p); // eat {}
}
else if (t.cs() == "LaTeXe") {
+ context.check_layout(os);
os << "LaTeX2e";
skip_braces(p); // eat {}
}
else if (t.cs() == "ldots") {
+ context.check_layout(os);
skip_braces(p);
os << "\\SpecialChar \\ldots{}\n";
}
else if (t.cs() == "lyxarrow") {
+ context.check_layout(os);
os << "\\SpecialChar \\menuseparator\n";
skip_braces(p);
}
- else if (t.cs() == "ldots") {
- os << "\\SpecialChar \\ldots{}\n";
+ else if (t.cs() == "textcompwordmark") {
+ context.check_layout(os);
+ os << "\\SpecialChar \\textcompwordmark{}\n";
skip_braces(p);
}
else if (t.cs() == "@" && p.next_token().asInput() == ".") {
+ context.check_layout(os);
os << "\\SpecialChar \\@.\n";
p.get_token();
}
- else if (t.cs() == "-")
+ else if (t.cs() == "-") {
+ context.check_layout(os);
os << "\\SpecialChar \\-\n";
+ }
else if (t.cs() == "textasciitilde") {
+ context.check_layout(os);
os << '~';
skip_braces(p);
}
else if (t.cs() == "textasciicircum") {
+ context.check_layout(os);
os << '^';
skip_braces(p);
}
else if (t.cs() == "textbackslash") {
- os << "\n\\backslash \n";
+ context.check_layout(os);
+ os << "\n\\backslash\n";
skip_braces(p);
}
- else if (t.cs() == "_" || t.cs() == "&" || t.cs() == "#"
- || t.cs() == "$" || t.cs() == "{" || t.cs() == "}"
- || t.cs() == "%")
+ else if (t.cs() == "_" || t.cs() == "&" || t.cs() == "#"
+ || t.cs() == "$" || t.cs() == "{" || t.cs() == "}"
+ || t.cs() == "%") {
+ context.check_layout(os);
os << t.cs();
+ }
else if (t.cs() == "char") {
+ context.check_layout(os);
if (p.next_token().character() == '`') {
p.get_token();
if (p.next_token().cs() == "\"") {
os << '"';
skip_braces(p);
} else {
- handle_ert(os, "\\char`");
+ handle_ert(os, "\\char`", context);
}
} else {
- handle_ert(os, "\\char");
+ handle_ert(os, "\\char", context);
}
}
+ else if (t.cs() == "verb") {
+ context.check_layout(os);
+ char const delimiter = p.next_token().character();
+ string const arg = p.getArg(delimiter, delimiter);
+ ostringstream oss;
+ oss << "\\verb" << delimiter << arg << delimiter;
+ handle_ert(os, oss.str(), context);
+ }
+
else if (t.cs() == "\"") {
+ context.check_layout(os);
string const name = p.verbatim_item();
if (name == "a") os << 'ä';
else if (name == "o") os << 'ö';
else if (name == "A") os << 'Ä';
else if (name == "O") os << 'Ö';
else if (name == "U") os << 'Ü';
- else handle_ert(os, "\"{" + name + "}");
+ else handle_ert(os, "\"{" + name + "}", context);
}
- else if (t.cs() == "=" || t.cs() == "H" || t.cs() == "c"
- || t.cs() == "^" || t.cs() == "'" || t.cs() == "~") {
+ // Problem: \= creates a tabstop inside the tabbing environment
+ // and else an accent. In the latter case we really would want
+ // \={o} instead of \= o.
+ else if (t.cs() == "=" && (flags & FLAG_TABBING))
+ handle_ert(os, t.asInput(), context);
+
+ else if (t.cs() == "H" || t.cs() == "c" || t.cs() == "^"
+ || t.cs() == "'" || t.cs() == "`"
+ || t.cs() == "~" || t.cs() == "." || t.cs() == "=") {
// we need the trim as the LyX parser chokes on such spaces
- os << "\n\\i \\" << t.cs() << "{"
- << trim(parse_text(p, FLAG_ITEM, outer, textclass), " ") << "}\n";
+ context.check_layout(os);
+ os << "\\i \\" << t.cs() << "{"
+ << trim(parse_text_snippet(p, FLAG_ITEM, outer, context), " ")
+ << "}\n";
}
- else if (t.cs() == "ss")
+ else if (t.cs() == "ss") {
+ context.check_layout(os);
os << "ß";
+ skip_braces(p); // eat {}
+ }
+
+ else if (t.cs() == "i" || t.cs() == "j" || t.cs() == "l" ||
+ t.cs() == "L") {
+ context.check_layout(os);
+ os << "\\i \\" << t.cs() << "{}\n";
+ skip_braces(p); // eat {}
+ }
+
+ else if (t.cs() == "\\") {
+ context.check_layout(os);
+ string const next = p.next_token().asInput();
+ if (next == "[")
+ handle_ert(os, "\\\\" + p.getOpt(), context);
+ else if (next == "*") {
+ p.get_token();
+ handle_ert(os, "\\\\*" + p.getOpt(), context);
+ }
+ else {
+ os << "\n\\newline\n";
+ }
+ }
+
+ else if (t.cs() == "input" || t.cs() == "include"
+ || t.cs() == "verbatiminput") {
+ string name = '\\' + t.cs();
+ if (t.cs() == "verbatiminput"
+ && p.next_token().asInput() == "*")
+ name += p.get_token().asInput();
+ context.check_layout(os);
+ begin_inset(os, "Include ");
+ string filename(normalize_filename(p.getArg('{', '}')));
+ string const path = getMasterFilePath();
+ // We want to preserve relative / absolute filenames,
+ // therefore path is only used for testing
+ // FIXME UNICODE encoding of filename and path may be
+ // wrong (makeAbsPath expects utf8)
+ if (t.cs() == "include" &&
+ !fs::exists(makeAbsPath(filename, path).toFilesystemEncoding())) {
+ // The file extension is probably missing.
+ // Now try to find it out.
+ string const tex_name =
+ find_file(filename, path,
+ known_tex_extensions);
+ if (!tex_name.empty())
+ filename = tex_name;
+ }
+ // FIXME UNICODE encoding of filename and path may be
+ // wrong (makeAbsPath expects utf8)
+ if (fs::exists(makeAbsPath(filename, path).toFilesystemEncoding())) {
+ string const abstexname =
+ makeAbsPath(filename, path).absFilename();
+ string const abslyxname =
+ changeExtension(abstexname, ".lyx");
+ fix_relative_filename(filename);
+ string const lyxname =
+ changeExtension(filename, ".lyx");
+ if (t.cs() != "verbatiminput" &&
+ tex2lyx(abstexname, FileName(abslyxname))) {
+ os << name << '{' << lyxname << "}\n";
+ } else {
+ os << name << '{' << filename << "}\n";
+ }
+ } else {
+ cerr << "Warning: Could not find included file '"
+ << filename << "'." << endl;
+ os << name << '{' << filename << "}\n";
+ }
+ os << "preview false\n";
+ end_inset(os);
+ }
+
+ else if (t.cs() == "bibliographystyle") {
+ // store new bibliographystyle
+ bibliographystyle = p.verbatim_item();
+ // output new bibliographystyle.
+ // This is only necessary if used in some other macro than \bibliography.
+ handle_ert(os, "\\bibliographystyle{" + bibliographystyle + "}", context);
+ }
+
+ else if (t.cs() == "bibliography") {
+ context.check_layout(os);
+ begin_inset(os, "LatexCommand ");
+ os << "\\bibtex";
+ // Do we have a bibliographystyle set?
+ if (!bibliographystyle.empty()) {
+ os << '[' << bibliographystyle << ']';
+ }
+ os << '{' << p.verbatim_item() << "}\n";
+ end_inset(os);
+ }
+
+ else if (t.cs() == "parbox")
+ parse_box(p, os, FLAG_ITEM, outer, context, true);
+
+ else if (t.cs() == "smallskip" ||
+ t.cs() == "medskip" ||
+ t.cs() == "bigskip" ||
+ t.cs() == "vfill") {
+ context.check_layout(os);
+ begin_inset(os, "VSpace ");
+ os << t.cs();
+ end_inset(os);
+ skip_braces(p);
+ }
- else if (t.cs() == "i" || t.cs() == "j")
- os << "\\" << t.cs() << ' ';
+ else if (is_known(t.cs(), known_spaces)) {
+ char const * const * where = is_known(t.cs(), known_spaces);
+ context.check_layout(os);
+ begin_inset(os, "InsetSpace ");
+ os << '\\' << known_coded_spaces[where - known_spaces]
+ << '\n';
+ // LaTeX swallows whitespace after all spaces except
+ // "\\,". We have to do that here, too, because LyX
+ // adds "{}" which would make the spaces significant.
+ if (t.cs() != ",")
+ eat_whitespace(p, os, context, false);
+ // LyX adds "{}" after all spaces except "\\ " and
+ // "\\,", so we have to remove "{}".
+ // "\\,{}" is equivalent to "\\," in LaTeX, so we
+ // remove the braces after "\\,", too.
+ if (t.cs() != " ")
+ skip_braces(p);
+ }
- else if (t.cs() == "\\")
- os << "\n\\newline \n";
-
- else if (t.cs() == "input")
- handle_ert(os, "\\input{" + p.verbatim_item() + "}\n");
+ else if (t.cs() == "newpage" ||
+ t.cs() == "clearpage" ||
+ t.cs() == "cleardoublepage") {
+ context.check_layout(os);
+ // FIXME: what about \\pagebreak?
+ os << "\n\\" << t.cs() << "\n";
+ skip_braces(p); // eat {}
+ }
- else if (t.cs() == "fancyhead") {
- ostringstream ss;
- ss << "\\fancyhead";
- ss << p.getOpt();
- ss << '{' << p.verbatim_item() << "}\n";
- handle_ert(os, ss.str());
+ else if (t.cs() == "newcommand" ||
+ t.cs() == "providecommand" ||
+ t.cs() == "renewcommand") {
+ // these could be handled by parse_command(), but
+ // we need to call add_known_command() here.
+ string name = t.asInput();
+ if (p.next_token().asInput() == "*") {
+ // Starred form. Eat '*'
+ p.get_token();
+ name += '*';
+ }
+ string const command = p.verbatim_item();
+ string const opt1 = p.getOpt();
+ string const opt2 = p.getFullOpt();
+ add_known_command(command, opt1, !opt2.empty());
+ string const ert = name + '{' + command + '}' +
+ opt1 + opt2 +
+ '{' + p.verbatim_item() + '}';
+ handle_ert(os, ert, context);
+ }
+
+ else if (t.cs() == "vspace") {
+ bool starred = false;
+ if (p.next_token().asInput() == "*") {
+ p.get_token();
+ starred = true;
+ }
+ string const length = p.verbatim_item();
+ string unit;
+ string valstring;
+ bool valid = splitLatexLength(length, valstring, unit);
+ bool known_vspace = false;
+ bool known_unit = false;
+ double value;
+ if (valid) {
+ istringstream iss(valstring);
+ iss >> value;
+ if (value == 1.0) {
+ if (unit == "\\smallskipamount") {
+ unit = "smallskip";
+ known_vspace = true;
+ } else if (unit == "\\medskipamount") {
+ unit = "medskip";
+ known_vspace = true;
+ } else if (unit == "\\bigskipamount") {
+ unit = "bigskip";
+ known_vspace = true;
+ } else if (unit == "\\fill") {
+ unit = "vfill";
+ known_vspace = true;
+ }
+ }
+ if (!known_vspace) {
+ switch (unitFromString(unit)) {
+ case LyXLength::SP:
+ case LyXLength::PT:
+ case LyXLength::BP:
+ case LyXLength::DD:
+ case LyXLength::MM:
+ case LyXLength::PC:
+ case LyXLength::CC:
+ case LyXLength::CM:
+ case LyXLength::IN:
+ case LyXLength::EX:
+ case LyXLength::EM:
+ case LyXLength::MU:
+ known_unit = true;
+ break;
+ default:
+ break;
+ }
+ }
+ }
+
+ if (known_unit || known_vspace) {
+ // Literal length or known variable
+ context.check_layout(os);
+ begin_inset(os, "VSpace ");
+ if (known_unit)
+ os << value;
+ os << unit;
+ if (starred)
+ os << '*';
+ end_inset(os);
+ } else {
+ // LyX can't handle other length variables in Inset VSpace
+ string name = t.asInput();
+ if (starred)
+ name += '*';
+ if (valid) {
+ if (value == 1.0)
+ handle_ert(os, name + '{' + unit + '}', context);
+ else if (value == -1.0)
+ handle_ert(os, name + "{-" + unit + '}', context);
+ else
+ handle_ert(os, name + '{' + valstring + unit + '}', context);
+ } else
+ handle_ert(os, name + '{' + length + '}', context);
+ }
}
else {
z = p.verbatim_item();
}
cerr << "found ERT: " << s << endl;
- handle_ert(os, s + ' ');
+ handle_ert(os, s + ' ', context);
*/
- handle_ert(os, t.asInput() + ' ');
+ string name = t.asInput();
+ if (p.next_token().asInput() == "*") {
+ // Starred commands like \vspace*{}
+ p.get_token(); // Eat '*'
+ name += '*';
+ }
+ if (! parse_command(name, p, os, outer, context))
+ handle_ert(os, name, context);
}
if (flags & FLAG_LEAVE) {
}
}
-
-string parse_text(Parser & p, unsigned flags, const bool outer,
- LyXTextClass const & textclass)
-{
- ostringstream os;
- parse_text(p, os, flags, outer, textclass);
- return os.str();
-}
+// }])
-// }])
+} // namespace lyx