#include <sstream>
#include <vector>
-using lyx::support::ChangeExtension;
-using lyx::support::MakeAbsPath;
-using lyx::support::MakeRelPath;
-using lyx::support::rtrim;
-using lyx::support::suffixIs;
-using lyx::support::contains;
-using lyx::support::subst;
+
+namespace lyx {
+
+using support::addExtension;
+using support::changeExtension;
+using support::FileName;
+using support::makeAbsPath;
+using support::makeRelPath;
+using support::rtrim;
+using support::suffixIs;
+using support::contains;
+using support::subst;
using std::cerr;
using std::endl;
* No starred form other than "cite*" known.
*/
char const * const known_jurabib_commands[] = { "cite", "citet", "citep",
-"citealt", "citealp", "citeauthor", "citeyear", "citeyearpar", "fullcite",
+"citealt", "citealp", "citeauthor", "citeyear", "citeyearpar",
// jurabib commands not (yet) supported by LyX:
+// "fullcite",
// "footcite", "footcitet", "footcitep", "footcitealt", "footcitealp",
// "footciteauthor", "footciteyear", "footciteyearpar",
"citefield", "citetitle", "cite*", 0 };
/// the same as known_quotes with .lyx names
char const * const known_coded_quotes[] = { "gld", "grd", "gld",
-"grd", "gls", "fls", "frd", 0};
+"grd", "gls", "fls", "frs", 0};
/// LaTeX names for font sizes
char const * const known_sizes[] = { "tiny", "scriptsize", "footnotesize",
*/
char const * const known_tex_extensions[] = {"tex", 0};
+/// spaces known by InsetSpace
+char const * const known_spaces[] = { " ", "space", ",", "thinspace", "quad",
+"qquad", "enspace", "enskip", "negthinspace", 0};
+
+/// the same as known_spaces with .lyx names
+char const * const known_coded_spaces[] = { "space{}", "space{}",
+"thinspace{}", "thinspace{}", "quad{}", "qquad{}", "enspace{}", "enskip{}",
+"negthinspace{}", 0};
+
/// splits "x=z, y=b" into a map
map<string, string> split_map(string const & s)
string find_file(string const & name, string const & path,
char const * const * extensions)
{
+ // FIXME UNICODE encoding of name and path may be wrong (makeAbsPath
+ // expects utf8)
for (char const * const * what = extensions; *what; ++what) {
- // We don't use ChangeExtension() because it does the wrong
- // thing if name contains a dot.
- string const trial = name + '.' + (*what);
- if (fs::exists(MakeAbsPath(trial, path)))
+ string const trial = addExtension(name, *what);
+ if (fs::exists(makeAbsPath(trial, path).toFilesystemEncoding()))
return trial;
}
return string();
LyXLayout_ptr findLayout(LyXTextClass const & textclass,
string const & name)
{
- LyXTextClass::const_iterator beg = textclass.begin();
+ LyXTextClass::const_iterator beg = textclass.begin();
LyXTextClass::const_iterator end = textclass.end();
LyXTextClass::const_iterator
case required:
// This argument contains regular LaTeX
handle_ert(os, ert + '{', context);
+ eat_whitespace(p, os, context, false);
parse_text(p, os, FLAG_ITEM, outer, context);
ert = "}";
break;
/// convention (relative to .lyx file) if it is relative
void fix_relative_filename(string & name)
{
- if (lyx::support::AbsolutePath(name))
+ if (lyx::support::absolutePath(name))
return;
- name = MakeRelPath(MakeAbsPath(name, getMasterFilePath()),
+ // FIXME UNICODE encoding of name may be wrong (makeAbsPath expects
+ // utf8)
+ name = makeRelPath(makeAbsPath(name, getMasterFilePath()).absFilename(),
getParentFilePath());
}
+
+/// Parse a NoWeb Scrap section. The initial "<<" is already parsed.
+void parse_noweb(Parser & p, ostream & os, Context & context)
+{
+ // assemble the rest of the keyword
+ string name("<<");
+ bool scrap = false;
+ while (p.good()) {
+ Token const & t = p.get_token();
+ if (t.asInput() == ">" && p.next_token().asInput() == ">") {
+ name += ">>";
+ p.get_token();
+ scrap = (p.good() && p.next_token().asInput() == "=");
+ if (scrap)
+ name += p.get_token().asInput();
+ break;
+ }
+ name += t.asInput();
+ }
+
+ if (!scrap || !context.new_layout_allowed ||
+ !context.textclass.hasLayout("Scrap")) {
+ cerr << "Warning: Could not interpret '" << name
+ << "'. Ignoring it." << endl;
+ return;
+ }
+
+ // We use new_paragraph instead of check_end_layout because the stuff
+ // following the noweb chunk needs to start with a \begin_layout.
+ // This may create a new paragraph even if there was none in the
+ // noweb file, but the alternative is an invalid LyX file. Since
+ // noweb code chunks are implemented with a layout style in LyX they
+ // always must be in an own paragraph.
+ context.new_paragraph(os);
+ Context newcontext(true, context.textclass, context.textclass["Scrap"]);
+ newcontext.check_layout(os);
+ os << name;
+ while (p.good()) {
+ Token const & t = p.get_token();
+ // We abuse the parser a bit, because this is no TeX syntax
+ // at all.
+ if (t.cat() == catEscape)
+ os << subst(t.asInput(), "\\", "\n\\backslash\n");
+ else
+ os << subst(t.asInput(), "\n", "\n\\newline\n");
+ // The scrap chunk is ended by an @ at the beginning of a line.
+ // After the @ the line may contain a comment and/or
+ // whitespace, but nothing else.
+ if (t.asInput() == "@" && p.prev_token().cat() == catNewline &&
+ (p.next_token().cat() == catSpace ||
+ p.next_token().cat() == catNewline ||
+ p.next_token().cat() == catComment)) {
+ while (p.good() && p.next_token().cat() == catSpace)
+ os << p.get_token().asInput();
+ if (p.next_token().cat() == catComment)
+ // The comment includes a final '\n'
+ os << p.get_token().asInput();
+ else {
+ if (p.next_token().cat() == catNewline)
+ p.get_token();
+ os << '\n';
+ }
+ break;
+ }
+ }
+ newcontext.check_end_layout(os);
+}
+
} // anonymous namespace
skip_braces(p);
}
+ else if (t.asInput() == "<"
+ && p.next_token().asInput() == "<" && noweb_mode) {
+ p.get_token();
+ parse_noweb(p, os, context);
+ }
+
else if (t.cat() == catSpace || (t.cat() == catNewline && ! p.isParagraph()))
check_space(p, os, context);
+ else if (t.character() == '[' && noweb_mode &&
+ p.next_token().character() == '[') {
+ // These can contain underscores
+ p.putback();
+ string const s = p.getFullOpt() + ']';
+ if (p.next_token().character() == ']')
+ p.get_token();
+ else
+ cerr << "Warning: Inserting missing ']' in '"
+ << s << "'." << endl;
+ handle_ert(os, s, context);
+ }
+
else if (t.cat() == catLetter ||
t.cat() == catOther ||
t.cat() == catAlign ||
string const path = getMasterFilePath();
// We want to preserve relative / absolute filenames,
// therefore path is only used for testing
- if (!fs::exists(MakeAbsPath(name, path))) {
+ // FIXME UNICODE encoding of name and path may be
+ // wrong (makeAbsPath expects utf8)
+ if (!fs::exists(makeAbsPath(name, path).toFilesystemEncoding())) {
// The file extension is probably missing.
// Now try to find it out.
string const dvips_name =
name = pdftex_name;
}
- if (fs::exists(MakeAbsPath(name, path)))
+ // FIXME UNICODE encoding of name and path may be
+ // wrong (makeAbsPath expects utf8)
+ if (fs::exists(makeAbsPath(name, path).toFilesystemEncoding()))
fix_relative_filename(name);
else
cerr << "Warning: Could not find graphics file '"
|| t.cs() == "~" || t.cs() == "." || t.cs() == "=") {
// we need the trim as the LyX parser chokes on such spaces
context.check_layout(os);
- os << "\n\\i \\" << t.cs() << "{"
+ os << "\\i \\" << t.cs() << "{"
<< trim(parse_text_snippet(p, FLAG_ITEM, outer, context), " ")
<< "}\n";
}
skip_braces(p); // eat {}
}
- else if (t.cs() == "i" || t.cs() == "j") {
+ else if (t.cs() == "i" || t.cs() == "j" || t.cs() == "l" ||
+ t.cs() == "L") {
context.check_layout(os);
- os << "\\" << t.cs() << ' ';
+ os << "\\i \\" << t.cs() << "{}\n";
skip_braces(p); // eat {}
}
string const path = getMasterFilePath();
// We want to preserve relative / absolute filenames,
// therefore path is only used for testing
+ // FIXME UNICODE encoding of filename and path may be
+ // wrong (makeAbsPath expects utf8)
if (t.cs() == "include" &&
- !fs::exists(MakeAbsPath(filename, path))) {
+ !fs::exists(makeAbsPath(filename, path).toFilesystemEncoding())) {
// The file extension is probably missing.
// Now try to find it out.
string const tex_name =
if (!tex_name.empty())
filename = tex_name;
}
- if (fs::exists(MakeAbsPath(filename, path))) {
+ // FIXME UNICODE encoding of filename and path may be
+ // wrong (makeAbsPath expects utf8)
+ if (fs::exists(makeAbsPath(filename, path).toFilesystemEncoding())) {
string const abstexname =
- MakeAbsPath(filename, path);
+ makeAbsPath(filename, path).absFilename();
string const abslyxname =
- ChangeExtension(abstexname, ".lyx");
+ changeExtension(abstexname, ".lyx");
fix_relative_filename(filename);
string const lyxname =
- ChangeExtension(filename, ".lyx");
+ changeExtension(filename, ".lyx");
if (t.cs() != "verbatiminput" &&
- tex2lyx(abstexname, abslyxname)) {
+ tex2lyx(abstexname, FileName(abslyxname))) {
os << name << '{' << lyxname << "}\n";
} else {
os << name << '{' << filename << "}\n";
skip_braces(p);
}
- else if (t.cs() == "newpage") {
+ else if (is_known(t.cs(), known_spaces)) {
+ char const * const * where = is_known(t.cs(), known_spaces);
context.check_layout(os);
- // FIXME: what about \\clearpage and \\pagebreak?
- os << "\n\\newpage\n";
+ begin_inset(os, "InsetSpace ");
+ os << '\\' << known_coded_spaces[where - known_spaces]
+ << '\n';
+ // LaTeX swallows whitespace after all spaces except
+ // "\\,". We have to do that here, too, because LyX
+ // adds "{}" which would make the spaces significant.
+ if (t.cs() != ",")
+ eat_whitespace(p, os, context, false);
+ // LyX adds "{}" after all spaces except "\\ " and
+ // "\\,", so we have to remove "{}".
+ // "\\,{}" is equivalent to "\\," in LaTeX, so we
+ // remove the braces after "\\,", too.
+ if (t.cs() != " ")
+ skip_braces(p);
+ }
+
+ else if (t.cs() == "newpage" ||
+ t.cs() == "clearpage" ||
+ t.cs() == "cleardoublepage") {
+ context.check_layout(os);
+ // FIXME: what about \\pagebreak?
+ os << "\n\\" << t.cs() << "\n";
skip_braces(p); // eat {}
}
}
// }])
+
+
+} // namespace lyx