#include "Buffer.h"
#include "Author.h"
-#include "BaseClassList.h"
+#include "LayoutFile.h"
#include "BiblioInfo.h"
#include "BranchList.h"
#include "buffer_funcs.h"
#include "Converter.h"
#include "Counters.h"
#include "DocIterator.h"
-#include "EmbeddedFiles.h"
#include "Encoding.h"
#include "ErrorList.h"
#include "Exporter.h"
#include "ParagraphParameters.h"
#include "ParIterator.h"
#include "PDFOptions.h"
-#include "Session.h"
#include "sgml.h"
#include "TexRow.h"
#include "TexStream.h"
#include "graphics/Previews.h"
+#include "support/lassert.h"
#include "support/convert.h"
#include "support/debug.h"
#include "support/ExceptionMessage.h"
#include "support/Path.h"
#include "support/textutils.h"
#include "support/types.h"
-#include "support/FileZipListDir.h"
#include <boost/bind.hpp>
#include <boost/shared_ptr.hpp>
#include <algorithm>
+#include <fstream>
#include <iomanip>
-#include <stack>
+#include <map>
#include <sstream>
-#include <fstream>
+#include <stack>
+#include <vector>
using namespace std;
using namespace lyx::support;
namespace {
-int const LYX_FORMAT = 315; // Richard Heck: column separation
-
-} // namespace anon
-
+int const LYX_FORMAT = 340; //jamatos: add plain layout
typedef map<string, bool> DepClean;
+typedef map<docstring, pair<InsetLabel const *, Buffer::References> > RefCache;
+
+} // namespace anon
class Buffer::Impl
{
wa_->closeAll();
delete wa_;
}
+ delete inset;
}
-
+
BufferParams params;
LyXVC lyxvc;
FileName temppath;
*/
bool file_fully_loaded;
- /// our Text that should be wrapped in an InsetText
- InsetText inset;
-
///
mutable TocBackend toc_backend;
typedef pair<DocIterator, MacroData> ScopeMacro;
typedef map<DocIterator, ScopeMacro> PositionScopeMacroMap;
typedef map<docstring, PositionScopeMacroMap> NamePositionScopeMacroMap;
+ /// map from the macro name to the position map,
+ /// which maps the macro definition position to the scope and the MacroData.
NamePositionScopeMacroMap macros;
bool macro_lock;
-
+
/// positions of child buffers in the buffer
typedef map<Buffer const * const, DocIterator> BufferPositionMap;
typedef pair<DocIterator, Buffer const *> ScopeBuffer;
typedef map<DocIterator, ScopeBuffer> PositionScopeBufferMap;
+ /// position of children buffers in this buffer
BufferPositionMap children_positions;
+ /// map from children inclusion positions to their scope and their buffer
PositionScopeBufferMap position_to_children;
/// Container for all sort of Buffer dependant errors.
map<string, ErrorList> errorLists;
- /// all embedded files of this buffer
- EmbeddedFileList embedded_files;
-
/// timestamp and checksum used to test if the file has been externally
/// modified. (Used to properly enable 'File->Revert to saved', bug 4114).
time_t timestamp_;
/// A cache for the bibfiles (including bibfiles of loaded child
/// documents), needed for appropriate update of natbib labels.
- mutable EmbeddedFileList bibfilesCache_;
+ mutable support::FileNameList bibfilesCache_;
+
+ // FIXME The caching mechanism could be improved. At present, we have a
+ // cache for each Buffer, that caches all the bibliography info for that
+ // Buffer. A more efficient solution would be to have a global cache per
+ // file, and then to construct the Buffer's bibinfo from that.
+ /// A cache for bibliography info
+ mutable BiblioInfo bibinfo_;
+ /// whether the bibinfo cache is valid
+ bool bibinfoCacheValid_;
+ /// Cache of timestamps of .bib files
+ map<FileName, time_t> bibfileStatus_;
+
+ mutable RefCache ref_cache_;
+
+ /// our Text that should be wrapped in an InsetText
+ InsetText * inset;
};
+
/// Creates the per buffer temporary directory
static FileName createBufferTmpDir()
{
Buffer::Impl::Impl(Buffer & parent, FileName const & file, bool readonly_)
: parent_buffer(0), lyx_clean(true), bak_clean(true), unnamed(false),
read_only(readonly_), filename(file), file_fully_loaded(false),
- inset(params), toc_backend(&parent), macro_lock(false),
- embedded_files(), timestamp_(0), checksum_(0), wa_(0),
- undo_(parent)
+ toc_backend(&parent), macro_lock(false), timestamp_(0),
+ checksum_(0), wa_(0), undo_(parent), bibinfoCacheValid_(false)
{
temppath = createBufferTmpDir();
- inset.setBuffer(parent);
- inset.setAutoBreakRows(true);
lyxvc.setBuffer(&parent);
if (use_gui)
wa_ = new frontend::WorkAreaManager;
{
LYXERR(Debug::INFO, "Buffer::Buffer()");
- d->inset.getText(0)->setMacrocontextPosition(par_iterator_begin());
+ d->inset = new InsetText(*this);
+ d->inset->setAutoBreakRows(true);
+ d->inset->getText(0)->setMacrocontextPosition(par_iterator_begin());
}
// GuiView already destroyed
gui_ = 0;
- Buffer const * master = masterBuffer();
- if (master != this && use_gui) {
- // We are closing buf which was a child document so we
- // must update the labels and section numbering of its master
- // Buffer.
- updateLabels(*master);
- master->updateMacros();
- }
- resetChildDocuments(false);
+ // loop over children
+ Impl::BufferPositionMap::iterator it = d->children_positions.begin();
+ Impl::BufferPositionMap::iterator end = d->children_positions.end();
+ for (; it != end; ++it)
+ theBufferList().releaseChild(this, const_cast<Buffer *>(it->first));
+
+ // clear references to children in macro tables
+ d->children_positions.clear();
+ d->position_to_children.clear();
if (!d->temppath.destroyDirectory()) {
Alert::warning(_("Could not remove temporary directory"),
}
// Remove any previewed LaTeX snippets associated with this buffer.
- graphics::Previews::get().removeLoader(*this);
+ thePreviews().removeLoader(*this);
delete d;
}
frontend::WorkAreaManager & Buffer::workAreaManager() const
{
- BOOST_ASSERT(d->wa_);
+ LASSERT(d->wa_, /**/);
return *d->wa_;
}
Text & Buffer::text() const
{
- return const_cast<Text &>(d->inset.text_);
+ return d->inset->text();
}
Inset & Buffer::inset() const
{
- return const_cast<InsetText &>(d->inset);
+ return *d->inset;
}
}
-EmbeddedFileList & Buffer::embeddedFiles()
-{
- return d->embedded_files;
-}
-
-
-EmbeddedFileList const & Buffer::embeddedFiles() const
-{
- return d->embedded_files;
-}
-
-
-bool Buffer::embedded() const
-{
- return params().embedded;
-}
-
-
Undo & Buffer::undo()
{
return d->undo_;
params().branchlist().clear();
params().preamble.erase();
params().options.erase();
+ params().master.erase();
params().float_placement.erase();
params().paperwidth.erase();
params().paperheight.erase();
params().headsep.erase();
params().footskip.erase();
params().columnsep.erase();
+ params().fontsCJK.erase();
params().listings_params.clear();
params().clearLayoutModules();
+ params().clearRemovedModules();
params().pdfoptions().clear();
-
+
for (int i = 0; i < 4; ++i) {
params().user_defined_bullet(i) = ITEMIZE_DEFAULTS[i];
params().temp_bullet(i) = ITEMIZE_DEFAULTS[i];
ErrorList & errorList = d->errorLists["Parse"];
while (lex.isOK()) {
- lex.next();
- string const token = lex.getString();
+ string token;
+ lex >> token;
if (token.empty())
continue;
errorList.push_back(ErrorItem(_("Document header error"),
s, -1, 0, 0));
}
-
+
params().makeDocumentClass();
return unknown_tokens;
ErrorList & errorList = d->errorLists["Parse"];
errorList.clear();
- lex.next();
- string const token = lex.getString();
- if (token != "\\begin_document") {
+ if (!lex.checkFor("\\begin_document")) {
docstring const s = _("\\begin_document is missing");
errorList.push_back(ErrorItem(_("Document header error"),
s, -1, 0, 0));
}
// we are reading in a brand new document
- BOOST_ASSERT(paragraphs().empty());
+ LASSERT(paragraphs().empty(), /**/);
readHeader(lex);
}
}
+ if (!params().master.empty()) {
+ FileName const master_file = makeAbsPath(params().master,
+ onlyPath(absFileName()));
+ if (isLyXFilename(master_file.absFilename())) {
+ Buffer * master = checkAndLoadLyXFile(master_file);
+ d->parent_buffer = master;
+ }
+ }
+
// read main text
- bool const res = text().read(*this, lex, errorList, &(d->inset));
+ bool const res = text().read(*this, lex, errorList, d->inset);
updateMacros();
updateMacroInstances();
if (*cit == '\n') {
if (autobreakrows && (!par.empty() || par.allowEmpty())) {
breakParagraph(params(), pars, pit, pos,
- par.layout()->isEnvironment());
+ par.layout().isEnvironment());
++pit;
pos = 0;
space_inserted = true;
// remove dummy empty par
paragraphs().clear();
- Lexer lex(0, 0);
+ Lexer lex;
istringstream is(s);
lex.setStream(is);
- FileName const name = FileName::tempName();
+ FileName const name = FileName::tempName("Buffer_readString");
switch (readFile(lex, name, true)) {
case failure:
return false;
bool Buffer::readFile(FileName const & filename)
{
FileName fname(filename);
- // Check if the file is compressed.
- string format = filename.guessFormatFromContents();
- if (format == "zip") {
- // decompress to a temp directory
- LYXERR(Debug::FILES, filename << " is in zip format. Unzip to " << temppath());
- ::unzipToDir(filename.toFilesystemEncoding(), temppath());
- //
- FileName lyxfile(addName(temppath(), "content.lyx"));
- // if both manifest.txt and file.lyx exist, this is am embedded file
- if (lyxfile.exists()) {
- // if in bundled format, save checksum of the compressed file, not content.lyx
- saveCheckSum(filename);
- params().embedded = true;
- fname = lyxfile;
- }
- }
- // The embedded lyx file can also be compressed, for backward compatibility
- format = fname.guessFormatFromContents();
- if (format == "gzip" || format == "zip" || format == "compress")
- params().compressed = true;
// remove dummy empty par
paragraphs().clear();
- Lexer lex(0, 0);
+ Lexer lex;
lex.setFile(fname);
if (readFile(lex, fname) != success)
return false;
Buffer::ReadStatus Buffer::readFile(Lexer & lex, FileName const & filename,
bool fromstring)
{
- BOOST_ASSERT(!filename.empty());
-
- if (!lex.isOK()) {
- Alert::error(_("Document could not be read"),
- bformat(_("%1$s could not be read."), from_utf8(filename.absFilename())));
- return failure;
- }
-
- lex.next();
- string const token = lex.getString();
-
- if (!lex) {
- Alert::error(_("Document could not be read"),
- bformat(_("%1$s could not be read."), from_utf8(filename.absFilename())));
- return failure;
- }
-
- // the first token _must_ be...
- if (token != "\\lyxformat") {
- lyxerr << "Token: " << token << endl;
+ LASSERT(!filename.empty(), /**/);
+ // the first (non-comment) token _must_ be...
+ if (!lex.checkFor("\\lyxformat")) {
Alert::error(_("Document format failure"),
- bformat(_("%1$s is not a LyX document."),
+ bformat(_("%1$s is not a readable LyX document."),
from_utf8(filename.absFilename())));
return failure;
}
- lex.next();
- string tmp_format = lex.getString();
+ string tmp_format;
+ lex >> tmp_format;
//lyxerr << "LyX Format: `" << tmp_format << '\'' << endl;
// if present remove ".," from string.
- string::size_type dot = tmp_format.find_first_of(".,");
+ size_t dot = tmp_format.find_first_of(".,");
//lyxerr << " dot found at " << dot << endl;
if (dot != string::npos)
tmp_format.erase(dot, 1);
// lyx2lyx would fail
return wrongversion;
- FileName const tmpfile = FileName::tempName();
+ FileName const tmpfile = FileName::tempName("Buffer_readFile");
if (tmpfile.empty()) {
Alert::error(_("Conversion failed"),
bformat(_("%1$s is from a different"
bool retval = false;
- FileName content;
- if (params().embedded)
- // first write the .lyx file to the temporary directory
- content = FileName(addName(temppath(), "content.lyx"));
- else
- content = fname;
-
docstring const str = bformat(_("Saving document %1$s..."),
- makeDisplayPath(content.absFilename()));
+ makeDisplayPath(fname.absFilename()));
message(str);
if (params().compressed) {
- gz::ogzstream ofs(content.toFilesystemEncoding().c_str(), ios::out|ios::trunc);
+ gz::ogzstream ofs(fname.toFilesystemEncoding().c_str(), ios::out|ios::trunc);
retval = ofs && write(ofs);
} else {
- ofstream ofs(content.toFilesystemEncoding().c_str(), ios::out|ios::trunc);
+ ofstream ofs(fname.toFilesystemEncoding().c_str(), ios::out|ios::trunc);
retval = ofs && write(ofs);
}
if (!retval) {
- message(str + _(" could not write file!."));
+ message(str + _(" could not write file!"));
return false;
}
removeAutosaveFile(d->filename.absFilename());
- if (params().embedded) {
- message(str + _(" writing embedded files!."));
- // if embedding is enabled, write file.lyx and all the embedded files
- // to the zip file fname.
- if (!d->embedded_files.writeFile(fname, *this)) {
- message(str + _(" could not write embedded files!."));
- return false;
- }
- }
saveCheckSum(d->filename);
message(str + _(" done."));
"representable in the chosen encoding.\n"
"Changing the document encoding to utf8 could help."),
e.par_id, e.pos, e.pos + 1));
- failed_export = true;
+ failed_export = true;
}
catch (iconv_codecvt_facet_exception & e) {
errorList.push_back(ErrorItem(_("iconv conversion failed"),
}
catch (...) {
lyxerr << "Caught some really weird exception..." << endl;
- LyX::cref().exit(1);
+ lyx_exit(1);
}
ofs.close();
OutputParams const & runparams_in,
bool const output_preamble, bool const output_body) const
{
+ // The child documents, if any, shall be already loaded at this point.
+
OutputParams runparams = runparams_in;
+ // Classify the unicode characters appearing in math insets
+ Encodings::initUnicodeMath(*this);
+
// validate the buffer.
LYXERR(Debug::LATEX, " Validating buffer...");
LaTeXFeatures features(*this, params(), runparams);
d->texrow.newline();
}
LYXERR(Debug::INFO, "lyx document header finished");
+
+ // Don't move this behind the parent_buffer=0 code below,
+ // because then the macros will not get the right "redefinition"
+ // flag as they don't see the parent macros which are output before.
+ updateMacros();
+
+ // fold macros if possible, still with parent buffer as the
+ // macros will be put in the prefix anyway.
+ updateMacroInstances();
+
// There are a few differences between nice LaTeX and usual files:
// usual is \batchmode and has a
// special input@path to allow the including of figures
d->texrow.newline();
}
+ // get parent macros (if this buffer has a parent) which will be
+ // written at the document begin further down.
+ MacroSet parentMacros;
+ listParentMacros(parentMacros, features);
+
// Write the preamble
runparams.use_babel = params().writeLaTeX(os, features, d->texrow);
+ runparams.use_japanese = features.isRequired("japanese");
+
if (!output_body)
return;
// make the body.
os << "\\begin{document}\n";
d->texrow.newline();
+
+ // output the parent macros
+ MacroSet::iterator it = parentMacros.begin();
+ MacroSet::iterator end = parentMacros.end();
+ for (; it != end; ++it)
+ (*it)->write(os, true);
} // output_preamble
d->texrow.start(paragraphs().begin()->id(), 0);
-
- LYXERR(Debug::INFO, "preamble finished, now the body.");
-
- // load children, if not already done.
- // This includes an updateMacro() call.
- // Don't move this behind the parent_buffer=0 code below,
- // because then the macros will not get the right "redefinition"
- // flag as they don't see the parent macros which are output before.
- loadChildDocuments();
- // fold macros if possible, still with parent buffer as the
- // macros will be put in the prefix anyway.
- updateMacroInstances();
+ LYXERR(Debug::INFO, "preamble finished, now the body.");
// if we are doing a real file with body, even if this is the
// child of some other buffer, let's cut the link here.
// This happens for example if only a child document is printed.
Buffer const * save_parent = 0;
if (output_preamble) {
- // output the macros visible for this buffer
- writeParentMacros(os);
-
save_parent = d->parent_buffer;
d->parent_buffer = 0;
}
latexParagraphs(*this, text(), os, d->texrow, runparams);
// Restore the parenthood if needed
- if (output_preamble) {
+ if (output_preamble)
d->parent_buffer = save_parent;
- // restore macros with correct parent buffer (especially
- // important for the redefinition flag which depends on the
- // parent)
- updateMacros();
- }
-
// add this just in case after all the paragraphs
os << endl;
d->texrow.newline();
if (runparams.flavor == OutputParams::XML)
top += params().language->code();
else
- top += params().language->code().substr(0,2);
+ top += params().language->code().substr(0, 2);
top += '"';
if (!params().options.empty()) {
params().documentClass().counters().reset();
- loadChildDocuments();
+ updateMacros();
sgml::openTag(os, top);
os << '\n';
{
params().validate(features);
- loadChildDocuments();
+ updateMacros();
for_each(paragraphs().begin(), paragraphs().end(),
boost::bind(&Paragraph::validate, _1, boost::ref(features)));
void Buffer::getLabelList(vector<docstring> & list) const
{
- /// if this is a child document and the parent is already loaded
- /// Use the parent's list instead [ale990407]
- Buffer const * tmp = masterBuffer();
- if (!tmp) {
- lyxerr << "masterBuffer() failed!" << endl;
- BOOST_ASSERT(tmp);
- }
- if (tmp != this) {
- tmp->getLabelList(list);
+ // If this is a child document, use the parent's list instead.
+ if (d->parent_buffer) {
+ d->parent_buffer->getLabelList(list);
return;
}
- loadChildDocuments();
-
- for (InsetIterator it = inset_iterator_begin(inset()); it; ++it)
- it.nextInset()->getLabelList(list);
+ list.clear();
+ Toc & toc = d->toc_backend.toc("label");
+ TocIterator toc_it = toc.begin();
+ TocIterator end = toc.end();
+ for (; toc_it != end; ++toc_it) {
+ if (toc_it->depth() == 0)
+ list.push_back(toc_it->str());
+ }
}
void Buffer::updateBibfilesCache() const
{
- // if this is a child document and the parent is already loaded
- // update the parent's cache instead
- Buffer const * tmp = masterBuffer();
- BOOST_ASSERT(tmp);
- if (tmp != this) {
- tmp->updateBibfilesCache();
+ // If this is a child document, use the parent's cache instead.
+ if (d->parent_buffer) {
+ d->parent_buffer->updateBibfilesCache();
return;
}
if (it->lyxCode() == BIBTEX_CODE) {
InsetBibtex const & inset =
static_cast<InsetBibtex const &>(*it);
- EmbeddedFileList const bibfiles = inset.embeddedFiles();
+ support::FileNameList const bibfiles = inset.getBibFiles();
d->bibfilesCache_.insert(d->bibfilesCache_.end(),
bibfiles.begin(),
bibfiles.end());
InsetInclude & inset =
static_cast<InsetInclude &>(*it);
inset.updateBibfilesCache();
- EmbeddedFileList const & bibfiles =
+ support::FileNameList const & bibfiles =
inset.getBibfilesCache(*this);
d->bibfilesCache_.insert(d->bibfilesCache_.end(),
bibfiles.begin(),
bibfiles.end());
}
}
+ // the bibinfo cache is now invalid
+ d->bibinfoCacheValid_ = false;
}
-EmbeddedFileList const & Buffer::getBibfilesCache() const
+void Buffer::invalidateBibinfoCache()
{
- // if this is a child document and the parent is already loaded
- // use the parent's cache instead
- Buffer const * tmp = masterBuffer();
- BOOST_ASSERT(tmp);
- if (tmp != this)
- return tmp->getBibfilesCache();
+ d->bibinfoCacheValid_ = false;
+}
+
+
+support::FileNameList const & Buffer::getBibfilesCache() const
+{
+ // If this is a child document, use the parent's cache instead.
+ if (d->parent_buffer)
+ return d->parent_buffer->getBibfilesCache();
// We update the cache when first used instead of at loading time.
if (d->bibfilesCache_.empty())
}
+BiblioInfo const & Buffer::masterBibInfo() const
+{
+ // if this is a child document and the parent is already loaded
+ // use the parent's list instead [ale990412]
+ Buffer const * const tmp = masterBuffer();
+ LASSERT(tmp, /**/);
+ if (tmp != this)
+ return tmp->masterBibInfo();
+ return localBibInfo();
+}
+
+
+BiblioInfo const & Buffer::localBibInfo() const
+{
+ if (d->bibinfoCacheValid_) {
+ support::FileNameList const & bibfilesCache = getBibfilesCache();
+ // compare the cached timestamps with the actual ones.
+ support::FileNameList::const_iterator ei = bibfilesCache.begin();
+ support::FileNameList::const_iterator en = bibfilesCache.end();
+ for (; ei != en; ++ ei) {
+ time_t lastw = ei->lastModified();
+ if (lastw != d->bibfileStatus_[*ei]) {
+ d->bibinfoCacheValid_ = false;
+ d->bibfileStatus_[*ei] = lastw;
+ break;
+ }
+ }
+ }
+
+ if (!d->bibinfoCacheValid_) {
+ d->bibinfo_.clear();
+ for (InsetIterator it = inset_iterator_begin(inset()); it; ++it)
+ it->fillWithBibKeys(d->bibinfo_, it);
+ d->bibinfoCacheValid_ = true;
+ }
+ return d->bibinfo_;
+}
+
+
bool Buffer::isDepClean(string const & name) const
{
DepClean::const_iterator const it = d->dep_clean.find(name);
break;
}
+ case LFUN_BRANCH_ACTIVATE:
+ case LFUN_BRANCH_DEACTIVATE: {
+ BranchList & branchList = params().branchlist();
+ docstring const branchName = func.argument();
+ Branch * branch = branchList.find(branchName);
+ if (!branch)
+ LYXERR0("Branch " << branchName << " does not exist.");
+ else
+ branch->setSelected(func.action == LFUN_BRANCH_ACTIVATE);
+ if (result)
+ *result = true;
+ }
+
default:
dispatched = false;
}
void Buffer::changeLanguage(Language const * from, Language const * to)
{
- BOOST_ASSERT(from);
- BOOST_ASSERT(to);
+ LASSERT(from, /**/);
+ LASSERT(to, /**/);
for_each(par_iterator_begin(),
par_iterator_end(),
bool Buffer::isExternallyModified(CheckMethod method) const
{
- BOOST_ASSERT(d->filename.exists());
+ LASSERT(d->filename.exists(), /**/);
// if method == timestamp, check timestamp before checksum
- return (method == checksum_method
+ return (method == checksum_method
|| d->timestamp_ != d->filename.lastModified())
&& d->checksum_ != d->filename.checksum();
}
{
if (!d->parent_buffer)
return this;
-
+
return d->parent_buffer->masterBuffer();
}
+bool Buffer::isChild(Buffer * child) const
+{
+ return d->children_positions.find(child) != d->children_positions.end();
+}
+
+
template<typename M>
typename M::iterator greatest_below(M & m, typename M::key_type const & x)
{
return m.end();
it--;
- return it;
+ return it;
}
-MacroData const * Buffer::getBufferMacro(docstring const & name,
+MacroData const * Buffer::getBufferMacro(docstring const & name,
DocIterator const & pos) const
{
LYXERR(Debug::MACROS, "Searching for " << to_ascii(name) << " at " << pos);
// we haven't found anything yet
DocIterator bestPos = par_iterator_begin();
MacroData const * bestData = 0;
-
+
// find macro definitions for name
Impl::NamePositionScopeMacroMap::iterator nameIt
= d->macros.find(name);
bestData = &it->second.second;
break;
}
-
+
// try previous macro if there is one
if (it == nameIt->second.begin())
break;
break;
--it;
}
-
+
// return the best macro we have found
return bestData;
}
DocIterator const & pos, bool global) const
{
if (d->macro_lock)
- return 0;
+ return 0;
// query buffer macros
MacroData const * data = getBufferMacro(name, pos);
// If there is a master buffer, query that
if (d->parent_buffer) {
d->macro_lock = true;
- MacroData const * macro
- = d->parent_buffer->getMacro(name, *this, false);
+ MacroData const * macro = d->parent_buffer->getMacro(
+ name, *this, false);
d->macro_lock = false;
if (macro)
return macro;
}
-MacroData const * Buffer::getMacro(docstring const & name, Buffer const & child, bool global) const
+MacroData const * Buffer::getMacro(docstring const & name,
+ Buffer const & child, bool global) const
{
// look where the child buffer is included first
- Impl::BufferPositionMap::iterator it
- = d->children_positions.find(&child);
+ Impl::BufferPositionMap::iterator it = d->children_positions.find(&child);
if (it == d->children_positions.end())
return 0;
}
-void Buffer::updateEnvironmentMacros(DocIterator & it,
- pit_type lastpit,
- DocIterator & scope) const
+void Buffer::updateMacros(DocIterator & it, DocIterator & scope) const
{
- Paragraph & par = it.paragraph();
- depth_type depth
- = par.params().depth();
- Length const & leftIndent
- = par.params().leftIndent();
+ pit_type lastpit = it.lastpit();
// look for macros in each paragraph
while (it.pit() <= lastpit) {
Paragraph & par = it.paragraph();
- // increased depth?
- if ((par.params().depth() > depth
- || par.params().leftIndent() != leftIndent)
- && par.layout()->isEnvironment()) {
- updateBlockMacros(it, scope);
- continue;
- }
-
// iterate over the insets of the current paragraph
InsetList const & insets = par.insetList();
InsetList::const_iterator iit = insets.begin();
InsetList::const_iterator end = insets.end();
for (; iit != end; ++iit) {
it.pos() = iit->pos;
-
+
// is it a nested text inset?
if (iit->inset->asInsetText()) {
// Inset needs its own scope?
- InsetText const * itext
+ InsetText const * itext
= iit->inset->asInsetText();
bool newScope = itext->isMacroScope();
- // scope which ends just behind the inset
+ // scope which ends just behind the inset
DocIterator insetScope = it;
++insetScope.pos();
// collect macros in inset
it.push_back(CursorSlice(*iit->inset));
- updateInsetMacros(it, newScope ? insetScope : scope);
+ updateMacros(it, newScope ? insetScope : scope);
it.pop_back();
continue;
}
-
+
// is it an external file?
if (iit->inset->lyxCode() == INCLUDE_CODE) {
// get buffer of external file
- InsetCommand const & inset
- = static_cast<InsetCommand const &>(*iit->inset);
+ InsetCommand const & inset
+ = static_cast<InsetCommand const &>(*iit->inset);
InsetCommandParams const & ip = inset.params();
d->macro_lock = true;
Buffer * child = loadIfNeeded(*this, ip);
d->macro_lock = false;
if (!child)
- continue;
+ continue;
// register its position, but only when it is
// included first in the buffer
if (d->children_positions.find(child)
- == d->children_positions.end())
+ == d->children_positions.end())
d->children_positions[child] = it;
-
+
// register child with its scope
d->position_to_children[it] = Impl::ScopeBuffer(scope, child);
-
continue;
}
if (iit->inset->lyxCode() != MATHMACRO_CODE)
continue;
-
+
// get macro data
MathMacroTemplate & macroTemplate
= static_cast<MathMacroTemplate &>(*iit->inset);
continue;
// register macro
- d->macros[macroTemplate.name()][it]
- = Impl::ScopeMacro(scope, MacroData(*this, it));
+ d->macros[macroTemplate.name()][it] =
+ Impl::ScopeMacro(scope, MacroData(*this, it));
}
// next paragraph
}
-void Buffer::updateBlockMacros(DocIterator & it, DocIterator & scope) const
-{
- Paragraph & par = it.paragraph();
-
- // set scope for macros in this paragraph:
- // * either the "old" outer scope
- // * or the scope ending after the environment
- if (par.layout()->isEnvironment()) {
- // find end of environment block,
- DocIterator envEnd = it;
- pit_type n = it.lastpit() + 1;
- depth_type depth = par.params().depth();
- Length const & length = par.params().leftIndent();
- // looping through the paragraph, basically until
- // the layout changes or the depth gets smaller.
- // (the logic of output_latex.cpp's TeXEnvironment)
- do {
- envEnd.pit()++;
- if (envEnd.pit() == n)
- break;
- } while (par.layout() == envEnd.paragraph().layout()
- || depth < envEnd.paragraph().params().depth()
- || length != envEnd.paragraph().params().leftIndent());
-
- // collect macros from environment block
- updateEnvironmentMacros(it, envEnd.pit() - 1, envEnd);
- } else {
- // collect macros from paragraph
- updateEnvironmentMacros(it, it.pit(), scope);
- }
-}
-
-
-void Buffer::updateInsetMacros(DocIterator & it, DocIterator & scope) const
-{
- // look for macros in each paragraph
- pit_type n = it.lastpit() + 1;
- while (it.pit() < n)
- updateBlockMacros(it, scope);
-}
-
-
void Buffer::updateMacros() const
{
if (d->macro_lock)
DocIterator it = par_iterator_begin();
DocIterator outerScope = it;
outerScope.pit() = outerScope.lastpit() + 2;
- updateInsetMacros(it, outerScope);
+ updateMacros(it, outerScope);
}
return;
d->macro_lock = true;
-
+
// loop over macro names
- Impl::NamePositionScopeMacroMap::iterator nameIt
- = d->macros.begin();
- Impl::NamePositionScopeMacroMap::iterator nameEnd
- = d->macros.end();
+ Impl::NamePositionScopeMacroMap::iterator nameIt = d->macros.begin();
+ Impl::NamePositionScopeMacroMap::iterator nameEnd = d->macros.end();
for (; nameIt != nameEnd; ++nameIt)
macros.insert(nameIt->first);
// loop over children
- Impl::BufferPositionMap::iterator it
- = d->children_positions.begin();
- Impl::BufferPositionMap::iterator end
- = d->children_positions.end();
+ Impl::BufferPositionMap::iterator it = d->children_positions.begin();
+ Impl::BufferPositionMap::iterator end = d->children_positions.end();
for (; it != end; ++it)
it->first->listMacroNames(macros);
if (d->parent_buffer)
d->parent_buffer->listMacroNames(macros);
- d->macro_lock = false;
+ d->macro_lock = false;
}
-void Buffer::writeParentMacros(odocstream & os) const
+void Buffer::listParentMacros(MacroSet & macros, LaTeXFeatures & features) const
{
if (!d->parent_buffer)
return;
- // collect macro names
MacroNameSet names;
d->parent_buffer->listMacroNames(names);
- // resolve and output them
+ // resolve macros
MacroNameSet::iterator it = names.begin();
MacroNameSet::iterator end = names.end();
for (; it != end; ++it) {
// defined?
- MacroData const * data =
+ MacroData const * data =
d->parent_buffer->getMacro(*it, *this, false);
- if (data)
- data->write(os, true);
+ if (data) {
+ macros.insert(data);
+
+ // we cannot access the original MathMacroTemplate anymore
+ // here to calls validate method. So we do its work here manually.
+ // FIXME: somehow make the template accessible here.
+ if (data->optionals() > 0)
+ features.require("xargs");
+ }
}
}
+Buffer::References & Buffer::references(docstring const & label)
+{
+ if (d->parent_buffer)
+ return const_cast<Buffer *>(masterBuffer())->references(label);
+
+ RefCache::iterator it = d->ref_cache_.find(label);
+ if (it != d->ref_cache_.end())
+ return it->second.second;
+
+ static InsetLabel const * dummy_il = 0;
+ static References const dummy_refs;
+ it = d->ref_cache_.insert(
+ make_pair(label, make_pair(dummy_il, dummy_refs))).first;
+ return it->second.second;
+}
+
+
+Buffer::References const & Buffer::references(docstring const & label) const
+{
+ return const_cast<Buffer *>(this)->references(label);
+}
+
+
+void Buffer::setInsetLabel(docstring const & label, InsetLabel const * il)
+{
+ masterBuffer()->d->ref_cache_[label].first = il;
+}
+
+
+InsetLabel const * Buffer::insetLabel(docstring const & label) const
+{
+ return masterBuffer()->d->ref_cache_[label].first;
+}
+
+
+void Buffer::clearReferenceCache() const
+{
+ if (!d->parent_buffer)
+ d->ref_cache_.clear();
+}
+
+
void Buffer::changeRefsIfUnique(docstring const & from, docstring const & to,
InsetCode code)
{
//FIXME: This does not work for child documents yet.
- BOOST_ASSERT(code == CITE_CODE || code == REF_CODE);
+ LASSERT(code == CITE_CODE, /**/);
// Check if the label 'from' appears more than once
vector<docstring> labels;
-
string paramName;
- if (code == CITE_CODE) {
- BiblioInfo keys;
- keys.fillWithBibKeys(this);
- BiblioInfo::const_iterator bit = keys.begin();
- BiblioInfo::const_iterator bend = keys.end();
+ BiblioInfo const & keys = masterBibInfo();
+ BiblioInfo::const_iterator bit = keys.begin();
+ BiblioInfo::const_iterator bend = keys.end();
- for (; bit != bend; ++bit)
- // FIXME UNICODE
- labels.push_back(bit->first);
- paramName = "key";
- } else {
- getLabelList(labels);
- paramName = "reference";
- }
+ for (; bit != bend; ++bit)
+ // FIXME UNICODE
+ labels.push_back(bit->first);
+ paramName = "key";
if (count(labels.begin(), labels.end(), from) > 1)
return;
void Buffer::getSourceCode(odocstream & os, pit_type par_begin,
- pit_type par_end, bool full_source)
+ pit_type par_end, bool full_source) const
{
OutputParams runparams(¶ms().encoding());
runparams.nice = true;
d->texrow.newline();
if (isLatex())
writeLaTeXSource(os, filePath(), runparams, true, true);
- else {
+ else
writeDocBookSource(os, absFileName(), runparams, false);
- }
} else {
runparams.par_begin = par_begin;
runparams.par_end = par_end;
- if (par_begin + 1 == par_end)
+ if (par_begin + 1 == par_end) {
os << "% "
<< bformat(_("Preview source code for paragraph %1$d"), par_begin)
<< "\n\n";
- else
+ } else {
os << "% "
<< bformat(_("Preview source code from paragraph %1$s to %2$s"),
convert<docstring>(par_begin),
convert<docstring>(par_end - 1))
<< "\n\n";
+ }
d->texrow.newline();
d->texrow.newline();
// output paragraphs
- if (isLatex()) {
+ if (isLatex())
latexParagraphs(*this, text(), os, d->texrow, runparams);
- } else {
+ else
// DocBook
docbookParagraphs(paragraphs(), *this, os, runparams);
- }
}
}
}
+void Buffer::updateTocItem(std::string const & type,
+ DocIterator const & dit) const
+{
+ if (gui_)
+ gui_->updateTocItem(type, dit);
+}
+
+
void Buffer::structureChanged() const
{
if (gui_)
///
int start()
{
- command_ = to_utf8(bformat(_("Auto-saving %1$s"),
+ command_ = to_utf8(bformat(_("Auto-saving %1$s"),
from_utf8(fname_.absFilename())));
return run(DontWait);
}
}
-void Buffer::resetChildDocuments(bool close_them) const
-{
- if (text().empty())
- return;
-
- for (InsetIterator it = inset_iterator_begin(inset()); it; ++it) {
- if (it->lyxCode() != INCLUDE_CODE)
- continue;
- InsetCommand const & inset = static_cast<InsetCommand const &>(*it);
- InsetCommandParams const & ip = inset.params();
-
- resetParentBuffer(this, ip, close_them);
- }
-
- if (use_gui && masterBuffer() == this)
- updateLabels(*this);
-
- // clear references to children in macro tables
- d->children_positions.clear();
- d->position_to_children.clear();
-}
-
-
-void Buffer::loadChildDocuments() const
-{
- bool parse_error = false;
-
- for (InsetIterator it = inset_iterator_begin(inset()); it; ++it) {
- if (it->lyxCode() != INCLUDE_CODE)
- continue;
- InsetCommand const & inset = static_cast<InsetCommand const &>(*it);
- InsetCommandParams const & ip = inset.params();
- Buffer * child = loadIfNeeded(*this, ip);
- if (!child)
- continue;
- parse_error |= !child->errorList("Parse").empty();
- child->loadChildDocuments();
- }
-
- if (use_gui && masterBuffer() == this)
- updateLabels(*this);
-
- updateMacros();
-}
-
-
string Buffer::bufferFormat() const
{
if (isDocBook())
return "docbook";
if (isLiterate())
return "literate";
+ if (params().encoding().package() == Encoding::japanese)
+ return "platex";
return "latex";
}
string const error_type = (format == "program")
? "Build" : bufferFormat();
+ ErrorList & error_list = d->errorLists[error_type];
string const ext = formats.extension(format);
FileName const tmp_result_file(changeExtension(filename, ext));
bool const success = theConverters().convert(this, FileName(filename),
tmp_result_file, FileName(absFileName()), backend_format, format,
- errorList(error_type));
+ error_list);
// Emit the signal to show the error list.
if (format != backend_format)
errors(error_type);
}
}
+
} // namespace lyx