X-Git-Url: https://git.lyx.org/gitweb/?a=blobdiff_plain;f=lib%2Flyx2lyx%2Fparser_tools.py;h=31948fe30c90d6f25b540b7bf6585c7e20d439b0;hb=3da2eb41aa617f68b95d46c6d3c6f011d5dd8c77;hp=465e174d3a863dce9645e04948f926a7105bd2a5;hpb=2b2b83803a4bf3ad90e2ca28da4f1f4a76183716;p=lyx.git diff --git a/lib/lyx2lyx/parser_tools.py b/lib/lyx2lyx/parser_tools.py index 465e174d3a..31948fe30c 100644 --- a/lib/lyx2lyx/parser_tools.py +++ b/lib/lyx2lyx/parser_tools.py @@ -1,5 +1,7 @@ # This file is part of lyx2lyx -# Copyright (C) 2002 Dekel Tsur , José Matos +# -*- coding: utf-8 -*- +# Copyright (C) 2002-2011 Dekel Tsur , +# José Matos , Richard Heck # # This program is free software; you can redistribute it and/or # modify it under the terms of the GNU General Public License @@ -13,173 +15,430 @@ # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software -# Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. +# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA -import string +''' +This modules offer several free functions to help parse lines. +More documentaton is below, but here is a quick guide to what +they do. Optional arguments are marked by brackets. + +find_token(lines, token, start[, end[, ignorews]]): + Returns the first line i, start <= i < end, on which + token is found at the beginning. Returns -1 if not + found. + If ignorews is (given and) True, then differences + in whitespace do not count, except that there must be no + extra whitespace following token itself. + +find_token_exact(lines, token, start[, end]): + As find_token, but with ignorews True. + +find_tokens(lines, tokens, start[, end[, ignorews]]): + Returns the first line i, start <= i < end, on which + oen of the tokens in tokens is found at the beginning. + Returns -1 if not found. + If ignorews is (given and) True, then differences + in whitespace do not count, except that there must be no + extra whitespace following token itself. + +find_tokens_exact(lines, token, start[, end]): + As find_tokens, but with ignorews True. + +find_token_backwards(lines, token, start): +find_tokens_backwards(lines, tokens, start): + As before, but look backwards. + +find_re(lines, rexp, start[, end]): + As find_token, but rexp is a regular expression object, + so it has to be passed as e.g.: re.compile(r'...'). + +get_value(lines, token, start[, end[, default]): + Similar to find_token, but it returns what follows the + token on the found line. Example: + get_value(document.header, "\use_xetex", 0) + will find a line like: + \use_xetex true + and, in that case, return "true". (Note that whitespace + is stripped.) The final argument, default, defaults to "", + and is what is returned if we do not find anything. So you + can use that to set a default. + +get_quoted_value(lines, token, start[, end[, default]): + Similar to get_value, but it will strip quotes off the + value, if they are present. So use this one for cases + where the value is normally quoted. + +get_option_value(line, option): + This assumes we have a line with something like: + option="value" + and returns value. Returns "" if not found. + +del_token(lines, token, start[, end]): + Like find_token, but deletes the line if it finds one. + Returns True if a line got deleted, otherwise False. + +find_beginning_of(lines, i, start_token, end_token): + Here, start_token and end_token are meant to be a matching + pair, like "\begin_layout" and "\end_layout". We look for + the start_token that pairs with the end_token that occurs + on or after line i. Returns -1 if not found. + So, in the layout case, this would find the \begin_layout + for the layout line i is in. + Example: + ec = find_token(document.body, " bool -# We need to check that the char after the token is space, but I think -# we can ignore this -def find_token(lines, token, start, end = 0): - if end == 0: - end = len(lines) + Return True if token is present in line and is the first element + else returns False.""" + + return line[:len(token)] == token + + +def is_nonempty_line(line): + """ is_nonempty_line(line) -> bool + + Return False if line is either empty or it has only whitespaces, + else return True.""" + return line != " "*len(line) + + +# Utilities for a list of lines +def find_token(lines, token, start, end = 0, ignorews = False): + """ find_token(lines, token, start[[, end], ignorews]) -> int + + Return the lowest line where token is found, and is the first + element, in lines[start, end]. + + If ignorews is True (default is False), then differences in + whitespace are ignored, except that there must be no extra + whitespace following token itself. + + Return -1 on failure.""" + + if end == 0 or end > len(lines): + end = len(lines) m = len(token) for i in xrange(start, end): - if lines[i][:m] == token: - return i + if ignorews: + x = lines[i].split() + y = token.split() + if len(x) < len(y): + continue + if x[:len(y)] == y: + return i + else: + if lines[i][:m] == token: + return i return -1 -def find_token2(lines, token, start, end = 0): - if end == 0: - end = len(lines) - for i in xrange(start, end): - x = string.split(lines[i]) - if len(x) > 0 and x[0] == token: - return i - return -1 -def find_tokens(lines, tokens, start, end = 0): - if end == 0: - end = len(lines) +def find_token_exact(lines, token, start, end = 0): + return find_token(lines, token, start, end, True) + + +def find_tokens(lines, tokens, start, end = 0, ignorews = False): + """ find_tokens(lines, tokens, start[[, end], ignorews]) -> int + + Return the lowest line where one token in tokens is found, and is + the first element, in lines[start, end]. + + Return -1 on failure.""" + if end == 0 or end > len(lines): + end = len(lines) + for i in xrange(start, end): - line = lines[i] - for token in tokens: - if line[:len(token)] == token: - return i + for token in tokens: + if ignorews: + x = lines[i].split() + y = token.split() + if len(x) < len(y): + continue + if x[:len(y)] == y: + return i + else: + if lines[i][:len(token)] == token: + return i return -1 + +def find_tokens_exact(lines, tokens, start, end = 0): + return find_tokens(lines, tokens, start, end, True) + + def find_re(lines, rexp, start, end = 0): - if end == 0: - end = len(lines) + """ find_token_re(lines, rexp, start[, end]) -> int + + Return the lowest line where rexp, a regular expression, is found + in lines[start, end]. + + Return -1 on failure.""" + + if end == 0 or end > len(lines): + end = len(lines) for i in xrange(start, end): - if rexp.match(lines[i]): - return i + if rexp.match(lines[i]): + return i return -1 + def find_token_backwards(lines, token, start): + """ find_token_backwards(lines, token, start) -> int + + Return the highest line where token is found, and is the first + element, in lines[start, end]. + + Return -1 on failure.""" m = len(token) for i in xrange(start, -1, -1): - line = lines[i] - if line[:m] == token: - return i + line = lines[i] + if line[:m] == token: + return i return -1 + def find_tokens_backwards(lines, tokens, start): + """ find_tokens_backwards(lines, token, start) -> int + + Return the highest line where token is found, and is the first + element, in lines[end, start]. + + Return -1 on failure.""" for i in xrange(start, -1, -1): - line = lines[i] - for token in tokens: - if line[:len(token)] == token: - return i + line = lines[i] + for token in tokens: + if line[:len(token)] == token: + return i return -1 -def get_value(lines, token, start, end = 0): - i = find_token2(lines, token, start, end) + +def get_value(lines, token, start, end = 0, default = ""): + """ get_value(lines, token, start[[, end], default]) -> string + + Find the next line that looks like: + token followed by other stuff + Returns "followed by other stuff" with leading and trailing + whitespace removed. + """ + + i = find_token_exact(lines, token, start, end) if i == -1: - return "" - if len(string.split(lines[i])) > 1: - return string.split(lines[i])[1] - else: - return "" - -def del_token(lines, token, i, j): - k = find_token2(lines, token, i, j) + return default + l = lines[i].split(None, 1) + if len(l) > 1: + return l[1].strip() + return default + + +def get_quoted_value(lines, token, start, end = 0, default = ""): + """ get_quoted_value(lines, token, start[[, end], default]) -> string + + Find the next line that looks like: + token "followed by other stuff" + Returns "followed by other stuff" with leading and trailing + whitespace and quotes removed. If there are no quotes, that is OK too. + So use get_value to preserve possible quotes, this one to remove them, + if they are there. + Note that we will NOT strip quotes from default! + """ + val = get_value(lines, token, start, end, "") + if not val: + return default + return val.strip('"') + + +def get_option_value(line, option): + rx = option + '\s*=\s*"([^"]+)"' + rx = re.compile(rx) + m = rx.search(line) + if not m: + return "" + return m.group(1) + + +def del_token(lines, token, start, end = 0): + """ del_token(lines, token, start, end) -> int + + Find the first line in lines where token is the first element + and delete that line. Returns True if we deleted a line, False + if we did not.""" + + k = find_token_exact(lines, token, start, end) if k == -1: - return j - else: - del lines[k] - return j-1 - -# Finds the paragraph that contains line i. -def get_paragraph(lines, i): - while i != -1: - i = find_tokens_backwards(lines, ["\\end_inset", "\\layout"], i) - if i == -1: return -1 - if check_token(lines[i], "\\layout"): - return i - i = find_beginning_of_inset(lines, i) - -# Finds the paragraph after the paragraph that contains line i. -def get_next_paragraph(lines, i): - while i != -1: - i = find_tokens(lines, ["\\begin_inset", "\\layout", "\\end_float", "\\the_end"], i) - if not check_token(lines[i], "\\begin_inset"): - return i - i = find_end_of_inset(lines, i) + return False + del lines[k] + return True + + +def find_beginning_of(lines, i, start_token, end_token): + count = 1 + while i > 0: + i = find_tokens_backwards(lines, [start_token, end_token], i-1) + if i == -1: + return -1 + if check_token(lines[i], end_token): + count = count+1 + else: + count = count-1 + if count == 0: + return i + return -1 + def find_end_of(lines, i, start_token, end_token): count = 1 n = len(lines) while i < n: - i = find_tokens(lines, [end_token, start_token], i+1) - if check_token(lines[i], start_token): - count = count+1 - else: - count = count-1 - if count == 0: - return i + i = find_tokens(lines, [end_token, start_token], i+1) + if i == -1: + return -1 + if check_token(lines[i], start_token): + count = count+1 + else: + count = count-1 + if count == 0: + return i return -1 -# Finds the matching \end_inset -def find_beginning_of(lines, i, start_token, end_token): - count = 1 - while i > 0: - i = find_tokens_backwards(lines, [start_token, end_token], i-1) - if check_token(lines[i], end_token): - count = count+1 - else: - count = count-1 - if count == 0: - return i + +def find_nonempty_line(lines, start, end = 0): + if end == 0: + end = len(lines) + for i in xrange(start, end): + if is_nonempty_line(lines[i]): + return i return -1 -# Finds the matching \end_inset + def find_end_of_inset(lines, i): + " Find end of inset, where lines[i] is included." return find_end_of(lines, i, "\\begin_inset", "\\end_inset") -# Finds the matching \end_inset -def find_beginning_of_inset(lines, i): - return find_beginning_of(lines, i, "\\begin_inset", "\\end_inset") -def find_end_of_tabular(lines, i): - return find_end_of(lines, i, "