X-Git-Url: https://git.donarmstrong.com/?a=blobdiff_plain;f=python%2Fauxiliar%2Fpostprocess_html.py;h=b268db1d25924de6201ac93b4138e4f01cf2cd54;hb=5b54b66d529f663196e755ff5bb9650752a04539;hp=c26d7473e2efb8c3274e2299238c0686c571b95d;hpb=375c0be42300202d2f4b631f5302d817b8a91c60;p=lilypond.git diff --git a/python/auxiliar/postprocess_html.py b/python/auxiliar/postprocess_html.py index c26d7473e2..b268db1d25 100644 --- a/python/auxiliar/postprocess_html.py +++ b/python/auxiliar/postprocess_html.py @@ -16,10 +16,12 @@ import langdefs non_copied_pages = ['Documentation/out-www/notation-big-page', 'Documentation/out-www/internals-big-page', 'Documentation/out-www/learning-big-page', - 'Documentation/out-www/application-big-page', + 'Documentation/out-www/usage-big-page', 'Documentation/out-www/music-glossary-big-page', 'Documentation/out-www/contributor', - 'Documentation/out-www/changes', + 'Documentation/out-www/changes-big-page', + 'Documentation/out-www/essay-big-page', + 'Documentation/out-www/extending-big-page', 'Documentation/out-www/snippets', 'out-www/examples', 'Documentation/topdocs', @@ -100,7 +102,7 @@ def source_links_replace (m, source_val): return 'href="' + os.path.join (source_val, m.group (1)) + '"' # More hardcoding, yay! -splitted_docs_re = re.compile ('(Documentation/out-www/(automated-engraving|essay|notation|music-glossary|application|general|learning|snippets))/') +splitted_docs_re = re.compile('(Documentation/out-www/(automated-engraving|essay|notation|changes|music-glossary|usage|web|learning|snippets))/') lily_snippets_re = re.compile ('(href|src)="([0-9a-f]{2}/lily-.*?)"') pictures_re = re.compile ('src="(pictures/.*?)"') @@ -160,10 +162,8 @@ def add_header (s, prefix): if not n: s = header + s - s = header_tag + '\n' + s - if doctype_re.search (s) == None: - s = doctype + s + s = doctype + header_tag + '\n' + s if css_re.search (s) == None: depth = (prefix.count ('/') - 1) * '../' @@ -218,7 +218,7 @@ online_links_re = re.compile ('''(href|src)=['"]\ ((?!Compiling-from-source.html")[^/][.]*[^.:'"]*)\ ([.]html)(#[^"']*|)['"]''') offline_links_re = re.compile ('href=[\'"]\ -((?!Compiling-from-source.html")[^/][.]*[^.:\'"]*)([.]html)(#[^"\']*|)[\'"]') +((?!Compiling-from-source.html")(?![.]{2}/contributor)[^/][.]*[^.:\'"]*)([.]html)(#[^"\']*|)[\'"]') big_page_name_re = re.compile ('''(.+?)-big-page''') def process_i18n_big_page_links (match, prefix, lang_ext): @@ -348,7 +348,7 @@ def process_html_files (package_name = '', ### add footer if footer_tag_re.search (s) == None: - if 'general' in file_name: + if 'web' in file_name: s = add_footer (s, footer_tag + web_footer) else: s = add_footer (s, footer_tag + footer)