aboutsummaryrefslogtreecommitdiffhomepage
path: root/org/output_xmls.org
diff options
context:
space:
mode:
Diffstat (limited to 'org/output_xmls.org')
-rw-r--r--org/output_xmls.org50
1 files changed, 29 insertions, 21 deletions
diff --git a/org/output_xmls.org b/org/output_xmls.org
index 18a0315..2e9e72c 100644
--- a/org/output_xmls.org
+++ b/org/output_xmls.org
@@ -1110,7 +1110,7 @@ void scroll(D,I)(
foreach (obj; doc_abstraction[part]) {
string _txt = xhtml_format.special_characters(obj, obj.text);
switch (obj.use) {
- case "frontmatter":
+ case "frontmatter": assert(part == "head" || "toc_scroll");
switch (obj.is_of) {
case "para":
switch (obj.is_a) {
@@ -1134,7 +1134,8 @@ void scroll(D,I)(
break;
}
break;
- case "body":
+ case "body": // assert(part == "body");
+ assert(part == "body" || "head"); // surprise
switch (obj.is_of) {
case "para":
switch (obj.is_a) {
@@ -1188,25 +1189,26 @@ void scroll(D,I)(
}
break;
case "backmatter":
+ assert(part == "endnotes" || "glossary" || "bibliography" || "bookindex_scroll" || "blurb" || "tail");
switch (obj.is_of) {
case "para":
switch (obj.is_a) {
case "heading":
doc_html ~= xhtml_format.heading_scroll(obj, _txt, suffix);
break;
- case "endnote":
+ case "endnote": assert(part == "endnotes");
doc_html ~= xhtml_format.para_scroll(obj, _txt, suffix);
break;
- case "glossary":
+ case "glossary": assert(part == "glossary");
doc_html ~= xhtml_format.para_scroll(obj, _txt, suffix);
break;
- case "bibliography":
+ case "bibliography": assert(part == "bibliography");
doc_html ~= xhtml_format.para_scroll(obj, _txt, suffix);
break;
- case "bookindex":
+ case "bookindex": assert(part == "bookindex_scroll");
doc_html ~= xhtml_format.para_scroll(obj, _txt, suffix);
break;
- case "blurb":
+ case "blurb": assert(part == "blurb");
doc_html ~= xhtml_format.para_scroll(obj, _txt, suffix);
break;
default:
@@ -1293,6 +1295,7 @@ void seg(D,I)(
foreach (obj; doc_abstraction[part]) {
string _txt = xhtml_format.special_characters(obj, obj.text);
if (obj.is_a == "heading") {
+ assert(part == "head" || "toc_seg" || "body" || "endnotes" || "glossary" || "bibliography" || "bookindex_seg" || "blurb" || "tail");
switch (obj.heading_lev_markup) {
case 0: .. case 3:
/+ fill buffer, and replace with new levels from 1 to 3 +/
@@ -1350,8 +1353,9 @@ void seg(D,I)(
break;
}
} else {
+ assert(part == "head" || "toc_seg" || "body" || "endnotes" || "glossary" || "bibliography" || "bookindex_seg" || "blurb" || "tail");
switch (obj.use) {
- case "frontmatter":
+ case "frontmatter": assert(part == "head" || "toc_seg");
switch (obj.is_of) {
case "para":
switch (obj.is_a) {
@@ -1373,7 +1377,7 @@ void seg(D,I)(
break;
}
break;
- case "body":
+ case "body": assert(part == "body");
switch (obj.is_of) {
case "para":
switch (obj.is_a) {
@@ -1435,29 +1439,30 @@ void seg(D,I)(
}
break;
case "backmatter":
+ assert(part == "endnotes" || "glossary" || "bibliography" || "bookindex_seg" || "blurb" || "tail");
switch (obj.is_of) {
case "para":
switch (obj.is_a) {
- case "endnote":
+ case "endnote": assert(part == "endnotes");
auto t = xhtml_format.para_seg(obj, _txt, suffix);
doc_html[segment_filename] ~= t[0];
break;
- case "glossary":
+ case "glossary": assert(part == "glossary");
auto t = xhtml_format.para_seg(obj, _txt, suffix);
doc_html[segment_filename] ~= t[0];
doc_html_endnotes[segment_filename] ~= t[1];
break;
- case "bibliography":
+ case "bibliography": assert(part == "bibliography");
auto t = xhtml_format.para_seg(obj, _txt, suffix);
doc_html[segment_filename] ~= t[0];
doc_html_endnotes[segment_filename] ~= t[1];
break;
- case "bookindex":
+ case "bookindex": assert(part == "bookindex_seg");
auto t = xhtml_format.para_seg(obj, _txt, suffix);
doc_html[segment_filename] ~= t[0];
doc_html_endnotes[segment_filename] ~= t[1];
break;
- case "blurb":
+ case "blurb": assert(part == "blurb");
auto t = xhtml_format.para_seg(obj, _txt, suffix);
doc_html[segment_filename] ~= t[0];
doc_html_endnotes[segment_filename] ~= t[1];
@@ -1912,6 +1917,7 @@ void outputEPub3(D,I)(
foreach (obj; doc_abstraction[part]) {
string _txt = xhtml_format.special_characters(obj, obj.text);
if (obj.is_a == "heading") {
+ assert(part == "head" || "toc_seg" || "body" || "endnotes" || "glossary" || "bibliography" || "bookindex_seg" || "blurb" || "tail");
switch (obj.heading_lev_markup) {
case 0: .. case 3:
/+ fill buffer, and replace with new levels from 1 to 3 +/
@@ -1968,8 +1974,9 @@ void outputEPub3(D,I)(
break;
}
} else {
+ assert(part == "head" || "toc_seg" || "body" || "endnotes" || "glossary" || "bibliography" || "bookindex_seg" || "blurb" || "tail");
switch (obj.use) {
- case "frontmatter":
+ case "frontmatter": assert(part == "head" || "toc_seg");
switch (obj.is_of) {
case "para":
switch (obj.is_a) {
@@ -1992,7 +1999,7 @@ void outputEPub3(D,I)(
break;
}
break;
- case "body":
+ case "body": assert(part == "body");
switch (obj.is_of) {
case "para":
switch (obj.is_a) {
@@ -2055,29 +2062,30 @@ void outputEPub3(D,I)(
}
break;
case "backmatter":
+ assert(part == "endnotes" || "glossary" || "bibliography" || "bookindex_seg" || "blurb" || "tail");
switch (obj.is_of) {
case "para":
switch (obj.is_a) {
- case "endnote":
+ case "endnote": assert(part == "endnotes");
auto t = xhtml_format.para_seg(obj, _txt, suffix);
doc_epub3[segment_filename] ~= t[0];
break;
- case "glossary":
+ case "glossary": assert(part == "glossary");
auto t = xhtml_format.para_seg(obj, _txt, suffix);
doc_epub3[segment_filename] ~= t[0];
doc_epub3_endnotes[segment_filename] ~= t[1];
break;
- case "bibliography":
+ case "bibliography": assert(part == "bibliography");
auto t = xhtml_format.para_seg(obj, _txt, suffix);
doc_epub3[segment_filename] ~= t[0];
doc_epub3_endnotes[segment_filename] ~= t[1];
break;
- case "bookindex":
+ case "bookindex": assert(part == "bookindex_seg");
auto t = xhtml_format.para_seg(obj, _txt, suffix);
doc_epub3[segment_filename] ~= t[0];
doc_epub3_endnotes[segment_filename] ~= t[1];
break;
- case "blurb":
+ case "blurb": assert(part == "blurb");
auto t = xhtml_format.para_seg(obj, _txt, suffix);
doc_epub3[segment_filename] ~= t[0];
doc_epub3_endnotes[segment_filename] ~= t[1];