@natlibfi/marc-record-validators-melinda 12.0.7 → 12.0.8

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (101) hide show
  1. package/dist/addMissingField041.js +6 -3
  2. package/dist/addMissingField041.js.map +2 -2
  3. package/dist/addMissingField336.js +7 -4
  4. package/dist/addMissingField336.js.map +2 -2
  5. package/dist/addMissingField337.js +6 -3
  6. package/dist/addMissingField337.js.map +2 -2
  7. package/dist/addMissingField338.js +8 -5
  8. package/dist/addMissingField338.js.map +2 -2
  9. package/dist/cyrillux-usemarcon-replacement.js +5 -2
  10. package/dist/cyrillux-usemarcon-replacement.js.map +2 -2
  11. package/dist/cyrillux.js +10 -7
  12. package/dist/cyrillux.js.map +2 -2
  13. package/dist/disambiguateSeriesStatements.js +2 -1
  14. package/dist/disambiguateSeriesStatements.js.map +2 -2
  15. package/dist/drop-terms.js +5 -4
  16. package/dist/drop-terms.js.map +2 -2
  17. package/dist/fix-33X.js +7 -4
  18. package/dist/fix-33X.js.map +2 -2
  19. package/dist/fix-country-codes.js +5 -0
  20. package/dist/fix-country-codes.js.map +2 -2
  21. package/dist/fix-language-codes.js +5 -1
  22. package/dist/fix-language-codes.js.map +2 -2
  23. package/dist/fix-sami-041.js +11 -10
  24. package/dist/fix-sami-041.js.map +2 -2
  25. package/dist/indicator-fixes.js +5 -1
  26. package/dist/indicator-fixes.js.map +2 -2
  27. package/dist/merge-fields/counterpartField.js +6 -6
  28. package/dist/merge-fields/counterpartField.js.map +2 -2
  29. package/dist/merge-fields/mergableIndicator.js +0 -3
  30. package/dist/merge-fields/mergableIndicator.js.map +2 -2
  31. package/dist/merge-fields/worldKnowledge.js.map +2 -2
  32. package/dist/mergeRelatorTermFields.js +9 -6
  33. package/dist/mergeRelatorTermFields.js.map +2 -2
  34. package/dist/normalize-dashes.js +7 -4
  35. package/dist/normalize-dashes.js.map +2 -2
  36. package/dist/normalize-identifiers.js.map +2 -2
  37. package/dist/normalize-utf8-diacritics.js.map +2 -2
  38. package/dist/normalizeFieldForComparison.js.map +1 -1
  39. package/dist/normalizeSubfieldValueForComparison.js.map +1 -1
  40. package/dist/punctuation2.js +5 -2
  41. package/dist/punctuation2.js.map +2 -2
  42. package/dist/reindexSubfield6OccurenceNumbers.js +11 -10
  43. package/dist/reindexSubfield6OccurenceNumbers.js.map +2 -2
  44. package/dist/removeDuplicateDataFields.js +3 -2
  45. package/dist/removeDuplicateDataFields.js.map +2 -2
  46. package/dist/removeInferiorDataFields.js.map +2 -2
  47. package/dist/resolveOrphanedSubfield6s.js +3 -2
  48. package/dist/resolveOrphanedSubfield6s.js.map +2 -2
  49. package/dist/sortSubfields.js +1 -1
  50. package/dist/sortSubfields.js.map +2 -2
  51. package/dist/stripPunctuation.js +4 -3
  52. package/dist/stripPunctuation.js.map +2 -2
  53. package/dist/subfield6Utils.js +4 -1
  54. package/dist/subfield6Utils.js.map +2 -2
  55. package/dist/subfield8Utils.js.map +2 -2
  56. package/dist/translate-terms.js +4 -3
  57. package/dist/translate-terms.js.map +2 -2
  58. package/dist/typeOfDate-008.js +3 -1
  59. package/dist/typeOfDate-008.js.map +2 -2
  60. package/dist/update-field-540.js.map +2 -2
  61. package/dist/urn.js +13 -12
  62. package/dist/urn.js.map +2 -2
  63. package/package.json +7 -7
  64. package/src/addMissingField041.js +8 -4
  65. package/src/addMissingField336.js +10 -5
  66. package/src/addMissingField337.js +9 -5
  67. package/src/addMissingField338.js +11 -6
  68. package/src/cyrillux-usemarcon-replacement.js +9 -5
  69. package/src/cyrillux.js +18 -12
  70. package/src/disambiguateSeriesStatements.js +4 -1
  71. package/src/drop-terms.js +8 -6
  72. package/src/fix-33X.js +10 -6
  73. package/src/fix-country-codes.js +7 -3
  74. package/src/fix-language-codes.js +8 -4
  75. package/src/fix-sami-041.js +13 -11
  76. package/src/indicator-fixes.js +10 -7
  77. package/src/merge-fields/counterpartField.js +10 -10
  78. package/src/merge-fields/mergableIndicator.js +3 -3
  79. package/src/merge-fields/worldKnowledge.js +11 -6
  80. package/src/mergeRelatorTermFields.js +12 -11
  81. package/src/normalize-dashes.js +11 -5
  82. package/src/normalize-identifiers.js +12 -19
  83. package/src/normalize-utf8-diacritics.js +6 -3
  84. package/src/normalizeFieldForComparison.js +2 -2
  85. package/src/normalizeSubfieldValueForComparison.js +2 -2
  86. package/src/punctuation2.js +34 -30
  87. package/src/reindexSubfield6OccurenceNumbers.js +13 -11
  88. package/src/removeDuplicateDataFields.js +29 -27
  89. package/src/removeInferiorDataFields.js +28 -24
  90. package/src/resolveOrphanedSubfield6s.js +6 -4
  91. package/src/sortSubfields.js +5 -5
  92. package/src/stripPunctuation.js +5 -3
  93. package/src/subfield6Utils.js +33 -35
  94. package/src/subfield8Utils.js +10 -7
  95. package/src/translate-terms.js +13 -9
  96. package/src/typeOfDate-008.js +4 -1
  97. package/src/update-field-540.js +7 -5
  98. package/src/urn.js +17 -13
  99. package/test-fixtures/drop-terms/02/metadata.json +1 -1
  100. package/test-fixtures/drop-terms/03/metadata.json +1 -1
  101. package/test-fixtures/drop-terms/04/metadata.json +1 -1
@@ -1,7 +1,7 @@
1
1
  {
2
2
  "version": 3,
3
3
  "sources": ["../src/normalizeFieldForComparison.js"],
4
- "sourcesContent": ["/*\n Note that this file contains very powerful normalizations and spells that are:\n - meant for comparing similarity/mergability of two fields (clone, normalize, compare),\n - and NOT for modifying the actual field!\n\n This is mainly used by melinda-marc-record-merge-reducers. However, also removeInferiorDataFields fixer also used this.\n Thus it is here. However, most of the testing is done via merge-reducers...\n*/\nimport clone from 'clone';\nimport {fieldStripPunctuation} from './punctuation2.js';\nimport {fieldToString, isContentSubfieldCode} from './utils.js';\n\nimport {fieldNormalizeControlNumbers/*, normalizeControlSubfieldValue*/} from './normalize-identifiers.js';\nimport createDebugLogger from 'debug';\nimport {normalizePartData, subfieldContainsPartData} from './normalizeSubfieldValueForComparison.js';\nimport {isEnnakkotietoSubfield} from './prepublicationUtils.js';\nimport {getSynonym} from './merge-fields/worldKnowledge.js';\n\nconst debug = createDebugLogger('@natlibfi/melinda-marc-record-merge-reducers:normalizeFieldForComparison');\n//const debugData = debug.extend('data');\nconst debugDev = debug.extend('dev');\n\nfunction debugFieldComparison(oldField, newField) { // NB: Debug-only function!\n /*\n // We may drop certain subfields:\n if (oldField.subfields.length === newField.subfields.length) {\n oldField.subfields.forEach((subfield, index) => {\n const newValue = newField.subfields[index].value;\n if (subfield.value !== newValue) {\n nvdebug(`NORMALIZE SUBFIELD: '${subfield.value}' => '${newValue}'`, debugDev);\n }\n });\n }\n */\n const oldString = fieldToString(oldField);\n const newString = fieldToString(newField);\n if (oldString === newString) {\n return;\n }\n //nvdebug(`NORMALIZE FIELD:\\n '${fieldToString(oldField)}' =>\\n '${fieldToString(newField)}'`, debugDev);\n}\n\nfunction containsHumanName(tag = '???', subfieldCode = undefined) {\n // NB! This set is for bibs! Auth has 400... What else...\n if (['100', '600', '700', '800'].includes(tag)) {\n if (subfieldCode === undefined || subfieldCode === 'a') {\n return true;\n }\n }\n // Others?\n return false;\n}\n\nfunction containsCorporateName(tag = '???', subfieldCode = undefined) {\n // NB! This set is for bibs! Auth has 410... What else...\n if (['110', '610', '710', '810'].includes(tag)) {\n if (subfieldCode === undefined || subfieldCode === 'a') {\n return true;\n }\n }\n // Others?\n return false;\n}\n\nfunction skipAllSubfieldNormalizations(value, subfieldCode, tag) {\n\n if (isEnnakkotietoSubfield({'code': subfieldCode, value})) {\n return true;\n }\n\n if (tag === '035' && ['a', 'z'].includes(subfieldCode)) {\n return true;\n }\n\n if (!isContentSubfieldCode(subfieldCode, tag)) {\n return true;\n }\n return false;\n}\n\nfunction skipSubfieldLowercase(value, subfieldCode, tag) {\n // These may contain Roman Numerals...\n if (subfieldContainsPartData(tag, subfieldCode)) {\n return true;\n }\n\n return skipAllSubfieldNormalizations(value, subfieldCode, tag);\n}\n\nfunction skipAllFieldNormalizations(tag) {\n if (['LOW', 'SID'].includes(tag)) {\n return true;\n }\n return false;\n}\n\n\nfunction subfieldValueLowercase(value, subfieldCode, tag) {\n if (skipSubfieldLowercase(value, subfieldCode, tag)) {\n return value;\n }\n\n //return value.toLowerCase();\n const newValue = value.toLowerCase();\n if (newValue !== value) {\n //nvdebug(`SVL ${tag} $${subfieldCode} '${value}' =>`, debugDev);\n //nvdebug(`SVL ${tag} $${subfieldCode} '${newValue}'`, debugDev);\n return newValue;\n }\n return value;\n}\n\nfunction subfieldLowercase(sf, tag) {\n if (valuelessSubfield(sf)) {\n return;\n }\n sf.value = subfieldValueLowercase(sf.value, sf.code, tag);\n}\n\nfunction fieldLowercase(field) {\n if (skipFieldLowercase(field)) {\n return;\n }\n\n field.subfields.forEach(sf => subfieldLowercase(sf, field.tag));\n\n function skipFieldLowercase(field) {\n if (skipAllFieldNormalizations(field.tag)) {\n return true;\n }\n // Skip non-interesting fields\n if (!containsHumanName(field.tag) && !containsCorporateName(field.tag) && !['240', '245', '630'].includes(field.tag)) {\n return true;\n }\n\n return false;\n }\n}\n\n\nfunction hack490SubfieldA(field) {\n if (field.tag !== '490') {\n return;\n }\n field.subfields.forEach(sf => removeSarja(sf));\n\n // NB! This won't work, if the punctuation has not been stripped beforehand!\n function removeSarja(subfield) {\n if (valuelessSubfield(subfield)) {\n return;\n }\n\n if (subfield.code !== 'a') {\n return;\n }\n const tmp = subfield.value.replace(/ ?-(?:[a-z]|\u00E4|\u00F6)*sarja$/u, '');\n if (tmp.length > 0) {\n subfield.value = tmp;\n return;\n }\n }\n}\n\nexport function tagAndSubfieldCodeReferToIsbn(tag, subfieldCode) {\n // NB! We don't do this to 020$z!\n if (subfieldCode === 'z' && ['765', '767', '770', '772', '773', '774', '776', '777', '780', '785', '786', '787'].includes(tag)) {\n return true;\n }\n if (tag === '020' && subfieldCode === 'a') {\n return true;\n }\n return false;\n}\n\nfunction looksLikeIsbn(value) {\n // Does not check validity!\n if (value.match(/^(?:[0-9]-?){9}(?:[0-9]-?[0-9]-?[0-9]-?)?[0-9Xx]$/u)) {\n return true;\n }\n return false;\n}\n\nfunction normalizeISBN(field) {\n if (!field.subfields) {\n return;\n }\n\n //nvdebug(`ISBN-field? ${fieldToString(field)}`);\n const relevantSubfields = field.subfields.filter(sf => tagAndSubfieldCodeReferToIsbn(field.tag, sf.code) && looksLikeIsbn(sf.value));\n relevantSubfields.forEach(sf => normalizeIsbnSubfield(sf));\n\n function normalizeIsbnSubfield(sf) {\n if (valuelessSubfield(sf)) {\n return;\n }\n //nvdebug(` ISBN-subfield? ${subfieldToString(sf)}`);\n sf.value = sf.value.replace(/-/ug, '');\n sf.value = sf.value.replace(/x/u, 'X');\n }\n\n}\n\nfunction fieldSpecificHacks(field) {\n normalizeISBN(field); // 020$a, not $z!\n hack490SubfieldA(field);\n}\n\nexport function fieldTrimSubfieldValues(field) {\n field.subfields?.forEach((sf) => {\n if (valuelessSubfield(sf)) {\n return;\n }\n sf.value = sf.value.replace(/^[ \\t\\n]+/u, '');\n sf.value = sf.value.replace(/[ \\t\\n]+$/u, '');\n sf.value = sf.value.replace(/[ \\t\\n]+/gu, ' ');\n });\n}\n\nfunction fieldRemoveDecomposedDiacritics(field) {\n // Raison d'\u00EAtre/motivation: \"Sir\u00E9n\" and diacriticless \"Siren\" might refer to a same surname, so this normalization\n // allows us to compare authors and avoid duplicate fields.\n field.subfields.forEach((sf) => {\n if (valuelessSubfield(sf)) {\n return;\n }\n sf.value = removeDecomposedDiacritics(sf.value);\n });\n}\n\nfunction removeDecomposedDiacritics(value = '') {\n // NB #1: Does nothing to precomposed letters. Do String.normalize('NFD') first, if you want to handle them.\n // NB #2: Finnish letters '\u00E5', '\u00E4', '\u00F6', '\u00C5', \u00C4', and '\u00D6' should be handled (=precomposed) before calling this. (= keep them as is)\n // NB #3: Calling our very own fixComposition() before this function handles both #1 and #2.\n return String(value).replace(/\\p{Diacritic}/gu, '');\n}\n\nfunction normalizeSubfieldValue(value, subfieldCode, tag) {\n // NB! For comparison of values only\n /* beslint-disable */\n value = removeCharsThatDontCarryMeaning(value, tag, subfieldCode);\n value = getSynonym(tag, subfieldCode, value); // Must be done before punc stripping and lowercasing...\n value = subfieldValueLowercase(value, subfieldCode, tag);\n\n\n // Normalize: s. = sivut = pp.\n value = normalizePartData(value, subfieldCode, tag);\n value = value.replace(/^\\[([^[\\]]+)\\]/gu, '$1');\n\n if (['130', '730'].includes(tag) && subfieldCode === 'a') {\n value = value.replace(' : ', ', '); // \"Halloween ends (elokuva, 2022)\" vs \"Halloween ends (elokuva : 2023)\"\n }\n /* beslint-enable */\n\n // Not going to do these in the foreseeable future, but keeping them here for discussion:\n // Possible normalizations include but are not limited to:\n // \u00F8 => \u00F6? Might be language dependent: 041 $a fin => \u00F6, 041 $a eng => o?\n // \u00D8 => \u00D6?\n // \u00DF => ss\n // \u00FE => th (NB! Both upper and lower case)\n // ...\n // Probably nots:\n // \u00FC => y (probably not, though this correlates with Finnish letter-to-sound rules)\n // w => v (OK for Finnish sorting in certain cases, but we are not here, are we?)\n // I guess we should use decomposed values in code here. (Not sure what composition my examples above use.)\n return value;\n}\n\nexport function cloneAndRemovePunctuation(field) {\n const clonedField = clone(field);\n if (fieldSkipNormalization(field)) {\n return clonedField;\n }\n fieldStripPunctuation(clonedField);\n fieldTrimSubfieldValues(clonedField);\n debugDev('PUNC');\n debugFieldComparison(field, clonedField);\n\n return clonedField;\n}\n\nfunction removeCharsThatDontCarryMeaning(value, tag, subfieldCode) {\n if (tag === '080') {\n return value;\n }\n\n // 3\" refers to inches, but as this is for comparison only we don't mind...\n value = value.replace(/['\u2018\u2019\"\u201E\u201C\u201D\u00AB\u00BB]/gu, ''); // MET-570 et al. Subset of https://hexdocs.pm/ex_unicode/Unicode.Category.QuoteMarks.html\n // MRA-273: Handle X00$a name initials.\n // NB #1: that we remove spaces for comparison (as it simpler), though actually space should be used. Doesn't matter as this is comparison only.\n // NB #2: we might/should eventually write a validator/fixer that adds those spaces. After that point, this expection should become obsolete.\n if (subfieldCode === 'a' && ['100', '400', '600', '700', '800'].includes(tag)) { // 400 is used in auth records. It's not a bib field at all.\n value = value.replace(/([A-Z]|\u00C5|\u00C4|\u00D6)\\. +/ugi, '$1.');\n }\n\n return value;\n}\n\nfunction normalizeField(field) {\n //sf.value = removeDecomposedDiacritics(sf.value);\n fieldStripPunctuation(field);\n fieldLowercase(field);\n fieldNormalizeControlNumbers(field); // FIN11 vs FI-MELINDA etc.\n return field;\n}\n\nexport function cloneAndNormalizeFieldForComparison(field) {\n // NB! This new field is for comparison purposes only.\n // Some of the normalizations might be considered a bit overkill for other purposes.\n const clonedField = cloneAndRemovePunctuation(field); // was only clone(field)\n\n if (fieldSkipNormalization(field)) {\n return clonedField;\n }\n clonedField.subfields.forEach((sf) => { // Do this for all fields or some fields?\n if (valuelessSubfield(sf)) {\n return;\n }\n\n sf.value = normalizeSubfieldValue(sf.value, sf.code, field.tag);\n //sf.value = normalizeForSamenessCheck(field.tag, sf.code, sf.value);\n\n\n });\n\n normalizeField(clonedField);\n fieldRemoveDecomposedDiacritics(clonedField);\n fieldSpecificHacks(clonedField);\n fieldTrimSubfieldValues(clonedField);\n\n\n debugFieldComparison(field, clonedField); // For debugging purposes only\n\n return clonedField;\n}\n\nfunction fieldSkipNormalization(field) {\n if (!field.subfields || ['018', '066', '080', '083'].includes(field.tag)) {\n return true;\n }\n return false;\n}\n\nfunction valuelessSubfield(sf) {\n return sf.value === undefined;\n}"],
4
+ "sourcesContent": ["/*\n Note that this file contains very powerful normalizations and spells that are:\n - meant for comparing similarity/mergability of two fields (clone, normalize, compare),\n - and NOT for modifying the actual field!\n\n This is mainly used by melinda-marc-record-merge-reducers. However, also removeInferiorDataFields fixer also used this.\n Thus it is here. However, most of the testing is done via merge-reducers...\n*/\nimport clone from 'clone';\nimport {fieldStripPunctuation} from './punctuation2.js';\nimport {fieldToString, isContentSubfieldCode} from './utils.js';\n\nimport {fieldNormalizeControlNumbers/*, normalizeControlSubfieldValue*/} from './normalize-identifiers.js';\nimport createDebugLogger from 'debug';\nimport {normalizePartData, subfieldContainsPartData} from './normalizeSubfieldValueForComparison.js';\nimport {isEnnakkotietoSubfield} from './prepublicationUtils.js';\nimport {getSynonym} from './merge-fields/worldKnowledge.js';\n\nconst debug = createDebugLogger('@natlibfi/melinda-marc-record-merge-reducers:normalizeFieldForComparison');\n//const debugData = debug.extend('data');\nconst debugDev = debug.extend('dev');\n\nfunction debugFieldComparison(oldField, newField) { // NB: Debug-only function!\n /*\n // We may drop certain subfields:\n if (oldField.subfields.length === newField.subfields.length) {\n oldField.subfields.forEach((subfield, index) => {\n const newValue = newField.subfields[index].value;\n if (subfield.value !== newValue) {\n nvdebug(`NORMALIZE SUBFIELD: '${subfield.value}' => '${newValue}'`, debugDev);\n }\n });\n }\n */\n const oldString = fieldToString(oldField);\n const newString = fieldToString(newField);\n if (oldString === newString) {\n return;\n }\n //nvdebug(`NORMALIZE FIELD:\\n '${fieldToString(oldField)}' =>\\n '${fieldToString(newField)}'`, debugDev);\n}\n\nfunction containsHumanName(tag = '???', subfieldCode = undefined) {\n // NB! This set is for bibs! Auth has 400... What else...\n if (['100', '600', '700', '800'].includes(tag)) {\n if (subfieldCode === undefined || subfieldCode === 'a') {\n return true;\n }\n }\n // Others?\n return false;\n}\n\nfunction containsCorporateName(tag = '???', subfieldCode = undefined) {\n // NB! This set is for bibs! Auth has 410... What else...\n if (['110', '610', '710', '810'].includes(tag)) {\n if (subfieldCode === undefined || subfieldCode === 'a') {\n return true;\n }\n }\n // Others?\n return false;\n}\n\nfunction skipAllSubfieldNormalizations(value, subfieldCode, tag) {\n\n if (isEnnakkotietoSubfield({'code': subfieldCode, value})) {\n return true;\n }\n\n if (tag === '035' && ['a', 'z'].includes(subfieldCode)) {\n return true;\n }\n\n if (!isContentSubfieldCode(subfieldCode, tag)) {\n return true;\n }\n return false;\n}\n\nfunction skipSubfieldLowercase(value, subfieldCode, tag) {\n // These may contain Roman Numerals...\n if (subfieldContainsPartData(tag, subfieldCode)) {\n return true;\n }\n\n return skipAllSubfieldNormalizations(value, subfieldCode, tag);\n}\n\nfunction skipAllFieldNormalizations(tag) {\n if (['LOW', 'SID'].includes(tag)) {\n return true;\n }\n return false;\n}\n\n\nfunction subfieldValueLowercase(value, subfieldCode, tag) {\n if (skipSubfieldLowercase(value, subfieldCode, tag)) {\n return value;\n }\n\n //return value.toLowerCase();\n const newValue = value.toLowerCase();\n if (newValue !== value) {\n //nvdebug(`SVL ${tag} $${subfieldCode} '${value}' =>`, debugDev);\n //nvdebug(`SVL ${tag} $${subfieldCode} '${newValue}'`, debugDev);\n return newValue;\n }\n return value;\n}\n\nfunction subfieldLowercase(sf, tag) {\n if (valuelessSubfield(sf)) {\n return;\n }\n sf.value = subfieldValueLowercase(sf.value, sf.code, tag);\n}\n\nfunction fieldLowercase(field) {\n if (skipFieldLowercase(field)) {\n return;\n }\n\n field.subfields.forEach(sf => subfieldLowercase(sf, field.tag));\n\n function skipFieldLowercase(field) {\n if (skipAllFieldNormalizations(field.tag)) {\n return true;\n }\n // Skip non-interesting fields\n if (!containsHumanName(field.tag) && !containsCorporateName(field.tag) && !['240', '245', '630'].includes(field.tag)) {\n return true;\n }\n\n return false;\n }\n}\n\n\nfunction hack490SubfieldA(field) {\n if (field.tag !== '490') {\n return;\n }\n field.subfields.forEach(sf => removeSarja(sf));\n\n // NB! This won't work, if the punctuation has not been stripped beforehand!\n function removeSarja(subfield) {\n if (valuelessSubfield(subfield)) {\n return;\n }\n\n if (subfield.code !== 'a') {\n return;\n }\n const tmp = subfield.value.replace(/ ?-(?:[a-z]|\u00E4|\u00F6)*sarja$/u, '');\n if (tmp.length > 0) {\n subfield.value = tmp;\n return;\n }\n }\n}\n\nexport function tagAndSubfieldCodeReferToIsbn(tag, subfieldCode) {\n // NB! We don't do this to 020$z!\n if (subfieldCode === 'z' && ['765', '767', '770', '772', '773', '774', '776', '777', '780', '785', '786', '787'].includes(tag)) {\n return true;\n }\n if (tag === '020' && subfieldCode === 'a') {\n return true;\n }\n return false;\n}\n\nfunction looksLikeIsbn(value) {\n // Does not check validity!\n if (value.match(/^(?:[0-9]-?){9}(?:[0-9]-?[0-9]-?[0-9]-?)?[0-9Xx]$/u)) {\n return true;\n }\n return false;\n}\n\nfunction normalizeISBN(field) {\n if (!field.subfields) {\n return;\n }\n\n //nvdebug(`ISBN-field? ${fieldToString(field)}`, debugDev);\n const relevantSubfields = field.subfields.filter(sf => tagAndSubfieldCodeReferToIsbn(field.tag, sf.code) && looksLikeIsbn(sf.value));\n relevantSubfields.forEach(sf => normalizeIsbnSubfield(sf));\n\n function normalizeIsbnSubfield(sf) {\n if (valuelessSubfield(sf)) {\n return;\n }\n //nvdebug(` ISBN-subfield? ${subfieldToString(sf)}`, debugDev);\n sf.value = sf.value.replace(/-/ug, '');\n sf.value = sf.value.replace(/x/u, 'X');\n }\n\n}\n\nfunction fieldSpecificHacks(field) {\n normalizeISBN(field); // 020$a, not $z!\n hack490SubfieldA(field);\n}\n\nexport function fieldTrimSubfieldValues(field) {\n field.subfields?.forEach((sf) => {\n if (valuelessSubfield(sf)) {\n return;\n }\n sf.value = sf.value.replace(/^[ \\t\\n]+/u, '');\n sf.value = sf.value.replace(/[ \\t\\n]+$/u, '');\n sf.value = sf.value.replace(/[ \\t\\n]+/gu, ' ');\n });\n}\n\nfunction fieldRemoveDecomposedDiacritics(field) {\n // Raison d'\u00EAtre/motivation: \"Sir\u00E9n\" and diacriticless \"Siren\" might refer to a same surname, so this normalization\n // allows us to compare authors and avoid duplicate fields.\n field.subfields.forEach((sf) => {\n if (valuelessSubfield(sf)) {\n return;\n }\n sf.value = removeDecomposedDiacritics(sf.value);\n });\n}\n\nfunction removeDecomposedDiacritics(value = '') {\n // NB #1: Does nothing to precomposed letters. Do String.normalize('NFD') first, if you want to handle them.\n // NB #2: Finnish letters '\u00E5', '\u00E4', '\u00F6', '\u00C5', \u00C4', and '\u00D6' should be handled (=precomposed) before calling this. (= keep them as is)\n // NB #3: Calling our very own fixComposition() before this function handles both #1 and #2.\n return String(value).replace(/\\p{Diacritic}/gu, '');\n}\n\nfunction normalizeSubfieldValue(value, subfieldCode, tag) {\n // NB! For comparison of values only\n /* beslint-disable */\n value = removeCharsThatDontCarryMeaning(value, tag, subfieldCode);\n value = getSynonym(tag, subfieldCode, value); // Must be done before punc stripping and lowercasing...\n value = subfieldValueLowercase(value, subfieldCode, tag);\n\n\n // Normalize: s. = sivut = pp.\n value = normalizePartData(value, subfieldCode, tag);\n value = value.replace(/^\\[([^[\\]]+)\\]/gu, '$1');\n\n if (['130', '730'].includes(tag) && subfieldCode === 'a') {\n value = value.replace(' : ', ', '); // \"Halloween ends (elokuva, 2022)\" vs \"Halloween ends (elokuva : 2023)\"\n }\n /* beslint-enable */\n\n // Not going to do these in the foreseeable future, but keeping them here for discussion:\n // Possible normalizations include but are not limited to:\n // \u00F8 => \u00F6? Might be language dependent: 041 $a fin => \u00F6, 041 $a eng => o?\n // \u00D8 => \u00D6?\n // \u00DF => ss\n // \u00FE => th (NB! Both upper and lower case)\n // ...\n // Probably nots:\n // \u00FC => y (probably not, though this correlates with Finnish letter-to-sound rules)\n // w => v (OK for Finnish sorting in certain cases, but we are not here, are we?)\n // I guess we should use decomposed values in code here. (Not sure what composition my examples above use.)\n return value;\n}\n\nexport function cloneAndRemovePunctuation(field) {\n const clonedField = clone(field);\n if (fieldSkipNormalization(field)) {\n return clonedField;\n }\n fieldStripPunctuation(clonedField);\n fieldTrimSubfieldValues(clonedField);\n debugDev('PUNC');\n debugFieldComparison(field, clonedField);\n\n return clonedField;\n}\n\nfunction removeCharsThatDontCarryMeaning(value, tag, subfieldCode) {\n if (tag === '080') {\n return value;\n }\n\n // 3\" refers to inches, but as this is for comparison only we don't mind...\n value = value.replace(/['\u2018\u2019\"\u201E\u201C\u201D\u00AB\u00BB]/gu, ''); // MET-570 et al. Subset of https://hexdocs.pm/ex_unicode/Unicode.Category.QuoteMarks.html\n // MRA-273: Handle X00$a name initials.\n // NB #1: that we remove spaces for comparison (as it simpler), though actually space should be used. Doesn't matter as this is comparison only.\n // NB #2: we might/should eventually write a validator/fixer that adds those spaces. After that point, this expection should become obsolete.\n if (subfieldCode === 'a' && ['100', '400', '600', '700', '800'].includes(tag)) { // 400 is used in auth records. It's not a bib field at all.\n value = value.replace(/([A-Z]|\u00C5|\u00C4|\u00D6)\\. +/ugi, '$1.');\n }\n\n return value;\n}\n\nfunction normalizeField(field) {\n //sf.value = removeDecomposedDiacritics(sf.value);\n fieldStripPunctuation(field);\n fieldLowercase(field);\n fieldNormalizeControlNumbers(field); // FIN11 vs FI-MELINDA etc.\n return field;\n}\n\nexport function cloneAndNormalizeFieldForComparison(field) {\n // NB! This new field is for comparison purposes only.\n // Some of the normalizations might be considered a bit overkill for other purposes.\n const clonedField = cloneAndRemovePunctuation(field); // was only clone(field)\n\n if (fieldSkipNormalization(field)) {\n return clonedField;\n }\n clonedField.subfields.forEach((sf) => { // Do this for all fields or some fields?\n if (valuelessSubfield(sf)) {\n return;\n }\n\n sf.value = normalizeSubfieldValue(sf.value, sf.code, field.tag);\n //sf.value = normalizeForSamenessCheck(field.tag, sf.code, sf.value);\n\n\n });\n\n normalizeField(clonedField);\n fieldRemoveDecomposedDiacritics(clonedField);\n fieldSpecificHacks(clonedField);\n fieldTrimSubfieldValues(clonedField);\n\n\n debugFieldComparison(field, clonedField); // For debugging purposes only\n\n return clonedField;\n}\n\nfunction fieldSkipNormalization(field) {\n if (!field.subfields || ['018', '066', '080', '083'].includes(field.tag)) {\n return true;\n }\n return false;\n}\n\nfunction valuelessSubfield(sf) {\n return sf.value === undefined;\n}"],
5
5
  "mappings": "AAQA,OAAO,WAAW;AAClB,SAAQ,6BAA4B;AACpC,SAAQ,eAAe,6BAA4B;AAEnD,SAAQ,oCAAsE;AAC9E,OAAO,uBAAuB;AAC9B,SAAQ,mBAAmB,gCAA+B;AAC1D,SAAQ,8BAA6B;AACrC,SAAQ,kBAAiB;AAEzB,MAAM,QAAQ,kBAAkB,0EAA0E;AAE1G,MAAM,WAAW,MAAM,OAAO,KAAK;AAEnC,SAAS,qBAAqB,UAAU,UAAU;AAYhD,QAAM,YAAY,cAAc,QAAQ;AACxC,QAAM,YAAY,cAAc,QAAQ;AACxC,MAAI,cAAc,WAAW;AAC3B;AAAA,EACF;AAEF;AAEA,SAAS,kBAAkB,MAAM,OAAO,eAAe,QAAW;AAEhE,MAAI,CAAC,OAAO,OAAO,OAAO,KAAK,EAAE,SAAS,GAAG,GAAG;AAC9C,QAAI,iBAAiB,UAAa,iBAAiB,KAAK;AACtD,aAAO;AAAA,IACT;AAAA,EACF;AAEA,SAAO;AACT;AAEA,SAAS,sBAAsB,MAAM,OAAO,eAAe,QAAW;AAEpE,MAAI,CAAC,OAAO,OAAO,OAAO,KAAK,EAAE,SAAS,GAAG,GAAG;AAC9C,QAAI,iBAAiB,UAAa,iBAAiB,KAAK;AACtD,aAAO;AAAA,IACT;AAAA,EACF;AAEA,SAAO;AACT;AAEA,SAAS,8BAA8B,OAAO,cAAc,KAAK;AAE/D,MAAI,uBAAuB,EAAC,QAAQ,cAAc,MAAK,CAAC,GAAG;AACzD,WAAO;AAAA,EACT;AAEA,MAAI,QAAQ,SAAS,CAAC,KAAK,GAAG,EAAE,SAAS,YAAY,GAAG;AACtD,WAAO;AAAA,EACT;AAEA,MAAI,CAAC,sBAAsB,cAAc,GAAG,GAAG;AAC7C,WAAO;AAAA,EACT;AACA,SAAO;AACT;AAEA,SAAS,sBAAsB,OAAO,cAAc,KAAK;AAEvD,MAAI,yBAAyB,KAAK,YAAY,GAAG;AAC/C,WAAO;AAAA,EACT;AAEA,SAAO,8BAA8B,OAAO,cAAc,GAAG;AAC/D;AAEA,SAAS,2BAA2B,KAAK;AACvC,MAAI,CAAC,OAAO,KAAK,EAAE,SAAS,GAAG,GAAG;AAChC,WAAO;AAAA,EACT;AACA,SAAO;AACT;AAGA,SAAS,uBAAuB,OAAO,cAAc,KAAK;AACxD,MAAI,sBAAsB,OAAO,cAAc,GAAG,GAAG;AACnD,WAAO;AAAA,EACT;AAGA,QAAM,WAAW,MAAM,YAAY;AACnC,MAAI,aAAa,OAAO;AAGtB,WAAO;AAAA,EACT;AACA,SAAO;AACT;AAEA,SAAS,kBAAkB,IAAI,KAAK;AAClC,MAAI,kBAAkB,EAAE,GAAG;AACzB;AAAA,EACF;AACA,KAAG,QAAQ,uBAAuB,GAAG,OAAO,GAAG,MAAM,GAAG;AAC1D;AAEA,SAAS,eAAe,OAAO;AAC7B,MAAI,mBAAmB,KAAK,GAAG;AAC7B;AAAA,EACF;AAEA,QAAM,UAAU,QAAQ,QAAM,kBAAkB,IAAI,MAAM,GAAG,CAAC;AAE9D,WAAS,mBAAmBA,QAAO;AACjC,QAAI,2BAA2BA,OAAM,GAAG,GAAG;AACzC,aAAO;AAAA,IACT;AAEA,QAAI,CAAC,kBAAkBA,OAAM,GAAG,KAAK,CAAC,sBAAsBA,OAAM,GAAG,KAAK,CAAC,CAAC,OAAO,OAAO,KAAK,EAAE,SAASA,OAAM,GAAG,GAAG;AACpH,aAAO;AAAA,IACT;AAEA,WAAO;AAAA,EACT;AACF;AAGA,SAAS,iBAAiB,OAAO;AAC/B,MAAI,MAAM,QAAQ,OAAO;AACvB;AAAA,EACF;AACA,QAAM,UAAU,QAAQ,QAAM,YAAY,EAAE,CAAC;AAG7C,WAAS,YAAY,UAAU;AAC7B,QAAI,kBAAkB,QAAQ,GAAG;AAC/B;AAAA,IACF;AAEA,QAAI,SAAS,SAAS,KAAK;AACzB;AAAA,IACF;AACA,UAAM,MAAM,SAAS,MAAM,QAAQ,4BAA4B,EAAE;AACjE,QAAI,IAAI,SAAS,GAAG;AAClB,eAAS,QAAQ;AACjB;AAAA,IACF;AAAA,EACF;AACF;AAEO,gBAAS,8BAA8B,KAAK,cAAc;AAE/D,MAAI,iBAAiB,OAAO,CAAC,OAAO,OAAO,OAAO,OAAO,OAAO,OAAO,OAAO,OAAO,OAAO,OAAO,OAAO,KAAK,EAAE,SAAS,GAAG,GAAG;AAC9H,WAAO;AAAA,EACT;AACA,MAAI,QAAQ,SAAS,iBAAiB,KAAK;AACzC,WAAO;AAAA,EACT;AACA,SAAO;AACT;AAEA,SAAS,cAAc,OAAO;AAE5B,MAAI,MAAM,MAAM,oDAAoD,GAAG;AACrE,WAAO;AAAA,EACT;AACA,SAAO;AACT;AAEA,SAAS,cAAc,OAAO;AAC5B,MAAI,CAAC,MAAM,WAAW;AACpB;AAAA,EACF;AAGA,QAAM,oBAAoB,MAAM,UAAU,OAAO,QAAM,8BAA8B,MAAM,KAAK,GAAG,IAAI,KAAK,cAAc,GAAG,KAAK,CAAC;AACnI,oBAAkB,QAAQ,QAAM,sBAAsB,EAAE,CAAC;AAEzD,WAAS,sBAAsB,IAAI;AACjC,QAAI,kBAAkB,EAAE,GAAG;AACzB;AAAA,IACF;AAEA,OAAG,QAAQ,GAAG,MAAM,QAAQ,OAAO,EAAE;AACrC,OAAG,QAAQ,GAAG,MAAM,QAAQ,MAAM,GAAG;AAAA,EACvC;AAEF;AAEA,SAAS,mBAAmB,OAAO;AACjC,gBAAc,KAAK;AACnB,mBAAiB,KAAK;AACxB;AAEO,gBAAS,wBAAwB,OAAO;AAC7C,QAAM,WAAW,QAAQ,CAAC,OAAO;AAC/B,QAAI,kBAAkB,EAAE,GAAG;AACzB;AAAA,IACF;AACA,OAAG,QAAQ,GAAG,MAAM,QAAQ,cAAc,EAAE;AAC5C,OAAG,QAAQ,GAAG,MAAM,QAAQ,cAAc,EAAE;AAC5C,OAAG,QAAQ,GAAG,MAAM,QAAQ,cAAc,GAAG;AAAA,EAC/C,CAAC;AACH;AAEA,SAAS,gCAAgC,OAAO;AAG9C,QAAM,UAAU,QAAQ,CAAC,OAAO;AAC9B,QAAI,kBAAkB,EAAE,GAAG;AACvB;AAAA,IACJ;AACA,OAAG,QAAQ,2BAA2B,GAAG,KAAK;AAAA,EAChD,CAAC;AACH;AAEA,SAAS,2BAA2B,QAAQ,IAAI;AAI9C,SAAO,OAAO,KAAK,EAAE,QAAQ,mBAAmB,EAAE;AACpD;AAEA,SAAS,uBAAuB,OAAO,cAAc,KAAK;AAGxD,UAAQ,gCAAgC,OAAO,KAAK,YAAY;AAChE,UAAQ,WAAW,KAAK,cAAc,KAAK;AAC3C,UAAQ,uBAAuB,OAAO,cAAc,GAAG;AAIvD,UAAQ,kBAAkB,OAAO,cAAc,GAAG;AAClD,UAAQ,MAAM,QAAQ,oBAAoB,IAAI;AAE9C,MAAI,CAAC,OAAO,KAAK,EAAE,SAAS,GAAG,KAAK,iBAAiB,KAAK;AACxD,YAAQ,MAAM,QAAQ,OAAO,IAAI;AAAA,EACnC;AAcA,SAAO;AACT;AAEO,gBAAS,0BAA0B,OAAO;AAC/C,QAAM,cAAc,MAAM,KAAK;AAC/B,MAAI,uBAAuB,KAAK,GAAG;AACjC,WAAO;AAAA,EACT;AACA,wBAAsB,WAAW;AACjC,0BAAwB,WAAW;AACnC,WAAS,MAAM;AACf,uBAAqB,OAAO,WAAW;AAEvC,SAAO;AACT;AAEA,SAAS,gCAAgC,OAAO,KAAK,cAAc;AACjE,MAAI,QAAQ,OAAO;AACjB,WAAO;AAAA,EACT;AAGA,UAAQ,MAAM,QAAQ,iBAAiB,EAAE;AAIzC,MAAI,iBAAiB,OAAO,CAAC,OAAO,OAAO,OAAO,OAAO,KAAK,EAAE,SAAS,GAAG,GAAG;AAC7E,YAAQ,MAAM,QAAQ,wBAAwB,KAAK;AAAA,EACrD;AAEA,SAAO;AACT;AAEA,SAAS,eAAe,OAAO;AAE7B,wBAAsB,KAAK;AAC3B,iBAAe,KAAK;AACpB,+BAA6B,KAAK;AAClC,SAAO;AACT;AAEO,gBAAS,oCAAoC,OAAO;AAGzD,QAAM,cAAc,0BAA0B,KAAK;AAEnD,MAAI,uBAAuB,KAAK,GAAG;AACjC,WAAO;AAAA,EACT;AACA,cAAY,UAAU,QAAQ,CAAC,OAAO;AACpC,QAAI,kBAAkB,EAAE,GAAG;AACzB;AAAA,IACF;AAEA,OAAG,QAAQ,uBAAuB,GAAG,OAAO,GAAG,MAAM,MAAM,GAAG;AAAA,EAIhE,CAAC;AAED,iBAAe,WAAW;AAC1B,kCAAgC,WAAW;AAC3C,qBAAmB,WAAW;AAC9B,0BAAwB,WAAW;AAGnC,uBAAqB,OAAO,WAAW;AAEvC,SAAO;AACT;AAEA,SAAS,uBAAuB,OAAO;AACrC,MAAI,CAAC,MAAM,aAAa,CAAC,OAAO,OAAO,OAAO,KAAK,EAAE,SAAS,MAAM,GAAG,GAAG;AACxE,WAAO;AAAA,EACT;AACA,SAAO;AACT;AAEA,SAAS,kBAAkB,IAAI;AAC7B,SAAO,GAAG,UAAU;AACtB;",
6
6
  "names": ["field"]
7
7
  }
@@ -1,7 +1,7 @@
1
1
  {
2
2
  "version": 3,
3
3
  "sources": ["../src/normalizeSubfieldValueForComparison.js"],
4
- "sourcesContent": ["import {nvdebug} from './utils.js';\nimport createDebugLogger from 'debug';\n\n// Normalizes at least 490$v and 773$g which contain information such as \"Raita 5\" vs \"5\", and \"Osa 3\" vs \"Osa III\".\n\nconst debug = createDebugLogger('@natlibfi/melinda-marc-record-merge-reducers:normalizeSubfieldValueForComparison');\n//const debugData = debug.extend('data');\nconst debugDev = debug.extend('dev');\n\nexport function subfieldContainsPartData(tag, subfieldCode) {\n // NB! Used by reducers' mergeSubield.js\n if (subfieldCode === 'v' && ['490', '800', '810', '811', '830'].includes(tag)) {\n return true;\n }\n if (tag === '773' && subfieldCode === 'g') {\n return true;\n }\n return false;\n}\n\nfunction splitPartData(originalValue) {\n // This a very hacky function, but cand really help it, as the the data is very iffy as well...\n // Remove punctuation and brackets:\n const value = originalValue.replace(/[-.,:; ]+$/ui, '').replace(/^\\[([0-9]+)\\]$/ui, '$1');\n\n const [year, rest] = extractYear(value);\n\n const splitPoint = rest.lastIndexOf(' '); // MRA-627: \"5, 2017\" should be split here. Think of this later on...\n if (splitPoint === -1) {\n return [undefined, year, rest];\n }\n const lhs = rest.substr(0, splitPoint);\n const rhs = rest.substr(splitPoint + 1);\n return [lhs, year, rhs];\n\n function extractYear(value) {\n // NB! Note that this is far for perfect. It cover just some very common cases...\n\n // \"2023, 3\" => [\"2023\", \"3\"]\n if (value.match(/^(?:1[89][0-9][0-9]|20[012][0-9]), (?:nro |n:o)?[1-9][0-9]{0,2}$/ui)) {\n return [value.substr(0, 4), value.substr(6)];\n }\n // \"2023/12\" => [\"2023\", \"12\"]\n if (value.match(/^(?:1[89][0-9][0-9]|20[012][0-9])[/:][1-9][0-9]{0,2}$/u)) {\n return [value.substr(0, 4), value.substr(5)];\n }\n // \"Vol. 3/2023\" => [\"2023\", \"Vol. 3\"]\n if (value.match(/^[^0-9]*[1-9][0-9]{0,2}\\/(?:1[89][0-9][0-9]|20[012][0-9])$/u)) {\n const len = value.length;\n return [value.substr(len - 4), value.substr(0, len - 5)];\n }\n\n\n return [undefined, value];\n }\n}\n\nfunction normalizePartType(originalValue) {\n if (originalValue === undefined) {\n return undefined;\n }\n const value = originalValue.toLowerCase();\n\n // Return Finnish singular nominative. Choise of language is arbitrary. This is best-ish for debug purposes...\n if (['n:o', 'no', 'nr', 'nro', 'number', 'numero', 'nummer'].includes(value)) {\n return 'numero';\n }\n if (['band', 'bd', 'h\u00E4fte', 'nide', 'osa', 'part', 'teil', 'vol', 'vol.', 'volume'].includes(value)) {\n return 'osa';\n }\n\n if (['p.', 'page', 'pages', 'pp.', 's.', 'sidor', 'sivu', 'sivut'].includes(value)) {\n return 'sivu';\n }\n\n return value;\n}\n\nconst romanNumbers = {'I': '1', 'II': '2', 'III': '3', 'IV': '4', 'V': '5', 'VI': '6', 'X': '10'};\n\nfunction normalizePartNumber(value) {\n // Should we handle all Roman numbers or some range of them?\n // There's probably a library for our purposes..\n if (value in romanNumbers) {\n const arabicValue = romanNumbers[value];\n nvdebug(` MAP ${value} to ${arabicValue}`, debugDev);\n return arabicValue;\n }\n return value.toLowerCase();\n}\n\nfunction splitAndNormalizePartData(value) {\n // This is just a stub. Does not handle eg. \"Levy 2, raita 15\"\n const [partType, partYear, partNumber] = splitPartData(value);\n //nvdebug(` LHS: '${lhs}'`, debugDev);\n //nvdebug(` RHS: '${rhs}'`, debugDev);\n return [normalizePartType(partType), partYear, normalizePartNumber(partNumber)];\n}\n\nexport function partsAgree(value1, value2, tag, subfieldCode) {\n // Note, that parts can not be normalized away, as \"2\" can agree with \"Part 2\" and \"Raita 2\" and \"Volume 2\"...\n // NB! Used by reducers' mergeSubield.js\n if (!subfieldContainsPartData(tag, subfieldCode)) {\n return false;\n }\n const [partType1, partYear1, partNumber1] = splitAndNormalizePartData(value1);\n const [partType2, partYear2, partNumber2] = splitAndNormalizePartData(value2);\n //nvdebug(`P1: ${partType1} | ${partYear1} | ${partNumber1}`);\n //nvdebug(`P2: ${partType2} | ${partYear2} | ${partNumber2}`);\n if (partNumber1 !== partNumber2) {\n return false;\n }\n if (partType1 !== undefined && partType2 !== undefined && partType1 !== partType2) {\n return false;\n }\n if (partYear1 !== undefined && partYear2 !== undefined && partYear1 !== partYear2) {\n return false;\n }\n\n\n return true;\n}\n\nexport function normalizePartData(value, subfieldCode, tag) {\n // This is for normalizing values for equality comparison only!\n if (!subfieldContainsPartData(tag, subfieldCode)) {\n return value;\n }\n\n const [partType, partYear, partNumber] = splitAndNormalizePartData(value);\n if (partType === undefined) {\n if (partYear === undefined) {\n return partNumber;\n }\n return `${partNumber}/${partYear}`;\n }\n if (partYear === undefined) {\n return `${partType} ${partNumber}`;\n }\n return `${partType} ${partNumber}/${partYear}`;\n}\n"],
4
+ "sourcesContent": ["import {nvdebug} from './utils.js';\nimport createDebugLogger from 'debug';\n\n// Normalizes at least 490$v and 773$g which contain information such as \"Raita 5\" vs \"5\", and \"Osa 3\" vs \"Osa III\".\n\nconst debug = createDebugLogger('@natlibfi/melinda-marc-record-merge-reducers:normalizeSubfieldValueForComparison');\n//const debugData = debug.extend('data');\nconst debugDev = debug.extend('dev');\n\nexport function subfieldContainsPartData(tag, subfieldCode) {\n // NB! Used by reducers' mergeSubield.js\n if (subfieldCode === 'v' && ['490', '800', '810', '811', '830'].includes(tag)) {\n return true;\n }\n if (tag === '773' && subfieldCode === 'g') {\n return true;\n }\n return false;\n}\n\nfunction splitPartData(originalValue) {\n // This a very hacky function, but cand really help it, as the the data is very iffy as well...\n // Remove punctuation and brackets:\n const value = originalValue.replace(/[-.,:; ]+$/ui, '').replace(/^\\[([0-9]+)\\]$/ui, '$1');\n\n const [year, rest] = extractYear(value);\n\n const splitPoint = rest.lastIndexOf(' '); // MRA-627: \"5, 2017\" should be split here. Think of this later on...\n if (splitPoint === -1) {\n return [undefined, year, rest];\n }\n const lhs = rest.substr(0, splitPoint);\n const rhs = rest.substr(splitPoint + 1);\n return [lhs, year, rhs];\n\n function extractYear(value) {\n // NB! Note that this is far for perfect. It cover just some very common cases...\n\n // \"2023, 3\" => [\"2023\", \"3\"]\n if (value.match(/^(?:1[89][0-9][0-9]|20[012][0-9]), (?:nro |n:o)?[1-9][0-9]{0,2}$/ui)) {\n return [value.substr(0, 4), value.substr(6)];\n }\n // \"2023/12\" => [\"2023\", \"12\"]\n if (value.match(/^(?:1[89][0-9][0-9]|20[012][0-9])[/:][1-9][0-9]{0,2}$/u)) {\n return [value.substr(0, 4), value.substr(5)];\n }\n // \"Vol. 3/2023\" => [\"2023\", \"Vol. 3\"]\n if (value.match(/^[^0-9]*[1-9][0-9]{0,2}\\/(?:1[89][0-9][0-9]|20[012][0-9])$/u)) {\n const len = value.length;\n return [value.substr(len - 4), value.substr(0, len - 5)];\n }\n\n\n return [undefined, value];\n }\n}\n\nfunction normalizePartType(originalValue) {\n if (originalValue === undefined) {\n return undefined;\n }\n const value = originalValue.toLowerCase();\n\n // Return Finnish singular nominative. Choise of language is arbitrary. This is best-ish for debug purposes...\n if (['n:o', 'no', 'nr', 'nro', 'number', 'numero', 'nummer'].includes(value)) {\n return 'numero';\n }\n if (['band', 'bd', 'h\u00E4fte', 'nide', 'osa', 'part', 'teil', 'vol', 'vol.', 'volume'].includes(value)) {\n return 'osa';\n }\n\n if (['p.', 'page', 'pages', 'pp.', 's.', 'sidor', 'sivu', 'sivut'].includes(value)) {\n return 'sivu';\n }\n\n return value;\n}\n\nconst romanNumbers = {'I': '1', 'II': '2', 'III': '3', 'IV': '4', 'V': '5', 'VI': '6', 'X': '10'};\n\nfunction normalizePartNumber(value) {\n // Should we handle all Roman numbers or some range of them?\n // There's probably a library for our purposes..\n if (value in romanNumbers) {\n const arabicValue = romanNumbers[value];\n nvdebug(` MAP ${value} to ${arabicValue}`, debugDev);\n return arabicValue;\n }\n return value.toLowerCase();\n}\n\nfunction splitAndNormalizePartData(value) {\n // This is just a stub. Does not handle eg. \"Levy 2, raita 15\"\n const [partType, partYear, partNumber] = splitPartData(value);\n //nvdebug(` LHS: '${lhs}'`, debugDev);\n //nvdebug(` RHS: '${rhs}'`, debugDev);\n return [normalizePartType(partType), partYear, normalizePartNumber(partNumber)];\n}\n\nexport function partsAgree(value1, value2, tag, subfieldCode) {\n // Note, that parts can not be normalized away, as \"2\" can agree with \"Part 2\" and \"Raita 2\" and \"Volume 2\"...\n // NB! Used by reducers' mergeSubield.js\n if (!subfieldContainsPartData(tag, subfieldCode)) {\n return false;\n }\n const [partType1, partYear1, partNumber1] = splitAndNormalizePartData(value1);\n const [partType2, partYear2, partNumber2] = splitAndNormalizePartData(value2);\n //nvdebug(`P1: ${partType1} | ${partYear1} | ${partNumber1}`, debugDev);\n //nvdebug(`P2: ${partType2} | ${partYear2} | ${partNumber2}`, debugDev);\n if (partNumber1 !== partNumber2) {\n return false;\n }\n if (partType1 !== undefined && partType2 !== undefined && partType1 !== partType2) {\n return false;\n }\n if (partYear1 !== undefined && partYear2 !== undefined && partYear1 !== partYear2) {\n return false;\n }\n\n\n return true;\n}\n\nexport function normalizePartData(value, subfieldCode, tag) {\n // This is for normalizing values for equality comparison only!\n if (!subfieldContainsPartData(tag, subfieldCode)) {\n return value;\n }\n\n const [partType, partYear, partNumber] = splitAndNormalizePartData(value);\n if (partType === undefined) {\n if (partYear === undefined) {\n return partNumber;\n }\n return `${partNumber}/${partYear}`;\n }\n if (partYear === undefined) {\n return `${partType} ${partNumber}`;\n }\n return `${partType} ${partNumber}/${partYear}`;\n}\n"],
5
5
  "mappings": "AAAA,SAAQ,eAAc;AACtB,OAAO,uBAAuB;AAI9B,MAAM,QAAQ,kBAAkB,kFAAkF;AAElH,MAAM,WAAW,MAAM,OAAO,KAAK;AAE5B,gBAAS,yBAAyB,KAAK,cAAc;AAE1D,MAAI,iBAAiB,OAAO,CAAC,OAAO,OAAO,OAAO,OAAO,KAAK,EAAE,SAAS,GAAG,GAAG;AAC7E,WAAO;AAAA,EACT;AACA,MAAI,QAAQ,SAAS,iBAAiB,KAAK;AACzC,WAAO;AAAA,EACT;AACA,SAAO;AACT;AAEA,SAAS,cAAc,eAAe;AAGpC,QAAM,QAAQ,cAAc,QAAQ,gBAAgB,EAAE,EAAE,QAAQ,oBAAoB,IAAI;AAExF,QAAM,CAAC,MAAM,IAAI,IAAI,YAAY,KAAK;AAEtC,QAAM,aAAa,KAAK,YAAY,GAAG;AACvC,MAAI,eAAe,IAAI;AACrB,WAAO,CAAC,QAAW,MAAM,IAAI;AAAA,EAC/B;AACA,QAAM,MAAM,KAAK,OAAO,GAAG,UAAU;AACrC,QAAM,MAAM,KAAK,OAAO,aAAa,CAAC;AACtC,SAAO,CAAC,KAAK,MAAM,GAAG;AAEtB,WAAS,YAAYA,QAAO;AAI1B,QAAIA,OAAM,MAAM,oEAAoE,GAAG;AACrF,aAAO,CAACA,OAAM,OAAO,GAAG,CAAC,GAAGA,OAAM,OAAO,CAAC,CAAC;AAAA,IAC7C;AAEA,QAAIA,OAAM,MAAM,wDAAwD,GAAG;AACzE,aAAO,CAACA,OAAM,OAAO,GAAG,CAAC,GAAGA,OAAM,OAAO,CAAC,CAAC;AAAA,IAC7C;AAEA,QAAIA,OAAM,MAAM,6DAA6D,GAAG;AAC9E,YAAM,MAAMA,OAAM;AAClB,aAAO,CAACA,OAAM,OAAO,MAAM,CAAC,GAAGA,OAAM,OAAO,GAAG,MAAM,CAAC,CAAC;AAAA,IACzD;AAGA,WAAO,CAAC,QAAWA,MAAK;AAAA,EAC1B;AACF;AAEA,SAAS,kBAAkB,eAAe;AACxC,MAAI,kBAAkB,QAAW;AAC/B,WAAO;AAAA,EACT;AACA,QAAM,QAAQ,cAAc,YAAY;AAGxC,MAAI,CAAC,OAAO,MAAM,MAAM,OAAO,UAAU,UAAU,QAAQ,EAAE,SAAS,KAAK,GAAG;AAC5E,WAAO;AAAA,EACT;AACA,MAAI,CAAC,QAAQ,MAAM,YAAS,QAAQ,OAAO,QAAQ,QAAQ,OAAO,QAAQ,QAAQ,EAAE,SAAS,KAAK,GAAG;AACnG,WAAO;AAAA,EACT;AAEA,MAAI,CAAC,MAAM,QAAQ,SAAS,OAAO,MAAM,SAAS,QAAQ,OAAO,EAAE,SAAS,KAAK,GAAG;AAClF,WAAO;AAAA,EACT;AAEA,SAAO;AACT;AAEA,MAAM,eAAe,EAAC,KAAK,KAAK,MAAM,KAAK,OAAO,KAAK,MAAM,KAAK,KAAK,KAAK,MAAM,KAAK,KAAK,KAAI;AAEhG,SAAS,oBAAoB,OAAO;AAGlC,MAAI,SAAS,cAAc;AACzB,UAAM,cAAc,aAAa,KAAK;AACtC,YAAQ,QAAQ,KAAK,OAAO,WAAW,IAAI,QAAQ;AACnD,WAAO;AAAA,EACT;AACA,SAAO,MAAM,YAAY;AAC3B;AAEA,SAAS,0BAA0B,OAAO;AAExC,QAAM,CAAC,UAAU,UAAU,UAAU,IAAI,cAAc,KAAK;AAG5D,SAAO,CAAC,kBAAkB,QAAQ,GAAG,UAAU,oBAAoB,UAAU,CAAC;AAChF;AAEO,gBAAS,WAAW,QAAQ,QAAQ,KAAK,cAAc;AAG5D,MAAI,CAAC,yBAAyB,KAAK,YAAY,GAAG;AAChD,WAAO;AAAA,EACT;AACA,QAAM,CAAC,WAAW,WAAW,WAAW,IAAI,0BAA0B,MAAM;AAC5E,QAAM,CAAC,WAAW,WAAW,WAAW,IAAI,0BAA0B,MAAM;AAG5E,MAAI,gBAAgB,aAAa;AAC/B,WAAO;AAAA,EACT;AACA,MAAI,cAAc,UAAa,cAAc,UAAa,cAAc,WAAW;AACjF,WAAO;AAAA,EACT;AACA,MAAI,cAAc,UAAa,cAAc,UAAa,cAAc,WAAW;AACjF,WAAO;AAAA,EACT;AAGA,SAAO;AACT;AAEO,gBAAS,kBAAkB,OAAO,cAAc,KAAK;AAE1D,MAAI,CAAC,yBAAyB,KAAK,YAAY,GAAG;AAChD,WAAO;AAAA,EACT;AAEA,QAAM,CAAC,UAAU,UAAU,UAAU,IAAI,0BAA0B,KAAK;AACxE,MAAI,aAAa,QAAW;AAC1B,QAAI,aAAa,QAAW;AAC1B,aAAO;AAAA,IACT;AACA,WAAO,GAAG,UAAU,IAAI,QAAQ;AAAA,EAClC;AACA,MAAI,aAAa,QAAW;AAC1B,WAAO,GAAG,QAAQ,IAAI,UAAU;AAAA,EAClC;AACA,SAAO,GAAG,QAAQ,IAAI,UAAU,IAAI,QAAQ;AAC9C;",
6
6
  "names": ["value"]
7
7
  }
@@ -1,8 +1,11 @@
1
1
  import { validateSingleField } from "./ending-punctuation.js";
2
2
  import { tagToDataProvenanceSubfieldCode } from "./dataProvenanceUtils.js";
3
3
  import { fieldGetUnambiguousTag } from "./subfield6Utils.js";
4
+ import createDebugLogger from "debug";
4
5
  import { fieldToString, isContentSubfieldCode, nvdebug } from "./utils.js";
5
6
  import clone from "clone";
7
+ const debug = createDebugLogger("@natlibfi/marc-record-validators-melinda:punctuation2");
8
+ const debugDev = debug.extend("dev");
6
9
  const descriptionString = "Remove invalid and add valid punctuation to data fields";
7
10
  export default function() {
8
11
  return {
@@ -11,13 +14,13 @@ export default function() {
11
14
  fix
12
15
  };
13
16
  function fix(record) {
14
- nvdebug(`${descriptionString}: fixer`);
17
+ nvdebug(`${descriptionString}: fixer`, debugDev);
15
18
  const res = { message: [], fix: [], valid: true };
16
19
  record.fields.forEach((f) => fieldFixPunctuation(f));
17
20
  return res;
18
21
  }
19
22
  function validate(record) {
20
- nvdebug(`${descriptionString}: validate`);
23
+ nvdebug(`${descriptionString}: validate`, debugDev);
21
24
  const fieldsNeedingModification = record.fields.filter((f) => fieldNeedsModification(f, true));
22
25
  const values = fieldsNeedingModification.map((f) => fieldToString(f));
23
26
  const newValues = fieldsNeedingModification.map((f) => fieldGetFixedString(f, true));
@@ -1,7 +1,7 @@
1
1
  {
2
2
  "version": 3,
3
3
  "sources": ["../src/punctuation2.js"],
4
- "sourcesContent": ["/*\n* punctuation.js -- try and fix a marc field punctuation\n*\n* Author(s): Nicholas Volk <nicholas.volk@helsinki.fi>\n*\n* NOTE #1: https://www.kiwi.fi/display/kumea/Loppupisteohje is implemented via another validator/fixer (ending-punctuation).\n* This file has some support but it's now yet thorough. (And mmight never be.)\n* NOTE #2: Validator/fixer punctuation does similar stuff, but focuses on X00 fields.\n* NOTE #3: As of 2023-06-05 control subfields ($0...$9) are obsolete. Don't use them in rules.\n* (They are jumped over when looking for next (non-controlfield subfield)\n*/\nimport {validateSingleField} from './ending-punctuation.js';\nimport {tagToDataProvenanceSubfieldCode} from './dataProvenanceUtils.js';\nimport {fieldGetUnambiguousTag} from './subfield6Utils.js';\n//import createDebugLogger from 'debug';\nimport {fieldToString, isContentSubfieldCode, nvdebug} from './utils.js';\nimport clone from 'clone';\n\n//const debug = createDebugLogger('debug/punctuation2');\n\nconst descriptionString = 'Remove invalid and add valid punctuation to data fields';\nexport default function () {\n return {\n description: descriptionString,\n validate, fix\n };\n\n function fix(record) {\n nvdebug(`${descriptionString}: fixer`);\n const res = {message: [], fix: [], valid: true};\n record.fields.forEach(f => fieldFixPunctuation(f));\n return res;\n }\n\n function validate(record) {\n nvdebug(`${descriptionString}: validate`);\n\n const fieldsNeedingModification = record.fields.filter(f => fieldNeedsModification(f, true));\n\n\n const values = fieldsNeedingModification.map(f => fieldToString(f));\n const newValues = fieldsNeedingModification.map(f => fieldGetFixedString(f, true));\n\n const messages = values.map((val, i) => `'${val}' => '${newValues[i]}'`);\n\n const res = {message: messages};\n\n res.valid = res.message.length < 1;\n return res;\n }\n}\n\n\n\nfunction isIrrelevantSubfield(subfield, tag) {\n const dataProvenanceSubfieldCode = tagToDataProvenanceSubfieldCode(tag);\n if (subfield.code === dataProvenanceSubfieldCode) {\n return true;\n }\n return !isContentSubfieldCode(subfield.code); // Currently this contains other stuff as well ($3, $4, $7, $9...)\n}\n\n\nfunction getNextRelevantSubfield(field, currSubfieldIndex) {\n return field.subfields.find((subfield, index) => index > currSubfieldIndex && !isIrrelevantSubfield(subfield, field.tag));\n}\n\nexport function fieldGetFixedString(field, add = true) {\n const cloneField = clone(field);\n const operation = add ? subfieldFixPunctuation : subfieldStripPunctuation;\n cloneField.subfields.forEach((sf, i) => {\n // NB! instead of next subfield, we should actually get next *non-control-subfield*!!!\n // (In plain English: We should skip $0 - $9 at least, maybe $w as well...)\n operation(cloneField, sf, getNextRelevantSubfield(cloneField, i));\n });\n return fieldToString(cloneField);\n}\n\nexport function fieldNeedsModification(field, add = true) {\n if (!field.subfields) {\n return false;\n }\n\n const originalFieldAsString = fieldToString(field);\n const modifiedFieldAsString = fieldGetFixedString(field, add);\n\n return modifiedFieldAsString !== originalFieldAsString;\n}\n\n/////////////////////////////////////////////////////////////////////////////////////\n// <= Above code is written for the validator logic <= //\n// => Everything below was originally transferred from reducers' punctuation.js => //\n/////////////////////////////////////////////////////////////////////////////////////\n\n\n//const stripCrap = / *[-;:,+]+$/u;\nconst needsPuncAfterAlphanumeric = /(?:[a-z0-9A-Z]|\u00E5|\u00E4|\u00F6|\u00C5|\u00C4|\u00D6)$/u;\nconst defaultNeedsPuncAfter2 = /(?:[\\]a-zA-Z0-9)]|\u00E4|\u00E5|\u00F6|\u00C5|\u00C4|\u00D6)$/u;\nconst doesNotEndInPunc = /[^!?.:;,]$/u; // non-punc for pre-240/700/XXX $, note that '.' comes if preceded by ')'\nconst blocksPuncRHS = /^(?:\\()/u;\nconst allowsPuncRHS = /^(?:[A-Za-z0-9]|\u00E5|\u00E4|\u00F6|\u00C5|\u00C4|\u00D6)/u;\nconst aToZ = 'abcdefghijklmnopqrstuvwxyz';\n\n\nconst dotIsProbablyPunc = /(?:[a-z0-9)]|\u00E5|\u00E4|\u00F6|(?:[A-Za-z0-9]|\u00C5|\u00C4|\u00D6)(?:[A-Z]|\u00C5|\u00C4|\u00D6))\\.$/u;\nconst puncIsProbablyPunc = /(?:[a-z0-9)]|\u00E5|\u00E4|\u00F6) ?[.,:;]$/u;\n// NB! 65X: Finnish terms don't use punctuation, but international ones do. Neither one is currently (2021-11-08) coded here.\n\n// Will unfortunately trigger \"Sukunimi, Th.\" type:\nconst removeColons = {'code': 'abcdefghijklmnopqrstuvwxyz', 'remove': / *[;:]$/u};\nconst removeX00Comma = {'code': 'abcdejnqt', 'followedBy': 'abcdenqtv#', 'context': /.,$/u, 'remove': /,$/u};\nconst cleanRHS = {'code': 'abcd', 'followedBy': 'bcde', 'context': /(?:(?:[a-z0-9]|\u00E5|\u00E4|\u00F6)\\.|,)$/u, 'contextRHS': blocksPuncRHS, 'remove': /[.,]$/u};\nconst cleanX00dCommaOrDot = {'code': 'd', 'followedBy': 'et#', 'context': /[0-9]-[,.]$/u, 'remove': /[,.]$/u};\nconst cleanX00aDot = {'code': 'abcde', 'followedBy': 'cdegj', 'context': dotIsProbablyPunc, 'remove': /\\.$/u};\nconst cleanCorruption = {'code': 'abcdefghijklmnopqrstuvwxyz', 'remove': / \\.$/u};\n// These $e dot removals are tricky: before removing the comma, we should know that it ain't an abbreviation such as \"esitt.\"...\nconst cleanX00eDot = {'code': 'e', 'followedBy': 'egj#', 'context': /(?:[ai]ja|j\u00E4)[.,]$/u, 'remove': /\\.$/u};\nconst cleanX11jDot = {'code': 'e', 'followedBy': 'egj#', 'context': /(?:[ai]ja|j\u00E4)[.,]$/u, 'remove': /\\.$/u};\nconst removeCommaBeforeLanguageSubfieldL = {'followedBy': 'l', 'remove': /,$/u};\nconst removeCommaBeforeTitleSubfieldT = {'followedBy': 't', 'remove': /,$/u};\n\nconst X00RemoveDotAfterBracket = {'code': 'cq', 'context': /\\)\\.$/u, 'remove': /\\.$/u};\n// 390, 800, 810, 830...\nconst cleanPuncBeforeLanguage = {'code': 'atvxyz', 'followedBy': 'l', 'context': puncIsProbablyPunc, 'remove': / *[.,:;]$/u};\n\nconst addX00aComma = {'add': ',', 'code': 'abcqejt', 'followedBy': 'cdegnr', 'context': doesNotEndInPunc, 'contextRHS': allowsPuncRHS};\nconst addX00dComma = {'name': 'X00$d ending in \"-\" does not get comma', 'add': ',', 'code': 'd', 'followedBy': 'cdeg', 'context': /[^-,.!]$/u, 'contextRHS': allowsPuncRHS};\nconst addX00aComma2 = {'add': ',', 'code': 'abcdej', 'followedBy': 'cdeg', 'context': /(?:[A-Z]|\u00C5|\u00C4|\u00D6)\\.$/u, 'contextRHS': allowsPuncRHS};\nconst addX00Dot = {'add': '.', 'code': 'abcdetv', 'followedBy': 'fklptu', 'context': needsPuncAfterAlphanumeric};\nconst addEntryFieldFinalDot = {'name': 'X00 final dot', 'add': '.', 'code': 'abcdefghijklmnopqrstuvwxyz', 'followedBy': '#', 'context': /[^.)!?-]$/u};\n\n\nconst addXX0iColon = {name: 'Punctuate relationship information', add: ':', code: 'i', context: defaultNeedsPuncAfter2}; // Not explicitly checking it, but this should always be followed by 'a' or 't'\nconst addX10bDot = {'name': 'Add X10 pre-$b dot', 'add': '.', 'code': 'ab', 'followedBy': 'b', 'context': defaultNeedsPuncAfter2};\nconst addX10Comma = {'add': ',', 'code': 'abet', 'followedBy': 'en', 'context': defaultNeedsPuncAfter2};\nconst addX10Dot = {'name': 'Add X10 final dot', 'add': '.', 'code': 'abet', 'followedBy': 'tu#', 'context': needsPuncAfterAlphanumeric};\nconst addColonToRelationshipInformation = {'name': 'Add \\':\\' to 7X0 $i relationship info', 'add': ':', 'code': 'i', 'context': defaultNeedsPuncAfter2};\n\nconst addX11Spacecolon = {name: '611 space colon(y :-)', add: ' :', code: 'nd', followedBy: 'dc', 'context': defaultNeedsPuncAfter2};\n\nconst addDotBeforeLanguageSubfieldL = {'name': 'Add dot before $l', 'add': '.', 'code': 'abepst', 'followedBy': 'l', 'context': doesNotEndInPunc};\n\n// 490:\nconst addSemicolonBeforeVolumeDesignation = {'name': 'Add \" ;\" before $v', 'add': ' ;', 'code': 'atxyz', 'followedBy': 'v', 'context': /[^;]$/u};\n\nconst NONE = 0;\nconst ADD = 2;\nconst REMOVE = 1;\nconst REMOVE_AND_ADD = 3;\n\n// Crappy punctuation consists of various crap that is somewhat common.\n// We strip crap for merge decisions. We are not trying to actively remove crap here.\n\nconst removeCrapFromAllEntryFields = [removeCommaBeforeLanguageSubfieldL, removeCommaBeforeTitleSubfieldT];\n\nconst removeX00Whatever = [removeX00Comma, cleanX00aDot, cleanX00eDot, cleanCorruption, cleanX00dCommaOrDot, cleanRHS, X00RemoveDotAfterBracket, removeColons, cleanPuncBeforeLanguage, ...removeCrapFromAllEntryFields];\nconst removeX10Whatever = [removeX00Comma, cleanX00aDot, cleanX00eDot, cleanCorruption, removeColons, cleanPuncBeforeLanguage, ...removeCrapFromAllEntryFields];\nconst removeX11Whatever = [removeX00Comma, cleanX11jDot, ...removeCrapFromAllEntryFields];\nconst removeX30Whatever = removeCrapFromAllEntryFields;\n\nconst remove490And830Whatever = [{'code': 'axyzv', 'followedBy': 'axyzv', 'remove': /(?: *;| *=|,)$/u}];\n\nconst linkingEntryRemoveWhatever = [\n {'code': 'i', 'followedBy': 'at', 'remove': / ?:$/u}, // ':'\n {'code': 'at', 'remove': /\\.$/u},\n // Only \". -\" separator is still used in music. We can strip it, but can only create the non-music punctuation!\n {'code': 'abdghiklmnopqrstuwxyz', 'followedBy': 'abdghiklmnopqrstuwxyz#', 'remove': /\\. -$/u}\n];\n\n\n// '!' means negation, thus '!b' means any other subfield but 'b'.\n// 'followedBy': '#' means that current subfield is the last subfield.\n// NB! Note that control subfields are ignored in punctuation rules.\n// NB #2! Control field ignorance causes issues with field 257: https://wiki.helsinki.fi/display/rdasovellusohje/Loppupisteohje\n// Might need to work on that at some point. NOT a top priority though.\n// NB #3! Final punctuation creation is/should be handled by endind-punctuation.js validator!\n\nconst crappy24X = [\n {'code': 'abnp', 'followedBy': '!c', 'remove': / \\/$/u},\n {'code': 'abn', 'followedBy': 'c', 'remove': /\\.$/u, 'context': dotIsProbablyPunc},\n {'code': 'abn', 'followedBy': 'c', 'remove': /\\.$/u, 'context': dotIsProbablyPunc},\n {'code': 'abc', 'followedBy': '#', 'remove': /\\.$/u, 'context': dotIsProbablyPunc},\n {'code': 'abfghinp', 'followedBy': '#', 'remove': /\\.$/u, 'context': dotIsProbablyPunc},\n {'code': 'n', 'followedBy': 'p', 'remove': /\\.$/u, 'context': dotIsProbablyPunc}, // MELINDA-8817\n {'code': 'p', 'followedBy': 'pc', 'remove': /\\.$/u, 'context': dotIsProbablyPunc}, // MELINDA-8817\n removeCommaBeforeLanguageSubfieldL\n];\n\n\nconst cleanCrappyPunctuationRules = {\n '100': removeX00Whatever,\n '110': removeX10Whatever,\n '111': removeX11Whatever,\n '130': removeX30Whatever,\n '240': crappy24X,\n '245': crappy24X,\n '246': crappy24X,\n '300': [\n {'code': 'a', 'followedBy': '!b', 'remove': / *:$/u},\n {'code': 'a', 'followedBy': 'b', 'remove': /:$/u, 'context': /[^ ]:$/u},\n {'code': 'ab', 'followedBy': '!c', 'remove': / *;$/u},\n {'code': 'ab', 'followedBy': 'c', 'remove': /;$/u, 'context': /[^ ];$/u},\n {'code': 'abc', 'followedBy': '!e', 'remove': / *\\+$/u} // Removes both valid (with one space) and invalid (spaceless et al) puncs\n\n ],\n\n '490': remove490And830Whatever,\n '600': removeX00Whatever,\n '610': removeX10Whatever,\n '611': removeX11Whatever,\n '630': removeX30Whatever,\n '700': removeX00Whatever,\n '710': removeX10Whatever,\n '711': removeX11Whatever,\n '730': removeX30Whatever,\n '773': linkingEntryRemoveWhatever,\n '774': linkingEntryRemoveWhatever,\n '776': linkingEntryRemoveWhatever,\n '787': linkingEntryRemoveWhatever,\n '800': removeX00Whatever,\n '810': removeX10Whatever,\n '830': remove490And830Whatever,\n '946': crappy24X\n};\n\nconst cleanLegalX00Comma = {'code': 'abcdetn', 'followedBy': 'cdegjnr', 'context': /.,$/u, 'remove': /,$/u};\n// Accept upper case letters in X00$b, since they are probably Roman numerals.\nconst cleanLegalX00bDot = {'code': 'b', 'followedBy': 't#', context: /^[IVXLCDM]+\\.$/u, 'remove': /\\.$/u};\nconst cleanLegalX00iColon = {'code': 'i', 'followedBy': 'a', 'remove': / *:$/u}; // NB! context is not needed\nconst cleanLegalX00Dot = {'code': 'abcdetkvl', 'followedBy': 'tklu#', 'context': /(?:[a-z0-9)]|\u00E5|\u00E4|\u00F6)\\.$/u, 'remove': /\\.$/u};\nconst cleanDotBeforeLanguageSubfieldL = {'name': 'pre-language-$l dot', 'followedBy': 'l', 'context': /.\\.$/u, 'remove': /\\.$/u};\n\nconst legalEntryField = [cleanDotBeforeLanguageSubfieldL];\n\nconst legalX11SpaceColon = {name: 'legal X11 spacecolony', code: 'nd', followedBy: 'dc', context: / :$/u, remove: / :$/u};\nconst legalX00punc = [cleanLegalX00Comma, cleanLegalX00iColon, cleanLegalX00bDot, cleanLegalX00Dot, ...legalEntryField];\n\nconst cleanLegalX10Comma = {'name': 'X10comma', 'code': 'abe', 'followedBy': 'e', 'context': /.,$/u, 'remove': /,$/u};\nconst cleanLegalX10Dot = {'name': 'X10dot', 'code': 'abt', 'followedBy': 'bst#', 'context': /.\\.$/u, 'remove': /\\.$/u};\n\nconst legalX10punc = [cleanLegalX10Comma, cleanLegalX10Dot, cleanX00eDot, ...legalEntryField];\n\nconst cleanLegalSeriesTitle = [ // 490 and 830\n {'code': 'a', 'followedBy': 'a', 'remove': / =$/u},\n {'code': 'axyz', 'followedBy': 'xyz', 'remove': /,$/u, 'context': /.,$/u},\n {'code': 'axyz', 'followedBy': 'v', 'remove': / *;$/u}\n];\n\nconst clean24X = [\n {'name': 'I:A', 'code': 'i', 'followedBy': 'a', 'remove': / *:$/u},\n {'name': 'A:B', 'code': 'a', 'followedBy': 'b', 'remove': / [:;=]$/u},\n {'name': 'AB:K', 'code': 'ab', 'followedBy': 'k', 'remove': / :$/u},\n {'name': 'ABK:F', 'code': 'abk', 'followedBy': 'f', 'remove': /,$/u},\n {'name': 'ABFNP:C', 'code': 'abfnp', 'followedBy': 'c', 'remove': / \\/$/u},\n {'name': 'ABN:N', 'code': 'abn', 'followedBy': 'n', 'remove': /\\.$/u},\n {'name': 'ABNP:#', 'code': 'abnp', 'followedBy': '#', 'remove': /\\.$/u},\n {'name': 'N:P', 'code': 'n', 'followedBy': 'p', 'remove': /,$/u},\n cleanDotBeforeLanguageSubfieldL\n];\n\nconst legalX11Punc = [...legalEntryField, legalX11SpaceColon];\n\nconst cleanValidPunctuationRules = {\n '100': legalX00punc,\n '110': legalX10punc,\n '111': legalX11Punc,\n '130': legalEntryField,\n '240': clean24X,\n '243': clean24X,\n '245': clean24X,\n '246': clean24X,\n '260': [\n {'code': 'abc', 'followedBy': 'a', 'remove': / ;$/u},\n {'code': 'a', 'followedBy': 'b', 'remove': / :$/u},\n {'code': 'b', 'followedBy': 'c', 'remove': /,$/u},\n {'code': 'c', 'followedBy': '#', 'remove': /\\.$/u},\n {'code': 'd', 'followedBy': 'e', 'remove': / :$/u},\n {'code': 'e', 'followedBy': 'f', 'remove': /,$/u},\n {'code': 'f', 'followedBy': '#', 'remove': /\\.$/u} // Probably ')' but should it be removed?\n ],\n '264': [\n {'code': 'a', 'followedBy': 'b', 'remove': / :$/u},\n {'code': 'b', 'followedBy': 'c', 'remove': /,$/u},\n {'code': 'c', 'followedBy': '#', 'remove': /\\.$/u}\n ],\n '300': [\n // NB! Remove crap as well, thus the '*' in / *:$/\n {'code': 'a', 'followedBy': 'b', 'remove': / :$/u},\n {'code': 'ab', 'followedBy': 'c', 'remove': / ;$/u},\n {'code': 'abc', 'followedBy': 'e', 'remove': / \\+$/u}\n ],\n '490': cleanLegalSeriesTitle,\n '534': [{'code': 'p', 'followedBy': 'c', 'remove': /:$/u}],\n '600': legalX00punc,\n '610': legalX10punc,\n '611': legalX11Punc,\n '630': legalEntryField,\n // Experimental, MET366-ish (end punc in internationally valid, but we don't use it here in Finland):\n '648': [{'code': 'a', 'content': /^[0-9]+\\.$/u, 'ind2': ['4'], 'remove': /\\.$/u}],\n '700': legalX00punc,\n '710': legalX10punc,\n '711': legalX11Punc,\n '730': legalEntryField,\n '800': legalX00punc,\n '810': legalX10punc,\n '811': legalX11Punc,\n '830': [...legalEntryField, ...cleanLegalSeriesTitle],\n '946': clean24X\n};\n\n\n// Overgeneralizes a bit: eg. addColonToRelationshipInformation only applies to 700/710 but as others don't have $i, it's fine.\nconst addToAllEntryFields = [addDotBeforeLanguageSubfieldL, addSemicolonBeforeVolumeDesignation, addColonToRelationshipInformation, addEntryFieldFinalDot];\n\n\nconst addX00 = [addXX0iColon, addX00aComma, addX00aComma2, addX00Dot, addX00dComma, ...addToAllEntryFields];\nconst addX10 = [addXX0iColon, addX10bDot, addX10Comma, addX10Dot, ...addToAllEntryFields];\nconst addX11 = [...addToAllEntryFields, addX11Spacecolon];\nconst addX30 = [...addToAllEntryFields];\n\nconst add24X = [\n {'code': 'i', 'followedBy': 'a', 'add': ':', 'context': needsPuncAfterAlphanumeric},\n {'code': 'a', 'followedBy': 'b', 'add': ' :', 'context': '[^:]$'},\n {'code': 'abk', 'followedBy': 'f', 'add': ',', 'context': needsPuncAfterAlphanumeric},\n {'code': 'abfnp', 'followedBy': 'c', 'add': ' /', 'context': '[^/]$'},\n addDotBeforeLanguageSubfieldL\n];\n\nconst add245 = [\n ...add24X,\n // Blah! Also \"$a = $b\" and \"$a ; $b\" can be valid... But ' :' is better than nothing, I guess...\n {'code': 'ab', 'followedBy': 'n', 'add': '.', 'context': needsPuncAfterAlphanumeric},\n {'code': 'n', 'followedBy': 'p', 'add': ',', 'context': defaultNeedsPuncAfter2},\n {'code': 'abnpc', 'followedBy': '#', 'add': '.', 'context': needsPuncAfterAlphanumeric} // Stepping on \"punctuation validator's\" toes\n];\n\nconst addSeriesTitle = [ // 490 and 830\n {'code': 'a', 'followedBy': 'a', 'add': ' =', 'context': defaultNeedsPuncAfter2},\n {'code': 'axyz', 'followedBy': 'xy', 'add': ',', 'context': defaultNeedsPuncAfter2},\n addSemicolonBeforeVolumeDesignation // eg. 490$axyz-$v\n];\n\nconst addLinkingEntry = [ // NB! Music 773 uses different punctuation rules, that are not implement here (can they even be?)\n {'code': 'i', 'followedBy': aToZ, 'add': ':', 'context': defaultNeedsPuncAfter2},\n {'code': 'a', 'followedBy': 't', 'add': '.', 'context': defaultNeedsPuncAfter2},\n {'code': 't', 'followedBy': 'dghoz', 'add': '.', 'context': defaultNeedsPuncAfter2}\n];\n\nconst addPairedPunctuationRules = {\n '100': addX00,\n '110': addX10,\n '111': addX11,\n '130': addX30,\n '240': add24X,\n '243': add24X,\n '245': add245,\n '246': add24X,\n '260': [\n {'code': 'a', 'followedBy': 'b', 'add': ' :', 'context': defaultNeedsPuncAfter2},\n {'code': 'ab', 'followedBy': 'c', 'add': ',', 'context': defaultNeedsPuncAfter2},\n {'code': 'abc', 'followedBy': 'a', 'add': ' ;', 'context': defaultNeedsPuncAfter2},\n {'code': 'e', 'followedBy': 'f', 'add': ' :', 'context': defaultNeedsPuncAfter2},\n {'code': 'f', 'followedBy': 'g', 'add': ',', 'context': defaultNeedsPuncAfter2}\n ],\n '264': [\n {'code': 'a', 'followedBy': 'b', 'add': ' :', 'context': defaultNeedsPuncAfter2},\n {'code': 'b', 'followedBy': 'c', 'add': ',', 'context': defaultNeedsPuncAfter2},\n // NB! The $c rule messes dotless exception \"264 #4 $c p1983\" up\n // We'll need to add a hacky postprocessor for this? Add 'hasInd1': '0123' etc?\n {'code': 'c', 'followedBy': '#', 'add': '.', 'context': needsPuncAfterAlphanumeric, 'ind2': ['0', '1', '2', '3']}\n ],\n '300': [\n {'code': 'a', 'followedBy': 'b', 'add': ' :', 'context': defaultNeedsPuncAfter2},\n {'code': 'ab', 'followedBy': 'c', 'add': ' ;', 'context': defaultNeedsPuncAfter2},\n {'code': 'abc', 'followedBy': 'e', 'add': ' +', 'context': defaultNeedsPuncAfter2}\n ],\n '490': addSeriesTitle,\n '506': [{'code': 'a', 'followedBy': '#', 'add': '.', 'context': defaultNeedsPuncAfter2}],\n '534': [{'code': 'p', 'followedBy': 'c', 'add': ':', 'context': defaultNeedsPuncAfter2}],\n '600': addX00,\n '610': addX10,\n '611': addX11,\n '630': addX30,\n '700': addX00,\n '710': addX10,\n '711': addX11,\n '730': addX30,\n '773': addLinkingEntry,\n '787': addLinkingEntry,\n '800': addX00,\n '810': addX10,\n '811': addX11,\n '830': [...addX30, ...addSeriesTitle],\n '946': [{'code': 'i', 'followedBy': 'a', 'add': ':', 'context': defaultNeedsPuncAfter2}]\n};\n\n/*\nfunction debugRule(rule) {\n //nvdebug('');\n nvdebug(`NAME ${rule.name ? rule.name : '<unnamed>'}`);\n nvdebug(`SUBFIELD CODE '${rule.code}' FOLLOWED BY SUBFIELD CODE '${rule.followedBy}'`);\n if ('add' in rule) {\n nvdebug(`ADD '${rule.add}'`);\n }\n if ('remove' in rule) {\n nvdebug(`REMOVE '${rule.remove}'`);\n }\n if ('context' in rule) {\n nvdebug(`CONTEXT '${rule.context.toString()}'`);\n }\n //nvdebug('');\n}\n*/\n\nfunction ruleAppliesToSubfieldCode(targetSubfieldCodes, currSubfieldCode) {\n if (!targetSubfieldCodes) { // We are not interested in what subfield precedes 240$l, ',' is removed anyway\n return true;\n }\n const negation = targetSubfieldCodes.includes('!');\n if (negation) {\n return !targetSubfieldCodes.includes(currSubfieldCode);\n }\n return targetSubfieldCodes.includes(currSubfieldCode);\n}\n\n\nfunction ruleAppliesToField(rule, field) {\n if ('ind1' in rule && !rule.ind1.includes(field.ind1)) {\n return false;\n }\n\n if ('ind2' in rule && !rule.ind2.includes(field.ind2)) {\n return false;\n }\n\n // If we want to check, say, $2, it should be implemented here!\n\n return true;\n}\n\n\nfunction ruleAppliesToCurrentSubfield(rule, subfield) {\n //nvdebug(` Apply rule on LHS?`);\n if (!ruleAppliesToSubfieldCode(rule.code, subfield.code)) {\n //nvdebug(` Reject rule!`);\n return false;\n }\n if ('context' in rule) {\n //nvdebug(` Check '${subfield.value}' versus '${rule.context.toString()}'`);\n if (!subfield.value.match(rule.context)) { // njsscan-ignore: regex_injection_dos\n //nvdebug(` Reject rule!`);\n return false;\n }\n }\n //nvdebug(` Apply rule!`);\n return true;\n}\n\nfunction ruleAppliesToNextSubfield(rule, nextSubfield) {\n if (!('followedBy' in rule)) { // Return true, if we are not interested in the next subfield\n return true;\n }\n // The '#' existence check applies only to the RHS field. LHS always exists.\n if (!nextSubfield) {\n const negation = rule.followedBy.includes('!');\n if (negation) {\n return !rule.followedBy.includes('#');\n }\n return rule.followedBy.includes('#');\n }\n\n if (!ruleAppliesToSubfieldCode(rule.followedBy, nextSubfield.code)) {\n return false;\n }\n if ('contextRHS' in rule && !nextSubfield.value.match(rule.contextRHS)) { // njsscan-ignore: regex_injection_dos\n return false;\n }\n return true;\n}\n\nfunction checkRule(rule, field, subfield1, subfield2) {\n if (!ruleAppliesToField(rule, field)) {\n //nvdebug(`FAIL ON WHOLE FIELD: '${fieldToString(field)}`);\n return false;\n }\n //const name = rule.name || 'UNNAMED';\n if (!ruleAppliesToCurrentSubfield(rule, subfield1)) {\n //nvdebug(`${name}: FAIL ON LHS SUBFIELD: '$${subfield1.code} ${subfield1.value}', SF=${rule.code}`, debug);\n return false;\n }\n\n // NB! This is not a perfect solution. We might have $e$0$e where $e$0 punctuation should actually be based on $e$e rules\n if (!ruleAppliesToNextSubfield(rule, subfield2)) {\n //const msg = subfield2 ? `${name}: FAIL ON RHS SUBFIELD '${subfield2.code}' not in [${rule.followedBy}]` : `${name}: FAIL ON RHS FIELD`;\n //nvdebug(msg, debug);\n return false;\n }\n\n //nvdebug(`${rule.name ? rule.name : '<unnamed>'}: ACCEPT ${rule.code} (${subfield1.code}), SF2=${rule.followedBy} (${subfield2 ? subfield2.code : '#'})`, debug);\n return true;\n}\n\n\nfunction applyPunctuationRules(field, subfield1, subfield2, ruleArray = null, operation = NONE) {\n if (operation === NONE || ruleArray === null) { // !fieldIsApplicable(field, ruleArray)) {\n return;\n }\n const tag2 = field.tag === '880' ? fieldGetUnambiguousTag(field) : field.tag;\n if (!tag2) {\n return;\n }\n if (!(`${tag2}` in ruleArray)) {\n return;\n }\n\n //nvdebug(`PUNCTUATE ${field.tag}/${tag2} '${subfieldToString(subfield1)}' XXX '${subfield2 ? subfieldToString(subfield2) : '#'} }`);\n\n //nvdebug(`OP=${operation} ${tag2}: '${subfield1.code}: ${subfield1.value}' ??? '${subfield2 ? subfield2.code : '#'}'`);\n const candRules = ruleArray[tag2];\n candRules.every(rule => { // uses \"every\", not \"forEach\", so that only one rule is applies to the given subfields\n //debugRule(rule);\n if (!checkRule(rule, field, subfield1, subfield2)) {\n return true;\n }\n\n //const originalValue = subfield1.value;\n if (rule.remove && [REMOVE, REMOVE_AND_ADD].includes(operation) && subfield1.value.match(rule.remove)) {\n //nvdebug(` PUNC REMOVAL TO BE PERFORMED FOR $${subfield1.code} '${subfield1.value}'`, debug);\n subfield1.value = subfield1.value.replace(rule.remove, '');\n //nvdebug(` PUNC REMOVAL PERFORMED FOR '${subfield1.value}'`);\n return false;\n }\n if (rule.add && [ADD, REMOVE_AND_ADD].includes(operation)) {\n subfield1.value += rule.add;\n //nvdebug(` ADDED '${rule.add}' TO FORM '${subfield1.value}' USING RULE ${rule.name}`);\n return false;\n }\n\n /*\n if (subfield1.value !== originalValue) {\n nvdebug(` PROCESS PUNC: '\u2021${subfield1.code} ${originalValue}' => '\u2021${subfield1.code} ${subfield1.value}'`, debug);\n }\n */\n\n return true;\n });\n}\n\nfunction subfieldFixPunctuation(field, subfield1, subfield2) {\n applyPunctuationRules(field, subfield1, subfield2, cleanCrappyPunctuationRules, REMOVE);\n applyPunctuationRules(field, subfield1, subfield2, addPairedPunctuationRules, ADD);\n}\n\nfunction subfieldStripPunctuation(field, subfield1, subfield2) {\n //nvdebug(`FSP1: '${subfield1.value}'`);\n applyPunctuationRules(field, subfield1, subfield2, cleanValidPunctuationRules, REMOVE);\n //nvdebug(`FSP2: '${subfield1.value}'`);\n applyPunctuationRules(field, subfield1, subfield2, cleanCrappyPunctuationRules, REMOVE);\n //nvdebug(`FSP3: '${subfield1.value}'`);\n\n}\n\nexport function fieldStripPunctuation(field) {\n if (!field.subfields) {\n return field;\n }\n\n field.subfields.forEach((sf, i) => {\n // NB! instead of next subfield, we should actually get next *non-control-subfield*!!!\n // (In plain English: We should skip $0 - $9 at least, maybe $w as well...)\n subfieldStripPunctuation(field, sf, getNextRelevantSubfield(field, i));\n\n });\n return field;\n}\n\nexport function fieldFixPunctuation(field) {\n if (!field.subfields) {\n return field;\n }\n //nvdebug(`################### fieldFixPunctuation() TEST ${fieldToString(field)}`);\n\n field.subfields.forEach((sf, i) => {\n // NB! instead of next subfield, we should actually get next *non-control-subfield*!!!\n // (In plain English: We should skip $0 - $9 at least, maybe $w as well...)\n // We'll need some magic for field 257 here, do we? (Also Finnish lexicons vs global lexicons in 65X fields)\n subfieldFixPunctuation(field, sf, getNextRelevantSubfield(field, i));\n });\n\n // Use shared code for final punctuation (sadly this does not fix intermediate punc):\n if (field.useExternalEndPunctuation) {\n // addFinalPunctuation(field); // local version. use shared code instead.\n validateSingleField(field, false, true); // NB! Don't use field.tag as second argument! It's a string, not an int. 3rd arg must be true (=fix)\n }\n return field;\n}\n"],
5
- "mappings": "AAWA,SAAQ,2BAA0B;AAClC,SAAQ,uCAAsC;AAC9C,SAAQ,8BAA6B;AAErC,SAAQ,eAAe,uBAAuB,eAAc;AAC5D,OAAO,WAAW;AAIlB,MAAM,oBAAoB;AAC1B,0BAA2B;AACzB,SAAO;AAAA,IACL,aAAa;AAAA,IACb;AAAA,IAAU;AAAA,EACZ;AAEA,WAAS,IAAI,QAAQ;AACnB,YAAQ,GAAG,iBAAiB,SAAS;AACrC,UAAM,MAAM,EAAC,SAAS,CAAC,GAAG,KAAK,CAAC,GAAG,OAAO,KAAI;AAC9C,WAAO,OAAO,QAAQ,OAAK,oBAAoB,CAAC,CAAC;AACjD,WAAO;AAAA,EACT;AAEA,WAAS,SAAS,QAAQ;AACxB,YAAQ,GAAG,iBAAiB,YAAY;AAExC,UAAM,4BAA4B,OAAO,OAAO,OAAO,OAAK,uBAAuB,GAAG,IAAI,CAAC;AAG3F,UAAM,SAAS,0BAA0B,IAAI,OAAK,cAAc,CAAC,CAAC;AAClE,UAAM,YAAY,0BAA0B,IAAI,OAAK,oBAAoB,GAAG,IAAI,CAAC;AAEjF,UAAM,WAAW,OAAO,IAAI,CAAC,KAAK,MAAM,IAAI,GAAG,SAAS,UAAU,CAAC,CAAC,GAAG;AAEvE,UAAM,MAAM,EAAC,SAAS,SAAQ;AAE9B,QAAI,QAAQ,IAAI,QAAQ,SAAS;AACjC,WAAO;AAAA,EACT;AACF;AAIA,SAAS,qBAAqB,UAAU,KAAK;AAC3C,QAAM,6BAA6B,gCAAgC,GAAG;AACtE,MAAI,SAAS,SAAS,4BAA4B;AAChD,WAAO;AAAA,EACT;AACA,SAAO,CAAC,sBAAsB,SAAS,IAAI;AAC7C;AAGA,SAAS,wBAAwB,OAAO,mBAAmB;AACzD,SAAO,MAAM,UAAU,KAAK,CAAC,UAAU,UAAU,QAAQ,qBAAqB,CAAC,qBAAqB,UAAU,MAAM,GAAG,CAAC;AAC1H;AAEO,gBAAS,oBAAoB,OAAO,MAAM,MAAM;AACrD,QAAM,aAAa,MAAM,KAAK;AAC9B,QAAM,YAAY,MAAM,yBAAyB;AACjD,aAAW,UAAU,QAAQ,CAAC,IAAI,MAAM;AAGtC,cAAU,YAAY,IAAI,wBAAwB,YAAY,CAAC,CAAC;AAAA,EAClE,CAAC;AACD,SAAO,cAAc,UAAU;AACjC;AAEO,gBAAS,uBAAuB,OAAO,MAAM,MAAM;AACxD,MAAI,CAAC,MAAM,WAAW;AACpB,WAAO;AAAA,EACT;AAEA,QAAM,wBAAwB,cAAc,KAAK;AACjD,QAAM,wBAAwB,oBAAoB,OAAO,GAAG;AAE5D,SAAO,0BAA0B;AACnC;AASA,MAAM,6BAA6B;AACnC,MAAM,yBAAyB;AAC/B,MAAM,mBAAmB;AACzB,MAAM,gBAAgB;AACtB,MAAM,gBAAgB;AACtB,MAAM,OAAO;AAGb,MAAM,oBAAoB;AAC1B,MAAM,qBAAqB;AAI3B,MAAM,eAAe,EAAC,QAAQ,8BAA8B,UAAU,WAAU;AAChF,MAAM,iBAAiB,EAAC,QAAQ,aAAa,cAAc,cAAc,WAAW,QAAQ,UAAU,MAAK;AAC3G,MAAM,WAAW,EAAC,QAAQ,QAAQ,cAAc,QAAQ,WAAW,gCAAgC,cAAc,eAAe,UAAU,SAAQ;AAClJ,MAAM,sBAAsB,EAAC,QAAQ,KAAK,cAAc,OAAO,WAAW,gBAAgB,UAAU,SAAQ;AAC5G,MAAM,eAAe,EAAC,QAAQ,SAAS,cAAc,SAAS,WAAW,mBAAmB,UAAU,OAAM;AAC5G,MAAM,kBAAkB,EAAC,QAAQ,8BAA8B,UAAU,QAAO;AAEhF,MAAM,eAAe,EAAC,QAAQ,KAAK,cAAc,QAAQ,WAAW,uBAAuB,UAAU,OAAM;AAC3G,MAAM,eAAe,EAAC,QAAQ,KAAK,cAAc,QAAQ,WAAW,uBAAuB,UAAU,OAAM;AAC3G,MAAM,qCAAqC,EAAC,cAAc,KAAK,UAAU,MAAK;AAC9E,MAAM,kCAAkC,EAAC,cAAc,KAAK,UAAU,MAAK;AAE3E,MAAM,2BAA2B,EAAC,QAAQ,MAAM,WAAW,UAAU,UAAU,OAAM;AAErF,MAAM,0BAA0B,EAAC,QAAQ,UAAU,cAAc,KAAK,WAAW,oBAAoB,UAAU,aAAY;AAE3H,MAAM,eAAe,EAAC,OAAO,KAAK,QAAQ,WAAW,cAAc,UAAU,WAAW,kBAAkB,cAAc,cAAa;AACrI,MAAM,eAAe,EAAC,QAAQ,0CAA0C,OAAO,KAAK,QAAQ,KAAK,cAAc,QAAQ,WAAW,aAAa,cAAc,cAAa;AAC1K,MAAM,gBAAgB,EAAC,OAAO,KAAK,QAAQ,UAAU,cAAc,QAAQ,WAAW,uBAAuB,cAAc,cAAa;AACxI,MAAM,YAAY,EAAC,OAAO,KAAK,QAAQ,WAAW,cAAc,UAAU,WAAW,2BAA0B;AAC/G,MAAM,wBAAwB,EAAC,QAAQ,iBAAiB,OAAO,KAAK,QAAQ,8BAA8B,cAAc,KAAK,WAAW,aAAY;AAGpJ,MAAM,eAAe,EAAC,MAAM,sCAAsC,KAAK,KAAK,MAAM,KAAK,SAAS,uBAAsB;AACtH,MAAM,aAAa,EAAC,QAAQ,sBAAsB,OAAO,KAAK,QAAQ,MAAM,cAAc,KAAK,WAAW,uBAAsB;AAChI,MAAM,cAAc,EAAC,OAAO,KAAK,QAAQ,QAAQ,cAAc,MAAM,WAAW,uBAAsB;AACtG,MAAM,YAAY,EAAC,QAAQ,qBAAqB,OAAO,KAAK,QAAQ,QAAQ,cAAc,OAAO,WAAW,2BAA0B;AACtI,MAAM,oCAAoC,EAAC,QAAQ,uCAAyC,OAAO,KAAK,QAAQ,KAAK,WAAW,uBAAsB;AAEtJ,MAAM,mBAAmB,EAAC,MAAM,yBAAyB,KAAK,MAAM,MAAM,MAAM,YAAY,MAAM,WAAW,uBAAsB;AAEnI,MAAM,gCAAgC,EAAC,QAAQ,qBAAqB,OAAO,KAAK,QAAQ,UAAU,cAAc,KAAK,WAAW,iBAAgB;AAGhJ,MAAM,sCAAsC,EAAC,QAAQ,sBAAsB,OAAO,MAAM,QAAQ,SAAS,cAAc,KAAK,WAAW,SAAQ;AAE/I,MAAM,OAAO;AACb,MAAM,MAAM;AACZ,MAAM,SAAS;AACf,MAAM,iBAAiB;AAKvB,MAAM,+BAA+B,CAAC,oCAAoC,+BAA+B;AAEzG,MAAM,oBAAoB,CAAC,gBAAgB,cAAc,cAAc,iBAAiB,qBAAqB,UAAU,0BAA0B,cAAc,yBAAyB,GAAG,4BAA4B;AACvN,MAAM,oBAAoB,CAAC,gBAAgB,cAAc,cAAc,iBAAiB,cAAc,yBAAyB,GAAG,4BAA4B;AAC9J,MAAM,oBAAoB,CAAC,gBAAgB,cAAc,GAAG,4BAA4B;AACxF,MAAM,oBAAoB;AAE1B,MAAM,0BAA0B,CAAC,EAAC,QAAQ,SAAS,cAAc,SAAS,UAAU,kBAAiB,CAAC;AAEtG,MAAM,6BAA6B;AAAA,EACjC,EAAC,QAAQ,KAAK,cAAc,MAAM,UAAU,QAAO;AAAA;AAAA,EACnD,EAAC,QAAQ,MAAM,UAAU,OAAM;AAAA;AAAA,EAE/B,EAAC,QAAQ,yBAAyB,cAAc,0BAA0B,UAAU,SAAQ;AAC9F;AAUA,MAAM,YAAY;AAAA,EAChB,EAAC,QAAQ,QAAQ,cAAc,MAAM,UAAU,QAAO;AAAA,EACtD,EAAC,QAAQ,OAAO,cAAc,KAAK,UAAU,QAAQ,WAAW,kBAAiB;AAAA,EACjF,EAAC,QAAQ,OAAO,cAAc,KAAK,UAAU,QAAQ,WAAW,kBAAiB;AAAA,EACjF,EAAC,QAAQ,OAAO,cAAc,KAAK,UAAU,QAAQ,WAAW,kBAAiB;AAAA,EACjF,EAAC,QAAQ,YAAY,cAAc,KAAK,UAAU,QAAQ,WAAW,kBAAiB;AAAA,EACtF,EAAC,QAAQ,KAAK,cAAc,KAAK,UAAU,QAAQ,WAAW,kBAAiB;AAAA;AAAA,EAC/E,EAAC,QAAQ,KAAK,cAAc,MAAM,UAAU,QAAQ,WAAW,kBAAiB;AAAA;AAAA,EAChF;AACF;AAGA,MAAM,8BAA8B;AAAA,EAClC,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,IACL,EAAC,QAAQ,KAAK,cAAc,MAAM,UAAU,QAAO;AAAA,IACnD,EAAC,QAAQ,KAAK,cAAc,KAAK,UAAU,OAAO,WAAW,UAAS;AAAA,IACtE,EAAC,QAAQ,MAAM,cAAc,MAAM,UAAU,QAAO;AAAA,IACpD,EAAC,QAAQ,MAAM,cAAc,KAAK,UAAU,OAAO,WAAW,UAAS;AAAA,IACvE,EAAC,QAAQ,OAAO,cAAc,MAAM,UAAU,SAAQ;AAAA;AAAA,EAExD;AAAA,EAEA,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AACT;AAEA,MAAM,qBAAqB,EAAC,QAAQ,WAAW,cAAc,WAAW,WAAW,QAAQ,UAAU,MAAK;AAE1G,MAAM,oBAAoB,EAAC,QAAQ,KAAK,cAAc,MAAM,SAAS,mBAAmB,UAAU,OAAM;AACxG,MAAM,sBAAsB,EAAC,QAAQ,KAAK,cAAc,KAAK,UAAU,QAAO;AAC9E,MAAM,mBAAmB,EAAC,QAAQ,aAAa,cAAc,SAAS,WAAW,2BAA2B,UAAU,OAAM;AAC5H,MAAM,kCAAkC,EAAC,QAAQ,uBAAuB,cAAc,KAAK,WAAW,SAAS,UAAU,OAAM;AAE/H,MAAM,kBAAkB,CAAC,+BAA+B;AAExD,MAAM,qBAAqB,EAAC,MAAM,yBAAyB,MAAM,MAAM,YAAY,MAAM,SAAS,QAAQ,QAAQ,OAAM;AACxH,MAAM,eAAe,CAAC,oBAAoB,qBAAqB,mBAAmB,kBAAkB,GAAG,eAAe;AAEtH,MAAM,qBAAqB,EAAC,QAAQ,YAAY,QAAQ,OAAO,cAAc,KAAK,WAAW,QAAQ,UAAU,MAAK;AACpH,MAAM,mBAAmB,EAAC,QAAQ,UAAU,QAAQ,OAAO,cAAc,QAAQ,WAAW,SAAS,UAAU,OAAM;AAErH,MAAM,eAAe,CAAC,oBAAoB,kBAAkB,cAAc,GAAG,eAAe;AAE5F,MAAM,wBAAwB;AAAA;AAAA,EAC5B,EAAC,QAAQ,KAAK,cAAc,KAAK,UAAU,OAAM;AAAA,EACjD,EAAC,QAAQ,QAAQ,cAAc,OAAO,UAAU,OAAO,WAAW,OAAM;AAAA,EACxE,EAAC,QAAQ,QAAQ,cAAc,KAAK,UAAU,QAAO;AACvD;AAEA,MAAM,WAAW;AAAA,EACf,EAAC,QAAQ,OAAO,QAAQ,KAAK,cAAc,KAAK,UAAU,QAAO;AAAA,EACjE,EAAC,QAAQ,OAAO,QAAQ,KAAK,cAAc,KAAK,UAAU,WAAU;AAAA,EACpE,EAAC,QAAQ,QAAQ,QAAQ,MAAM,cAAc,KAAK,UAAU,OAAM;AAAA,EAClE,EAAC,QAAQ,SAAS,QAAQ,OAAO,cAAc,KAAK,UAAU,MAAK;AAAA,EACnE,EAAC,QAAQ,WAAW,QAAQ,SAAS,cAAc,KAAK,UAAU,QAAO;AAAA,EACzE,EAAC,QAAQ,SAAS,QAAQ,OAAO,cAAc,KAAK,UAAU,OAAM;AAAA,EACpE,EAAC,QAAQ,UAAU,QAAQ,QAAQ,cAAc,KAAK,UAAU,OAAM;AAAA,EACtE,EAAC,QAAQ,OAAO,QAAQ,KAAK,cAAc,KAAK,UAAU,MAAK;AAAA,EAC/D;AACF;AAEA,MAAM,eAAe,CAAC,GAAG,iBAAiB,kBAAkB;AAE5D,MAAM,6BAA6B;AAAA,EACjC,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,IACL,EAAC,QAAQ,OAAO,cAAc,KAAK,UAAU,OAAM;AAAA,IACnD,EAAC,QAAQ,KAAK,cAAc,KAAK,UAAU,OAAM;AAAA,IACjD,EAAC,QAAQ,KAAK,cAAc,KAAK,UAAU,MAAK;AAAA,IAChD,EAAC,QAAQ,KAAK,cAAc,KAAK,UAAU,OAAM;AAAA,IACjD,EAAC,QAAQ,KAAK,cAAc,KAAK,UAAU,OAAM;AAAA,IACjD,EAAC,QAAQ,KAAK,cAAc,KAAK,UAAU,MAAK;AAAA,IAChD,EAAC,QAAQ,KAAK,cAAc,KAAK,UAAU,OAAM;AAAA;AAAA,EACnD;AAAA,EACA,OAAO;AAAA,IACL,EAAC,QAAQ,KAAK,cAAc,KAAK,UAAU,OAAM;AAAA,IACjD,EAAC,QAAQ,KAAK,cAAc,KAAK,UAAU,MAAK;AAAA,IAChD,EAAC,QAAQ,KAAK,cAAc,KAAK,UAAU,OAAM;AAAA,EACnD;AAAA,EACA,OAAO;AAAA;AAAA,IAEL,EAAC,QAAQ,KAAK,cAAc,KAAK,UAAU,OAAM;AAAA,IACjD,EAAC,QAAQ,MAAM,cAAc,KAAK,UAAU,OAAM;AAAA,IAClD,EAAC,QAAQ,OAAO,cAAc,KAAK,UAAU,QAAO;AAAA,EACtD;AAAA,EACA,OAAO;AAAA,EACP,OAAO,CAAC,EAAC,QAAQ,KAAK,cAAc,KAAK,UAAU,MAAK,CAAC;AAAA,EACzD,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA;AAAA,EAEP,OAAO,CAAC,EAAC,QAAQ,KAAK,WAAW,eAAe,QAAQ,CAAC,GAAG,GAAG,UAAU,OAAM,CAAC;AAAA,EAChF,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO,CAAC,GAAG,iBAAiB,GAAG,qBAAqB;AAAA,EACpD,OAAO;AACT;AAIA,MAAM,sBAAsB,CAAC,+BAA+B,qCAAqC,mCAAmC,qBAAqB;AAGzJ,MAAM,SAAS,CAAC,cAAc,cAAc,eAAe,WAAW,cAAc,GAAG,mBAAmB;AAC1G,MAAM,SAAS,CAAC,cAAc,YAAY,aAAa,WAAW,GAAG,mBAAmB;AACxF,MAAM,SAAS,CAAC,GAAG,qBAAqB,gBAAgB;AACxD,MAAM,SAAS,CAAC,GAAG,mBAAmB;AAEtC,MAAM,SAAS;AAAA,EACb,EAAC,QAAQ,KAAK,cAAc,KAAK,OAAO,KAAK,WAAW,2BAA0B;AAAA,EAClF,EAAC,QAAQ,KAAK,cAAc,KAAK,OAAO,MAAM,WAAW,QAAO;AAAA,EAChE,EAAC,QAAQ,OAAO,cAAc,KAAK,OAAO,KAAK,WAAW,2BAA0B;AAAA,EACpF,EAAC,QAAQ,SAAS,cAAc,KAAK,OAAO,MAAM,WAAW,QAAO;AAAA,EACpE;AACF;AAEA,MAAM,SAAS;AAAA,EACb,GAAG;AAAA;AAAA,EAEH,EAAC,QAAQ,MAAM,cAAc,KAAK,OAAO,KAAK,WAAW,2BAA0B;AAAA,EACnF,EAAC,QAAQ,KAAK,cAAc,KAAK,OAAO,KAAK,WAAW,uBAAsB;AAAA,EAC9E,EAAC,QAAQ,SAAS,cAAc,KAAK,OAAO,KAAK,WAAW,2BAA0B;AAAA;AACxF;AAEA,MAAM,iBAAiB;AAAA;AAAA,EACrB,EAAC,QAAQ,KAAK,cAAc,KAAK,OAAO,MAAM,WAAW,uBAAsB;AAAA,EAC/E,EAAC,QAAQ,QAAQ,cAAc,MAAM,OAAO,KAAK,WAAW,uBAAsB;AAAA,EAClF;AAAA;AACF;AAEA,MAAM,kBAAkB;AAAA;AAAA,EACtB,EAAC,QAAQ,KAAK,cAAc,MAAM,OAAO,KAAK,WAAW,uBAAsB;AAAA,EAC/E,EAAC,QAAQ,KAAK,cAAc,KAAK,OAAO,KAAK,WAAW,uBAAsB;AAAA,EAC9E,EAAC,QAAQ,KAAK,cAAc,SAAS,OAAO,KAAK,WAAW,uBAAsB;AACpF;AAEA,MAAM,4BAA4B;AAAA,EAChC,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,IACL,EAAC,QAAQ,KAAK,cAAc,KAAK,OAAO,MAAM,WAAW,uBAAsB;AAAA,IAC/E,EAAC,QAAQ,MAAM,cAAc,KAAK,OAAO,KAAK,WAAW,uBAAsB;AAAA,IAC/E,EAAC,QAAQ,OAAO,cAAc,KAAK,OAAO,MAAM,WAAW,uBAAsB;AAAA,IACjF,EAAC,QAAQ,KAAK,cAAc,KAAK,OAAO,MAAM,WAAW,uBAAsB;AAAA,IAC/E,EAAC,QAAQ,KAAK,cAAc,KAAK,OAAO,KAAK,WAAW,uBAAsB;AAAA,EAChF;AAAA,EACA,OAAO;AAAA,IACL,EAAC,QAAQ,KAAK,cAAc,KAAK,OAAO,MAAM,WAAW,uBAAsB;AAAA,IAC/E,EAAC,QAAQ,KAAK,cAAc,KAAK,OAAO,KAAK,WAAW,uBAAsB;AAAA;AAAA;AAAA,IAG9E,EAAC,QAAQ,KAAK,cAAc,KAAK,OAAO,KAAK,WAAW,4BAA4B,QAAQ,CAAC,KAAK,KAAK,KAAK,GAAG,EAAC;AAAA,EAClH;AAAA,EACA,OAAO;AAAA,IACL,EAAC,QAAQ,KAAK,cAAc,KAAK,OAAO,MAAM,WAAW,uBAAsB;AAAA,IAC/E,EAAC,QAAQ,MAAM,cAAc,KAAK,OAAO,MAAM,WAAW,uBAAsB;AAAA,IAChF,EAAC,QAAQ,OAAO,cAAc,KAAK,OAAO,MAAM,WAAW,uBAAsB;AAAA,EACnF;AAAA,EACA,OAAO;AAAA,EACP,OAAO,CAAC,EAAC,QAAQ,KAAK,cAAc,KAAK,OAAO,KAAK,WAAW,uBAAsB,CAAC;AAAA,EACvF,OAAO,CAAC,EAAC,QAAQ,KAAK,cAAc,KAAK,OAAO,KAAK,WAAW,uBAAsB,CAAC;AAAA,EACvF,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO,CAAC,GAAG,QAAQ,GAAG,cAAc;AAAA,EACpC,OAAO,CAAC,EAAC,QAAQ,KAAK,cAAc,KAAK,OAAO,KAAK,WAAW,uBAAsB,CAAC;AACzF;AAoBA,SAAS,0BAA0B,qBAAqB,kBAAkB;AACxE,MAAI,CAAC,qBAAqB;AACxB,WAAO;AAAA,EACT;AACA,QAAM,WAAW,oBAAoB,SAAS,GAAG;AACjD,MAAI,UAAU;AACZ,WAAO,CAAC,oBAAoB,SAAS,gBAAgB;AAAA,EACvD;AACA,SAAO,oBAAoB,SAAS,gBAAgB;AACtD;AAGA,SAAS,mBAAmB,MAAM,OAAO;AACvC,MAAI,UAAU,QAAQ,CAAC,KAAK,KAAK,SAAS,MAAM,IAAI,GAAG;AACrD,WAAO;AAAA,EACT;AAEA,MAAI,UAAU,QAAQ,CAAC,KAAK,KAAK,SAAS,MAAM,IAAI,GAAG;AACrD,WAAO;AAAA,EACT;AAIA,SAAO;AACT;AAGA,SAAS,6BAA6B,MAAM,UAAU;AAEpD,MAAI,CAAC,0BAA0B,KAAK,MAAM,SAAS,IAAI,GAAG;AAExD,WAAO;AAAA,EACT;AACA,MAAI,aAAa,MAAM;AAErB,QAAI,CAAC,SAAS,MAAM,MAAM,KAAK,OAAO,GAAG;AAEvC,aAAO;AAAA,IACT;AAAA,EACF;AAEA,SAAO;AACT;AAEA,SAAS,0BAA0B,MAAM,cAAc;AACrD,MAAI,EAAE,gBAAgB,OAAO;AAC3B,WAAO;AAAA,EACT;AAEA,MAAI,CAAC,cAAc;AACjB,UAAM,WAAW,KAAK,WAAW,SAAS,GAAG;AAC7C,QAAI,UAAU;AACZ,aAAO,CAAC,KAAK,WAAW,SAAS,GAAG;AAAA,IACtC;AACA,WAAO,KAAK,WAAW,SAAS,GAAG;AAAA,EACrC;AAEA,MAAI,CAAC,0BAA0B,KAAK,YAAY,aAAa,IAAI,GAAG;AAClE,WAAO;AAAA,EACT;AACA,MAAI,gBAAgB,QAAQ,CAAC,aAAa,MAAM,MAAM,KAAK,UAAU,GAAG;AACtE,WAAO;AAAA,EACT;AACA,SAAO;AACT;AAEA,SAAS,UAAU,MAAM,OAAO,WAAW,WAAW;AACpD,MAAI,CAAC,mBAAmB,MAAM,KAAK,GAAG;AAEpC,WAAO;AAAA,EACT;AAEA,MAAI,CAAC,6BAA6B,MAAM,SAAS,GAAG;AAElD,WAAO;AAAA,EACT;AAGA,MAAI,CAAC,0BAA0B,MAAM,SAAS,GAAG;AAG/C,WAAO;AAAA,EACT;AAGA,SAAO;AACT;AAGA,SAAS,sBAAsB,OAAO,WAAW,WAAW,YAAY,MAAM,YAAY,MAAM;AAC9F,MAAI,cAAc,QAAQ,cAAc,MAAM;AAC5C;AAAA,EACF;AACA,QAAM,OAAO,MAAM,QAAQ,QAAQ,uBAAuB,KAAK,IAAI,MAAM;AACzE,MAAI,CAAC,MAAM;AACT;AAAA,EACF;AACA,MAAI,EAAE,GAAG,IAAI,MAAM,YAAY;AAC7B;AAAA,EACF;AAKA,QAAM,YAAY,UAAU,IAAI;AAChC,YAAU,MAAM,UAAQ;AAEtB,QAAI,CAAC,UAAU,MAAM,OAAO,WAAW,SAAS,GAAG;AACjD,aAAO;AAAA,IACT;AAGA,QAAI,KAAK,UAAU,CAAC,QAAQ,cAAc,EAAE,SAAS,SAAS,KAAK,UAAU,MAAM,MAAM,KAAK,MAAM,GAAG;AAErG,gBAAU,QAAQ,UAAU,MAAM,QAAQ,KAAK,QAAQ,EAAE;AAEzD,aAAO;AAAA,IACT;AACA,QAAI,KAAK,OAAO,CAAC,KAAK,cAAc,EAAE,SAAS,SAAS,GAAG;AACzD,gBAAU,SAAS,KAAK;AAExB,aAAO;AAAA,IACT;AAQA,WAAO;AAAA,EACT,CAAC;AACH;AAEA,SAAS,uBAAuB,OAAO,WAAW,WAAW;AAC3D,wBAAsB,OAAO,WAAW,WAAW,6BAA6B,MAAM;AACtF,wBAAsB,OAAO,WAAW,WAAW,2BAA2B,GAAG;AACnF;AAEA,SAAS,yBAAyB,OAAO,WAAW,WAAW;AAE7D,wBAAsB,OAAO,WAAW,WAAW,4BAA4B,MAAM;AAErF,wBAAsB,OAAO,WAAW,WAAW,6BAA6B,MAAM;AAGxF;AAEO,gBAAS,sBAAsB,OAAO;AAC3C,MAAI,CAAC,MAAM,WAAW;AACpB,WAAO;AAAA,EACT;AAEA,QAAM,UAAU,QAAQ,CAAC,IAAI,MAAM;AAGjC,6BAAyB,OAAO,IAAI,wBAAwB,OAAO,CAAC,CAAC;AAAA,EAEvE,CAAC;AACD,SAAO;AACT;AAEO,gBAAS,oBAAoB,OAAO;AACzC,MAAI,CAAC,MAAM,WAAW;AACpB,WAAO;AAAA,EACT;AAGA,QAAM,UAAU,QAAQ,CAAC,IAAI,MAAM;AAIjC,2BAAuB,OAAO,IAAI,wBAAwB,OAAO,CAAC,CAAC;AAAA,EACrE,CAAC;AAGD,MAAI,MAAM,2BAA2B;AAEnC,wBAAoB,OAAO,OAAO,IAAI;AAAA,EACxC;AACA,SAAO;AACT;",
4
+ "sourcesContent": ["/* eslint-disable max-lines */\n/*\n* punctuation.js -- try and fix a marc field punctuation\n*\n* Author(s): Nicholas Volk <nicholas.volk@helsinki.fi>\n*\n* NOTE #1: https://www.kiwi.fi/display/kumea/Loppupisteohje is implemented via another validator/fixer (ending-punctuation).\n* This file has some support but it's now yet thorough. (And mmight never be.)\n* NOTE #2: Validator/fixer punctuation does similar stuff, but focuses on X00 fields.\n* NOTE #3: As of 2023-06-05 control subfields ($0...$9) are obsolete. Don't use them in rules.\n* (They are jumped over when looking for next (non-controlfield subfield)\n*/\nimport {validateSingleField} from './ending-punctuation.js';\nimport {tagToDataProvenanceSubfieldCode} from './dataProvenanceUtils.js';\nimport {fieldGetUnambiguousTag} from './subfield6Utils.js';\nimport createDebugLogger from 'debug';\nimport {fieldToString, isContentSubfieldCode, nvdebug} from './utils.js';\nimport clone from 'clone';\n\nconst debug = createDebugLogger('@natlibfi/marc-record-validators-melinda:punctuation2');\n//const debugData = debug.extend('data');\nconst debugDev = debug.extend('dev');\n\nconst descriptionString = 'Remove invalid and add valid punctuation to data fields';\nexport default function () {\n return {\n description: descriptionString,\n validate, fix\n };\n\n function fix(record) {\n nvdebug(`${descriptionString}: fixer`, debugDev);\n const res = {message: [], fix: [], valid: true};\n record.fields.forEach(f => fieldFixPunctuation(f));\n return res;\n }\n\n function validate(record) {\n nvdebug(`${descriptionString}: validate`, debugDev);\n\n const fieldsNeedingModification = record.fields.filter(f => fieldNeedsModification(f, true));\n\n\n const values = fieldsNeedingModification.map(f => fieldToString(f));\n const newValues = fieldsNeedingModification.map(f => fieldGetFixedString(f, true));\n\n const messages = values.map((val, i) => `'${val}' => '${newValues[i]}'`);\n\n const res = {message: messages};\n\n res.valid = res.message.length < 1;\n return res;\n }\n}\n\n\n\nfunction isIrrelevantSubfield(subfield, tag) {\n const dataProvenanceSubfieldCode = tagToDataProvenanceSubfieldCode(tag);\n if (subfield.code === dataProvenanceSubfieldCode) {\n return true;\n }\n return !isContentSubfieldCode(subfield.code); // Currently this contains other stuff as well ($3, $4, $7, $9...)\n}\n\n\nfunction getNextRelevantSubfield(field, currSubfieldIndex) {\n return field.subfields.find((subfield, index) => index > currSubfieldIndex && !isIrrelevantSubfield(subfield, field.tag));\n}\n\nexport function fieldGetFixedString(field, add = true) {\n const cloneField = clone(field);\n const operation = add ? subfieldFixPunctuation : subfieldStripPunctuation;\n cloneField.subfields.forEach((sf, i) => {\n // NB! instead of next subfield, we should actually get next *non-control-subfield*!!!\n // (In plain English: We should skip $0 - $9 at least, maybe $w as well...)\n operation(cloneField, sf, getNextRelevantSubfield(cloneField, i));\n });\n return fieldToString(cloneField);\n}\n\nexport function fieldNeedsModification(field, add = true) {\n if (!field.subfields) {\n return false;\n }\n\n const originalFieldAsString = fieldToString(field);\n const modifiedFieldAsString = fieldGetFixedString(field, add);\n\n return modifiedFieldAsString !== originalFieldAsString;\n}\n\n/////////////////////////////////////////////////////////////////////////////////////\n// <= Above code is written for the validator logic <= //\n// => Everything below was originally transferred from reducers' punctuation.js => //\n/////////////////////////////////////////////////////////////////////////////////////\n\n\n//const stripCrap = / *[-;:,+]+$/u;\nconst needsPuncAfterAlphanumeric = /(?:[a-z0-9A-Z]|\u00E5|\u00E4|\u00F6|\u00C5|\u00C4|\u00D6)$/u;\nconst defaultNeedsPuncAfter2 = /(?:[\\]a-zA-Z0-9)]|\u00E4|\u00E5|\u00F6|\u00C5|\u00C4|\u00D6)$/u;\nconst doesNotEndInPunc = /[^!?.:;,]$/u; // non-punc for pre-240/700/XXX $, note that '.' comes if preceded by ')'\nconst blocksPuncRHS = /^(?:\\()/u;\nconst allowsPuncRHS = /^(?:[A-Za-z0-9]|\u00E5|\u00E4|\u00F6|\u00C5|\u00C4|\u00D6)/u;\nconst aToZ = 'abcdefghijklmnopqrstuvwxyz';\n\n\nconst dotIsProbablyPunc = /(?:[a-z0-9)]|\u00E5|\u00E4|\u00F6|(?:[A-Za-z0-9]|\u00C5|\u00C4|\u00D6)(?:[A-Z]|\u00C5|\u00C4|\u00D6))\\.$/u;\nconst puncIsProbablyPunc = /(?:[a-z0-9)]|\u00E5|\u00E4|\u00F6) ?[.,:;]$/u;\n// NB! 65X: Finnish terms don't use punctuation, but international ones do. Neither one is currently (2021-11-08) coded here.\n\n// Will unfortunately trigger \"Sukunimi, Th.\" type:\nconst removeColons = {'code': 'abcdefghijklmnopqrstuvwxyz', 'remove': / *[;:]$/u};\nconst removeX00Comma = {'code': 'abcdejnqt', 'followedBy': 'abcdenqtv#', 'context': /.,$/u, 'remove': /,$/u};\nconst cleanRHS = {'code': 'abcd', 'followedBy': 'bcde', 'context': /(?:(?:[a-z0-9]|\u00E5|\u00E4|\u00F6)\\.|,)$/u, 'contextRHS': blocksPuncRHS, 'remove': /[.,]$/u};\nconst cleanX00dCommaOrDot = {'code': 'd', 'followedBy': 'et#', 'context': /[0-9]-[,.]$/u, 'remove': /[,.]$/u};\nconst cleanX00aDot = {'code': 'abcde', 'followedBy': 'cdegj', 'context': dotIsProbablyPunc, 'remove': /\\.$/u};\nconst cleanCorruption = {'code': 'abcdefghijklmnopqrstuvwxyz', 'remove': / \\.$/u};\n// These $e dot removals are tricky: before removing the comma, we should know that it ain't an abbreviation such as \"esitt.\"...\nconst cleanX00eDot = {'code': 'e', 'followedBy': 'egj#', 'context': /(?:[ai]ja|j\u00E4)[.,]$/u, 'remove': /\\.$/u};\nconst cleanX11jDot = {'code': 'e', 'followedBy': 'egj#', 'context': /(?:[ai]ja|j\u00E4)[.,]$/u, 'remove': /\\.$/u};\nconst removeCommaBeforeLanguageSubfieldL = {'followedBy': 'l', 'remove': /,$/u};\nconst removeCommaBeforeTitleSubfieldT = {'followedBy': 't', 'remove': /,$/u};\n\nconst X00RemoveDotAfterBracket = {'code': 'cq', 'context': /\\)\\.$/u, 'remove': /\\.$/u};\n// 390, 800, 810, 830...\nconst cleanPuncBeforeLanguage = {'code': 'atvxyz', 'followedBy': 'l', 'context': puncIsProbablyPunc, 'remove': / *[.,:;]$/u};\n\nconst addX00aComma = {'add': ',', 'code': 'abcqejt', 'followedBy': 'cdegnr', 'context': doesNotEndInPunc, 'contextRHS': allowsPuncRHS};\nconst addX00dComma = {'name': 'X00$d ending in \"-\" does not get comma', 'add': ',', 'code': 'd', 'followedBy': 'cdeg', 'context': /[^-,.!]$/u, 'contextRHS': allowsPuncRHS};\nconst addX00aComma2 = {'add': ',', 'code': 'abcdej', 'followedBy': 'cdeg', 'context': /(?:[A-Z]|\u00C5|\u00C4|\u00D6)\\.$/u, 'contextRHS': allowsPuncRHS};\nconst addX00Dot = {'add': '.', 'code': 'abcdetv', 'followedBy': 'fklptu', 'context': needsPuncAfterAlphanumeric};\nconst addEntryFieldFinalDot = {'name': 'X00 final dot', 'add': '.', 'code': 'abcdefghijklmnopqrstuvwxyz', 'followedBy': '#', 'context': /[^.)!?-]$/u};\n\n\nconst addXX0iColon = {name: 'Punctuate relationship information', add: ':', code: 'i', context: defaultNeedsPuncAfter2}; // Not explicitly checking it, but this should always be followed by 'a' or 't'\nconst addX10bDot = {'name': 'Add X10 pre-$b dot', 'add': '.', 'code': 'ab', 'followedBy': 'b', 'context': defaultNeedsPuncAfter2};\nconst addX10Comma = {'add': ',', 'code': 'abet', 'followedBy': 'en', 'context': defaultNeedsPuncAfter2};\nconst addX10Dot = {'name': 'Add X10 final dot', 'add': '.', 'code': 'abet', 'followedBy': 'tu#', 'context': needsPuncAfterAlphanumeric};\nconst addColonToRelationshipInformation = {'name': 'Add \\':\\' to 7X0 $i relationship info', 'add': ':', 'code': 'i', 'context': defaultNeedsPuncAfter2};\n\nconst addX11Spacecolon = {name: '611 space colon(y :-)', add: ' :', code: 'nd', followedBy: 'dc', 'context': defaultNeedsPuncAfter2};\n\nconst addDotBeforeLanguageSubfieldL = {'name': 'Add dot before $l', 'add': '.', 'code': 'abepst', 'followedBy': 'l', 'context': doesNotEndInPunc};\n\n// 490:\nconst addSemicolonBeforeVolumeDesignation = {'name': 'Add \" ;\" before $v', 'add': ' ;', 'code': 'atxyz', 'followedBy': 'v', 'context': /[^;]$/u};\n\nconst NONE = 0;\nconst ADD = 2;\nconst REMOVE = 1;\nconst REMOVE_AND_ADD = 3;\n\n// Crappy punctuation consists of various crap that is somewhat common.\n// We strip crap for merge decisions. We are not trying to actively remove crap here.\n\nconst removeCrapFromAllEntryFields = [removeCommaBeforeLanguageSubfieldL, removeCommaBeforeTitleSubfieldT];\n\nconst removeX00Whatever = [removeX00Comma, cleanX00aDot, cleanX00eDot, cleanCorruption, cleanX00dCommaOrDot, cleanRHS, X00RemoveDotAfterBracket, removeColons, cleanPuncBeforeLanguage, ...removeCrapFromAllEntryFields];\nconst removeX10Whatever = [removeX00Comma, cleanX00aDot, cleanX00eDot, cleanCorruption, removeColons, cleanPuncBeforeLanguage, ...removeCrapFromAllEntryFields];\nconst removeX11Whatever = [removeX00Comma, cleanX11jDot, ...removeCrapFromAllEntryFields];\nconst removeX30Whatever = removeCrapFromAllEntryFields;\n\nconst remove490And830Whatever = [{'code': 'axyzv', 'followedBy': 'axyzv', 'remove': /(?: *;| *=|,)$/u}];\n\nconst linkingEntryRemoveWhatever = [\n {'code': 'i', 'followedBy': 'at', 'remove': / ?:$/u}, // ':'\n {'code': 'at', 'remove': /\\.$/u},\n // Only \". -\" separator is still used in music. We can strip it, but can only create the non-music punctuation!\n {'code': 'abdghiklmnopqrstuwxyz', 'followedBy': 'abdghiklmnopqrstuwxyz#', 'remove': /\\. -$/u}\n];\n\n\n// '!' means negation, thus '!b' means any other subfield but 'b'.\n// 'followedBy': '#' means that current subfield is the last subfield.\n// NB! Note that control subfields are ignored in punctuation rules.\n// NB #2! Control field ignorance causes issues with field 257: https://wiki.helsinki.fi/display/rdasovellusohje/Loppupisteohje\n// Might need to work on that at some point. NOT a top priority though.\n// NB #3! Final punctuation creation is/should be handled by endind-punctuation.js validator!\n\nconst crappy24X = [\n {'code': 'abnp', 'followedBy': '!c', 'remove': / \\/$/u},\n {'code': 'abn', 'followedBy': 'c', 'remove': /\\.$/u, 'context': dotIsProbablyPunc},\n {'code': 'abn', 'followedBy': 'c', 'remove': /\\.$/u, 'context': dotIsProbablyPunc},\n {'code': 'abc', 'followedBy': '#', 'remove': /\\.$/u, 'context': dotIsProbablyPunc},\n {'code': 'abfghinp', 'followedBy': '#', 'remove': /\\.$/u, 'context': dotIsProbablyPunc},\n {'code': 'n', 'followedBy': 'p', 'remove': /\\.$/u, 'context': dotIsProbablyPunc}, // MELINDA-8817\n {'code': 'p', 'followedBy': 'pc', 'remove': /\\.$/u, 'context': dotIsProbablyPunc}, // MELINDA-8817\n removeCommaBeforeLanguageSubfieldL\n];\n\n\nconst cleanCrappyPunctuationRules = {\n '100': removeX00Whatever,\n '110': removeX10Whatever,\n '111': removeX11Whatever,\n '130': removeX30Whatever,\n '240': crappy24X,\n '245': crappy24X,\n '246': crappy24X,\n '300': [\n {'code': 'a', 'followedBy': '!b', 'remove': / *:$/u},\n {'code': 'a', 'followedBy': 'b', 'remove': /:$/u, 'context': /[^ ]:$/u},\n {'code': 'ab', 'followedBy': '!c', 'remove': / *;$/u},\n {'code': 'ab', 'followedBy': 'c', 'remove': /;$/u, 'context': /[^ ];$/u},\n {'code': 'abc', 'followedBy': '!e', 'remove': / *\\+$/u} // Removes both valid (with one space) and invalid (spaceless et al) puncs\n\n ],\n\n '490': remove490And830Whatever,\n '600': removeX00Whatever,\n '610': removeX10Whatever,\n '611': removeX11Whatever,\n '630': removeX30Whatever,\n '700': removeX00Whatever,\n '710': removeX10Whatever,\n '711': removeX11Whatever,\n '730': removeX30Whatever,\n '773': linkingEntryRemoveWhatever,\n '774': linkingEntryRemoveWhatever,\n '776': linkingEntryRemoveWhatever,\n '787': linkingEntryRemoveWhatever,\n '800': removeX00Whatever,\n '810': removeX10Whatever,\n '830': remove490And830Whatever,\n '946': crappy24X\n};\n\nconst cleanLegalX00Comma = {'code': 'abcdetn', 'followedBy': 'cdegjnr', 'context': /.,$/u, 'remove': /,$/u};\n// Accept upper case letters in X00$b, since they are probably Roman numerals.\nconst cleanLegalX00bDot = {'code': 'b', 'followedBy': 't#', context: /^[IVXLCDM]+\\.$/u, 'remove': /\\.$/u};\nconst cleanLegalX00iColon = {'code': 'i', 'followedBy': 'a', 'remove': / *:$/u}; // NB! context is not needed\nconst cleanLegalX00Dot = {'code': 'abcdetkvl', 'followedBy': 'tklu#', 'context': /(?:[a-z0-9)]|\u00E5|\u00E4|\u00F6)\\.$/u, 'remove': /\\.$/u};\nconst cleanDotBeforeLanguageSubfieldL = {'name': 'pre-language-$l dot', 'followedBy': 'l', 'context': /.\\.$/u, 'remove': /\\.$/u};\n\nconst legalEntryField = [cleanDotBeforeLanguageSubfieldL];\n\nconst legalX11SpaceColon = {name: 'legal X11 spacecolony', code: 'nd', followedBy: 'dc', context: / :$/u, remove: / :$/u};\nconst legalX00punc = [cleanLegalX00Comma, cleanLegalX00iColon, cleanLegalX00bDot, cleanLegalX00Dot, ...legalEntryField];\n\nconst cleanLegalX10Comma = {'name': 'X10comma', 'code': 'abe', 'followedBy': 'e', 'context': /.,$/u, 'remove': /,$/u};\nconst cleanLegalX10Dot = {'name': 'X10dot', 'code': 'abt', 'followedBy': 'bst#', 'context': /.\\.$/u, 'remove': /\\.$/u};\n\nconst legalX10punc = [cleanLegalX10Comma, cleanLegalX10Dot, cleanX00eDot, ...legalEntryField];\n\nconst cleanLegalSeriesTitle = [ // 490 and 830\n {'code': 'a', 'followedBy': 'a', 'remove': / =$/u},\n {'code': 'axyz', 'followedBy': 'xyz', 'remove': /,$/u, 'context': /.,$/u},\n {'code': 'axyz', 'followedBy': 'v', 'remove': / *;$/u}\n];\n\nconst clean24X = [\n {'name': 'I:A', 'code': 'i', 'followedBy': 'a', 'remove': / *:$/u},\n {'name': 'A:B', 'code': 'a', 'followedBy': 'b', 'remove': / [:;=]$/u},\n {'name': 'AB:K', 'code': 'ab', 'followedBy': 'k', 'remove': / :$/u},\n {'name': 'ABK:F', 'code': 'abk', 'followedBy': 'f', 'remove': /,$/u},\n {'name': 'ABFNP:C', 'code': 'abfnp', 'followedBy': 'c', 'remove': / \\/$/u},\n {'name': 'ABN:N', 'code': 'abn', 'followedBy': 'n', 'remove': /\\.$/u},\n {'name': 'ABNP:#', 'code': 'abnp', 'followedBy': '#', 'remove': /\\.$/u},\n {'name': 'N:P', 'code': 'n', 'followedBy': 'p', 'remove': /,$/u},\n cleanDotBeforeLanguageSubfieldL\n];\n\nconst legalX11Punc = [...legalEntryField, legalX11SpaceColon];\n\nconst cleanValidPunctuationRules = {\n '100': legalX00punc,\n '110': legalX10punc,\n '111': legalX11Punc,\n '130': legalEntryField,\n '240': clean24X,\n '243': clean24X,\n '245': clean24X,\n '246': clean24X,\n '260': [\n {'code': 'abc', 'followedBy': 'a', 'remove': / ;$/u},\n {'code': 'a', 'followedBy': 'b', 'remove': / :$/u},\n {'code': 'b', 'followedBy': 'c', 'remove': /,$/u},\n {'code': 'c', 'followedBy': '#', 'remove': /\\.$/u},\n {'code': 'd', 'followedBy': 'e', 'remove': / :$/u},\n {'code': 'e', 'followedBy': 'f', 'remove': /,$/u},\n {'code': 'f', 'followedBy': '#', 'remove': /\\.$/u} // Probably ')' but should it be removed?\n ],\n '264': [\n {'code': 'a', 'followedBy': 'b', 'remove': / :$/u},\n {'code': 'b', 'followedBy': 'c', 'remove': /,$/u},\n {'code': 'c', 'followedBy': '#', 'remove': /\\.$/u}\n ],\n '300': [\n // NB! Remove crap as well, thus the '*' in / *:$/\n {'code': 'a', 'followedBy': 'b', 'remove': / :$/u},\n {'code': 'ab', 'followedBy': 'c', 'remove': / ;$/u},\n {'code': 'abc', 'followedBy': 'e', 'remove': / \\+$/u}\n ],\n '490': cleanLegalSeriesTitle,\n '534': [{'code': 'p', 'followedBy': 'c', 'remove': /:$/u}],\n '600': legalX00punc,\n '610': legalX10punc,\n '611': legalX11Punc,\n '630': legalEntryField,\n // Experimental, MET366-ish (end punc in internationally valid, but we don't use it here in Finland):\n '648': [{'code': 'a', 'content': /^[0-9]+\\.$/u, 'ind2': ['4'], 'remove': /\\.$/u}],\n '700': legalX00punc,\n '710': legalX10punc,\n '711': legalX11Punc,\n '730': legalEntryField,\n '800': legalX00punc,\n '810': legalX10punc,\n '811': legalX11Punc,\n '830': [...legalEntryField, ...cleanLegalSeriesTitle],\n '946': clean24X\n};\n\n\n// Overgeneralizes a bit: eg. addColonToRelationshipInformation only applies to 700/710 but as others don't have $i, it's fine.\nconst addToAllEntryFields = [addDotBeforeLanguageSubfieldL, addSemicolonBeforeVolumeDesignation, addColonToRelationshipInformation, addEntryFieldFinalDot];\n\n\nconst addX00 = [addXX0iColon, addX00aComma, addX00aComma2, addX00Dot, addX00dComma, ...addToAllEntryFields];\nconst addX10 = [addXX0iColon, addX10bDot, addX10Comma, addX10Dot, ...addToAllEntryFields];\nconst addX11 = [...addToAllEntryFields, addX11Spacecolon];\nconst addX30 = [...addToAllEntryFields];\n\nconst add24X = [\n {'code': 'i', 'followedBy': 'a', 'add': ':', 'context': needsPuncAfterAlphanumeric},\n {'code': 'a', 'followedBy': 'b', 'add': ' :', 'context': '[^:]$'},\n {'code': 'abk', 'followedBy': 'f', 'add': ',', 'context': needsPuncAfterAlphanumeric},\n {'code': 'abfnp', 'followedBy': 'c', 'add': ' /', 'context': '[^/]$'},\n addDotBeforeLanguageSubfieldL\n];\n\nconst add245 = [\n ...add24X,\n // Blah! Also \"$a = $b\" and \"$a ; $b\" can be valid... But ' :' is better than nothing, I guess...\n {'code': 'ab', 'followedBy': 'n', 'add': '.', 'context': needsPuncAfterAlphanumeric},\n {'code': 'n', 'followedBy': 'p', 'add': ',', 'context': defaultNeedsPuncAfter2},\n {'code': 'abnpc', 'followedBy': '#', 'add': '.', 'context': needsPuncAfterAlphanumeric} // Stepping on \"punctuation validator's\" toes\n];\n\nconst addSeriesTitle = [ // 490 and 830\n {'code': 'a', 'followedBy': 'a', 'add': ' =', 'context': defaultNeedsPuncAfter2},\n {'code': 'axyz', 'followedBy': 'xy', 'add': ',', 'context': defaultNeedsPuncAfter2},\n addSemicolonBeforeVolumeDesignation // eg. 490$axyz-$v\n];\n\nconst addLinkingEntry = [ // NB! Music 773 uses different punctuation rules, that are not implement here (can they even be?)\n {'code': 'i', 'followedBy': aToZ, 'add': ':', 'context': defaultNeedsPuncAfter2},\n {'code': 'a', 'followedBy': 't', 'add': '.', 'context': defaultNeedsPuncAfter2},\n {'code': 't', 'followedBy': 'dghoz', 'add': '.', 'context': defaultNeedsPuncAfter2}\n];\n\nconst addPairedPunctuationRules = {\n '100': addX00,\n '110': addX10,\n '111': addX11,\n '130': addX30,\n '240': add24X,\n '243': add24X,\n '245': add245,\n '246': add24X,\n '260': [\n {'code': 'a', 'followedBy': 'b', 'add': ' :', 'context': defaultNeedsPuncAfter2},\n {'code': 'ab', 'followedBy': 'c', 'add': ',', 'context': defaultNeedsPuncAfter2},\n {'code': 'abc', 'followedBy': 'a', 'add': ' ;', 'context': defaultNeedsPuncAfter2},\n {'code': 'e', 'followedBy': 'f', 'add': ' :', 'context': defaultNeedsPuncAfter2},\n {'code': 'f', 'followedBy': 'g', 'add': ',', 'context': defaultNeedsPuncAfter2}\n ],\n '264': [\n {'code': 'a', 'followedBy': 'b', 'add': ' :', 'context': defaultNeedsPuncAfter2},\n {'code': 'b', 'followedBy': 'c', 'add': ',', 'context': defaultNeedsPuncAfter2},\n // NB! The $c rule messes dotless exception \"264 #4 $c p1983\" up\n // We'll need to add a hacky postprocessor for this? Add 'hasInd1': '0123' etc?\n {'code': 'c', 'followedBy': '#', 'add': '.', 'context': needsPuncAfterAlphanumeric, 'ind2': ['0', '1', '2', '3']}\n ],\n '300': [\n {'code': 'a', 'followedBy': 'b', 'add': ' :', 'context': defaultNeedsPuncAfter2},\n {'code': 'ab', 'followedBy': 'c', 'add': ' ;', 'context': defaultNeedsPuncAfter2},\n {'code': 'abc', 'followedBy': 'e', 'add': ' +', 'context': defaultNeedsPuncAfter2}\n ],\n '490': addSeriesTitle,\n '506': [{'code': 'a', 'followedBy': '#', 'add': '.', 'context': defaultNeedsPuncAfter2}],\n '534': [{'code': 'p', 'followedBy': 'c', 'add': ':', 'context': defaultNeedsPuncAfter2}],\n '600': addX00,\n '610': addX10,\n '611': addX11,\n '630': addX30,\n '700': addX00,\n '710': addX10,\n '711': addX11,\n '730': addX30,\n '773': addLinkingEntry,\n '787': addLinkingEntry,\n '800': addX00,\n '810': addX10,\n '811': addX11,\n '830': [...addX30, ...addSeriesTitle],\n '946': [{'code': 'i', 'followedBy': 'a', 'add': ':', 'context': defaultNeedsPuncAfter2}]\n};\n\n/*\nfunction debugRule(rule) {\n //nvdebug('', debugDev);\n nvdebug(`NAME ${rule.name ? rule.name : '<unnamed>'}`, debugDev);\n nvdebug(`SUBFIELD CODE '${rule.code}' FOLLOWED BY SUBFIELD CODE '${rule.followedBy}'`, debugDev);\n if ('add' in rule) {\n nvdebug(`ADD '${rule.add}'`, debugDev);\n }\n if ('remove' in rule) {\n nvdebug(`REMOVE '${rule.remove}'`, debugDev);\n }\n if ('context' in rule) {\n nvdebug(`CONTEXT '${rule.context.toString()}'`, debugDev);\n }\n //nvdebug('', debugDev);\n}\n*/\n\nfunction ruleAppliesToSubfieldCode(targetSubfieldCodes, currSubfieldCode) {\n if (!targetSubfieldCodes) { // We are not interested in what subfield precedes 240$l, ',' is removed anyway\n return true;\n }\n const negation = targetSubfieldCodes.includes('!');\n if (negation) {\n return !targetSubfieldCodes.includes(currSubfieldCode);\n }\n return targetSubfieldCodes.includes(currSubfieldCode);\n}\n\n\nfunction ruleAppliesToField(rule, field) {\n if ('ind1' in rule && !rule.ind1.includes(field.ind1)) {\n return false;\n }\n\n if ('ind2' in rule && !rule.ind2.includes(field.ind2)) {\n return false;\n }\n\n // If we want to check, say, $2, it should be implemented here!\n\n return true;\n}\n\n\nfunction ruleAppliesToCurrentSubfield(rule, subfield) {\n //nvdebug(` Apply rule on LHS?`, debugDev);\n if (!ruleAppliesToSubfieldCode(rule.code, subfield.code)) {\n //nvdebug(` Reject rule!`, debugDev);\n return false;\n }\n if ('context' in rule) {\n //nvdebug(` Check '${subfield.value}' versus '${rule.context.toString()}'`, debugDev);\n if (!subfield.value.match(rule.context)) { // njsscan-ignore: regex_injection_dos\n //nvdebug(` Reject rule!`, debugDev);\n return false;\n }\n }\n //nvdebug(` Apply rule!`, debugDev);\n return true;\n}\n\nfunction ruleAppliesToNextSubfield(rule, nextSubfield) {\n if (!('followedBy' in rule)) { // Return true, if we are not interested in the next subfield\n return true;\n }\n // The '#' existence check applies only to the RHS field. LHS always exists.\n if (!nextSubfield) {\n const negation = rule.followedBy.includes('!');\n if (negation) {\n return !rule.followedBy.includes('#');\n }\n return rule.followedBy.includes('#');\n }\n\n if (!ruleAppliesToSubfieldCode(rule.followedBy, nextSubfield.code)) {\n return false;\n }\n if ('contextRHS' in rule && !nextSubfield.value.match(rule.contextRHS)) { // njsscan-ignore: regex_injection_dos\n return false;\n }\n return true;\n}\n\nfunction checkRule(rule, field, subfield1, subfield2) {\n if (!ruleAppliesToField(rule, field)) {\n //nvdebug(`FAIL ON WHOLE FIELD: '${fieldToString(field)}`, debugDev);\n return false;\n }\n //const name = rule.name || 'UNNAMED';\n if (!ruleAppliesToCurrentSubfield(rule, subfield1)) {\n //nvdebug(`${name}: FAIL ON LHS SUBFIELD: '$${subfield1.code} ${subfield1.value}', SF=${rule.code}`, debugDev);\n return false;\n }\n\n // NB! This is not a perfect solution. We might have $e$0$e where $e$0 punctuation should actually be based on $e$e rules\n if (!ruleAppliesToNextSubfield(rule, subfield2)) {\n //const msg = subfield2 ? `${name}: FAIL ON RHS SUBFIELD '${subfield2.code}' not in [${rule.followedBy}]` : `${name}: FAIL ON RHS FIELD`;\n //nvdebug(msg, debugDev);\n return false;\n }\n\n \n //nvdebug(`${rule.name ? rule.name : '<unnamed>'}: ACCEPT ${rule.code} (${subfield1.code}), SF2=${rule.followedBy} (${subfield2 ? subfield2.code : '#'})`, debugDev);\n return true;\n}\n\n\nfunction applyPunctuationRules(field, subfield1, subfield2, ruleArray = null, operation = NONE) {\n if (operation === NONE || ruleArray === null) { // !fieldIsApplicable(field, ruleArray)) {\n return;\n }\n const tag2 = field.tag === '880' ? fieldGetUnambiguousTag(field) : field.tag;\n if (!tag2) {\n return;\n }\n if (!(`${tag2}` in ruleArray)) {\n return;\n }\n\n //nvdebug(`PUNCTUATE ${field.tag}/${tag2} '${subfieldToString(subfield1)}' XXX '${subfield2 ? subfieldToString(subfield2) : '#'} }`, debugDev);\n\n //nvdebug(`OP=${operation} ${tag2}: '${subfield1.code}: ${subfield1.value}' ??? '${subfield2 ? subfield2.code : '#'}'`, debugDev);\n const candRules = ruleArray[tag2];\n candRules.every(rule => { // uses \"every\", not \"forEach\", so that only one rule is applies to the given subfields\n //debugRule(rule);\n if (!checkRule(rule, field, subfield1, subfield2)) {\n return true;\n }\n\n //const originalValue = subfield1.value;\n if (rule.remove && [REMOVE, REMOVE_AND_ADD].includes(operation) && subfield1.value.match(rule.remove)) {\n //nvdebug(` PUNC REMOVAL TO BE PERFORMED FOR $${subfield1.code} '${subfield1.value}'`, debugDev);\n subfield1.value = subfield1.value.replace(rule.remove, '');\n //nvdebug(` PUNC REMOVAL PERFORMED FOR '${subfield1.value}'`, debugDev);\n return false;\n }\n if (rule.add && [ADD, REMOVE_AND_ADD].includes(operation)) {\n subfield1.value += rule.add;\n //nvdebug(` ADDED '${rule.add}' TO FORM '${subfield1.value}' USING RULE ${rule.name}`, debugDev);\n return false;\n }\n\n /*\n if (subfield1.value !== originalValue) {\n nvdebug(` PROCESS PUNC: '\u2021${subfield1.code} ${originalValue}' => '\u2021${subfield1.code} ${subfield1.value}'`, debugDev);\n }\n */\n\n return true;\n });\n}\n\nfunction subfieldFixPunctuation(field, subfield1, subfield2) {\n applyPunctuationRules(field, subfield1, subfield2, cleanCrappyPunctuationRules, REMOVE);\n applyPunctuationRules(field, subfield1, subfield2, addPairedPunctuationRules, ADD);\n}\n\nfunction subfieldStripPunctuation(field, subfield1, subfield2) {\n //nvdebug(`FSP1: '${subfield1.value}'`, debugDev);\n applyPunctuationRules(field, subfield1, subfield2, cleanValidPunctuationRules, REMOVE);\n //nvdebug(`FSP2: '${subfield1.value}'`, debugDev);\n applyPunctuationRules(field, subfield1, subfield2, cleanCrappyPunctuationRules, REMOVE);\n //nvdebug(`FSP3: '${subfield1.value}'`, debugDev);\n\n}\n\nexport function fieldStripPunctuation(field) {\n if (!field.subfields) {\n return field;\n }\n\n field.subfields.forEach((sf, i) => {\n // NB! instead of next subfield, we should actually get next *non-control-subfield*!!!\n // (In plain English: We should skip $0 - $9 at least, maybe $w as well...)\n subfieldStripPunctuation(field, sf, getNextRelevantSubfield(field, i));\n\n });\n return field;\n}\n\nexport function fieldFixPunctuation(field) {\n if (!field.subfields) {\n return field;\n }\n //nvdebug(`################### fieldFixPunctuation() TEST ${fieldToString(field)}`, debugDev);\n\n field.subfields.forEach((sf, i) => {\n // NB! instead of next subfield, we should actually get next *non-control-subfield*!!!\n // (In plain English: We should skip $0 - $9 at least, maybe $w as well...)\n // We'll need some magic for field 257 here, do we? (Also Finnish lexicons vs global lexicons in 65X fields)\n subfieldFixPunctuation(field, sf, getNextRelevantSubfield(field, i));\n });\n\n // Use shared code for final punctuation (sadly this does not fix intermediate punc):\n if (field.useExternalEndPunctuation) {\n // addFinalPunctuation(field); // local version. use shared code instead.\n validateSingleField(field, false, true); // NB! Don't use field.tag as second argument! It's a string, not an int. 3rd arg must be true (=fix)\n }\n return field;\n}\n"],
5
+ "mappings": "AAYA,SAAQ,2BAA0B;AAClC,SAAQ,uCAAsC;AAC9C,SAAQ,8BAA6B;AACrC,OAAO,uBAAuB;AAC9B,SAAQ,eAAe,uBAAuB,eAAc;AAC5D,OAAO,WAAW;AAElB,MAAM,QAAQ,kBAAkB,uDAAuD;AAEvF,MAAM,WAAW,MAAM,OAAO,KAAK;AAEnC,MAAM,oBAAoB;AAC1B,0BAA2B;AACzB,SAAO;AAAA,IACL,aAAa;AAAA,IACb;AAAA,IAAU;AAAA,EACZ;AAEA,WAAS,IAAI,QAAQ;AACnB,YAAQ,GAAG,iBAAiB,WAAW,QAAQ;AAC/C,UAAM,MAAM,EAAC,SAAS,CAAC,GAAG,KAAK,CAAC,GAAG,OAAO,KAAI;AAC9C,WAAO,OAAO,QAAQ,OAAK,oBAAoB,CAAC,CAAC;AACjD,WAAO;AAAA,EACT;AAEA,WAAS,SAAS,QAAQ;AACxB,YAAQ,GAAG,iBAAiB,cAAc,QAAQ;AAElD,UAAM,4BAA4B,OAAO,OAAO,OAAO,OAAK,uBAAuB,GAAG,IAAI,CAAC;AAG3F,UAAM,SAAS,0BAA0B,IAAI,OAAK,cAAc,CAAC,CAAC;AAClE,UAAM,YAAY,0BAA0B,IAAI,OAAK,oBAAoB,GAAG,IAAI,CAAC;AAEjF,UAAM,WAAW,OAAO,IAAI,CAAC,KAAK,MAAM,IAAI,GAAG,SAAS,UAAU,CAAC,CAAC,GAAG;AAEvE,UAAM,MAAM,EAAC,SAAS,SAAQ;AAE9B,QAAI,QAAQ,IAAI,QAAQ,SAAS;AACjC,WAAO;AAAA,EACT;AACF;AAIA,SAAS,qBAAqB,UAAU,KAAK;AAC3C,QAAM,6BAA6B,gCAAgC,GAAG;AACtE,MAAI,SAAS,SAAS,4BAA4B;AAChD,WAAO;AAAA,EACT;AACA,SAAO,CAAC,sBAAsB,SAAS,IAAI;AAC7C;AAGA,SAAS,wBAAwB,OAAO,mBAAmB;AACzD,SAAO,MAAM,UAAU,KAAK,CAAC,UAAU,UAAU,QAAQ,qBAAqB,CAAC,qBAAqB,UAAU,MAAM,GAAG,CAAC;AAC1H;AAEO,gBAAS,oBAAoB,OAAO,MAAM,MAAM;AACrD,QAAM,aAAa,MAAM,KAAK;AAC9B,QAAM,YAAY,MAAM,yBAAyB;AACjD,aAAW,UAAU,QAAQ,CAAC,IAAI,MAAM;AAGtC,cAAU,YAAY,IAAI,wBAAwB,YAAY,CAAC,CAAC;AAAA,EAClE,CAAC;AACD,SAAO,cAAc,UAAU;AACjC;AAEO,gBAAS,uBAAuB,OAAO,MAAM,MAAM;AACxD,MAAI,CAAC,MAAM,WAAW;AACpB,WAAO;AAAA,EACT;AAEA,QAAM,wBAAwB,cAAc,KAAK;AACjD,QAAM,wBAAwB,oBAAoB,OAAO,GAAG;AAE5D,SAAO,0BAA0B;AACnC;AASA,MAAM,6BAA6B;AACnC,MAAM,yBAAyB;AAC/B,MAAM,mBAAmB;AACzB,MAAM,gBAAgB;AACtB,MAAM,gBAAgB;AACtB,MAAM,OAAO;AAGb,MAAM,oBAAoB;AAC1B,MAAM,qBAAqB;AAI3B,MAAM,eAAe,EAAC,QAAQ,8BAA8B,UAAU,WAAU;AAChF,MAAM,iBAAiB,EAAC,QAAQ,aAAa,cAAc,cAAc,WAAW,QAAQ,UAAU,MAAK;AAC3G,MAAM,WAAW,EAAC,QAAQ,QAAQ,cAAc,QAAQ,WAAW,gCAAgC,cAAc,eAAe,UAAU,SAAQ;AAClJ,MAAM,sBAAsB,EAAC,QAAQ,KAAK,cAAc,OAAO,WAAW,gBAAgB,UAAU,SAAQ;AAC5G,MAAM,eAAe,EAAC,QAAQ,SAAS,cAAc,SAAS,WAAW,mBAAmB,UAAU,OAAM;AAC5G,MAAM,kBAAkB,EAAC,QAAQ,8BAA8B,UAAU,QAAO;AAEhF,MAAM,eAAe,EAAC,QAAQ,KAAK,cAAc,QAAQ,WAAW,uBAAuB,UAAU,OAAM;AAC3G,MAAM,eAAe,EAAC,QAAQ,KAAK,cAAc,QAAQ,WAAW,uBAAuB,UAAU,OAAM;AAC3G,MAAM,qCAAqC,EAAC,cAAc,KAAK,UAAU,MAAK;AAC9E,MAAM,kCAAkC,EAAC,cAAc,KAAK,UAAU,MAAK;AAE3E,MAAM,2BAA2B,EAAC,QAAQ,MAAM,WAAW,UAAU,UAAU,OAAM;AAErF,MAAM,0BAA0B,EAAC,QAAQ,UAAU,cAAc,KAAK,WAAW,oBAAoB,UAAU,aAAY;AAE3H,MAAM,eAAe,EAAC,OAAO,KAAK,QAAQ,WAAW,cAAc,UAAU,WAAW,kBAAkB,cAAc,cAAa;AACrI,MAAM,eAAe,EAAC,QAAQ,0CAA0C,OAAO,KAAK,QAAQ,KAAK,cAAc,QAAQ,WAAW,aAAa,cAAc,cAAa;AAC1K,MAAM,gBAAgB,EAAC,OAAO,KAAK,QAAQ,UAAU,cAAc,QAAQ,WAAW,uBAAuB,cAAc,cAAa;AACxI,MAAM,YAAY,EAAC,OAAO,KAAK,QAAQ,WAAW,cAAc,UAAU,WAAW,2BAA0B;AAC/G,MAAM,wBAAwB,EAAC,QAAQ,iBAAiB,OAAO,KAAK,QAAQ,8BAA8B,cAAc,KAAK,WAAW,aAAY;AAGpJ,MAAM,eAAe,EAAC,MAAM,sCAAsC,KAAK,KAAK,MAAM,KAAK,SAAS,uBAAsB;AACtH,MAAM,aAAa,EAAC,QAAQ,sBAAsB,OAAO,KAAK,QAAQ,MAAM,cAAc,KAAK,WAAW,uBAAsB;AAChI,MAAM,cAAc,EAAC,OAAO,KAAK,QAAQ,QAAQ,cAAc,MAAM,WAAW,uBAAsB;AACtG,MAAM,YAAY,EAAC,QAAQ,qBAAqB,OAAO,KAAK,QAAQ,QAAQ,cAAc,OAAO,WAAW,2BAA0B;AACtI,MAAM,oCAAoC,EAAC,QAAQ,uCAAyC,OAAO,KAAK,QAAQ,KAAK,WAAW,uBAAsB;AAEtJ,MAAM,mBAAmB,EAAC,MAAM,yBAAyB,KAAK,MAAM,MAAM,MAAM,YAAY,MAAM,WAAW,uBAAsB;AAEnI,MAAM,gCAAgC,EAAC,QAAQ,qBAAqB,OAAO,KAAK,QAAQ,UAAU,cAAc,KAAK,WAAW,iBAAgB;AAGhJ,MAAM,sCAAsC,EAAC,QAAQ,sBAAsB,OAAO,MAAM,QAAQ,SAAS,cAAc,KAAK,WAAW,SAAQ;AAE/I,MAAM,OAAO;AACb,MAAM,MAAM;AACZ,MAAM,SAAS;AACf,MAAM,iBAAiB;AAKvB,MAAM,+BAA+B,CAAC,oCAAoC,+BAA+B;AAEzG,MAAM,oBAAoB,CAAC,gBAAgB,cAAc,cAAc,iBAAiB,qBAAqB,UAAU,0BAA0B,cAAc,yBAAyB,GAAG,4BAA4B;AACvN,MAAM,oBAAoB,CAAC,gBAAgB,cAAc,cAAc,iBAAiB,cAAc,yBAAyB,GAAG,4BAA4B;AAC9J,MAAM,oBAAoB,CAAC,gBAAgB,cAAc,GAAG,4BAA4B;AACxF,MAAM,oBAAoB;AAE1B,MAAM,0BAA0B,CAAC,EAAC,QAAQ,SAAS,cAAc,SAAS,UAAU,kBAAiB,CAAC;AAEtG,MAAM,6BAA6B;AAAA,EACjC,EAAC,QAAQ,KAAK,cAAc,MAAM,UAAU,QAAO;AAAA;AAAA,EACnD,EAAC,QAAQ,MAAM,UAAU,OAAM;AAAA;AAAA,EAE/B,EAAC,QAAQ,yBAAyB,cAAc,0BAA0B,UAAU,SAAQ;AAC9F;AAUA,MAAM,YAAY;AAAA,EAChB,EAAC,QAAQ,QAAQ,cAAc,MAAM,UAAU,QAAO;AAAA,EACtD,EAAC,QAAQ,OAAO,cAAc,KAAK,UAAU,QAAQ,WAAW,kBAAiB;AAAA,EACjF,EAAC,QAAQ,OAAO,cAAc,KAAK,UAAU,QAAQ,WAAW,kBAAiB;AAAA,EACjF,EAAC,QAAQ,OAAO,cAAc,KAAK,UAAU,QAAQ,WAAW,kBAAiB;AAAA,EACjF,EAAC,QAAQ,YAAY,cAAc,KAAK,UAAU,QAAQ,WAAW,kBAAiB;AAAA,EACtF,EAAC,QAAQ,KAAK,cAAc,KAAK,UAAU,QAAQ,WAAW,kBAAiB;AAAA;AAAA,EAC/E,EAAC,QAAQ,KAAK,cAAc,MAAM,UAAU,QAAQ,WAAW,kBAAiB;AAAA;AAAA,EAChF;AACF;AAGA,MAAM,8BAA8B;AAAA,EAClC,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,IACL,EAAC,QAAQ,KAAK,cAAc,MAAM,UAAU,QAAO;AAAA,IACnD,EAAC,QAAQ,KAAK,cAAc,KAAK,UAAU,OAAO,WAAW,UAAS;AAAA,IACtE,EAAC,QAAQ,MAAM,cAAc,MAAM,UAAU,QAAO;AAAA,IACpD,EAAC,QAAQ,MAAM,cAAc,KAAK,UAAU,OAAO,WAAW,UAAS;AAAA,IACvE,EAAC,QAAQ,OAAO,cAAc,MAAM,UAAU,SAAQ;AAAA;AAAA,EAExD;AAAA,EAEA,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AACT;AAEA,MAAM,qBAAqB,EAAC,QAAQ,WAAW,cAAc,WAAW,WAAW,QAAQ,UAAU,MAAK;AAE1G,MAAM,oBAAoB,EAAC,QAAQ,KAAK,cAAc,MAAM,SAAS,mBAAmB,UAAU,OAAM;AACxG,MAAM,sBAAsB,EAAC,QAAQ,KAAK,cAAc,KAAK,UAAU,QAAO;AAC9E,MAAM,mBAAmB,EAAC,QAAQ,aAAa,cAAc,SAAS,WAAW,2BAA2B,UAAU,OAAM;AAC5H,MAAM,kCAAkC,EAAC,QAAQ,uBAAuB,cAAc,KAAK,WAAW,SAAS,UAAU,OAAM;AAE/H,MAAM,kBAAkB,CAAC,+BAA+B;AAExD,MAAM,qBAAqB,EAAC,MAAM,yBAAyB,MAAM,MAAM,YAAY,MAAM,SAAS,QAAQ,QAAQ,OAAM;AACxH,MAAM,eAAe,CAAC,oBAAoB,qBAAqB,mBAAmB,kBAAkB,GAAG,eAAe;AAEtH,MAAM,qBAAqB,EAAC,QAAQ,YAAY,QAAQ,OAAO,cAAc,KAAK,WAAW,QAAQ,UAAU,MAAK;AACpH,MAAM,mBAAmB,EAAC,QAAQ,UAAU,QAAQ,OAAO,cAAc,QAAQ,WAAW,SAAS,UAAU,OAAM;AAErH,MAAM,eAAe,CAAC,oBAAoB,kBAAkB,cAAc,GAAG,eAAe;AAE5F,MAAM,wBAAwB;AAAA;AAAA,EAC5B,EAAC,QAAQ,KAAK,cAAc,KAAK,UAAU,OAAM;AAAA,EACjD,EAAC,QAAQ,QAAQ,cAAc,OAAO,UAAU,OAAO,WAAW,OAAM;AAAA,EACxE,EAAC,QAAQ,QAAQ,cAAc,KAAK,UAAU,QAAO;AACvD;AAEA,MAAM,WAAW;AAAA,EACf,EAAC,QAAQ,OAAO,QAAQ,KAAK,cAAc,KAAK,UAAU,QAAO;AAAA,EACjE,EAAC,QAAQ,OAAO,QAAQ,KAAK,cAAc,KAAK,UAAU,WAAU;AAAA,EACpE,EAAC,QAAQ,QAAQ,QAAQ,MAAM,cAAc,KAAK,UAAU,OAAM;AAAA,EAClE,EAAC,QAAQ,SAAS,QAAQ,OAAO,cAAc,KAAK,UAAU,MAAK;AAAA,EACnE,EAAC,QAAQ,WAAW,QAAQ,SAAS,cAAc,KAAK,UAAU,QAAO;AAAA,EACzE,EAAC,QAAQ,SAAS,QAAQ,OAAO,cAAc,KAAK,UAAU,OAAM;AAAA,EACpE,EAAC,QAAQ,UAAU,QAAQ,QAAQ,cAAc,KAAK,UAAU,OAAM;AAAA,EACtE,EAAC,QAAQ,OAAO,QAAQ,KAAK,cAAc,KAAK,UAAU,MAAK;AAAA,EAC/D;AACF;AAEA,MAAM,eAAe,CAAC,GAAG,iBAAiB,kBAAkB;AAE5D,MAAM,6BAA6B;AAAA,EACjC,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,IACL,EAAC,QAAQ,OAAO,cAAc,KAAK,UAAU,OAAM;AAAA,IACnD,EAAC,QAAQ,KAAK,cAAc,KAAK,UAAU,OAAM;AAAA,IACjD,EAAC,QAAQ,KAAK,cAAc,KAAK,UAAU,MAAK;AAAA,IAChD,EAAC,QAAQ,KAAK,cAAc,KAAK,UAAU,OAAM;AAAA,IACjD,EAAC,QAAQ,KAAK,cAAc,KAAK,UAAU,OAAM;AAAA,IACjD,EAAC,QAAQ,KAAK,cAAc,KAAK,UAAU,MAAK;AAAA,IAChD,EAAC,QAAQ,KAAK,cAAc,KAAK,UAAU,OAAM;AAAA;AAAA,EACnD;AAAA,EACA,OAAO;AAAA,IACL,EAAC,QAAQ,KAAK,cAAc,KAAK,UAAU,OAAM;AAAA,IACjD,EAAC,QAAQ,KAAK,cAAc,KAAK,UAAU,MAAK;AAAA,IAChD,EAAC,QAAQ,KAAK,cAAc,KAAK,UAAU,OAAM;AAAA,EACnD;AAAA,EACA,OAAO;AAAA;AAAA,IAEL,EAAC,QAAQ,KAAK,cAAc,KAAK,UAAU,OAAM;AAAA,IACjD,EAAC,QAAQ,MAAM,cAAc,KAAK,UAAU,OAAM;AAAA,IAClD,EAAC,QAAQ,OAAO,cAAc,KAAK,UAAU,QAAO;AAAA,EACtD;AAAA,EACA,OAAO;AAAA,EACP,OAAO,CAAC,EAAC,QAAQ,KAAK,cAAc,KAAK,UAAU,MAAK,CAAC;AAAA,EACzD,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA;AAAA,EAEP,OAAO,CAAC,EAAC,QAAQ,KAAK,WAAW,eAAe,QAAQ,CAAC,GAAG,GAAG,UAAU,OAAM,CAAC;AAAA,EAChF,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO,CAAC,GAAG,iBAAiB,GAAG,qBAAqB;AAAA,EACpD,OAAO;AACT;AAIA,MAAM,sBAAsB,CAAC,+BAA+B,qCAAqC,mCAAmC,qBAAqB;AAGzJ,MAAM,SAAS,CAAC,cAAc,cAAc,eAAe,WAAW,cAAc,GAAG,mBAAmB;AAC1G,MAAM,SAAS,CAAC,cAAc,YAAY,aAAa,WAAW,GAAG,mBAAmB;AACxF,MAAM,SAAS,CAAC,GAAG,qBAAqB,gBAAgB;AACxD,MAAM,SAAS,CAAC,GAAG,mBAAmB;AAEtC,MAAM,SAAS;AAAA,EACb,EAAC,QAAQ,KAAK,cAAc,KAAK,OAAO,KAAK,WAAW,2BAA0B;AAAA,EAClF,EAAC,QAAQ,KAAK,cAAc,KAAK,OAAO,MAAM,WAAW,QAAO;AAAA,EAChE,EAAC,QAAQ,OAAO,cAAc,KAAK,OAAO,KAAK,WAAW,2BAA0B;AAAA,EACpF,EAAC,QAAQ,SAAS,cAAc,KAAK,OAAO,MAAM,WAAW,QAAO;AAAA,EACpE;AACF;AAEA,MAAM,SAAS;AAAA,EACb,GAAG;AAAA;AAAA,EAEH,EAAC,QAAQ,MAAM,cAAc,KAAK,OAAO,KAAK,WAAW,2BAA0B;AAAA,EACnF,EAAC,QAAQ,KAAK,cAAc,KAAK,OAAO,KAAK,WAAW,uBAAsB;AAAA,EAC9E,EAAC,QAAQ,SAAS,cAAc,KAAK,OAAO,KAAK,WAAW,2BAA0B;AAAA;AACxF;AAEA,MAAM,iBAAiB;AAAA;AAAA,EACrB,EAAC,QAAQ,KAAK,cAAc,KAAK,OAAO,MAAM,WAAW,uBAAsB;AAAA,EAC/E,EAAC,QAAQ,QAAQ,cAAc,MAAM,OAAO,KAAK,WAAW,uBAAsB;AAAA,EAClF;AAAA;AACF;AAEA,MAAM,kBAAkB;AAAA;AAAA,EACtB,EAAC,QAAQ,KAAK,cAAc,MAAM,OAAO,KAAK,WAAW,uBAAsB;AAAA,EAC/E,EAAC,QAAQ,KAAK,cAAc,KAAK,OAAO,KAAK,WAAW,uBAAsB;AAAA,EAC9E,EAAC,QAAQ,KAAK,cAAc,SAAS,OAAO,KAAK,WAAW,uBAAsB;AACpF;AAEA,MAAM,4BAA4B;AAAA,EAChC,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,IACL,EAAC,QAAQ,KAAK,cAAc,KAAK,OAAO,MAAM,WAAW,uBAAsB;AAAA,IAC/E,EAAC,QAAQ,MAAM,cAAc,KAAK,OAAO,KAAK,WAAW,uBAAsB;AAAA,IAC/E,EAAC,QAAQ,OAAO,cAAc,KAAK,OAAO,MAAM,WAAW,uBAAsB;AAAA,IACjF,EAAC,QAAQ,KAAK,cAAc,KAAK,OAAO,MAAM,WAAW,uBAAsB;AAAA,IAC/E,EAAC,QAAQ,KAAK,cAAc,KAAK,OAAO,KAAK,WAAW,uBAAsB;AAAA,EAChF;AAAA,EACA,OAAO;AAAA,IACL,EAAC,QAAQ,KAAK,cAAc,KAAK,OAAO,MAAM,WAAW,uBAAsB;AAAA,IAC/E,EAAC,QAAQ,KAAK,cAAc,KAAK,OAAO,KAAK,WAAW,uBAAsB;AAAA;AAAA;AAAA,IAG9E,EAAC,QAAQ,KAAK,cAAc,KAAK,OAAO,KAAK,WAAW,4BAA4B,QAAQ,CAAC,KAAK,KAAK,KAAK,GAAG,EAAC;AAAA,EAClH;AAAA,EACA,OAAO;AAAA,IACL,EAAC,QAAQ,KAAK,cAAc,KAAK,OAAO,MAAM,WAAW,uBAAsB;AAAA,IAC/E,EAAC,QAAQ,MAAM,cAAc,KAAK,OAAO,MAAM,WAAW,uBAAsB;AAAA,IAChF,EAAC,QAAQ,OAAO,cAAc,KAAK,OAAO,MAAM,WAAW,uBAAsB;AAAA,EACnF;AAAA,EACA,OAAO;AAAA,EACP,OAAO,CAAC,EAAC,QAAQ,KAAK,cAAc,KAAK,OAAO,KAAK,WAAW,uBAAsB,CAAC;AAAA,EACvF,OAAO,CAAC,EAAC,QAAQ,KAAK,cAAc,KAAK,OAAO,KAAK,WAAW,uBAAsB,CAAC;AAAA,EACvF,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO;AAAA,EACP,OAAO,CAAC,GAAG,QAAQ,GAAG,cAAc;AAAA,EACpC,OAAO,CAAC,EAAC,QAAQ,KAAK,cAAc,KAAK,OAAO,KAAK,WAAW,uBAAsB,CAAC;AACzF;AAoBA,SAAS,0BAA0B,qBAAqB,kBAAkB;AACxE,MAAI,CAAC,qBAAqB;AACxB,WAAO;AAAA,EACT;AACA,QAAM,WAAW,oBAAoB,SAAS,GAAG;AACjD,MAAI,UAAU;AACZ,WAAO,CAAC,oBAAoB,SAAS,gBAAgB;AAAA,EACvD;AACA,SAAO,oBAAoB,SAAS,gBAAgB;AACtD;AAGA,SAAS,mBAAmB,MAAM,OAAO;AACvC,MAAI,UAAU,QAAQ,CAAC,KAAK,KAAK,SAAS,MAAM,IAAI,GAAG;AACrD,WAAO;AAAA,EACT;AAEA,MAAI,UAAU,QAAQ,CAAC,KAAK,KAAK,SAAS,MAAM,IAAI,GAAG;AACrD,WAAO;AAAA,EACT;AAIA,SAAO;AACT;AAGA,SAAS,6BAA6B,MAAM,UAAU;AAEpD,MAAI,CAAC,0BAA0B,KAAK,MAAM,SAAS,IAAI,GAAG;AAExD,WAAO;AAAA,EACT;AACA,MAAI,aAAa,MAAM;AAErB,QAAI,CAAC,SAAS,MAAM,MAAM,KAAK,OAAO,GAAG;AAEvC,aAAO;AAAA,IACT;AAAA,EACF;AAEA,SAAO;AACT;AAEA,SAAS,0BAA0B,MAAM,cAAc;AACrD,MAAI,EAAE,gBAAgB,OAAO;AAC3B,WAAO;AAAA,EACT;AAEA,MAAI,CAAC,cAAc;AACjB,UAAM,WAAW,KAAK,WAAW,SAAS,GAAG;AAC7C,QAAI,UAAU;AACZ,aAAO,CAAC,KAAK,WAAW,SAAS,GAAG;AAAA,IACtC;AACA,WAAO,KAAK,WAAW,SAAS,GAAG;AAAA,EACrC;AAEA,MAAI,CAAC,0BAA0B,KAAK,YAAY,aAAa,IAAI,GAAG;AAClE,WAAO;AAAA,EACT;AACA,MAAI,gBAAgB,QAAQ,CAAC,aAAa,MAAM,MAAM,KAAK,UAAU,GAAG;AACtE,WAAO;AAAA,EACT;AACA,SAAO;AACT;AAEA,SAAS,UAAU,MAAM,OAAO,WAAW,WAAW;AACpD,MAAI,CAAC,mBAAmB,MAAM,KAAK,GAAG;AAEpC,WAAO;AAAA,EACT;AAEA,MAAI,CAAC,6BAA6B,MAAM,SAAS,GAAG;AAElD,WAAO;AAAA,EACT;AAGA,MAAI,CAAC,0BAA0B,MAAM,SAAS,GAAG;AAG/C,WAAO;AAAA,EACT;AAIA,SAAO;AACT;AAGA,SAAS,sBAAsB,OAAO,WAAW,WAAW,YAAY,MAAM,YAAY,MAAM;AAC9F,MAAI,cAAc,QAAQ,cAAc,MAAM;AAC5C;AAAA,EACF;AACA,QAAM,OAAO,MAAM,QAAQ,QAAQ,uBAAuB,KAAK,IAAI,MAAM;AACzE,MAAI,CAAC,MAAM;AACT;AAAA,EACF;AACA,MAAI,EAAE,GAAG,IAAI,MAAM,YAAY;AAC7B;AAAA,EACF;AAKA,QAAM,YAAY,UAAU,IAAI;AAChC,YAAU,MAAM,UAAQ;AAEtB,QAAI,CAAC,UAAU,MAAM,OAAO,WAAW,SAAS,GAAG;AACjD,aAAO;AAAA,IACT;AAGA,QAAI,KAAK,UAAU,CAAC,QAAQ,cAAc,EAAE,SAAS,SAAS,KAAK,UAAU,MAAM,MAAM,KAAK,MAAM,GAAG;AAErG,gBAAU,QAAQ,UAAU,MAAM,QAAQ,KAAK,QAAQ,EAAE;AAEzD,aAAO;AAAA,IACT;AACA,QAAI,KAAK,OAAO,CAAC,KAAK,cAAc,EAAE,SAAS,SAAS,GAAG;AACzD,gBAAU,SAAS,KAAK;AAExB,aAAO;AAAA,IACT;AAQA,WAAO;AAAA,EACT,CAAC;AACH;AAEA,SAAS,uBAAuB,OAAO,WAAW,WAAW;AAC3D,wBAAsB,OAAO,WAAW,WAAW,6BAA6B,MAAM;AACtF,wBAAsB,OAAO,WAAW,WAAW,2BAA2B,GAAG;AACnF;AAEA,SAAS,yBAAyB,OAAO,WAAW,WAAW;AAE7D,wBAAsB,OAAO,WAAW,WAAW,4BAA4B,MAAM;AAErF,wBAAsB,OAAO,WAAW,WAAW,6BAA6B,MAAM;AAGxF;AAEO,gBAAS,sBAAsB,OAAO;AAC3C,MAAI,CAAC,MAAM,WAAW;AACpB,WAAO;AAAA,EACT;AAEA,QAAM,UAAU,QAAQ,CAAC,IAAI,MAAM;AAGjC,6BAAyB,OAAO,IAAI,wBAAwB,OAAO,CAAC,CAAC;AAAA,EAEvE,CAAC;AACD,SAAO;AACT;AAEO,gBAAS,oBAAoB,OAAO;AACzC,MAAI,CAAC,MAAM,WAAW;AACpB,WAAO;AAAA,EACT;AAGA,QAAM,UAAU,QAAQ,CAAC,IAAI,MAAM;AAIjC,2BAAuB,OAAO,IAAI,wBAAwB,OAAO,CAAC,CAAC;AAAA,EACrE,CAAC;AAGD,MAAI,MAAM,2BAA2B;AAEnC,wBAAoB,OAAO,OAAO,IAAI;AAAA,EACxC;AACA,SAAO;AACT;",
6
6
  "names": []
7
7
  }
@@ -12,6 +12,7 @@ import {
12
12
  subfield6ResetOccurrenceNumber
13
13
  } from "./subfield6Utils.js";
14
14
  const debug = createDebugLogger("@natlibfi/marc-record-validators-melinda:reindexSubfield6OccurrenceNumbers");
15
+ const debugDev = debug.extend("dev");
15
16
  export default function() {
16
17
  return {
17
18
  description: "Reindex occurrence numbers in $6 subfield so that they start from 01 and end in NN",
@@ -19,7 +20,7 @@ export default function() {
19
20
  fix
20
21
  };
21
22
  function fix(record) {
22
- nvdebug("Fix SF6 occurrence numbers", debug);
23
+ nvdebug("Fix SF6 occurrence numbers", debugDev);
23
24
  const res = { message: [], fix: [], valid: true };
24
25
  recordDisambiguateSharedSubfield6OccurrenceNumbers(record);
25
26
  recordResetSubfield6OccurrenceNumbers(record);
@@ -27,11 +28,11 @@ export default function() {
27
28
  }
28
29
  function validate(record) {
29
30
  const res = { message: [] };
30
- nvdebug("Validate SF6 occurrence number multiuses", debug);
31
+ nvdebug("Validate SF6 occurrence number multiuses", debugDev);
31
32
  if (recordGetSharedOccurrenceNumbers(record).length) {
32
33
  res.message.push(`Multi-use of occurrence number(s) detected`);
33
34
  }
34
- nvdebug("Validate SF6 occurrence number (max vs n instances)", debug);
35
+ nvdebug("Validate SF6 occurrence number (max vs n instances)", debugDev);
35
36
  const max = recordGetMaxSubfield6OccurrenceNumberAsInteger(record);
36
37
  const size = recordGetNumberOfUniqueSubfield6OccurrenceNumbers(record);
37
38
  if (max !== size) {
@@ -67,26 +68,26 @@ function recordDisambiguateSharedSubfield6OccurrenceNumbers(record) {
67
68
  if (sharedOccurrenceNumberFields.length < 2) {
68
69
  return;
69
70
  }
70
- nvdebug(`Disambiguate occurrence numbers (N=${sharedOccurrenceNumberFields.length}) in...`, debug);
71
+ nvdebug(`Disambiguate occurrence numbers (N=${sharedOccurrenceNumberFields.length}) in...`, debugDev);
71
72
  sharedOccurrenceNumberFields.forEach((field) => disambiguateOccurrenceNumber(field));
72
73
  function disambiguateable(field) {
73
74
  if (field.tag === "880") {
74
75
  return false;
75
76
  }
76
77
  const occurrenceNumber = fieldGetUnambiguousOccurrenceNumber(field);
77
- nvdebug(` Trying to disambiguate ${occurrenceNumber} in '${fieldToString(field)}`);
78
+ nvdebug(` Trying to disambiguate ${occurrenceNumber} in '${fieldToString(field)}`, debugDev);
78
79
  if (occurrenceNumber === void 0) {
79
80
  return false;
80
81
  }
81
82
  const allRelevantFields = getPotentialSharedOccurrenceNumberFields(occurrenceNumber, sharedOccurrenceNumberFields);
82
83
  if (allRelevantFields.length < 2) {
83
- nvdebug(` Currently only ${allRelevantFields.length} field(s) use occurrence number ${occurrenceNumber}. No action required.`);
84
+ nvdebug(` Currently only ${allRelevantFields.length} field(s) use occurrence number ${occurrenceNumber}. No action required.`, debugDev);
84
85
  return false;
85
86
  }
86
- nvdebug(` Currently ${allRelevantFields.length} field(s) use occurrence number ${occurrenceNumber}. ACTION REQUIRED!`);
87
+ nvdebug(` Currently ${allRelevantFields.length} field(s) use occurrence number ${occurrenceNumber}. ACTION REQUIRED!`, debugDev);
87
88
  const relevantFieldsWithCurrFieldTag = allRelevantFields.filter((candField) => field.tag === candField.tag);
88
89
  if (relevantFieldsWithCurrFieldTag.length !== 1) {
89
- nvdebug(` Number of them using tag ${field.tag} is ${relevantFieldsWithCurrFieldTag.length}. Can not disambiguate!`);
90
+ nvdebug(` Number of them using tag ${field.tag} is ${relevantFieldsWithCurrFieldTag.length}. Can not disambiguate!`, debugDev);
90
91
  return false;
91
92
  }
92
93
  return true;
@@ -99,7 +100,7 @@ function recordDisambiguateSharedSubfield6OccurrenceNumbers(record) {
99
100
  const newOccurrenceNumberAsInt = recordGetMaxSubfield6OccurrenceNumberAsInteger(record) + 1;
100
101
  const newOccurrenceNumber = intToOccurrenceNumberString(newOccurrenceNumberAsInt);
101
102
  const pairedFields = fieldGetOccurrenceNumberPairs(field, record.fields);
102
- nvdebug(` Reindex '${fieldToString(field)}' occurrence number and it's ${pairedFields.length} pair(s) using '${newOccurrenceNumber}'`, debug);
103
+ nvdebug(` Reindex '${fieldToString(field)}' occurrence number and it's ${pairedFields.length} pair(s) using '${newOccurrenceNumber}'`, debugDev);
103
104
  fieldResetOccurrenceNumber(field, newOccurrenceNumber, occurrenceNumber);
104
105
  pairedFields.forEach((pairedField) => fieldResetOccurrenceNumber(pairedField, newOccurrenceNumber, occurrenceNumber));
105
106
  }
@@ -132,7 +133,7 @@ export function recordResetSubfield6OccurrenceNumbers(record) {
132
133
  let oldtoNewCache = {};
133
134
  record.fields.forEach((field) => fieldResetSubfield6(field));
134
135
  function fieldResetSubfield6(field) {
135
- nvdebug(`fieldResetSubfield6(${fieldToString(field)}), CURR:${currentInt}`, debug);
136
+ nvdebug(`fieldResetSubfield6(${fieldToString(field)}), CURR:${currentInt}`, debugDev);
136
137
  if (!field.subfields) {
137
138
  return;
138
139
  }
@@ -1,7 +1,7 @@
1
1
  {
2
2
  "version": 3,
3
3
  "sources": ["../src/reindexSubfield6OccurenceNumbers.js"],
4
- "sourcesContent": ["import createDebugLogger from 'debug';\nimport {fieldHasSubfield, fieldToString, nvdebug} from './utils.js';\nimport {fieldGetOccurrenceNumberPairs, fieldGetUnambiguousOccurrenceNumber, fieldResetOccurrenceNumber, intToOccurrenceNumberString, isValidSubfield6,\n recordGetMaxSubfield6OccurrenceNumberAsInteger,\n subfield6GetOccurrenceNumber, subfield6GetOccurrenceNumberAsInteger, subfield6ResetOccurrenceNumber} from './subfield6Utils.js';\n\n// Relocated from melinda-marc-record-merge-reducers (and renamed)\n\nconst debug = createDebugLogger('@natlibfi/marc-record-validators-melinda:reindexSubfield6OccurrenceNumbers');\n\n\n// NB! This validator/fixer has two functionalities:\n// 1) normal reindexing of occurrence numbers\n// 2) disambiguation (when possible) of unambiguous occurrence numbers\n\nexport default function () {\n return {\n description: 'Reindex occurrence numbers in $6 subfield so that they start from 01 and end in NN',\n validate, fix\n };\n\n function fix(record) {\n nvdebug('Fix SF6 occurrence numbers', debug);\n const res = {message: [], fix: [], valid: true};\n //message.fix = [];\n\n // This can not really fail...\n\n recordDisambiguateSharedSubfield6OccurrenceNumbers(record);\n recordResetSubfield6OccurrenceNumbers(record);\n\n // message.valid = !(message.message.length >= 1);\n return res;\n }\n\n function validate(record) {\n const res = {message: []};\n\n nvdebug('Validate SF6 occurrence number multiuses', debug);\n if (recordGetSharedOccurrenceNumbers(record).length) {\n res.message.push(`Multi-use of occurrence number(s) detected`);\n }\n\n // Check max, and check number of different indexes\n nvdebug('Validate SF6 occurrence number (max vs n instances)', debug);\n const max = recordGetMaxSubfield6OccurrenceNumberAsInteger(record);\n const size = recordGetNumberOfUniqueSubfield6OccurrenceNumbers(record);\n\n\n if (max !== size) {\n res.message.push(`Gaps detected in occurrence numbers: found ${size}, seen max ${max}`);\n }\n res.valid = res.message.length < 1;\n return res;\n }\n}\n\nfunction getPotentialSharedOccurrenceNumberFields(occurrenceNumber, fields) {\n return fields.filter(f => f.tag !== '880' && f.subfields.some(sf => subfield6GetOccurrenceNumber(sf) === occurrenceNumber));\n}\n\nfunction subfieldHasSharedOccurrenceNumber(subfield, candFields) {\n const occurrenceNumber = subfield6GetOccurrenceNumber(subfield);\n if (!occurrenceNumber || occurrenceNumber === '00') {\n return false;\n }\n const relevantFields = getPotentialSharedOccurrenceNumberFields(occurrenceNumber, candFields);\n // record.fields.filter(f => f.tag !== '880' && fieldHasOccurrenceNumber(f, occurrenceNumber));\n return relevantFields.length > 1;\n}\n\nfunction fieldHasSharedOccurrenceNumber(field, candFields) {\n if (!field.subfields || field.tag === '880') { // Should not happen\n return false;\n }\n\n // What if there are multiple $6s in a given field? Should not be, but...\n return field.subfields.some(subfield => subfieldHasSharedOccurrenceNumber(subfield, candFields));\n\n}\n\nfunction recordGetSharedOccurrenceNumbers(record) {\n const fieldsContainingSubfield6 = record.fields.filter(field => field.tag !== '880' && fieldHasSubfield(field, '6'));\n // fieldsContainingSubfield6.some(field => fieldHasSharedOccurrenceNumber(field, fieldsContainingSubfield6)))\n return fieldsContainingSubfield6.filter(field => fieldHasSharedOccurrenceNumber(field, fieldsContainingSubfield6));\n}\n\nfunction recordDisambiguateSharedSubfield6OccurrenceNumbers(record) {\n const sharedOccurrenceNumberFields = recordGetSharedOccurrenceNumbers(record);\n if (sharedOccurrenceNumberFields.length < 2) {\n return;\n }\n nvdebug(`Disambiguate occurrence numbers (N=${sharedOccurrenceNumberFields.length}) in...`, debug);\n sharedOccurrenceNumberFields.forEach(field => disambiguateOccurrenceNumber(field));\n\n function disambiguateable(field) {\n if (field.tag === '880') { // Not needed, already filtered...\n return false;\n }\n const occurrenceNumber = fieldGetUnambiguousOccurrenceNumber(field);\n nvdebug(` Trying to disambiguate ${occurrenceNumber} in '${fieldToString(field)}`);\n if (occurrenceNumber === undefined) {\n return false;\n }\n const allRelevantFields = getPotentialSharedOccurrenceNumberFields(occurrenceNumber, sharedOccurrenceNumberFields);\n if (allRelevantFields.length < 2) {\n nvdebug(` Currently only ${allRelevantFields.length} field(s) use occurrence number ${occurrenceNumber}. No action required.`);\n return false;\n }\n nvdebug(` Currently ${allRelevantFields.length} field(s) use occurrence number ${occurrenceNumber}. ACTION REQUIRED!`);\n const relevantFieldsWithCurrFieldTag = allRelevantFields.filter(candField => field.tag === candField.tag);\n\n if (relevantFieldsWithCurrFieldTag.length !== 1) {\n nvdebug(` Number of them using tag ${field.tag} is ${relevantFieldsWithCurrFieldTag.length}. Can not disambiguate!`);\n return false;\n }\n\n return true;\n }\n\n function disambiguateOccurrenceNumber(field) {\n if (!disambiguateable(field)) {\n return;\n }\n // Reset field:\n const occurrenceNumber = fieldGetUnambiguousOccurrenceNumber(field);\n const newOccurrenceNumberAsInt = recordGetMaxSubfield6OccurrenceNumberAsInteger(record) + 1;\n const newOccurrenceNumber = intToOccurrenceNumberString(newOccurrenceNumberAsInt);\n const pairedFields = fieldGetOccurrenceNumberPairs(field, record.fields);\n\n nvdebug(` Reindex '${fieldToString(field)}' occurrence number and it's ${pairedFields.length} pair(s) using '${newOccurrenceNumber}'`, debug);\n\n fieldResetOccurrenceNumber(field, newOccurrenceNumber, occurrenceNumber);\n pairedFields.forEach(pairedField => fieldResetOccurrenceNumber(pairedField, newOccurrenceNumber, occurrenceNumber));\n\n }\n\n\n}\nfunction recordGetNumberOfUniqueSubfield6OccurrenceNumbers(record) {\n // Calculates the number of used different occurrence numbers\n let indexArray = [];\n record.fields.forEach(field => gatherFieldData(field));\n\n function gatherFieldData(field) {\n if (!field.subfields) {\n return;\n }\n field.subfields.forEach(subfield => gatherSubfieldData(subfield));\n }\n\n function gatherSubfieldData(subfield) {\n if (!isValidSubfield6(subfield)) {\n return;\n }\n const i = subfield6GetOccurrenceNumberAsInteger(subfield);\n if (i === 0) {\n return\n }\n indexArray[i] = 1;\n }\n let n = 0;\n indexArray.forEach(elem => n+= elem);\n return n;\n}\n\nexport function recordResetSubfield6OccurrenceNumbers(record) { // Remove gaps\n /* eslint-disable */\n let currentInt = 1;\n let oldtoNewCache = {};\n\n record.fields.forEach(field => fieldResetSubfield6(field));\n\n function fieldResetSubfield6(field) {\n nvdebug(`fieldResetSubfield6(${fieldToString(field)}), CURR:${currentInt}`, debug);\n if (!field.subfields) {\n return;\n }\n field.subfields.forEach(subfield => subfieldReset6(subfield));\n }\n\n function subfieldReset6(subfield) {\n if (!isValidSubfield6(subfield)) {\n return;\n }\n const currIndex = subfield6GetOccurrenceNumber(subfield);\n if (currIndex === undefined || currIndex === '00') {\n return;\n }\n\n const newIndex = mapCurrIndexToNewIndex(currIndex);\n //nvdebug(`subfieldReset6(${subfieldToString(subfield)}): ${newIndex}`, debug);\n subfield6ResetOccurrenceNumber(subfield, newIndex);\n }\n\n function mapCurrIndexToNewIndex(currIndex) {\n if(currIndex in oldtoNewCache) {\n return oldtoNewCache[currIndex];\n }\n const newIndex = intToOccurrenceNumberString(currentInt);\n oldtoNewCache[currIndex] = newIndex;\n currentInt++;\n return newIndex;\n }\n\n /* eslint-enable */\n\n}\n"],
5
- "mappings": "AAAA,OAAO,uBAAuB;AAC9B,SAAQ,kBAAkB,eAAe,eAAc;AACvD;AAAA,EAAQ;AAAA,EAA+B;AAAA,EAAqC;AAAA,EAA4B;AAAA,EAA6B;AAAA,EACnI;AAAA,EACA;AAAA,EAA8B;AAAA,EAAuC;AAAA,OAAqC;AAI5G,MAAM,QAAQ,kBAAkB,4EAA4E;AAO5G,0BAA2B;AACzB,SAAO;AAAA,IACL,aAAa;AAAA,IACb;AAAA,IAAU;AAAA,EACZ;AAEA,WAAS,IAAI,QAAQ;AACnB,YAAQ,8BAA8B,KAAK;AAC3C,UAAM,MAAM,EAAC,SAAS,CAAC,GAAG,KAAK,CAAC,GAAG,OAAO,KAAI;AAK9C,uDAAmD,MAAM;AACzD,0CAAsC,MAAM;AAG5C,WAAO;AAAA,EACT;AAEA,WAAS,SAAS,QAAQ;AACxB,UAAM,MAAM,EAAC,SAAS,CAAC,EAAC;AAExB,YAAQ,4CAA4C,KAAK;AACzD,QAAI,iCAAiC,MAAM,EAAE,QAAQ;AACnD,UAAI,QAAQ,KAAK,4CAA4C;AAAA,IAC/D;AAGA,YAAQ,uDAAuD,KAAK;AACpE,UAAM,MAAM,+CAA+C,MAAM;AACjE,UAAM,OAAO,kDAAkD,MAAM;AAGrE,QAAI,QAAQ,MAAM;AAChB,UAAI,QAAQ,KAAK,8CAA8C,IAAI,cAAc,GAAG,EAAE;AAAA,IACxF;AACA,QAAI,QAAQ,IAAI,QAAQ,SAAS;AACjC,WAAO;AAAA,EACT;AACF;AAEA,SAAS,yCAAyC,kBAAkB,QAAQ;AAC1E,SAAO,OAAO,OAAO,OAAK,EAAE,QAAQ,SAAS,EAAE,UAAU,KAAK,QAAM,6BAA6B,EAAE,MAAM,gBAAgB,CAAC;AAC5H;AAEA,SAAS,kCAAkC,UAAU,YAAY;AAC/D,QAAM,mBAAmB,6BAA6B,QAAQ;AAC9D,MAAI,CAAC,oBAAoB,qBAAqB,MAAM;AAClD,WAAO;AAAA,EACT;AACA,QAAM,iBAAiB,yCAAyC,kBAAkB,UAAU;AAE5F,SAAO,eAAe,SAAS;AACjC;AAEA,SAAS,+BAA+B,OAAO,YAAY;AACzD,MAAI,CAAC,MAAM,aAAa,MAAM,QAAQ,OAAO;AAC3C,WAAO;AAAA,EACT;AAGA,SAAO,MAAM,UAAU,KAAK,cAAY,kCAAkC,UAAU,UAAU,CAAC;AAEjG;AAEA,SAAS,iCAAiC,QAAQ;AAChD,QAAM,4BAA4B,OAAO,OAAO,OAAO,WAAS,MAAM,QAAQ,SAAS,iBAAiB,OAAO,GAAG,CAAC;AAEnH,SAAO,0BAA0B,OAAO,WAAS,+BAA+B,OAAO,yBAAyB,CAAC;AACnH;AAEA,SAAS,mDAAmD,QAAQ;AAClE,QAAM,+BAA+B,iCAAiC,MAAM;AAC5E,MAAI,6BAA6B,SAAS,GAAG;AAC3C;AAAA,EACF;AACA,UAAQ,sCAAsC,6BAA6B,MAAM,WAAW,KAAK;AACjG,+BAA6B,QAAQ,WAAS,6BAA6B,KAAK,CAAC;AAEjF,WAAS,iBAAiB,OAAO;AAC/B,QAAI,MAAM,QAAQ,OAAO;AACvB,aAAO;AAAA,IACT;AACA,UAAM,mBAAmB,oCAAoC,KAAK;AAClE,YAAQ,2BAA2B,gBAAgB,QAAQ,cAAc,KAAK,CAAC,EAAE;AACjF,QAAI,qBAAqB,QAAW;AAClC,aAAO;AAAA,IACT;AACA,UAAM,oBAAoB,yCAAyC,kBAAkB,4BAA4B;AACjH,QAAI,kBAAkB,SAAS,GAAG;AAChC,cAAQ,mBAAmB,kBAAkB,MAAM,mCAAmC,gBAAgB,uBAAuB;AAC7H,aAAO;AAAA,IACT;AACA,YAAQ,cAAc,kBAAkB,MAAM,mCAAmC,gBAAgB,oBAAoB;AACrH,UAAM,iCAAiC,kBAAkB,OAAO,eAAa,MAAM,QAAQ,UAAU,GAAG;AAExG,QAAI,+BAA+B,WAAW,GAAG;AAC/C,cAAQ,6BAA6B,MAAM,GAAG,OAAO,+BAA+B,MAAM,yBAAyB;AACnH,aAAO;AAAA,IACT;AAEA,WAAO;AAAA,EACT;AAEA,WAAS,6BAA6B,OAAO;AAC3C,QAAI,CAAC,iBAAiB,KAAK,GAAG;AAC5B;AAAA,IACF;AAEA,UAAM,mBAAmB,oCAAoC,KAAK;AAClE,UAAM,2BAA2B,+CAA+C,MAAM,IAAI;AAC1F,UAAM,sBAAsB,4BAA4B,wBAAwB;AAChF,UAAM,eAAe,8BAA8B,OAAO,OAAO,MAAM;AAEvE,YAAQ,aAAa,cAAc,KAAK,CAAC,gCAAgC,aAAa,MAAM,mBAAmB,mBAAmB,KAAK,KAAK;AAE5I,+BAA2B,OAAO,qBAAqB,gBAAgB;AACvE,iBAAa,QAAQ,iBAAe,2BAA2B,aAAa,qBAAqB,gBAAgB,CAAC;AAAA,EAEpH;AAGF;AACA,SAAS,kDAAkD,QAAQ;AAEjE,MAAI,aAAa,CAAC;AAClB,SAAO,OAAO,QAAQ,WAAS,gBAAgB,KAAK,CAAC;AAErD,WAAS,gBAAgB,OAAO;AAC9B,QAAI,CAAC,MAAM,WAAW;AACpB;AAAA,IACF;AACA,UAAM,UAAU,QAAQ,cAAY,mBAAmB,QAAQ,CAAC;AAAA,EAClE;AAEA,WAAS,mBAAmB,UAAU;AACpC,QAAI,CAAC,iBAAiB,QAAQ,GAAG;AAC/B;AAAA,IACF;AACA,UAAM,IAAI,sCAAsC,QAAQ;AACxD,QAAI,MAAM,GAAG;AACX;AAAA,IACF;AACA,eAAW,CAAC,IAAI;AAAA,EAClB;AACA,MAAI,IAAI;AACR,aAAW,QAAQ,UAAQ,KAAI,IAAI;AACnC,SAAO;AACT;AAEO,gBAAS,sCAAsC,QAAQ;AAE5D,MAAI,aAAa;AACjB,MAAI,gBAAgB,CAAC;AAErB,SAAO,OAAO,QAAQ,WAAS,oBAAoB,KAAK,CAAC;AAEzD,WAAS,oBAAoB,OAAO;AAClC,YAAQ,uBAAuB,cAAc,KAAK,CAAC,WAAW,UAAU,IAAI,KAAK;AACjF,QAAI,CAAC,MAAM,WAAW;AACpB;AAAA,IACF;AACA,UAAM,UAAU,QAAQ,cAAY,eAAe,QAAQ,CAAC;AAAA,EAC9D;AAEA,WAAS,eAAe,UAAU;AAChC,QAAI,CAAC,iBAAiB,QAAQ,GAAG;AAC/B;AAAA,IACF;AACA,UAAM,YAAY,6BAA6B,QAAQ;AACvD,QAAI,cAAc,UAAa,cAAc,MAAM;AACjD;AAAA,IACF;AAEA,UAAM,WAAW,uBAAuB,SAAS;AAEjD,mCAA+B,UAAU,QAAQ;AAAA,EACnD;AAEA,WAAS,uBAAuB,WAAW;AACzC,QAAG,aAAa,eAAe;AAC7B,aAAO,cAAc,SAAS;AAAA,IAChC;AACA,UAAM,WAAW,4BAA4B,UAAU;AACvD,kBAAc,SAAS,IAAI;AAC3B;AACA,WAAO;AAAA,EACT;AAIF;",
4
+ "sourcesContent": ["import createDebugLogger from 'debug';\nimport {fieldHasSubfield, fieldToString, nvdebug} from './utils.js';\nimport {fieldGetOccurrenceNumberPairs, fieldGetUnambiguousOccurrenceNumber, fieldResetOccurrenceNumber, intToOccurrenceNumberString, isValidSubfield6,\n recordGetMaxSubfield6OccurrenceNumberAsInteger,\n subfield6GetOccurrenceNumber, subfield6GetOccurrenceNumberAsInteger, subfield6ResetOccurrenceNumber} from './subfield6Utils.js';\n\n// Relocated from melinda-marc-record-merge-reducers (and renamed)\n\nconst debug = createDebugLogger('@natlibfi/marc-record-validators-melinda:reindexSubfield6OccurrenceNumbers');\n//const debugData = debug.extend('data');\nconst debugDev = debug.extend('dev');\n\n\n// NB! This validator/fixer has two functionalities:\n// 1) normal reindexing of occurrence numbers\n// 2) disambiguation (when possible) of unambiguous occurrence numbers\n\nexport default function () {\n return {\n description: 'Reindex occurrence numbers in $6 subfield so that they start from 01 and end in NN',\n validate, fix\n };\n\n function fix(record) {\n nvdebug('Fix SF6 occurrence numbers', debugDev);\n const res = {message: [], fix: [], valid: true};\n //message.fix = [];\n\n // This can not really fail...\n\n recordDisambiguateSharedSubfield6OccurrenceNumbers(record);\n recordResetSubfield6OccurrenceNumbers(record);\n\n // message.valid = !(message.message.length >= 1);\n return res;\n }\n\n function validate(record) {\n const res = {message: []};\n\n nvdebug('Validate SF6 occurrence number multiuses', debugDev);\n if (recordGetSharedOccurrenceNumbers(record).length) {\n res.message.push(`Multi-use of occurrence number(s) detected`);\n }\n\n // Check max, and check number of different indexes\n nvdebug('Validate SF6 occurrence number (max vs n instances)', debugDev);\n const max = recordGetMaxSubfield6OccurrenceNumberAsInteger(record);\n const size = recordGetNumberOfUniqueSubfield6OccurrenceNumbers(record);\n\n\n if (max !== size) {\n res.message.push(`Gaps detected in occurrence numbers: found ${size}, seen max ${max}`);\n }\n res.valid = res.message.length < 1;\n return res;\n }\n}\n\nfunction getPotentialSharedOccurrenceNumberFields(occurrenceNumber, fields) {\n return fields.filter(f => f.tag !== '880' && f.subfields.some(sf => subfield6GetOccurrenceNumber(sf) === occurrenceNumber));\n}\n\nfunction subfieldHasSharedOccurrenceNumber(subfield, candFields) {\n const occurrenceNumber = subfield6GetOccurrenceNumber(subfield);\n if (!occurrenceNumber || occurrenceNumber === '00') {\n return false;\n }\n const relevantFields = getPotentialSharedOccurrenceNumberFields(occurrenceNumber, candFields);\n // record.fields.filter(f => f.tag !== '880' && fieldHasOccurrenceNumber(f, occurrenceNumber));\n return relevantFields.length > 1;\n}\n\nfunction fieldHasSharedOccurrenceNumber(field, candFields) {\n if (!field.subfields || field.tag === '880') { // Should not happen\n return false;\n }\n\n // What if there are multiple $6s in a given field? Should not be, but...\n return field.subfields.some(subfield => subfieldHasSharedOccurrenceNumber(subfield, candFields));\n\n}\n\nfunction recordGetSharedOccurrenceNumbers(record) {\n const fieldsContainingSubfield6 = record.fields.filter(field => field.tag !== '880' && fieldHasSubfield(field, '6'));\n // fieldsContainingSubfield6.some(field => fieldHasSharedOccurrenceNumber(field, fieldsContainingSubfield6)))\n return fieldsContainingSubfield6.filter(field => fieldHasSharedOccurrenceNumber(field, fieldsContainingSubfield6));\n}\n\nfunction recordDisambiguateSharedSubfield6OccurrenceNumbers(record) {\n const sharedOccurrenceNumberFields = recordGetSharedOccurrenceNumbers(record);\n if (sharedOccurrenceNumberFields.length < 2) {\n return;\n }\n nvdebug(`Disambiguate occurrence numbers (N=${sharedOccurrenceNumberFields.length}) in...`, debugDev);\n sharedOccurrenceNumberFields.forEach(field => disambiguateOccurrenceNumber(field));\n\n function disambiguateable(field) {\n if (field.tag === '880') { // Not needed, already filtered...\n return false;\n }\n const occurrenceNumber = fieldGetUnambiguousOccurrenceNumber(field);\n nvdebug(` Trying to disambiguate ${occurrenceNumber} in '${fieldToString(field)}`, debugDev);\n if (occurrenceNumber === undefined) {\n return false;\n }\n const allRelevantFields = getPotentialSharedOccurrenceNumberFields(occurrenceNumber, sharedOccurrenceNumberFields);\n if (allRelevantFields.length < 2) {\n nvdebug(` Currently only ${allRelevantFields.length} field(s) use occurrence number ${occurrenceNumber}. No action required.`, debugDev);\n return false;\n }\n nvdebug(` Currently ${allRelevantFields.length} field(s) use occurrence number ${occurrenceNumber}. ACTION REQUIRED!`, debugDev);\n const relevantFieldsWithCurrFieldTag = allRelevantFields.filter(candField => field.tag === candField.tag);\n\n if (relevantFieldsWithCurrFieldTag.length !== 1) {\n nvdebug(` Number of them using tag ${field.tag} is ${relevantFieldsWithCurrFieldTag.length}. Can not disambiguate!`, debugDev);\n return false;\n }\n\n return true;\n }\n\n function disambiguateOccurrenceNumber(field) {\n if (!disambiguateable(field)) {\n return;\n }\n // Reset field:\n const occurrenceNumber = fieldGetUnambiguousOccurrenceNumber(field);\n const newOccurrenceNumberAsInt = recordGetMaxSubfield6OccurrenceNumberAsInteger(record) + 1;\n const newOccurrenceNumber = intToOccurrenceNumberString(newOccurrenceNumberAsInt);\n const pairedFields = fieldGetOccurrenceNumberPairs(field, record.fields);\n\n nvdebug(` Reindex '${fieldToString(field)}' occurrence number and it's ${pairedFields.length} pair(s) using '${newOccurrenceNumber}'`, debugDev);\n\n fieldResetOccurrenceNumber(field, newOccurrenceNumber, occurrenceNumber);\n pairedFields.forEach(pairedField => fieldResetOccurrenceNumber(pairedField, newOccurrenceNumber, occurrenceNumber));\n\n }\n\n\n}\nfunction recordGetNumberOfUniqueSubfield6OccurrenceNumbers(record) {\n // Calculates the number of used different occurrence numbers\n let indexArray = [];\n record.fields.forEach(field => gatherFieldData(field));\n\n function gatherFieldData(field) {\n if (!field.subfields) {\n return;\n }\n field.subfields.forEach(subfield => gatherSubfieldData(subfield));\n }\n\n function gatherSubfieldData(subfield) {\n if (!isValidSubfield6(subfield)) {\n return;\n }\n const i = subfield6GetOccurrenceNumberAsInteger(subfield);\n if (i === 0) {\n return\n }\n indexArray[i] = 1;\n }\n let n = 0;\n indexArray.forEach(elem => n+= elem);\n return n;\n}\n\nexport function recordResetSubfield6OccurrenceNumbers(record) { // Remove gaps\n /* eslint-disable */\n let currentInt = 1;\n let oldtoNewCache = {};\n\n record.fields.forEach(field => fieldResetSubfield6(field));\n\n function fieldResetSubfield6(field) {\n nvdebug(`fieldResetSubfield6(${fieldToString(field)}), CURR:${currentInt}`, debugDev);\n if (!field.subfields) {\n return;\n }\n field.subfields.forEach(subfield => subfieldReset6(subfield));\n }\n\n function subfieldReset6(subfield) {\n if (!isValidSubfield6(subfield)) {\n return;\n }\n const currIndex = subfield6GetOccurrenceNumber(subfield);\n if (currIndex === undefined || currIndex === '00') {\n return;\n }\n\n const newIndex = mapCurrIndexToNewIndex(currIndex);\n //nvdebug(`subfieldReset6(${subfieldToString(subfield)}): ${newIndex}`, debugDev);\n subfield6ResetOccurrenceNumber(subfield, newIndex);\n }\n\n function mapCurrIndexToNewIndex(currIndex) {\n if(currIndex in oldtoNewCache) {\n return oldtoNewCache[currIndex];\n }\n const newIndex = intToOccurrenceNumberString(currentInt);\n oldtoNewCache[currIndex] = newIndex;\n currentInt++;\n return newIndex;\n }\n\n /* eslint-enable */\n\n}\n"],
5
+ "mappings": "AAAA,OAAO,uBAAuB;AAC9B,SAAQ,kBAAkB,eAAe,eAAc;AACvD;AAAA,EAAQ;AAAA,EAA+B;AAAA,EAAqC;AAAA,EAA4B;AAAA,EAA6B;AAAA,EACnI;AAAA,EACA;AAAA,EAA8B;AAAA,EAAuC;AAAA,OAAqC;AAI5G,MAAM,QAAQ,kBAAkB,4EAA4E;AAE5G,MAAM,WAAW,MAAM,OAAO,KAAK;AAOnC,0BAA2B;AACzB,SAAO;AAAA,IACL,aAAa;AAAA,IACb;AAAA,IAAU;AAAA,EACZ;AAEA,WAAS,IAAI,QAAQ;AACnB,YAAQ,8BAA8B,QAAQ;AAC9C,UAAM,MAAM,EAAC,SAAS,CAAC,GAAG,KAAK,CAAC,GAAG,OAAO,KAAI;AAK9C,uDAAmD,MAAM;AACzD,0CAAsC,MAAM;AAG5C,WAAO;AAAA,EACT;AAEA,WAAS,SAAS,QAAQ;AACxB,UAAM,MAAM,EAAC,SAAS,CAAC,EAAC;AAExB,YAAQ,4CAA4C,QAAQ;AAC5D,QAAI,iCAAiC,MAAM,EAAE,QAAQ;AACnD,UAAI,QAAQ,KAAK,4CAA4C;AAAA,IAC/D;AAGA,YAAQ,uDAAuD,QAAQ;AACvE,UAAM,MAAM,+CAA+C,MAAM;AACjE,UAAM,OAAO,kDAAkD,MAAM;AAGrE,QAAI,QAAQ,MAAM;AAChB,UAAI,QAAQ,KAAK,8CAA8C,IAAI,cAAc,GAAG,EAAE;AAAA,IACxF;AACA,QAAI,QAAQ,IAAI,QAAQ,SAAS;AACjC,WAAO;AAAA,EACT;AACF;AAEA,SAAS,yCAAyC,kBAAkB,QAAQ;AAC1E,SAAO,OAAO,OAAO,OAAK,EAAE,QAAQ,SAAS,EAAE,UAAU,KAAK,QAAM,6BAA6B,EAAE,MAAM,gBAAgB,CAAC;AAC5H;AAEA,SAAS,kCAAkC,UAAU,YAAY;AAC/D,QAAM,mBAAmB,6BAA6B,QAAQ;AAC9D,MAAI,CAAC,oBAAoB,qBAAqB,MAAM;AAClD,WAAO;AAAA,EACT;AACA,QAAM,iBAAiB,yCAAyC,kBAAkB,UAAU;AAE5F,SAAO,eAAe,SAAS;AACjC;AAEA,SAAS,+BAA+B,OAAO,YAAY;AACzD,MAAI,CAAC,MAAM,aAAa,MAAM,QAAQ,OAAO;AAC3C,WAAO;AAAA,EACT;AAGA,SAAO,MAAM,UAAU,KAAK,cAAY,kCAAkC,UAAU,UAAU,CAAC;AAEjG;AAEA,SAAS,iCAAiC,QAAQ;AAChD,QAAM,4BAA4B,OAAO,OAAO,OAAO,WAAS,MAAM,QAAQ,SAAS,iBAAiB,OAAO,GAAG,CAAC;AAEnH,SAAO,0BAA0B,OAAO,WAAS,+BAA+B,OAAO,yBAAyB,CAAC;AACnH;AAEA,SAAS,mDAAmD,QAAQ;AAClE,QAAM,+BAA+B,iCAAiC,MAAM;AAC5E,MAAI,6BAA6B,SAAS,GAAG;AAC3C;AAAA,EACF;AACA,UAAQ,sCAAsC,6BAA6B,MAAM,WAAW,QAAQ;AACpG,+BAA6B,QAAQ,WAAS,6BAA6B,KAAK,CAAC;AAEjF,WAAS,iBAAiB,OAAO;AAC/B,QAAI,MAAM,QAAQ,OAAO;AACvB,aAAO;AAAA,IACT;AACA,UAAM,mBAAmB,oCAAoC,KAAK;AAClE,YAAQ,2BAA2B,gBAAgB,QAAQ,cAAc,KAAK,CAAC,IAAI,QAAQ;AAC3F,QAAI,qBAAqB,QAAW;AAClC,aAAO;AAAA,IACT;AACA,UAAM,oBAAoB,yCAAyC,kBAAkB,4BAA4B;AACjH,QAAI,kBAAkB,SAAS,GAAG;AAChC,cAAQ,mBAAmB,kBAAkB,MAAM,mCAAmC,gBAAgB,yBAAyB,QAAQ;AACvI,aAAO;AAAA,IACT;AACA,YAAQ,cAAc,kBAAkB,MAAM,mCAAmC,gBAAgB,sBAAsB,QAAQ;AAC/H,UAAM,iCAAiC,kBAAkB,OAAO,eAAa,MAAM,QAAQ,UAAU,GAAG;AAExG,QAAI,+BAA+B,WAAW,GAAG;AAC/C,cAAQ,6BAA6B,MAAM,GAAG,OAAO,+BAA+B,MAAM,2BAA2B,QAAQ;AAC7H,aAAO;AAAA,IACT;AAEA,WAAO;AAAA,EACT;AAEA,WAAS,6BAA6B,OAAO;AAC3C,QAAI,CAAC,iBAAiB,KAAK,GAAG;AAC5B;AAAA,IACF;AAEA,UAAM,mBAAmB,oCAAoC,KAAK;AAClE,UAAM,2BAA2B,+CAA+C,MAAM,IAAI;AAC1F,UAAM,sBAAsB,4BAA4B,wBAAwB;AAChF,UAAM,eAAe,8BAA8B,OAAO,OAAO,MAAM;AAEvE,YAAQ,aAAa,cAAc,KAAK,CAAC,gCAAgC,aAAa,MAAM,mBAAmB,mBAAmB,KAAK,QAAQ;AAE/I,+BAA2B,OAAO,qBAAqB,gBAAgB;AACvE,iBAAa,QAAQ,iBAAe,2BAA2B,aAAa,qBAAqB,gBAAgB,CAAC;AAAA,EAEpH;AAGF;AACA,SAAS,kDAAkD,QAAQ;AAEjE,MAAI,aAAa,CAAC;AAClB,SAAO,OAAO,QAAQ,WAAS,gBAAgB,KAAK,CAAC;AAErD,WAAS,gBAAgB,OAAO;AAC9B,QAAI,CAAC,MAAM,WAAW;AACpB;AAAA,IACF;AACA,UAAM,UAAU,QAAQ,cAAY,mBAAmB,QAAQ,CAAC;AAAA,EAClE;AAEA,WAAS,mBAAmB,UAAU;AACpC,QAAI,CAAC,iBAAiB,QAAQ,GAAG;AAC/B;AAAA,IACF;AACA,UAAM,IAAI,sCAAsC,QAAQ;AACxD,QAAI,MAAM,GAAG;AACX;AAAA,IACF;AACA,eAAW,CAAC,IAAI;AAAA,EAClB;AACA,MAAI,IAAI;AACR,aAAW,QAAQ,UAAQ,KAAI,IAAI;AACnC,SAAO;AACT;AAEO,gBAAS,sCAAsC,QAAQ;AAE5D,MAAI,aAAa;AACjB,MAAI,gBAAgB,CAAC;AAErB,SAAO,OAAO,QAAQ,WAAS,oBAAoB,KAAK,CAAC;AAEzD,WAAS,oBAAoB,OAAO;AAClC,YAAQ,uBAAuB,cAAc,KAAK,CAAC,WAAW,UAAU,IAAI,QAAQ;AACpF,QAAI,CAAC,MAAM,WAAW;AACpB;AAAA,IACF;AACA,UAAM,UAAU,QAAQ,cAAY,eAAe,QAAQ,CAAC;AAAA,EAC9D;AAEA,WAAS,eAAe,UAAU;AAChC,QAAI,CAAC,iBAAiB,QAAQ,GAAG;AAC/B;AAAA,IACF;AACA,UAAM,YAAY,6BAA6B,QAAQ;AACvD,QAAI,cAAc,UAAa,cAAc,MAAM;AACjD;AAAA,IACF;AAEA,UAAM,WAAW,uBAAuB,SAAS;AAEjD,mCAA+B,UAAU,QAAQ;AAAA,EACnD;AAEA,WAAS,uBAAuB,WAAW;AACzC,QAAG,aAAa,eAAe;AAC7B,aAAO,cAAc,SAAS;AAAA,IAChC;AACA,UAAM,WAAW,4BAA4B,UAAU;AACvD,kBAAc,SAAS,IAAI;AAC3B;AACA,WAAO;AAAA,EACT;AAIF;",
6
6
  "names": []
7
7
  }
@@ -6,6 +6,7 @@ const LINK_ROOT = 4;
6
6
  const LINKED_AND_PROCESSED = 2;
7
7
  const LINKED_NOT_PROCESSED = 1;
8
8
  const debug = createDebugLogger("@natlibfi/marc-record-validators-melinda:removeDuplicateDataFields");
9
+ const debugDev = debug.extend("dev");
9
10
  export default function() {
10
11
  return {
11
12
  description: "Remove duplicate data fields. Certain exceptions apply, mainly too complited chained fields",
@@ -13,13 +14,13 @@ export default function() {
13
14
  fix
14
15
  };
15
16
  function fix(record) {
16
- nvdebug("Remove duplicate data fields");
17
+ nvdebug("Remove duplicate data fields", debugDev);
17
18
  const res = { message: [], fix: [], valid: true };
18
19
  removeDuplicateDatafields(record, true);
19
20
  return res;
20
21
  }
21
22
  function validate(record) {
22
- nvdebug("Validate record: duplicate data fields cause (t)error", debug);
23
+ nvdebug("Validate record: duplicate data fields cause (t)error", debugDev);
23
24
  const duplicates = removeDuplicateDatafields(record, false);
24
25
  const res = { message: duplicates };
25
26
  res.valid = res.message.length < 1;