@natlibfi/marc-record-validators-melinda 12.0.0 → 12.0.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/disambiguateSeriesStatements.test.js +1 -7
- package/dist/disambiguateSeriesStatements.test.js.map +2 -2
- package/dist/drop-terms.js +122 -0
- package/dist/drop-terms.js.map +7 -0
- package/dist/drop-terms.test.js +56 -0
- package/dist/drop-terms.test.js.map +7 -0
- package/dist/field-008-18-34-character-groups.test.js +1 -7
- package/dist/field-008-18-34-character-groups.test.js.map +2 -2
- package/dist/field-505-separators.test.js +1 -7
- package/dist/field-505-separators.test.js.map +2 -2
- package/dist/field-521-fix.test.js +1 -7
- package/dist/field-521-fix.test.js.map +2 -2
- package/dist/fix-33X.test.js +1 -1
- package/dist/fix-33X.test.js.map +1 -1
- package/dist/fix-country-codes.test.js +1 -7
- package/dist/fix-country-codes.test.js.map +2 -2
- package/dist/fix-sami-041.js +87 -0
- package/dist/fix-sami-041.js.map +7 -0
- package/dist/fix-sami-041.test.js +40 -0
- package/dist/fix-sami-041.test.js.map +7 -0
- package/dist/fixRelatorTerms.test.js +1 -7
- package/dist/fixRelatorTerms.test.js.map +2 -2
- package/dist/index.js +7 -1
- package/dist/index.js.map +2 -2
- package/dist/indicator-fixes.js +10 -0
- package/dist/indicator-fixes.js.map +2 -2
- package/dist/indicator-fixes.test.js +1 -7
- package/dist/indicator-fixes.test.js.map +2 -2
- package/dist/merge-fields.test.js +2 -7
- package/dist/merge-fields.test.js.map +2 -2
- package/dist/mergeField500Lisapainokset.test.js +1 -7
- package/dist/mergeField500Lisapainokset.test.js.map +2 -2
- package/dist/mergeRelatorTermFields.test.js +1 -7
- package/dist/mergeRelatorTermFields.test.js.map +2 -2
- package/dist/multiple-subfield-0.test.js +1 -7
- package/dist/multiple-subfield-0.test.js.map +2 -2
- package/dist/normalize-dashes.test.js +1 -7
- package/dist/normalize-dashes.test.js.map +2 -2
- package/dist/normalize-identifiers.test.js +1 -7
- package/dist/normalize-identifiers.test.js.map +2 -2
- package/dist/normalize-qualifying-information.test.js +1 -7
- package/dist/normalize-qualifying-information.test.js.map +2 -2
- package/dist/normalize-utf8-diacritics.test.js +1 -7
- package/dist/normalize-utf8-diacritics.test.js.map +2 -2
- package/dist/punctuation.test.js +1 -7
- package/dist/punctuation.test.js.map +2 -2
- package/dist/punctuation2.test.js +1 -7
- package/dist/punctuation2.test.js.map +2 -2
- package/dist/reindexSubfield6OccurenceNumbers.test.js +1 -7
- package/dist/reindexSubfield6OccurenceNumbers.test.js.map +2 -2
- package/dist/remove-041-zxx.js +56 -0
- package/dist/remove-041-zxx.js.map +7 -0
- package/dist/remove-041-zxx.test.js +40 -0
- package/dist/remove-041-zxx.test.js.map +7 -0
- package/dist/removeDuplicateDataFields.test.js +1 -7
- package/dist/removeDuplicateDataFields.test.js.map +2 -2
- package/dist/removeInferiorDataFields.js +1 -9
- package/dist/removeInferiorDataFields.js.map +2 -2
- package/dist/removeInferiorDataFields.test.js +1 -7
- package/dist/removeInferiorDataFields.test.js.map +2 -2
- package/dist/resolveOrphanedSubfield6s.js +1 -3
- package/dist/resolveOrphanedSubfield6s.js.map +2 -2
- package/dist/resolveOrphanedSubfield6s.test.js +1 -7
- package/dist/resolveOrphanedSubfield6s.test.js.map +2 -2
- package/dist/sanitize-vocabulary-source-codes.test.js +1 -7
- package/dist/sanitize-vocabulary-source-codes.test.js.map +2 -2
- package/dist/sortFields.js +16 -1
- package/dist/sortFields.js.map +2 -2
- package/dist/sortFields.test.js +1 -7
- package/dist/sortFields.test.js.map +2 -2
- package/dist/sortRelatorTerms.test.js +1 -7
- package/dist/sortRelatorTerms.test.js.map +2 -2
- package/dist/sortSubfields.js +3 -1
- package/dist/sortSubfields.js.map +2 -2
- package/dist/sortSubfields.test.js +1 -7
- package/dist/sortSubfields.test.js.map +2 -2
- package/dist/stripPunctuation.js +7 -4
- package/dist/stripPunctuation.js.map +2 -2
- package/dist/stripPunctuation.test.js +1 -7
- package/dist/stripPunctuation.test.js.map +2 -2
- package/dist/subfield6Utils.js +1 -13
- package/dist/subfield6Utils.js.map +2 -2
- package/dist/subfieldValueNormalizations.test.js +1 -7
- package/dist/subfieldValueNormalizations.test.js.map +2 -2
- package/dist/sync-007-and-300.test.js +1 -7
- package/dist/sync-007-and-300.test.js.map +2 -2
- package/dist/sync-language.js +103 -0
- package/dist/sync-language.js.map +7 -0
- package/dist/sync-language.test.js +40 -0
- package/dist/sync-language.test.js.map +7 -0
- package/dist/translate-terms.js +121 -85
- package/dist/translate-terms.js.map +3 -3
- package/dist/translate-terms.test.js +5 -8
- package/dist/translate-terms.test.js.map +2 -2
- package/dist/update-field-540.test.js +1 -7
- package/dist/update-field-540.test.js.map +2 -2
- package/dist/urn.test.js +1 -7
- package/dist/urn.test.js.map +2 -2
- package/package.json +8 -6
- package/src/disambiguateSeriesStatements.test.js +3 -8
- package/src/drop-terms.js +162 -0
- package/src/drop-terms.test.js +81 -0
- package/src/field-008-18-34-character-groups.test.js +3 -8
- package/src/field-505-separators.test.js +3 -8
- package/src/field-521-fix.test.js +3 -8
- package/src/fix-33X.test.js +1 -1
- package/src/fix-country-codes.test.js +3 -8
- package/src/fix-sami-041.js +113 -0
- package/src/fix-sami-041.test.js +52 -0
- package/src/fixRelatorTerms.test.js +3 -8
- package/src/index.js +8 -1
- package/src/indicator-fixes.js +12 -0
- package/src/indicator-fixes.test.js +3 -8
- package/src/merge-fields.test.js +3 -8
- package/src/mergeField500Lisapainokset.test.js +3 -8
- package/src/mergeRelatorTermFields.test.js +3 -8
- package/src/multiple-subfield-0.test.js +3 -8
- package/src/normalize-dashes.test.js +3 -8
- package/src/normalize-identifiers.test.js +3 -8
- package/src/normalize-qualifying-information.test.js +3 -8
- package/src/normalize-utf8-diacritics.test.js +3 -8
- package/src/punctuation.test.js +3 -8
- package/src/punctuation2.test.js +3 -8
- package/src/reindexSubfield6OccurenceNumbers.test.js +3 -8
- package/src/remove-041-zxx.js +85 -0
- package/src/remove-041-zxx.test.js +52 -0
- package/src/removeDuplicateDataFields.test.js +3 -8
- package/src/removeInferiorDataFields.js +7 -7
- package/src/removeInferiorDataFields.test.js +3 -8
- package/src/resolveOrphanedSubfield6s.js +3 -3
- package/src/resolveOrphanedSubfield6s.test.js +3 -8
- package/src/sanitize-vocabulary-source-codes.test.js +3 -8
- package/src/sortFields.js +20 -1
- package/src/sortFields.test.js +3 -8
- package/src/sortRelatorTerms.test.js +3 -8
- package/src/sortSubfields.js +3 -1
- package/src/sortSubfields.test.js +3 -8
- package/src/stripPunctuation.js +9 -6
- package/src/stripPunctuation.test.js +3 -8
- package/src/subfield6Utils.js +13 -13
- package/src/subfieldValueNormalizations.test.js +3 -8
- package/src/sync-007-and-300.test.js +3 -8
- package/src/sync-language.js +148 -0
- package/src/sync-language.test.js +52 -0
- package/src/translate-terms.js +158 -103
- package/src/translate-terms.test.js +12 -16
- package/src/update-field-540.test.js +3 -8
- package/src/urn.test.js +3 -8
- package/test-fixtures/drop-terms/01/expectedResult.json +31 -0
- package/test-fixtures/drop-terms/01/metadata.json +6 -0
- package/test-fixtures/drop-terms/01/record.json +35 -0
- package/test-fixtures/drop-terms/02/expectedResult.json +7 -0
- package/test-fixtures/drop-terms/02/metadata.json +6 -0
- package/test-fixtures/drop-terms/02/record.json +40 -0
- package/test-fixtures/drop-terms/03/expectedResult.json +6 -0
- package/test-fixtures/drop-terms/03/metadata.json +18 -0
- package/test-fixtures/drop-terms/03/record.json +39 -0
- package/test-fixtures/drop-terms/04/expectedResult.json +6 -0
- package/test-fixtures/drop-terms/04/metadata.json +19 -0
- package/test-fixtures/drop-terms/04/record.json +24 -0
- package/test-fixtures/fix-language-codes/02/metadata.json +1 -1
- package/test-fixtures/fix-sami-041/01/expectedResult.json +6 -0
- package/test-fixtures/fix-sami-041/01/metadata.json +4 -0
- package/test-fixtures/fix-sami-041/01/record.json +13 -0
- package/test-fixtures/fix-sami-041/02/expectedResult.json +10 -0
- package/test-fixtures/fix-sami-041/02/metadata.json +4 -0
- package/test-fixtures/fix-sami-041/02/record.json +8 -0
- package/test-fixtures/fix-sami-041/03/expectedResult.json +5 -0
- package/test-fixtures/fix-sami-041/03/metadata.json +5 -0
- package/test-fixtures/fix-sami-041/03/record.json +8 -0
- package/test-fixtures/fix-sami-041/04/expectedResult.json +7 -0
- package/test-fixtures/fix-sami-041/04/metadata.json +4 -0
- package/test-fixtures/fix-sami-041/04/record.json +10 -0
- package/test-fixtures/fix-sami-041/05/expectedResult.json +10 -0
- package/test-fixtures/fix-sami-041/05/metadata.json +6 -0
- package/test-fixtures/fix-sami-041/05/record.json +8 -0
- package/test-fixtures/indicator-fixes/11/expectedResult.json +10 -0
- package/test-fixtures/indicator-fixes/11/metadata.json +4 -0
- package/test-fixtures/indicator-fixes/11/record.json +10 -0
- package/test-fixtures/merge-fields/f05/metadata.json +1 -1
- package/test-fixtures/remove-041-zxx/01/expectedResult.json +5 -0
- package/test-fixtures/remove-041-zxx/01/metadata.json +5 -0
- package/test-fixtures/remove-041-zxx/01/record.json +10 -0
- package/test-fixtures/remove-041-zxx/02/expectedResult.json +7 -0
- package/test-fixtures/remove-041-zxx/02/metadata.json +5 -0
- package/test-fixtures/remove-041-zxx/02/record.json +9 -0
- package/test-fixtures/remove-041-zxx/11/expectedResult.json +10 -0
- package/test-fixtures/remove-041-zxx/11/metadata.json +5 -0
- package/test-fixtures/remove-041-zxx/11/record.json +9 -0
- package/test-fixtures/remove-041-zxx/12/expectedResult.json +10 -0
- package/test-fixtures/remove-041-zxx/12/metadata.json +5 -0
- package/test-fixtures/remove-041-zxx/12/record.json +9 -0
- package/test-fixtures/sort-fields/15/input.json +9 -0
- package/test-fixtures/sort-fields/15/metadata.json +5 -0
- package/test-fixtures/sort-fields/15/result.json +10 -0
- package/test-fixtures/sync-language/01/expectedResult.json +5 -0
- package/test-fixtures/sync-language/01/metadata.json +5 -0
- package/test-fixtures/sync-language/01/record.json +7 -0
- package/test-fixtures/sync-language/02/expectedResult.json +6 -0
- package/test-fixtures/sync-language/02/metadata.json +5 -0
- package/test-fixtures/sync-language/02/record.json +10 -0
- package/test-fixtures/sync-language/03/expectedResult.json +6 -0
- package/test-fixtures/sync-language/03/metadata.json +5 -0
- package/test-fixtures/sync-language/03/record.json +6 -0
- package/test-fixtures/sync-language/10/expectedResult.json +10 -0
- package/test-fixtures/sync-language/10/metadata.json +5 -0
- package/test-fixtures/sync-language/10/record.json +8 -0
- package/test-fixtures/sync-language/11/expectedResult.json +10 -0
- package/test-fixtures/sync-language/11/metadata.json +5 -0
- package/test-fixtures/sync-language/11/record.json +7 -0
- package/test-fixtures/sync-language/12/expectedResult.json +9 -0
- package/test-fixtures/sync-language/12/metadata.json +6 -0
- package/test-fixtures/sync-language/12/record.json +6 -0
- package/test-fixtures/sync-language/13/expectedResult.json +10 -0
- package/test-fixtures/sync-language/13/metadata.json +5 -0
- package/test-fixtures/sync-language/13/record.json +8 -0
- package/test-fixtures/sync-language/14/expectedResult.json +9 -0
- package/test-fixtures/sync-language/14/metadata.json +5 -0
- package/test-fixtures/sync-language/14/record.json +7 -0
- package/test-fixtures/sync-language/15/expectedResult.json +9 -0
- package/test-fixtures/sync-language/15/metadata.json +5 -0
- package/test-fixtures/sync-language/15/record.json +7 -0
- package/test-fixtures/translate-terms/05/expectedResult.json +12 -0
- package/test-fixtures/translate-terms/05/metadata.json +7 -0
- package/test-fixtures/translate-terms/05/record.json +11 -0
- package/test-fixtures/translate-terms/06/expectedResult.json +12 -0
- package/test-fixtures/translate-terms/06/metadata.json +7 -0
- package/test-fixtures/translate-terms/06/record.json +11 -0
- package/test-fixtures/translate-terms-data.js +23 -0
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
{
|
|
2
2
|
"version": 3,
|
|
3
3
|
"sources": ["../src/removeInferiorDataFields.js"],
|
|
4
|
-
"sourcesContent": ["import createDebugLogger from 'debug';\nimport {fieldToChain, sameField} from './removeDuplicateDataFields.js';\nimport {fieldGetOccurrenceNumberPairs, fieldHasValidSubfield6, fieldSevenToOneOccurrenceNumber, fieldsToNormalizedString} from './subfield6Utils.js';\nimport {fieldHasSubfield, fieldsToString, fieldToString, nvdebug, uniqArray} from './utils.js';\nimport {fieldHasValidSubfield8} from './subfield8Utils.js';\nimport {encodingLevelIsBetterThanPrepublication, fieldRefersToKoneellisestiTuotettuTietue, getEncodingLevel} from './prepublicationUtils.js';\nimport {cloneAndNormalizeFieldForComparison} from './normalizeFieldForComparison.js';\nimport {fixComposition, precomposeFinnishLetters} from './normalize-utf8-diacritics.js';\n\n// Relocated from melinda-marc-record-merge-reducers (and renamed)\n\n// NB! This validator handles only full fields, and does not support subfield $8 removal.\n// Also, having multiple $8 subfields in same fields is not supported.\n// If this functionality is needed, see removeDuplicateDatafields.js for examples of subfield-only stuff.\nconst debug = createDebugLogger('@natlibfi/marc-record-validators-melinda:removeInferiorDataFields');\n\nexport default function () {\n return {\n description: 'Remove subset data fields. Certain exceptions apply, mainly too complicated for chained fields',\n validate, fix\n };\n\n function fix(record) {\n nvdebug('Fix record: remove inferior (eg. subset) data fields', debug);\n const res = {message: [], fix: [], valid: true};\n removeInferiorDatafields(record, true);\n // This can not really fail...\n return res;\n }\n\n function validate(record) {\n // Check max, and check number of different indexes\n nvdebug('Validate record: remove inferior (eg. subset) data fields', debug);\n\n const duplicates = removeInferiorDatafields(record, false);\n\n const res = {message: duplicates};\n\n res.valid = res.message.length < 1;\n return res;\n }\n}\n\n\nfunction deriveInferiorChains(fields, record) {\n //nvdebug(`======= GOT ${fields.length} FIELDS TO CHAINIFY`);\n const hash = {};\n\n fields.forEach(f => fieldToChainToDeletables(f));\n\n return hash;\n\n //nvdebug(`WP1: GOT ${todoList.length} CHAINS`);\n\n\n // here we map deletableStringObject[str] => field. The idea is to help debugging. We don't actually need the field object...\n //return deriveChainDeletables(todoList);\n\n function fieldToChainToDeletables(field) {\n const chain = fieldToChain(field, record);\n if (chain.length < 2) {\n return;\n }\n const chainAsString = fieldsToNormalizedString(chain, 0, true, true);\n const arr = deriveChainDeletables([chainAsString]);\n //nvdebug(`GOT ${arr.length} DELETABLES FOR ${chainAsString}`);\n arr.forEach(val => {\n if (!(val in hash)) {\n hash[val] = field;\n }\n });\n }\n\n function deriveChainDeletables(todoList, deletables = []) {\n const [chainAsString, ...stillToDo] = todoList;\n if (chainAsString === undefined) {\n return deletables;\n }\n\n // Fix MRA-476 (part 1): one $6 value can be worse than the other\n const withoutScriptIdentificationCode = chainAsString.replace(/( \u20216 [0-9X][0-9][0-9]-(?:XX|[0-9]+))\\/[^ ]+/u, '$1');\n\n // Remove keepless versions:\n const keepless = chainAsString.replace(/ \u20219 [A-Z]+<KEEP>/u, '');\n\n // MRA-433: 490 ind1=1 vs ind1=0: remove latter (luckily no 2nd indicator etc)\n const linked490Ind1 = chainAsString.replace(/^490 1/u, '490 0').replace(/\\t880 1/ug, '\\t880 0');\n const arr = [withoutScriptIdentificationCode, keepless, linked490Ind1].filter(val => val !== chainAsString);\n if (arr.length > 0) {\n return deriveChainDeletables([...stillToDo, ...arr], [...deletables, ...arr]);\n }\n\n return deriveChainDeletables(stillToDo, deletables);\n }\n\n}\n\nfunction isRelevantChain6(field, record) {\n //Can't be a chain:\n if (!fieldHasValidSubfield6(field) && !fieldHasValidSubfield8(field)) {\n return false;\n }\n // Too short to be a chain:\n const chain = fieldToChain(field, record);\n if (chain.length < 2) {\n return false;\n }\n // No field can contains no more than one subfield $6\n if (chain.some(f => f.subfields.filter(sf => sf.code === '6').length > 1)) {\n return false;\n }\n\n // Check whether our field is the head of a chain:\n return sameField(field, chain[0]);\n}\n\nexport function removeInferiorChains(record, fix = true) {\n const fields = record.fields.filter(f => isRelevantChain6(f, record));\n //nvdebug(`WP2.0: GOT ${fields.length} chain(s)`);\n\n const deletableChainsAsKeys = deriveInferiorChains(fields, record);\n const nChains = Object.keys(deletableChainsAsKeys).length;\n //nvdebug(`WP2: GOT ${nChains} chain(s)`);\n if (nChains === 0) {\n return [];\n }\n\n //nvdebug(`removeInferiorChains() has ${fields.length} fields-in-chain(s), and a list of ${nChains} deletable(s)`);\n\n return innerRemoveInferiorChains(fields);\n\n function innerRemoveInferiorChains(fields, deletedStringsArray = []) {\n const [currField, ...remainingFields] = fields;\n\n if (currField === undefined) {\n return deletedStringsArray;\n }\n\n const chain = fieldToChain(currField, record);\n if (chain.length === 0 || !sameField(currField, chain[0])) {\n return innerRemoveInferiorChains(remainingFields, deletedStringsArray);\n }\n\n const chainAsString = fieldsToNormalizedString(chain, 0, true, true);\n if (!(chainAsString in deletableChainsAsKeys)) {\n return innerRemoveInferiorChains(remainingFields, deletedStringsArray);\n }\n\n const triggeringField = deletableChainsAsKeys[chainAsString];\n const triggeringChain = fieldToChain(triggeringField, record);\n\n // If the inferior (deletable) chain is 1XX-based, convert the triggering better chain from 7XX to 1XX:\n if (chainContains1XX(chain)) {\n triggeringChain.forEach(f => sevenToOne(f, triggeringChain));\n }\n //nvdebug(`iRIS6C: ${chainAsString}`);\n const deletedString = fieldsToString(chain);\n const message = `DEL: '${deletedString}' REASON: '${fieldsToString(triggeringChain)}'`;\n if (fix) {\n //nvdebug(`INFERIOR $6 CHAIN REMOVAL: ${message}}`, debug);\n chain.forEach(field => record.removeField(field));\n }\n return innerRemoveInferiorChains(remainingFields, [...deletedStringsArray, message]);\n }\n\n function chainContains1XX(chain) {\n return chain.some(f => f.tag.substring(0, 1) === '1');\n }\n\n function sevenToOne(field, chain) { // Change 7XX field to 1XX field. Also handle the corresponding 880$6 7XX-NN subfields\n // NB! This function should be called only if the original 1XX gets deleted!\n if (!['700', '710', '711', '730'].includes(field.tag)) {\n return;\n }\n // Retag field 7XX as 1XX and fix corresponding occurrence numbers as well:\n const pairs = fieldGetOccurrenceNumberPairs(field, chain);\n field.tag = `1${field.tag.substring(1)}`;\n // There should always be one pair, but I'm not sanity-checking this\n pairs.forEach(pairedField => fieldSevenToOneOccurrenceNumber(pairedField));\n }\n\n}\n\n\nfunction getIdentifierlessAndKeeplessSubsets(fieldAsString) {\n // The rules below are not perfect (in complex cases they don't catch all permutations), but good enough:\n // Remove identifier(s) (MELKEHITYS-2383-ish):\n\n const identifierlessString = fieldAsString.replace(/ \u2021[01] [^\u2021]+($| \u2021)/u, '$1');\n const keeplessString = fieldAsString.replace(/ \u20219 [A-Z]+<KEEP>/u, '');\n\n return [identifierlessString, keeplessString].filter(val => val !== fieldAsString);\n}\n\nfunction deriveIndividualDeletables490(todoList, deletables = []) {\n const [fieldAsString, ...stillToDo] = todoList;\n if (fieldAsString === undefined) {\n return deletables;\n }\n //nvdebug(`PROCESS ${fieldAsString}`);\n if (!fieldAsString.match(/^490/u)) {\n return deriveIndividualDeletables490(stillToDo, deletables);\n }\n\n // $6-less version (keep this first)\n const sixless = fieldAsString.replace(/ \u20216 [^\u2021]+ \u2021/u, ' \u2021');\n\n // Without final $v or $x:\n const withoutFinalVOrX = fieldAsString.replace(/ *[;,] \u2021[vx] [^\u2021]+$/u, '');\n // Add intermediate $x-less version\n const xless = fieldAsString.replace(/, \u2021x [^\u2021]+(, \u2021x| ; \u2021v)/u, '$1');\n\n // Add $xv-less version (handled by recursion?)\n const xvless = fieldAsString.replace(/, \u2021x [^\u2021]+ \u2021v [^\u2021]+$/u, '');\n\n // MRA-433-ish (non-chain): 490 ind1=1 vs ind1=0: remove latter\n const modifiedInd2 = fieldAsString.match(/^490 1/u) ? `490 0${fieldAsString.substring(5)}` : fieldAsString;\n\n const arr = [sixless, withoutFinalVOrX, xless, xvless, modifiedInd2].filter(val => val !== fieldAsString);\n\n /*\n if (arr.length) {\n nvdebug(`${arr.length} derivation(s) for ${fieldAsString}`);\n nvdebug(arr.join('\\n'));\n }\n */\n return arr;\n}\n\nfunction deriveIndividualDeletables(record) {\n const todoList = record.fields.map(f => fieldToString(f));\n //const finishedRecord = encodingLevelIsBetterThanPrepublication(getEncodingLevel(record));\n\n const deletableStringsArray = processTodoList(todoList);\n\n const inferiorTerms = getInferiorTerms(record);\n\n return uniqArray([...deletableStringsArray, ...inferiorTerms]);\n\n function getInferiorTerms(record) {\n const inputFields = record.fields.filter(f => ['648', '650', '651'].includes(f.tag) && f.subfields);\n const result = inputFields.flatMap(f => fieldToInferiorFields(f));\n\n // console.log(result.join('\\n')); // eslint-disable-line no-console\n return result;\n }\n\n function fieldToInferiorFields(field) {\n const aArray = field.subfields.filter(sf => sf.code === 'a');\n if (field.tag === '650') {\n return aArray.flatMap(sf => [`653 ## \u2021a ${sf.value}`, `653 #0 \u2021a ${sf.value}`]);\n }\n return aArray.map(sf => `653 ## \u2021a ${sf.value}`);\n }\n\n function processTodoList(thingsToDo, deletables = []) {\n const [currString, ...stillToDo] = thingsToDo;\n\n if (currString === undefined) {\n return deletables;\n }\n\n const accentless = getAccentlessVersion(currString);\n const d490 = deriveIndividualDeletables490([currString]);\n const subsets = getIdentifierlessAndKeeplessSubsets(currString);\n const moreToDo = [...accentless, ...d490, ...subsets];\n\n\n if (currString.match(/^[1678]00/u)) {\n // Proof-of-concept rule. Should be improved eventually...\n if (currString.match(/, \u2021e [^\u2021]+\\.$/u)) {\n const tmp = currString.replace(/, \u2021e [^\u2021]+\\.$/u, '.');\n return processTodoList([tmp, ...stillToDo, ...moreToDo], [...deletables, tmp]);\n }\n }\n\n if (currString.match(/^500 ## \u2021a Lis\u00E4painokset: Lis\u00E4painos /u)) { // MET-569\n const tmp1 = currString.replace(' Lis\u00E4painos ', ' [Lis\u00E4painos] ');\n const tmp2 = currString.replace(' Lis\u00E4painos ', ' ');\n if (tmp1 !== currString && tmp2 !== currString) {\n return processTodoList([...stillToDo, ...moreToDo], [...deletables, tmp1, tmp2]);\n }\n }\n\n if (currString.match(/^500 ## \u2021a Lis\u00E4painokset: \\[Lis\u00E4painos\\] /u)) { // MET-569\n const tmp = currString.replace(' [Lis\u00E4painos] ', ' ');\n if (tmp !== currString) {\n return processTodoList([...stillToDo, ...moreToDo], [...deletables, tmp]);\n }\n }\n\n if (currString.match(/^500 ## \u2021a Ei vastaanotettu\\.$/u)) { // MELKEHITYS-3147\n return processTodoList([...stillToDo, ...moreToDo], [...deletables, '500 ## \u2021a EI VASTAANOTETTU.']);\n }\n if (currString.match(/^500 ## \u2021a Ei ilmesty\\.$/u)) { // MELKEHITYS-3147\n return processTodoList([...stillToDo, ...moreToDo], [...deletables, '500 ## \u2021a EI ILMESTY.']);\n }\n\n if (currString.match(/^505 .0.*-- \u2021t/u)) { // MRA-413-ish\n const tmp = currString.replace(/ -- \u2021t /gu, ' -- '). // remove non-initial $t subfields\n replace(/ \u2021[rg] /gu, ' '). // remove $r and $g subfields\n replace(/ \u2021t /u, ' \u2021a '). // change first $t to $a\n // ind2: '1' => '#':\n replace(/^505 (.)0/u, '505 $1#');\n if (tmp !== currString) {\n return processTodoList([tmp, ...stillToDo, ...moreToDo], [...deletables, tmp]);\n }\n //nvdebug(`505 ORIGINAL: '${fieldAsString}'`)\n //nvdebug(`505 DERIVATE: '${tmp}'`)\n }\n\n if (currString.match(/^594 ## \u2021a Ei vastaanotettu \u20215 FENNI$/u)) { // MELKEHITYS-3147\n return processTodoList([...stillToDo, ...moreToDo], [...deletables, '594 ## \u2021a EI VASTAANOTETTU \u20215 FENNI']);\n }\n if (currString.match(/^594 ## \u2021a Ei ilmesty \u20215 FENNI$/u)) { // MELKEHITYS-3147\n return processTodoList([...stillToDo, ...moreToDo], [...deletables, '594 ## \u2021a EI ILMESTY \u20215 FENNI']);\n }\n\n // MET-381: remove occurence number TAG-00, if TAG-NN existists\n if (currString.match(/^880.* \u20216 [0-9][0-9][0-9]-(?:[1-9][0-9]|0[1-9])/u)) {\n const tmp = currString.replace(/( \u20216 [0-9][0-9][0-9])-[0-9]+/u, '$1-00');\n //nvdebug(`MET-381: ADD TO DELETABLES: '${tmp}'`);\n //deletableStringsArray.push(tmp);\n if (tmp.match(/ \u20216 [0-9][0-9][0-9]-00\\/[^ ]+ /u)) {\n const tmp2 = tmp.replace(/( \u20216 [0-9][0-9][0-9]-00)[^ ]+/u, '$1');\n //nvdebug(`MET-381: ADD TO DELETABLES: '${tmp2}'`);\n return processTodoList([...stillToDo, ...moreToDo], [...deletables, tmp, tmp2]);\n }\n return processTodoList([...stillToDo, ...moreToDo], [...deletables, tmp]);\n }\n\n // MET-575 (merge: applies in postprocessing)\n const inferiorTerms = getPrepublicationTerms(currString);\n\n // MELKEHITYS-3277-ish: non-AI is better than AI (a rare case where longer version is inferior):\n const aiBased = `${currString} \u20217 (dpenmw)AI`;\n\n const newDeletables = [...deletables, ...subsets, ...accentless, ...d490, ...inferiorTerms, aiBased];\n\n if (subsets.length) {\n return processTodoList([...stillToDo, ...moreToDo], newDeletables);\n }\n\n return processTodoList([...stillToDo, ...moreToDo], newDeletables);\n }\n\n function getAccentlessVersion(string) { // MET-527\n //nvdebug(`START: '${string}`);\n // This is a sanity check: if precomposition does something, there's something wrong, and we don't want to proceed..\n if (string !== precomposeFinnishLetters(string)) {\n return [];\n }\n const accentless = String(fixComposition(string)).replace(/\\p{Diacritic}/gu, '');\n //nvdebug(`FROM '${string}'\\n TO '${accentless}'`);\n if (accentless === string) { // Don't self-destruct\n return [];\n }\n return [accentless];\n }\n\n function getPrepublicationTerms(fieldAsString) {\n if (fieldAsString.match(/^653./u)) {\n // MET-528 (extented by MET-575)\n return [`${fieldAsString} \u2021g ENNAKKOTIETO`, `${fieldAsString} \u2021g Ennakkotieto`, `${fieldAsString} \u2021g ennakkotieto`, `${fieldAsString} \u2021g ENNAKKOTIETO.`, `${fieldAsString} \u2021g Ennakkotieto.`, `${fieldAsString} \u2021g ennakkotieto.`];\n }\n\n return [];\n }\n\n}\n\nfunction fieldToNormalizedString(field) {\n const normalizedField = cloneAndNormalizeFieldForComparison(field);\n return fieldToString(normalizedField);\n}\n\nfunction deriveIndividualNormalizedDeletables(record) { // MET-461:\n const encodingLevel = getEncodingLevel(record);\n const recordIsFinished = encodingLevelIsBetterThanPrepublication(encodingLevel);\n const met495 = encodingLevel === '2' && record.fields.some(f => f.tag === '500' && fieldRefersToKoneellisestiTuotettuTietue(f));\n if (!recordIsFinished || met495) {\n return [];\n }\n const relevantFields = record.fields.filter(f => ['245', '246'].includes(f.tag) && fieldHasSubfield(f, 'a'));\n\n return deriveDeletable946s(relevantFields);\n\n function deriveDeletable946s(fields, results = []) {\n const [currField, ...remainingFields] = fields;\n if (currField === undefined) {\n return results;\n }\n\n const fieldAsNormalizedString = fieldToNormalizedString(currField);\n const tmp = fieldAsNormalizedString.replace(/^(?:...) ../u, '946 ##'). // <= Change tag to 946 and indicators to '##'\n replace(' \u2021a ', ' \u2021i nimeke onixissa \u2021a '). // Add $i before $a. NB! This is added in the normalized lower-cased form!\n replace(/(?: \\/)? \u2021c[^\u2021]+$/u, ''); // Remove $c. (Can $c be non-last?)\n const candArray = [tmp, `${tmp} \u20215 MELINDA`].filter(val => val !== fieldAsNormalizedString);\n if (candArray.length) {\n return deriveDeletable946s(remainingFields, [...results, ...candArray]);\n }\n return deriveDeletable946s(remainingFields, results);\n }\n}\n\nexport function removeIndividualInferiorDatafields(record, fix = true) { // No $6 nor $8 in field\n const deletableFieldsAsStrings = deriveIndividualDeletables(record);\n const deletableFieldsAsNormalizedStrings = deriveIndividualNormalizedDeletables(record);\n\n // nvdebug(`Deletables:\\n ${deletableFieldsAsStrings.join('\\n ')}`);\n // nvdebug(`Normalized deletables:\\n ${deletableFieldsAsNormalizedStrings.join('\\n ')}`);\n\n const hits = record.fields.filter(field => isDeletableField(field));\n\n const deletedFieldsAsStrings = hits.map(f => fieldToString(f));\n\n if (fix) {\n hits.forEach(field => {\n //nvdebug(`Remove inferior field: ${fieldToString(field)}`, debug);\n record.removeField(field);\n });\n }\n\n return deletedFieldsAsStrings;\n\n function isDeletableField(field) {\n const fieldAsString = fieldToString(field);\n if (deletableFieldsAsStrings.includes(fieldAsString)) {\n return true;\n }\n const fieldAsNormalizedString = fieldToNormalizedString(field);\n if (deletableFieldsAsNormalizedStrings.includes(fieldAsNormalizedString)) {\n return true;\n }\n\n return false;\n }\n}\n\n\nexport function removeInferiorDatafields(record, fix = true) {\n const removables = removeIndividualInferiorDatafields(record, fix); // Lone fields\n //const removables8 = removeDuplicateSubfield8Chains(record, fix); // Lone subfield $8 chains\n const removables6 = removeInferiorChains(record, fix); // Lone subfield $6 chains\n // HOW TO HANDLE $6+$8 combos? Skipping is relatively OK.\n\n nvdebug(`REMOVABLES:\\n ${removables.join('\\n ')}`, debug);\n nvdebug(`REMOVABLES 6:\\n ${removables6.join('\\n ')}`, debug);\n\n const removablesAll = removables.concat(removables6); //.concat(removables8);\n\n return removablesAll;\n}\n"],
|
|
5
|
-
"mappings": "
|
|
4
|
+
"sourcesContent": ["//import createDebugLogger from 'debug';\nimport {fieldToChain, sameField} from './removeDuplicateDataFields.js';\nimport {fieldGetOccurrenceNumberPairs, fieldHasValidSubfield6, fieldSevenToOneOccurrenceNumber, fieldsToNormalizedString} from './subfield6Utils.js';\nimport {fieldHasSubfield, fieldsToString, fieldToString, /*nvdebug,*/ uniqArray} from './utils.js';\nimport {fieldHasValidSubfield8} from './subfield8Utils.js';\nimport {encodingLevelIsBetterThanPrepublication, fieldRefersToKoneellisestiTuotettuTietue, getEncodingLevel} from './prepublicationUtils.js';\nimport {cloneAndNormalizeFieldForComparison} from './normalizeFieldForComparison.js';\nimport {fixComposition, precomposeFinnishLetters} from './normalize-utf8-diacritics.js';\n\n// Relocated from melinda-marc-record-merge-reducers (and renamed)\n\n// NB! This validator handles only full fields, and does not support subfield $8 removal.\n// Also, having multiple $8 subfields in same fields is not supported.\n// If this functionality is needed, see removeDuplicateDatafields.js for examples of subfield-only stuff.\n// const debug = createDebugLogger('@natlibfi/marc-record-validators-melinda:removeInferiorDataFields');\n\nexport default function () {\n return {\n description: 'Remove subset data fields. Certain exceptions apply, mainly too complicated for chained fields',\n validate, fix\n };\n\n function fix(record) {\n //nvdebug('Fix record: remove inferior (eg. subset) data fields', debug);\n const res = {message: [], fix: [], valid: true};\n removeInferiorDatafields(record, true);\n // This can not really fail...\n return res;\n }\n\n function validate(record) {\n // Check max, and check number of different indexes\n //nvdebug('Validate record: remove inferior (eg. subset) data fields', debug);\n\n const duplicates = removeInferiorDatafields(record, false);\n\n const res = {message: duplicates};\n\n res.valid = res.message.length < 1;\n return res;\n }\n}\n\n\nfunction deriveInferiorChains(fields, record) {\n //nvdebug(`======= GOT ${fields.length} FIELDS TO CHAINIFY`);\n const hash = {};\n\n fields.forEach(f => fieldToChainToDeletables(f));\n\n return hash;\n\n //nvdebug(`WP1: GOT ${todoList.length} CHAINS`);\n\n\n // here we map deletableStringObject[str] => field. The idea is to help debugging. We don't actually need the field object...\n //return deriveChainDeletables(todoList);\n\n function fieldToChainToDeletables(field) {\n const chain = fieldToChain(field, record);\n if (chain.length < 2) {\n return;\n }\n const chainAsString = fieldsToNormalizedString(chain, 0, true, true);\n const arr = deriveChainDeletables([chainAsString]);\n //nvdebug(`GOT ${arr.length} DELETABLES FOR ${chainAsString}`);\n arr.forEach(val => {\n if (!(val in hash)) {\n hash[val] = field;\n }\n });\n }\n\n function deriveChainDeletables(todoList, deletables = []) {\n const [chainAsString, ...stillToDo] = todoList;\n if (chainAsString === undefined) {\n return deletables;\n }\n\n // Fix MRA-476 (part 1): one $6 value can be worse than the other\n const withoutScriptIdentificationCode = chainAsString.replace(/( \u20216 [0-9X][0-9][0-9]-(?:XX|[0-9]+))\\/[^ ]+/u, '$1');\n\n // Remove keepless versions:\n const keepless = chainAsString.replace(/ \u20219 [A-Z]+<KEEP>/u, '');\n\n // MRA-433: 490 ind1=1 vs ind1=0: remove latter (luckily no 2nd indicator etc)\n const linked490Ind1 = chainAsString.replace(/^490 1/u, '490 0').replace(/\\t880 1/ug, '\\t880 0');\n const arr = [withoutScriptIdentificationCode, keepless, linked490Ind1].filter(val => val !== chainAsString);\n if (arr.length > 0) {\n return deriveChainDeletables([...stillToDo, ...arr], [...deletables, ...arr]);\n }\n\n return deriveChainDeletables(stillToDo, deletables);\n }\n\n}\n\nfunction isRelevantChain6(field, record) {\n //Can't be a chain:\n if (!fieldHasValidSubfield6(field) && !fieldHasValidSubfield8(field)) {\n return false;\n }\n // Too short to be a chain:\n const chain = fieldToChain(field, record);\n if (chain.length < 2) {\n return false;\n }\n // No field can contains no more than one subfield $6\n if (chain.some(f => f.subfields.filter(sf => sf.code === '6').length > 1)) {\n return false;\n }\n\n // Check whether our field is the head of a chain:\n return sameField(field, chain[0]);\n}\n\nexport function removeInferiorChains(record, fix = true) {\n const fields = record.fields.filter(f => isRelevantChain6(f, record));\n //nvdebug(`WP2.0: GOT ${fields.length} chain(s)`);\n\n const deletableChainsAsKeys = deriveInferiorChains(fields, record);\n const nChains = Object.keys(deletableChainsAsKeys).length;\n //nvdebug(`WP2: GOT ${nChains} chain(s)`);\n if (nChains === 0) {\n return [];\n }\n\n //nvdebug(`removeInferiorChains() has ${fields.length} fields-in-chain(s), and a list of ${nChains} deletable(s)`);\n\n return innerRemoveInferiorChains(fields);\n\n function innerRemoveInferiorChains(fields, deletedStringsArray = []) {\n const [currField, ...remainingFields] = fields;\n\n if (currField === undefined) {\n return deletedStringsArray;\n }\n\n const chain = fieldToChain(currField, record);\n if (chain.length === 0 || !sameField(currField, chain[0])) {\n return innerRemoveInferiorChains(remainingFields, deletedStringsArray);\n }\n\n const chainAsString = fieldsToNormalizedString(chain, 0, true, true);\n if (!(chainAsString in deletableChainsAsKeys)) {\n return innerRemoveInferiorChains(remainingFields, deletedStringsArray);\n }\n\n const triggeringField = deletableChainsAsKeys[chainAsString];\n const triggeringChain = fieldToChain(triggeringField, record);\n\n // If the inferior (deletable) chain is 1XX-based, convert the triggering better chain from 7XX to 1XX:\n if (chainContains1XX(chain)) {\n triggeringChain.forEach(f => sevenToOne(f, triggeringChain));\n }\n //nvdebug(`iRIS6C: ${chainAsString}`);\n const deletedString = fieldsToString(chain);\n const message = `DEL: '${deletedString}' REASON: '${fieldsToString(triggeringChain)}'`;\n if (fix) {\n //nvdebug(`INFERIOR $6 CHAIN REMOVAL: ${message}}`, debug);\n chain.forEach(field => record.removeField(field));\n }\n return innerRemoveInferiorChains(remainingFields, [...deletedStringsArray, message]);\n }\n\n function chainContains1XX(chain) {\n return chain.some(f => f.tag.substring(0, 1) === '1');\n }\n\n function sevenToOne(field, chain) { // Change 7XX field to 1XX field. Also handle the corresponding 880$6 7XX-NN subfields\n // NB! This function should be called only if the original 1XX gets deleted!\n if (!['700', '710', '711', '730'].includes(field.tag)) {\n return;\n }\n // Retag field 7XX as 1XX and fix corresponding occurrence numbers as well:\n const pairs = fieldGetOccurrenceNumberPairs(field, chain);\n field.tag = `1${field.tag.substring(1)}`;\n // There should always be one pair, but I'm not sanity-checking this\n pairs.forEach(pairedField => fieldSevenToOneOccurrenceNumber(pairedField));\n }\n\n}\n\n\nfunction getIdentifierlessAndKeeplessSubsets(fieldAsString) {\n // The rules below are not perfect (in complex cases they don't catch all permutations), but good enough:\n // Remove identifier(s) (MELKEHITYS-2383-ish):\n\n const identifierlessString = fieldAsString.replace(/ \u2021[01] [^\u2021]+($| \u2021)/u, '$1');\n const keeplessString = fieldAsString.replace(/ \u20219 [A-Z]+<KEEP>/u, '');\n\n return [identifierlessString, keeplessString].filter(val => val !== fieldAsString);\n}\n\nfunction deriveIndividualDeletables490(todoList, deletables = []) {\n const [fieldAsString, ...stillToDo] = todoList;\n if (fieldAsString === undefined) {\n return deletables;\n }\n //nvdebug(`PROCESS ${fieldAsString}`);\n if (!fieldAsString.match(/^490/u)) {\n return deriveIndividualDeletables490(stillToDo, deletables);\n }\n\n // $6-less version (keep this first)\n const sixless = fieldAsString.replace(/ \u20216 [^\u2021]+ \u2021/u, ' \u2021');\n\n // Without final $v or $x:\n const withoutFinalVOrX = fieldAsString.replace(/ *[;,] \u2021[vx] [^\u2021]+$/u, '');\n // Add intermediate $x-less version\n const xless = fieldAsString.replace(/, \u2021x [^\u2021]+(, \u2021x| ; \u2021v)/u, '$1');\n\n // Add $xv-less version (handled by recursion?)\n const xvless = fieldAsString.replace(/, \u2021x [^\u2021]+ \u2021v [^\u2021]+$/u, '');\n\n // MRA-433-ish (non-chain): 490 ind1=1 vs ind1=0: remove latter\n const modifiedInd2 = fieldAsString.match(/^490 1/u) ? `490 0${fieldAsString.substring(5)}` : fieldAsString;\n\n const arr = [sixless, withoutFinalVOrX, xless, xvless, modifiedInd2].filter(val => val !== fieldAsString);\n\n /*\n if (arr.length) {\n nvdebug(`${arr.length} derivation(s) for ${fieldAsString}`);\n nvdebug(arr.join('\\n'));\n }\n */\n return arr;\n}\n\nfunction deriveIndividualDeletables(record) {\n const todoList = record.fields.map(f => fieldToString(f));\n //const finishedRecord = encodingLevelIsBetterThanPrepublication(getEncodingLevel(record));\n\n const deletableStringsArray = processTodoList(todoList);\n\n const inferiorTerms = getInferiorTerms(record);\n\n return uniqArray([...deletableStringsArray, ...inferiorTerms]);\n\n function getInferiorTerms(record) {\n const inputFields = record.fields.filter(f => ['648', '650', '651'].includes(f.tag) && f.subfields);\n const result = inputFields.flatMap(f => fieldToInferiorFields(f));\n\n // console.log(result.join('\\n')); // eslint-disable-line no-console\n return result;\n }\n\n function fieldToInferiorFields(field) {\n const aArray = field.subfields.filter(sf => sf.code === 'a');\n if (field.tag === '650') {\n return aArray.flatMap(sf => [`653 ## \u2021a ${sf.value}`, `653 #0 \u2021a ${sf.value}`]);\n }\n return aArray.map(sf => `653 ## \u2021a ${sf.value}`);\n }\n\n function processTodoList(thingsToDo, deletables = []) {\n const [currString, ...stillToDo] = thingsToDo;\n\n if (currString === undefined) {\n return deletables;\n }\n\n const accentless = getAccentlessVersion(currString);\n const d490 = deriveIndividualDeletables490([currString]);\n const subsets = getIdentifierlessAndKeeplessSubsets(currString);\n const moreToDo = [...accentless, ...d490, ...subsets];\n\n\n if (currString.match(/^[1678]00/u)) {\n // Proof-of-concept rule. Should be improved eventually...\n if (currString.match(/, \u2021e [^\u2021]+\\.$/u)) {\n const tmp = currString.replace(/, \u2021e [^\u2021]+\\.$/u, '.');\n return processTodoList([tmp, ...stillToDo, ...moreToDo], [...deletables, tmp]);\n }\n }\n\n if (currString.match(/^500 ## \u2021a Lis\u00E4painokset: Lis\u00E4painos /u)) { // MET-569\n const tmp1 = currString.replace(' Lis\u00E4painos ', ' [Lis\u00E4painos] ');\n const tmp2 = currString.replace(' Lis\u00E4painos ', ' ');\n if (tmp1 !== currString && tmp2 !== currString) {\n return processTodoList([...stillToDo, ...moreToDo], [...deletables, tmp1, tmp2]);\n }\n }\n\n if (currString.match(/^500 ## \u2021a Lis\u00E4painokset: \\[Lis\u00E4painos\\] /u)) { // MET-569\n const tmp = currString.replace(' [Lis\u00E4painos] ', ' ');\n if (tmp !== currString) {\n return processTodoList([...stillToDo, ...moreToDo], [...deletables, tmp]);\n }\n }\n\n if (currString.match(/^500 ## \u2021a Ei vastaanotettu\\.$/u)) { // MELKEHITYS-3147\n return processTodoList([...stillToDo, ...moreToDo], [...deletables, '500 ## \u2021a EI VASTAANOTETTU.']);\n }\n if (currString.match(/^500 ## \u2021a Ei ilmesty\\.$/u)) { // MELKEHITYS-3147\n return processTodoList([...stillToDo, ...moreToDo], [...deletables, '500 ## \u2021a EI ILMESTY.']);\n }\n\n if (currString.match(/^505 .0.*-- \u2021t/u)) { // MRA-413-ish\n const tmp = currString.replace(/ -- \u2021t /gu, ' -- '). // remove non-initial $t subfields\n replace(/ \u2021[rg] /gu, ' '). // remove $r and $g subfields\n replace(/ \u2021t /u, ' \u2021a '). // change first $t to $a\n // ind2: '1' => '#':\n replace(/^505 (.)0/u, '505 $1#');\n if (tmp !== currString) {\n return processTodoList([tmp, ...stillToDo, ...moreToDo], [...deletables, tmp]);\n }\n //nvdebug(`505 ORIGINAL: '${fieldAsString}'`)\n //nvdebug(`505 DERIVATE: '${tmp}'`)\n }\n\n if (currString.match(/^594 ## \u2021a Ei vastaanotettu \u20215 FENNI$/u)) { // MELKEHITYS-3147\n return processTodoList([...stillToDo, ...moreToDo], [...deletables, '594 ## \u2021a EI VASTAANOTETTU \u20215 FENNI']);\n }\n if (currString.match(/^594 ## \u2021a Ei ilmesty \u20215 FENNI$/u)) { // MELKEHITYS-3147\n return processTodoList([...stillToDo, ...moreToDo], [...deletables, '594 ## \u2021a EI ILMESTY \u20215 FENNI']);\n }\n\n // MET-381: remove occurence number TAG-00, if TAG-NN existists\n if (currString.match(/^880.* \u20216 [0-9][0-9][0-9]-(?:[1-9][0-9]|0[1-9])/u)) {\n const tmp = currString.replace(/( \u20216 [0-9][0-9][0-9])-[0-9]+/u, '$1-00');\n //nvdebug(`MET-381: ADD TO DELETABLES: '${tmp}'`);\n //deletableStringsArray.push(tmp);\n if (tmp.match(/ \u20216 [0-9][0-9][0-9]-00\\/[^ ]+ /u)) {\n const tmp2 = tmp.replace(/( \u20216 [0-9][0-9][0-9]-00)[^ ]+/u, '$1');\n //nvdebug(`MET-381: ADD TO DELETABLES: '${tmp2}'`);\n return processTodoList([...stillToDo, ...moreToDo], [...deletables, tmp, tmp2]);\n }\n return processTodoList([...stillToDo, ...moreToDo], [...deletables, tmp]);\n }\n\n // MET-575 (merge: applies in postprocessing)\n const inferiorTerms = getPrepublicationTerms(currString);\n\n // MELKEHITYS-3277-ish: non-AI is better than AI (a rare case where longer version is inferior):\n const aiBased = `${currString} \u20217 (dpenmw)AI`;\n\n const newDeletables = [...deletables, ...subsets, ...accentless, ...d490, ...inferiorTerms, aiBased];\n\n if (subsets.length) {\n return processTodoList([...stillToDo, ...moreToDo], newDeletables);\n }\n\n return processTodoList([...stillToDo, ...moreToDo], newDeletables);\n }\n\n function getAccentlessVersion(string) { // MET-527\n //nvdebug(`START: '${string}`);\n // This is a sanity check: if precomposition does something, there's something wrong, and we don't want to proceed..\n if (string !== precomposeFinnishLetters(string)) {\n return [];\n }\n const accentless = String(fixComposition(string)).replace(/\\p{Diacritic}/gu, '');\n //nvdebug(`FROM '${string}'\\n TO '${accentless}'`);\n if (accentless === string) { // Don't self-destruct\n return [];\n }\n return [accentless];\n }\n\n function getPrepublicationTerms(fieldAsString) {\n if (fieldAsString.match(/^653./u)) {\n // MET-528 (extented by MET-575)\n return [`${fieldAsString} \u2021g ENNAKKOTIETO`, `${fieldAsString} \u2021g Ennakkotieto`, `${fieldAsString} \u2021g ennakkotieto`, `${fieldAsString} \u2021g ENNAKKOTIETO.`, `${fieldAsString} \u2021g Ennakkotieto.`, `${fieldAsString} \u2021g ennakkotieto.`];\n }\n\n return [];\n }\n\n}\n\nfunction fieldToNormalizedString(field) {\n const normalizedField = cloneAndNormalizeFieldForComparison(field);\n return fieldToString(normalizedField);\n}\n\nfunction deriveIndividualNormalizedDeletables(record) { // MET-461:\n const encodingLevel = getEncodingLevel(record);\n const recordIsFinished = encodingLevelIsBetterThanPrepublication(encodingLevel);\n const met495 = encodingLevel === '2' && record.fields.some(f => f.tag === '500' && fieldRefersToKoneellisestiTuotettuTietue(f));\n if (!recordIsFinished || met495) {\n return [];\n }\n const relevantFields = record.fields.filter(f => ['245', '246'].includes(f.tag) && fieldHasSubfield(f, 'a'));\n\n return deriveDeletable946s(relevantFields);\n\n function deriveDeletable946s(fields, results = []) {\n const [currField, ...remainingFields] = fields;\n if (currField === undefined) {\n return results;\n }\n\n const fieldAsNormalizedString = fieldToNormalizedString(currField);\n const tmp = fieldAsNormalizedString.replace(/^(?:...) ../u, '946 ##'). // <= Change tag to 946 and indicators to '##'\n replace(' \u2021a ', ' \u2021i nimeke onixissa \u2021a '). // Add $i before $a. NB! This is added in the normalized lower-cased form!\n replace(/(?: \\/)? \u2021c[^\u2021]+$/u, ''); // Remove $c. (Can $c be non-last?)\n const candArray = [tmp, `${tmp} \u20215 MELINDA`].filter(val => val !== fieldAsNormalizedString);\n if (candArray.length) {\n return deriveDeletable946s(remainingFields, [...results, ...candArray]);\n }\n return deriveDeletable946s(remainingFields, results);\n }\n}\n\nexport function removeIndividualInferiorDatafields(record, fix = true) { // No $6 nor $8 in field\n const deletableFieldsAsStrings = deriveIndividualDeletables(record);\n const deletableFieldsAsNormalizedStrings = deriveIndividualNormalizedDeletables(record);\n\n // nvdebug(`Deletables:\\n ${deletableFieldsAsStrings.join('\\n ')}`);\n // nvdebug(`Normalized deletables:\\n ${deletableFieldsAsNormalizedStrings.join('\\n ')}`);\n\n const hits = record.fields.filter(field => isDeletableField(field));\n\n const deletedFieldsAsStrings = hits.map(f => fieldToString(f));\n\n if (fix) {\n hits.forEach(field => {\n //nvdebug(`Remove inferior field: ${fieldToString(field)}`, debug);\n record.removeField(field);\n });\n }\n\n return deletedFieldsAsStrings;\n\n function isDeletableField(field) {\n const fieldAsString = fieldToString(field);\n if (deletableFieldsAsStrings.includes(fieldAsString)) {\n return true;\n }\n const fieldAsNormalizedString = fieldToNormalizedString(field);\n if (deletableFieldsAsNormalizedStrings.includes(fieldAsNormalizedString)) {\n return true;\n }\n\n return false;\n }\n}\n\n\nexport function removeInferiorDatafields(record, fix = true) {\n const removables = removeIndividualInferiorDatafields(record, fix); // Lone fields\n //const removables8 = removeDuplicateSubfield8Chains(record, fix); // Lone subfield $8 chains\n const removables6 = removeInferiorChains(record, fix); // Lone subfield $6 chains\n // HOW TO HANDLE $6+$8 combos? Skipping is relatively OK.\n\n //nvdebug(`REMOVABLES:\\n ${removables.join('\\n ')}`, debug);\n //nvdebug(`REMOVABLES 6:\\n ${removables6.join('\\n ')}`, debug);\n\n const removablesAll = removables.concat(removables6); //.concat(removables8);\n\n return removablesAll;\n}\n"],
|
|
5
|
+
"mappings": "AACA,SAAQ,cAAc,iBAAgB;AACtC,SAAQ,+BAA+B,wBAAwB,iCAAiC,gCAA+B;AAC/H,SAAQ,kBAAkB,gBAAgB,eAA4B,iBAAgB;AACtF,SAAQ,8BAA6B;AACrC,SAAQ,yCAAyC,0CAA0C,wBAAuB;AAClH,SAAQ,2CAA0C;AAClD,SAAQ,gBAAgB,gCAA+B;AASvD,0BAA2B;AACzB,SAAO;AAAA,IACL,aAAa;AAAA,IACb;AAAA,IAAU;AAAA,EACZ;AAEA,WAAS,IAAI,QAAQ;AAEnB,UAAM,MAAM,EAAC,SAAS,CAAC,GAAG,KAAK,CAAC,GAAG,OAAO,KAAI;AAC9C,6BAAyB,QAAQ,IAAI;AAErC,WAAO;AAAA,EACT;AAEA,WAAS,SAAS,QAAQ;AAIxB,UAAM,aAAa,yBAAyB,QAAQ,KAAK;AAEzD,UAAM,MAAM,EAAC,SAAS,WAAU;AAEhC,QAAI,QAAQ,IAAI,QAAQ,SAAS;AACjC,WAAO;AAAA,EACT;AACF;AAGA,SAAS,qBAAqB,QAAQ,QAAQ;AAE5C,QAAM,OAAO,CAAC;AAEd,SAAO,QAAQ,OAAK,yBAAyB,CAAC,CAAC;AAE/C,SAAO;AAQP,WAAS,yBAAyB,OAAO;AACvC,UAAM,QAAQ,aAAa,OAAO,MAAM;AACxC,QAAI,MAAM,SAAS,GAAG;AACpB;AAAA,IACF;AACA,UAAM,gBAAgB,yBAAyB,OAAO,GAAG,MAAM,IAAI;AACnE,UAAM,MAAM,sBAAsB,CAAC,aAAa,CAAC;AAEjD,QAAI,QAAQ,SAAO;AACjB,UAAI,EAAE,OAAO,OAAO;AAClB,aAAK,GAAG,IAAI;AAAA,MACd;AAAA,IACF,CAAC;AAAA,EACH;AAEA,WAAS,sBAAsB,UAAU,aAAa,CAAC,GAAG;AACxD,UAAM,CAAC,eAAe,GAAG,SAAS,IAAI;AACtC,QAAI,kBAAkB,QAAW;AAC/B,aAAO;AAAA,IACT;AAGA,UAAM,kCAAkC,cAAc,QAAQ,gDAAgD,IAAI;AAGlH,UAAM,WAAW,cAAc,QAAQ,qBAAqB,EAAE;AAG9D,UAAM,gBAAgB,cAAc,QAAQ,WAAW,OAAO,EAAE,QAAQ,aAAa,QAAS;AAC9F,UAAM,MAAM,CAAC,iCAAiC,UAAU,aAAa,EAAE,OAAO,SAAO,QAAQ,aAAa;AAC1G,QAAI,IAAI,SAAS,GAAG;AAClB,aAAO,sBAAsB,CAAC,GAAG,WAAW,GAAG,GAAG,GAAG,CAAC,GAAG,YAAY,GAAG,GAAG,CAAC;AAAA,IAC9E;AAEA,WAAO,sBAAsB,WAAW,UAAU;AAAA,EACpD;AAEF;AAEA,SAAS,iBAAiB,OAAO,QAAQ;AAEvC,MAAI,CAAC,uBAAuB,KAAK,KAAK,CAAC,uBAAuB,KAAK,GAAG;AACpE,WAAO;AAAA,EACT;AAEA,QAAM,QAAQ,aAAa,OAAO,MAAM;AACxC,MAAI,MAAM,SAAS,GAAG;AACpB,WAAO;AAAA,EACT;AAEA,MAAI,MAAM,KAAK,OAAK,EAAE,UAAU,OAAO,QAAM,GAAG,SAAS,GAAG,EAAE,SAAS,CAAC,GAAG;AACzE,WAAO;AAAA,EACT;AAGA,SAAO,UAAU,OAAO,MAAM,CAAC,CAAC;AAClC;AAEO,gBAAS,qBAAqB,QAAQ,MAAM,MAAM;AACvD,QAAM,SAAS,OAAO,OAAO,OAAO,OAAK,iBAAiB,GAAG,MAAM,CAAC;AAGpE,QAAM,wBAAwB,qBAAqB,QAAQ,MAAM;AACjE,QAAM,UAAU,OAAO,KAAK,qBAAqB,EAAE;AAEnD,MAAI,YAAY,GAAG;AACjB,WAAO,CAAC;AAAA,EACV;AAIA,SAAO,0BAA0B,MAAM;AAEvC,WAAS,0BAA0BA,SAAQ,sBAAsB,CAAC,GAAG;AACnE,UAAM,CAAC,WAAW,GAAG,eAAe,IAAIA;AAExC,QAAI,cAAc,QAAW;AAC3B,aAAO;AAAA,IACT;AAEA,UAAM,QAAQ,aAAa,WAAW,MAAM;AAC5C,QAAI,MAAM,WAAW,KAAK,CAAC,UAAU,WAAW,MAAM,CAAC,CAAC,GAAG;AACzD,aAAO,0BAA0B,iBAAiB,mBAAmB;AAAA,IACvE;AAEA,UAAM,gBAAgB,yBAAyB,OAAO,GAAG,MAAM,IAAI;AACnE,QAAI,EAAE,iBAAiB,wBAAwB;AAC7C,aAAO,0BAA0B,iBAAiB,mBAAmB;AAAA,IACvE;AAEA,UAAM,kBAAkB,sBAAsB,aAAa;AAC3D,UAAM,kBAAkB,aAAa,iBAAiB,MAAM;AAG5D,QAAI,iBAAiB,KAAK,GAAG;AAC3B,sBAAgB,QAAQ,OAAK,WAAW,GAAG,eAAe,CAAC;AAAA,IAC7D;AAEA,UAAM,gBAAgB,eAAe,KAAK;AAC1C,UAAM,UAAU,SAAS,aAAa,eAAe,eAAe,eAAe,CAAC;AACpF,QAAI,KAAK;AAEP,YAAM,QAAQ,WAAS,OAAO,YAAY,KAAK,CAAC;AAAA,IAClD;AACA,WAAO,0BAA0B,iBAAiB,CAAC,GAAG,qBAAqB,OAAO,CAAC;AAAA,EACrF;AAEA,WAAS,iBAAiB,OAAO;AAC/B,WAAO,MAAM,KAAK,OAAK,EAAE,IAAI,UAAU,GAAG,CAAC,MAAM,GAAG;AAAA,EACtD;AAEA,WAAS,WAAW,OAAO,OAAO;AAEhC,QAAI,CAAC,CAAC,OAAO,OAAO,OAAO,KAAK,EAAE,SAAS,MAAM,GAAG,GAAG;AACrD;AAAA,IACF;AAEA,UAAM,QAAQ,8BAA8B,OAAO,KAAK;AACxD,UAAM,MAAM,IAAI,MAAM,IAAI,UAAU,CAAC,CAAC;AAEtC,UAAM,QAAQ,iBAAe,gCAAgC,WAAW,CAAC;AAAA,EAC3E;AAEF;AAGA,SAAS,oCAAoC,eAAe;AAI1D,QAAM,uBAAuB,cAAc,QAAQ,uBAAuB,IAAI;AAC9E,QAAM,iBAAiB,cAAc,QAAQ,qBAAqB,EAAE;AAEpE,SAAO,CAAC,sBAAsB,cAAc,EAAE,OAAO,SAAO,QAAQ,aAAa;AACnF;AAEA,SAAS,8BAA8B,UAAU,aAAa,CAAC,GAAG;AAChE,QAAM,CAAC,eAAe,GAAG,SAAS,IAAI;AACtC,MAAI,kBAAkB,QAAW;AAC/B,WAAO;AAAA,EACT;AAEA,MAAI,CAAC,cAAc,MAAM,OAAO,GAAG;AACjC,WAAO,8BAA8B,WAAW,UAAU;AAAA,EAC5D;AAGA,QAAM,UAAU,cAAc,QAAQ,gBAAgB,SAAI;AAG1D,QAAM,mBAAmB,cAAc,QAAQ,wBAAwB,EAAE;AAEzE,QAAM,QAAQ,cAAc,QAAQ,2BAA2B,IAAI;AAGnE,QAAM,SAAS,cAAc,QAAQ,yBAAyB,EAAE;AAGhE,QAAM,eAAe,cAAc,MAAM,SAAS,IAAI,QAAQ,cAAc,UAAU,CAAC,CAAC,KAAK;AAE7F,QAAM,MAAM,CAAC,SAAS,kBAAkB,OAAO,QAAQ,YAAY,EAAE,OAAO,SAAO,QAAQ,aAAa;AAQxG,SAAO;AACT;AAEA,SAAS,2BAA2B,QAAQ;AAC1C,QAAM,WAAW,OAAO,OAAO,IAAI,OAAK,cAAc,CAAC,CAAC;AAGxD,QAAM,wBAAwB,gBAAgB,QAAQ;AAEtD,QAAM,gBAAgB,iBAAiB,MAAM;AAE7C,SAAO,UAAU,CAAC,GAAG,uBAAuB,GAAG,aAAa,CAAC;AAE7D,WAAS,iBAAiBC,SAAQ;AAChC,UAAM,cAAcA,QAAO,OAAO,OAAO,OAAK,CAAC,OAAO,OAAO,KAAK,EAAE,SAAS,EAAE,GAAG,KAAK,EAAE,SAAS;AAClG,UAAM,SAAS,YAAY,QAAQ,OAAK,sBAAsB,CAAC,CAAC;AAGhE,WAAO;AAAA,EACT;AAEA,WAAS,sBAAsB,OAAO;AACpC,UAAM,SAAS,MAAM,UAAU,OAAO,QAAM,GAAG,SAAS,GAAG;AAC3D,QAAI,MAAM,QAAQ,OAAO;AACvB,aAAO,OAAO,QAAQ,QAAM,CAAC,kBAAa,GAAG,KAAK,IAAI,kBAAa,GAAG,KAAK,EAAE,CAAC;AAAA,IAChF;AACA,WAAO,OAAO,IAAI,QAAM,kBAAa,GAAG,KAAK,EAAE;AAAA,EACjD;AAEA,WAAS,gBAAgB,YAAY,aAAa,CAAC,GAAG;AACpD,UAAM,CAAC,YAAY,GAAG,SAAS,IAAI;AAEnC,QAAI,eAAe,QAAW;AAC5B,aAAO;AAAA,IACT;AAEA,UAAM,aAAa,qBAAqB,UAAU;AAClD,UAAM,OAAO,8BAA8B,CAAC,UAAU,CAAC;AACvD,UAAM,UAAU,oCAAoC,UAAU;AAC9D,UAAM,WAAW,CAAC,GAAG,YAAY,GAAG,MAAM,GAAG,OAAO;AAGpD,QAAI,WAAW,MAAM,YAAY,GAAG;AAElC,UAAI,WAAW,MAAM,gBAAgB,GAAG;AACtC,cAAM,MAAM,WAAW,QAAQ,kBAAkB,GAAG;AACpD,eAAO,gBAAgB,CAAC,KAAK,GAAG,WAAW,GAAG,QAAQ,GAAG,CAAC,GAAG,YAAY,GAAG,CAAC;AAAA,MAC/E;AAAA,IACF;AAEA,QAAI,WAAW,MAAM,wCAAwC,GAAG;AAC9D,YAAM,OAAO,WAAW,QAAQ,mBAAgB,mBAAgB;AAChE,YAAM,OAAO,WAAW,QAAQ,mBAAgB,GAAG;AACnD,UAAI,SAAS,cAAc,SAAS,YAAY;AAC9C,eAAO,gBAAgB,CAAC,GAAG,WAAW,GAAG,QAAQ,GAAG,CAAC,GAAG,YAAY,MAAM,IAAI,CAAC;AAAA,MACjF;AAAA,IACF;AAEA,QAAI,WAAW,MAAM,4CAA4C,GAAG;AAClE,YAAM,MAAM,WAAW,QAAQ,qBAAkB,GAAG;AACpD,UAAI,QAAQ,YAAY;AACtB,eAAO,gBAAgB,CAAC,GAAG,WAAW,GAAG,QAAQ,GAAG,CAAC,GAAG,YAAY,GAAG,CAAC;AAAA,MAC1E;AAAA,IACF;AAEA,QAAI,WAAW,MAAM,iCAAiC,GAAG;AACvD,aAAO,gBAAgB,CAAC,GAAG,WAAW,GAAG,QAAQ,GAAG,CAAC,GAAG,YAAY,kCAA6B,CAAC;AAAA,IACpG;AACA,QAAI,WAAW,MAAM,2BAA2B,GAAG;AACjD,aAAO,gBAAgB,CAAC,GAAG,WAAW,GAAG,QAAQ,GAAG,CAAC,GAAG,YAAY,4BAAuB,CAAC;AAAA,IAC9F;AAEA,QAAI,WAAW,MAAM,iBAAiB,GAAG;AACvC,YAAM,MAAM,WAAW,QAAQ,aAAa,MAAM,EAChD,QAAQ,aAAa,GAAG,EACxB,QAAQ,SAAS,WAAM,EAEvB,QAAQ,cAAc,SAAS;AACjC,UAAI,QAAQ,YAAY;AACtB,eAAO,gBAAgB,CAAC,KAAK,GAAG,WAAW,GAAG,QAAQ,GAAG,CAAC,GAAG,YAAY,GAAG,CAAC;AAAA,MAC/E;AAAA,IAGF;AAEA,QAAI,WAAW,MAAM,wCAAwC,GAAG;AAC9D,aAAO,gBAAgB,CAAC,GAAG,WAAW,GAAG,QAAQ,GAAG,CAAC,GAAG,YAAY,+CAAqC,CAAC;AAAA,IAC5G;AACA,QAAI,WAAW,MAAM,kCAAkC,GAAG;AACxD,aAAO,gBAAgB,CAAC,GAAG,WAAW,GAAG,QAAQ,GAAG,CAAC,GAAG,YAAY,yCAA+B,CAAC;AAAA,IACtG;AAGA,QAAI,WAAW,MAAM,kDAAkD,GAAG;AACxE,YAAM,MAAM,WAAW,QAAQ,iCAAiC,OAAO;AAGvE,UAAI,IAAI,MAAM,iCAAiC,GAAG;AAChD,cAAM,OAAO,IAAI,QAAQ,kCAAkC,IAAI;AAE/D,eAAO,gBAAgB,CAAC,GAAG,WAAW,GAAG,QAAQ,GAAG,CAAC,GAAG,YAAY,KAAK,IAAI,CAAC;AAAA,MAChF;AACA,aAAO,gBAAgB,CAAC,GAAG,WAAW,GAAG,QAAQ,GAAG,CAAC,GAAG,YAAY,GAAG,CAAC;AAAA,IAC1E;AAGA,UAAMC,iBAAgB,uBAAuB,UAAU;AAGvD,UAAM,UAAU,GAAG,UAAU;AAE7B,UAAM,gBAAgB,CAAC,GAAG,YAAY,GAAG,SAAS,GAAG,YAAY,GAAG,MAAM,GAAGA,gBAAe,OAAO;AAEnG,QAAI,QAAQ,QAAQ;AAClB,aAAO,gBAAgB,CAAC,GAAG,WAAW,GAAG,QAAQ,GAAG,aAAa;AAAA,IACnE;AAEA,WAAO,gBAAgB,CAAC,GAAG,WAAW,GAAG,QAAQ,GAAG,aAAa;AAAA,EACnE;AAEA,WAAS,qBAAqB,QAAQ;AAGpC,QAAI,WAAW,yBAAyB,MAAM,GAAG;AAC/C,aAAO,CAAC;AAAA,IACV;AACA,UAAM,aAAa,OAAO,eAAe,MAAM,CAAC,EAAE,QAAQ,mBAAmB,EAAE;AAE/E,QAAI,eAAe,QAAQ;AACzB,aAAO,CAAC;AAAA,IACV;AACA,WAAO,CAAC,UAAU;AAAA,EACpB;AAEA,WAAS,uBAAuB,eAAe;AAC7C,QAAI,cAAc,MAAM,QAAQ,GAAG;AAEjC,aAAO,CAAC,GAAG,aAAa,yBAAoB,GAAG,aAAa,yBAAoB,GAAG,aAAa,yBAAoB,GAAG,aAAa,0BAAqB,GAAG,aAAa,0BAAqB,GAAG,aAAa,wBAAmB;AAAA,IACnO;AAEA,WAAO,CAAC;AAAA,EACV;AAEF;AAEA,SAAS,wBAAwB,OAAO;AACtC,QAAM,kBAAkB,oCAAoC,KAAK;AACjE,SAAO,cAAc,eAAe;AACtC;AAEA,SAAS,qCAAqC,QAAQ;AACpD,QAAM,gBAAgB,iBAAiB,MAAM;AAC7C,QAAM,mBAAmB,wCAAwC,aAAa;AAC9E,QAAM,SAAS,kBAAkB,OAAO,OAAO,OAAO,KAAK,OAAK,EAAE,QAAQ,SAAS,yCAAyC,CAAC,CAAC;AAC9H,MAAI,CAAC,oBAAoB,QAAQ;AAC/B,WAAO,CAAC;AAAA,EACV;AACA,QAAM,iBAAiB,OAAO,OAAO,OAAO,OAAK,CAAC,OAAO,KAAK,EAAE,SAAS,EAAE,GAAG,KAAK,iBAAiB,GAAG,GAAG,CAAC;AAE3G,SAAO,oBAAoB,cAAc;AAEzC,WAAS,oBAAoB,QAAQ,UAAU,CAAC,GAAG;AACjD,UAAM,CAAC,WAAW,GAAG,eAAe,IAAI;AACxC,QAAI,cAAc,QAAW;AAC3B,aAAO;AAAA,IACT;AAEA,UAAM,0BAA0B,wBAAwB,SAAS;AACjE,UAAM,MAAM,wBAAwB,QAAQ,gBAAgB,QAAQ,EAClE,QAAQ,aAAQ,mCAAyB,EACzC,QAAQ,sBAAsB,EAAE;AAClC,UAAM,YAAY,CAAC,KAAK,GAAG,GAAG,kBAAa,EAAE,OAAO,SAAO,QAAQ,uBAAuB;AAC1F,QAAI,UAAU,QAAQ;AACpB,aAAO,oBAAoB,iBAAiB,CAAC,GAAG,SAAS,GAAG,SAAS,CAAC;AAAA,IACxE;AACA,WAAO,oBAAoB,iBAAiB,OAAO;AAAA,EACrD;AACF;AAEO,gBAAS,mCAAmC,QAAQ,MAAM,MAAM;AACrE,QAAM,2BAA2B,2BAA2B,MAAM;AAClE,QAAM,qCAAqC,qCAAqC,MAAM;AAKtF,QAAM,OAAO,OAAO,OAAO,OAAO,WAAS,iBAAiB,KAAK,CAAC;AAElE,QAAM,yBAAyB,KAAK,IAAI,OAAK,cAAc,CAAC,CAAC;AAE7D,MAAI,KAAK;AACP,SAAK,QAAQ,WAAS;AAEpB,aAAO,YAAY,KAAK;AAAA,IAC1B,CAAC;AAAA,EACH;AAEA,SAAO;AAEP,WAAS,iBAAiB,OAAO;AAC/B,UAAM,gBAAgB,cAAc,KAAK;AACzC,QAAI,yBAAyB,SAAS,aAAa,GAAG;AACpD,aAAO;AAAA,IACT;AACA,UAAM,0BAA0B,wBAAwB,KAAK;AAC7D,QAAI,mCAAmC,SAAS,uBAAuB,GAAG;AACxE,aAAO;AAAA,IACT;AAEA,WAAO;AAAA,EACT;AACF;AAGO,gBAAS,yBAAyB,QAAQ,MAAM,MAAM;AAC3D,QAAM,aAAa,mCAAmC,QAAQ,GAAG;AAEjE,QAAM,cAAc,qBAAqB,QAAQ,GAAG;AAMpD,QAAM,gBAAgB,WAAW,OAAO,WAAW;AAEnD,SAAO;AACT;",
|
|
6
6
|
"names": ["fields", "record", "inferiorTerms"]
|
|
7
7
|
}
|
|
@@ -3,7 +3,6 @@ import { MarcRecord } from "@natlibfi/marc-record";
|
|
|
3
3
|
import validatorFactory from "./removeInferiorDataFields.js";
|
|
4
4
|
import { READERS } from "@natlibfi/fixura";
|
|
5
5
|
import generateTests from "@natlibfi/fixugen";
|
|
6
|
-
import createDebugLogger from "debug";
|
|
7
6
|
generateTests({
|
|
8
7
|
callback,
|
|
9
8
|
path: [import.meta.dirname, "..", "test-fixtures", "remove-inferior-datafields"],
|
|
@@ -18,18 +17,13 @@ generateTests({
|
|
|
18
17
|
}
|
|
19
18
|
}
|
|
20
19
|
});
|
|
21
|
-
const debug = createDebugLogger("@natlibfi/marc-record-validators-melinda/removeInferiorDataFields:test");
|
|
22
20
|
async function testValidatorFactory() {
|
|
23
21
|
const validator = await validatorFactory();
|
|
24
22
|
assert.equal(typeof validator, "object");
|
|
25
23
|
assert.equal(typeof validator.description, "string");
|
|
26
24
|
assert.equal(typeof validator.validate).to.be.a("function");
|
|
27
25
|
}
|
|
28
|
-
async function callback({ getFixture,
|
|
29
|
-
if (enabled === false) {
|
|
30
|
-
debug("TEST SKIPPED!");
|
|
31
|
-
return;
|
|
32
|
-
}
|
|
26
|
+
async function callback({ getFixture, fix = false }) {
|
|
33
27
|
const validator = await validatorFactory();
|
|
34
28
|
const record = new MarcRecord(getFixture("record.json"));
|
|
35
29
|
const expectedResult = getFixture("expectedResult.json");
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
{
|
|
2
2
|
"version": 3,
|
|
3
3
|
"sources": ["../src/removeInferiorDataFields.test.js"],
|
|
4
|
-
"sourcesContent": ["import assert from 'node:assert';\nimport {MarcRecord} from '@natlibfi/marc-record';\nimport validatorFactory from './removeInferiorDataFields.js';\nimport {READERS} from '@natlibfi/fixura';\nimport generateTests from '@natlibfi/fixugen';\
|
|
5
|
-
"mappings": "AAAA,OAAO,YAAY;AACnB,SAAQ,kBAAiB;AACzB,OAAO,sBAAsB;AAC7B,SAAQ,eAAc;AACtB,OAAO,mBAAmB;
|
|
4
|
+
"sourcesContent": ["import assert from 'node:assert';\nimport {MarcRecord} from '@natlibfi/marc-record';\nimport validatorFactory from './removeInferiorDataFields.js';\nimport {READERS} from '@natlibfi/fixura';\nimport generateTests from '@natlibfi/fixugen';\n//import createDebugLogger from 'debug';\n\ngenerateTests({\n callback,\n path: [import.meta.dirname, '..', 'test-fixtures', 'remove-inferior-datafields'],\n useMetadataFile: true,\n recurse: false,\n fixura: {\n reader: READERS.JSON\n },\n hooks: {\n before: async () => {\n testValidatorFactory();\n }\n }\n});\n\n//const debug = createDebugLogger('@natlibfi/marc-record-validators-melinda/removeInferiorDataFields:test');\n\nasync function testValidatorFactory() {\n const validator = await validatorFactory();\n\n assert.equal(typeof validator, 'object');\n assert.equal(typeof validator.description, 'string');\n assert.equal(typeof validator.validate).to.be.a('function');\n}\n\nasync function callback({getFixture, fix = false}) {\n const validator = await validatorFactory();\n const record = new MarcRecord(getFixture('record.json'));\n const expectedResult = getFixture('expectedResult.json');\n // console.log(expectedResult); // eslint-disable-line\n\n if (!fix) {\n const result = await validator.validate(record);\n assert.deepEqual(result, expectedResult);\n return;\n }\n\n await validator.fix(record);\n assert.deepEqual(record, expectedResult);\n}\n"],
|
|
5
|
+
"mappings": "AAAA,OAAO,YAAY;AACnB,SAAQ,kBAAiB;AACzB,OAAO,sBAAsB;AAC7B,SAAQ,eAAc;AACtB,OAAO,mBAAmB;AAG1B,cAAc;AAAA,EACZ;AAAA,EACA,MAAM,CAAC,YAAY,SAAS,MAAM,iBAAiB,4BAA4B;AAAA,EAC/E,iBAAiB;AAAA,EACjB,SAAS;AAAA,EACT,QAAQ;AAAA,IACN,QAAQ,QAAQ;AAAA,EAClB;AAAA,EACA,OAAO;AAAA,IACL,QAAQ,YAAY;AAClB,2BAAqB;AAAA,IACvB;AAAA,EACF;AACF,CAAC;AAID,eAAe,uBAAuB;AACpC,QAAM,YAAY,MAAM,iBAAiB;AAEzC,SAAO,MAAM,OAAO,WAAW,QAAQ;AACvC,SAAO,MAAM,OAAO,UAAU,aAAa,QAAQ;AACnD,SAAO,MAAM,OAAO,UAAU,QAAQ,EAAE,GAAG,GAAG,EAAE,UAAU;AAC5D;AAEA,eAAe,SAAS,EAAC,YAAY,MAAM,MAAK,GAAG;AACjD,QAAM,YAAY,MAAM,iBAAiB;AACzC,QAAM,SAAS,IAAI,WAAW,WAAW,aAAa,CAAC;AACvD,QAAM,iBAAiB,WAAW,qBAAqB;AAGvD,MAAI,CAAC,KAAK;AACR,UAAM,SAAS,MAAM,UAAU,SAAS,MAAM;AAC9C,WAAO,UAAU,QAAQ,cAAc;AACvC;AAAA,EACF;AAEA,QAAM,UAAU,IAAI,MAAM;AAC1B,SAAO,UAAU,QAAQ,cAAc;AACzC;",
|
|
6
6
|
"names": []
|
|
7
7
|
}
|
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
import createDebugLogger from "debug";
|
|
2
|
-
import { fieldHasSubfield, fieldToString, nvdebug
|
|
2
|
+
import { fieldHasSubfield, fieldToString, nvdebug } from "./utils.js";
|
|
3
3
|
import { fieldHasWantedTagAndOccurrenceNumber, isValidSubfield6, subfield6GetOccurrenceNumber, subfield6ResetOccurrenceNumber } from "./subfield6Utils.js";
|
|
4
4
|
const debug = createDebugLogger("@natlibfi/marc-record-validators-melinda:resolveOrphanedSubfield6s");
|
|
5
5
|
export default function() {
|
|
@@ -53,7 +53,6 @@ function findPairForSubfield6OccurrenceNumber(subfield6, myTag, candPairFields)
|
|
|
53
53
|
if (!isValidSubfield6(subfield6)) {
|
|
54
54
|
return void 0;
|
|
55
55
|
}
|
|
56
|
-
nvdebug(`LOOKING FOR PAIR: ${myTag} ${subfieldToString(subfield6)}`);
|
|
57
56
|
candPairFields.forEach((field) => fieldToString(field));
|
|
58
57
|
const referredTag = subfield6.value.substring(0, 3);
|
|
59
58
|
const occurrenceNumber = subfield6GetOccurrenceNumber(subfield6);
|
|
@@ -61,7 +60,6 @@ function findPairForSubfield6OccurrenceNumber(subfield6, myTag, candPairFields)
|
|
|
61
60
|
return void 0;
|
|
62
61
|
}
|
|
63
62
|
const tagAndOccurrenceNumber = `${myTag}-${occurrenceNumber}`;
|
|
64
|
-
nvdebug(`Try to find occurrence number ${tagAndOccurrenceNumber} in field ${referredTag}...`);
|
|
65
63
|
const relevantFields = candPairFields.filter((field) => field.tag === referredTag && fieldHasWantedTagAndOccurrenceNumber(field, tagAndOccurrenceNumber));
|
|
66
64
|
if (relevantFields.length === 0) {
|
|
67
65
|
return void 0;
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
{
|
|
2
2
|
"version": 3,
|
|
3
3
|
"sources": ["../src/resolveOrphanedSubfield6s.js"],
|
|
4
|
-
"sourcesContent": ["import createDebugLogger from 'debug';\nimport {fieldHasSubfield, fieldToString, nvdebug
|
|
5
|
-
"mappings": "AAAA,OAAO,uBAAuB;AAC9B,SAAQ,kBAAkB,eAAe,
|
|
4
|
+
"sourcesContent": ["import createDebugLogger from 'debug';\nimport {fieldHasSubfield, fieldToString, nvdebug /*, subfieldToString*/ } from './utils.js';\nimport {fieldHasWantedTagAndOccurrenceNumber, isValidSubfield6, subfield6GetOccurrenceNumber, subfield6ResetOccurrenceNumber} from './subfield6Utils.js';\n\n// Relocated from melinda-marc-record-merge-reducers (and renamed)\n\nconst debug = createDebugLogger('@natlibfi/marc-record-validators-melinda:resolveOrphanedSubfield6s');\n\nexport default function () {\n return {\n description: 'Remove occurrence-number-orphaned $6 subfields. In field 880, occurrence number becomes 00',\n validate, fix\n };\n\n function fix(record) {\n nvdebug('Fix SF6 orphaned occurrence numbers');\n const res = {message: [], fix: [], valid: true};\n //message.fix = [];\n\n // This can not really fail...\n recordFixSubfield6OccurrenceNumbers(record);\n\n //message.valid = !(message.message.length >= 1);\n return res;\n }\n\n function validate(record) {\n // Check max, and check number of different indexes\n nvdebug('Validate SF6 orphaned occurrence numbers', debug);\n const fieldsContainingSubfield6 = record.fields.filter(field => fieldHasSubfield(field, '6'));\n\n const orphanedFields = getOrphanedFields(fieldsContainingSubfield6);\n\n const res = {message: []};\n\n if (orphanedFields.length > 0) {\n res.message = [`${orphanedFields.length} orphaned occurrence number field(s) detected`];\n }\n res.valid = res.message.length < 1;\n return res;\n }\n}\n\nfunction recordFixSubfield6OccurrenceNumbers(record) {\n const fieldsContainingSubfield6 = record.fields.filter(field => fieldHasSubfield(field, '6'));\n const orphanedFields = getOrphanedFields(fieldsContainingSubfield6);\n\n orphanedFields.forEach(field => fieldFixOrphanedSubfields(field));\n\n function fieldFixOrphanedSubfields(field) {\n // Field 880: orphaned $6 subfields: occurrence number is changed to '00':\n if (field.tag === '880') {\n field.subfields.forEach(sf => field880FixOrphanedSubfield(sf));\n return;\n }\n // Non-880 fields get their orphaned $6s removed:\n const remainingSubfields = field.subfields.filter(sf => !isOrphanedSubfield(sf, field.tag, fieldsContainingSubfield6));\n if (remainingSubfields.length === 0) {\n record.removeField(field);\n return;\n }\n field.subfields = remainingSubfields;\n }\n\n function field880FixOrphanedSubfield(subfield) {\n if (!isOrphanedSubfield(subfield, '880', fieldsContainingSubfield6)) {\n return;\n }\n // convert occurrence number to 00\n subfield6ResetOccurrenceNumber(subfield, '00');\n }\n}\n\n\nfunction findPairForSubfield6OccurrenceNumber(subfield6, myTag, candPairFields) {\n // We keep the crap!\n if (!isValidSubfield6(subfield6)) {\n return undefined;\n }\n //nvdebug(`LOOKING FOR PAIR: ${myTag} ${subfieldToString(subfield6)}`);\n candPairFields.forEach(field => fieldToString(field));\n\n // Only valid $6 value that fails to map to another field is iffy...\n const referredTag = subfield6.value.substring(0, 3);\n\n const occurrenceNumber = subfield6GetOccurrenceNumber(subfield6);\n if (occurrenceNumber === '00') {\n return undefined;\n }\n const tagAndOccurrenceNumber = `${myTag}-${occurrenceNumber}`;\n //nvdebug(`Try to find occurrence number ${tagAndOccurrenceNumber} in field ${referredTag}...`);\n //const relevantFields = fields.filter(field => field.tag === referredTag && field.subfields.some(sf => subfield6GetOccurrenceNumber(sf) === occurrenceNumber));\n const relevantFields = candPairFields.filter(field => field.tag === referredTag && fieldHasWantedTagAndOccurrenceNumber(field, tagAndOccurrenceNumber));\n if (relevantFields.length === 0) {\n return undefined;\n }\n // This should always return just one (not sanity checking this for now):\n return relevantFields[0];\n}\n\nfunction isOrphanedSubfield(subfield, tag, pairCandidateFields) {\n if (!isValidSubfield6(subfield) || subfield6GetOccurrenceNumber(subfield) === '00') {\n return false;\n }\n return !findPairForSubfield6OccurrenceNumber(subfield, tag, pairCandidateFields);\n}\n\n\nfunction isOrphanedField(field, candidatePairFields) {\n return field.subfields.some(sf => isOrphanedSubfield(sf, field.tag, candidatePairFields));\n}\n\nfunction getOrphanedFields(relevantFields) {\n return relevantFields.filter(field => isOrphanedField(field, relevantFields));\n}\n"],
|
|
5
|
+
"mappings": "AAAA,OAAO,uBAAuB;AAC9B,SAAQ,kBAAkB,eAAe,eAAsC;AAC/E,SAAQ,sCAAsC,kBAAkB,8BAA8B,sCAAqC;AAInI,MAAM,QAAQ,kBAAkB,oEAAoE;AAEpG,0BAA2B;AACzB,SAAO;AAAA,IACL,aAAa;AAAA,IACb;AAAA,IAAU;AAAA,EACZ;AAEA,WAAS,IAAI,QAAQ;AACnB,YAAQ,qCAAqC;AAC7C,UAAM,MAAM,EAAC,SAAS,CAAC,GAAG,KAAK,CAAC,GAAG,OAAO,KAAI;AAI9C,wCAAoC,MAAM;AAG1C,WAAO;AAAA,EACT;AAEA,WAAS,SAAS,QAAQ;AAExB,YAAQ,4CAA4C,KAAK;AACzD,UAAM,4BAA4B,OAAO,OAAO,OAAO,WAAS,iBAAiB,OAAO,GAAG,CAAC;AAE5F,UAAM,iBAAiB,kBAAkB,yBAAyB;AAElE,UAAM,MAAM,EAAC,SAAS,CAAC,EAAC;AAExB,QAAI,eAAe,SAAS,GAAG;AAC7B,UAAI,UAAU,CAAC,GAAG,eAAe,MAAM,+CAA+C;AAAA,IACxF;AACA,QAAI,QAAQ,IAAI,QAAQ,SAAS;AACjC,WAAO;AAAA,EACT;AACF;AAEA,SAAS,oCAAoC,QAAQ;AACnD,QAAM,4BAA4B,OAAO,OAAO,OAAO,WAAS,iBAAiB,OAAO,GAAG,CAAC;AAC5F,QAAM,iBAAiB,kBAAkB,yBAAyB;AAElE,iBAAe,QAAQ,WAAS,0BAA0B,KAAK,CAAC;AAEhE,WAAS,0BAA0B,OAAO;AAExC,QAAI,MAAM,QAAQ,OAAO;AACvB,YAAM,UAAU,QAAQ,QAAM,4BAA4B,EAAE,CAAC;AAC7D;AAAA,IACF;AAEA,UAAM,qBAAqB,MAAM,UAAU,OAAO,QAAM,CAAC,mBAAmB,IAAI,MAAM,KAAK,yBAAyB,CAAC;AACrH,QAAI,mBAAmB,WAAW,GAAG;AACnC,aAAO,YAAY,KAAK;AACxB;AAAA,IACF;AACA,UAAM,YAAY;AAAA,EACpB;AAEA,WAAS,4BAA4B,UAAU;AAC7C,QAAI,CAAC,mBAAmB,UAAU,OAAO,yBAAyB,GAAG;AACnE;AAAA,IACF;AAEA,mCAA+B,UAAU,IAAI;AAAA,EAC/C;AACF;AAGA,SAAS,qCAAqC,WAAW,OAAO,gBAAgB;AAE9E,MAAI,CAAC,iBAAiB,SAAS,GAAG;AAChC,WAAO;AAAA,EACT;AAEA,iBAAe,QAAQ,WAAS,cAAc,KAAK,CAAC;AAGpD,QAAM,cAAc,UAAU,MAAM,UAAU,GAAG,CAAC;AAElD,QAAM,mBAAmB,6BAA6B,SAAS;AAC/D,MAAI,qBAAqB,MAAM;AAC7B,WAAO;AAAA,EACT;AACA,QAAM,yBAAyB,GAAG,KAAK,IAAI,gBAAgB;AAG3D,QAAM,iBAAiB,eAAe,OAAO,WAAS,MAAM,QAAQ,eAAe,qCAAqC,OAAO,sBAAsB,CAAC;AACtJ,MAAI,eAAe,WAAW,GAAG;AAC/B,WAAO;AAAA,EACT;AAEA,SAAO,eAAe,CAAC;AACzB;AAEA,SAAS,mBAAmB,UAAU,KAAK,qBAAqB;AAC9D,MAAI,CAAC,iBAAiB,QAAQ,KAAK,6BAA6B,QAAQ,MAAM,MAAM;AAClF,WAAO;AAAA,EACT;AACA,SAAO,CAAC,qCAAqC,UAAU,KAAK,mBAAmB;AACjF;AAGA,SAAS,gBAAgB,OAAO,qBAAqB;AACnD,SAAO,MAAM,UAAU,KAAK,QAAM,mBAAmB,IAAI,MAAM,KAAK,mBAAmB,CAAC;AAC1F;AAEA,SAAS,kBAAkB,gBAAgB;AACzC,SAAO,eAAe,OAAO,WAAS,gBAAgB,OAAO,cAAc,CAAC;AAC9E;",
|
|
6
6
|
"names": []
|
|
7
7
|
}
|
|
@@ -3,7 +3,6 @@ import { MarcRecord } from "@natlibfi/marc-record";
|
|
|
3
3
|
import validatorFactory from "./resolveOrphanedSubfield6s.js";
|
|
4
4
|
import { READERS } from "@natlibfi/fixura";
|
|
5
5
|
import generateTests from "@natlibfi/fixugen";
|
|
6
|
-
import createDebugLogger from "debug";
|
|
7
6
|
generateTests({
|
|
8
7
|
callback,
|
|
9
8
|
path: [import.meta.dirname, "..", "test-fixtures", "remove-orphanded-sf6s"],
|
|
@@ -18,18 +17,13 @@ generateTests({
|
|
|
18
17
|
}
|
|
19
18
|
}
|
|
20
19
|
});
|
|
21
|
-
const debug = createDebugLogger("@natlibfi/marc-record-validators-melinda/resolveOrphanedSubfield6s:test");
|
|
22
20
|
async function testValidatorFactory() {
|
|
23
21
|
const validator = await validatorFactory();
|
|
24
22
|
assert.equal(typeof validator, "object");
|
|
25
23
|
assert.equal(typeof validator.description, "string");
|
|
26
24
|
assert.equal(typeof validator.validate, "function");
|
|
27
25
|
}
|
|
28
|
-
async function callback({ getFixture,
|
|
29
|
-
if (enabled === false) {
|
|
30
|
-
debug("TEST SKIPPED!");
|
|
31
|
-
return;
|
|
32
|
-
}
|
|
26
|
+
async function callback({ getFixture, fix = false }) {
|
|
33
27
|
const validator = await validatorFactory();
|
|
34
28
|
const record = new MarcRecord(getFixture("record.json"));
|
|
35
29
|
const expectedResult = getFixture("expectedResult.json");
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
{
|
|
2
2
|
"version": 3,
|
|
3
3
|
"sources": ["../src/resolveOrphanedSubfield6s.test.js"],
|
|
4
|
-
"sourcesContent": ["import assert from 'node:assert';\nimport {MarcRecord} from '@natlibfi/marc-record';\nimport validatorFactory from './resolveOrphanedSubfield6s.js';\nimport {READERS} from '@natlibfi/fixura';\nimport generateTests from '@natlibfi/fixugen';\
|
|
5
|
-
"mappings": "AAAA,OAAO,YAAY;AACnB,SAAQ,kBAAiB;AACzB,OAAO,sBAAsB;AAC7B,SAAQ,eAAc;AACtB,OAAO,mBAAmB;
|
|
4
|
+
"sourcesContent": ["import assert from 'node:assert';\nimport {MarcRecord} from '@natlibfi/marc-record';\nimport validatorFactory from './resolveOrphanedSubfield6s.js';\nimport {READERS} from '@natlibfi/fixura';\nimport generateTests from '@natlibfi/fixugen';\n//import createDebugLogger from 'debug';\n\ngenerateTests({\n callback,\n path: [import.meta.dirname, '..', 'test-fixtures', 'remove-orphanded-sf6s'],\n useMetadataFile: true,\n recurse: false,\n fixura: {\n reader: READERS.JSON\n },\n hooks: {\n before: async () => {\n testValidatorFactory();\n }\n }\n});\n\n//const debug = createDebugLogger('@natlibfi/marc-record-validators-melinda/resolveOrphanedSubfield6s:test');\n\nasync function testValidatorFactory() {\n const validator = await validatorFactory();\n\n assert.equal(typeof validator, 'object');\n assert.equal(typeof validator.description, 'string');\n assert.equal(typeof validator.validate, 'function');\n}\n\nasync function callback({getFixture, fix = false}) {\n const validator = await validatorFactory();\n const record = new MarcRecord(getFixture('record.json'));\n const expectedResult = getFixture('expectedResult.json');\n // console.log(expectedResult); // eslint-disable-line\n\n if (!fix) {\n const result = await validator.validate(record);\n assert.deepEqual(result, expectedResult);\n return;\n }\n\n await validator.fix(record);\n assert.deepEqual(record, expectedResult);\n}\n"],
|
|
5
|
+
"mappings": "AAAA,OAAO,YAAY;AACnB,SAAQ,kBAAiB;AACzB,OAAO,sBAAsB;AAC7B,SAAQ,eAAc;AACtB,OAAO,mBAAmB;AAG1B,cAAc;AAAA,EACZ;AAAA,EACA,MAAM,CAAC,YAAY,SAAS,MAAM,iBAAiB,uBAAuB;AAAA,EAC1E,iBAAiB;AAAA,EACjB,SAAS;AAAA,EACT,QAAQ;AAAA,IACN,QAAQ,QAAQ;AAAA,EAClB;AAAA,EACA,OAAO;AAAA,IACL,QAAQ,YAAY;AAClB,2BAAqB;AAAA,IACvB;AAAA,EACF;AACF,CAAC;AAID,eAAe,uBAAuB;AACpC,QAAM,YAAY,MAAM,iBAAiB;AAEzC,SAAO,MAAM,OAAO,WAAW,QAAQ;AACvC,SAAO,MAAM,OAAO,UAAU,aAAa,QAAQ;AACnD,SAAO,MAAM,OAAO,UAAU,UAAU,UAAU;AACpD;AAEA,eAAe,SAAS,EAAC,YAAY,MAAM,MAAK,GAAG;AACjD,QAAM,YAAY,MAAM,iBAAiB;AACzC,QAAM,SAAS,IAAI,WAAW,WAAW,aAAa,CAAC;AACvD,QAAM,iBAAiB,WAAW,qBAAqB;AAGvD,MAAI,CAAC,KAAK;AACR,UAAM,SAAS,MAAM,UAAU,SAAS,MAAM;AAC9C,WAAO,UAAU,QAAQ,cAAc;AACvC;AAAA,EACF;AAEA,QAAM,UAAU,IAAI,MAAM;AAC1B,SAAO,UAAU,QAAQ,cAAc;AACzC;",
|
|
6
6
|
"names": []
|
|
7
7
|
}
|
|
@@ -3,7 +3,6 @@ import { MarcRecord } from "@natlibfi/marc-record";
|
|
|
3
3
|
import validatorFactory from "./sanitize-vocabulary-source-codes.js";
|
|
4
4
|
import { READERS } from "@natlibfi/fixura";
|
|
5
5
|
import generateTests from "@natlibfi/fixugen";
|
|
6
|
-
import createDebugLogger from "debug";
|
|
7
6
|
generateTests({
|
|
8
7
|
callback,
|
|
9
8
|
path: [import.meta.dirname, "..", "test-fixtures", "sanitize-vocabulary-source-codes"],
|
|
@@ -18,18 +17,13 @@ generateTests({
|
|
|
18
17
|
}
|
|
19
18
|
}
|
|
20
19
|
});
|
|
21
|
-
const debug = createDebugLogger("@natlibfi/marc-record-validators-melinda/sanitize-vocabulary-source-codes:test");
|
|
22
20
|
async function testValidatorFactory() {
|
|
23
21
|
const validator = await validatorFactory();
|
|
24
22
|
assert.equal(typeof validator, "object");
|
|
25
23
|
assert.equal(typeof validator.description, "string");
|
|
26
24
|
assert.equal(typeof validator.validate, "function");
|
|
27
25
|
}
|
|
28
|
-
async function callback({ getFixture,
|
|
29
|
-
if (enabled === false) {
|
|
30
|
-
debug("TEST SKIPPED!");
|
|
31
|
-
return;
|
|
32
|
-
}
|
|
26
|
+
async function callback({ getFixture, fix = false }) {
|
|
33
27
|
const validator = await validatorFactory();
|
|
34
28
|
const recordFixture = getFixture("record.json");
|
|
35
29
|
const record = recordFixture._validationOptions ? new MarcRecord(recordFixture, recordFixture._validationOptions) : new MarcRecord(recordFixture);
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
{
|
|
2
2
|
"version": 3,
|
|
3
3
|
"sources": ["../src/sanitize-vocabulary-source-codes.test.js"],
|
|
4
|
-
"sourcesContent": ["import assert from 'node:assert';\nimport {MarcRecord} from '@natlibfi/marc-record';\nimport validatorFactory from './sanitize-vocabulary-source-codes.js';\nimport {READERS} from '@natlibfi/fixura';\nimport generateTests from '@natlibfi/fixugen';\
|
|
5
|
-
"mappings": "AAAA,OAAO,YAAY;AACnB,SAAQ,kBAAiB;AACzB,OAAO,sBAAsB;AAC7B,SAAQ,eAAc;AACtB,OAAO,mBAAmB;
|
|
4
|
+
"sourcesContent": ["import assert from 'node:assert';\nimport {MarcRecord} from '@natlibfi/marc-record';\nimport validatorFactory from './sanitize-vocabulary-source-codes.js';\nimport {READERS} from '@natlibfi/fixura';\nimport generateTests from '@natlibfi/fixugen';\n//import createDebugLogger from 'debug';\n\ngenerateTests({\n callback,\n path: [import.meta.dirname, '..', 'test-fixtures', 'sanitize-vocabulary-source-codes'],\n useMetadataFile: true,\n recurse: false,\n fixura: {\n reader: READERS.JSON\n },\n hooks: {\n before: async () => {\n testValidatorFactory();\n }\n }\n});\n//const debug = createDebugLogger('@natlibfi/marc-record-validators-melinda/sanitize-vocabulary-source-codes:test');\n\nasync function testValidatorFactory() {\n const validator = await validatorFactory();\n\n assert.equal(typeof validator, 'object');\n assert.equal(typeof validator.description, 'string');\n assert.equal(typeof validator.validate, 'function');\n}\n\nasync function callback({getFixture, fix = false}) {\n const validator = await validatorFactory();\n\n const recordFixture = getFixture('record.json');\n const record = recordFixture._validationOptions ? new MarcRecord(recordFixture, recordFixture._validationOptions) : new MarcRecord(recordFixture);\n\n const expectedResult = getFixture('expectedResult.json');\n // console.log(expectedResult); // eslint-disable-line\n\n if (!fix) {\n const result = await validator.validate(record);\n assert.deepEqual(result, expectedResult);\n return;\n }\n\n await validator.fix(record);\n assert.deepEqual(record, expectedResult);\n}\n"],
|
|
5
|
+
"mappings": "AAAA,OAAO,YAAY;AACnB,SAAQ,kBAAiB;AACzB,OAAO,sBAAsB;AAC7B,SAAQ,eAAc;AACtB,OAAO,mBAAmB;AAG1B,cAAc;AAAA,EACZ;AAAA,EACA,MAAM,CAAC,YAAY,SAAS,MAAM,iBAAiB,kCAAkC;AAAA,EACrF,iBAAiB;AAAA,EACjB,SAAS;AAAA,EACT,QAAQ;AAAA,IACN,QAAQ,QAAQ;AAAA,EAClB;AAAA,EACA,OAAO;AAAA,IACL,QAAQ,YAAY;AAClB,2BAAqB;AAAA,IACvB;AAAA,EACF;AACF,CAAC;AAGD,eAAe,uBAAuB;AACpC,QAAM,YAAY,MAAM,iBAAiB;AAEzC,SAAO,MAAM,OAAO,WAAW,QAAQ;AACvC,SAAO,MAAM,OAAO,UAAU,aAAa,QAAQ;AACnD,SAAO,MAAM,OAAO,UAAU,UAAU,UAAU;AACpD;AAEA,eAAe,SAAS,EAAC,YAAY,MAAM,MAAK,GAAG;AACjD,QAAM,YAAY,MAAM,iBAAiB;AAEzC,QAAM,gBAAgB,WAAW,aAAa;AAC9C,QAAM,SAAS,cAAc,qBAAqB,IAAI,WAAW,eAAe,cAAc,kBAAkB,IAAI,IAAI,WAAW,aAAa;AAEhJ,QAAM,iBAAiB,WAAW,qBAAqB;AAGvD,MAAI,CAAC,KAAK;AACR,UAAM,SAAS,MAAM,UAAU,SAAS,MAAM;AAC9C,WAAO,UAAU,QAAQ,cAAc;AACvC;AAAA,EACF;AAEA,QAAM,UAAU,IAAI,MAAM;AAC1B,SAAO,UAAU,QAAQ,cAAc;AACzC;",
|
|
6
6
|
"names": []
|
|
7
7
|
}
|
package/dist/sortFields.js
CHANGED
|
@@ -197,7 +197,7 @@ export function scoreRelatorTerm(value, typeOfMaterial = void 0) {
|
|
|
197
197
|
return 0;
|
|
198
198
|
}
|
|
199
199
|
export function fieldOrderComparator(fieldA, fieldB) {
|
|
200
|
-
const sorterFunctions = [sortByTag, sortByIndexTerms, sortAlphabetically, sortByRelatorTerm, sortBySubfield6, preferFenniKeep, sortByFieldLinkAndSequenceNumber];
|
|
200
|
+
const sorterFunctions = [sortByTag, sortByIndexTerms, sortAlphabetically, sortByRelatorTerm, sortBySubfield6, preferFenniKeep, sortByFieldLinkAndSequenceNumber, sortByHacks];
|
|
201
201
|
return globalFieldOrderComparator(fieldA, fieldB, sorterFunctions);
|
|
202
202
|
}
|
|
203
203
|
function sortByIndexTerms(fieldA, fieldB) {
|
|
@@ -335,6 +335,21 @@ function sortByFieldLinkAndSequenceNumber(fieldA, fieldB) {
|
|
|
335
335
|
}
|
|
336
336
|
return -1;
|
|
337
337
|
}
|
|
338
|
+
function sortByHacks(fieldA, fieldB) {
|
|
339
|
+
if (fieldA.tag === fieldB.tag) {
|
|
340
|
+
if (fieldA.tag === "041") {
|
|
341
|
+
const a2 = fieldA.subfields.some((sf) => sf.code === "2");
|
|
342
|
+
const b2 = fieldB.subfields.some((sf) => sf.code === "2");
|
|
343
|
+
if (a2 && !b2) {
|
|
344
|
+
return 1;
|
|
345
|
+
}
|
|
346
|
+
if (b2 && !a2) {
|
|
347
|
+
return -1;
|
|
348
|
+
}
|
|
349
|
+
}
|
|
350
|
+
}
|
|
351
|
+
return 0;
|
|
352
|
+
}
|
|
338
353
|
function sortBySubfield6(fieldA, fieldB) {
|
|
339
354
|
if (fieldA.tag !== "880" || fieldB.tag !== "880") {
|
|
340
355
|
return 0;
|
package/dist/sortFields.js.map
CHANGED
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
{
|
|
2
2
|
"version": 3,
|
|
3
3
|
"sources": ["../src/sortFields.js"],
|
|
4
|
-
"sourcesContent": ["// Taken from project marc-record-js, file marcSortFields.js as this contains more and more Melinda-specific rules.\n\nimport clone from 'clone';\n//import createDebugLogger from 'debug';\nimport {fieldHasSubfield, fieldToString} from './utils.js';\nimport {sortByTag, sortAlphabetically, fieldOrderComparator as globalFieldOrderComparator} from '@natlibfi/marc-record/dist/marcFieldSort.js';\nimport {isValidSubfield8} from './subfield8Utils.js';\nimport {fieldGetUnambiguousOccurrenceNumber, fieldGetUnambiguousTag} from './subfield6Utils.js';\n\n//const debug = createDebugLogger('@natlibfi/marc-record-validators-melinda:sortFields');\n//const debugData = debug.extend('data');\n//const debugDev = debug.extend('dev');\n\nconst BIG_BAD_NUMBER = 999999999;\nexport default function () {\n\n return {\n description: 'Sort fields using both generic and Melinda specific rules',\n validate, fix\n };\n\n function fix(record) {\n const res = {message: [], fix: [], valid: true};\n\n record.fields.sort(fieldOrderComparator);\n\n return res;\n }\n\n function validate(record) {\n const res = {message: []};\n\n const fields = record.fields.map(f => clone(f));\n fields.sort(fieldOrderComparator);\n\n\n const relocatedFields = fields.filter((f, i) => fieldToString(f) !== fieldToString(record.fields[i]));\n\n if (relocatedFields.length > 0) {\n res.message.push(`${relocatedFields.length} field(s) in new places`);\n }\n\n res.valid = !(res.message.length >= 1);\n return res;\n }\n}\n\n\nconst relatorTermScore = { // Here bigger is better\n // NB! This is exportable as field internal $e sorting in marc-record-validators-js uses this.\n // NB! The more abstract, the earlier it appears.\n // Note that terms with same abstraction level might also have order preferences\n // We should 1) check the order of these, and 2) add translations (support Swedish at the very least)\n // work/teos > expression/ekspressio > manifestation/manifestaatin,\n // Work https://finto.fi/mts/fi/page/m1298\n 's\u00E4velt\u00E4j\u00E4': 100, 'composer': 100,\n 'kirjoittaja': 100, 'author': 100,\n 'libretisti': 100,\n 'sarjakuvantekij\u00E4': 100, 'soitonoppaan tekij\u00E4': 100,\n 'kartantekij\u00E4': 99,\n 'taiteilija': 98,\n 'designer': 90,\n 'sanoittaja': 90,\n 'k\u00E4sikirjoittaja': 90,\n\n 'kuvaaja': 89, 'valokuvaaja': 89,\n 'kokoaja': 86,\n 'alkuper\u00E4isidean luoja': 85,\n 'teoksella kunnioitettu': 84, 'gratulaation kohde': 84,\n 'julkaisija': 82,\n 'tuottaja': 81,\n // expression: https://finto.fi/mts/fi/page/m153\n 'sovittaja': 79, 'arranger': 79,\n 'toimittaja': 78, 'editor': 78,\n 'kuvittaja': 77,\n 'esipuheen kirjoittaja': 76,\n 'alkusanojen kirjoittaja': 75, 'loppusanojen kirjoittaja': 75,\n\n 'esitt\u00E4j\u00E4': 74,\n 'johtaja': 73, // orkesterinjohtaja\n 'editointi': 71, // for music, editor/toimittaja is another thing\n 'k\u00E4\u00E4nt\u00E4j\u00E4': 70,\n 'lukija': 61, 'kertoja': 61,\n // Manifestation level: https://finto.fi/mts/fi/page/m491\n 'graafinen suunnittelija': 50,\n 'kustantaja': 41\n // Item level: https://finto.fi/mts/fi/page/m1157\n};\n\nconst relatorTermScoreBk = {\n // https://finto.fi/mts/fi/page/m34 100-81\n 'libretisti': 100, 'sarjakuvantekij\u00E4': 100,\n 'kirjoittaja': 99, 'author': 99, 'soitonoppaan tekij\u00E4': 99,\n 'kuvaaja': 98, 'valokuvaaja': 98,\n 'kokoaja': 86, 'designer': 86,\n 'alkuper\u00E4isidean luoja': 85,\n 'teoksella kunnioitettu': 84, 'gratulaation kohde': 84,\n 's\u00E4velt\u00E4j\u00E4': 83, // if 300$e has CD etc\n 'sanoittaja': 82,\n 'julkaisija': 81,\n // expression: https://finto.fi/mts/fi/page/m153\n 'toimittaja': 78,\n 'kuvittaja': 77,\n 'esipuheen kirjoittaja': 76,\n 'alkusanojen kirjoittaja': 75, 'loppusanojen kirjoittaja': 75,\n 'k\u00E4\u00E4nt\u00E4j\u00E4': 70,\n 'sovittaja': 50,\n // manifestaatio\n 'graafinen suunnittelija': 40,\n // kappale/item\n 'sitoja': 25,\n 'gratulaation kirjoittaja': 24\n};\n\nconst relatorTermScoreMu = {\n 's\u00E4velt\u00E4j\u00E4': 100,\n 'sanoittaja': 99,\n 'soitonoppaan tekij\u00E4': 98,\n 'alkuper\u00E4isidean luoja': 85,\n 'teoksella kunnioitettu': 81,\n // expression: https://finto.fi/mts/fi/page/m153\n 'sovittaja': 79,\n 'johtaja': 78,\n 'esitt\u00E4j\u00E4': 77,\n 'lukija': 76,\n 'miksaaja': 75,\n // manifestaatio\n 'esitt\u00E4j\u00E4 (manifestaatio)': 69,\n\n 'graafinen suunnittelija': 50,\n 'valmistaja': 41,\n 'jakaja': 40\n // kappale/item\n\n};\n\nconst relatorTermScoreVm = { // Visual Material\n // Work\n 'ohjaaja': 100,\n 'kirjoittaja': 99, 'author': 99, // Here we assume that film/whatever is based on author's book\n 'k\u00E4sikirjoittaja': 98, 'designer': 98,\n 'kuvaaja': 89,\n 's\u00E4velt\u00E4j\u00E4': 86, // Volatile. John Williams?\n 'alkuper\u00E4isidean luoja': 85,\n 'julkaisija': 82,\n 'tuottaja': 81,\n // Expression\n 'leikkaaja': 80,\n 'sovittaja': 79,\n 'johtaja': 78,\n 'esitt\u00E4j\u00E4': 77,\n 'lukija': 76,\n 'miksaaja': 75,\n 'k\u00E4\u00E4nt\u00E4j\u00E4': 70,\n // Manifestation\n 'kaivertaja': 60,\n 'graafinen suunnittelija': 59,\n 'kustantaja': 42,\n 'elokuvan jakelija': 41, 'jakaja': 41\n\n // Item\n};\n\nfunction normalizeValue(value) {\n // Removing last punc char is good enough for our purposes.\n // We don't handle abbreviations here etc.\n // Brackets should not happen either, should they?\n return value.replace(/[.,]$/u, '');\n}\n\nfunction scoreRelatorTermBk(normalizedValue) {\n if (normalizedValue in relatorTermScoreBk) {\n return relatorTermScoreBk[normalizedValue];\n }\n return 0;\n}\n\nfunction scoreRelatorTermMu(normalizedValue) {\n if (normalizedValue in relatorTermScoreMu) {\n return relatorTermScoreMu[normalizedValue];\n }\n return 0;\n}\n\nfunction scoreRelatorTermVm(normalizedValue) {\n if (normalizedValue in relatorTermScoreVm) {\n return relatorTermScoreVm[normalizedValue];\n }\n return 0;\n}\n\nexport function scoreRelatorTerm(value, typeOfMaterial = undefined) {\n // NB! We are currently using type of material only for sorting relator terms, not 7XX fields!!!\n const normalizedValue = normalizeValue(value);\n if (typeOfMaterial === 'BK') { // Books\n return scoreRelatorTermBk(normalizedValue);\n }\n if (typeOfMaterial === 'MU') { // Music (NB: audio books should be BK in this context!)\n return scoreRelatorTermMu(normalizedValue);\n }\n if (typeOfMaterial === 'VM') { // video material\n return scoreRelatorTermVm(normalizedValue);\n }\n if (normalizedValue in relatorTermScore) {\n return relatorTermScore[normalizedValue];\n }\n return 0;\n}\n\nexport function fieldOrderComparator(fieldA, fieldB) {\n\n //const sorterFunctions = [sortByTag, sortByIndexTerms, sortAlphabetically, sortByRelatorTerm, sortByOccurrenceNumber, preferFenniKeep, sortByFieldLinkAndSequenceNumber];\n\n const sorterFunctions = [sortByTag, sortByIndexTerms, sortAlphabetically, sortByRelatorTerm, sortBySubfield6, preferFenniKeep, sortByFieldLinkAndSequenceNumber];\n //const sorterFunctions = [sortByIndexTerms, sortByRelatorTerm, sortByOccurrenceNumber, preferFenniKeep, sortByFieldLinkAndSequenceNumber];\n\n return globalFieldOrderComparator(fieldA, fieldB, sorterFunctions);\n}\n\nfunction sortByIndexTerms(fieldA, fieldB) {\n\n const indexTermFields = ['600', '610', '611', '630', '648', '650', '651', '652', '653', '654', '655', '656', '657', '658', '659', '662'];\n\n function scoreInd2(val) {\n const ind2Score = {'0': 0, '1': 1, '2': 2, '3': 3, '4': 8, '5': 5, '6': 6, '7': 7};\n\n if (val in ind2Score) {\n return ind2Score[val];\n }\n return 9;\n }\n\n // ATM this is not needed.\n // You may need this, if you change compare function order in sorterFunctions\n // istanbul ignore next\n if (fieldA.tag !== fieldB.tag) {\n return 0;\n }\n\n if (!indexTermFields.includes(fieldA.tag)) {\n return 0;\n }\n\n // Puts ind2=4 last\n if (scoreInd2(fieldA.ind2) > scoreInd2(fieldB.ind2)) {\n return 1;\n }\n if (scoreInd2(fieldA.ind2) < scoreInd2(fieldB.ind2)) {\n return -1;\n }\n\n function scoreDictionary(dictionary) {\n const dictionarySortIndex = {\n 'yso/fin': 0,\n 'yso/swe': 1,\n 'yso/eng': 2,\n 'slm/fin': 0.1,\n 'slm/swe': 1.1,\n 'kauno/fin': 2.1,\n 'kauno/swe': 2.2,\n 'kaunokki': 4,\n 'bella': 5\n };\n\n if (dictionary in dictionarySortIndex) {\n return dictionarySortIndex[dictionary];\n }\n return BIG_BAD_NUMBER;\n }\n\n const dictionaryA = selectFirstValue(fieldA, '2');\n const dictionaryB = selectFirstValue(fieldB, '2');\n\n const dictScoreA = scoreDictionary(dictionaryA);\n const dictScoreB = scoreDictionary(dictionaryB);\n // Use priority order for listed dictionaries:\n if (dictScoreA > dictScoreB) {\n return 1;\n }\n if (dictScoreA < dictScoreB) {\n return -1;\n }\n // Unlisted dictionaries: sort $2 value alphabetically:\n //if (dictScoreA === BIG_BAD_NUMBER) {\n if (dictionaryA > dictionaryB) {\n return 1;\n }\n if (dictionaryA < dictionaryB) {\n return -1;\n }\n //}\n return 0;\n}\n\n\nfunction preferKeep(fieldA, fieldB, keepOwner = 'FENNI') {\n const hasKeepA = fieldHasSubfield(fieldA, '9', `${keepOwner}<KEEP>`);\n const hasKeepB = fieldHasSubfield(fieldB, '9', `${keepOwner}<KEEP>`);\n\n if (hasKeepA && !hasKeepB) {\n return -1;\n }\n if (!hasKeepA && hasKeepB) {\n return 1;\n }\n\n return 0;\n}\n\n\nfunction preferFenniKeep(fieldA, fieldB) {\n const fenniPreference = preferKeep(fieldA, fieldB, 'FENNI');\n if (fenniPreference !== 0) {\n return fenniPreference;\n }\n const violaPreference = preferKeep(fieldA, fieldB, 'VIOLA');\n if (violaPreference !== 0) {\n return violaPreference;\n }\n return preferKeep(fieldA, fieldB, 'FIKKA');\n}\n\nfunction sortByRelatorTerm(fieldA, fieldB) {\n //if (!['600', '610', '611', '630', '700', '710', '711', '730', '800', '810', '811', '830'].includes(fieldA.tag)) {\n if (!['700', '710', '711', '730'].includes(fieldA.tag)) {\n return 0;\n }\n\n function fieldGetMaxRelatorTermScore(field) {\n if (!field.subfields) {\n return -2;\n }\n // If field has $t, it's a teos-nimeke-auktoriteetti, and thus meaningless. These should follow all $t-less fields...\n if (fieldHasSubfield(field, 't')) {\n return -2;\n }\n const relatorSubfieldCode = ['611', '711', '811'].includes(field.tag) ? 'j' : 'e';\n const e = field.subfields.filter(sf => sf.code === relatorSubfieldCode);\n if (e.length === 0) { // No $e is still better than having a $t\n return -1;\n }\n const scores = e.map(sf => scoreRelatorTerm(sf.value));\n //debugDev(`RELATOR SCORE FOR '${fieldToString(field)}': ${scores.join(', ')}`);\n return Math.max(...scores);\n }\n\n const scoreA = fieldGetMaxRelatorTermScore(fieldA);\n const scoreB = fieldGetMaxRelatorTermScore(fieldB);\n\n if (scoreA < scoreB) {\n return 1;\n }\n if (scoreA > scoreB) {\n return -1;\n }\n return 0;\n}\n\n\nfunction fieldGetMinLinkAndSequenceNumber(field) {\n if (!field.subfields) {\n return BIG_BAD_NUMBER;\n }\n const relevantSubfields = field.subfields.filter(sf => isValidSubfield8(sf));\n // If val is something like \"1.2\\x\" parseFloat() would give a syntax erro because of hex-like escape sequnce (at least on Chrome). Thus remove tail:\n const scores = relevantSubfields.map(sf => parseFloat(sf.value.replace(/\\\\.*$/u, '')));\n if (scores.length === 0) {\n return BIG_BAD_NUMBER;\n }\n return Math.min(...scores);\n}\n\nfunction sortByFieldLinkAndSequenceNumber(fieldA, fieldB) { // Sort by subfield $8 that is...\n const scoreA = fieldGetMinLinkAndSequenceNumber(fieldA);\n const scoreB = fieldGetMinLinkAndSequenceNumber(fieldB);\n //debugDev(` sf-8-A-score for '${fieldToString(fieldA)}: ${scoreA}`);\n //debugDev(` sf-8-B-score for '${fieldToString(fieldB)}: ${scoreB}`);\n if (scoreA === scoreB) {\n return 0;\n }\n if (scoreB === 0) {\n return 1;\n }\n if (scoreA === 0) {\n return -1;\n }\n if (scoreA > scoreB) { // smaller is better\n return 1;\n }\n return -1;\n}\n\n\nfunction sortBySubfield6(fieldA, fieldB) { // Sort by subfield $6, ex-sortByOccurrenceNumber...\n if (fieldA.tag !== '880' || fieldB.tag !== '880') {\n return 0;\n }\n\n function compareLinkingTags() {\n const tagStringA = fieldGetUnambiguousTag(fieldA);\n const tagStringB = fieldGetUnambiguousTag(fieldB);\n if (tagStringA === tagStringB || !tagStringA || !tagStringB) {\n return 0;\n }\n if (tagStringA > tagStringB) {\n return 1;\n }\n return -1;\n }\n\n function compareOccurrenceNumbers() {\n const stringA = fieldGetUnambiguousOccurrenceNumber(fieldA);\n const stringB = fieldGetUnambiguousOccurrenceNumber(fieldB);\n if (stringA === stringB) { // No action required here\n return 0;\n }\n\n // Handle expections: no occurrence number, occurrence number '00':\n if (!stringB || stringB === '00') {\n if (!stringA || stringA === '00') {\n return 0;\n }\n return -1;\n }\n if (!stringA || stringA === '00') {\n return 1;\n }\n\n // NB! We need compare ints as occurrence number can exceed 99 and be a three-digit value!\n const scoreA = parseInt(stringA, 10);\n const scoreB = parseInt(stringB, 10);\n\n if (scoreA > scoreB) { // smaller is better, thus '00' is the best\n return 1;\n }\n return -1;\n }\n\n const linkingTagComparisonResult = compareLinkingTags();\n if (linkingTagComparisonResult !== 0) {\n return linkingTagComparisonResult;\n }\n\n return compareOccurrenceNumbers();\n}\n\n\nfunction selectFirstValue(field, subcode) {\n return field.subfields\n .filter(subfield => subcode === subfield.code)\n .map(subfield => subfield.value)\n .slice(0, 1);\n}\n\n"],
|
|
5
|
-
"mappings": "AAEA,OAAO,WAAW;AAElB,SAAQ,kBAAkB,qBAAoB;AAC9C,SAAQ,WAAW,oBAAoB,wBAAwB,kCAAiC;AAChG,SAAQ,wBAAuB;AAC/B,SAAQ,qCAAqC,8BAA6B;AAM1E,MAAM,iBAAiB;AACvB,0BAA2B;AAEzB,SAAO;AAAA,IACL,aAAa;AAAA,IACb;AAAA,IAAU;AAAA,EACZ;AAEA,WAAS,IAAI,QAAQ;AACnB,UAAM,MAAM,EAAC,SAAS,CAAC,GAAG,KAAK,CAAC,GAAG,OAAO,KAAI;AAE9C,WAAO,OAAO,KAAK,oBAAoB;AAEvC,WAAO;AAAA,EACT;AAEA,WAAS,SAAS,QAAQ;AACxB,UAAM,MAAM,EAAC,SAAS,CAAC,EAAC;AAExB,UAAM,SAAS,OAAO,OAAO,IAAI,OAAK,MAAM,CAAC,CAAC;AAC9C,WAAO,KAAK,oBAAoB;AAGhC,UAAM,kBAAkB,OAAO,OAAO,CAAC,GAAG,MAAM,cAAc,CAAC,MAAM,cAAc,OAAO,OAAO,CAAC,CAAC,CAAC;AAEpG,QAAI,gBAAgB,SAAS,GAAG;AAC9B,UAAI,QAAQ,KAAK,GAAG,gBAAgB,MAAM,yBAAyB;AAAA,IACrE;AAEA,QAAI,QAAQ,EAAE,IAAI,QAAQ,UAAU;AACpC,WAAO;AAAA,EACT;AACF;AAGA,MAAM,mBAAmB;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAOvB,sBAAa;AAAA,EAAK,YAAY;AAAA,EAC9B,eAAe;AAAA,EAAK,UAAU;AAAA,EAC9B,cAAc;AAAA,EACd,uBAAoB;AAAA,EAAK,0BAAuB;AAAA,EAChD,mBAAgB;AAAA,EAChB,cAAc;AAAA,EACd,YAAY;AAAA,EACZ,cAAc;AAAA,EACd,sBAAmB;AAAA,EAEnB,WAAW;AAAA,EAAI,eAAe;AAAA,EAC9B,WAAW;AAAA,EACX,4BAAyB;AAAA,EACzB,0BAA0B;AAAA,EAAI,sBAAsB;AAAA,EACpD,cAAc;AAAA,EACd,YAAY;AAAA;AAAA,EAEZ,aAAa;AAAA,EAAI,YAAY;AAAA,EAC7B,cAAc;AAAA,EAAI,UAAU;AAAA,EAC5B,aAAa;AAAA,EACb,yBAAyB;AAAA,EACzB,2BAA2B;AAAA,EAAI,4BAA4B;AAAA,EAE3D,kBAAY;AAAA,EACZ,WAAW;AAAA;AAAA,EACX,aAAa;AAAA;AAAA,EACb,wBAAY;AAAA,EACZ,UAAU;AAAA,EAAI,WAAW;AAAA;AAAA,EAEzB,2BAA2B;AAAA,EAC3B,cAAc;AAAA;AAEhB;AAEA,MAAM,qBAAqB;AAAA;AAAA,EAEzB,cAAc;AAAA,EAAK,uBAAoB;AAAA,EACvC,eAAe;AAAA,EAAI,UAAU;AAAA,EAAI,0BAAuB;AAAA,EACxD,WAAW;AAAA,EAAI,eAAe;AAAA,EAC9B,WAAW;AAAA,EAAI,YAAY;AAAA,EAC3B,4BAAyB;AAAA,EACzB,0BAA0B;AAAA,EAAI,sBAAsB;AAAA,EACpD,sBAAa;AAAA;AAAA,EACb,cAAc;AAAA,EACd,cAAc;AAAA;AAAA,EAEd,cAAc;AAAA,EACd,aAAa;AAAA,EACb,yBAAyB;AAAA,EACzB,2BAA2B;AAAA,EAAI,4BAA4B;AAAA,EAC3D,wBAAY;AAAA,EACZ,aAAa;AAAA;AAAA,EAEb,2BAA2B;AAAA;AAAA,EAE3B,UAAU;AAAA,EACV,4BAA4B;AAC9B;AAEA,MAAM,qBAAqB;AAAA,EACzB,sBAAa;AAAA,EACb,cAAc;AAAA,EACd,0BAAuB;AAAA,EACvB,4BAAyB;AAAA,EACzB,0BAA0B;AAAA;AAAA,EAE1B,aAAa;AAAA,EACb,WAAW;AAAA,EACX,kBAAY;AAAA,EACZ,UAAU;AAAA,EACV,YAAY;AAAA;AAAA,EAEZ,kCAA4B;AAAA,EAE5B,2BAA2B;AAAA,EAC3B,cAAc;AAAA,EACd,UAAU;AAAA;AAGZ;AAEA,MAAM,qBAAqB;AAAA;AAAA;AAAA,EAEzB,WAAW;AAAA,EACX,eAAe;AAAA,EAAI,UAAU;AAAA;AAAA,EAC7B,sBAAmB;AAAA,EAAI,YAAY;AAAA,EACnC,WAAW;AAAA,EACX,sBAAa;AAAA;AAAA,EACb,4BAAyB;AAAA,EACzB,cAAc;AAAA,EACd,YAAY;AAAA;AAAA,EAEZ,aAAa;AAAA,EACb,aAAa;AAAA,EACb,WAAW;AAAA,EACX,kBAAY;AAAA,EACZ,UAAU;AAAA,EACV,YAAY;AAAA,EACZ,wBAAY;AAAA;AAAA,EAEZ,cAAc;AAAA,EACd,2BAA2B;AAAA,EAC3B,cAAc;AAAA,EACd,qBAAqB;AAAA,EAAI,UAAU;AAAA;AAGrC;AAEA,SAAS,eAAe,OAAO;AAI7B,SAAO,MAAM,QAAQ,UAAU,EAAE;AACnC;AAEA,SAAS,mBAAmB,iBAAiB;AAC3C,MAAI,mBAAmB,oBAAoB;AACzC,WAAO,mBAAmB,eAAe;AAAA,EAC3C;AACA,SAAO;AACT;AAEA,SAAS,mBAAmB,iBAAiB;AAC3C,MAAI,mBAAmB,oBAAoB;AACzC,WAAO,mBAAmB,eAAe;AAAA,EAC3C;AACA,SAAO;AACT;AAEA,SAAS,mBAAmB,iBAAiB;AAC3C,MAAI,mBAAmB,oBAAoB;AACzC,WAAO,mBAAmB,eAAe;AAAA,EAC3C;AACA,SAAO;AACT;AAEO,gBAAS,iBAAiB,OAAO,iBAAiB,QAAW;AAElE,QAAM,kBAAkB,eAAe,KAAK;AAC5C,MAAI,mBAAmB,MAAM;AAC3B,WAAO,mBAAmB,eAAe;AAAA,EAC3C;AACA,MAAI,mBAAmB,MAAM;AAC3B,WAAO,mBAAmB,eAAe;AAAA,EAC3C;AACA,MAAI,mBAAmB,MAAM;AAC3B,WAAO,mBAAmB,eAAe;AAAA,EAC3C;AACA,MAAI,mBAAmB,kBAAkB;AACvC,WAAO,iBAAiB,eAAe;AAAA,EACzC;AACA,SAAO;AACT;AAEO,gBAAS,qBAAqB,QAAQ,QAAQ;AAInD,QAAM,kBAAkB,CAAC,WAAW,kBAAkB,oBAAoB,mBAAmB,iBAAiB,iBAAiB,
|
|
4
|
+
"sourcesContent": ["// Taken from project marc-record-js, file marcSortFields.js as this contains more and more Melinda-specific rules.\n\nimport clone from 'clone';\n//import createDebugLogger from 'debug';\nimport {fieldHasSubfield, fieldToString} from './utils.js';\nimport {sortByTag, sortAlphabetically, fieldOrderComparator as globalFieldOrderComparator} from '@natlibfi/marc-record/dist/marcFieldSort.js';\nimport {isValidSubfield8} from './subfield8Utils.js';\nimport {fieldGetUnambiguousOccurrenceNumber, fieldGetUnambiguousTag} from './subfield6Utils.js';\n\n//const debug = createDebugLogger('@natlibfi/marc-record-validators-melinda:sortFields');\n//const debugData = debug.extend('data');\n//const debugDev = debug.extend('dev');\n\nconst BIG_BAD_NUMBER = 999999999;\nexport default function () {\n\n return {\n description: 'Sort fields using both generic and Melinda specific rules',\n validate, fix\n };\n\n function fix(record) {\n const res = {message: [], fix: [], valid: true};\n\n record.fields.sort(fieldOrderComparator);\n\n return res;\n }\n\n function validate(record) {\n const res = {message: []};\n\n const fields = record.fields.map(f => clone(f));\n fields.sort(fieldOrderComparator);\n\n\n const relocatedFields = fields.filter((f, i) => fieldToString(f) !== fieldToString(record.fields[i]));\n\n if (relocatedFields.length > 0) {\n res.message.push(`${relocatedFields.length} field(s) in new places`);\n }\n\n res.valid = !(res.message.length >= 1);\n return res;\n }\n}\n\n\nconst relatorTermScore = { // Here bigger is better\n // NB! This is exportable as field internal $e sorting in marc-record-validators-js uses this.\n // NB! The more abstract, the earlier it appears.\n // Note that terms with same abstraction level might also have order preferences\n // We should 1) check the order of these, and 2) add translations (support Swedish at the very least)\n // work/teos > expression/ekspressio > manifestation/manifestaatin,\n // Work https://finto.fi/mts/fi/page/m1298\n 's\u00E4velt\u00E4j\u00E4': 100, 'composer': 100,\n 'kirjoittaja': 100, 'author': 100,\n 'libretisti': 100,\n 'sarjakuvantekij\u00E4': 100, 'soitonoppaan tekij\u00E4': 100,\n 'kartantekij\u00E4': 99,\n 'taiteilija': 98,\n 'designer': 90,\n 'sanoittaja': 90,\n 'k\u00E4sikirjoittaja': 90,\n\n 'kuvaaja': 89, 'valokuvaaja': 89,\n 'kokoaja': 86,\n 'alkuper\u00E4isidean luoja': 85,\n 'teoksella kunnioitettu': 84, 'gratulaation kohde': 84,\n 'julkaisija': 82,\n 'tuottaja': 81,\n // expression: https://finto.fi/mts/fi/page/m153\n 'sovittaja': 79, 'arranger': 79,\n 'toimittaja': 78, 'editor': 78,\n 'kuvittaja': 77,\n 'esipuheen kirjoittaja': 76,\n 'alkusanojen kirjoittaja': 75, 'loppusanojen kirjoittaja': 75,\n\n 'esitt\u00E4j\u00E4': 74,\n 'johtaja': 73, // orkesterinjohtaja\n 'editointi': 71, // for music, editor/toimittaja is another thing\n 'k\u00E4\u00E4nt\u00E4j\u00E4': 70,\n 'lukija': 61, 'kertoja': 61,\n // Manifestation level: https://finto.fi/mts/fi/page/m491\n 'graafinen suunnittelija': 50,\n 'kustantaja': 41\n // Item level: https://finto.fi/mts/fi/page/m1157\n};\n\nconst relatorTermScoreBk = {\n // https://finto.fi/mts/fi/page/m34 100-81\n 'libretisti': 100, 'sarjakuvantekij\u00E4': 100,\n 'kirjoittaja': 99, 'author': 99, 'soitonoppaan tekij\u00E4': 99,\n 'kuvaaja': 98, 'valokuvaaja': 98,\n 'kokoaja': 86, 'designer': 86,\n 'alkuper\u00E4isidean luoja': 85,\n 'teoksella kunnioitettu': 84, 'gratulaation kohde': 84,\n 's\u00E4velt\u00E4j\u00E4': 83, // if 300$e has CD etc\n 'sanoittaja': 82,\n 'julkaisija': 81,\n // expression: https://finto.fi/mts/fi/page/m153\n 'toimittaja': 78,\n 'kuvittaja': 77,\n 'esipuheen kirjoittaja': 76,\n 'alkusanojen kirjoittaja': 75, 'loppusanojen kirjoittaja': 75,\n 'k\u00E4\u00E4nt\u00E4j\u00E4': 70,\n 'sovittaja': 50,\n // manifestaatio\n 'graafinen suunnittelija': 40,\n // kappale/item\n 'sitoja': 25,\n 'gratulaation kirjoittaja': 24\n};\n\nconst relatorTermScoreMu = {\n 's\u00E4velt\u00E4j\u00E4': 100,\n 'sanoittaja': 99,\n 'soitonoppaan tekij\u00E4': 98,\n 'alkuper\u00E4isidean luoja': 85,\n 'teoksella kunnioitettu': 81,\n // expression: https://finto.fi/mts/fi/page/m153\n 'sovittaja': 79,\n 'johtaja': 78,\n 'esitt\u00E4j\u00E4': 77,\n 'lukija': 76,\n 'miksaaja': 75,\n // manifestaatio\n 'esitt\u00E4j\u00E4 (manifestaatio)': 69,\n\n 'graafinen suunnittelija': 50,\n 'valmistaja': 41,\n 'jakaja': 40\n // kappale/item\n\n};\n\nconst relatorTermScoreVm = { // Visual Material\n // Work\n 'ohjaaja': 100,\n 'kirjoittaja': 99, 'author': 99, // Here we assume that film/whatever is based on author's book\n 'k\u00E4sikirjoittaja': 98, 'designer': 98,\n 'kuvaaja': 89,\n 's\u00E4velt\u00E4j\u00E4': 86, // Volatile. John Williams?\n 'alkuper\u00E4isidean luoja': 85,\n 'julkaisija': 82,\n 'tuottaja': 81,\n // Expression\n 'leikkaaja': 80,\n 'sovittaja': 79,\n 'johtaja': 78,\n 'esitt\u00E4j\u00E4': 77,\n 'lukija': 76,\n 'miksaaja': 75,\n 'k\u00E4\u00E4nt\u00E4j\u00E4': 70,\n // Manifestation\n 'kaivertaja': 60,\n 'graafinen suunnittelija': 59,\n 'kustantaja': 42,\n 'elokuvan jakelija': 41, 'jakaja': 41\n\n // Item\n};\n\nfunction normalizeValue(value) {\n // Removing last punc char is good enough for our purposes.\n // We don't handle abbreviations here etc.\n // Brackets should not happen either, should they?\n return value.replace(/[.,]$/u, '');\n}\n\nfunction scoreRelatorTermBk(normalizedValue) {\n if (normalizedValue in relatorTermScoreBk) {\n return relatorTermScoreBk[normalizedValue];\n }\n return 0;\n}\n\nfunction scoreRelatorTermMu(normalizedValue) {\n if (normalizedValue in relatorTermScoreMu) {\n return relatorTermScoreMu[normalizedValue];\n }\n return 0;\n}\n\nfunction scoreRelatorTermVm(normalizedValue) {\n if (normalizedValue in relatorTermScoreVm) {\n return relatorTermScoreVm[normalizedValue];\n }\n return 0;\n}\n\nexport function scoreRelatorTerm(value, typeOfMaterial = undefined) {\n // NB! We are currently using type of material only for sorting relator terms, not 7XX fields!!!\n const normalizedValue = normalizeValue(value);\n if (typeOfMaterial === 'BK') { // Books\n return scoreRelatorTermBk(normalizedValue);\n }\n if (typeOfMaterial === 'MU') { // Music (NB: audio books should be BK in this context!)\n return scoreRelatorTermMu(normalizedValue);\n }\n if (typeOfMaterial === 'VM') { // video material\n return scoreRelatorTermVm(normalizedValue);\n }\n if (normalizedValue in relatorTermScore) {\n return relatorTermScore[normalizedValue];\n }\n return 0;\n}\n\nexport function fieldOrderComparator(fieldA, fieldB) {\n\n //const sorterFunctions = [sortByTag, sortByIndexTerms, sortAlphabetically, sortByRelatorTerm, sortByOccurrenceNumber, preferFenniKeep, sortByFieldLinkAndSequenceNumber];\n\n const sorterFunctions = [sortByTag, sortByIndexTerms, sortAlphabetically, sortByRelatorTerm, sortBySubfield6, preferFenniKeep, sortByFieldLinkAndSequenceNumber, sortByHacks];\n //const sorterFunctions = [sortByIndexTerms, sortByRelatorTerm, sortByOccurrenceNumber, preferFenniKeep, sortByFieldLinkAndSequenceNumber];\n\n return globalFieldOrderComparator(fieldA, fieldB, sorterFunctions);\n}\n\nfunction sortByIndexTerms(fieldA, fieldB) {\n\n const indexTermFields = ['600', '610', '611', '630', '648', '650', '651', '652', '653', '654', '655', '656', '657', '658', '659', '662'];\n\n function scoreInd2(val) {\n const ind2Score = {'0': 0, '1': 1, '2': 2, '3': 3, '4': 8, '5': 5, '6': 6, '7': 7};\n\n if (val in ind2Score) {\n return ind2Score[val];\n }\n return 9;\n }\n\n // ATM this is not needed.\n // You may need this, if you change compare function order in sorterFunctions\n // istanbul ignore next\n if (fieldA.tag !== fieldB.tag) {\n return 0;\n }\n\n if (!indexTermFields.includes(fieldA.tag)) {\n return 0;\n }\n\n // Puts ind2=4 last\n if (scoreInd2(fieldA.ind2) > scoreInd2(fieldB.ind2)) {\n return 1;\n }\n if (scoreInd2(fieldA.ind2) < scoreInd2(fieldB.ind2)) {\n return -1;\n }\n\n function scoreDictionary(dictionary) {\n const dictionarySortIndex = {\n 'yso/fin': 0,\n 'yso/swe': 1,\n 'yso/eng': 2,\n 'slm/fin': 0.1,\n 'slm/swe': 1.1,\n 'kauno/fin': 2.1,\n 'kauno/swe': 2.2,\n 'kaunokki': 4,\n 'bella': 5\n };\n\n if (dictionary in dictionarySortIndex) {\n return dictionarySortIndex[dictionary];\n }\n return BIG_BAD_NUMBER;\n }\n\n const dictionaryA = selectFirstValue(fieldA, '2');\n const dictionaryB = selectFirstValue(fieldB, '2');\n\n const dictScoreA = scoreDictionary(dictionaryA);\n const dictScoreB = scoreDictionary(dictionaryB);\n // Use priority order for listed dictionaries:\n if (dictScoreA > dictScoreB) {\n return 1;\n }\n if (dictScoreA < dictScoreB) {\n return -1;\n }\n // Unlisted dictionaries: sort $2 value alphabetically:\n //if (dictScoreA === BIG_BAD_NUMBER) {\n if (dictionaryA > dictionaryB) {\n return 1;\n }\n if (dictionaryA < dictionaryB) {\n return -1;\n }\n //}\n return 0;\n}\n\n\nfunction preferKeep(fieldA, fieldB, keepOwner = 'FENNI') {\n const hasKeepA = fieldHasSubfield(fieldA, '9', `${keepOwner}<KEEP>`);\n const hasKeepB = fieldHasSubfield(fieldB, '9', `${keepOwner}<KEEP>`);\n\n if (hasKeepA && !hasKeepB) {\n return -1;\n }\n if (!hasKeepA && hasKeepB) {\n return 1;\n }\n\n return 0;\n}\n\n\nfunction preferFenniKeep(fieldA, fieldB) {\n const fenniPreference = preferKeep(fieldA, fieldB, 'FENNI');\n if (fenniPreference !== 0) {\n return fenniPreference;\n }\n const violaPreference = preferKeep(fieldA, fieldB, 'VIOLA');\n if (violaPreference !== 0) {\n return violaPreference;\n }\n return preferKeep(fieldA, fieldB, 'FIKKA');\n}\n\nfunction sortByRelatorTerm(fieldA, fieldB) {\n //if (!['600', '610', '611', '630', '700', '710', '711', '730', '800', '810', '811', '830'].includes(fieldA.tag)) {\n if (!['700', '710', '711', '730'].includes(fieldA.tag)) {\n return 0;\n }\n\n function fieldGetMaxRelatorTermScore(field) {\n if (!field.subfields) {\n return -2;\n }\n // If field has $t, it's a teos-nimeke-auktoriteetti, and thus meaningless. These should follow all $t-less fields...\n if (fieldHasSubfield(field, 't')) {\n return -2;\n }\n const relatorSubfieldCode = ['611', '711', '811'].includes(field.tag) ? 'j' : 'e';\n const e = field.subfields.filter(sf => sf.code === relatorSubfieldCode);\n if (e.length === 0) { // No $e is still better than having a $t\n return -1;\n }\n const scores = e.map(sf => scoreRelatorTerm(sf.value));\n //debugDev(`RELATOR SCORE FOR '${fieldToString(field)}': ${scores.join(', ')}`);\n return Math.max(...scores);\n }\n\n const scoreA = fieldGetMaxRelatorTermScore(fieldA);\n const scoreB = fieldGetMaxRelatorTermScore(fieldB);\n\n if (scoreA < scoreB) {\n return 1;\n }\n if (scoreA > scoreB) {\n return -1;\n }\n return 0;\n}\n\n\nfunction fieldGetMinLinkAndSequenceNumber(field) {\n if (!field.subfields) {\n return BIG_BAD_NUMBER;\n }\n const relevantSubfields = field.subfields.filter(sf => isValidSubfield8(sf));\n // If val is something like \"1.2\\x\" parseFloat() would give a syntax erro because of hex-like escape sequnce (at least on Chrome). Thus remove tail:\n const scores = relevantSubfields.map(sf => parseFloat(sf.value.replace(/\\\\.*$/u, '')));\n if (scores.length === 0) {\n return BIG_BAD_NUMBER;\n }\n return Math.min(...scores);\n}\n\nfunction sortByFieldLinkAndSequenceNumber(fieldA, fieldB) { // Sort by subfield $8 that is...\n const scoreA = fieldGetMinLinkAndSequenceNumber(fieldA);\n const scoreB = fieldGetMinLinkAndSequenceNumber(fieldB);\n //debugDev(` sf-8-A-score for '${fieldToString(fieldA)}: ${scoreA}`);\n //debugDev(` sf-8-B-score for '${fieldToString(fieldB)}: ${scoreB}`);\n if (scoreA === scoreB) {\n return 0;\n }\n if (scoreB === 0) {\n return 1;\n }\n if (scoreA === 0) {\n return -1;\n }\n if (scoreA > scoreB) { // smaller is better\n return 1;\n }\n return -1;\n}\n\nfunction sortByHacks(fieldA, fieldB) {\n // Place for various hackier rules\n\n if (fieldA.tag === fieldB.tag) {\n if (fieldA.tag === '041') {\n // If we have multiple f041, we want to put the $2-less first; Values from $2 whatever can not be copied to 008/35-37. (See sync-language.js.)\n const a2 = fieldA.subfields.some(sf => sf.code === '2');\n const b2 = fieldB.subfields.some(sf => sf.code === '2');\n if (a2 && !b2) {\n return 1;\n }\n if (b2 && !a2) {\n return -1;\n }\n }\n\n }\n return 0;\n}\n\nfunction sortBySubfield6(fieldA, fieldB) { // Sort by subfield $6, ex-sortByOccurrenceNumber...\n if (fieldA.tag !== '880' || fieldB.tag !== '880') {\n return 0;\n }\n\n function compareLinkingTags() {\n const tagStringA = fieldGetUnambiguousTag(fieldA);\n const tagStringB = fieldGetUnambiguousTag(fieldB);\n if (tagStringA === tagStringB || !tagStringA || !tagStringB) {\n return 0;\n }\n if (tagStringA > tagStringB) {\n return 1;\n }\n return -1;\n }\n\n function compareOccurrenceNumbers() {\n const stringA = fieldGetUnambiguousOccurrenceNumber(fieldA);\n const stringB = fieldGetUnambiguousOccurrenceNumber(fieldB);\n if (stringA === stringB) { // No action required here\n return 0;\n }\n\n // Handle expections: no occurrence number, occurrence number '00':\n if (!stringB || stringB === '00') {\n if (!stringA || stringA === '00') {\n return 0;\n }\n return -1;\n }\n if (!stringA || stringA === '00') {\n return 1;\n }\n\n // NB! We need compare ints as occurrence number can exceed 99 and be a three-digit value!\n const scoreA = parseInt(stringA, 10);\n const scoreB = parseInt(stringB, 10);\n\n if (scoreA > scoreB) { // smaller is better, thus '00' is the best\n return 1;\n }\n return -1;\n }\n\n const linkingTagComparisonResult = compareLinkingTags();\n if (linkingTagComparisonResult !== 0) {\n return linkingTagComparisonResult;\n }\n\n return compareOccurrenceNumbers();\n}\n\n\nfunction selectFirstValue(field, subcode) {\n return field.subfields\n .filter(subfield => subcode === subfield.code)\n .map(subfield => subfield.value)\n .slice(0, 1);\n}\n\n"],
|
|
5
|
+
"mappings": "AAEA,OAAO,WAAW;AAElB,SAAQ,kBAAkB,qBAAoB;AAC9C,SAAQ,WAAW,oBAAoB,wBAAwB,kCAAiC;AAChG,SAAQ,wBAAuB;AAC/B,SAAQ,qCAAqC,8BAA6B;AAM1E,MAAM,iBAAiB;AACvB,0BAA2B;AAEzB,SAAO;AAAA,IACL,aAAa;AAAA,IACb;AAAA,IAAU;AAAA,EACZ;AAEA,WAAS,IAAI,QAAQ;AACnB,UAAM,MAAM,EAAC,SAAS,CAAC,GAAG,KAAK,CAAC,GAAG,OAAO,KAAI;AAE9C,WAAO,OAAO,KAAK,oBAAoB;AAEvC,WAAO;AAAA,EACT;AAEA,WAAS,SAAS,QAAQ;AACxB,UAAM,MAAM,EAAC,SAAS,CAAC,EAAC;AAExB,UAAM,SAAS,OAAO,OAAO,IAAI,OAAK,MAAM,CAAC,CAAC;AAC9C,WAAO,KAAK,oBAAoB;AAGhC,UAAM,kBAAkB,OAAO,OAAO,CAAC,GAAG,MAAM,cAAc,CAAC,MAAM,cAAc,OAAO,OAAO,CAAC,CAAC,CAAC;AAEpG,QAAI,gBAAgB,SAAS,GAAG;AAC9B,UAAI,QAAQ,KAAK,GAAG,gBAAgB,MAAM,yBAAyB;AAAA,IACrE;AAEA,QAAI,QAAQ,EAAE,IAAI,QAAQ,UAAU;AACpC,WAAO;AAAA,EACT;AACF;AAGA,MAAM,mBAAmB;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA,EAOvB,sBAAa;AAAA,EAAK,YAAY;AAAA,EAC9B,eAAe;AAAA,EAAK,UAAU;AAAA,EAC9B,cAAc;AAAA,EACd,uBAAoB;AAAA,EAAK,0BAAuB;AAAA,EAChD,mBAAgB;AAAA,EAChB,cAAc;AAAA,EACd,YAAY;AAAA,EACZ,cAAc;AAAA,EACd,sBAAmB;AAAA,EAEnB,WAAW;AAAA,EAAI,eAAe;AAAA,EAC9B,WAAW;AAAA,EACX,4BAAyB;AAAA,EACzB,0BAA0B;AAAA,EAAI,sBAAsB;AAAA,EACpD,cAAc;AAAA,EACd,YAAY;AAAA;AAAA,EAEZ,aAAa;AAAA,EAAI,YAAY;AAAA,EAC7B,cAAc;AAAA,EAAI,UAAU;AAAA,EAC5B,aAAa;AAAA,EACb,yBAAyB;AAAA,EACzB,2BAA2B;AAAA,EAAI,4BAA4B;AAAA,EAE3D,kBAAY;AAAA,EACZ,WAAW;AAAA;AAAA,EACX,aAAa;AAAA;AAAA,EACb,wBAAY;AAAA,EACZ,UAAU;AAAA,EAAI,WAAW;AAAA;AAAA,EAEzB,2BAA2B;AAAA,EAC3B,cAAc;AAAA;AAEhB;AAEA,MAAM,qBAAqB;AAAA;AAAA,EAEzB,cAAc;AAAA,EAAK,uBAAoB;AAAA,EACvC,eAAe;AAAA,EAAI,UAAU;AAAA,EAAI,0BAAuB;AAAA,EACxD,WAAW;AAAA,EAAI,eAAe;AAAA,EAC9B,WAAW;AAAA,EAAI,YAAY;AAAA,EAC3B,4BAAyB;AAAA,EACzB,0BAA0B;AAAA,EAAI,sBAAsB;AAAA,EACpD,sBAAa;AAAA;AAAA,EACb,cAAc;AAAA,EACd,cAAc;AAAA;AAAA,EAEd,cAAc;AAAA,EACd,aAAa;AAAA,EACb,yBAAyB;AAAA,EACzB,2BAA2B;AAAA,EAAI,4BAA4B;AAAA,EAC3D,wBAAY;AAAA,EACZ,aAAa;AAAA;AAAA,EAEb,2BAA2B;AAAA;AAAA,EAE3B,UAAU;AAAA,EACV,4BAA4B;AAC9B;AAEA,MAAM,qBAAqB;AAAA,EACzB,sBAAa;AAAA,EACb,cAAc;AAAA,EACd,0BAAuB;AAAA,EACvB,4BAAyB;AAAA,EACzB,0BAA0B;AAAA;AAAA,EAE1B,aAAa;AAAA,EACb,WAAW;AAAA,EACX,kBAAY;AAAA,EACZ,UAAU;AAAA,EACV,YAAY;AAAA;AAAA,EAEZ,kCAA4B;AAAA,EAE5B,2BAA2B;AAAA,EAC3B,cAAc;AAAA,EACd,UAAU;AAAA;AAGZ;AAEA,MAAM,qBAAqB;AAAA;AAAA;AAAA,EAEzB,WAAW;AAAA,EACX,eAAe;AAAA,EAAI,UAAU;AAAA;AAAA,EAC7B,sBAAmB;AAAA,EAAI,YAAY;AAAA,EACnC,WAAW;AAAA,EACX,sBAAa;AAAA;AAAA,EACb,4BAAyB;AAAA,EACzB,cAAc;AAAA,EACd,YAAY;AAAA;AAAA,EAEZ,aAAa;AAAA,EACb,aAAa;AAAA,EACb,WAAW;AAAA,EACX,kBAAY;AAAA,EACZ,UAAU;AAAA,EACV,YAAY;AAAA,EACZ,wBAAY;AAAA;AAAA,EAEZ,cAAc;AAAA,EACd,2BAA2B;AAAA,EAC3B,cAAc;AAAA,EACd,qBAAqB;AAAA,EAAI,UAAU;AAAA;AAGrC;AAEA,SAAS,eAAe,OAAO;AAI7B,SAAO,MAAM,QAAQ,UAAU,EAAE;AACnC;AAEA,SAAS,mBAAmB,iBAAiB;AAC3C,MAAI,mBAAmB,oBAAoB;AACzC,WAAO,mBAAmB,eAAe;AAAA,EAC3C;AACA,SAAO;AACT;AAEA,SAAS,mBAAmB,iBAAiB;AAC3C,MAAI,mBAAmB,oBAAoB;AACzC,WAAO,mBAAmB,eAAe;AAAA,EAC3C;AACA,SAAO;AACT;AAEA,SAAS,mBAAmB,iBAAiB;AAC3C,MAAI,mBAAmB,oBAAoB;AACzC,WAAO,mBAAmB,eAAe;AAAA,EAC3C;AACA,SAAO;AACT;AAEO,gBAAS,iBAAiB,OAAO,iBAAiB,QAAW;AAElE,QAAM,kBAAkB,eAAe,KAAK;AAC5C,MAAI,mBAAmB,MAAM;AAC3B,WAAO,mBAAmB,eAAe;AAAA,EAC3C;AACA,MAAI,mBAAmB,MAAM;AAC3B,WAAO,mBAAmB,eAAe;AAAA,EAC3C;AACA,MAAI,mBAAmB,MAAM;AAC3B,WAAO,mBAAmB,eAAe;AAAA,EAC3C;AACA,MAAI,mBAAmB,kBAAkB;AACvC,WAAO,iBAAiB,eAAe;AAAA,EACzC;AACA,SAAO;AACT;AAEO,gBAAS,qBAAqB,QAAQ,QAAQ;AAInD,QAAM,kBAAkB,CAAC,WAAW,kBAAkB,oBAAoB,mBAAmB,iBAAiB,iBAAiB,kCAAkC,WAAW;AAG5K,SAAO,2BAA2B,QAAQ,QAAQ,eAAe;AACnE;AAEA,SAAS,iBAAiB,QAAQ,QAAQ;AAExC,QAAM,kBAAkB,CAAC,OAAO,OAAO,OAAO,OAAO,OAAO,OAAO,OAAO,OAAO,OAAO,OAAO,OAAO,OAAO,OAAO,OAAO,OAAO,KAAK;AAEvI,WAAS,UAAU,KAAK;AACtB,UAAM,YAAY,EAAC,KAAK,GAAG,KAAK,GAAG,KAAK,GAAG,KAAK,GAAG,KAAK,GAAG,KAAK,GAAG,KAAK,GAAG,KAAK,EAAC;AAEjF,QAAI,OAAO,WAAW;AACpB,aAAO,UAAU,GAAG;AAAA,IACtB;AACA,WAAO;AAAA,EACT;AAKA,MAAI,OAAO,QAAQ,OAAO,KAAK;AAC7B,WAAO;AAAA,EACT;AAEA,MAAI,CAAC,gBAAgB,SAAS,OAAO,GAAG,GAAG;AACzC,WAAO;AAAA,EACT;AAGA,MAAI,UAAU,OAAO,IAAI,IAAI,UAAU,OAAO,IAAI,GAAG;AACnD,WAAO;AAAA,EACT;AACA,MAAI,UAAU,OAAO,IAAI,IAAI,UAAU,OAAO,IAAI,GAAG;AACnD,WAAO;AAAA,EACT;AAEA,WAAS,gBAAgB,YAAY;AACnC,UAAM,sBAAsB;AAAA,MAC1B,WAAW;AAAA,MACX,WAAW;AAAA,MACX,WAAW;AAAA,MACX,WAAW;AAAA,MACX,WAAW;AAAA,MACX,aAAa;AAAA,MACb,aAAa;AAAA,MACb,YAAY;AAAA,MACZ,SAAS;AAAA,IACX;AAEA,QAAI,cAAc,qBAAqB;AACrC,aAAO,oBAAoB,UAAU;AAAA,IACvC;AACA,WAAO;AAAA,EACT;AAEA,QAAM,cAAc,iBAAiB,QAAQ,GAAG;AAChD,QAAM,cAAc,iBAAiB,QAAQ,GAAG;AAEhD,QAAM,aAAa,gBAAgB,WAAW;AAC9C,QAAM,aAAa,gBAAgB,WAAW;AAE9C,MAAI,aAAa,YAAY;AAC3B,WAAO;AAAA,EACT;AACA,MAAI,aAAa,YAAY;AAC3B,WAAO;AAAA,EACT;AAGA,MAAI,cAAc,aAAa;AAC7B,WAAO;AAAA,EACT;AACA,MAAI,cAAc,aAAa;AAC7B,WAAO;AAAA,EACT;AAEA,SAAO;AACT;AAGA,SAAS,WAAW,QAAQ,QAAQ,YAAY,SAAS;AACvD,QAAM,WAAW,iBAAiB,QAAQ,KAAK,GAAG,SAAS,QAAQ;AACnE,QAAM,WAAW,iBAAiB,QAAQ,KAAK,GAAG,SAAS,QAAQ;AAEnE,MAAI,YAAY,CAAC,UAAU;AACzB,WAAO;AAAA,EACT;AACA,MAAI,CAAC,YAAY,UAAU;AACzB,WAAO;AAAA,EACT;AAEA,SAAO;AACT;AAGA,SAAS,gBAAgB,QAAQ,QAAQ;AACvC,QAAM,kBAAkB,WAAW,QAAQ,QAAQ,OAAO;AAC1D,MAAI,oBAAoB,GAAG;AACzB,WAAO;AAAA,EACT;AACA,QAAM,kBAAkB,WAAW,QAAQ,QAAQ,OAAO;AAC1D,MAAI,oBAAoB,GAAG;AACzB,WAAO;AAAA,EACT;AACA,SAAO,WAAW,QAAQ,QAAQ,OAAO;AAC3C;AAEA,SAAS,kBAAkB,QAAQ,QAAQ;AAEzC,MAAI,CAAC,CAAC,OAAO,OAAO,OAAO,KAAK,EAAE,SAAS,OAAO,GAAG,GAAG;AACtD,WAAO;AAAA,EACT;AAEA,WAAS,4BAA4B,OAAO;AAC1C,QAAI,CAAC,MAAM,WAAW;AACpB,aAAO;AAAA,IACT;AAEA,QAAI,iBAAiB,OAAO,GAAG,GAAG;AAChC,aAAO;AAAA,IACT;AACA,UAAM,sBAAsB,CAAC,OAAO,OAAO,KAAK,EAAE,SAAS,MAAM,GAAG,IAAI,MAAM;AAC9E,UAAM,IAAI,MAAM,UAAU,OAAO,QAAM,GAAG,SAAS,mBAAmB;AACtE,QAAI,EAAE,WAAW,GAAG;AAClB,aAAO;AAAA,IACT;AACA,UAAM,SAAS,EAAE,IAAI,QAAM,iBAAiB,GAAG,KAAK,CAAC;AAErD,WAAO,KAAK,IAAI,GAAG,MAAM;AAAA,EAC3B;AAEA,QAAM,SAAS,4BAA4B,MAAM;AACjD,QAAM,SAAS,4BAA4B,MAAM;AAEjD,MAAI,SAAS,QAAQ;AACnB,WAAO;AAAA,EACT;AACA,MAAI,SAAS,QAAQ;AACnB,WAAO;AAAA,EACT;AACA,SAAO;AACT;AAGA,SAAS,iCAAiC,OAAO;AAC/C,MAAI,CAAC,MAAM,WAAW;AACpB,WAAO;AAAA,EACT;AACA,QAAM,oBAAoB,MAAM,UAAU,OAAO,QAAM,iBAAiB,EAAE,CAAC;AAE3E,QAAM,SAAS,kBAAkB,IAAI,QAAM,WAAW,GAAG,MAAM,QAAQ,UAAU,EAAE,CAAC,CAAC;AACrF,MAAI,OAAO,WAAW,GAAG;AACvB,WAAO;AAAA,EACT;AACA,SAAO,KAAK,IAAI,GAAG,MAAM;AAC3B;AAEA,SAAS,iCAAiC,QAAQ,QAAQ;AACxD,QAAM,SAAS,iCAAiC,MAAM;AACtD,QAAM,SAAS,iCAAiC,MAAM;AAGtD,MAAI,WAAW,QAAQ;AACrB,WAAO;AAAA,EACT;AACA,MAAI,WAAW,GAAG;AAChB,WAAO;AAAA,EACT;AACA,MAAI,WAAW,GAAG;AAChB,WAAO;AAAA,EACT;AACA,MAAI,SAAS,QAAQ;AACnB,WAAO;AAAA,EACT;AACA,SAAO;AACT;AAEA,SAAS,YAAY,QAAQ,QAAQ;AAGnC,MAAI,OAAO,QAAQ,OAAO,KAAK;AAC7B,QAAI,OAAO,QAAQ,OAAO;AAExB,YAAM,KAAK,OAAO,UAAU,KAAK,QAAM,GAAG,SAAS,GAAG;AACtD,YAAM,KAAK,OAAO,UAAU,KAAK,QAAM,GAAG,SAAS,GAAG;AACtD,UAAI,MAAM,CAAC,IAAI;AACb,eAAO;AAAA,MACT;AACA,UAAI,MAAM,CAAC,IAAI;AACb,eAAO;AAAA,MACT;AAAA,IACF;AAAA,EAEF;AACA,SAAO;AACT;AAEA,SAAS,gBAAgB,QAAQ,QAAQ;AACvC,MAAI,OAAO,QAAQ,SAAS,OAAO,QAAQ,OAAO;AAChD,WAAO;AAAA,EACT;AAEA,WAAS,qBAAqB;AAC5B,UAAM,aAAa,uBAAuB,MAAM;AAChD,UAAM,aAAa,uBAAuB,MAAM;AAChD,QAAI,eAAe,cAAc,CAAC,cAAc,CAAC,YAAY;AAC3D,aAAO;AAAA,IACT;AACA,QAAI,aAAa,YAAY;AAC3B,aAAO;AAAA,IACT;AACA,WAAO;AAAA,EACT;AAEA,WAAS,2BAA2B;AAClC,UAAM,UAAU,oCAAoC,MAAM;AAC1D,UAAM,UAAU,oCAAoC,MAAM;AAC1D,QAAI,YAAY,SAAS;AACvB,aAAO;AAAA,IACT;AAGA,QAAI,CAAC,WAAW,YAAY,MAAM;AAChC,UAAI,CAAC,WAAW,YAAY,MAAM;AAChC,eAAO;AAAA,MACT;AACA,aAAO;AAAA,IACT;AACA,QAAI,CAAC,WAAW,YAAY,MAAM;AAChC,aAAO;AAAA,IACT;AAGA,UAAM,SAAS,SAAS,SAAS,EAAE;AACnC,UAAM,SAAS,SAAS,SAAS,EAAE;AAEnC,QAAI,SAAS,QAAQ;AACnB,aAAO;AAAA,IACT;AACA,WAAO;AAAA,EACT;AAEA,QAAM,6BAA6B,mBAAmB;AACtD,MAAI,+BAA+B,GAAG;AACpC,WAAO;AAAA,EACT;AAEA,SAAO,yBAAyB;AAClC;AAGA,SAAS,iBAAiB,OAAO,SAAS;AACxC,SAAO,MAAM,UACV,OAAO,cAAY,YAAY,SAAS,IAAI,EAC5C,IAAI,cAAY,SAAS,KAAK,EAC9B,MAAM,GAAG,CAAC;AACf;",
|
|
6
6
|
"names": []
|
|
7
7
|
}
|
package/dist/sortFields.test.js
CHANGED
|
@@ -3,7 +3,6 @@ import { MarcRecord } from "@natlibfi/marc-record";
|
|
|
3
3
|
import validatorFactory from "./sortFields.js";
|
|
4
4
|
import { READERS } from "@natlibfi/fixura";
|
|
5
5
|
import generateTests from "@natlibfi/fixugen";
|
|
6
|
-
import createDebugLogger from "debug";
|
|
7
6
|
generateTests({
|
|
8
7
|
callback,
|
|
9
8
|
path: [import.meta.dirname, "..", "test-fixtures", "sort-fields"],
|
|
@@ -18,18 +17,13 @@ generateTests({
|
|
|
18
17
|
}
|
|
19
18
|
}
|
|
20
19
|
});
|
|
21
|
-
const debug = createDebugLogger("@natlibfi/marc-record-validators-melinda/sortFields:test");
|
|
22
20
|
async function testValidatorFactory() {
|
|
23
21
|
const validator = await validatorFactory();
|
|
24
22
|
assert.equal(typeof validator, "object");
|
|
25
23
|
assert.equal(typeof validator.description, "string");
|
|
26
24
|
assert.equal(typeof validator.validate, "function");
|
|
27
25
|
}
|
|
28
|
-
async function callback({ getFixture,
|
|
29
|
-
if (enabled === false) {
|
|
30
|
-
debug("TEST SKIPPED!");
|
|
31
|
-
return;
|
|
32
|
-
}
|
|
26
|
+
async function callback({ getFixture, fix = true }) {
|
|
33
27
|
const validator = await validatorFactory();
|
|
34
28
|
const record = new MarcRecord(getFixture("input.json"));
|
|
35
29
|
const expectedResult = getFixture("result.json");
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
{
|
|
2
2
|
"version": 3,
|
|
3
3
|
"sources": ["../src/sortFields.test.js"],
|
|
4
|
-
"sourcesContent": ["import assert from 'node:assert';\nimport {MarcRecord} from '@natlibfi/marc-record';\nimport validatorFactory from './sortFields.js';\nimport {READERS} from '@natlibfi/fixura';\nimport generateTests from '@natlibfi/fixugen';\
|
|
5
|
-
"mappings": "AAAA,OAAO,YAAY;AACnB,SAAQ,kBAAiB;AACzB,OAAO,sBAAsB;AAC7B,SAAQ,eAAc;AACtB,OAAO,mBAAmB;
|
|
4
|
+
"sourcesContent": ["import assert from 'node:assert';\nimport {MarcRecord} from '@natlibfi/marc-record';\nimport validatorFactory from './sortFields.js';\nimport {READERS} from '@natlibfi/fixura';\nimport generateTests from '@natlibfi/fixugen';\n//import createDebugLogger from 'debug';\n\ngenerateTests({\n callback,\n path: [import.meta.dirname, '..', 'test-fixtures', 'sort-fields'],\n useMetadataFile: true,\n recurse: false,\n fixura: {\n reader: READERS.JSON\n },\n hooks: {\n before: async () => {\n testValidatorFactory();\n }\n }\n});\n//const debug = createDebugLogger('@natlibfi/marc-record-validators-melinda/sortFields:test');\n\nasync function testValidatorFactory() {\n const validator = await validatorFactory();\n\n assert.equal(typeof validator, 'object');\n assert.equal(typeof validator.description, 'string');\n assert.equal(typeof validator.validate, 'function');\n}\n\nasync function callback({getFixture, fix = true}) {\n const validator = await validatorFactory();\n const record = new MarcRecord(getFixture('input.json'));\n const expectedResult = getFixture('result.json');\n // console.log(expectedResult); // eslint-disable-line\n\n if (!fix) {\n const result = await validator.validate(record);\n assert.deepEqual(result, expectedResult);\n return;\n }\n\n await validator.fix(record);\n assert.deepEqual(record, expectedResult);\n}\n"],
|
|
5
|
+
"mappings": "AAAA,OAAO,YAAY;AACnB,SAAQ,kBAAiB;AACzB,OAAO,sBAAsB;AAC7B,SAAQ,eAAc;AACtB,OAAO,mBAAmB;AAG1B,cAAc;AAAA,EACZ;AAAA,EACA,MAAM,CAAC,YAAY,SAAS,MAAM,iBAAiB,aAAa;AAAA,EAChE,iBAAiB;AAAA,EACjB,SAAS;AAAA,EACT,QAAQ;AAAA,IACN,QAAQ,QAAQ;AAAA,EAClB;AAAA,EACA,OAAO;AAAA,IACL,QAAQ,YAAY;AAClB,2BAAqB;AAAA,IACvB;AAAA,EACF;AACF,CAAC;AAGD,eAAe,uBAAuB;AACpC,QAAM,YAAY,MAAM,iBAAiB;AAEzC,SAAO,MAAM,OAAO,WAAW,QAAQ;AACvC,SAAO,MAAM,OAAO,UAAU,aAAa,QAAQ;AACnD,SAAO,MAAM,OAAO,UAAU,UAAU,UAAU;AACpD;AAEA,eAAe,SAAS,EAAC,YAAY,MAAM,KAAI,GAAG;AAChD,QAAM,YAAY,MAAM,iBAAiB;AACzC,QAAM,SAAS,IAAI,WAAW,WAAW,YAAY,CAAC;AACtD,QAAM,iBAAiB,WAAW,aAAa;AAG/C,MAAI,CAAC,KAAK;AACR,UAAM,SAAS,MAAM,UAAU,SAAS,MAAM;AAC9C,WAAO,UAAU,QAAQ,cAAc;AACvC;AAAA,EACF;AAEA,QAAM,UAAU,IAAI,MAAM;AAC1B,SAAO,UAAU,QAAQ,cAAc;AACzC;",
|
|
6
6
|
"names": []
|
|
7
7
|
}
|
|
@@ -3,7 +3,6 @@ import { MarcRecord } from "@natlibfi/marc-record";
|
|
|
3
3
|
import validatorFactory from "./sortRelatorTerms.js";
|
|
4
4
|
import { READERS } from "@natlibfi/fixura";
|
|
5
5
|
import generateTests from "@natlibfi/fixugen";
|
|
6
|
-
import createDebugLogger from "debug";
|
|
7
6
|
generateTests({
|
|
8
7
|
callback,
|
|
9
8
|
path: [import.meta.dirname, "..", "test-fixtures", "sort-relator-terms"],
|
|
@@ -18,18 +17,13 @@ generateTests({
|
|
|
18
17
|
}
|
|
19
18
|
}
|
|
20
19
|
});
|
|
21
|
-
const debug = createDebugLogger("@natlibfi/marc-record-validators-melinda/sortRelatorTerms:test");
|
|
22
20
|
async function testValidatorFactory() {
|
|
23
21
|
const validator = await validatorFactory();
|
|
24
22
|
assert.equal(typeof validator, object);
|
|
25
23
|
assert.equal(typeof validator.description, "string");
|
|
26
24
|
assert.equal(typeof validator.validate, "function");
|
|
27
25
|
}
|
|
28
|
-
async function callback({ getFixture,
|
|
29
|
-
if (enabled === false) {
|
|
30
|
-
debug("TEST SKIPPED!");
|
|
31
|
-
return;
|
|
32
|
-
}
|
|
26
|
+
async function callback({ getFixture, fix = false }) {
|
|
33
27
|
const validator = await validatorFactory();
|
|
34
28
|
const record = new MarcRecord(getFixture("record.json"));
|
|
35
29
|
const expectedResult = getFixture("expectedResult.json");
|