@webiny/lexical-converter 5.38.1 → 5.38.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
|
@@ -32,11 +32,12 @@ const passthroughMapper = node => node;
|
|
|
32
32
|
*/
|
|
33
33
|
const createHtmlToLexicalParser = (config = {}) => {
|
|
34
34
|
return domDocument => {
|
|
35
|
+
var _config$editorConfig;
|
|
35
36
|
const normalizeTextNodes = config.normalizeTextNodes ?? true;
|
|
36
37
|
const textNodeNormalizer = normalizeTextNodes ? textNodeParentNormalizer : passthroughMapper;
|
|
37
38
|
const customNodeMapper = config.nodeMapper ?? passthroughMapper;
|
|
38
39
|
const editor = (0, _headless.createHeadlessEditor)((0, _objectSpread2.default)((0, _objectSpread2.default)({}, config.editorConfig), {}, {
|
|
39
|
-
nodes: [..._lexicalNodes.allNodes, ...(config.editorConfig
|
|
40
|
+
nodes: [..._lexicalNodes.allNodes, ...(((_config$editorConfig = config.editorConfig) === null || _config$editorConfig === void 0 ? void 0 : _config$editorConfig.nodes) || [])]
|
|
40
41
|
}));
|
|
41
42
|
let parsingError;
|
|
42
43
|
editor.update(() => {
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"names":["_headless","require","_html","_lexical","_lexicalNodes","textNodeParentNormalizer","node","getType","getParent","$createParagraphNode","append","passthroughMapper","createHtmlToLexicalParser","config","domDocument","normalizeTextNodes","textNodeNormalizer","customNodeMapper","nodeMapper","editor","createHeadlessEditor","_objectSpread2","default","editorConfig","nodes","allNodes","parsingError","update","lexicalNodes","$generateNodesFromDOM","map","$getRoot","select","selection","$getSelection","insertNodes","err","discrete","getEditorState","toJSON","exports"],"sources":["createHtmlToLexicalParser.ts"],"sourcesContent":["import { createHeadlessEditor } from \"@lexical/headless\";\nimport { $generateNodesFromDOM } from \"@lexical/html\";\nimport { $getRoot, $getSelection } from \"lexical\";\nimport { allNodes, $createParagraphNode } from \"@webiny/lexical-nodes\";\nimport { NodeMapper, ParserConfigurationOptions } from \"~/types\";\n\n/**\n * By itself, \"text\" node without a parent node (like \"paragraph\"), is not a valid node. Lexical will simply ignore these elements.\n * To fix this issue, we wrap the text node with a paragraph node.\n *\n * EXAMPLE:\n * When we parse DOM, sometimes, 'span' html tag doesn't have parent elements that match the\n * lexical node elements (there's no Node class that can handle that HTML element), like paragraph or headings.\n * In this case, Lexical will parse the 'span' tag as a text node, but without a parent element.\n */\nconst textNodeParentNormalizer: NodeMapper = node => {\n if (node.getType() === \"text\" && node.getParent() === null) {\n return $createParagraphNode().append(node);\n }\n return node;\n};\n\nconst passthroughMapper: NodeMapper = node => node;\n\n/**\n * Parse html string to lexical JSON object.\n */\nexport const createHtmlToLexicalParser = (config: ParserConfigurationOptions = {}) => {\n return (domDocument: Document): Record<string, any> | null => {\n const normalizeTextNodes = config.normalizeTextNodes ?? true;\n const textNodeNormalizer = normalizeTextNodes\n ? textNodeParentNormalizer\n : passthroughMapper;\n\n const customNodeMapper: NodeMapper = config.nodeMapper ?? passthroughMapper;\n\n const editor = createHeadlessEditor({\n ...config.editorConfig,\n nodes: [...allNodes, ...(config.editorConfig?.nodes || [])]\n });\n\n let parsingError;\n\n editor.update(\n () => {\n // Convert to lexical node objects that can be stored in db.\n const lexicalNodes = $generateNodesFromDOM(editor, domDocument)\n .map(textNodeNormalizer)\n .map(customNodeMapper);\n\n // Select the root\n $getRoot().select();\n\n // Insert the nodes at a selection.\n const selection = $getSelection();\n if (selection) {\n try {\n selection.insertNodes(lexicalNodes);\n } catch (err) {\n parsingError = err;\n }\n }\n },\n /**\n * Prevents this update from being batched, forcing it to run synchronously.\n */\n { discrete: true }\n );\n\n if (parsingError) {\n throw parsingError;\n }\n\n return editor.getEditorState().toJSON();\n };\n};\n"],"mappings":";;;;;;;;AAAA,IAAAA,SAAA,GAAAC,OAAA;AACA,IAAAC,KAAA,GAAAD,OAAA;AACA,IAAAE,QAAA,GAAAF,OAAA;AACA,IAAAG,aAAA,GAAAH,OAAA;AAGA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA,MAAMI,wBAAoC,GAAGC,IAAI,IAAI;EACjD,IAAIA,IAAI,CAACC,OAAO,CAAC,CAAC,KAAK,MAAM,IAAID,IAAI,CAACE,SAAS,CAAC,CAAC,KAAK,IAAI,EAAE;IACxD,OAAO,IAAAC,kCAAoB,EAAC,CAAC,CAACC,MAAM,CAACJ,IAAI,CAAC;EAC9C;EACA,OAAOA,IAAI;AACf,CAAC;AAED,MAAMK,iBAA6B,GAAGL,IAAI,IAAIA,IAAI;;AAElD;AACA;AACA;AACO,MAAMM,yBAAyB,GAAGA,CAACC,MAAkC,GAAG,CAAC,CAAC,KAAK;EAClF,OAAQC,WAAqB,IAAiC;IAC1D,MAAMC,kBAAkB,
|
|
1
|
+
{"version":3,"names":["_headless","require","_html","_lexical","_lexicalNodes","textNodeParentNormalizer","node","getType","getParent","$createParagraphNode","append","passthroughMapper","createHtmlToLexicalParser","config","domDocument","_config$editorConfig","normalizeTextNodes","textNodeNormalizer","customNodeMapper","nodeMapper","editor","createHeadlessEditor","_objectSpread2","default","editorConfig","nodes","allNodes","parsingError","update","lexicalNodes","$generateNodesFromDOM","map","$getRoot","select","selection","$getSelection","insertNodes","err","discrete","getEditorState","toJSON","exports"],"sources":["createHtmlToLexicalParser.ts"],"sourcesContent":["import { createHeadlessEditor } from \"@lexical/headless\";\nimport { $generateNodesFromDOM } from \"@lexical/html\";\nimport { $getRoot, $getSelection } from \"lexical\";\nimport { allNodes, $createParagraphNode } from \"@webiny/lexical-nodes\";\nimport { NodeMapper, ParserConfigurationOptions } from \"~/types\";\n\n/**\n * By itself, \"text\" node without a parent node (like \"paragraph\"), is not a valid node. Lexical will simply ignore these elements.\n * To fix this issue, we wrap the text node with a paragraph node.\n *\n * EXAMPLE:\n * When we parse DOM, sometimes, 'span' html tag doesn't have parent elements that match the\n * lexical node elements (there's no Node class that can handle that HTML element), like paragraph or headings.\n * In this case, Lexical will parse the 'span' tag as a text node, but without a parent element.\n */\nconst textNodeParentNormalizer: NodeMapper = node => {\n if (node.getType() === \"text\" && node.getParent() === null) {\n return $createParagraphNode().append(node);\n }\n return node;\n};\n\nconst passthroughMapper: NodeMapper = node => node;\n\n/**\n * Parse html string to lexical JSON object.\n */\nexport const createHtmlToLexicalParser = (config: ParserConfigurationOptions = {}) => {\n return (domDocument: Document): Record<string, any> | null => {\n const normalizeTextNodes = config.normalizeTextNodes ?? true;\n const textNodeNormalizer = normalizeTextNodes\n ? textNodeParentNormalizer\n : passthroughMapper;\n\n const customNodeMapper: NodeMapper = config.nodeMapper ?? passthroughMapper;\n\n const editor = createHeadlessEditor({\n ...config.editorConfig,\n nodes: [...allNodes, ...(config.editorConfig?.nodes || [])]\n });\n\n let parsingError;\n\n editor.update(\n () => {\n // Convert to lexical node objects that can be stored in db.\n const lexicalNodes = $generateNodesFromDOM(editor, domDocument)\n .map(textNodeNormalizer)\n .map(customNodeMapper);\n\n // Select the root\n $getRoot().select();\n\n // Insert the nodes at a selection.\n const selection = $getSelection();\n if (selection) {\n try {\n selection.insertNodes(lexicalNodes);\n } catch (err) {\n parsingError = err;\n }\n }\n },\n /**\n * Prevents this update from being batched, forcing it to run synchronously.\n */\n { discrete: true }\n );\n\n if (parsingError) {\n throw parsingError;\n }\n\n return editor.getEditorState().toJSON();\n };\n};\n"],"mappings":";;;;;;;;AAAA,IAAAA,SAAA,GAAAC,OAAA;AACA,IAAAC,KAAA,GAAAD,OAAA;AACA,IAAAE,QAAA,GAAAF,OAAA;AACA,IAAAG,aAAA,GAAAH,OAAA;AAGA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA;AACA,MAAMI,wBAAoC,GAAGC,IAAI,IAAI;EACjD,IAAIA,IAAI,CAACC,OAAO,CAAC,CAAC,KAAK,MAAM,IAAID,IAAI,CAACE,SAAS,CAAC,CAAC,KAAK,IAAI,EAAE;IACxD,OAAO,IAAAC,kCAAoB,EAAC,CAAC,CAACC,MAAM,CAACJ,IAAI,CAAC;EAC9C;EACA,OAAOA,IAAI;AACf,CAAC;AAED,MAAMK,iBAA6B,GAAGL,IAAI,IAAIA,IAAI;;AAElD;AACA;AACA;AACO,MAAMM,yBAAyB,GAAGA,CAACC,MAAkC,GAAG,CAAC,CAAC,KAAK;EAClF,OAAQC,WAAqB,IAAiC;IAAA,IAAAC,oBAAA;IAC1D,MAAMC,kBAAkB,GAAGH,MAAM,CAACG,kBAAkB,IAAI,IAAI;IAC5D,MAAMC,kBAAkB,GAAGD,kBAAkB,GACvCX,wBAAwB,GACxBM,iBAAiB;IAEvB,MAAMO,gBAA4B,GAAGL,MAAM,CAACM,UAAU,IAAIR,iBAAiB;IAE3E,MAAMS,MAAM,GAAG,IAAAC,8BAAoB,MAAAC,cAAA,CAAAC,OAAA,MAAAD,cAAA,CAAAC,OAAA,MAC5BV,MAAM,CAACW,YAAY;MACtBC,KAAK,EAAE,CAAC,GAAGC,sBAAQ,EAAE,IAAI,EAAAX,oBAAA,GAAAF,MAAM,CAACW,YAAY,cAAAT,oBAAA,uBAAnBA,oBAAA,CAAqBU,KAAK,KAAI,EAAE,CAAC;IAAC,EAC9D,CAAC;IAEF,IAAIE,YAAY;IAEhBP,MAAM,CAACQ,MAAM,CACT,MAAM;MACF;MACA,MAAMC,YAAY,GAAG,IAAAC,2BAAqB,EAACV,MAAM,EAAEN,WAAW,CAAC,CAC1DiB,GAAG,CAACd,kBAAkB,CAAC,CACvBc,GAAG,CAACb,gBAAgB,CAAC;;MAE1B;MACA,IAAAc,iBAAQ,EAAC,CAAC,CAACC,MAAM,CAAC,CAAC;;MAEnB;MACA,MAAMC,SAAS,GAAG,IAAAC,sBAAa,EAAC,CAAC;MACjC,IAAID,SAAS,EAAE;QACX,IAAI;UACAA,SAAS,CAACE,WAAW,CAACP,YAAY,CAAC;QACvC,CAAC,CAAC,OAAOQ,GAAG,EAAE;UACVV,YAAY,GAAGU,GAAG;QACtB;MACJ;IACJ,CAAC;IACD;AACZ;AACA;IACY;MAAEC,QAAQ,EAAE;IAAK,CACrB,CAAC;IAED,IAAIX,YAAY,EAAE;MACd,MAAMA,YAAY;IACtB;IAEA,OAAOP,MAAM,CAACmB,cAAc,CAAC,CAAC,CAACC,MAAM,CAAC,CAAC;EAC3C,CAAC;AACL,CAAC;AAACC,OAAA,CAAA7B,yBAAA,GAAAA,yBAAA"}
|
|
@@ -13,9 +13,10 @@ var _headless = require("@lexical/headless");
|
|
|
13
13
|
var _lexicalNodes = require("@webiny/lexical-nodes");
|
|
14
14
|
class LexicalStateTransformer {
|
|
15
15
|
constructor(config = {}) {
|
|
16
|
+
var _config$editorConfig;
|
|
16
17
|
(0, _defineProperty2.default)(this, "editor", void 0);
|
|
17
18
|
this.editor = (0, _headless.createHeadlessEditor)((0, _objectSpread2.default)((0, _objectSpread2.default)({}, config.editorConfig), {}, {
|
|
18
|
-
nodes: [..._lexicalNodes.allNodes, ...(config.editorConfig
|
|
19
|
+
nodes: [..._lexicalNodes.allNodes, ...(((_config$editorConfig = config.editorConfig) === null || _config$editorConfig === void 0 ? void 0 : _config$editorConfig.nodes) || [])]
|
|
19
20
|
}));
|
|
20
21
|
}
|
|
21
22
|
flatten(state) {
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"names":["_lexical","require","_html","_headless","_lexicalNodes","LexicalStateTransformer","constructor","config","_defineProperty2","default","editor","createHeadlessEditor","_objectSpread2","editorConfig","nodes","allNodes","flatten","state","editorState","parseEditorState","setEditorState","flattenedNodes","update","children","$getRoot","getChildren","map","childNode","selection","$createNodeSelection","add","getKey","getNodeDescendants","forEach","node","html","$generateHtmlFromNodes","toHtml","$isElementNode","child","flat","createLexicalStateTransformer","exports"],"sources":["createLexicalStateTransformer.ts"],"sourcesContent":["import {\n CreateEditorArgs,\n LexicalEditor,\n SerializedEditorState,\n $getRoot,\n $createNodeSelection,\n $isElementNode,\n LexicalNode\n} from \"lexical\";\nimport { $generateHtmlFromNodes } from \"@lexical/html\";\nimport { createHeadlessEditor } from \"@lexical/headless\";\nimport { allNodes } from \"@webiny/lexical-nodes\";\n\ninterface LexicalStateTransformerConfig {\n editorConfig?: Pick<CreateEditorArgs, \"nodes\" | \"theme\">;\n}\n\nexport type FlatStateWithHTML = Array<{ node: LexicalNode; html: string }>;\n\nclass LexicalStateTransformer {\n private readonly editor: LexicalEditor;\n\n constructor(config: LexicalStateTransformerConfig = {}) {\n this.editor = createHeadlessEditor({\n ...config.editorConfig,\n nodes: [...allNodes, ...(config.editorConfig?.nodes || [])]\n });\n }\n\n public flatten(state: string | SerializedEditorState) {\n const editorState = this.editor.parseEditorState(state);\n this.editor.setEditorState(editorState);\n\n let flattenedNodes: FlatStateWithHTML = [];\n\n this.editor.update(() => {\n const children = $getRoot().getChildren();\n\n flattenedNodes = children.map(childNode => {\n const selection = $createNodeSelection();\n selection.add(childNode.getKey());\n\n this.getNodeDescendants(childNode).forEach(node => {\n selection.add(node.getKey());\n });\n\n const html = $generateHtmlFromNodes(this.editor, selection);\n\n return {\n node: childNode,\n html\n };\n });\n });\n\n return flattenedNodes;\n }\n\n public toHtml(state: string | SerializedEditorState) {\n const editorState = this.editor.parseEditorState(state);\n this.editor.setEditorState(editorState);\n\n let html = \"\";\n\n this.editor.update(() => {\n html = $generateHtmlFromNodes(this.editor);\n });\n\n return html;\n }\n\n private getNodeDescendants(node: LexicalNode): LexicalNode[] {\n if (!$isElementNode(node)) {\n return [];\n }\n const children = node.getChildren();\n return [...children, ...children.map(child => this.getNodeDescendants(child)).flat()];\n }\n}\n\nexport const createLexicalStateTransformer = (config?: LexicalStateTransformerConfig) => {\n return new LexicalStateTransformer(config);\n};\n"],"mappings":";;;;;;;;;AAAA,IAAAA,QAAA,GAAAC,OAAA;AASA,IAAAC,KAAA,GAAAD,OAAA;AACA,IAAAE,SAAA,GAAAF,OAAA;AACA,IAAAG,aAAA,GAAAH,OAAA;AAQA,MAAMI,uBAAuB,CAAC;EAG1BC,WAAWA,CAACC,MAAqC,GAAG,CAAC,CAAC,EAAE;IAAA,IAAAC,gBAAA,CAAAC,OAAA;IACpD,IAAI,CAACC,MAAM,GAAG,IAAAC,8BAAoB,MAAAC,cAAA,CAAAH,OAAA,MAAAG,cAAA,CAAAH,OAAA,
|
|
1
|
+
{"version":3,"names":["_lexical","require","_html","_headless","_lexicalNodes","LexicalStateTransformer","constructor","config","_config$editorConfig","_defineProperty2","default","editor","createHeadlessEditor","_objectSpread2","editorConfig","nodes","allNodes","flatten","state","editorState","parseEditorState","setEditorState","flattenedNodes","update","children","$getRoot","getChildren","map","childNode","selection","$createNodeSelection","add","getKey","getNodeDescendants","forEach","node","html","$generateHtmlFromNodes","toHtml","$isElementNode","child","flat","createLexicalStateTransformer","exports"],"sources":["createLexicalStateTransformer.ts"],"sourcesContent":["import {\n CreateEditorArgs,\n LexicalEditor,\n SerializedEditorState,\n $getRoot,\n $createNodeSelection,\n $isElementNode,\n LexicalNode\n} from \"lexical\";\nimport { $generateHtmlFromNodes } from \"@lexical/html\";\nimport { createHeadlessEditor } from \"@lexical/headless\";\nimport { allNodes } from \"@webiny/lexical-nodes\";\n\ninterface LexicalStateTransformerConfig {\n editorConfig?: Pick<CreateEditorArgs, \"nodes\" | \"theme\">;\n}\n\nexport type FlatStateWithHTML = Array<{ node: LexicalNode; html: string }>;\n\nclass LexicalStateTransformer {\n private readonly editor: LexicalEditor;\n\n constructor(config: LexicalStateTransformerConfig = {}) {\n this.editor = createHeadlessEditor({\n ...config.editorConfig,\n nodes: [...allNodes, ...(config.editorConfig?.nodes || [])]\n });\n }\n\n public flatten(state: string | SerializedEditorState) {\n const editorState = this.editor.parseEditorState(state);\n this.editor.setEditorState(editorState);\n\n let flattenedNodes: FlatStateWithHTML = [];\n\n this.editor.update(() => {\n const children = $getRoot().getChildren();\n\n flattenedNodes = children.map(childNode => {\n const selection = $createNodeSelection();\n selection.add(childNode.getKey());\n\n this.getNodeDescendants(childNode).forEach(node => {\n selection.add(node.getKey());\n });\n\n const html = $generateHtmlFromNodes(this.editor, selection);\n\n return {\n node: childNode,\n html\n };\n });\n });\n\n return flattenedNodes;\n }\n\n public toHtml(state: string | SerializedEditorState) {\n const editorState = this.editor.parseEditorState(state);\n this.editor.setEditorState(editorState);\n\n let html = \"\";\n\n this.editor.update(() => {\n html = $generateHtmlFromNodes(this.editor);\n });\n\n return html;\n }\n\n private getNodeDescendants(node: LexicalNode): LexicalNode[] {\n if (!$isElementNode(node)) {\n return [];\n }\n const children = node.getChildren();\n return [...children, ...children.map(child => this.getNodeDescendants(child)).flat()];\n }\n}\n\nexport const createLexicalStateTransformer = (config?: LexicalStateTransformerConfig) => {\n return new LexicalStateTransformer(config);\n};\n"],"mappings":";;;;;;;;;AAAA,IAAAA,QAAA,GAAAC,OAAA;AASA,IAAAC,KAAA,GAAAD,OAAA;AACA,IAAAE,SAAA,GAAAF,OAAA;AACA,IAAAG,aAAA,GAAAH,OAAA;AAQA,MAAMI,uBAAuB,CAAC;EAG1BC,WAAWA,CAACC,MAAqC,GAAG,CAAC,CAAC,EAAE;IAAA,IAAAC,oBAAA;IAAA,IAAAC,gBAAA,CAAAC,OAAA;IACpD,IAAI,CAACC,MAAM,GAAG,IAAAC,8BAAoB,MAAAC,cAAA,CAAAH,OAAA,MAAAG,cAAA,CAAAH,OAAA,MAC3BH,MAAM,CAACO,YAAY;MACtBC,KAAK,EAAE,CAAC,GAAGC,sBAAQ,EAAE,IAAI,EAAAR,oBAAA,GAAAD,MAAM,CAACO,YAAY,cAAAN,oBAAA,uBAAnBA,oBAAA,CAAqBO,KAAK,KAAI,EAAE,CAAC;IAAC,EAC9D,CAAC;EACN;EAEOE,OAAOA,CAACC,KAAqC,EAAE;IAClD,MAAMC,WAAW,GAAG,IAAI,CAACR,MAAM,CAACS,gBAAgB,CAACF,KAAK,CAAC;IACvD,IAAI,CAACP,MAAM,CAACU,cAAc,CAACF,WAAW,CAAC;IAEvC,IAAIG,cAAiC,GAAG,EAAE;IAE1C,IAAI,CAACX,MAAM,CAACY,MAAM,CAAC,MAAM;MACrB,MAAMC,QAAQ,GAAG,IAAAC,iBAAQ,EAAC,CAAC,CAACC,WAAW,CAAC,CAAC;MAEzCJ,cAAc,GAAGE,QAAQ,CAACG,GAAG,CAACC,SAAS,IAAI;QACvC,MAAMC,SAAS,GAAG,IAAAC,6BAAoB,EAAC,CAAC;QACxCD,SAAS,CAACE,GAAG,CAACH,SAAS,CAACI,MAAM,CAAC,CAAC,CAAC;QAEjC,IAAI,CAACC,kBAAkB,CAACL,SAAS,CAAC,CAACM,OAAO,CAACC,IAAI,IAAI;UAC/CN,SAAS,CAACE,GAAG,CAACI,IAAI,CAACH,MAAM,CAAC,CAAC,CAAC;QAChC,CAAC,CAAC;QAEF,MAAMI,IAAI,GAAG,IAAAC,4BAAsB,EAAC,IAAI,CAAC1B,MAAM,EAAEkB,SAAS,CAAC;QAE3D,OAAO;UACHM,IAAI,EAAEP,SAAS;UACfQ;QACJ,CAAC;MACL,CAAC,CAAC;IACN,CAAC,CAAC;IAEF,OAAOd,cAAc;EACzB;EAEOgB,MAAMA,CAACpB,KAAqC,EAAE;IACjD,MAAMC,WAAW,GAAG,IAAI,CAACR,MAAM,CAACS,gBAAgB,CAACF,KAAK,CAAC;IACvD,IAAI,CAACP,MAAM,CAACU,cAAc,CAACF,WAAW,CAAC;IAEvC,IAAIiB,IAAI,GAAG,EAAE;IAEb,IAAI,CAACzB,MAAM,CAACY,MAAM,CAAC,MAAM;MACrBa,IAAI,GAAG,IAAAC,4BAAsB,EAAC,IAAI,CAAC1B,MAAM,CAAC;IAC9C,CAAC,CAAC;IAEF,OAAOyB,IAAI;EACf;EAEQH,kBAAkBA,CAACE,IAAiB,EAAiB;IACzD,IAAI,CAAC,IAAAI,uBAAc,EAACJ,IAAI,CAAC,EAAE;MACvB,OAAO,EAAE;IACb;IACA,MAAMX,QAAQ,GAAGW,IAAI,CAACT,WAAW,CAAC,CAAC;IACnC,OAAO,CAAC,GAAGF,QAAQ,EAAE,GAAGA,QAAQ,CAACG,GAAG,CAACa,KAAK,IAAI,IAAI,CAACP,kBAAkB,CAACO,KAAK,CAAC,CAAC,CAACC,IAAI,CAAC,CAAC,CAAC;EACzF;AACJ;AAEO,MAAMC,6BAA6B,GAAInC,MAAsC,IAAK;EACrF,OAAO,IAAIF,uBAAuB,CAACE,MAAM,CAAC;AAC9C,CAAC;AAACoC,OAAA,CAAAD,6BAAA,GAAAA,6BAAA"}
|
package/package.json
CHANGED
|
@@ -1,16 +1,16 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@webiny/lexical-converter",
|
|
3
|
-
"version": "5.38.
|
|
3
|
+
"version": "5.38.2",
|
|
4
4
|
"dependencies": {
|
|
5
5
|
"@lexical/headless": "0.12.2",
|
|
6
6
|
"@lexical/html": "0.12.2",
|
|
7
|
-
"@webiny/lexical-nodes": "5.38.
|
|
7
|
+
"@webiny/lexical-nodes": "5.38.2",
|
|
8
8
|
"lexical": "0.12.2"
|
|
9
9
|
},
|
|
10
10
|
"devDependencies": {
|
|
11
11
|
"@types/jsdom": "21.1.3",
|
|
12
|
-
"@webiny/cli": "5.38.
|
|
13
|
-
"@webiny/project-utils": "5.38.
|
|
12
|
+
"@webiny/cli": "5.38.2",
|
|
13
|
+
"@webiny/project-utils": "5.38.2",
|
|
14
14
|
"jsdom": "21.1.2"
|
|
15
15
|
},
|
|
16
16
|
"publishConfig": {
|
|
@@ -22,5 +22,5 @@
|
|
|
22
22
|
"watch": "yarn webiny run watch"
|
|
23
23
|
},
|
|
24
24
|
"test": "jest --verbose --runInBand --detectOpenHandles --forceExit",
|
|
25
|
-
"gitHead": "
|
|
25
|
+
"gitHead": "443e107cc474882f558e0783a2d9068b7692bd00"
|
|
26
26
|
}
|