@dxos/echo-pipeline 0.5.3-main.5a4243f → 0.5.3-main.f752aaa
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/lib/browser/index.mjs.map +1 -1
- package/dist/lib/node/index.cjs.map +1 -1
- package/dist/types/src/automerge/automerge-host.d.ts +2 -2
- package/dist/types/src/automerge/leveldb-storage-adapter.d.ts +2 -2
- package/dist/types/src/automerge/migrations.d.ts +2 -2
- package/package.json +33 -33
- package/src/automerge/automerge-host.ts +3 -3
- package/src/automerge/leveldb-storage-adapter.ts +2 -2
- package/src/automerge/migrations.ts +2 -2
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
{
|
|
2
2
|
"version": 3,
|
|
3
3
|
"sources": ["../../../src/automerge/automerge-host.ts", "../../../src/automerge/echo-network-adapter.ts", "../../../src/automerge/leveldb-storage-adapter.ts", "../../../src/automerge/local-host-network-adapter.ts", "../../../src/automerge/mesh-network-adapter.ts", "../../../src/automerge/migrations.ts", "../../../src/automerge/automerge-storage-adapter.ts", "../../../src/automerge/automerge-doc-loader.ts"],
|
|
4
|
-
"sourcesContent": ["//\n// Copyright 2023 DXOS.org\n//\n\nimport { Event } from '@dxos/async';\nimport { type Doc, next as automerge, getBackend, type Heads, getHeads } from '@dxos/automerge/automerge';\nimport {\n type DocHandle,\n Repo,\n type DocumentId,\n type PeerId,\n type StorageAdapterInterface,\n type DocHandleChangePayload,\n} from '@dxos/automerge/automerge-repo';\nimport { type Stream } from '@dxos/codec-protobuf';\nimport { Context, type Lifecycle } from '@dxos/context';\nimport { type SpaceDoc } from '@dxos/echo-protocol';\nimport { type IndexMetadataStore } from '@dxos/indexing';\nimport { invariant } from '@dxos/invariant';\nimport { PublicKey } from '@dxos/keys';\nimport { type SubLevelDB } from '@dxos/kv-store';\nimport { log } from '@dxos/log';\nimport { idCodec } from '@dxos/protocols';\nimport {\n type FlushRequest,\n type HostInfo,\n type SyncRepoRequest,\n type SyncRepoResponse,\n} from '@dxos/protocols/proto/dxos/echo/service';\nimport { type Directory } from '@dxos/random-access-storage';\nimport { type AutomergeReplicator } from '@dxos/teleport-extension-automerge-replicator';\nimport { trace } from '@dxos/tracing';\nimport { ComplexMap, ComplexSet, defaultMap, mapValues } from '@dxos/util';\n\nimport { EchoNetworkAdapter } from './echo-network-adapter';\nimport { type EchoReplicator } from './echo-replicator';\nimport { type BeforeSaveParams, LevelDBStorageAdapter } from './leveldb-storage-adapter';\nimport { LocalHostNetworkAdapter } from './local-host-network-adapter';\nimport { MeshNetworkAdapter } from './mesh-network-adapter';\nimport { levelMigration } from './migrations';\n\n// TODO: Remove\nexport type { DocumentId };\n\nexport type AutomergeHostParams = {\n db: SubLevelDB;\n /**\n * For migration purposes.\n */\n directory?: Directory;\n\n indexMetadataStore: IndexMetadataStore;\n};\n\n@trace.resource()\nexport class AutomergeHost {\n private readonly _indexMetadataStore: IndexMetadataStore;\n private readonly _ctx = new Context();\n private readonly _directory?: Directory;\n private readonly _db: SubLevelDB;\n private readonly _echoNetworkAdapter = new EchoNetworkAdapter();\n\n private _repo!: Repo;\n private _meshNetwork!: MeshNetworkAdapter;\n private _clientNetwork!: LocalHostNetworkAdapter;\n private _storage!: StorageAdapterInterface & Lifecycle;\n\n @trace.info()\n private _peerId!: string;\n\n /**\n * spaceKey -> deviceKey[]\n */\n private readonly _authorizedDevices = new ComplexMap<PublicKey, ComplexSet<PublicKey>>(PublicKey.hash);\n\n public _requestedDocs = new Set<string>();\n\n constructor({ directory, db, indexMetadataStore }: AutomergeHostParams) {\n this._directory = directory;\n this._db = db;\n this._indexMetadataStore = indexMetadataStore;\n }\n\n async open() {\n // TODO(mykola): remove this before 0.6 release.\n this._directory && (await levelMigration({ db: this._db, directory: this._directory }));\n this._storage = new LevelDBStorageAdapter({\n db: this._db,\n callbacks: {\n beforeSave: async (params) => this._beforeSave(params),\n afterSave: async () => this._afterSave(),\n },\n });\n await this._storage.open?.();\n this._peerId = `host-${PublicKey.random().toHex()}` as PeerId;\n\n this._meshNetwork = new MeshNetworkAdapter();\n this._clientNetwork = new LocalHostNetworkAdapter();\n\n this._repo = new Repo({\n peerId: this._peerId as PeerId,\n network: [this._clientNetwork, this._meshNetwork, this._echoNetworkAdapter],\n storage: this._storage,\n\n // TODO(dmaretskyi): Share based on HALO permissions and space affinity.\n // Hosts, running in the worker, don't share documents unless requested by other peers.\n sharePolicy: async (peerId /* device key */, documentId /* space key */) => {\n if (peerId.startsWith('client-')) {\n return false; // Only send docs to clients if they are requested.\n }\n\n if (!documentId) {\n return false;\n }\n\n const peerMetadata = this.repo.peerMetadataByPeerId[peerId];\n if ((peerMetadata as any)?.dxos_peerSource === 'EchoNetworkAdapter') {\n return this._echoNetworkAdapter.shouldAdvertize(peerId, { documentId });\n }\n\n const doc = this._repo.handles[documentId]?.docSync();\n if (!doc) {\n const isRequested = this._requestedDocs.has(`automerge:${documentId}`);\n log('doc share policy check', { peerId, documentId, isRequested });\n return isRequested;\n }\n\n try {\n const spaceKey = getSpaceKeyFromDoc(doc);\n if (!spaceKey) {\n log('space key not found for share policy check', { peerId, documentId });\n return false;\n }\n\n const authorizedDevices = this._authorizedDevices.get(PublicKey.from(spaceKey));\n\n // TODO(mykola): Hack, stop abusing `peerMetadata` field.\n const deviceKeyHex = (peerMetadata as any)?.dxos_deviceKey;\n if (!deviceKeyHex) {\n log('device key not found for share policy check', { peerId, documentId });\n return false;\n }\n const deviceKey = PublicKey.from(deviceKeyHex);\n\n const isAuthorized = authorizedDevices?.has(deviceKey) ?? false;\n log('share policy check', {\n localPeer: this._peerId,\n remotePeer: peerId,\n documentId,\n deviceKey,\n spaceKey,\n isAuthorized,\n });\n return isAuthorized;\n } catch (err) {\n log.catch(err);\n return false;\n }\n },\n });\n this._clientNetwork.ready();\n this._meshNetwork.ready();\n await this._echoNetworkAdapter.open();\n\n await this._clientNetwork.whenConnected();\n await this._echoNetworkAdapter.whenConnected();\n }\n\n async close() {\n await this._storage.close?.();\n await this._clientNetwork.close();\n await this._echoNetworkAdapter.close();\n await this._ctx.dispose();\n }\n\n get repo(): Repo {\n return this._repo;\n }\n\n async addReplicator(replicator: EchoReplicator) {\n await this._echoNetworkAdapter.addReplicator(replicator);\n }\n\n async removeReplicator(replicator: EchoReplicator) {\n await this._echoNetworkAdapter.removeReplicator(replicator);\n }\n\n private async _beforeSave({ path, batch }: BeforeSaveParams) {\n const handle = this._repo.handles[path[0] as DocumentId];\n if (!handle) {\n return;\n }\n const doc = handle.docSync();\n if (!doc) {\n return;\n }\n\n const lastAvailableHash = getHeads(doc);\n\n const objectIds = Object.keys(doc.objects ?? {});\n const encodedIds = objectIds.map((objectId) => idCodec.encode({ documentId: handle.documentId, objectId }));\n const idToLastHash = new Map(encodedIds.map((id) => [id, lastAvailableHash]));\n this._indexMetadataStore.markDirty(idToLastHash, batch);\n }\n\n /**\n * Called by AutomergeStorageAdapter after levelDB batch commit.\n */\n private async _afterSave() {\n this._indexMetadataStore.notifyMarkedDirty();\n }\n\n @trace.info({ depth: null })\n private _automergeDocs() {\n return mapValues(this._repo.handles, (handle) => ({\n state: handle.state,\n hasDoc: !!handle.docSync(),\n heads: handle.docSync() ? automerge.getHeads(handle.docSync()) : null,\n data:\n handle.docSync() &&\n mapValues(handle.docSync(), (value, key) => {\n try {\n switch (key) {\n case 'access':\n case 'links':\n return value;\n case 'objects':\n return Object.keys(value as any);\n default:\n return `${value}`;\n }\n } catch (err) {\n return `${err}`;\n }\n }),\n }));\n }\n\n @trace.info({ depth: null })\n private _automergePeers() {\n return this._repo.peers;\n }\n\n //\n // Methods for client-services.\n //\n @trace.span({ showInBrowserTimeline: true })\n async flush({ states }: FlushRequest): Promise<void> {\n // Note: Wait for all requested documents to be loaded/synced from thin-client.\n await Promise.all(\n states?.map(async ({ heads, documentId }) => {\n invariant(heads, 'heads are required for flush');\n const handle = this.repo.handles[documentId as DocumentId] ?? this._repo.find(documentId as DocumentId);\n await waitForHeads(handle, heads);\n }) ?? [],\n );\n\n await this._repo.flush(states?.map(({ documentId }) => documentId as DocumentId));\n }\n\n syncRepo(request: SyncRepoRequest): Stream<SyncRepoResponse> {\n return this._clientNetwork.syncRepo(request);\n }\n\n sendSyncMessage(request: SyncRepoRequest): Promise<void> {\n return this._clientNetwork.sendSyncMessage(request);\n }\n\n async getHostInfo(): Promise<HostInfo> {\n return this._clientNetwork.getHostInfo();\n }\n\n //\n // Mesh replication.\n //\n\n createExtension(): AutomergeReplicator {\n return this._meshNetwork.createExtension();\n }\n\n authorizeDevice(spaceKey: PublicKey, deviceKey: PublicKey) {\n log('authorizeDevice', { spaceKey, deviceKey });\n defaultMap(this._authorizedDevices, spaceKey, () => new ComplexSet(PublicKey.hash)).add(deviceKey);\n }\n}\n\nexport const getSpaceKeyFromDoc = (doc: any): string | null => {\n // experimental_spaceKey is set on old documents, new ones are created with doc.access.spaceKey\n const rawSpaceKey = doc.access?.spaceKey ?? doc.experimental_spaceKey;\n if (rawSpaceKey == null) {\n return null;\n }\n\n return String(rawSpaceKey);\n};\n\nconst waitForHeads = async (handle: DocHandle<SpaceDoc>, heads: Heads) => {\n await handle.whenReady();\n const unavailableHeads = new Set(heads);\n\n await Event.wrap<DocHandleChangePayload<SpaceDoc>>(handle, 'change').waitForCondition(() => {\n // Check if unavailable heads became available.\n for (const changeHash of unavailableHeads.values()) {\n if (changeIsPresentInDoc(handle.docSync(), changeHash)) {\n unavailableHeads.delete(changeHash);\n }\n }\n\n if (unavailableHeads.size === 0) {\n return true;\n }\n return false;\n });\n};\n\nconst changeIsPresentInDoc = (doc: Doc<any>, changeHash: string): boolean => {\n return !!getBackend(doc).getChangeByHash(changeHash);\n};\n", "//\n// Copyright 2024 DXOS.org\n//\n\nimport { Trigger, synchronized } from '@dxos/async';\nimport { type Message, NetworkAdapter, type PeerId, type PeerMetadata } from '@dxos/automerge/automerge-repo';\nimport { LifecycleState } from '@dxos/context';\nimport { invariant } from '@dxos/invariant';\nimport { log } from '@dxos/log';\n\nimport { type EchoReplicator, type ReplicatorConnection, type ShouldAdvertizeParams } from './echo-replicator';\n\n/**\n * Manages a set of {@link EchoReplicator} instances.\n */\nexport class EchoNetworkAdapter extends NetworkAdapter {\n private readonly _replicators = new Set<EchoReplicator>();\n /**\n * Remote peer id -> connection.\n */\n private readonly _connections = new Map<PeerId, ConnectionEntry>();\n private _lifecycleState: LifecycleState = LifecycleState.CLOSED;\n private readonly _connected = new Trigger();\n\n override connect(peerId: PeerId, peerMetadata?: PeerMetadata | undefined): void {\n this.peerId = peerId;\n this.peerMetadata = peerMetadata;\n this._connected.wake();\n }\n\n override send(message: Message): void {\n const connectionEntry = this._connections.get(message.targetId);\n if (!connectionEntry) {\n throw new Error('Connection not found.');\n }\n\n // TODO(dmaretskyi): Find a way to enforce backpressure on AM-repo.\n connectionEntry.writer.write(message).catch((err) => {\n if (connectionEntry.isOpen) {\n log.catch(err);\n }\n });\n }\n\n override disconnect(): void {\n // No-op\n }\n\n @synchronized\n async open() {\n invariant(this._lifecycleState === LifecycleState.CLOSED);\n this._lifecycleState = LifecycleState.OPEN;\n\n this.emit('ready', {\n network: this,\n });\n }\n\n @synchronized\n async close() {\n invariant(this._lifecycleState === LifecycleState.OPEN);\n\n for (const replicator of this._replicators) {\n await replicator.disconnect();\n }\n this._replicators.clear();\n\n this._lifecycleState = LifecycleState.CLOSED;\n }\n\n async whenConnected() {\n await this._connected.wait({ timeout: 10_000 });\n }\n\n @synchronized\n async addReplicator(replicator: EchoReplicator) {\n invariant(this.peerId);\n invariant(!this._replicators.has(replicator));\n\n await replicator.connect({\n peerId: this.peerId,\n onConnectionOpen: this._onConnectionOpen.bind(this),\n onConnectionClosed: this._onConnectionClosed.bind(this),\n });\n }\n\n @synchronized\n async removeReplicator(replicator: EchoReplicator) {\n invariant(this._replicators.has(replicator));\n await replicator.disconnect();\n }\n\n async shouldAdvertize(peerId: PeerId, params: ShouldAdvertizeParams): Promise<boolean> {\n const connection = this._connections.get(peerId);\n if (!connection) {\n return false;\n }\n\n return connection.connection.shouldAdvertize(params);\n }\n\n private _onConnectionOpen(connection: ReplicatorConnection) {\n invariant(!this._connections.has(connection.peerId as PeerId));\n const reader = connection.readable.getReader();\n const writer = connection.writable.getWriter();\n const connectionEntry: ConnectionEntry = { connection, reader, writer, isOpen: true };\n this._connections.set(connection.peerId as PeerId, connectionEntry);\n\n queueMicrotask(async () => {\n try {\n while (true) {\n // TODO(dmaretskyi): Find a way to enforce backpressure on AM-repo.\n const { done, value } = await reader.read();\n if (done) {\n break;\n }\n\n this.emit('message', value);\n }\n } catch (err) {\n if (connectionEntry.isOpen) {\n log.catch(err);\n }\n }\n });\n\n this.emit('peer-candidate', {\n peerId: connection.peerId as PeerId,\n peerMetadata: {\n // TODO(dmaretskyi): Refactor this.\n dxos_peerSource: 'EchoNetworkAdapter',\n } as any,\n });\n }\n\n private _onConnectionClosed(connection: ReplicatorConnection) {\n const entry = this._connections.get(connection.peerId as PeerId);\n invariant(entry);\n\n entry.isOpen = false;\n this.emit('peer-disconnected', { peerId: connection.peerId as PeerId });\n\n void entry.reader.cancel().catch((err) => log.catch(err));\n void entry.writer.abort().catch((err) => log.catch(err));\n\n this._connections.delete(connection.peerId as PeerId);\n }\n}\n\ntype ConnectionEntry = {\n connection: ReplicatorConnection;\n reader: ReadableStreamDefaultReader<Message>;\n writer: WritableStreamDefaultWriter<Message>;\n isOpen: boolean;\n};\n", "//\n// Copyright 2024 DXOS.org\n// s\n\nimport { type MixedEncoding } from 'level-transcoder';\n\nimport { type StorageAdapterInterface, type Chunk, type StorageKey } from '@dxos/automerge/automerge-repo';\nimport { LifecycleState, Resource } from '@dxos/context';\nimport { type BatchLevel, type SubLevelDB } from '@dxos/kv-store';\nimport { type MaybePromise } from '@dxos/util';\n\nexport type LevelDBStorageAdapterParams = {\n db: SubLevelDB;\n callbacks?: StorageCallbacks;\n};\n\nexport type BeforeSaveParams = { path: StorageKey; batch: BatchLevel };\n\nexport interface StorageCallbacks {\n beforeSave(params: BeforeSaveParams): MaybePromise<void>;\n afterSave(path: StorageKey): MaybePromise<void>;\n}\n\nexport class LevelDBStorageAdapter extends Resource implements StorageAdapterInterface {\n constructor(private readonly _params: LevelDBStorageAdapterParams) {\n super();\n }\n\n async load(keyArray: StorageKey): Promise<Uint8Array | undefined> {\n try {\n if (this._lifecycleState !== LifecycleState.OPEN) {\n // TODO(mykola): this should be an error.\n return undefined;\n }\n return await this._params.db.get<StorageKey, Uint8Array>(keyArray, { ...encodingOptions });\n } catch (err: any) {\n if (isLevelDbNotFoundError(err)) {\n return undefined;\n }\n throw err;\n }\n }\n\n async save(keyArray: StorageKey, binary: Uint8Array): Promise<void> {\n if (this._lifecycleState !== LifecycleState.OPEN) {\n return undefined;\n }\n const batch = this._params.db.batch();\n\n await this._params.callbacks?.beforeSave?.({ path: keyArray, batch });\n batch.put<StorageKey, Uint8Array>(keyArray, Buffer.from(binary), {\n ...encodingOptions,\n });\n await batch.write();\n\n await this._params.callbacks?.afterSave?.(keyArray);\n }\n\n async remove(keyArray: StorageKey): Promise<void> {\n if (this._lifecycleState !== LifecycleState.OPEN) {\n return undefined;\n }\n await this._params.db.del<StorageKey>(keyArray, { ...encodingOptions });\n }\n\n async loadRange(keyPrefix: StorageKey): Promise<Chunk[]> {\n if (this._lifecycleState !== LifecycleState.OPEN) {\n return [];\n }\n const result: Chunk[] = [];\n for await (const [key, value] of this._params.db.iterator<StorageKey, Uint8Array>({\n gte: keyPrefix,\n lte: [...keyPrefix, '\\uffff'],\n ...encodingOptions,\n })) {\n result.push({\n key,\n data: value,\n });\n }\n return result;\n }\n\n async removeRange(keyPrefix: StorageKey): Promise<void> {\n if (this._lifecycleState !== LifecycleState.OPEN) {\n return undefined;\n }\n const batch = this._params.db.batch();\n\n for await (const [key] of this._params.db.iterator<StorageKey, Uint8Array>({\n gte: keyPrefix,\n lte: [...keyPrefix, '\\uffff'],\n ...encodingOptions,\n })) {\n batch.del<StorageKey>(key, { ...encodingOptions });\n }\n await batch.write();\n }\n}\n\nconst keyEncoder: MixedEncoding<StorageKey, Uint8Array, StorageKey> = {\n encode: (key: StorageKey): Uint8Array =>\n Buffer.from(key.map((k) => k.replaceAll('%', '%25').replaceAll('-', '%2D')).join('-')),\n decode: (key: Uint8Array): StorageKey =>\n Buffer.from(key)\n .toString()\n .split('-')\n .map((k) => k.replaceAll('%2D', '-').replaceAll('%25', '%')),\n format: 'buffer',\n};\n\nexport const encodingOptions = {\n keyEncoding: keyEncoder,\n valueEncoding: 'buffer',\n};\n\nconst isLevelDbNotFoundError = (err: any): boolean => err.code === 'LEVEL_NOT_FOUND';\n", "//\n// Copyright 2024 DXOS.org\n//\n\nimport { Trigger } from '@dxos/async';\nimport { NetworkAdapter, type Message, type PeerId, cbor } from '@dxos/automerge/automerge-repo';\nimport { Stream } from '@dxos/codec-protobuf';\nimport { invariant } from '@dxos/invariant';\nimport { type HostInfo, type SyncRepoRequest, type SyncRepoResponse } from '@dxos/protocols/proto/dxos/echo/service';\n\ntype ClientSyncState = {\n connected: boolean;\n send: (message: Message) => void;\n disconnect: () => void;\n};\n\n/**\n * Used to replicate with apps running on the same device.\n */\nexport class LocalHostNetworkAdapter extends NetworkAdapter {\n private readonly _peers: Map<PeerId, ClientSyncState> = new Map();\n\n /**\n * Emits `ready` event. That signals to `Repo` that it can start using the adapter.\n */\n ready() {\n // NOTE: Emitting `ready` event in NetworkAdapter`s constructor causes a race condition\n // because `Repo` waits for `ready` event (which it never receives) before it starts using the adapter.\n this.emit('ready', {\n network: this,\n });\n }\n\n private readonly _connected = new Trigger();\n private _isConnected: boolean = false;\n\n /**\n * Called by `Repo` to connect to the network.\n *\n * @param peerId Our peer Id.\n */\n override connect(peerId: PeerId): void {\n this.peerId = peerId;\n this._isConnected = true;\n this._connected.wake();\n // No-op. Client always connects first\n }\n\n override send(message: Message): void {\n const peer = this._peers.get(message.targetId);\n invariant(peer, 'Peer not found.');\n peer.send(message);\n }\n\n async close() {\n this._peers.forEach((peer) => peer.disconnect());\n this.emit('close');\n }\n\n override disconnect(): void {\n // TODO(mykola): `disconnect` is not used anywhere in `Repo` from `@automerge/automerge-repo`. Should we remove it?\n // No-op\n }\n\n async whenConnected(): Promise<void> {\n await this._connected.wait({ timeout: 10_000 });\n }\n\n syncRepo({ id, syncMessage }: SyncRepoRequest): Stream<SyncRepoResponse> {\n const peerId = this._getPeerId(id);\n\n return new Stream(({ next, close }) => {\n invariant(!this._peers.has(peerId), 'Peer already connected.');\n this._peers.set(peerId, {\n connected: true,\n send: (message) => {\n next({\n syncMessage: cbor.encode(message),\n });\n },\n disconnect: () => {\n this._peers.delete(peerId);\n close();\n this.emit('peer-disconnected', {\n peerId,\n });\n },\n });\n\n invariant(this._isConnected);\n this.emit('peer-candidate', {\n peerMetadata: {},\n peerId,\n });\n });\n }\n\n async sendSyncMessage({ id, syncMessage }: SyncRepoRequest): Promise<void> {\n invariant(this._isConnected);\n const message = cbor.decode(syncMessage!) as Message;\n this.emit('message', message);\n }\n\n async getHostInfo(): Promise<HostInfo> {\n invariant(this._isConnected);\n invariant(this.peerId, 'Peer id not set.');\n return {\n peerId: this.peerId,\n };\n }\n\n private _getPeerId(id: string): PeerId {\n return id as PeerId;\n }\n}\n", "//\n// Copyright 2024 DXOS.org\n//\n\nimport { Trigger } from '@dxos/async';\nimport { NetworkAdapter, type Message, type PeerId, cbor } from '@dxos/automerge/automerge-repo';\nimport { invariant } from '@dxos/invariant';\nimport { log } from '@dxos/log';\nimport { type PeerInfo } from '@dxos/protocols/proto/dxos/mesh/teleport/automerge';\nimport { AutomergeReplicator } from '@dxos/teleport-extension-automerge-replicator';\n\n/**\n * Used to replicate with other peers over the network.\n */\nexport class MeshNetworkAdapter extends NetworkAdapter {\n private readonly _extensions: Map<string, AutomergeReplicator> = new Map();\n private _connected = new Trigger();\n\n /**\n * Emits `ready` event. That signals to `Repo` that it can start using the adapter.\n */\n ready() {\n // NOTE: Emitting `ready` event in NetworkAdapter`s constructor causes a race condition\n // because `Repo` waits for `ready` event (which it never receives) before it starts using the adapter.\n this.emit('ready', {\n network: this,\n });\n }\n\n override connect(peerId: PeerId): void {\n this.peerId = peerId;\n this._connected.wake();\n }\n\n override send(message: Message): void {\n const receiverId = message.targetId;\n const extension = this._extensions.get(receiverId);\n invariant(extension, 'Extension not found.');\n extension.sendSyncMessage({ payload: cbor.encode(message) }).catch((err) => log.catch(err));\n }\n\n override disconnect(): void {\n // No-op\n }\n\n createExtension(): AutomergeReplicator {\n invariant(this.peerId, 'Peer id not set.');\n\n let peerInfo: PeerInfo;\n const extension = new AutomergeReplicator(\n {\n peerId: this.peerId,\n },\n {\n onStartReplication: async (info, remotePeerId /** Teleport ID */) => {\n await this._connected.wait();\n\n // Note: We store only one extension per peer.\n // There can be a case where two connected peers have more than one teleport connection between them\n // and each of them uses different teleport connections to send messages.\n // It works because we receive messages from all teleport connections and Automerge Repo dedup them.\n // TODO(mykola): Use only one teleport connection per peer.\n\n // TODO(dmaretskyi): Critical bug.\n // - two peers get connected via swarm 1\n // - they get connected via swarm 2\n // - swarm 1 gets disconnected\n // - automerge repo thinks that peer 2 got disconnected even though swarm 2 is still active\n\n log('onStartReplication', { id: info.id, thisPeerId: this.peerId, remotePeerId: remotePeerId.toHex() });\n if (!this._extensions.has(info.id)) {\n peerInfo = info;\n // TODO(mykola): Fix race condition?\n this._extensions.set(info.id, extension);\n\n log('peer-candidate', { id: info.id, thisPeerId: this.peerId, remotePeerId: remotePeerId.toHex() });\n this.emit('peer-candidate', {\n // TODO(mykola): Hack, stop abusing `peerMetadata` field.\n peerMetadata: {\n dxos_deviceKey: remotePeerId.toHex(),\n } as any,\n peerId: info.id as PeerId,\n });\n }\n },\n onSyncMessage: async ({ payload }) => {\n if (!peerInfo) {\n return;\n }\n const message = cbor.decode(payload) as Message;\n // Note: automerge Repo dedup messages.\n this.emit('message', message);\n },\n onClose: async () => {\n if (!peerInfo) {\n return;\n }\n this.emit('peer-disconnected', {\n peerId: peerInfo.id as PeerId,\n });\n this._extensions.delete(peerInfo.id);\n },\n },\n );\n return extension;\n }\n}\n", "//\n// Copyright 2024 DXOS.org\n//\n\nimport { type StorageKey } from '@dxos/automerge/automerge-repo';\nimport { IndexedDBStorageAdapter } from '@dxos/automerge/automerge-repo-storage-indexeddb';\nimport { type SubLevelDB } from '@dxos/kv-store';\nimport { log } from '@dxos/log';\nimport { StorageType, type Directory } from '@dxos/random-access-storage';\n\nimport { AutomergeStorageAdapter } from './automerge-storage-adapter';\nimport { encodingOptions } from './leveldb-storage-adapter';\n\nexport const levelMigration = async ({ db, directory }: { db: SubLevelDB; directory: Directory }) => {\n // Note: Make automigration from previous storage to leveldb here.\n const isNewLevel = !(await db\n .iterator<StorageKey, Uint8Array>({\n ...encodingOptions,\n })\n .next());\n\n if (!isNewLevel) {\n return;\n }\n\n const oldStorageAdapter =\n directory.type === StorageType.IDB\n ? new IndexedDBStorageAdapter(directory.path, 'data')\n : new AutomergeStorageAdapter(directory);\n\n const chunks = await oldStorageAdapter.loadRange([]);\n if (chunks.length === 0) {\n return;\n }\n const batch = db.batch();\n log.info('found chunks on old storage adapter', { chunks: chunks.length });\n for (const { key, data } of await oldStorageAdapter.loadRange([])) {\n data && batch.put<StorageKey, Uint8Array>(key, data, { ...encodingOptions });\n }\n await batch.write();\n};\n", "//\n// Copyright 2024 DXOS.org\n//\n//\n// Copyright 2023 DXOS.org\n//\n\nimport { type Chunk, type StorageKey, type StorageAdapterInterface } from '@dxos/automerge/automerge-repo';\nimport { type Directory } from '@dxos/random-access-storage';\nimport { arrayToBuffer, bufferToArray } from '@dxos/util';\n\nexport class AutomergeStorageAdapter implements StorageAdapterInterface {\n // TODO(mykola): Hack for restricting automerge Repo to access storage if Host is `closed`.\n // Automerge Repo do not have any lifetime management.\n private _state: 'opened' | 'closed' = 'opened';\n\n constructor(private readonly _directory: Directory) {}\n\n async load(key: StorageKey): Promise<Uint8Array | undefined> {\n if (this._state !== 'opened') {\n return undefined;\n }\n const filename = this._getFilename(key);\n const file = this._directory.getOrCreateFile(filename);\n const { size } = await file.stat();\n if (!size || size === 0) {\n return undefined;\n }\n const buffer = await file.read(0, size);\n return bufferToArray(buffer);\n }\n\n async save(key: StorageKey, data: Uint8Array): Promise<void> {\n if (this._state !== 'opened') {\n return undefined;\n }\n const filename = this._getFilename(key);\n const file = this._directory.getOrCreateFile(filename);\n await file.write(0, arrayToBuffer(data));\n await file.truncate?.(data.length);\n\n await file.flush?.();\n }\n\n async remove(key: StorageKey): Promise<void> {\n if (this._state !== 'opened') {\n return undefined;\n }\n // TODO(dmaretskyi): Better deletion.\n const filename = this._getFilename(key);\n const file = this._directory.getOrCreateFile(filename);\n await file.destroy();\n }\n\n async loadRange(keyPrefix: StorageKey): Promise<Chunk[]> {\n if (this._state !== 'opened') {\n return [];\n }\n const filename = this._getFilename(keyPrefix);\n const entries = await this._directory.list();\n return Promise.all(\n entries\n .filter((entry) => entry.startsWith(filename))\n .map(async (entry): Promise<Chunk> => {\n const file = this._directory.getOrCreateFile(entry);\n const { size } = await file.stat();\n const buffer = await file.read(0, size);\n return {\n key: this._getKeyFromFilename(entry),\n data: bufferToArray(buffer),\n };\n }),\n );\n }\n\n async removeRange(keyPrefix: StorageKey): Promise<void> {\n if (this._state !== 'opened') {\n return undefined;\n }\n const filename = this._getFilename(keyPrefix);\n const entries = await this._directory.list();\n await Promise.all(\n entries\n .filter((entry) => entry.startsWith(filename))\n .map(async (entry): Promise<void> => {\n const file = this._directory.getOrCreateFile(entry);\n await file.destroy();\n }),\n );\n }\n\n async close(): Promise<void> {\n this._state = 'closed';\n }\n\n private _getFilename(key: StorageKey): string {\n return key.map((k) => k.replaceAll('%', '%25').replaceAll('-', '%2D')).join('-');\n }\n\n private _getKeyFromFilename(filename: string): StorageKey {\n return filename.split('-').map((k) => k.replaceAll('%2D', '-').replaceAll('%25', '%'));\n }\n}\n", "//\n// Copyright 2024 DXOS.org\n//\n\nimport { Event } from '@dxos/async';\nimport { type DocHandle, type AutomergeUrl, type DocumentId, type Repo } from '@dxos/automerge/automerge-repo';\nimport { cancelWithContext, type Context } from '@dxos/context';\nimport { warnAfterTimeout } from '@dxos/debug';\nimport { type SpaceState, type SpaceDoc } from '@dxos/echo-protocol';\nimport { invariant } from '@dxos/invariant';\nimport { type PublicKey } from '@dxos/keys';\nimport { log } from '@dxos/log';\nimport { trace } from '@dxos/tracing';\n\ntype SpaceDocumentLinks = SpaceDoc['links'];\n\nexport interface AutomergeDocumentLoader {\n onObjectDocumentLoaded: Event<ObjectDocumentLoaded>;\n\n getAllHandles(): DocHandle<SpaceDoc>[];\n\n loadSpaceRootDocHandle(ctx: Context, spaceState: SpaceState): Promise<void>;\n loadObjectDocument(objectId: string | string[]): void;\n getSpaceRootDocHandle(): DocHandle<SpaceDoc>;\n createDocumentForObject(objectId: string): DocHandle<SpaceDoc>;\n onObjectLinksUpdated(links: SpaceDocumentLinks): void;\n onObjectBoundToDocument(handle: DocHandle<SpaceDoc>, objectId: string): void;\n\n /**\n * @returns objectIds for which we had document handles or were loading one.\n */\n clearHandleReferences(): string[];\n}\n\n/**\n * Manages object <-> docHandle binding and automerge document loading.\n */\n@trace.resource()\nexport class AutomergeDocumentLoaderImpl implements AutomergeDocumentLoader {\n private _spaceRootDocHandle: DocHandle<SpaceDoc> | null = null;\n /**\n * An object id pointer to a handle of the document where the object is stored inline.\n */\n private readonly _objectDocumentHandles = new Map<string, DocHandle<SpaceDoc>>();\n /**\n * If object was requested via loadObjectDocument but root document links weren't updated yet\n * loading will be triggered in onObjectLinksUpdated callback.\n */\n private readonly _objectsPendingDocumentLoad = new Set<string>();\n\n public readonly onObjectDocumentLoaded = new Event<ObjectDocumentLoaded>();\n\n constructor(\n private readonly _spaceKey: PublicKey,\n private readonly _repo: Repo,\n ) {}\n\n getAllHandles(): DocHandle<SpaceDoc>[] {\n return this._spaceRootDocHandle != null\n ? [this._spaceRootDocHandle, ...new Set(this._objectDocumentHandles.values())]\n : [];\n }\n\n @trace.span({ showInBrowserTimeline: true })\n public async loadSpaceRootDocHandle(ctx: Context, spaceState: SpaceState): Promise<void> {\n if (this._spaceRootDocHandle != null) {\n return;\n }\n if (!spaceState.rootUrl) {\n log.error('Database opened with no rootUrl', { spaceKey: this._spaceKey });\n this._createContextBoundSpaceRootDocument(ctx);\n } else {\n const existingDocHandle = await this._initDocHandle(ctx, spaceState.rootUrl);\n const doc = existingDocHandle.docSync();\n invariant(doc);\n if (doc.access == null) {\n this._initDocAccess(existingDocHandle);\n }\n this._spaceRootDocHandle = existingDocHandle;\n }\n }\n\n public loadObjectDocument(objectIdOrMany: string | string[]) {\n const objectIds = Array.isArray(objectIdOrMany) ? objectIdOrMany : [objectIdOrMany];\n let hasUrlsToLoad = false;\n const urlsToLoad: SpaceDoc['links'] = {};\n for (const objectId of objectIds) {\n invariant(this._spaceRootDocHandle);\n if (this._objectDocumentHandles.has(objectId) || this._objectsPendingDocumentLoad.has(objectId)) {\n continue;\n }\n const spaceRootDoc = this._spaceRootDocHandle.docSync();\n invariant(spaceRootDoc);\n const documentUrl = (spaceRootDoc.links ?? {})[objectId];\n if (documentUrl == null) {\n this._objectsPendingDocumentLoad.add(objectId);\n log.info('loading delayed until object links are initialized', { objectId });\n } else {\n urlsToLoad[objectId] = documentUrl;\n hasUrlsToLoad = true;\n }\n }\n if (hasUrlsToLoad) {\n this._loadLinkedObjects(urlsToLoad);\n }\n }\n\n public onObjectLinksUpdated(links: SpaceDocumentLinks) {\n if (!links) {\n return;\n }\n const linksAwaitingLoad = Object.entries(links).filter(([objectId]) =>\n this._objectsPendingDocumentLoad.has(objectId),\n );\n this._loadLinkedObjects(Object.fromEntries(linksAwaitingLoad));\n linksAwaitingLoad.forEach(([objectId]) => this._objectsPendingDocumentLoad.delete(objectId));\n }\n\n public getSpaceRootDocHandle(): DocHandle<SpaceDoc> {\n invariant(this._spaceRootDocHandle);\n return this._spaceRootDocHandle;\n }\n\n public createDocumentForObject(objectId: string): DocHandle<SpaceDoc> {\n invariant(this._spaceRootDocHandle);\n const spaceDocHandle = this._repo.create<SpaceDoc>();\n this._initDocAccess(spaceDocHandle);\n this.onObjectBoundToDocument(spaceDocHandle, objectId);\n this._spaceRootDocHandle.change((newDoc: SpaceDoc) => {\n newDoc.links ??= {};\n newDoc.links[objectId] = spaceDocHandle.url;\n });\n return spaceDocHandle;\n }\n\n public onObjectBoundToDocument(handle: DocHandle<SpaceDoc>, objectId: string) {\n this._objectDocumentHandles.set(objectId, handle);\n }\n\n public clearHandleReferences(): string[] {\n const objectsWithHandles = [...this._objectDocumentHandles.keys()];\n this._objectDocumentHandles.clear();\n this._spaceRootDocHandle = null;\n return objectsWithHandles;\n }\n\n private _loadLinkedObjects(links: SpaceDocumentLinks) {\n if (!links) {\n return;\n }\n for (const [objectId, automergeUrl] of Object.entries(links)) {\n const logMeta = { objectId, automergeUrl };\n const objectDocumentHandle = this._objectDocumentHandles.get(objectId);\n if (objectDocumentHandle != null && objectDocumentHandle.url !== automergeUrl) {\n log.warn('object already inlined in a different document, ignoring the link', {\n ...logMeta,\n actualDocumentUrl: objectDocumentHandle.url,\n });\n continue;\n }\n if (objectDocumentHandle?.url === automergeUrl) {\n log.warn('object document was already loaded', logMeta);\n continue;\n }\n const handle = this._repo.find<SpaceDoc>(automergeUrl as DocumentId);\n log.debug('document loading triggered', logMeta);\n this._objectDocumentHandles.set(objectId, handle);\n void this._createObjectOnDocumentLoad(handle, objectId);\n }\n }\n\n private async _initDocHandle(ctx: Context, url: string) {\n const docHandle = this._repo.find<SpaceDoc>(url as DocumentId);\n while (true) {\n try {\n await warnAfterTimeout(5_000, 'Automerge root doc load timeout (AutomergeDb)', async () => {\n await cancelWithContext(ctx, docHandle.whenReady()); // TODO(dmaretskyi): Temporary 5s timeout for debugging.\n });\n break;\n } catch (err) {\n if (`${err}`.includes('Timeout')) {\n log.info('wraparound', { id: docHandle.documentId, state: docHandle.state });\n continue;\n }\n\n throw err;\n }\n }\n\n if (docHandle.state === 'unavailable') {\n throw new Error('Automerge document is unavailable');\n }\n\n return docHandle;\n }\n\n private _createContextBoundSpaceRootDocument(ctx: Context) {\n const docHandle = this._repo.create<SpaceDoc>();\n this._spaceRootDocHandle = docHandle;\n ctx.onDispose(() => {\n docHandle.delete();\n this._spaceRootDocHandle = null;\n });\n }\n\n private _initDocAccess(handle: DocHandle<SpaceDoc>) {\n handle.change((newDoc: SpaceDoc) => {\n newDoc.access ??= { spaceKey: this._spaceKey.toHex() };\n newDoc.access.spaceKey = this._spaceKey.toHex();\n });\n }\n\n private async _createObjectOnDocumentLoad(handle: DocHandle<SpaceDoc>, objectId: string) {\n try {\n await handle.doc(['ready']);\n const logMeta = { objectId, docUrl: handle.url };\n if (this.onObjectDocumentLoaded.listenerCount() === 0) {\n log.info('document loaded after all listeners were removed', logMeta);\n return;\n }\n const objectDocHandle = this._objectDocumentHandles.get(objectId);\n if (objectDocHandle?.url !== handle.url) {\n log.warn('object was rebound while a document was loading, discarding handle', logMeta);\n return;\n }\n this.onObjectDocumentLoaded.emit({ handle, objectId });\n } catch (err) {\n const shouldRetryLoading = this.onObjectDocumentLoaded.listenerCount() > 0;\n log.warn('failed to load a document', {\n objectId,\n automergeUrl: handle.url,\n retryLoading: shouldRetryLoading,\n err,\n });\n if (shouldRetryLoading) {\n await this._createObjectOnDocumentLoad(handle, objectId);\n }\n }\n }\n}\n\nexport interface ObjectDocumentLoaded {\n handle: DocHandle<SpaceDoc>;\n objectId: string;\n}\n\nexport interface DocumentChanges {\n createdObjectIds: string[];\n updatedObjectIds: string[];\n objectsToRebind: string[];\n linkedDocuments: {\n [echoId: string]: AutomergeUrl;\n };\n}\n"],
|
|
4
|
+
"sourcesContent": ["//\n// Copyright 2023 DXOS.org\n//\n\nimport { Event } from '@dxos/async';\nimport { type Doc, next as automerge, getBackend, type Heads, getHeads } from '@dxos/automerge/automerge';\nimport {\n type DocHandle,\n Repo,\n type DocumentId,\n type PeerId,\n type StorageAdapterInterface,\n type DocHandleChangePayload,\n} from '@dxos/automerge/automerge-repo';\nimport { type Stream } from '@dxos/codec-protobuf';\nimport { Context, type Lifecycle } from '@dxos/context';\nimport { type SpaceDoc } from '@dxos/echo-protocol';\nimport { type IndexMetadataStore } from '@dxos/indexing';\nimport { invariant } from '@dxos/invariant';\nimport { PublicKey } from '@dxos/keys';\nimport { type SublevelDB } from '@dxos/kv-store';\nimport { log } from '@dxos/log';\nimport { idCodec } from '@dxos/protocols';\nimport {\n type FlushRequest,\n type HostInfo,\n type SyncRepoRequest,\n type SyncRepoResponse,\n} from '@dxos/protocols/proto/dxos/echo/service';\nimport { type Directory } from '@dxos/random-access-storage';\nimport { type AutomergeReplicator } from '@dxos/teleport-extension-automerge-replicator';\nimport { trace } from '@dxos/tracing';\nimport { ComplexMap, ComplexSet, defaultMap, mapValues } from '@dxos/util';\n\nimport { EchoNetworkAdapter } from './echo-network-adapter';\nimport { type EchoReplicator } from './echo-replicator';\nimport { type BeforeSaveParams, LevelDBStorageAdapter } from './leveldb-storage-adapter';\nimport { LocalHostNetworkAdapter } from './local-host-network-adapter';\nimport { MeshNetworkAdapter } from './mesh-network-adapter';\nimport { levelMigration } from './migrations';\n\n// TODO: Remove\nexport type { DocumentId };\n\nexport type AutomergeHostParams = {\n db: SublevelDB;\n /**\n * For migration purposes.\n */\n directory?: Directory;\n\n indexMetadataStore: IndexMetadataStore;\n};\n\n@trace.resource()\nexport class AutomergeHost {\n private readonly _indexMetadataStore: IndexMetadataStore;\n private readonly _ctx = new Context();\n private readonly _directory?: Directory;\n private readonly _db: SublevelDB;\n private readonly _echoNetworkAdapter = new EchoNetworkAdapter();\n\n private _repo!: Repo;\n private _meshNetwork!: MeshNetworkAdapter;\n private _clientNetwork!: LocalHostNetworkAdapter;\n private _storage!: StorageAdapterInterface & Lifecycle;\n\n @trace.info()\n private _peerId!: string;\n\n /**\n * spaceKey -> deviceKey[]\n */\n private readonly _authorizedDevices = new ComplexMap<PublicKey, ComplexSet<PublicKey>>(PublicKey.hash);\n\n public _requestedDocs = new Set<string>();\n\n constructor({ directory, db, indexMetadataStore }: AutomergeHostParams) {\n this._directory = directory;\n this._db = db;\n this._indexMetadataStore = indexMetadataStore;\n }\n\n async open() {\n // TODO(mykola): remove this before 0.6 release.\n this._directory && (await levelMigration({ db: this._db, directory: this._directory }));\n this._storage = new LevelDBStorageAdapter({\n db: this._db,\n callbacks: {\n beforeSave: async (params) => this._beforeSave(params),\n afterSave: async () => this._afterSave(),\n },\n });\n await this._storage.open?.();\n this._peerId = `host-${PublicKey.random().toHex()}` as PeerId;\n\n this._meshNetwork = new MeshNetworkAdapter();\n this._clientNetwork = new LocalHostNetworkAdapter();\n\n this._repo = new Repo({\n peerId: this._peerId as PeerId,\n network: [this._clientNetwork, this._meshNetwork, this._echoNetworkAdapter],\n storage: this._storage,\n\n // TODO(dmaretskyi): Share based on HALO permissions and space affinity.\n // Hosts, running in the worker, don't share documents unless requested by other peers.\n sharePolicy: async (peerId /* device key */, documentId /* space key */) => {\n if (peerId.startsWith('client-')) {\n return false; // Only send docs to clients if they are requested.\n }\n\n if (!documentId) {\n return false;\n }\n\n const peerMetadata = this.repo.peerMetadataByPeerId[peerId];\n if ((peerMetadata as any)?.dxos_peerSource === 'EchoNetworkAdapter') {\n return this._echoNetworkAdapter.shouldAdvertize(peerId, { documentId });\n }\n\n const doc = this._repo.handles[documentId]?.docSync();\n if (!doc) {\n const isRequested = this._requestedDocs.has(`automerge:${documentId}`);\n log('doc share policy check', { peerId, documentId, isRequested });\n return isRequested;\n }\n\n try {\n const spaceKey = getSpaceKeyFromDoc(doc);\n if (!spaceKey) {\n log('space key not found for share policy check', { peerId, documentId });\n return false;\n }\n\n const authorizedDevices = this._authorizedDevices.get(PublicKey.from(spaceKey));\n\n // TODO(mykola): Hack, stop abusing `peerMetadata` field.\n const deviceKeyHex = (peerMetadata as any)?.dxos_deviceKey;\n if (!deviceKeyHex) {\n log('device key not found for share policy check', { peerId, documentId });\n return false;\n }\n const deviceKey = PublicKey.from(deviceKeyHex);\n\n const isAuthorized = authorizedDevices?.has(deviceKey) ?? false;\n log('share policy check', {\n localPeer: this._peerId,\n remotePeer: peerId,\n documentId,\n deviceKey,\n spaceKey,\n isAuthorized,\n });\n return isAuthorized;\n } catch (err) {\n log.catch(err);\n return false;\n }\n },\n });\n this._clientNetwork.ready();\n this._meshNetwork.ready();\n await this._echoNetworkAdapter.open();\n\n await this._clientNetwork.whenConnected();\n await this._echoNetworkAdapter.whenConnected();\n }\n\n async close() {\n await this._storage.close?.();\n await this._clientNetwork.close();\n await this._echoNetworkAdapter.close();\n await this._ctx.dispose();\n }\n\n get repo(): Repo {\n return this._repo;\n }\n\n async addReplicator(replicator: EchoReplicator) {\n await this._echoNetworkAdapter.addReplicator(replicator);\n }\n\n async removeReplicator(replicator: EchoReplicator) {\n await this._echoNetworkAdapter.removeReplicator(replicator);\n }\n\n private async _beforeSave({ path, batch }: BeforeSaveParams) {\n const handle = this._repo.handles[path[0] as DocumentId];\n if (!handle) {\n return;\n }\n const doc = handle.docSync();\n if (!doc) {\n return;\n }\n\n const lastAvailableHash = getHeads(doc);\n\n const objectIds = Object.keys(doc.objects ?? {});\n const encodedIds = objectIds.map((objectId) => idCodec.encode({ documentId: handle.documentId, objectId }));\n const idToLastHash = new Map(encodedIds.map((id) => [id, lastAvailableHash]));\n this._indexMetadataStore.markDirty(idToLastHash, batch);\n }\n\n /**\n * Called by AutomergeStorageAdapter after levelDB batch commit.\n */\n private async _afterSave() {\n this._indexMetadataStore.notifyMarkedDirty();\n }\n\n @trace.info({ depth: null })\n private _automergeDocs() {\n return mapValues(this._repo.handles, (handle) => ({\n state: handle.state,\n hasDoc: !!handle.docSync(),\n heads: handle.docSync() ? automerge.getHeads(handle.docSync()) : null,\n data:\n handle.docSync() &&\n mapValues(handle.docSync(), (value, key) => {\n try {\n switch (key) {\n case 'access':\n case 'links':\n return value;\n case 'objects':\n return Object.keys(value as any);\n default:\n return `${value}`;\n }\n } catch (err) {\n return `${err}`;\n }\n }),\n }));\n }\n\n @trace.info({ depth: null })\n private _automergePeers() {\n return this._repo.peers;\n }\n\n //\n // Methods for client-services.\n //\n @trace.span({ showInBrowserTimeline: true })\n async flush({ states }: FlushRequest): Promise<void> {\n // Note: Wait for all requested documents to be loaded/synced from thin-client.\n await Promise.all(\n states?.map(async ({ heads, documentId }) => {\n invariant(heads, 'heads are required for flush');\n const handle = this.repo.handles[documentId as DocumentId] ?? this._repo.find(documentId as DocumentId);\n await waitForHeads(handle, heads);\n }) ?? [],\n );\n\n await this._repo.flush(states?.map(({ documentId }) => documentId as DocumentId));\n }\n\n syncRepo(request: SyncRepoRequest): Stream<SyncRepoResponse> {\n return this._clientNetwork.syncRepo(request);\n }\n\n sendSyncMessage(request: SyncRepoRequest): Promise<void> {\n return this._clientNetwork.sendSyncMessage(request);\n }\n\n async getHostInfo(): Promise<HostInfo> {\n return this._clientNetwork.getHostInfo();\n }\n\n //\n // Mesh replication.\n //\n\n createExtension(): AutomergeReplicator {\n return this._meshNetwork.createExtension();\n }\n\n authorizeDevice(spaceKey: PublicKey, deviceKey: PublicKey) {\n log('authorizeDevice', { spaceKey, deviceKey });\n defaultMap(this._authorizedDevices, spaceKey, () => new ComplexSet(PublicKey.hash)).add(deviceKey);\n }\n}\n\nexport const getSpaceKeyFromDoc = (doc: any): string | null => {\n // experimental_spaceKey is set on old documents, new ones are created with doc.access.spaceKey\n const rawSpaceKey = doc.access?.spaceKey ?? doc.experimental_spaceKey;\n if (rawSpaceKey == null) {\n return null;\n }\n\n return String(rawSpaceKey);\n};\n\nconst waitForHeads = async (handle: DocHandle<SpaceDoc>, heads: Heads) => {\n await handle.whenReady();\n const unavailableHeads = new Set(heads);\n\n await Event.wrap<DocHandleChangePayload<SpaceDoc>>(handle, 'change').waitForCondition(() => {\n // Check if unavailable heads became available.\n for (const changeHash of unavailableHeads.values()) {\n if (changeIsPresentInDoc(handle.docSync(), changeHash)) {\n unavailableHeads.delete(changeHash);\n }\n }\n\n if (unavailableHeads.size === 0) {\n return true;\n }\n return false;\n });\n};\n\nconst changeIsPresentInDoc = (doc: Doc<any>, changeHash: string): boolean => {\n return !!getBackend(doc).getChangeByHash(changeHash);\n};\n", "//\n// Copyright 2024 DXOS.org\n//\n\nimport { Trigger, synchronized } from '@dxos/async';\nimport { type Message, NetworkAdapter, type PeerId, type PeerMetadata } from '@dxos/automerge/automerge-repo';\nimport { LifecycleState } from '@dxos/context';\nimport { invariant } from '@dxos/invariant';\nimport { log } from '@dxos/log';\n\nimport { type EchoReplicator, type ReplicatorConnection, type ShouldAdvertizeParams } from './echo-replicator';\n\n/**\n * Manages a set of {@link EchoReplicator} instances.\n */\nexport class EchoNetworkAdapter extends NetworkAdapter {\n private readonly _replicators = new Set<EchoReplicator>();\n /**\n * Remote peer id -> connection.\n */\n private readonly _connections = new Map<PeerId, ConnectionEntry>();\n private _lifecycleState: LifecycleState = LifecycleState.CLOSED;\n private readonly _connected = new Trigger();\n\n override connect(peerId: PeerId, peerMetadata?: PeerMetadata | undefined): void {\n this.peerId = peerId;\n this.peerMetadata = peerMetadata;\n this._connected.wake();\n }\n\n override send(message: Message): void {\n const connectionEntry = this._connections.get(message.targetId);\n if (!connectionEntry) {\n throw new Error('Connection not found.');\n }\n\n // TODO(dmaretskyi): Find a way to enforce backpressure on AM-repo.\n connectionEntry.writer.write(message).catch((err) => {\n if (connectionEntry.isOpen) {\n log.catch(err);\n }\n });\n }\n\n override disconnect(): void {\n // No-op\n }\n\n @synchronized\n async open() {\n invariant(this._lifecycleState === LifecycleState.CLOSED);\n this._lifecycleState = LifecycleState.OPEN;\n\n this.emit('ready', {\n network: this,\n });\n }\n\n @synchronized\n async close() {\n invariant(this._lifecycleState === LifecycleState.OPEN);\n\n for (const replicator of this._replicators) {\n await replicator.disconnect();\n }\n this._replicators.clear();\n\n this._lifecycleState = LifecycleState.CLOSED;\n }\n\n async whenConnected() {\n await this._connected.wait({ timeout: 10_000 });\n }\n\n @synchronized\n async addReplicator(replicator: EchoReplicator) {\n invariant(this.peerId);\n invariant(!this._replicators.has(replicator));\n\n await replicator.connect({\n peerId: this.peerId,\n onConnectionOpen: this._onConnectionOpen.bind(this),\n onConnectionClosed: this._onConnectionClosed.bind(this),\n });\n }\n\n @synchronized\n async removeReplicator(replicator: EchoReplicator) {\n invariant(this._replicators.has(replicator));\n await replicator.disconnect();\n }\n\n async shouldAdvertize(peerId: PeerId, params: ShouldAdvertizeParams): Promise<boolean> {\n const connection = this._connections.get(peerId);\n if (!connection) {\n return false;\n }\n\n return connection.connection.shouldAdvertize(params);\n }\n\n private _onConnectionOpen(connection: ReplicatorConnection) {\n invariant(!this._connections.has(connection.peerId as PeerId));\n const reader = connection.readable.getReader();\n const writer = connection.writable.getWriter();\n const connectionEntry: ConnectionEntry = { connection, reader, writer, isOpen: true };\n this._connections.set(connection.peerId as PeerId, connectionEntry);\n\n queueMicrotask(async () => {\n try {\n while (true) {\n // TODO(dmaretskyi): Find a way to enforce backpressure on AM-repo.\n const { done, value } = await reader.read();\n if (done) {\n break;\n }\n\n this.emit('message', value);\n }\n } catch (err) {\n if (connectionEntry.isOpen) {\n log.catch(err);\n }\n }\n });\n\n this.emit('peer-candidate', {\n peerId: connection.peerId as PeerId,\n peerMetadata: {\n // TODO(dmaretskyi): Refactor this.\n dxos_peerSource: 'EchoNetworkAdapter',\n } as any,\n });\n }\n\n private _onConnectionClosed(connection: ReplicatorConnection) {\n const entry = this._connections.get(connection.peerId as PeerId);\n invariant(entry);\n\n entry.isOpen = false;\n this.emit('peer-disconnected', { peerId: connection.peerId as PeerId });\n\n void entry.reader.cancel().catch((err) => log.catch(err));\n void entry.writer.abort().catch((err) => log.catch(err));\n\n this._connections.delete(connection.peerId as PeerId);\n }\n}\n\ntype ConnectionEntry = {\n connection: ReplicatorConnection;\n reader: ReadableStreamDefaultReader<Message>;\n writer: WritableStreamDefaultWriter<Message>;\n isOpen: boolean;\n};\n", "//\n// Copyright 2024 DXOS.org\n// s\n\nimport { type MixedEncoding } from 'level-transcoder';\n\nimport { type StorageAdapterInterface, type Chunk, type StorageKey } from '@dxos/automerge/automerge-repo';\nimport { LifecycleState, Resource } from '@dxos/context';\nimport { type BatchLevel, type SublevelDB } from '@dxos/kv-store';\nimport { type MaybePromise } from '@dxos/util';\n\nexport type LevelDBStorageAdapterParams = {\n db: SublevelDB;\n callbacks?: StorageCallbacks;\n};\n\nexport type BeforeSaveParams = { path: StorageKey; batch: BatchLevel };\n\nexport interface StorageCallbacks {\n beforeSave(params: BeforeSaveParams): MaybePromise<void>;\n afterSave(path: StorageKey): MaybePromise<void>;\n}\n\nexport class LevelDBStorageAdapter extends Resource implements StorageAdapterInterface {\n constructor(private readonly _params: LevelDBStorageAdapterParams) {\n super();\n }\n\n async load(keyArray: StorageKey): Promise<Uint8Array | undefined> {\n try {\n if (this._lifecycleState !== LifecycleState.OPEN) {\n // TODO(mykola): this should be an error.\n return undefined;\n }\n return await this._params.db.get<StorageKey, Uint8Array>(keyArray, { ...encodingOptions });\n } catch (err: any) {\n if (isLevelDbNotFoundError(err)) {\n return undefined;\n }\n throw err;\n }\n }\n\n async save(keyArray: StorageKey, binary: Uint8Array): Promise<void> {\n if (this._lifecycleState !== LifecycleState.OPEN) {\n return undefined;\n }\n const batch = this._params.db.batch();\n\n await this._params.callbacks?.beforeSave?.({ path: keyArray, batch });\n batch.put<StorageKey, Uint8Array>(keyArray, Buffer.from(binary), {\n ...encodingOptions,\n });\n await batch.write();\n\n await this._params.callbacks?.afterSave?.(keyArray);\n }\n\n async remove(keyArray: StorageKey): Promise<void> {\n if (this._lifecycleState !== LifecycleState.OPEN) {\n return undefined;\n }\n await this._params.db.del<StorageKey>(keyArray, { ...encodingOptions });\n }\n\n async loadRange(keyPrefix: StorageKey): Promise<Chunk[]> {\n if (this._lifecycleState !== LifecycleState.OPEN) {\n return [];\n }\n const result: Chunk[] = [];\n for await (const [key, value] of this._params.db.iterator<StorageKey, Uint8Array>({\n gte: keyPrefix,\n lte: [...keyPrefix, '\\uffff'],\n ...encodingOptions,\n })) {\n result.push({\n key,\n data: value,\n });\n }\n return result;\n }\n\n async removeRange(keyPrefix: StorageKey): Promise<void> {\n if (this._lifecycleState !== LifecycleState.OPEN) {\n return undefined;\n }\n const batch = this._params.db.batch();\n\n for await (const [key] of this._params.db.iterator<StorageKey, Uint8Array>({\n gte: keyPrefix,\n lte: [...keyPrefix, '\\uffff'],\n ...encodingOptions,\n })) {\n batch.del<StorageKey>(key, { ...encodingOptions });\n }\n await batch.write();\n }\n}\n\nconst keyEncoder: MixedEncoding<StorageKey, Uint8Array, StorageKey> = {\n encode: (key: StorageKey): Uint8Array =>\n Buffer.from(key.map((k) => k.replaceAll('%', '%25').replaceAll('-', '%2D')).join('-')),\n decode: (key: Uint8Array): StorageKey =>\n Buffer.from(key)\n .toString()\n .split('-')\n .map((k) => k.replaceAll('%2D', '-').replaceAll('%25', '%')),\n format: 'buffer',\n};\n\nexport const encodingOptions = {\n keyEncoding: keyEncoder,\n valueEncoding: 'buffer',\n};\n\nconst isLevelDbNotFoundError = (err: any): boolean => err.code === 'LEVEL_NOT_FOUND';\n", "//\n// Copyright 2024 DXOS.org\n//\n\nimport { Trigger } from '@dxos/async';\nimport { NetworkAdapter, type Message, type PeerId, cbor } from '@dxos/automerge/automerge-repo';\nimport { Stream } from '@dxos/codec-protobuf';\nimport { invariant } from '@dxos/invariant';\nimport { type HostInfo, type SyncRepoRequest, type SyncRepoResponse } from '@dxos/protocols/proto/dxos/echo/service';\n\ntype ClientSyncState = {\n connected: boolean;\n send: (message: Message) => void;\n disconnect: () => void;\n};\n\n/**\n * Used to replicate with apps running on the same device.\n */\nexport class LocalHostNetworkAdapter extends NetworkAdapter {\n private readonly _peers: Map<PeerId, ClientSyncState> = new Map();\n\n /**\n * Emits `ready` event. That signals to `Repo` that it can start using the adapter.\n */\n ready() {\n // NOTE: Emitting `ready` event in NetworkAdapter`s constructor causes a race condition\n // because `Repo` waits for `ready` event (which it never receives) before it starts using the adapter.\n this.emit('ready', {\n network: this,\n });\n }\n\n private readonly _connected = new Trigger();\n private _isConnected: boolean = false;\n\n /**\n * Called by `Repo` to connect to the network.\n *\n * @param peerId Our peer Id.\n */\n override connect(peerId: PeerId): void {\n this.peerId = peerId;\n this._isConnected = true;\n this._connected.wake();\n // No-op. Client always connects first\n }\n\n override send(message: Message): void {\n const peer = this._peers.get(message.targetId);\n invariant(peer, 'Peer not found.');\n peer.send(message);\n }\n\n async close() {\n this._peers.forEach((peer) => peer.disconnect());\n this.emit('close');\n }\n\n override disconnect(): void {\n // TODO(mykola): `disconnect` is not used anywhere in `Repo` from `@automerge/automerge-repo`. Should we remove it?\n // No-op\n }\n\n async whenConnected(): Promise<void> {\n await this._connected.wait({ timeout: 10_000 });\n }\n\n syncRepo({ id, syncMessage }: SyncRepoRequest): Stream<SyncRepoResponse> {\n const peerId = this._getPeerId(id);\n\n return new Stream(({ next, close }) => {\n invariant(!this._peers.has(peerId), 'Peer already connected.');\n this._peers.set(peerId, {\n connected: true,\n send: (message) => {\n next({\n syncMessage: cbor.encode(message),\n });\n },\n disconnect: () => {\n this._peers.delete(peerId);\n close();\n this.emit('peer-disconnected', {\n peerId,\n });\n },\n });\n\n invariant(this._isConnected);\n this.emit('peer-candidate', {\n peerMetadata: {},\n peerId,\n });\n });\n }\n\n async sendSyncMessage({ id, syncMessage }: SyncRepoRequest): Promise<void> {\n invariant(this._isConnected);\n const message = cbor.decode(syncMessage!) as Message;\n this.emit('message', message);\n }\n\n async getHostInfo(): Promise<HostInfo> {\n invariant(this._isConnected);\n invariant(this.peerId, 'Peer id not set.');\n return {\n peerId: this.peerId,\n };\n }\n\n private _getPeerId(id: string): PeerId {\n return id as PeerId;\n }\n}\n", "//\n// Copyright 2024 DXOS.org\n//\n\nimport { Trigger } from '@dxos/async';\nimport { NetworkAdapter, type Message, type PeerId, cbor } from '@dxos/automerge/automerge-repo';\nimport { invariant } from '@dxos/invariant';\nimport { log } from '@dxos/log';\nimport { type PeerInfo } from '@dxos/protocols/proto/dxos/mesh/teleport/automerge';\nimport { AutomergeReplicator } from '@dxos/teleport-extension-automerge-replicator';\n\n/**\n * Used to replicate with other peers over the network.\n */\nexport class MeshNetworkAdapter extends NetworkAdapter {\n private readonly _extensions: Map<string, AutomergeReplicator> = new Map();\n private _connected = new Trigger();\n\n /**\n * Emits `ready` event. That signals to `Repo` that it can start using the adapter.\n */\n ready() {\n // NOTE: Emitting `ready` event in NetworkAdapter`s constructor causes a race condition\n // because `Repo` waits for `ready` event (which it never receives) before it starts using the adapter.\n this.emit('ready', {\n network: this,\n });\n }\n\n override connect(peerId: PeerId): void {\n this.peerId = peerId;\n this._connected.wake();\n }\n\n override send(message: Message): void {\n const receiverId = message.targetId;\n const extension = this._extensions.get(receiverId);\n invariant(extension, 'Extension not found.');\n extension.sendSyncMessage({ payload: cbor.encode(message) }).catch((err) => log.catch(err));\n }\n\n override disconnect(): void {\n // No-op\n }\n\n createExtension(): AutomergeReplicator {\n invariant(this.peerId, 'Peer id not set.');\n\n let peerInfo: PeerInfo;\n const extension = new AutomergeReplicator(\n {\n peerId: this.peerId,\n },\n {\n onStartReplication: async (info, remotePeerId /** Teleport ID */) => {\n await this._connected.wait();\n\n // Note: We store only one extension per peer.\n // There can be a case where two connected peers have more than one teleport connection between them\n // and each of them uses different teleport connections to send messages.\n // It works because we receive messages from all teleport connections and Automerge Repo dedup them.\n // TODO(mykola): Use only one teleport connection per peer.\n\n // TODO(dmaretskyi): Critical bug.\n // - two peers get connected via swarm 1\n // - they get connected via swarm 2\n // - swarm 1 gets disconnected\n // - automerge repo thinks that peer 2 got disconnected even though swarm 2 is still active\n\n log('onStartReplication', { id: info.id, thisPeerId: this.peerId, remotePeerId: remotePeerId.toHex() });\n if (!this._extensions.has(info.id)) {\n peerInfo = info;\n // TODO(mykola): Fix race condition?\n this._extensions.set(info.id, extension);\n\n log('peer-candidate', { id: info.id, thisPeerId: this.peerId, remotePeerId: remotePeerId.toHex() });\n this.emit('peer-candidate', {\n // TODO(mykola): Hack, stop abusing `peerMetadata` field.\n peerMetadata: {\n dxos_deviceKey: remotePeerId.toHex(),\n } as any,\n peerId: info.id as PeerId,\n });\n }\n },\n onSyncMessage: async ({ payload }) => {\n if (!peerInfo) {\n return;\n }\n const message = cbor.decode(payload) as Message;\n // Note: automerge Repo dedup messages.\n this.emit('message', message);\n },\n onClose: async () => {\n if (!peerInfo) {\n return;\n }\n this.emit('peer-disconnected', {\n peerId: peerInfo.id as PeerId,\n });\n this._extensions.delete(peerInfo.id);\n },\n },\n );\n return extension;\n }\n}\n", "//\n// Copyright 2024 DXOS.org\n//\n\nimport { type StorageKey } from '@dxos/automerge/automerge-repo';\nimport { IndexedDBStorageAdapter } from '@dxos/automerge/automerge-repo-storage-indexeddb';\nimport { type SublevelDB } from '@dxos/kv-store';\nimport { log } from '@dxos/log';\nimport { StorageType, type Directory } from '@dxos/random-access-storage';\n\nimport { AutomergeStorageAdapter } from './automerge-storage-adapter';\nimport { encodingOptions } from './leveldb-storage-adapter';\n\nexport const levelMigration = async ({ db, directory }: { db: SublevelDB; directory: Directory }) => {\n // Note: Make automigration from previous storage to leveldb here.\n const isNewLevel = !(await db\n .iterator<StorageKey, Uint8Array>({\n ...encodingOptions,\n })\n .next());\n\n if (!isNewLevel) {\n return;\n }\n\n const oldStorageAdapter =\n directory.type === StorageType.IDB\n ? new IndexedDBStorageAdapter(directory.path, 'data')\n : new AutomergeStorageAdapter(directory);\n\n const chunks = await oldStorageAdapter.loadRange([]);\n if (chunks.length === 0) {\n return;\n }\n const batch = db.batch();\n log.info('found chunks on old storage adapter', { chunks: chunks.length });\n for (const { key, data } of await oldStorageAdapter.loadRange([])) {\n data && batch.put<StorageKey, Uint8Array>(key, data, { ...encodingOptions });\n }\n await batch.write();\n};\n", "//\n// Copyright 2024 DXOS.org\n//\n//\n// Copyright 2023 DXOS.org\n//\n\nimport { type Chunk, type StorageKey, type StorageAdapterInterface } from '@dxos/automerge/automerge-repo';\nimport { type Directory } from '@dxos/random-access-storage';\nimport { arrayToBuffer, bufferToArray } from '@dxos/util';\n\nexport class AutomergeStorageAdapter implements StorageAdapterInterface {\n // TODO(mykola): Hack for restricting automerge Repo to access storage if Host is `closed`.\n // Automerge Repo do not have any lifetime management.\n private _state: 'opened' | 'closed' = 'opened';\n\n constructor(private readonly _directory: Directory) {}\n\n async load(key: StorageKey): Promise<Uint8Array | undefined> {\n if (this._state !== 'opened') {\n return undefined;\n }\n const filename = this._getFilename(key);\n const file = this._directory.getOrCreateFile(filename);\n const { size } = await file.stat();\n if (!size || size === 0) {\n return undefined;\n }\n const buffer = await file.read(0, size);\n return bufferToArray(buffer);\n }\n\n async save(key: StorageKey, data: Uint8Array): Promise<void> {\n if (this._state !== 'opened') {\n return undefined;\n }\n const filename = this._getFilename(key);\n const file = this._directory.getOrCreateFile(filename);\n await file.write(0, arrayToBuffer(data));\n await file.truncate?.(data.length);\n\n await file.flush?.();\n }\n\n async remove(key: StorageKey): Promise<void> {\n if (this._state !== 'opened') {\n return undefined;\n }\n // TODO(dmaretskyi): Better deletion.\n const filename = this._getFilename(key);\n const file = this._directory.getOrCreateFile(filename);\n await file.destroy();\n }\n\n async loadRange(keyPrefix: StorageKey): Promise<Chunk[]> {\n if (this._state !== 'opened') {\n return [];\n }\n const filename = this._getFilename(keyPrefix);\n const entries = await this._directory.list();\n return Promise.all(\n entries\n .filter((entry) => entry.startsWith(filename))\n .map(async (entry): Promise<Chunk> => {\n const file = this._directory.getOrCreateFile(entry);\n const { size } = await file.stat();\n const buffer = await file.read(0, size);\n return {\n key: this._getKeyFromFilename(entry),\n data: bufferToArray(buffer),\n };\n }),\n );\n }\n\n async removeRange(keyPrefix: StorageKey): Promise<void> {\n if (this._state !== 'opened') {\n return undefined;\n }\n const filename = this._getFilename(keyPrefix);\n const entries = await this._directory.list();\n await Promise.all(\n entries\n .filter((entry) => entry.startsWith(filename))\n .map(async (entry): Promise<void> => {\n const file = this._directory.getOrCreateFile(entry);\n await file.destroy();\n }),\n );\n }\n\n async close(): Promise<void> {\n this._state = 'closed';\n }\n\n private _getFilename(key: StorageKey): string {\n return key.map((k) => k.replaceAll('%', '%25').replaceAll('-', '%2D')).join('-');\n }\n\n private _getKeyFromFilename(filename: string): StorageKey {\n return filename.split('-').map((k) => k.replaceAll('%2D', '-').replaceAll('%25', '%'));\n }\n}\n", "//\n// Copyright 2024 DXOS.org\n//\n\nimport { Event } from '@dxos/async';\nimport { type DocHandle, type AutomergeUrl, type DocumentId, type Repo } from '@dxos/automerge/automerge-repo';\nimport { cancelWithContext, type Context } from '@dxos/context';\nimport { warnAfterTimeout } from '@dxos/debug';\nimport { type SpaceState, type SpaceDoc } from '@dxos/echo-protocol';\nimport { invariant } from '@dxos/invariant';\nimport { type PublicKey } from '@dxos/keys';\nimport { log } from '@dxos/log';\nimport { trace } from '@dxos/tracing';\n\ntype SpaceDocumentLinks = SpaceDoc['links'];\n\nexport interface AutomergeDocumentLoader {\n onObjectDocumentLoaded: Event<ObjectDocumentLoaded>;\n\n getAllHandles(): DocHandle<SpaceDoc>[];\n\n loadSpaceRootDocHandle(ctx: Context, spaceState: SpaceState): Promise<void>;\n loadObjectDocument(objectId: string | string[]): void;\n getSpaceRootDocHandle(): DocHandle<SpaceDoc>;\n createDocumentForObject(objectId: string): DocHandle<SpaceDoc>;\n onObjectLinksUpdated(links: SpaceDocumentLinks): void;\n onObjectBoundToDocument(handle: DocHandle<SpaceDoc>, objectId: string): void;\n\n /**\n * @returns objectIds for which we had document handles or were loading one.\n */\n clearHandleReferences(): string[];\n}\n\n/**\n * Manages object <-> docHandle binding and automerge document loading.\n */\n@trace.resource()\nexport class AutomergeDocumentLoaderImpl implements AutomergeDocumentLoader {\n private _spaceRootDocHandle: DocHandle<SpaceDoc> | null = null;\n /**\n * An object id pointer to a handle of the document where the object is stored inline.\n */\n private readonly _objectDocumentHandles = new Map<string, DocHandle<SpaceDoc>>();\n /**\n * If object was requested via loadObjectDocument but root document links weren't updated yet\n * loading will be triggered in onObjectLinksUpdated callback.\n */\n private readonly _objectsPendingDocumentLoad = new Set<string>();\n\n public readonly onObjectDocumentLoaded = new Event<ObjectDocumentLoaded>();\n\n constructor(\n private readonly _spaceKey: PublicKey,\n private readonly _repo: Repo,\n ) {}\n\n getAllHandles(): DocHandle<SpaceDoc>[] {\n return this._spaceRootDocHandle != null\n ? [this._spaceRootDocHandle, ...new Set(this._objectDocumentHandles.values())]\n : [];\n }\n\n @trace.span({ showInBrowserTimeline: true })\n public async loadSpaceRootDocHandle(ctx: Context, spaceState: SpaceState): Promise<void> {\n if (this._spaceRootDocHandle != null) {\n return;\n }\n if (!spaceState.rootUrl) {\n log.error('Database opened with no rootUrl', { spaceKey: this._spaceKey });\n this._createContextBoundSpaceRootDocument(ctx);\n } else {\n const existingDocHandle = await this._initDocHandle(ctx, spaceState.rootUrl);\n const doc = existingDocHandle.docSync();\n invariant(doc);\n if (doc.access == null) {\n this._initDocAccess(existingDocHandle);\n }\n this._spaceRootDocHandle = existingDocHandle;\n }\n }\n\n public loadObjectDocument(objectIdOrMany: string | string[]) {\n const objectIds = Array.isArray(objectIdOrMany) ? objectIdOrMany : [objectIdOrMany];\n let hasUrlsToLoad = false;\n const urlsToLoad: SpaceDoc['links'] = {};\n for (const objectId of objectIds) {\n invariant(this._spaceRootDocHandle);\n if (this._objectDocumentHandles.has(objectId) || this._objectsPendingDocumentLoad.has(objectId)) {\n continue;\n }\n const spaceRootDoc = this._spaceRootDocHandle.docSync();\n invariant(spaceRootDoc);\n const documentUrl = (spaceRootDoc.links ?? {})[objectId];\n if (documentUrl == null) {\n this._objectsPendingDocumentLoad.add(objectId);\n log.info('loading delayed until object links are initialized', { objectId });\n } else {\n urlsToLoad[objectId] = documentUrl;\n hasUrlsToLoad = true;\n }\n }\n if (hasUrlsToLoad) {\n this._loadLinkedObjects(urlsToLoad);\n }\n }\n\n public onObjectLinksUpdated(links: SpaceDocumentLinks) {\n if (!links) {\n return;\n }\n const linksAwaitingLoad = Object.entries(links).filter(([objectId]) =>\n this._objectsPendingDocumentLoad.has(objectId),\n );\n this._loadLinkedObjects(Object.fromEntries(linksAwaitingLoad));\n linksAwaitingLoad.forEach(([objectId]) => this._objectsPendingDocumentLoad.delete(objectId));\n }\n\n public getSpaceRootDocHandle(): DocHandle<SpaceDoc> {\n invariant(this._spaceRootDocHandle);\n return this._spaceRootDocHandle;\n }\n\n public createDocumentForObject(objectId: string): DocHandle<SpaceDoc> {\n invariant(this._spaceRootDocHandle);\n const spaceDocHandle = this._repo.create<SpaceDoc>();\n this._initDocAccess(spaceDocHandle);\n this.onObjectBoundToDocument(spaceDocHandle, objectId);\n this._spaceRootDocHandle.change((newDoc: SpaceDoc) => {\n newDoc.links ??= {};\n newDoc.links[objectId] = spaceDocHandle.url;\n });\n return spaceDocHandle;\n }\n\n public onObjectBoundToDocument(handle: DocHandle<SpaceDoc>, objectId: string) {\n this._objectDocumentHandles.set(objectId, handle);\n }\n\n public clearHandleReferences(): string[] {\n const objectsWithHandles = [...this._objectDocumentHandles.keys()];\n this._objectDocumentHandles.clear();\n this._spaceRootDocHandle = null;\n return objectsWithHandles;\n }\n\n private _loadLinkedObjects(links: SpaceDocumentLinks) {\n if (!links) {\n return;\n }\n for (const [objectId, automergeUrl] of Object.entries(links)) {\n const logMeta = { objectId, automergeUrl };\n const objectDocumentHandle = this._objectDocumentHandles.get(objectId);\n if (objectDocumentHandle != null && objectDocumentHandle.url !== automergeUrl) {\n log.warn('object already inlined in a different document, ignoring the link', {\n ...logMeta,\n actualDocumentUrl: objectDocumentHandle.url,\n });\n continue;\n }\n if (objectDocumentHandle?.url === automergeUrl) {\n log.warn('object document was already loaded', logMeta);\n continue;\n }\n const handle = this._repo.find<SpaceDoc>(automergeUrl as DocumentId);\n log.debug('document loading triggered', logMeta);\n this._objectDocumentHandles.set(objectId, handle);\n void this._createObjectOnDocumentLoad(handle, objectId);\n }\n }\n\n private async _initDocHandle(ctx: Context, url: string) {\n const docHandle = this._repo.find<SpaceDoc>(url as DocumentId);\n while (true) {\n try {\n await warnAfterTimeout(5_000, 'Automerge root doc load timeout (AutomergeDb)', async () => {\n await cancelWithContext(ctx, docHandle.whenReady()); // TODO(dmaretskyi): Temporary 5s timeout for debugging.\n });\n break;\n } catch (err) {\n if (`${err}`.includes('Timeout')) {\n log.info('wraparound', { id: docHandle.documentId, state: docHandle.state });\n continue;\n }\n\n throw err;\n }\n }\n\n if (docHandle.state === 'unavailable') {\n throw new Error('Automerge document is unavailable');\n }\n\n return docHandle;\n }\n\n private _createContextBoundSpaceRootDocument(ctx: Context) {\n const docHandle = this._repo.create<SpaceDoc>();\n this._spaceRootDocHandle = docHandle;\n ctx.onDispose(() => {\n docHandle.delete();\n this._spaceRootDocHandle = null;\n });\n }\n\n private _initDocAccess(handle: DocHandle<SpaceDoc>) {\n handle.change((newDoc: SpaceDoc) => {\n newDoc.access ??= { spaceKey: this._spaceKey.toHex() };\n newDoc.access.spaceKey = this._spaceKey.toHex();\n });\n }\n\n private async _createObjectOnDocumentLoad(handle: DocHandle<SpaceDoc>, objectId: string) {\n try {\n await handle.doc(['ready']);\n const logMeta = { objectId, docUrl: handle.url };\n if (this.onObjectDocumentLoaded.listenerCount() === 0) {\n log.info('document loaded after all listeners were removed', logMeta);\n return;\n }\n const objectDocHandle = this._objectDocumentHandles.get(objectId);\n if (objectDocHandle?.url !== handle.url) {\n log.warn('object was rebound while a document was loading, discarding handle', logMeta);\n return;\n }\n this.onObjectDocumentLoaded.emit({ handle, objectId });\n } catch (err) {\n const shouldRetryLoading = this.onObjectDocumentLoaded.listenerCount() > 0;\n log.warn('failed to load a document', {\n objectId,\n automergeUrl: handle.url,\n retryLoading: shouldRetryLoading,\n err,\n });\n if (shouldRetryLoading) {\n await this._createObjectOnDocumentLoad(handle, objectId);\n }\n }\n }\n}\n\nexport interface ObjectDocumentLoaded {\n handle: DocHandle<SpaceDoc>;\n objectId: string;\n}\n\nexport interface DocumentChanges {\n createdObjectIds: string[];\n updatedObjectIds: string[];\n objectsToRebind: string[];\n linkedDocuments: {\n [echoId: string]: AutomergeUrl;\n };\n}\n"],
|
|
5
5
|
"mappings": ";;;;;;;;;;;;;;;;;;;;;;;;;;;AAIA,SAASA,aAAa;AACtB,SAAmBC,QAAQC,WAAWC,YAAwBC,gBAAgB;AAC9E,SAEEC,YAKK;AAEP,SAASC,eAA+B;AAGxC,SAASC,aAAAA,kBAAiB;AAC1B,SAASC,iBAAiB;AAE1B,SAASC,OAAAA,YAAW;AACpB,SAASC,eAAe;AASxB,SAASC,aAAa;AACtB,SAASC,YAAYC,YAAYC,YAAYC,iBAAiB;;;AC5B9D,SAASC,SAASC,oBAAoB;AACtC,SAAuBC,sBAAsD;AAC7E,SAASC,sBAAsB;AAC/B,SAASC,iBAAiB;AAC1B,SAASC,WAAW;;;;;;;;;;;;AAOb,IAAMC,qBAAN,cAAiCJ,eAAAA;EAAjC;;AACYK,wBAAe,oBAAIC,IAAAA;AAInBC;;;wBAAe,oBAAIC,IAAAA;AAC5BC,2BAAkCR,eAAeS;AACxCC,sBAAa,IAAIb,QAAAA;;EAEzBc,QAAQC,QAAgBC,cAA+C;AAC9E,SAAKD,SAASA;AACd,SAAKC,eAAeA;AACpB,SAAKH,WAAWI,KAAI;EACtB;EAESC,KAAKC,SAAwB;AACpC,UAAMC,kBAAkB,KAAKX,aAAaY,IAAIF,QAAQG,QAAQ;AAC9D,QAAI,CAACF,iBAAiB;AACpB,YAAM,IAAIG,MAAM,uBAAA;IAClB;AAGAH,oBAAgBI,OAAOC,MAAMN,OAAAA,EAASO,MAAM,CAACC,QAAAA;AAC3C,UAAIP,gBAAgBQ,QAAQ;AAC1BvB,YAAIqB,MAAMC,KAAAA,QAAAA;;;;;;MACZ;IACF,CAAA;EACF;EAESE,aAAmB;EAE5B;EAEA,MACMC,OAAO;AACX1B,cAAU,KAAKO,oBAAoBR,eAAeS,QAAM,QAAA;;;;;;;;;AACxD,SAAKD,kBAAkBR,eAAe4B;AAEtC,SAAKC,KAAK,SAAS;MACjBC,SAAS;IACX,CAAA;EACF;EAEA,MACMC,QAAQ;AACZ9B,cAAU,KAAKO,oBAAoBR,eAAe4B,MAAI,QAAA;;;;;;;;;AAEtD,eAAWI,cAAc,KAAK5B,cAAc;AAC1C,YAAM4B,WAAWN,WAAU;IAC7B;AACA,SAAKtB,aAAa6B,MAAK;AAEvB,SAAKzB,kBAAkBR,eAAeS;EACxC;EAEA,MAAMyB,gBAAgB;AACpB,UAAM,KAAKxB,WAAWyB,KAAK;MAAEC,SAAS;IAAO,CAAA;EAC/C;EAEA,MACMC,cAAcL,YAA4B;AAC9C/B,cAAU,KAAKW,QAAM,QAAA;;;;;;;;;AACrBX,cAAU,CAAC,KAAKG,aAAakC,IAAIN,UAAAA,GAAAA,QAAAA;;;;;;;;;AAEjC,UAAMA,WAAWrB,QAAQ;MACvBC,QAAQ,KAAKA;MACb2B,kBAAkB,KAAKC,kBAAkBC,KAAK,IAAI;MAClDC,oBAAoB,KAAKC,oBAAoBF,KAAK,IAAI;IACxD,CAAA;EACF;EAEA,MACMG,iBAAiBZ,YAA4B;AACjD/B,cAAU,KAAKG,aAAakC,IAAIN,UAAAA,GAAAA,QAAAA;;;;;;;;;AAChC,UAAMA,WAAWN,WAAU;EAC7B;EAEA,MAAMmB,gBAAgBjC,QAAgBkC,QAAiD;AACrF,UAAMC,aAAa,KAAKzC,aAAaY,IAAIN,MAAAA;AACzC,QAAI,CAACmC,YAAY;AACf,aAAO;IACT;AAEA,WAAOA,WAAWA,WAAWF,gBAAgBC,MAAAA;EAC/C;EAEQN,kBAAkBO,YAAkC;AAC1D9C,cAAU,CAAC,KAAKK,aAAagC,IAAIS,WAAWnC,MAAM,GAAA,QAAA;;;;;;;;;AAClD,UAAMoC,SAASD,WAAWE,SAASC,UAAS;AAC5C,UAAM7B,SAAS0B,WAAWI,SAASC,UAAS;AAC5C,UAAMnC,kBAAmC;MAAE8B;MAAYC;MAAQ3B;MAAQI,QAAQ;IAAK;AACpF,SAAKnB,aAAa+C,IAAIN,WAAWnC,QAAkBK,eAAAA;AAEnDqC,mBAAe,YAAA;AACb,UAAI;AACF,eAAO,MAAM;AAEX,gBAAM,EAAEC,MAAMC,MAAK,IAAK,MAAMR,OAAOS,KAAI;AACzC,cAAIF,MAAM;AACR;UACF;AAEA,eAAK1B,KAAK,WAAW2B,KAAAA;QACvB;MACF,SAAShC,KAAK;AACZ,YAAIP,gBAAgBQ,QAAQ;AAC1BvB,cAAIqB,MAAMC,KAAAA,QAAAA;;;;;;QACZ;MACF;IACF,CAAA;AAEA,SAAKK,KAAK,kBAAkB;MAC1BjB,QAAQmC,WAAWnC;MACnBC,cAAc;;QAEZ6C,iBAAiB;MACnB;IACF,CAAA;EACF;EAEQf,oBAAoBI,YAAkC;AAC5D,UAAMY,QAAQ,KAAKrD,aAAaY,IAAI6B,WAAWnC,MAAM;AACrDX,cAAU0D,OAAAA,QAAAA;;;;;;;;;AAEVA,UAAMlC,SAAS;AACf,SAAKI,KAAK,qBAAqB;MAAEjB,QAAQmC,WAAWnC;IAAiB,CAAA;AAErE,SAAK+C,MAAMX,OAAOY,OAAM,EAAGrC,MAAM,CAACC,QAAQtB,IAAIqB,MAAMC,KAAAA,QAAAA;;;;;;AACpD,SAAKmC,MAAMtC,OAAOwC,MAAK,EAAGtC,MAAM,CAACC,QAAQtB,IAAIqB,MAAMC,KAAAA,QAAAA;;;;;;AAEnD,SAAKlB,aAAawD,OAAOf,WAAWnC,MAAM;EAC5C;AACF;;EAnGGd;GAjCUK,mBAAAA,WAAAA,QAAAA,IAAAA;;EA2CVL;GA3CUK,mBAAAA,WAAAA,SAAAA,IAAAA;;EA2DVL;GA3DUK,mBAAAA,WAAAA,iBAAAA,IAAAA;;EAuEVL;GAvEUK,mBAAAA,WAAAA,oBAAAA,IAAAA;;;ACRb,SAAS4D,kBAAAA,iBAAgBC,gBAAgB;AAgBlC,IAAMC,wBAAN,cAAoCC,SAAAA;EACzCC,YAA6BC,SAAsC;AACjE,UAAK;SADsBA,UAAAA;EAE7B;EAEA,MAAMC,KAAKC,UAAuD;AAChE,QAAI;AACF,UAAI,KAAKC,oBAAoBC,gBAAeC,MAAM;AAEhD,eAAOC;MACT;AACA,aAAO,MAAM,KAAKN,QAAQO,GAAGC,IAA4BN,UAAU;QAAE,GAAGO;MAAgB,CAAA;IAC1F,SAASC,KAAU;AACjB,UAAIC,uBAAuBD,GAAAA,GAAM;AAC/B,eAAOJ;MACT;AACA,YAAMI;IACR;EACF;EAEA,MAAME,KAAKV,UAAsBW,QAAmC;AAClE,QAAI,KAAKV,oBAAoBC,gBAAeC,MAAM;AAChD,aAAOC;IACT;AACA,UAAMQ,QAAQ,KAAKd,QAAQO,GAAGO,MAAK;AAEnC,UAAM,KAAKd,QAAQe,WAAWC,aAAa;MAAEC,MAAMf;MAAUY;IAAM,CAAA;AACnEA,UAAMI,IAA4BhB,UAAUiB,OAAOC,KAAKP,MAAAA,GAAS;MAC/D,GAAGJ;IACL,CAAA;AACA,UAAMK,MAAMO,MAAK;AAEjB,UAAM,KAAKrB,QAAQe,WAAWO,YAAYpB,QAAAA;EAC5C;EAEA,MAAMqB,OAAOrB,UAAqC;AAChD,QAAI,KAAKC,oBAAoBC,gBAAeC,MAAM;AAChD,aAAOC;IACT;AACA,UAAM,KAAKN,QAAQO,GAAGiB,IAAgBtB,UAAU;MAAE,GAAGO;IAAgB,CAAA;EACvE;EAEA,MAAMgB,UAAUC,WAAyC;AACvD,QAAI,KAAKvB,oBAAoBC,gBAAeC,MAAM;AAChD,aAAO,CAAA;IACT;AACA,UAAMsB,SAAkB,CAAA;AACxB,qBAAiB,CAACC,KAAKC,KAAAA,KAAU,KAAK7B,QAAQO,GAAGuB,SAAiC;MAChFC,KAAKL;MACLM,KAAK;WAAIN;QAAW;;MACpB,GAAGjB;IACL,CAAA,GAAI;AACFkB,aAAOM,KAAK;QACVL;QACAM,MAAML;MACR,CAAA;IACF;AACA,WAAOF;EACT;EAEA,MAAMQ,YAAYT,WAAsC;AACtD,QAAI,KAAKvB,oBAAoBC,gBAAeC,MAAM;AAChD,aAAOC;IACT;AACA,UAAMQ,QAAQ,KAAKd,QAAQO,GAAGO,MAAK;AAEnC,qBAAiB,CAACc,GAAAA,KAAQ,KAAK5B,QAAQO,GAAGuB,SAAiC;MACzEC,KAAKL;MACLM,KAAK;WAAIN;QAAW;;MACpB,GAAGjB;IACL,CAAA,GAAI;AACFK,YAAMU,IAAgBI,KAAK;QAAE,GAAGnB;MAAgB,CAAA;IAClD;AACA,UAAMK,MAAMO,MAAK;EACnB;AACF;AAEA,IAAMe,aAAgE;EACpEC,QAAQ,CAACT,QACPT,OAAOC,KAAKQ,IAAIU,IAAI,CAACC,MAAMA,EAAEC,WAAW,KAAK,KAAA,EAAOA,WAAW,KAAK,KAAA,CAAA,EAAQC,KAAK,GAAA,CAAA;EACnFC,QAAQ,CAACd,QACPT,OAAOC,KAAKQ,GAAAA,EACTe,SAAQ,EACRC,MAAM,GAAA,EACNN,IAAI,CAACC,MAAMA,EAAEC,WAAW,OAAO,GAAA,EAAKA,WAAW,OAAO,GAAA,CAAA;EAC3DK,QAAQ;AACV;AAEO,IAAMpC,kBAAkB;EAC7BqC,aAAaV;EACbW,eAAe;AACjB;AAEA,IAAMpC,yBAAyB,CAACD,QAAsBA,IAAIsC,SAAS;;;AChHnE,SAASC,WAAAA,gBAAe;AACxB,SAASC,kBAAAA,iBAA2CC,YAAY;AAChE,SAASC,cAAc;AACvB,SAASC,aAAAA,kBAAiB;;AAYnB,IAAMC,0BAAN,cAAsCJ,gBAAAA;EAAtC;;AACYK,kBAAuC,oBAAIC,IAAAA;AAa3CC,sBAAa,IAAIR,SAAAA;AAC1BS,wBAAwB;;;;;EAThCC,QAAQ;AAGN,SAAKC,KAAK,SAAS;MACjBC,SAAS;IACX,CAAA;EACF;;;;;;EAUSC,QAAQC,QAAsB;AACrC,SAAKA,SAASA;AACd,SAAKL,eAAe;AACpB,SAAKD,WAAWO,KAAI;EAEtB;EAESC,KAAKC,SAAwB;AACpC,UAAMC,OAAO,KAAKZ,OAAOa,IAAIF,QAAQG,QAAQ;AAC7ChB,IAAAA,WAAUc,MAAM,mBAAA;;;;;;;;;AAChBA,SAAKF,KAAKC,OAAAA;EACZ;EAEA,MAAMI,QAAQ;AACZ,SAAKf,OAAOgB,QAAQ,CAACJ,SAASA,KAAKK,WAAU,CAAA;AAC7C,SAAKZ,KAAK,OAAA;EACZ;EAESY,aAAmB;EAG5B;EAEA,MAAMC,gBAA+B;AACnC,UAAM,KAAKhB,WAAWiB,KAAK;MAAEC,SAAS;IAAO,CAAA;EAC/C;EAEAC,SAAS,EAAEC,IAAIC,YAAW,GAA+C;AACvE,UAAMf,SAAS,KAAKgB,WAAWF,EAAAA;AAE/B,WAAO,IAAIzB,OAAO,CAAC,EAAE4B,MAAMV,MAAK,MAAE;AAChCjB,MAAAA,WAAU,CAAC,KAAKE,OAAO0B,IAAIlB,MAAAA,GAAS,2BAAA;;;;;;;;;AACpC,WAAKR,OAAO2B,IAAInB,QAAQ;QACtBoB,WAAW;QACXlB,MAAM,CAACC,YAAAA;AACLc,eAAK;YACHF,aAAa3B,KAAKiC,OAAOlB,OAAAA;UAC3B,CAAA;QACF;QACAM,YAAY,MAAA;AACV,eAAKjB,OAAO8B,OAAOtB,MAAAA;AACnBO,gBAAAA;AACA,eAAKV,KAAK,qBAAqB;YAC7BG;UACF,CAAA;QACF;MACF,CAAA;AAEAV,MAAAA,WAAU,KAAKK,cAAY,QAAA;;;;;;;;;AAC3B,WAAKE,KAAK,kBAAkB;QAC1B0B,cAAc,CAAC;QACfvB;MACF,CAAA;IACF,CAAA;EACF;EAEA,MAAMwB,gBAAgB,EAAEV,IAAIC,YAAW,GAAoC;AACzEzB,IAAAA,WAAU,KAAKK,cAAY,QAAA;;;;;;;;;AAC3B,UAAMQ,UAAUf,KAAKqC,OAAOV,WAAAA;AAC5B,SAAKlB,KAAK,WAAWM,OAAAA;EACvB;EAEA,MAAMuB,cAAiC;AACrCpC,IAAAA,WAAU,KAAKK,cAAY,QAAA;;;;;;;;;AAC3BL,IAAAA,WAAU,KAAKU,QAAQ,oBAAA;;;;;;;;;AACvB,WAAO;MACLA,QAAQ,KAAKA;IACf;EACF;EAEQgB,WAAWF,IAAoB;AACrC,WAAOA;EACT;AACF;;;AC9GA,SAASa,WAAAA,gBAAe;AACxB,SAASC,kBAAAA,iBAA2CC,QAAAA,aAAY;AAChE,SAASC,aAAAA,kBAAiB;AAC1B,SAASC,OAAAA,YAAW;AAEpB,SAASC,2BAA2B;;AAK7B,IAAMC,qBAAN,cAAiCL,gBAAAA;EAAjC;;AACYM,uBAAgD,oBAAIC,IAAAA;AAC7DC,sBAAa,IAAIT,SAAAA;;;;;EAKzBU,QAAQ;AAGN,SAAKC,KAAK,SAAS;MACjBC,SAAS;IACX,CAAA;EACF;EAESC,QAAQC,QAAsB;AACrC,SAAKA,SAASA;AACd,SAAKL,WAAWM,KAAI;EACtB;EAESC,KAAKC,SAAwB;AACpC,UAAMC,aAAaD,QAAQE;AAC3B,UAAMC,YAAY,KAAKb,YAAYc,IAAIH,UAAAA;AACvCf,IAAAA,WAAUiB,WAAW,wBAAA;;;;;;;;;AACrBA,cAAUE,gBAAgB;MAAEC,SAASrB,MAAKsB,OAAOP,OAAAA;IAAS,CAAA,EAAGQ,MAAM,CAACC,QAAQtB,KAAIqB,MAAMC,KAAAA,QAAAA;;;;;;EACxF;EAESC,aAAmB;EAE5B;EAEAC,kBAAuC;AACrCzB,IAAAA,WAAU,KAAKW,QAAQ,oBAAA;;;;;;;;;AAEvB,QAAIe;AACJ,UAAMT,YAAY,IAAIf,oBACpB;MACES,QAAQ,KAAKA;IACf,GACA;MACEgB,oBAAoB,OAAOC,MAAMC,iBAA6B;AAC5D,cAAM,KAAKvB,WAAWwB,KAAI;AAc1B7B,QAAAA,KAAI,sBAAsB;UAAE8B,IAAIH,KAAKG;UAAIC,YAAY,KAAKrB;UAAQkB,cAAcA,aAAaI,MAAK;QAAG,GAAA;;;;;;AACrG,YAAI,CAAC,KAAK7B,YAAY8B,IAAIN,KAAKG,EAAE,GAAG;AAClCL,qBAAWE;AAEX,eAAKxB,YAAY+B,IAAIP,KAAKG,IAAId,SAAAA;AAE9BhB,UAAAA,KAAI,kBAAkB;YAAE8B,IAAIH,KAAKG;YAAIC,YAAY,KAAKrB;YAAQkB,cAAcA,aAAaI,MAAK;UAAG,GAAA;;;;;;AACjG,eAAKzB,KAAK,kBAAkB;;YAE1B4B,cAAc;cACZC,gBAAgBR,aAAaI,MAAK;YACpC;YACAtB,QAAQiB,KAAKG;UACf,CAAA;QACF;MACF;MACAO,eAAe,OAAO,EAAElB,QAAO,MAAE;AAC/B,YAAI,CAACM,UAAU;AACb;QACF;AACA,cAAMZ,UAAUf,MAAKwC,OAAOnB,OAAAA;AAE5B,aAAKZ,KAAK,WAAWM,OAAAA;MACvB;MACA0B,SAAS,YAAA;AACP,YAAI,CAACd,UAAU;AACb;QACF;AACA,aAAKlB,KAAK,qBAAqB;UAC7BG,QAAQe,SAASK;QACnB,CAAA;AACA,aAAK3B,YAAYqC,OAAOf,SAASK,EAAE;MACrC;IACF,CAAA;AAEF,WAAOd;EACT;AACF;;;ACrGA,SAASyB,+BAA+B;AAExC,SAASC,OAAAA,YAAW;AACpB,SAASC,mBAAmC;;;ACC5C,SAASC,eAAeC,qBAAqB;AAEtC,IAAMC,0BAAN,MAAMA;EAKXC,YAA6BC,YAAuB;SAAvBA,aAAAA;SAFrBC,SAA8B;EAEe;EAErD,MAAMC,KAAKC,KAAkD;AAC3D,QAAI,KAAKF,WAAW,UAAU;AAC5B,aAAOG;IACT;AACA,UAAMC,WAAW,KAAKC,aAAaH,GAAAA;AACnC,UAAMI,OAAO,KAAKP,WAAWQ,gBAAgBH,QAAAA;AAC7C,UAAM,EAAEI,KAAI,IAAK,MAAMF,KAAKG,KAAI;AAChC,QAAI,CAACD,QAAQA,SAAS,GAAG;AACvB,aAAOL;IACT;AACA,UAAMO,SAAS,MAAMJ,KAAKK,KAAK,GAAGH,IAAAA;AAClC,WAAOI,cAAcF,MAAAA;EACvB;EAEA,MAAMG,KAAKX,KAAiBY,MAAiC;AAC3D,QAAI,KAAKd,WAAW,UAAU;AAC5B,aAAOG;IACT;AACA,UAAMC,WAAW,KAAKC,aAAaH,GAAAA;AACnC,UAAMI,OAAO,KAAKP,WAAWQ,gBAAgBH,QAAAA;AAC7C,UAAME,KAAKS,MAAM,GAAGC,cAAcF,IAAAA,CAAAA;AAClC,UAAMR,KAAKW,WAAWH,KAAKI,MAAM;AAEjC,UAAMZ,KAAKa,QAAK;EAClB;EAEA,MAAMC,OAAOlB,KAAgC;AAC3C,QAAI,KAAKF,WAAW,UAAU;AAC5B,aAAOG;IACT;AAEA,UAAMC,WAAW,KAAKC,aAAaH,GAAAA;AACnC,UAAMI,OAAO,KAAKP,WAAWQ,gBAAgBH,QAAAA;AAC7C,UAAME,KAAKe,QAAO;EACpB;EAEA,MAAMC,UAAUC,WAAyC;AACvD,QAAI,KAAKvB,WAAW,UAAU;AAC5B,aAAO,CAAA;IACT;AACA,UAAMI,WAAW,KAAKC,aAAakB,SAAAA;AACnC,UAAMC,UAAU,MAAM,KAAKzB,WAAW0B,KAAI;AAC1C,WAAOC,QAAQC,IACbH,QACGI,OAAO,CAACC,UAAUA,MAAMC,WAAW1B,QAAAA,CAAAA,EACnC2B,IAAI,OAAOF,UAAAA;AACV,YAAMvB,OAAO,KAAKP,WAAWQ,gBAAgBsB,KAAAA;AAC7C,YAAM,EAAErB,KAAI,IAAK,MAAMF,KAAKG,KAAI;AAChC,YAAMC,SAAS,MAAMJ,KAAKK,KAAK,GAAGH,IAAAA;AAClC,aAAO;QACLN,KAAK,KAAK8B,oBAAoBH,KAAAA;QAC9Bf,MAAMF,cAAcF,MAAAA;MACtB;IACF,CAAA,CAAA;EAEN;EAEA,MAAMuB,YAAYV,WAAsC;AACtD,QAAI,KAAKvB,WAAW,UAAU;AAC5B,aAAOG;IACT;AACA,UAAMC,WAAW,KAAKC,aAAakB,SAAAA;AACnC,UAAMC,UAAU,MAAM,KAAKzB,WAAW0B,KAAI;AAC1C,UAAMC,QAAQC,IACZH,QACGI,OAAO,CAACC,UAAUA,MAAMC,WAAW1B,QAAAA,CAAAA,EACnC2B,IAAI,OAAOF,UAAAA;AACV,YAAMvB,OAAO,KAAKP,WAAWQ,gBAAgBsB,KAAAA;AAC7C,YAAMvB,KAAKe,QAAO;IACpB,CAAA,CAAA;EAEN;EAEA,MAAMa,QAAuB;AAC3B,SAAKlC,SAAS;EAChB;EAEQK,aAAaH,KAAyB;AAC5C,WAAOA,IAAI6B,IAAI,CAACI,MAAMA,EAAEC,WAAW,KAAK,KAAA,EAAOA,WAAW,KAAK,KAAA,CAAA,EAAQC,KAAK,GAAA;EAC9E;EAEQL,oBAAoB5B,UAA8B;AACxD,WAAOA,SAASkC,MAAM,GAAA,EAAKP,IAAI,CAACI,MAAMA,EAAEC,WAAW,OAAO,GAAA,EAAKA,WAAW,OAAO,GAAA,CAAA;EACnF;AACF;;;;ADzFO,IAAMG,iBAAiB,OAAO,EAAEC,IAAIC,UAAS,MAA4C;AAE9F,QAAMC,aAAa,CAAE,MAAMF,GACxBG,SAAiC;IAChC,GAAGC;EACL,CAAA,EACCC,KAAI;AAEP,MAAI,CAACH,YAAY;AACf;EACF;AAEA,QAAMI,oBACJL,UAAUM,SAASC,YAAYC,MAC3B,IAAIC,wBAAwBT,UAAUU,MAAM,MAAA,IAC5C,IAAIC,wBAAwBX,SAAAA;AAElC,QAAMY,SAAS,MAAMP,kBAAkBQ,UAAU,CAAA,CAAE;AACnD,MAAID,OAAOE,WAAW,GAAG;AACvB;EACF;AACA,QAAMC,QAAQhB,GAAGgB,MAAK;AACtBC,EAAAA,KAAIC,KAAK,uCAAuC;IAAEL,QAAQA,OAAOE;EAAO,GAAA;;;;;;AACxE,aAAW,EAAEI,KAAKC,KAAI,KAAM,MAAMd,kBAAkBQ,UAAU,CAAA,CAAE,GAAG;AACjEM,YAAQJ,MAAMK,IAA4BF,KAAKC,MAAM;MAAE,GAAGhB;IAAgB,CAAA;EAC5E;AACA,QAAMY,MAAMM,MAAK;AACnB;;;;;;;;;;;;;;ALeO,IAAMC,gBAAN,MAAMA;EAsBXC,YAAY,EAAEC,WAAWC,IAAIC,mBAAkB,GAAyB;AApBvDC,gBAAO,IAAIC,QAAAA;AAGXC,+BAAsB,IAAIC,mBAAAA;AAa1BC;;;8BAAqB,IAAIC,WAA6CC,UAAUC,IAAI;AAE9FC,0BAAiB,oBAAIC,IAAAA;AAG1B,SAAKC,aAAab;AAClB,SAAKc,MAAMb;AACX,SAAKc,sBAAsBb;EAC7B;EAEA,MAAMc,OAAO;AAEX,SAAKH,cAAe,MAAMI,eAAe;MAAEhB,IAAI,KAAKa;MAAKd,WAAW,KAAKa;IAAW,CAAA;AACpF,SAAKK,WAAW,IAAIC,sBAAsB;MACxClB,IAAI,KAAKa;MACTM,WAAW;QACTC,YAAY,OAAOC,WAAW,KAAKC,YAAYD,MAAAA;QAC/CE,WAAW,YAAY,KAAKC,WAAU;MACxC;IACF,CAAA;AACA,UAAM,KAAKP,SAASF,OAAI;AACxB,SAAKU,UAAU,QAAQjB,UAAUkB,OAAM,EAAGC,MAAK,CAAA;AAE/C,SAAKC,eAAe,IAAIC,mBAAAA;AACxB,SAAKC,iBAAiB,IAAIC,wBAAAA;AAE1B,SAAKC,QAAQ,IAAIC,KAAK;MACpBC,QAAQ,KAAKT;MACbU,SAAS;QAAC,KAAKL;QAAgB,KAAKF;QAAc,KAAKxB;;MACvDgC,SAAS,KAAKnB;;;MAIdoB,aAAa,OAAOH,QAAyBI,eAAwB;AACnE,YAAIJ,OAAOK,WAAW,SAAA,GAAY;AAChC,iBAAO;QACT;AAEA,YAAI,CAACD,YAAY;AACf,iBAAO;QACT;AAEA,cAAME,eAAe,KAAKC,KAAKC,qBAAqBR,MAAAA;AACpD,YAAKM,cAAsBG,oBAAoB,sBAAsB;AACnE,iBAAO,KAAKvC,oBAAoBwC,gBAAgBV,QAAQ;YAAEI;UAAW,CAAA;QACvE;AAEA,cAAMO,MAAM,KAAKb,MAAMc,QAAQR,UAAAA,GAAaS,QAAAA;AAC5C,YAAI,CAACF,KAAK;AACR,gBAAMG,cAAc,KAAKtC,eAAeuC,IAAI,aAAaX,UAAAA,EAAY;AACrEY,UAAAA,KAAI,0BAA0B;YAAEhB;YAAQI;YAAYU;UAAY,GAAA;;;;;;AAChE,iBAAOA;QACT;AAEA,YAAI;AACF,gBAAMG,WAAWC,mBAAmBP,GAAAA;AACpC,cAAI,CAACM,UAAU;AACbD,YAAAA,KAAI,8CAA8C;cAAEhB;cAAQI;YAAW,GAAA;;;;;;AACvE,mBAAO;UACT;AAEA,gBAAMe,oBAAoB,KAAK/C,mBAAmBgD,IAAI9C,UAAU+C,KAAKJ,QAAAA,CAAAA;AAGrE,gBAAMK,eAAgBhB,cAAsBiB;AAC5C,cAAI,CAACD,cAAc;AACjBN,YAAAA,KAAI,+CAA+C;cAAEhB;cAAQI;YAAW,GAAA;;;;;;AACxE,mBAAO;UACT;AACA,gBAAMoB,YAAYlD,UAAU+C,KAAKC,YAAAA;AAEjC,gBAAMG,eAAeN,mBAAmBJ,IAAIS,SAAAA,KAAc;AAC1DR,UAAAA,KAAI,sBAAsB;YACxBU,WAAW,KAAKnC;YAChBoC,YAAY3B;YACZI;YACAoB;YACAP;YACAQ;UACF,GAAA;;;;;;AACA,iBAAOA;QACT,SAASG,KAAK;AACZZ,UAAAA,KAAIa,MAAMD,KAAAA,QAAAA;;;;;;AACV,iBAAO;QACT;MACF;IACF,CAAA;AACA,SAAKhC,eAAekC,MAAK;AACzB,SAAKpC,aAAaoC,MAAK;AACvB,UAAM,KAAK5D,oBAAoBW,KAAI;AAEnC,UAAM,KAAKe,eAAemC,cAAa;AACvC,UAAM,KAAK7D,oBAAoB6D,cAAa;EAC9C;EAEA,MAAMC,QAAQ;AACZ,UAAM,KAAKjD,SAASiD,QAAK;AACzB,UAAM,KAAKpC,eAAeoC,MAAK;AAC/B,UAAM,KAAK9D,oBAAoB8D,MAAK;AACpC,UAAM,KAAKhE,KAAKiE,QAAO;EACzB;EAEA,IAAI1B,OAAa;AACf,WAAO,KAAKT;EACd;EAEA,MAAMoC,cAAcC,YAA4B;AAC9C,UAAM,KAAKjE,oBAAoBgE,cAAcC,UAAAA;EAC/C;EAEA,MAAMC,iBAAiBD,YAA4B;AACjD,UAAM,KAAKjE,oBAAoBkE,iBAAiBD,UAAAA;EAClD;EAEA,MAAc/C,YAAY,EAAEiD,MAAMC,MAAK,GAAsB;AAC3D,UAAMC,SAAS,KAAKzC,MAAMc,QAAQyB,KAAK,CAAA,CAAE;AACzC,QAAI,CAACE,QAAQ;AACX;IACF;AACA,UAAM5B,MAAM4B,OAAO1B,QAAO;AAC1B,QAAI,CAACF,KAAK;AACR;IACF;AAEA,UAAM6B,oBAAoBC,SAAS9B,GAAAA;AAEnC,UAAM+B,YAAYC,OAAOC,KAAKjC,IAAIkC,WAAW,CAAC,CAAA;AAC9C,UAAMC,aAAaJ,UAAUK,IAAI,CAACC,aAAaC,QAAQC,OAAO;MAAE9C,YAAYmC,OAAOnC;MAAY4C;IAAS,CAAA,CAAA;AACxG,UAAMG,eAAe,IAAIC,IAAIN,WAAWC,IAAI,CAACM,OAAO;MAACA;MAAIb;KAAkB,CAAA;AAC3E,SAAK5D,oBAAoB0E,UAAUH,cAAcb,KAAAA;EACnD;;;;EAKA,MAAchD,aAAa;AACzB,SAAKV,oBAAoB2E,kBAAiB;EAC5C;EAGQC,iBAAiB;AACvB,WAAOC,UAAU,KAAK3D,MAAMc,SAAS,CAAC2B,YAAY;MAChDmB,OAAOnB,OAAOmB;MACdC,QAAQ,CAAC,CAACpB,OAAO1B,QAAO;MACxB+C,OAAOrB,OAAO1B,QAAO,IAAKgD,UAAUpB,SAASF,OAAO1B,QAAO,CAAA,IAAM;MACjEiD,MACEvB,OAAO1B,QAAO,KACd4C,UAAUlB,OAAO1B,QAAO,GAAI,CAACkD,OAAOC,QAAAA;AAClC,YAAI;AACF,kBAAQA,KAAAA;YACN,KAAK;YACL,KAAK;AACH,qBAAOD;YACT,KAAK;AACH,qBAAOpB,OAAOC,KAAKmB,KAAAA;YACrB;AACE,qBAAO,GAAGA,KAAAA;UACd;QACF,SAASnC,KAAK;AACZ,iBAAO,GAAGA,GAAAA;QACZ;MACF,CAAA;IACJ,EAAA;EACF;EAGQqC,kBAAkB;AACxB,WAAO,KAAKnE,MAAMoE;EACpB;;;;EAKA,MACMC,MAAM,EAAEC,OAAM,GAAiC;AAEnD,UAAMC,QAAQC,IACZF,QAAQrB,IAAI,OAAO,EAAEa,OAAOxD,WAAU,MAAE;AACtCmE,MAAAA,WAAUX,OAAO,gCAAA;;;;;;;;;AACjB,YAAMrB,SAAS,KAAKhC,KAAKK,QAAQR,UAAAA,KAA6B,KAAKN,MAAM0E,KAAKpE,UAAAA;AAC9E,YAAMqE,aAAalC,QAAQqB,KAAAA;IAC7B,CAAA,KAAM,CAAA,CAAE;AAGV,UAAM,KAAK9D,MAAMqE,MAAMC,QAAQrB,IAAI,CAAC,EAAE3C,WAAU,MAAOA,UAAAA,CAAAA;EACzD;EAEAsE,SAASC,SAAoD;AAC3D,WAAO,KAAK/E,eAAe8E,SAASC,OAAAA;EACtC;EAEAC,gBAAgBD,SAAyC;AACvD,WAAO,KAAK/E,eAAegF,gBAAgBD,OAAAA;EAC7C;EAEA,MAAME,cAAiC;AACrC,WAAO,KAAKjF,eAAeiF,YAAW;EACxC;;;;EAMAC,kBAAuC;AACrC,WAAO,KAAKpF,aAAaoF,gBAAe;EAC1C;EAEAC,gBAAgB9D,UAAqBO,WAAsB;AACzDR,IAAAA,KAAI,mBAAmB;MAAEC;MAAUO;IAAU,GAAA;;;;;;AAC7CwD,eAAW,KAAK5G,oBAAoB6C,UAAU,MAAM,IAAIgE,WAAW3G,UAAUC,IAAI,CAAA,EAAG2G,IAAI1D,SAAAA;EAC1F;AACF;;EAzNG2D,MAAMC,KAAI;GAZAzH,cAAAA,WAAAA,WAAAA,MAAAA;;EA6JVwH,MAAMC,KAAK;IAAEC,OAAO;EAAK,CAAA;GA7Jf1H,cAAAA,WAAAA,kBAAAA,IAAAA;;EAuLVwH,MAAMC,KAAK;IAAEC,OAAO;EAAK,CAAA;GAvLf1H,cAAAA,WAAAA,mBAAAA,IAAAA;;EA+LVwH,MAAMG,KAAK;IAAEC,uBAAuB;EAAK,CAAA;GA/L/B5H,cAAAA,WAAAA,SAAAA,IAAAA;AAAAA,gBAAAA,cAAAA;EADZwH,MAAMK,SAAQ;GACF7H,aAAAA;AAuON,IAAMuD,qBAAqB,CAACP,QAAAA;AAEjC,QAAM8E,cAAc9E,IAAI+E,QAAQzE,YAAYN,IAAIgF;AAChD,MAAIF,eAAe,MAAM;AACvB,WAAO;EACT;AAEA,SAAOG,OAAOH,WAAAA;AAChB;AAEA,IAAMhB,eAAe,OAAOlC,QAA6BqB,UAAAA;AACvD,QAAMrB,OAAOsD,UAAS;AACtB,QAAMC,mBAAmB,IAAIrH,IAAImF,KAAAA;AAEjC,QAAMmC,MAAMC,KAAuCzD,QAAQ,QAAA,EAAU0D,iBAAiB,MAAA;AAEpF,eAAWC,cAAcJ,iBAAiBK,OAAM,GAAI;AAClD,UAAIC,qBAAqB7D,OAAO1B,QAAO,GAAIqF,UAAAA,GAAa;AACtDJ,yBAAiBO,OAAOH,UAAAA;MAC1B;IACF;AAEA,QAAIJ,iBAAiBQ,SAAS,GAAG;AAC/B,aAAO;IACT;AACA,WAAO;EACT,CAAA;AACF;AAEA,IAAMF,uBAAuB,CAACzF,KAAeuF,eAAAA;AAC3C,SAAO,CAAC,CAACK,WAAW5F,GAAAA,EAAK6F,gBAAgBN,UAAAA;AAC3C;;;AOzTA,SAASO,SAAAA,cAAa;AAEtB,SAASC,yBAAuC;AAChD,SAASC,wBAAwB;AAEjC,SAASC,aAAAA,kBAAiB;AAE1B,SAASC,OAAAA,YAAW;AACpB,SAASC,SAAAA,cAAa;;;;;;;;;;;;AA0Bf,IAAMC,8BAAN,MAAMA;EAcXC,YACmBC,WACAC,OACjB;SAFiBD,YAAAA;SACAC,QAAAA;SAfXC,sBAAkD;SAIzCC,yBAAyB,oBAAIC,IAAAA;SAK7BC,8BAA8B,oBAAIC,IAAAA;SAEnCC,yBAAyB,IAAIf,OAAAA;EAK1C;EAEHgB,gBAAuC;AACrC,WAAO,KAAKN,uBAAuB,OAC/B;MAAC,KAAKA;SAAwB,IAAII,IAAI,KAAKH,uBAAuBM,OAAM,CAAA;QACxE,CAAA;EACN;EAEA,MACaC,uBAAuBC,KAAcC,YAAuC;AACvF,QAAI,KAAKV,uBAAuB,MAAM;AACpC;IACF;AACA,QAAI,CAACU,WAAWC,SAAS;AACvBjB,MAAAA,KAAIkB,MAAM,mCAAmC;QAAEC,UAAU,KAAKf;MAAU,GAAA;;;;;;AACxE,WAAKgB,qCAAqCL,GAAAA;IAC5C,OAAO;AACL,YAAMM,oBAAoB,MAAM,KAAKC,eAAeP,KAAKC,WAAWC,OAAO;AAC3E,YAAMM,MAAMF,kBAAkBG,QAAO;AACrCzB,MAAAA,WAAUwB,KAAAA,QAAAA;;;;;;;;;AACV,UAAIA,IAAIE,UAAU,MAAM;AACtB,aAAKC,eAAeL,iBAAAA;MACtB;AACA,WAAKf,sBAAsBe;IAC7B;EACF;EAEOM,mBAAmBC,gBAAmC;AAC3D,UAAMC,YAAYC,MAAMC,QAAQH,cAAAA,IAAkBA,iBAAiB;MAACA;;AACpE,QAAII,gBAAgB;AACpB,UAAMC,aAAgC,CAAC;AACvC,eAAWC,YAAYL,WAAW;AAChC9B,MAAAA,WAAU,KAAKO,qBAAmB,QAAA;;;;;;;;;AAClC,UAAI,KAAKC,uBAAuB4B,IAAID,QAAAA,KAAa,KAAKzB,4BAA4B0B,IAAID,QAAAA,GAAW;AAC/F;MACF;AACA,YAAME,eAAe,KAAK9B,oBAAoBkB,QAAO;AACrDzB,MAAAA,WAAUqC,cAAAA,QAAAA;;;;;;;;;AACV,YAAMC,eAAeD,aAAaE,SAAS,CAAC,GAAGJ,QAAAA;AAC/C,UAAIG,eAAe,MAAM;AACvB,aAAK5B,4BAA4B8B,IAAIL,QAAAA;AACrClC,QAAAA,KAAIwC,KAAK,sDAAsD;UAAEN;QAAS,GAAA;;;;;;MAC5E,OAAO;AACLD,mBAAWC,QAAAA,IAAYG;AACvBL,wBAAgB;MAClB;IACF;AACA,QAAIA,eAAe;AACjB,WAAKS,mBAAmBR,UAAAA;IAC1B;EACF;EAEOS,qBAAqBJ,OAA2B;AACrD,QAAI,CAACA,OAAO;AACV;IACF;AACA,UAAMK,oBAAoBC,OAAOC,QAAQP,KAAAA,EAAOQ,OAAO,CAAC,CAACZ,QAAAA,MACvD,KAAKzB,4BAA4B0B,IAAID,QAAAA,CAAAA;AAEvC,SAAKO,mBAAmBG,OAAOG,YAAYJ,iBAAAA,CAAAA;AAC3CA,sBAAkBK,QAAQ,CAAC,CAACd,QAAAA,MAAc,KAAKzB,4BAA4BwC,OAAOf,QAAAA,CAAAA;EACpF;EAEOgB,wBAA6C;AAClDnD,IAAAA,WAAU,KAAKO,qBAAmB,QAAA;;;;;;;;;AAClC,WAAO,KAAKA;EACd;EAEO6C,wBAAwBjB,UAAuC;AACpEnC,IAAAA,WAAU,KAAKO,qBAAmB,QAAA;;;;;;;;;AAClC,UAAM8C,iBAAiB,KAAK/C,MAAMgD,OAAM;AACxC,SAAK3B,eAAe0B,cAAAA;AACpB,SAAKE,wBAAwBF,gBAAgBlB,QAAAA;AAC7C,SAAK5B,oBAAoBiD,OAAO,CAACC,WAAAA;AAC/BA,aAAOlB,UAAU,CAAC;AAClBkB,aAAOlB,MAAMJ,QAAAA,IAAYkB,eAAeK;IAC1C,CAAA;AACA,WAAOL;EACT;EAEOE,wBAAwBI,QAA6BxB,UAAkB;AAC5E,SAAK3B,uBAAuBoD,IAAIzB,UAAUwB,MAAAA;EAC5C;EAEOE,wBAAkC;AACvC,UAAMC,qBAAqB;SAAI,KAAKtD,uBAAuBuD,KAAI;;AAC/D,SAAKvD,uBAAuBwD,MAAK;AACjC,SAAKzD,sBAAsB;AAC3B,WAAOuD;EACT;EAEQpB,mBAAmBH,OAA2B;AACpD,QAAI,CAACA,OAAO;AACV;IACF;AACA,eAAW,CAACJ,UAAU8B,YAAAA,KAAiBpB,OAAOC,QAAQP,KAAAA,GAAQ;AAC5D,YAAM2B,UAAU;QAAE/B;QAAU8B;MAAa;AACzC,YAAME,uBAAuB,KAAK3D,uBAAuB4D,IAAIjC,QAAAA;AAC7D,UAAIgC,wBAAwB,QAAQA,qBAAqBT,QAAQO,cAAc;AAC7EhE,QAAAA,KAAIoE,KAAK,qEAAqE;UAC5E,GAAGH;UACHI,mBAAmBH,qBAAqBT;QAC1C,GAAA;;;;;;AACA;MACF;AACA,UAAIS,sBAAsBT,QAAQO,cAAc;AAC9ChE,QAAAA,KAAIoE,KAAK,sCAAsCH,SAAAA;;;;;;AAC/C;MACF;AACA,YAAMP,SAAS,KAAKrD,MAAMiE,KAAeN,YAAAA;AACzChE,MAAAA,KAAIuE,MAAM,8BAA8BN,SAAAA;;;;;;AACxC,WAAK1D,uBAAuBoD,IAAIzB,UAAUwB,MAAAA;AAC1C,WAAK,KAAKc,4BAA4Bd,QAAQxB,QAAAA;IAChD;EACF;EAEA,MAAcZ,eAAeP,KAAc0C,KAAa;AACtD,UAAMgB,YAAY,KAAKpE,MAAMiE,KAAeb,GAAAA;AAC5C,WAAO,MAAM;AACX,UAAI;AACF,cAAM3D,iBAAiB,KAAO,iDAAiD,YAAA;AAC7E,gBAAMD,kBAAkBkB,KAAK0D,UAAUC,UAAS,CAAA;QAClD,CAAA;AACA;MACF,SAASC,KAAK;AACZ,YAAI,GAAGA,GAAAA,GAAMC,SAAS,SAAA,GAAY;AAChC5E,UAAAA,KAAIwC,KAAK,cAAc;YAAEqC,IAAIJ,UAAUK;YAAYC,OAAON,UAAUM;UAAM,GAAA;;;;;;AAC1E;QACF;AAEA,cAAMJ;MACR;IACF;AAEA,QAAIF,UAAUM,UAAU,eAAe;AACrC,YAAM,IAAIC,MAAM,mCAAA;IAClB;AAEA,WAAOP;EACT;EAEQrD,qCAAqCL,KAAc;AACzD,UAAM0D,YAAY,KAAKpE,MAAMgD,OAAM;AACnC,SAAK/C,sBAAsBmE;AAC3B1D,QAAIkE,UAAU,MAAA;AACZR,gBAAUxB,OAAM;AAChB,WAAK3C,sBAAsB;IAC7B,CAAA;EACF;EAEQoB,eAAegC,QAA6B;AAClDA,WAAOH,OAAO,CAACC,WAAAA;AACbA,aAAO/B,WAAW;QAAEN,UAAU,KAAKf,UAAU8E,MAAK;MAAG;AACrD1B,aAAO/B,OAAON,WAAW,KAAKf,UAAU8E,MAAK;IAC/C,CAAA;EACF;EAEA,MAAcV,4BAA4Bd,QAA6BxB,UAAkB;AACvF,QAAI;AACF,YAAMwB,OAAOnC,IAAI;QAAC;OAAQ;AAC1B,YAAM0C,UAAU;QAAE/B;QAAUiD,QAAQzB,OAAOD;MAAI;AAC/C,UAAI,KAAK9C,uBAAuByE,cAAa,MAAO,GAAG;AACrDpF,QAAAA,KAAIwC,KAAK,oDAAoDyB,SAAAA;;;;;;AAC7D;MACF;AACA,YAAMoB,kBAAkB,KAAK9E,uBAAuB4D,IAAIjC,QAAAA;AACxD,UAAImD,iBAAiB5B,QAAQC,OAAOD,KAAK;AACvCzD,QAAAA,KAAIoE,KAAK,sEAAsEH,SAAAA;;;;;;AAC/E;MACF;AACA,WAAKtD,uBAAuB2E,KAAK;QAAE5B;QAAQxB;MAAS,CAAA;IACtD,SAASyC,KAAK;AACZ,YAAMY,qBAAqB,KAAK5E,uBAAuByE,cAAa,IAAK;AACzEpF,MAAAA,KAAIoE,KAAK,6BAA6B;QACpClC;QACA8B,cAAcN,OAAOD;QACrB+B,cAAcD;QACdZ;MACF,GAAA;;;;;;AACA,UAAIY,oBAAoB;AACtB,cAAM,KAAKf,4BAA4Bd,QAAQxB,QAAAA;MACjD;IACF;EACF;AACF;;EAhLGjC,OAAMwF,KAAK;IAAEC,uBAAuB;EAAK,CAAA;GAzB/BxF,4BAAAA,WAAAA,0BAAAA,IAAAA;AAAAA,8BAAAA,cAAAA;EADZD,OAAM0F,SAAQ;GACFzF,2BAAAA;",
|
|
6
6
|
"names": ["Event", "next", "automerge", "getBackend", "getHeads", "Repo", "Context", "invariant", "PublicKey", "log", "idCodec", "trace", "ComplexMap", "ComplexSet", "defaultMap", "mapValues", "Trigger", "synchronized", "NetworkAdapter", "LifecycleState", "invariant", "log", "EchoNetworkAdapter", "_replicators", "Set", "_connections", "Map", "_lifecycleState", "CLOSED", "_connected", "connect", "peerId", "peerMetadata", "wake", "send", "message", "connectionEntry", "get", "targetId", "Error", "writer", "write", "catch", "err", "isOpen", "disconnect", "open", "OPEN", "emit", "network", "close", "replicator", "clear", "whenConnected", "wait", "timeout", "addReplicator", "has", "onConnectionOpen", "_onConnectionOpen", "bind", "onConnectionClosed", "_onConnectionClosed", "removeReplicator", "shouldAdvertize", "params", "connection", "reader", "readable", "getReader", "writable", "getWriter", "set", "queueMicrotask", "done", "value", "read", "dxos_peerSource", "entry", "cancel", "abort", "delete", "LifecycleState", "Resource", "LevelDBStorageAdapter", "Resource", "constructor", "_params", "load", "keyArray", "_lifecycleState", "LifecycleState", "OPEN", "undefined", "db", "get", "encodingOptions", "err", "isLevelDbNotFoundError", "save", "binary", "batch", "callbacks", "beforeSave", "path", "put", "Buffer", "from", "write", "afterSave", "remove", "del", "loadRange", "keyPrefix", "result", "key", "value", "iterator", "gte", "lte", "push", "data", "removeRange", "keyEncoder", "encode", "map", "k", "replaceAll", "join", "decode", "toString", "split", "format", "keyEncoding", "valueEncoding", "code", "Trigger", "NetworkAdapter", "cbor", "Stream", "invariant", "LocalHostNetworkAdapter", "_peers", "Map", "_connected", "_isConnected", "ready", "emit", "network", "connect", "peerId", "wake", "send", "message", "peer", "get", "targetId", "close", "forEach", "disconnect", "whenConnected", "wait", "timeout", "syncRepo", "id", "syncMessage", "_getPeerId", "next", "has", "set", "connected", "encode", "delete", "peerMetadata", "sendSyncMessage", "decode", "getHostInfo", "Trigger", "NetworkAdapter", "cbor", "invariant", "log", "AutomergeReplicator", "MeshNetworkAdapter", "_extensions", "Map", "_connected", "ready", "emit", "network", "connect", "peerId", "wake", "send", "message", "receiverId", "targetId", "extension", "get", "sendSyncMessage", "payload", "encode", "catch", "err", "disconnect", "createExtension", "peerInfo", "onStartReplication", "info", "remotePeerId", "wait", "id", "thisPeerId", "toHex", "has", "set", "peerMetadata", "dxos_deviceKey", "onSyncMessage", "decode", "onClose", "delete", "IndexedDBStorageAdapter", "log", "StorageType", "arrayToBuffer", "bufferToArray", "AutomergeStorageAdapter", "constructor", "_directory", "_state", "load", "key", "undefined", "filename", "_getFilename", "file", "getOrCreateFile", "size", "stat", "buffer", "read", "bufferToArray", "save", "data", "write", "arrayToBuffer", "truncate", "length", "flush", "remove", "destroy", "loadRange", "keyPrefix", "entries", "list", "Promise", "all", "filter", "entry", "startsWith", "map", "_getKeyFromFilename", "removeRange", "close", "k", "replaceAll", "join", "split", "levelMigration", "db", "directory", "isNewLevel", "iterator", "encodingOptions", "next", "oldStorageAdapter", "type", "StorageType", "IDB", "IndexedDBStorageAdapter", "path", "AutomergeStorageAdapter", "chunks", "loadRange", "length", "batch", "log", "info", "key", "data", "put", "write", "AutomergeHost", "constructor", "directory", "db", "indexMetadataStore", "_ctx", "Context", "_echoNetworkAdapter", "EchoNetworkAdapter", "_authorizedDevices", "ComplexMap", "PublicKey", "hash", "_requestedDocs", "Set", "_directory", "_db", "_indexMetadataStore", "open", "levelMigration", "_storage", "LevelDBStorageAdapter", "callbacks", "beforeSave", "params", "_beforeSave", "afterSave", "_afterSave", "_peerId", "random", "toHex", "_meshNetwork", "MeshNetworkAdapter", "_clientNetwork", "LocalHostNetworkAdapter", "_repo", "Repo", "peerId", "network", "storage", "sharePolicy", "documentId", "startsWith", "peerMetadata", "repo", "peerMetadataByPeerId", "dxos_peerSource", "shouldAdvertize", "doc", "handles", "docSync", "isRequested", "has", "log", "spaceKey", "getSpaceKeyFromDoc", "authorizedDevices", "get", "from", "deviceKeyHex", "dxos_deviceKey", "deviceKey", "isAuthorized", "localPeer", "remotePeer", "err", "catch", "ready", "whenConnected", "close", "dispose", "addReplicator", "replicator", "removeReplicator", "path", "batch", "handle", "lastAvailableHash", "getHeads", "objectIds", "Object", "keys", "objects", "encodedIds", "map", "objectId", "idCodec", "encode", "idToLastHash", "Map", "id", "markDirty", "notifyMarkedDirty", "_automergeDocs", "mapValues", "state", "hasDoc", "heads", "automerge", "data", "value", "key", "_automergePeers", "peers", "flush", "states", "Promise", "all", "invariant", "find", "waitForHeads", "syncRepo", "request", "sendSyncMessage", "getHostInfo", "createExtension", "authorizeDevice", "defaultMap", "ComplexSet", "add", "trace", "info", "depth", "span", "showInBrowserTimeline", "resource", "rawSpaceKey", "access", "experimental_spaceKey", "String", "whenReady", "unavailableHeads", "Event", "wrap", "waitForCondition", "changeHash", "values", "changeIsPresentInDoc", "delete", "size", "getBackend", "getChangeByHash", "Event", "cancelWithContext", "warnAfterTimeout", "invariant", "log", "trace", "AutomergeDocumentLoaderImpl", "constructor", "_spaceKey", "_repo", "_spaceRootDocHandle", "_objectDocumentHandles", "Map", "_objectsPendingDocumentLoad", "Set", "onObjectDocumentLoaded", "getAllHandles", "values", "loadSpaceRootDocHandle", "ctx", "spaceState", "rootUrl", "error", "spaceKey", "_createContextBoundSpaceRootDocument", "existingDocHandle", "_initDocHandle", "doc", "docSync", "access", "_initDocAccess", "loadObjectDocument", "objectIdOrMany", "objectIds", "Array", "isArray", "hasUrlsToLoad", "urlsToLoad", "objectId", "has", "spaceRootDoc", "documentUrl", "links", "add", "info", "_loadLinkedObjects", "onObjectLinksUpdated", "linksAwaitingLoad", "Object", "entries", "filter", "fromEntries", "forEach", "delete", "getSpaceRootDocHandle", "createDocumentForObject", "spaceDocHandle", "create", "onObjectBoundToDocument", "change", "newDoc", "url", "handle", "set", "clearHandleReferences", "objectsWithHandles", "keys", "clear", "automergeUrl", "logMeta", "objectDocumentHandle", "get", "warn", "actualDocumentUrl", "find", "debug", "_createObjectOnDocumentLoad", "docHandle", "whenReady", "err", "includes", "id", "documentId", "state", "Error", "onDispose", "toHex", "docUrl", "listenerCount", "objectDocHandle", "emit", "shouldRetryLoading", "retryLoading", "span", "showInBrowserTimeline", "resource"]
|
|
7
7
|
}
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
{
|
|
2
2
|
"version": 3,
|
|
3
3
|
"sources": ["../../../src/automerge/automerge-host.ts", "../../../src/automerge/echo-network-adapter.ts", "../../../src/automerge/leveldb-storage-adapter.ts", "../../../src/automerge/local-host-network-adapter.ts", "../../../src/automerge/mesh-network-adapter.ts", "../../../src/automerge/migrations.ts", "../../../src/automerge/automerge-storage-adapter.ts", "../../../src/automerge/automerge-doc-loader.ts"],
|
|
4
|
-
"sourcesContent": ["//\n// Copyright 2023 DXOS.org\n//\n\nimport { Event } from '@dxos/async';\nimport { type Doc, next as automerge, getBackend, type Heads, getHeads } from '@dxos/automerge/automerge';\nimport {\n type DocHandle,\n Repo,\n type DocumentId,\n type PeerId,\n type StorageAdapterInterface,\n type DocHandleChangePayload,\n} from '@dxos/automerge/automerge-repo';\nimport { type Stream } from '@dxos/codec-protobuf';\nimport { Context, type Lifecycle } from '@dxos/context';\nimport { type SpaceDoc } from '@dxos/echo-protocol';\nimport { type IndexMetadataStore } from '@dxos/indexing';\nimport { invariant } from '@dxos/invariant';\nimport { PublicKey } from '@dxos/keys';\nimport { type SubLevelDB } from '@dxos/kv-store';\nimport { log } from '@dxos/log';\nimport { idCodec } from '@dxos/protocols';\nimport {\n type FlushRequest,\n type HostInfo,\n type SyncRepoRequest,\n type SyncRepoResponse,\n} from '@dxos/protocols/proto/dxos/echo/service';\nimport { type Directory } from '@dxos/random-access-storage';\nimport { type AutomergeReplicator } from '@dxos/teleport-extension-automerge-replicator';\nimport { trace } from '@dxos/tracing';\nimport { ComplexMap, ComplexSet, defaultMap, mapValues } from '@dxos/util';\n\nimport { EchoNetworkAdapter } from './echo-network-adapter';\nimport { type EchoReplicator } from './echo-replicator';\nimport { type BeforeSaveParams, LevelDBStorageAdapter } from './leveldb-storage-adapter';\nimport { LocalHostNetworkAdapter } from './local-host-network-adapter';\nimport { MeshNetworkAdapter } from './mesh-network-adapter';\nimport { levelMigration } from './migrations';\n\n// TODO: Remove\nexport type { DocumentId };\n\nexport type AutomergeHostParams = {\n db: SubLevelDB;\n /**\n * For migration purposes.\n */\n directory?: Directory;\n\n indexMetadataStore: IndexMetadataStore;\n};\n\n@trace.resource()\nexport class AutomergeHost {\n private readonly _indexMetadataStore: IndexMetadataStore;\n private readonly _ctx = new Context();\n private readonly _directory?: Directory;\n private readonly _db: SubLevelDB;\n private readonly _echoNetworkAdapter = new EchoNetworkAdapter();\n\n private _repo!: Repo;\n private _meshNetwork!: MeshNetworkAdapter;\n private _clientNetwork!: LocalHostNetworkAdapter;\n private _storage!: StorageAdapterInterface & Lifecycle;\n\n @trace.info()\n private _peerId!: string;\n\n /**\n * spaceKey -> deviceKey[]\n */\n private readonly _authorizedDevices = new ComplexMap<PublicKey, ComplexSet<PublicKey>>(PublicKey.hash);\n\n public _requestedDocs = new Set<string>();\n\n constructor({ directory, db, indexMetadataStore }: AutomergeHostParams) {\n this._directory = directory;\n this._db = db;\n this._indexMetadataStore = indexMetadataStore;\n }\n\n async open() {\n // TODO(mykola): remove this before 0.6 release.\n this._directory && (await levelMigration({ db: this._db, directory: this._directory }));\n this._storage = new LevelDBStorageAdapter({\n db: this._db,\n callbacks: {\n beforeSave: async (params) => this._beforeSave(params),\n afterSave: async () => this._afterSave(),\n },\n });\n await this._storage.open?.();\n this._peerId = `host-${PublicKey.random().toHex()}` as PeerId;\n\n this._meshNetwork = new MeshNetworkAdapter();\n this._clientNetwork = new LocalHostNetworkAdapter();\n\n this._repo = new Repo({\n peerId: this._peerId as PeerId,\n network: [this._clientNetwork, this._meshNetwork, this._echoNetworkAdapter],\n storage: this._storage,\n\n // TODO(dmaretskyi): Share based on HALO permissions and space affinity.\n // Hosts, running in the worker, don't share documents unless requested by other peers.\n sharePolicy: async (peerId /* device key */, documentId /* space key */) => {\n if (peerId.startsWith('client-')) {\n return false; // Only send docs to clients if they are requested.\n }\n\n if (!documentId) {\n return false;\n }\n\n const peerMetadata = this.repo.peerMetadataByPeerId[peerId];\n if ((peerMetadata as any)?.dxos_peerSource === 'EchoNetworkAdapter') {\n return this._echoNetworkAdapter.shouldAdvertize(peerId, { documentId });\n }\n\n const doc = this._repo.handles[documentId]?.docSync();\n if (!doc) {\n const isRequested = this._requestedDocs.has(`automerge:${documentId}`);\n log('doc share policy check', { peerId, documentId, isRequested });\n return isRequested;\n }\n\n try {\n const spaceKey = getSpaceKeyFromDoc(doc);\n if (!spaceKey) {\n log('space key not found for share policy check', { peerId, documentId });\n return false;\n }\n\n const authorizedDevices = this._authorizedDevices.get(PublicKey.from(spaceKey));\n\n // TODO(mykola): Hack, stop abusing `peerMetadata` field.\n const deviceKeyHex = (peerMetadata as any)?.dxos_deviceKey;\n if (!deviceKeyHex) {\n log('device key not found for share policy check', { peerId, documentId });\n return false;\n }\n const deviceKey = PublicKey.from(deviceKeyHex);\n\n const isAuthorized = authorizedDevices?.has(deviceKey) ?? false;\n log('share policy check', {\n localPeer: this._peerId,\n remotePeer: peerId,\n documentId,\n deviceKey,\n spaceKey,\n isAuthorized,\n });\n return isAuthorized;\n } catch (err) {\n log.catch(err);\n return false;\n }\n },\n });\n this._clientNetwork.ready();\n this._meshNetwork.ready();\n await this._echoNetworkAdapter.open();\n\n await this._clientNetwork.whenConnected();\n await this._echoNetworkAdapter.whenConnected();\n }\n\n async close() {\n await this._storage.close?.();\n await this._clientNetwork.close();\n await this._echoNetworkAdapter.close();\n await this._ctx.dispose();\n }\n\n get repo(): Repo {\n return this._repo;\n }\n\n async addReplicator(replicator: EchoReplicator) {\n await this._echoNetworkAdapter.addReplicator(replicator);\n }\n\n async removeReplicator(replicator: EchoReplicator) {\n await this._echoNetworkAdapter.removeReplicator(replicator);\n }\n\n private async _beforeSave({ path, batch }: BeforeSaveParams) {\n const handle = this._repo.handles[path[0] as DocumentId];\n if (!handle) {\n return;\n }\n const doc = handle.docSync();\n if (!doc) {\n return;\n }\n\n const lastAvailableHash = getHeads(doc);\n\n const objectIds = Object.keys(doc.objects ?? {});\n const encodedIds = objectIds.map((objectId) => idCodec.encode({ documentId: handle.documentId, objectId }));\n const idToLastHash = new Map(encodedIds.map((id) => [id, lastAvailableHash]));\n this._indexMetadataStore.markDirty(idToLastHash, batch);\n }\n\n /**\n * Called by AutomergeStorageAdapter after levelDB batch commit.\n */\n private async _afterSave() {\n this._indexMetadataStore.notifyMarkedDirty();\n }\n\n @trace.info({ depth: null })\n private _automergeDocs() {\n return mapValues(this._repo.handles, (handle) => ({\n state: handle.state,\n hasDoc: !!handle.docSync(),\n heads: handle.docSync() ? automerge.getHeads(handle.docSync()) : null,\n data:\n handle.docSync() &&\n mapValues(handle.docSync(), (value, key) => {\n try {\n switch (key) {\n case 'access':\n case 'links':\n return value;\n case 'objects':\n return Object.keys(value as any);\n default:\n return `${value}`;\n }\n } catch (err) {\n return `${err}`;\n }\n }),\n }));\n }\n\n @trace.info({ depth: null })\n private _automergePeers() {\n return this._repo.peers;\n }\n\n //\n // Methods for client-services.\n //\n @trace.span({ showInBrowserTimeline: true })\n async flush({ states }: FlushRequest): Promise<void> {\n // Note: Wait for all requested documents to be loaded/synced from thin-client.\n await Promise.all(\n states?.map(async ({ heads, documentId }) => {\n invariant(heads, 'heads are required for flush');\n const handle = this.repo.handles[documentId as DocumentId] ?? this._repo.find(documentId as DocumentId);\n await waitForHeads(handle, heads);\n }) ?? [],\n );\n\n await this._repo.flush(states?.map(({ documentId }) => documentId as DocumentId));\n }\n\n syncRepo(request: SyncRepoRequest): Stream<SyncRepoResponse> {\n return this._clientNetwork.syncRepo(request);\n }\n\n sendSyncMessage(request: SyncRepoRequest): Promise<void> {\n return this._clientNetwork.sendSyncMessage(request);\n }\n\n async getHostInfo(): Promise<HostInfo> {\n return this._clientNetwork.getHostInfo();\n }\n\n //\n // Mesh replication.\n //\n\n createExtension(): AutomergeReplicator {\n return this._meshNetwork.createExtension();\n }\n\n authorizeDevice(spaceKey: PublicKey, deviceKey: PublicKey) {\n log('authorizeDevice', { spaceKey, deviceKey });\n defaultMap(this._authorizedDevices, spaceKey, () => new ComplexSet(PublicKey.hash)).add(deviceKey);\n }\n}\n\nexport const getSpaceKeyFromDoc = (doc: any): string | null => {\n // experimental_spaceKey is set on old documents, new ones are created with doc.access.spaceKey\n const rawSpaceKey = doc.access?.spaceKey ?? doc.experimental_spaceKey;\n if (rawSpaceKey == null) {\n return null;\n }\n\n return String(rawSpaceKey);\n};\n\nconst waitForHeads = async (handle: DocHandle<SpaceDoc>, heads: Heads) => {\n await handle.whenReady();\n const unavailableHeads = new Set(heads);\n\n await Event.wrap<DocHandleChangePayload<SpaceDoc>>(handle, 'change').waitForCondition(() => {\n // Check if unavailable heads became available.\n for (const changeHash of unavailableHeads.values()) {\n if (changeIsPresentInDoc(handle.docSync(), changeHash)) {\n unavailableHeads.delete(changeHash);\n }\n }\n\n if (unavailableHeads.size === 0) {\n return true;\n }\n return false;\n });\n};\n\nconst changeIsPresentInDoc = (doc: Doc<any>, changeHash: string): boolean => {\n return !!getBackend(doc).getChangeByHash(changeHash);\n};\n", "//\n// Copyright 2024 DXOS.org\n//\n\nimport { Trigger, synchronized } from '@dxos/async';\nimport { type Message, NetworkAdapter, type PeerId, type PeerMetadata } from '@dxos/automerge/automerge-repo';\nimport { LifecycleState } from '@dxos/context';\nimport { invariant } from '@dxos/invariant';\nimport { log } from '@dxos/log';\n\nimport { type EchoReplicator, type ReplicatorConnection, type ShouldAdvertizeParams } from './echo-replicator';\n\n/**\n * Manages a set of {@link EchoReplicator} instances.\n */\nexport class EchoNetworkAdapter extends NetworkAdapter {\n private readonly _replicators = new Set<EchoReplicator>();\n /**\n * Remote peer id -> connection.\n */\n private readonly _connections = new Map<PeerId, ConnectionEntry>();\n private _lifecycleState: LifecycleState = LifecycleState.CLOSED;\n private readonly _connected = new Trigger();\n\n override connect(peerId: PeerId, peerMetadata?: PeerMetadata | undefined): void {\n this.peerId = peerId;\n this.peerMetadata = peerMetadata;\n this._connected.wake();\n }\n\n override send(message: Message): void {\n const connectionEntry = this._connections.get(message.targetId);\n if (!connectionEntry) {\n throw new Error('Connection not found.');\n }\n\n // TODO(dmaretskyi): Find a way to enforce backpressure on AM-repo.\n connectionEntry.writer.write(message).catch((err) => {\n if (connectionEntry.isOpen) {\n log.catch(err);\n }\n });\n }\n\n override disconnect(): void {\n // No-op\n }\n\n @synchronized\n async open() {\n invariant(this._lifecycleState === LifecycleState.CLOSED);\n this._lifecycleState = LifecycleState.OPEN;\n\n this.emit('ready', {\n network: this,\n });\n }\n\n @synchronized\n async close() {\n invariant(this._lifecycleState === LifecycleState.OPEN);\n\n for (const replicator of this._replicators) {\n await replicator.disconnect();\n }\n this._replicators.clear();\n\n this._lifecycleState = LifecycleState.CLOSED;\n }\n\n async whenConnected() {\n await this._connected.wait({ timeout: 10_000 });\n }\n\n @synchronized\n async addReplicator(replicator: EchoReplicator) {\n invariant(this.peerId);\n invariant(!this._replicators.has(replicator));\n\n await replicator.connect({\n peerId: this.peerId,\n onConnectionOpen: this._onConnectionOpen.bind(this),\n onConnectionClosed: this._onConnectionClosed.bind(this),\n });\n }\n\n @synchronized\n async removeReplicator(replicator: EchoReplicator) {\n invariant(this._replicators.has(replicator));\n await replicator.disconnect();\n }\n\n async shouldAdvertize(peerId: PeerId, params: ShouldAdvertizeParams): Promise<boolean> {\n const connection = this._connections.get(peerId);\n if (!connection) {\n return false;\n }\n\n return connection.connection.shouldAdvertize(params);\n }\n\n private _onConnectionOpen(connection: ReplicatorConnection) {\n invariant(!this._connections.has(connection.peerId as PeerId));\n const reader = connection.readable.getReader();\n const writer = connection.writable.getWriter();\n const connectionEntry: ConnectionEntry = { connection, reader, writer, isOpen: true };\n this._connections.set(connection.peerId as PeerId, connectionEntry);\n\n queueMicrotask(async () => {\n try {\n while (true) {\n // TODO(dmaretskyi): Find a way to enforce backpressure on AM-repo.\n const { done, value } = await reader.read();\n if (done) {\n break;\n }\n\n this.emit('message', value);\n }\n } catch (err) {\n if (connectionEntry.isOpen) {\n log.catch(err);\n }\n }\n });\n\n this.emit('peer-candidate', {\n peerId: connection.peerId as PeerId,\n peerMetadata: {\n // TODO(dmaretskyi): Refactor this.\n dxos_peerSource: 'EchoNetworkAdapter',\n } as any,\n });\n }\n\n private _onConnectionClosed(connection: ReplicatorConnection) {\n const entry = this._connections.get(connection.peerId as PeerId);\n invariant(entry);\n\n entry.isOpen = false;\n this.emit('peer-disconnected', { peerId: connection.peerId as PeerId });\n\n void entry.reader.cancel().catch((err) => log.catch(err));\n void entry.writer.abort().catch((err) => log.catch(err));\n\n this._connections.delete(connection.peerId as PeerId);\n }\n}\n\ntype ConnectionEntry = {\n connection: ReplicatorConnection;\n reader: ReadableStreamDefaultReader<Message>;\n writer: WritableStreamDefaultWriter<Message>;\n isOpen: boolean;\n};\n", "//\n// Copyright 2024 DXOS.org\n// s\n\nimport { type MixedEncoding } from 'level-transcoder';\n\nimport { type StorageAdapterInterface, type Chunk, type StorageKey } from '@dxos/automerge/automerge-repo';\nimport { LifecycleState, Resource } from '@dxos/context';\nimport { type BatchLevel, type SubLevelDB } from '@dxos/kv-store';\nimport { type MaybePromise } from '@dxos/util';\n\nexport type LevelDBStorageAdapterParams = {\n db: SubLevelDB;\n callbacks?: StorageCallbacks;\n};\n\nexport type BeforeSaveParams = { path: StorageKey; batch: BatchLevel };\n\nexport interface StorageCallbacks {\n beforeSave(params: BeforeSaveParams): MaybePromise<void>;\n afterSave(path: StorageKey): MaybePromise<void>;\n}\n\nexport class LevelDBStorageAdapter extends Resource implements StorageAdapterInterface {\n constructor(private readonly _params: LevelDBStorageAdapterParams) {\n super();\n }\n\n async load(keyArray: StorageKey): Promise<Uint8Array | undefined> {\n try {\n if (this._lifecycleState !== LifecycleState.OPEN) {\n // TODO(mykola): this should be an error.\n return undefined;\n }\n return await this._params.db.get<StorageKey, Uint8Array>(keyArray, { ...encodingOptions });\n } catch (err: any) {\n if (isLevelDbNotFoundError(err)) {\n return undefined;\n }\n throw err;\n }\n }\n\n async save(keyArray: StorageKey, binary: Uint8Array): Promise<void> {\n if (this._lifecycleState !== LifecycleState.OPEN) {\n return undefined;\n }\n const batch = this._params.db.batch();\n\n await this._params.callbacks?.beforeSave?.({ path: keyArray, batch });\n batch.put<StorageKey, Uint8Array>(keyArray, Buffer.from(binary), {\n ...encodingOptions,\n });\n await batch.write();\n\n await this._params.callbacks?.afterSave?.(keyArray);\n }\n\n async remove(keyArray: StorageKey): Promise<void> {\n if (this._lifecycleState !== LifecycleState.OPEN) {\n return undefined;\n }\n await this._params.db.del<StorageKey>(keyArray, { ...encodingOptions });\n }\n\n async loadRange(keyPrefix: StorageKey): Promise<Chunk[]> {\n if (this._lifecycleState !== LifecycleState.OPEN) {\n return [];\n }\n const result: Chunk[] = [];\n for await (const [key, value] of this._params.db.iterator<StorageKey, Uint8Array>({\n gte: keyPrefix,\n lte: [...keyPrefix, '\\uffff'],\n ...encodingOptions,\n })) {\n result.push({\n key,\n data: value,\n });\n }\n return result;\n }\n\n async removeRange(keyPrefix: StorageKey): Promise<void> {\n if (this._lifecycleState !== LifecycleState.OPEN) {\n return undefined;\n }\n const batch = this._params.db.batch();\n\n for await (const [key] of this._params.db.iterator<StorageKey, Uint8Array>({\n gte: keyPrefix,\n lte: [...keyPrefix, '\\uffff'],\n ...encodingOptions,\n })) {\n batch.del<StorageKey>(key, { ...encodingOptions });\n }\n await batch.write();\n }\n}\n\nconst keyEncoder: MixedEncoding<StorageKey, Uint8Array, StorageKey> = {\n encode: (key: StorageKey): Uint8Array =>\n Buffer.from(key.map((k) => k.replaceAll('%', '%25').replaceAll('-', '%2D')).join('-')),\n decode: (key: Uint8Array): StorageKey =>\n Buffer.from(key)\n .toString()\n .split('-')\n .map((k) => k.replaceAll('%2D', '-').replaceAll('%25', '%')),\n format: 'buffer',\n};\n\nexport const encodingOptions = {\n keyEncoding: keyEncoder,\n valueEncoding: 'buffer',\n};\n\nconst isLevelDbNotFoundError = (err: any): boolean => err.code === 'LEVEL_NOT_FOUND';\n", "//\n// Copyright 2024 DXOS.org\n//\n\nimport { Trigger } from '@dxos/async';\nimport { NetworkAdapter, type Message, type PeerId, cbor } from '@dxos/automerge/automerge-repo';\nimport { Stream } from '@dxos/codec-protobuf';\nimport { invariant } from '@dxos/invariant';\nimport { type HostInfo, type SyncRepoRequest, type SyncRepoResponse } from '@dxos/protocols/proto/dxos/echo/service';\n\ntype ClientSyncState = {\n connected: boolean;\n send: (message: Message) => void;\n disconnect: () => void;\n};\n\n/**\n * Used to replicate with apps running on the same device.\n */\nexport class LocalHostNetworkAdapter extends NetworkAdapter {\n private readonly _peers: Map<PeerId, ClientSyncState> = new Map();\n\n /**\n * Emits `ready` event. That signals to `Repo` that it can start using the adapter.\n */\n ready() {\n // NOTE: Emitting `ready` event in NetworkAdapter`s constructor causes a race condition\n // because `Repo` waits for `ready` event (which it never receives) before it starts using the adapter.\n this.emit('ready', {\n network: this,\n });\n }\n\n private readonly _connected = new Trigger();\n private _isConnected: boolean = false;\n\n /**\n * Called by `Repo` to connect to the network.\n *\n * @param peerId Our peer Id.\n */\n override connect(peerId: PeerId): void {\n this.peerId = peerId;\n this._isConnected = true;\n this._connected.wake();\n // No-op. Client always connects first\n }\n\n override send(message: Message): void {\n const peer = this._peers.get(message.targetId);\n invariant(peer, 'Peer not found.');\n peer.send(message);\n }\n\n async close() {\n this._peers.forEach((peer) => peer.disconnect());\n this.emit('close');\n }\n\n override disconnect(): void {\n // TODO(mykola): `disconnect` is not used anywhere in `Repo` from `@automerge/automerge-repo`. Should we remove it?\n // No-op\n }\n\n async whenConnected(): Promise<void> {\n await this._connected.wait({ timeout: 10_000 });\n }\n\n syncRepo({ id, syncMessage }: SyncRepoRequest): Stream<SyncRepoResponse> {\n const peerId = this._getPeerId(id);\n\n return new Stream(({ next, close }) => {\n invariant(!this._peers.has(peerId), 'Peer already connected.');\n this._peers.set(peerId, {\n connected: true,\n send: (message) => {\n next({\n syncMessage: cbor.encode(message),\n });\n },\n disconnect: () => {\n this._peers.delete(peerId);\n close();\n this.emit('peer-disconnected', {\n peerId,\n });\n },\n });\n\n invariant(this._isConnected);\n this.emit('peer-candidate', {\n peerMetadata: {},\n peerId,\n });\n });\n }\n\n async sendSyncMessage({ id, syncMessage }: SyncRepoRequest): Promise<void> {\n invariant(this._isConnected);\n const message = cbor.decode(syncMessage!) as Message;\n this.emit('message', message);\n }\n\n async getHostInfo(): Promise<HostInfo> {\n invariant(this._isConnected);\n invariant(this.peerId, 'Peer id not set.');\n return {\n peerId: this.peerId,\n };\n }\n\n private _getPeerId(id: string): PeerId {\n return id as PeerId;\n }\n}\n", "//\n// Copyright 2024 DXOS.org\n//\n\nimport { Trigger } from '@dxos/async';\nimport { NetworkAdapter, type Message, type PeerId, cbor } from '@dxos/automerge/automerge-repo';\nimport { invariant } from '@dxos/invariant';\nimport { log } from '@dxos/log';\nimport { type PeerInfo } from '@dxos/protocols/proto/dxos/mesh/teleport/automerge';\nimport { AutomergeReplicator } from '@dxos/teleport-extension-automerge-replicator';\n\n/**\n * Used to replicate with other peers over the network.\n */\nexport class MeshNetworkAdapter extends NetworkAdapter {\n private readonly _extensions: Map<string, AutomergeReplicator> = new Map();\n private _connected = new Trigger();\n\n /**\n * Emits `ready` event. That signals to `Repo` that it can start using the adapter.\n */\n ready() {\n // NOTE: Emitting `ready` event in NetworkAdapter`s constructor causes a race condition\n // because `Repo` waits for `ready` event (which it never receives) before it starts using the adapter.\n this.emit('ready', {\n network: this,\n });\n }\n\n override connect(peerId: PeerId): void {\n this.peerId = peerId;\n this._connected.wake();\n }\n\n override send(message: Message): void {\n const receiverId = message.targetId;\n const extension = this._extensions.get(receiverId);\n invariant(extension, 'Extension not found.');\n extension.sendSyncMessage({ payload: cbor.encode(message) }).catch((err) => log.catch(err));\n }\n\n override disconnect(): void {\n // No-op\n }\n\n createExtension(): AutomergeReplicator {\n invariant(this.peerId, 'Peer id not set.');\n\n let peerInfo: PeerInfo;\n const extension = new AutomergeReplicator(\n {\n peerId: this.peerId,\n },\n {\n onStartReplication: async (info, remotePeerId /** Teleport ID */) => {\n await this._connected.wait();\n\n // Note: We store only one extension per peer.\n // There can be a case where two connected peers have more than one teleport connection between them\n // and each of them uses different teleport connections to send messages.\n // It works because we receive messages from all teleport connections and Automerge Repo dedup them.\n // TODO(mykola): Use only one teleport connection per peer.\n\n // TODO(dmaretskyi): Critical bug.\n // - two peers get connected via swarm 1\n // - they get connected via swarm 2\n // - swarm 1 gets disconnected\n // - automerge repo thinks that peer 2 got disconnected even though swarm 2 is still active\n\n log('onStartReplication', { id: info.id, thisPeerId: this.peerId, remotePeerId: remotePeerId.toHex() });\n if (!this._extensions.has(info.id)) {\n peerInfo = info;\n // TODO(mykola): Fix race condition?\n this._extensions.set(info.id, extension);\n\n log('peer-candidate', { id: info.id, thisPeerId: this.peerId, remotePeerId: remotePeerId.toHex() });\n this.emit('peer-candidate', {\n // TODO(mykola): Hack, stop abusing `peerMetadata` field.\n peerMetadata: {\n dxos_deviceKey: remotePeerId.toHex(),\n } as any,\n peerId: info.id as PeerId,\n });\n }\n },\n onSyncMessage: async ({ payload }) => {\n if (!peerInfo) {\n return;\n }\n const message = cbor.decode(payload) as Message;\n // Note: automerge Repo dedup messages.\n this.emit('message', message);\n },\n onClose: async () => {\n if (!peerInfo) {\n return;\n }\n this.emit('peer-disconnected', {\n peerId: peerInfo.id as PeerId,\n });\n this._extensions.delete(peerInfo.id);\n },\n },\n );\n return extension;\n }\n}\n", "//\n// Copyright 2024 DXOS.org\n//\n\nimport { type StorageKey } from '@dxos/automerge/automerge-repo';\nimport { IndexedDBStorageAdapter } from '@dxos/automerge/automerge-repo-storage-indexeddb';\nimport { type SubLevelDB } from '@dxos/kv-store';\nimport { log } from '@dxos/log';\nimport { StorageType, type Directory } from '@dxos/random-access-storage';\n\nimport { AutomergeStorageAdapter } from './automerge-storage-adapter';\nimport { encodingOptions } from './leveldb-storage-adapter';\n\nexport const levelMigration = async ({ db, directory }: { db: SubLevelDB; directory: Directory }) => {\n // Note: Make automigration from previous storage to leveldb here.\n const isNewLevel = !(await db\n .iterator<StorageKey, Uint8Array>({\n ...encodingOptions,\n })\n .next());\n\n if (!isNewLevel) {\n return;\n }\n\n const oldStorageAdapter =\n directory.type === StorageType.IDB\n ? new IndexedDBStorageAdapter(directory.path, 'data')\n : new AutomergeStorageAdapter(directory);\n\n const chunks = await oldStorageAdapter.loadRange([]);\n if (chunks.length === 0) {\n return;\n }\n const batch = db.batch();\n log.info('found chunks on old storage adapter', { chunks: chunks.length });\n for (const { key, data } of await oldStorageAdapter.loadRange([])) {\n data && batch.put<StorageKey, Uint8Array>(key, data, { ...encodingOptions });\n }\n await batch.write();\n};\n", "//\n// Copyright 2024 DXOS.org\n//\n//\n// Copyright 2023 DXOS.org\n//\n\nimport { type Chunk, type StorageKey, type StorageAdapterInterface } from '@dxos/automerge/automerge-repo';\nimport { type Directory } from '@dxos/random-access-storage';\nimport { arrayToBuffer, bufferToArray } from '@dxos/util';\n\nexport class AutomergeStorageAdapter implements StorageAdapterInterface {\n // TODO(mykola): Hack for restricting automerge Repo to access storage if Host is `closed`.\n // Automerge Repo do not have any lifetime management.\n private _state: 'opened' | 'closed' = 'opened';\n\n constructor(private readonly _directory: Directory) {}\n\n async load(key: StorageKey): Promise<Uint8Array | undefined> {\n if (this._state !== 'opened') {\n return undefined;\n }\n const filename = this._getFilename(key);\n const file = this._directory.getOrCreateFile(filename);\n const { size } = await file.stat();\n if (!size || size === 0) {\n return undefined;\n }\n const buffer = await file.read(0, size);\n return bufferToArray(buffer);\n }\n\n async save(key: StorageKey, data: Uint8Array): Promise<void> {\n if (this._state !== 'opened') {\n return undefined;\n }\n const filename = this._getFilename(key);\n const file = this._directory.getOrCreateFile(filename);\n await file.write(0, arrayToBuffer(data));\n await file.truncate?.(data.length);\n\n await file.flush?.();\n }\n\n async remove(key: StorageKey): Promise<void> {\n if (this._state !== 'opened') {\n return undefined;\n }\n // TODO(dmaretskyi): Better deletion.\n const filename = this._getFilename(key);\n const file = this._directory.getOrCreateFile(filename);\n await file.destroy();\n }\n\n async loadRange(keyPrefix: StorageKey): Promise<Chunk[]> {\n if (this._state !== 'opened') {\n return [];\n }\n const filename = this._getFilename(keyPrefix);\n const entries = await this._directory.list();\n return Promise.all(\n entries\n .filter((entry) => entry.startsWith(filename))\n .map(async (entry): Promise<Chunk> => {\n const file = this._directory.getOrCreateFile(entry);\n const { size } = await file.stat();\n const buffer = await file.read(0, size);\n return {\n key: this._getKeyFromFilename(entry),\n data: bufferToArray(buffer),\n };\n }),\n );\n }\n\n async removeRange(keyPrefix: StorageKey): Promise<void> {\n if (this._state !== 'opened') {\n return undefined;\n }\n const filename = this._getFilename(keyPrefix);\n const entries = await this._directory.list();\n await Promise.all(\n entries\n .filter((entry) => entry.startsWith(filename))\n .map(async (entry): Promise<void> => {\n const file = this._directory.getOrCreateFile(entry);\n await file.destroy();\n }),\n );\n }\n\n async close(): Promise<void> {\n this._state = 'closed';\n }\n\n private _getFilename(key: StorageKey): string {\n return key.map((k) => k.replaceAll('%', '%25').replaceAll('-', '%2D')).join('-');\n }\n\n private _getKeyFromFilename(filename: string): StorageKey {\n return filename.split('-').map((k) => k.replaceAll('%2D', '-').replaceAll('%25', '%'));\n }\n}\n", "//\n// Copyright 2024 DXOS.org\n//\n\nimport { Event } from '@dxos/async';\nimport { type DocHandle, type AutomergeUrl, type DocumentId, type Repo } from '@dxos/automerge/automerge-repo';\nimport { cancelWithContext, type Context } from '@dxos/context';\nimport { warnAfterTimeout } from '@dxos/debug';\nimport { type SpaceState, type SpaceDoc } from '@dxos/echo-protocol';\nimport { invariant } from '@dxos/invariant';\nimport { type PublicKey } from '@dxos/keys';\nimport { log } from '@dxos/log';\nimport { trace } from '@dxos/tracing';\n\ntype SpaceDocumentLinks = SpaceDoc['links'];\n\nexport interface AutomergeDocumentLoader {\n onObjectDocumentLoaded: Event<ObjectDocumentLoaded>;\n\n getAllHandles(): DocHandle<SpaceDoc>[];\n\n loadSpaceRootDocHandle(ctx: Context, spaceState: SpaceState): Promise<void>;\n loadObjectDocument(objectId: string | string[]): void;\n getSpaceRootDocHandle(): DocHandle<SpaceDoc>;\n createDocumentForObject(objectId: string): DocHandle<SpaceDoc>;\n onObjectLinksUpdated(links: SpaceDocumentLinks): void;\n onObjectBoundToDocument(handle: DocHandle<SpaceDoc>, objectId: string): void;\n\n /**\n * @returns objectIds for which we had document handles or were loading one.\n */\n clearHandleReferences(): string[];\n}\n\n/**\n * Manages object <-> docHandle binding and automerge document loading.\n */\n@trace.resource()\nexport class AutomergeDocumentLoaderImpl implements AutomergeDocumentLoader {\n private _spaceRootDocHandle: DocHandle<SpaceDoc> | null = null;\n /**\n * An object id pointer to a handle of the document where the object is stored inline.\n */\n private readonly _objectDocumentHandles = new Map<string, DocHandle<SpaceDoc>>();\n /**\n * If object was requested via loadObjectDocument but root document links weren't updated yet\n * loading will be triggered in onObjectLinksUpdated callback.\n */\n private readonly _objectsPendingDocumentLoad = new Set<string>();\n\n public readonly onObjectDocumentLoaded = new Event<ObjectDocumentLoaded>();\n\n constructor(\n private readonly _spaceKey: PublicKey,\n private readonly _repo: Repo,\n ) {}\n\n getAllHandles(): DocHandle<SpaceDoc>[] {\n return this._spaceRootDocHandle != null\n ? [this._spaceRootDocHandle, ...new Set(this._objectDocumentHandles.values())]\n : [];\n }\n\n @trace.span({ showInBrowserTimeline: true })\n public async loadSpaceRootDocHandle(ctx: Context, spaceState: SpaceState): Promise<void> {\n if (this._spaceRootDocHandle != null) {\n return;\n }\n if (!spaceState.rootUrl) {\n log.error('Database opened with no rootUrl', { spaceKey: this._spaceKey });\n this._createContextBoundSpaceRootDocument(ctx);\n } else {\n const existingDocHandle = await this._initDocHandle(ctx, spaceState.rootUrl);\n const doc = existingDocHandle.docSync();\n invariant(doc);\n if (doc.access == null) {\n this._initDocAccess(existingDocHandle);\n }\n this._spaceRootDocHandle = existingDocHandle;\n }\n }\n\n public loadObjectDocument(objectIdOrMany: string | string[]) {\n const objectIds = Array.isArray(objectIdOrMany) ? objectIdOrMany : [objectIdOrMany];\n let hasUrlsToLoad = false;\n const urlsToLoad: SpaceDoc['links'] = {};\n for (const objectId of objectIds) {\n invariant(this._spaceRootDocHandle);\n if (this._objectDocumentHandles.has(objectId) || this._objectsPendingDocumentLoad.has(objectId)) {\n continue;\n }\n const spaceRootDoc = this._spaceRootDocHandle.docSync();\n invariant(spaceRootDoc);\n const documentUrl = (spaceRootDoc.links ?? {})[objectId];\n if (documentUrl == null) {\n this._objectsPendingDocumentLoad.add(objectId);\n log.info('loading delayed until object links are initialized', { objectId });\n } else {\n urlsToLoad[objectId] = documentUrl;\n hasUrlsToLoad = true;\n }\n }\n if (hasUrlsToLoad) {\n this._loadLinkedObjects(urlsToLoad);\n }\n }\n\n public onObjectLinksUpdated(links: SpaceDocumentLinks) {\n if (!links) {\n return;\n }\n const linksAwaitingLoad = Object.entries(links).filter(([objectId]) =>\n this._objectsPendingDocumentLoad.has(objectId),\n );\n this._loadLinkedObjects(Object.fromEntries(linksAwaitingLoad));\n linksAwaitingLoad.forEach(([objectId]) => this._objectsPendingDocumentLoad.delete(objectId));\n }\n\n public getSpaceRootDocHandle(): DocHandle<SpaceDoc> {\n invariant(this._spaceRootDocHandle);\n return this._spaceRootDocHandle;\n }\n\n public createDocumentForObject(objectId: string): DocHandle<SpaceDoc> {\n invariant(this._spaceRootDocHandle);\n const spaceDocHandle = this._repo.create<SpaceDoc>();\n this._initDocAccess(spaceDocHandle);\n this.onObjectBoundToDocument(spaceDocHandle, objectId);\n this._spaceRootDocHandle.change((newDoc: SpaceDoc) => {\n newDoc.links ??= {};\n newDoc.links[objectId] = spaceDocHandle.url;\n });\n return spaceDocHandle;\n }\n\n public onObjectBoundToDocument(handle: DocHandle<SpaceDoc>, objectId: string) {\n this._objectDocumentHandles.set(objectId, handle);\n }\n\n public clearHandleReferences(): string[] {\n const objectsWithHandles = [...this._objectDocumentHandles.keys()];\n this._objectDocumentHandles.clear();\n this._spaceRootDocHandle = null;\n return objectsWithHandles;\n }\n\n private _loadLinkedObjects(links: SpaceDocumentLinks) {\n if (!links) {\n return;\n }\n for (const [objectId, automergeUrl] of Object.entries(links)) {\n const logMeta = { objectId, automergeUrl };\n const objectDocumentHandle = this._objectDocumentHandles.get(objectId);\n if (objectDocumentHandle != null && objectDocumentHandle.url !== automergeUrl) {\n log.warn('object already inlined in a different document, ignoring the link', {\n ...logMeta,\n actualDocumentUrl: objectDocumentHandle.url,\n });\n continue;\n }\n if (objectDocumentHandle?.url === automergeUrl) {\n log.warn('object document was already loaded', logMeta);\n continue;\n }\n const handle = this._repo.find<SpaceDoc>(automergeUrl as DocumentId);\n log.debug('document loading triggered', logMeta);\n this._objectDocumentHandles.set(objectId, handle);\n void this._createObjectOnDocumentLoad(handle, objectId);\n }\n }\n\n private async _initDocHandle(ctx: Context, url: string) {\n const docHandle = this._repo.find<SpaceDoc>(url as DocumentId);\n while (true) {\n try {\n await warnAfterTimeout(5_000, 'Automerge root doc load timeout (AutomergeDb)', async () => {\n await cancelWithContext(ctx, docHandle.whenReady()); // TODO(dmaretskyi): Temporary 5s timeout for debugging.\n });\n break;\n } catch (err) {\n if (`${err}`.includes('Timeout')) {\n log.info('wraparound', { id: docHandle.documentId, state: docHandle.state });\n continue;\n }\n\n throw err;\n }\n }\n\n if (docHandle.state === 'unavailable') {\n throw new Error('Automerge document is unavailable');\n }\n\n return docHandle;\n }\n\n private _createContextBoundSpaceRootDocument(ctx: Context) {\n const docHandle = this._repo.create<SpaceDoc>();\n this._spaceRootDocHandle = docHandle;\n ctx.onDispose(() => {\n docHandle.delete();\n this._spaceRootDocHandle = null;\n });\n }\n\n private _initDocAccess(handle: DocHandle<SpaceDoc>) {\n handle.change((newDoc: SpaceDoc) => {\n newDoc.access ??= { spaceKey: this._spaceKey.toHex() };\n newDoc.access.spaceKey = this._spaceKey.toHex();\n });\n }\n\n private async _createObjectOnDocumentLoad(handle: DocHandle<SpaceDoc>, objectId: string) {\n try {\n await handle.doc(['ready']);\n const logMeta = { objectId, docUrl: handle.url };\n if (this.onObjectDocumentLoaded.listenerCount() === 0) {\n log.info('document loaded after all listeners were removed', logMeta);\n return;\n }\n const objectDocHandle = this._objectDocumentHandles.get(objectId);\n if (objectDocHandle?.url !== handle.url) {\n log.warn('object was rebound while a document was loading, discarding handle', logMeta);\n return;\n }\n this.onObjectDocumentLoaded.emit({ handle, objectId });\n } catch (err) {\n const shouldRetryLoading = this.onObjectDocumentLoaded.listenerCount() > 0;\n log.warn('failed to load a document', {\n objectId,\n automergeUrl: handle.url,\n retryLoading: shouldRetryLoading,\n err,\n });\n if (shouldRetryLoading) {\n await this._createObjectOnDocumentLoad(handle, objectId);\n }\n }\n }\n}\n\nexport interface ObjectDocumentLoaded {\n handle: DocHandle<SpaceDoc>;\n objectId: string;\n}\n\nexport interface DocumentChanges {\n createdObjectIds: string[];\n updatedObjectIds: string[];\n objectsToRebind: string[];\n linkedDocuments: {\n [echoId: string]: AutomergeUrl;\n };\n}\n"],
|
|
4
|
+
"sourcesContent": ["//\n// Copyright 2023 DXOS.org\n//\n\nimport { Event } from '@dxos/async';\nimport { type Doc, next as automerge, getBackend, type Heads, getHeads } from '@dxos/automerge/automerge';\nimport {\n type DocHandle,\n Repo,\n type DocumentId,\n type PeerId,\n type StorageAdapterInterface,\n type DocHandleChangePayload,\n} from '@dxos/automerge/automerge-repo';\nimport { type Stream } from '@dxos/codec-protobuf';\nimport { Context, type Lifecycle } from '@dxos/context';\nimport { type SpaceDoc } from '@dxos/echo-protocol';\nimport { type IndexMetadataStore } from '@dxos/indexing';\nimport { invariant } from '@dxos/invariant';\nimport { PublicKey } from '@dxos/keys';\nimport { type SublevelDB } from '@dxos/kv-store';\nimport { log } from '@dxos/log';\nimport { idCodec } from '@dxos/protocols';\nimport {\n type FlushRequest,\n type HostInfo,\n type SyncRepoRequest,\n type SyncRepoResponse,\n} from '@dxos/protocols/proto/dxos/echo/service';\nimport { type Directory } from '@dxos/random-access-storage';\nimport { type AutomergeReplicator } from '@dxos/teleport-extension-automerge-replicator';\nimport { trace } from '@dxos/tracing';\nimport { ComplexMap, ComplexSet, defaultMap, mapValues } from '@dxos/util';\n\nimport { EchoNetworkAdapter } from './echo-network-adapter';\nimport { type EchoReplicator } from './echo-replicator';\nimport { type BeforeSaveParams, LevelDBStorageAdapter } from './leveldb-storage-adapter';\nimport { LocalHostNetworkAdapter } from './local-host-network-adapter';\nimport { MeshNetworkAdapter } from './mesh-network-adapter';\nimport { levelMigration } from './migrations';\n\n// TODO: Remove\nexport type { DocumentId };\n\nexport type AutomergeHostParams = {\n db: SublevelDB;\n /**\n * For migration purposes.\n */\n directory?: Directory;\n\n indexMetadataStore: IndexMetadataStore;\n};\n\n@trace.resource()\nexport class AutomergeHost {\n private readonly _indexMetadataStore: IndexMetadataStore;\n private readonly _ctx = new Context();\n private readonly _directory?: Directory;\n private readonly _db: SublevelDB;\n private readonly _echoNetworkAdapter = new EchoNetworkAdapter();\n\n private _repo!: Repo;\n private _meshNetwork!: MeshNetworkAdapter;\n private _clientNetwork!: LocalHostNetworkAdapter;\n private _storage!: StorageAdapterInterface & Lifecycle;\n\n @trace.info()\n private _peerId!: string;\n\n /**\n * spaceKey -> deviceKey[]\n */\n private readonly _authorizedDevices = new ComplexMap<PublicKey, ComplexSet<PublicKey>>(PublicKey.hash);\n\n public _requestedDocs = new Set<string>();\n\n constructor({ directory, db, indexMetadataStore }: AutomergeHostParams) {\n this._directory = directory;\n this._db = db;\n this._indexMetadataStore = indexMetadataStore;\n }\n\n async open() {\n // TODO(mykola): remove this before 0.6 release.\n this._directory && (await levelMigration({ db: this._db, directory: this._directory }));\n this._storage = new LevelDBStorageAdapter({\n db: this._db,\n callbacks: {\n beforeSave: async (params) => this._beforeSave(params),\n afterSave: async () => this._afterSave(),\n },\n });\n await this._storage.open?.();\n this._peerId = `host-${PublicKey.random().toHex()}` as PeerId;\n\n this._meshNetwork = new MeshNetworkAdapter();\n this._clientNetwork = new LocalHostNetworkAdapter();\n\n this._repo = new Repo({\n peerId: this._peerId as PeerId,\n network: [this._clientNetwork, this._meshNetwork, this._echoNetworkAdapter],\n storage: this._storage,\n\n // TODO(dmaretskyi): Share based on HALO permissions and space affinity.\n // Hosts, running in the worker, don't share documents unless requested by other peers.\n sharePolicy: async (peerId /* device key */, documentId /* space key */) => {\n if (peerId.startsWith('client-')) {\n return false; // Only send docs to clients if they are requested.\n }\n\n if (!documentId) {\n return false;\n }\n\n const peerMetadata = this.repo.peerMetadataByPeerId[peerId];\n if ((peerMetadata as any)?.dxos_peerSource === 'EchoNetworkAdapter') {\n return this._echoNetworkAdapter.shouldAdvertize(peerId, { documentId });\n }\n\n const doc = this._repo.handles[documentId]?.docSync();\n if (!doc) {\n const isRequested = this._requestedDocs.has(`automerge:${documentId}`);\n log('doc share policy check', { peerId, documentId, isRequested });\n return isRequested;\n }\n\n try {\n const spaceKey = getSpaceKeyFromDoc(doc);\n if (!spaceKey) {\n log('space key not found for share policy check', { peerId, documentId });\n return false;\n }\n\n const authorizedDevices = this._authorizedDevices.get(PublicKey.from(spaceKey));\n\n // TODO(mykola): Hack, stop abusing `peerMetadata` field.\n const deviceKeyHex = (peerMetadata as any)?.dxos_deviceKey;\n if (!deviceKeyHex) {\n log('device key not found for share policy check', { peerId, documentId });\n return false;\n }\n const deviceKey = PublicKey.from(deviceKeyHex);\n\n const isAuthorized = authorizedDevices?.has(deviceKey) ?? false;\n log('share policy check', {\n localPeer: this._peerId,\n remotePeer: peerId,\n documentId,\n deviceKey,\n spaceKey,\n isAuthorized,\n });\n return isAuthorized;\n } catch (err) {\n log.catch(err);\n return false;\n }\n },\n });\n this._clientNetwork.ready();\n this._meshNetwork.ready();\n await this._echoNetworkAdapter.open();\n\n await this._clientNetwork.whenConnected();\n await this._echoNetworkAdapter.whenConnected();\n }\n\n async close() {\n await this._storage.close?.();\n await this._clientNetwork.close();\n await this._echoNetworkAdapter.close();\n await this._ctx.dispose();\n }\n\n get repo(): Repo {\n return this._repo;\n }\n\n async addReplicator(replicator: EchoReplicator) {\n await this._echoNetworkAdapter.addReplicator(replicator);\n }\n\n async removeReplicator(replicator: EchoReplicator) {\n await this._echoNetworkAdapter.removeReplicator(replicator);\n }\n\n private async _beforeSave({ path, batch }: BeforeSaveParams) {\n const handle = this._repo.handles[path[0] as DocumentId];\n if (!handle) {\n return;\n }\n const doc = handle.docSync();\n if (!doc) {\n return;\n }\n\n const lastAvailableHash = getHeads(doc);\n\n const objectIds = Object.keys(doc.objects ?? {});\n const encodedIds = objectIds.map((objectId) => idCodec.encode({ documentId: handle.documentId, objectId }));\n const idToLastHash = new Map(encodedIds.map((id) => [id, lastAvailableHash]));\n this._indexMetadataStore.markDirty(idToLastHash, batch);\n }\n\n /**\n * Called by AutomergeStorageAdapter after levelDB batch commit.\n */\n private async _afterSave() {\n this._indexMetadataStore.notifyMarkedDirty();\n }\n\n @trace.info({ depth: null })\n private _automergeDocs() {\n return mapValues(this._repo.handles, (handle) => ({\n state: handle.state,\n hasDoc: !!handle.docSync(),\n heads: handle.docSync() ? automerge.getHeads(handle.docSync()) : null,\n data:\n handle.docSync() &&\n mapValues(handle.docSync(), (value, key) => {\n try {\n switch (key) {\n case 'access':\n case 'links':\n return value;\n case 'objects':\n return Object.keys(value as any);\n default:\n return `${value}`;\n }\n } catch (err) {\n return `${err}`;\n }\n }),\n }));\n }\n\n @trace.info({ depth: null })\n private _automergePeers() {\n return this._repo.peers;\n }\n\n //\n // Methods for client-services.\n //\n @trace.span({ showInBrowserTimeline: true })\n async flush({ states }: FlushRequest): Promise<void> {\n // Note: Wait for all requested documents to be loaded/synced from thin-client.\n await Promise.all(\n states?.map(async ({ heads, documentId }) => {\n invariant(heads, 'heads are required for flush');\n const handle = this.repo.handles[documentId as DocumentId] ?? this._repo.find(documentId as DocumentId);\n await waitForHeads(handle, heads);\n }) ?? [],\n );\n\n await this._repo.flush(states?.map(({ documentId }) => documentId as DocumentId));\n }\n\n syncRepo(request: SyncRepoRequest): Stream<SyncRepoResponse> {\n return this._clientNetwork.syncRepo(request);\n }\n\n sendSyncMessage(request: SyncRepoRequest): Promise<void> {\n return this._clientNetwork.sendSyncMessage(request);\n }\n\n async getHostInfo(): Promise<HostInfo> {\n return this._clientNetwork.getHostInfo();\n }\n\n //\n // Mesh replication.\n //\n\n createExtension(): AutomergeReplicator {\n return this._meshNetwork.createExtension();\n }\n\n authorizeDevice(spaceKey: PublicKey, deviceKey: PublicKey) {\n log('authorizeDevice', { spaceKey, deviceKey });\n defaultMap(this._authorizedDevices, spaceKey, () => new ComplexSet(PublicKey.hash)).add(deviceKey);\n }\n}\n\nexport const getSpaceKeyFromDoc = (doc: any): string | null => {\n // experimental_spaceKey is set on old documents, new ones are created with doc.access.spaceKey\n const rawSpaceKey = doc.access?.spaceKey ?? doc.experimental_spaceKey;\n if (rawSpaceKey == null) {\n return null;\n }\n\n return String(rawSpaceKey);\n};\n\nconst waitForHeads = async (handle: DocHandle<SpaceDoc>, heads: Heads) => {\n await handle.whenReady();\n const unavailableHeads = new Set(heads);\n\n await Event.wrap<DocHandleChangePayload<SpaceDoc>>(handle, 'change').waitForCondition(() => {\n // Check if unavailable heads became available.\n for (const changeHash of unavailableHeads.values()) {\n if (changeIsPresentInDoc(handle.docSync(), changeHash)) {\n unavailableHeads.delete(changeHash);\n }\n }\n\n if (unavailableHeads.size === 0) {\n return true;\n }\n return false;\n });\n};\n\nconst changeIsPresentInDoc = (doc: Doc<any>, changeHash: string): boolean => {\n return !!getBackend(doc).getChangeByHash(changeHash);\n};\n", "//\n// Copyright 2024 DXOS.org\n//\n\nimport { Trigger, synchronized } from '@dxos/async';\nimport { type Message, NetworkAdapter, type PeerId, type PeerMetadata } from '@dxos/automerge/automerge-repo';\nimport { LifecycleState } from '@dxos/context';\nimport { invariant } from '@dxos/invariant';\nimport { log } from '@dxos/log';\n\nimport { type EchoReplicator, type ReplicatorConnection, type ShouldAdvertizeParams } from './echo-replicator';\n\n/**\n * Manages a set of {@link EchoReplicator} instances.\n */\nexport class EchoNetworkAdapter extends NetworkAdapter {\n private readonly _replicators = new Set<EchoReplicator>();\n /**\n * Remote peer id -> connection.\n */\n private readonly _connections = new Map<PeerId, ConnectionEntry>();\n private _lifecycleState: LifecycleState = LifecycleState.CLOSED;\n private readonly _connected = new Trigger();\n\n override connect(peerId: PeerId, peerMetadata?: PeerMetadata | undefined): void {\n this.peerId = peerId;\n this.peerMetadata = peerMetadata;\n this._connected.wake();\n }\n\n override send(message: Message): void {\n const connectionEntry = this._connections.get(message.targetId);\n if (!connectionEntry) {\n throw new Error('Connection not found.');\n }\n\n // TODO(dmaretskyi): Find a way to enforce backpressure on AM-repo.\n connectionEntry.writer.write(message).catch((err) => {\n if (connectionEntry.isOpen) {\n log.catch(err);\n }\n });\n }\n\n override disconnect(): void {\n // No-op\n }\n\n @synchronized\n async open() {\n invariant(this._lifecycleState === LifecycleState.CLOSED);\n this._lifecycleState = LifecycleState.OPEN;\n\n this.emit('ready', {\n network: this,\n });\n }\n\n @synchronized\n async close() {\n invariant(this._lifecycleState === LifecycleState.OPEN);\n\n for (const replicator of this._replicators) {\n await replicator.disconnect();\n }\n this._replicators.clear();\n\n this._lifecycleState = LifecycleState.CLOSED;\n }\n\n async whenConnected() {\n await this._connected.wait({ timeout: 10_000 });\n }\n\n @synchronized\n async addReplicator(replicator: EchoReplicator) {\n invariant(this.peerId);\n invariant(!this._replicators.has(replicator));\n\n await replicator.connect({\n peerId: this.peerId,\n onConnectionOpen: this._onConnectionOpen.bind(this),\n onConnectionClosed: this._onConnectionClosed.bind(this),\n });\n }\n\n @synchronized\n async removeReplicator(replicator: EchoReplicator) {\n invariant(this._replicators.has(replicator));\n await replicator.disconnect();\n }\n\n async shouldAdvertize(peerId: PeerId, params: ShouldAdvertizeParams): Promise<boolean> {\n const connection = this._connections.get(peerId);\n if (!connection) {\n return false;\n }\n\n return connection.connection.shouldAdvertize(params);\n }\n\n private _onConnectionOpen(connection: ReplicatorConnection) {\n invariant(!this._connections.has(connection.peerId as PeerId));\n const reader = connection.readable.getReader();\n const writer = connection.writable.getWriter();\n const connectionEntry: ConnectionEntry = { connection, reader, writer, isOpen: true };\n this._connections.set(connection.peerId as PeerId, connectionEntry);\n\n queueMicrotask(async () => {\n try {\n while (true) {\n // TODO(dmaretskyi): Find a way to enforce backpressure on AM-repo.\n const { done, value } = await reader.read();\n if (done) {\n break;\n }\n\n this.emit('message', value);\n }\n } catch (err) {\n if (connectionEntry.isOpen) {\n log.catch(err);\n }\n }\n });\n\n this.emit('peer-candidate', {\n peerId: connection.peerId as PeerId,\n peerMetadata: {\n // TODO(dmaretskyi): Refactor this.\n dxos_peerSource: 'EchoNetworkAdapter',\n } as any,\n });\n }\n\n private _onConnectionClosed(connection: ReplicatorConnection) {\n const entry = this._connections.get(connection.peerId as PeerId);\n invariant(entry);\n\n entry.isOpen = false;\n this.emit('peer-disconnected', { peerId: connection.peerId as PeerId });\n\n void entry.reader.cancel().catch((err) => log.catch(err));\n void entry.writer.abort().catch((err) => log.catch(err));\n\n this._connections.delete(connection.peerId as PeerId);\n }\n}\n\ntype ConnectionEntry = {\n connection: ReplicatorConnection;\n reader: ReadableStreamDefaultReader<Message>;\n writer: WritableStreamDefaultWriter<Message>;\n isOpen: boolean;\n};\n", "//\n// Copyright 2024 DXOS.org\n// s\n\nimport { type MixedEncoding } from 'level-transcoder';\n\nimport { type StorageAdapterInterface, type Chunk, type StorageKey } from '@dxos/automerge/automerge-repo';\nimport { LifecycleState, Resource } from '@dxos/context';\nimport { type BatchLevel, type SublevelDB } from '@dxos/kv-store';\nimport { type MaybePromise } from '@dxos/util';\n\nexport type LevelDBStorageAdapterParams = {\n db: SublevelDB;\n callbacks?: StorageCallbacks;\n};\n\nexport type BeforeSaveParams = { path: StorageKey; batch: BatchLevel };\n\nexport interface StorageCallbacks {\n beforeSave(params: BeforeSaveParams): MaybePromise<void>;\n afterSave(path: StorageKey): MaybePromise<void>;\n}\n\nexport class LevelDBStorageAdapter extends Resource implements StorageAdapterInterface {\n constructor(private readonly _params: LevelDBStorageAdapterParams) {\n super();\n }\n\n async load(keyArray: StorageKey): Promise<Uint8Array | undefined> {\n try {\n if (this._lifecycleState !== LifecycleState.OPEN) {\n // TODO(mykola): this should be an error.\n return undefined;\n }\n return await this._params.db.get<StorageKey, Uint8Array>(keyArray, { ...encodingOptions });\n } catch (err: any) {\n if (isLevelDbNotFoundError(err)) {\n return undefined;\n }\n throw err;\n }\n }\n\n async save(keyArray: StorageKey, binary: Uint8Array): Promise<void> {\n if (this._lifecycleState !== LifecycleState.OPEN) {\n return undefined;\n }\n const batch = this._params.db.batch();\n\n await this._params.callbacks?.beforeSave?.({ path: keyArray, batch });\n batch.put<StorageKey, Uint8Array>(keyArray, Buffer.from(binary), {\n ...encodingOptions,\n });\n await batch.write();\n\n await this._params.callbacks?.afterSave?.(keyArray);\n }\n\n async remove(keyArray: StorageKey): Promise<void> {\n if (this._lifecycleState !== LifecycleState.OPEN) {\n return undefined;\n }\n await this._params.db.del<StorageKey>(keyArray, { ...encodingOptions });\n }\n\n async loadRange(keyPrefix: StorageKey): Promise<Chunk[]> {\n if (this._lifecycleState !== LifecycleState.OPEN) {\n return [];\n }\n const result: Chunk[] = [];\n for await (const [key, value] of this._params.db.iterator<StorageKey, Uint8Array>({\n gte: keyPrefix,\n lte: [...keyPrefix, '\\uffff'],\n ...encodingOptions,\n })) {\n result.push({\n key,\n data: value,\n });\n }\n return result;\n }\n\n async removeRange(keyPrefix: StorageKey): Promise<void> {\n if (this._lifecycleState !== LifecycleState.OPEN) {\n return undefined;\n }\n const batch = this._params.db.batch();\n\n for await (const [key] of this._params.db.iterator<StorageKey, Uint8Array>({\n gte: keyPrefix,\n lte: [...keyPrefix, '\\uffff'],\n ...encodingOptions,\n })) {\n batch.del<StorageKey>(key, { ...encodingOptions });\n }\n await batch.write();\n }\n}\n\nconst keyEncoder: MixedEncoding<StorageKey, Uint8Array, StorageKey> = {\n encode: (key: StorageKey): Uint8Array =>\n Buffer.from(key.map((k) => k.replaceAll('%', '%25').replaceAll('-', '%2D')).join('-')),\n decode: (key: Uint8Array): StorageKey =>\n Buffer.from(key)\n .toString()\n .split('-')\n .map((k) => k.replaceAll('%2D', '-').replaceAll('%25', '%')),\n format: 'buffer',\n};\n\nexport const encodingOptions = {\n keyEncoding: keyEncoder,\n valueEncoding: 'buffer',\n};\n\nconst isLevelDbNotFoundError = (err: any): boolean => err.code === 'LEVEL_NOT_FOUND';\n", "//\n// Copyright 2024 DXOS.org\n//\n\nimport { Trigger } from '@dxos/async';\nimport { NetworkAdapter, type Message, type PeerId, cbor } from '@dxos/automerge/automerge-repo';\nimport { Stream } from '@dxos/codec-protobuf';\nimport { invariant } from '@dxos/invariant';\nimport { type HostInfo, type SyncRepoRequest, type SyncRepoResponse } from '@dxos/protocols/proto/dxos/echo/service';\n\ntype ClientSyncState = {\n connected: boolean;\n send: (message: Message) => void;\n disconnect: () => void;\n};\n\n/**\n * Used to replicate with apps running on the same device.\n */\nexport class LocalHostNetworkAdapter extends NetworkAdapter {\n private readonly _peers: Map<PeerId, ClientSyncState> = new Map();\n\n /**\n * Emits `ready` event. That signals to `Repo` that it can start using the adapter.\n */\n ready() {\n // NOTE: Emitting `ready` event in NetworkAdapter`s constructor causes a race condition\n // because `Repo` waits for `ready` event (which it never receives) before it starts using the adapter.\n this.emit('ready', {\n network: this,\n });\n }\n\n private readonly _connected = new Trigger();\n private _isConnected: boolean = false;\n\n /**\n * Called by `Repo` to connect to the network.\n *\n * @param peerId Our peer Id.\n */\n override connect(peerId: PeerId): void {\n this.peerId = peerId;\n this._isConnected = true;\n this._connected.wake();\n // No-op. Client always connects first\n }\n\n override send(message: Message): void {\n const peer = this._peers.get(message.targetId);\n invariant(peer, 'Peer not found.');\n peer.send(message);\n }\n\n async close() {\n this._peers.forEach((peer) => peer.disconnect());\n this.emit('close');\n }\n\n override disconnect(): void {\n // TODO(mykola): `disconnect` is not used anywhere in `Repo` from `@automerge/automerge-repo`. Should we remove it?\n // No-op\n }\n\n async whenConnected(): Promise<void> {\n await this._connected.wait({ timeout: 10_000 });\n }\n\n syncRepo({ id, syncMessage }: SyncRepoRequest): Stream<SyncRepoResponse> {\n const peerId = this._getPeerId(id);\n\n return new Stream(({ next, close }) => {\n invariant(!this._peers.has(peerId), 'Peer already connected.');\n this._peers.set(peerId, {\n connected: true,\n send: (message) => {\n next({\n syncMessage: cbor.encode(message),\n });\n },\n disconnect: () => {\n this._peers.delete(peerId);\n close();\n this.emit('peer-disconnected', {\n peerId,\n });\n },\n });\n\n invariant(this._isConnected);\n this.emit('peer-candidate', {\n peerMetadata: {},\n peerId,\n });\n });\n }\n\n async sendSyncMessage({ id, syncMessage }: SyncRepoRequest): Promise<void> {\n invariant(this._isConnected);\n const message = cbor.decode(syncMessage!) as Message;\n this.emit('message', message);\n }\n\n async getHostInfo(): Promise<HostInfo> {\n invariant(this._isConnected);\n invariant(this.peerId, 'Peer id not set.');\n return {\n peerId: this.peerId,\n };\n }\n\n private _getPeerId(id: string): PeerId {\n return id as PeerId;\n }\n}\n", "//\n// Copyright 2024 DXOS.org\n//\n\nimport { Trigger } from '@dxos/async';\nimport { NetworkAdapter, type Message, type PeerId, cbor } from '@dxos/automerge/automerge-repo';\nimport { invariant } from '@dxos/invariant';\nimport { log } from '@dxos/log';\nimport { type PeerInfo } from '@dxos/protocols/proto/dxos/mesh/teleport/automerge';\nimport { AutomergeReplicator } from '@dxos/teleport-extension-automerge-replicator';\n\n/**\n * Used to replicate with other peers over the network.\n */\nexport class MeshNetworkAdapter extends NetworkAdapter {\n private readonly _extensions: Map<string, AutomergeReplicator> = new Map();\n private _connected = new Trigger();\n\n /**\n * Emits `ready` event. That signals to `Repo` that it can start using the adapter.\n */\n ready() {\n // NOTE: Emitting `ready` event in NetworkAdapter`s constructor causes a race condition\n // because `Repo` waits for `ready` event (which it never receives) before it starts using the adapter.\n this.emit('ready', {\n network: this,\n });\n }\n\n override connect(peerId: PeerId): void {\n this.peerId = peerId;\n this._connected.wake();\n }\n\n override send(message: Message): void {\n const receiverId = message.targetId;\n const extension = this._extensions.get(receiverId);\n invariant(extension, 'Extension not found.');\n extension.sendSyncMessage({ payload: cbor.encode(message) }).catch((err) => log.catch(err));\n }\n\n override disconnect(): void {\n // No-op\n }\n\n createExtension(): AutomergeReplicator {\n invariant(this.peerId, 'Peer id not set.');\n\n let peerInfo: PeerInfo;\n const extension = new AutomergeReplicator(\n {\n peerId: this.peerId,\n },\n {\n onStartReplication: async (info, remotePeerId /** Teleport ID */) => {\n await this._connected.wait();\n\n // Note: We store only one extension per peer.\n // There can be a case where two connected peers have more than one teleport connection between them\n // and each of them uses different teleport connections to send messages.\n // It works because we receive messages from all teleport connections and Automerge Repo dedup them.\n // TODO(mykola): Use only one teleport connection per peer.\n\n // TODO(dmaretskyi): Critical bug.\n // - two peers get connected via swarm 1\n // - they get connected via swarm 2\n // - swarm 1 gets disconnected\n // - automerge repo thinks that peer 2 got disconnected even though swarm 2 is still active\n\n log('onStartReplication', { id: info.id, thisPeerId: this.peerId, remotePeerId: remotePeerId.toHex() });\n if (!this._extensions.has(info.id)) {\n peerInfo = info;\n // TODO(mykola): Fix race condition?\n this._extensions.set(info.id, extension);\n\n log('peer-candidate', { id: info.id, thisPeerId: this.peerId, remotePeerId: remotePeerId.toHex() });\n this.emit('peer-candidate', {\n // TODO(mykola): Hack, stop abusing `peerMetadata` field.\n peerMetadata: {\n dxos_deviceKey: remotePeerId.toHex(),\n } as any,\n peerId: info.id as PeerId,\n });\n }\n },\n onSyncMessage: async ({ payload }) => {\n if (!peerInfo) {\n return;\n }\n const message = cbor.decode(payload) as Message;\n // Note: automerge Repo dedup messages.\n this.emit('message', message);\n },\n onClose: async () => {\n if (!peerInfo) {\n return;\n }\n this.emit('peer-disconnected', {\n peerId: peerInfo.id as PeerId,\n });\n this._extensions.delete(peerInfo.id);\n },\n },\n );\n return extension;\n }\n}\n", "//\n// Copyright 2024 DXOS.org\n//\n\nimport { type StorageKey } from '@dxos/automerge/automerge-repo';\nimport { IndexedDBStorageAdapter } from '@dxos/automerge/automerge-repo-storage-indexeddb';\nimport { type SublevelDB } from '@dxos/kv-store';\nimport { log } from '@dxos/log';\nimport { StorageType, type Directory } from '@dxos/random-access-storage';\n\nimport { AutomergeStorageAdapter } from './automerge-storage-adapter';\nimport { encodingOptions } from './leveldb-storage-adapter';\n\nexport const levelMigration = async ({ db, directory }: { db: SublevelDB; directory: Directory }) => {\n // Note: Make automigration from previous storage to leveldb here.\n const isNewLevel = !(await db\n .iterator<StorageKey, Uint8Array>({\n ...encodingOptions,\n })\n .next());\n\n if (!isNewLevel) {\n return;\n }\n\n const oldStorageAdapter =\n directory.type === StorageType.IDB\n ? new IndexedDBStorageAdapter(directory.path, 'data')\n : new AutomergeStorageAdapter(directory);\n\n const chunks = await oldStorageAdapter.loadRange([]);\n if (chunks.length === 0) {\n return;\n }\n const batch = db.batch();\n log.info('found chunks on old storage adapter', { chunks: chunks.length });\n for (const { key, data } of await oldStorageAdapter.loadRange([])) {\n data && batch.put<StorageKey, Uint8Array>(key, data, { ...encodingOptions });\n }\n await batch.write();\n};\n", "//\n// Copyright 2024 DXOS.org\n//\n//\n// Copyright 2023 DXOS.org\n//\n\nimport { type Chunk, type StorageKey, type StorageAdapterInterface } from '@dxos/automerge/automerge-repo';\nimport { type Directory } from '@dxos/random-access-storage';\nimport { arrayToBuffer, bufferToArray } from '@dxos/util';\n\nexport class AutomergeStorageAdapter implements StorageAdapterInterface {\n // TODO(mykola): Hack for restricting automerge Repo to access storage if Host is `closed`.\n // Automerge Repo do not have any lifetime management.\n private _state: 'opened' | 'closed' = 'opened';\n\n constructor(private readonly _directory: Directory) {}\n\n async load(key: StorageKey): Promise<Uint8Array | undefined> {\n if (this._state !== 'opened') {\n return undefined;\n }\n const filename = this._getFilename(key);\n const file = this._directory.getOrCreateFile(filename);\n const { size } = await file.stat();\n if (!size || size === 0) {\n return undefined;\n }\n const buffer = await file.read(0, size);\n return bufferToArray(buffer);\n }\n\n async save(key: StorageKey, data: Uint8Array): Promise<void> {\n if (this._state !== 'opened') {\n return undefined;\n }\n const filename = this._getFilename(key);\n const file = this._directory.getOrCreateFile(filename);\n await file.write(0, arrayToBuffer(data));\n await file.truncate?.(data.length);\n\n await file.flush?.();\n }\n\n async remove(key: StorageKey): Promise<void> {\n if (this._state !== 'opened') {\n return undefined;\n }\n // TODO(dmaretskyi): Better deletion.\n const filename = this._getFilename(key);\n const file = this._directory.getOrCreateFile(filename);\n await file.destroy();\n }\n\n async loadRange(keyPrefix: StorageKey): Promise<Chunk[]> {\n if (this._state !== 'opened') {\n return [];\n }\n const filename = this._getFilename(keyPrefix);\n const entries = await this._directory.list();\n return Promise.all(\n entries\n .filter((entry) => entry.startsWith(filename))\n .map(async (entry): Promise<Chunk> => {\n const file = this._directory.getOrCreateFile(entry);\n const { size } = await file.stat();\n const buffer = await file.read(0, size);\n return {\n key: this._getKeyFromFilename(entry),\n data: bufferToArray(buffer),\n };\n }),\n );\n }\n\n async removeRange(keyPrefix: StorageKey): Promise<void> {\n if (this._state !== 'opened') {\n return undefined;\n }\n const filename = this._getFilename(keyPrefix);\n const entries = await this._directory.list();\n await Promise.all(\n entries\n .filter((entry) => entry.startsWith(filename))\n .map(async (entry): Promise<void> => {\n const file = this._directory.getOrCreateFile(entry);\n await file.destroy();\n }),\n );\n }\n\n async close(): Promise<void> {\n this._state = 'closed';\n }\n\n private _getFilename(key: StorageKey): string {\n return key.map((k) => k.replaceAll('%', '%25').replaceAll('-', '%2D')).join('-');\n }\n\n private _getKeyFromFilename(filename: string): StorageKey {\n return filename.split('-').map((k) => k.replaceAll('%2D', '-').replaceAll('%25', '%'));\n }\n}\n", "//\n// Copyright 2024 DXOS.org\n//\n\nimport { Event } from '@dxos/async';\nimport { type DocHandle, type AutomergeUrl, type DocumentId, type Repo } from '@dxos/automerge/automerge-repo';\nimport { cancelWithContext, type Context } from '@dxos/context';\nimport { warnAfterTimeout } from '@dxos/debug';\nimport { type SpaceState, type SpaceDoc } from '@dxos/echo-protocol';\nimport { invariant } from '@dxos/invariant';\nimport { type PublicKey } from '@dxos/keys';\nimport { log } from '@dxos/log';\nimport { trace } from '@dxos/tracing';\n\ntype SpaceDocumentLinks = SpaceDoc['links'];\n\nexport interface AutomergeDocumentLoader {\n onObjectDocumentLoaded: Event<ObjectDocumentLoaded>;\n\n getAllHandles(): DocHandle<SpaceDoc>[];\n\n loadSpaceRootDocHandle(ctx: Context, spaceState: SpaceState): Promise<void>;\n loadObjectDocument(objectId: string | string[]): void;\n getSpaceRootDocHandle(): DocHandle<SpaceDoc>;\n createDocumentForObject(objectId: string): DocHandle<SpaceDoc>;\n onObjectLinksUpdated(links: SpaceDocumentLinks): void;\n onObjectBoundToDocument(handle: DocHandle<SpaceDoc>, objectId: string): void;\n\n /**\n * @returns objectIds for which we had document handles or were loading one.\n */\n clearHandleReferences(): string[];\n}\n\n/**\n * Manages object <-> docHandle binding and automerge document loading.\n */\n@trace.resource()\nexport class AutomergeDocumentLoaderImpl implements AutomergeDocumentLoader {\n private _spaceRootDocHandle: DocHandle<SpaceDoc> | null = null;\n /**\n * An object id pointer to a handle of the document where the object is stored inline.\n */\n private readonly _objectDocumentHandles = new Map<string, DocHandle<SpaceDoc>>();\n /**\n * If object was requested via loadObjectDocument but root document links weren't updated yet\n * loading will be triggered in onObjectLinksUpdated callback.\n */\n private readonly _objectsPendingDocumentLoad = new Set<string>();\n\n public readonly onObjectDocumentLoaded = new Event<ObjectDocumentLoaded>();\n\n constructor(\n private readonly _spaceKey: PublicKey,\n private readonly _repo: Repo,\n ) {}\n\n getAllHandles(): DocHandle<SpaceDoc>[] {\n return this._spaceRootDocHandle != null\n ? [this._spaceRootDocHandle, ...new Set(this._objectDocumentHandles.values())]\n : [];\n }\n\n @trace.span({ showInBrowserTimeline: true })\n public async loadSpaceRootDocHandle(ctx: Context, spaceState: SpaceState): Promise<void> {\n if (this._spaceRootDocHandle != null) {\n return;\n }\n if (!spaceState.rootUrl) {\n log.error('Database opened with no rootUrl', { spaceKey: this._spaceKey });\n this._createContextBoundSpaceRootDocument(ctx);\n } else {\n const existingDocHandle = await this._initDocHandle(ctx, spaceState.rootUrl);\n const doc = existingDocHandle.docSync();\n invariant(doc);\n if (doc.access == null) {\n this._initDocAccess(existingDocHandle);\n }\n this._spaceRootDocHandle = existingDocHandle;\n }\n }\n\n public loadObjectDocument(objectIdOrMany: string | string[]) {\n const objectIds = Array.isArray(objectIdOrMany) ? objectIdOrMany : [objectIdOrMany];\n let hasUrlsToLoad = false;\n const urlsToLoad: SpaceDoc['links'] = {};\n for (const objectId of objectIds) {\n invariant(this._spaceRootDocHandle);\n if (this._objectDocumentHandles.has(objectId) || this._objectsPendingDocumentLoad.has(objectId)) {\n continue;\n }\n const spaceRootDoc = this._spaceRootDocHandle.docSync();\n invariant(spaceRootDoc);\n const documentUrl = (spaceRootDoc.links ?? {})[objectId];\n if (documentUrl == null) {\n this._objectsPendingDocumentLoad.add(objectId);\n log.info('loading delayed until object links are initialized', { objectId });\n } else {\n urlsToLoad[objectId] = documentUrl;\n hasUrlsToLoad = true;\n }\n }\n if (hasUrlsToLoad) {\n this._loadLinkedObjects(urlsToLoad);\n }\n }\n\n public onObjectLinksUpdated(links: SpaceDocumentLinks) {\n if (!links) {\n return;\n }\n const linksAwaitingLoad = Object.entries(links).filter(([objectId]) =>\n this._objectsPendingDocumentLoad.has(objectId),\n );\n this._loadLinkedObjects(Object.fromEntries(linksAwaitingLoad));\n linksAwaitingLoad.forEach(([objectId]) => this._objectsPendingDocumentLoad.delete(objectId));\n }\n\n public getSpaceRootDocHandle(): DocHandle<SpaceDoc> {\n invariant(this._spaceRootDocHandle);\n return this._spaceRootDocHandle;\n }\n\n public createDocumentForObject(objectId: string): DocHandle<SpaceDoc> {\n invariant(this._spaceRootDocHandle);\n const spaceDocHandle = this._repo.create<SpaceDoc>();\n this._initDocAccess(spaceDocHandle);\n this.onObjectBoundToDocument(spaceDocHandle, objectId);\n this._spaceRootDocHandle.change((newDoc: SpaceDoc) => {\n newDoc.links ??= {};\n newDoc.links[objectId] = spaceDocHandle.url;\n });\n return spaceDocHandle;\n }\n\n public onObjectBoundToDocument(handle: DocHandle<SpaceDoc>, objectId: string) {\n this._objectDocumentHandles.set(objectId, handle);\n }\n\n public clearHandleReferences(): string[] {\n const objectsWithHandles = [...this._objectDocumentHandles.keys()];\n this._objectDocumentHandles.clear();\n this._spaceRootDocHandle = null;\n return objectsWithHandles;\n }\n\n private _loadLinkedObjects(links: SpaceDocumentLinks) {\n if (!links) {\n return;\n }\n for (const [objectId, automergeUrl] of Object.entries(links)) {\n const logMeta = { objectId, automergeUrl };\n const objectDocumentHandle = this._objectDocumentHandles.get(objectId);\n if (objectDocumentHandle != null && objectDocumentHandle.url !== automergeUrl) {\n log.warn('object already inlined in a different document, ignoring the link', {\n ...logMeta,\n actualDocumentUrl: objectDocumentHandle.url,\n });\n continue;\n }\n if (objectDocumentHandle?.url === automergeUrl) {\n log.warn('object document was already loaded', logMeta);\n continue;\n }\n const handle = this._repo.find<SpaceDoc>(automergeUrl as DocumentId);\n log.debug('document loading triggered', logMeta);\n this._objectDocumentHandles.set(objectId, handle);\n void this._createObjectOnDocumentLoad(handle, objectId);\n }\n }\n\n private async _initDocHandle(ctx: Context, url: string) {\n const docHandle = this._repo.find<SpaceDoc>(url as DocumentId);\n while (true) {\n try {\n await warnAfterTimeout(5_000, 'Automerge root doc load timeout (AutomergeDb)', async () => {\n await cancelWithContext(ctx, docHandle.whenReady()); // TODO(dmaretskyi): Temporary 5s timeout for debugging.\n });\n break;\n } catch (err) {\n if (`${err}`.includes('Timeout')) {\n log.info('wraparound', { id: docHandle.documentId, state: docHandle.state });\n continue;\n }\n\n throw err;\n }\n }\n\n if (docHandle.state === 'unavailable') {\n throw new Error('Automerge document is unavailable');\n }\n\n return docHandle;\n }\n\n private _createContextBoundSpaceRootDocument(ctx: Context) {\n const docHandle = this._repo.create<SpaceDoc>();\n this._spaceRootDocHandle = docHandle;\n ctx.onDispose(() => {\n docHandle.delete();\n this._spaceRootDocHandle = null;\n });\n }\n\n private _initDocAccess(handle: DocHandle<SpaceDoc>) {\n handle.change((newDoc: SpaceDoc) => {\n newDoc.access ??= { spaceKey: this._spaceKey.toHex() };\n newDoc.access.spaceKey = this._spaceKey.toHex();\n });\n }\n\n private async _createObjectOnDocumentLoad(handle: DocHandle<SpaceDoc>, objectId: string) {\n try {\n await handle.doc(['ready']);\n const logMeta = { objectId, docUrl: handle.url };\n if (this.onObjectDocumentLoaded.listenerCount() === 0) {\n log.info('document loaded after all listeners were removed', logMeta);\n return;\n }\n const objectDocHandle = this._objectDocumentHandles.get(objectId);\n if (objectDocHandle?.url !== handle.url) {\n log.warn('object was rebound while a document was loading, discarding handle', logMeta);\n return;\n }\n this.onObjectDocumentLoaded.emit({ handle, objectId });\n } catch (err) {\n const shouldRetryLoading = this.onObjectDocumentLoaded.listenerCount() > 0;\n log.warn('failed to load a document', {\n objectId,\n automergeUrl: handle.url,\n retryLoading: shouldRetryLoading,\n err,\n });\n if (shouldRetryLoading) {\n await this._createObjectOnDocumentLoad(handle, objectId);\n }\n }\n }\n}\n\nexport interface ObjectDocumentLoaded {\n handle: DocHandle<SpaceDoc>;\n objectId: string;\n}\n\nexport interface DocumentChanges {\n createdObjectIds: string[];\n updatedObjectIds: string[];\n objectsToRebind: string[];\n linkedDocuments: {\n [echoId: string]: AutomergeUrl;\n };\n}\n"],
|
|
5
5
|
"mappings": ";;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAIA,mBAAsB;AACtB,uBAA8E;AAC9E,4BAOO;AAEP,qBAAwC;AAGxC,uBAA0B;AAC1B,kBAA0B;AAE1B,iBAAoB;AACpB,uBAAwB;AASxB,qBAAsB;AACtB,kBAA8D;AC5B9D,IAAAA,gBAAsC;AACtC,IAAAC,yBAA6E;AAC7E,IAAAC,kBAA+B;AAC/B,IAAAC,oBAA0B;AAC1B,IAAAC,cAAoB;ACDpB,IAAAF,kBAAyC;ACHzC,IAAAF,gBAAwB;AACxB,IAAAC,yBAAgE;AAChE,4BAAuB;AACvB,IAAAE,oBAA0B;ACH1B,IAAAH,gBAAwB;AACxB,IAAAC,yBAAgE;AAChE,IAAAE,oBAA0B;AAC1B,IAAAC,cAAoB;AAEpB,qDAAoC;ACJpC,8CAAwC;AAExC,IAAAA,cAAoB;AACpB,mCAA4C;ACC5C,IAAAC,eAA6C;ACL7C,IAAAL,gBAAsB;AAEtB,IAAAE,kBAAgD;AAChD,mBAAiC;AAEjC,IAAAC,oBAA0B;AAE1B,IAAAC,cAAoB;AACpB,IAAAE,kBAAsB;;;;;;;;;;;;ANGf,IAAMC,qBAAN,cAAiCC,sCAAAA;EAAjC,cAAA;;AACYC,SAAAA,eAAe,oBAAIC,IAAAA;wBAIJ,oBAAIC,IAAAA;AAC5BC,SAAAA,kBAAkCC,+BAAeC;AACxCC,SAAAA,aAAa,IAAIC,sBAAAA;;EAEzBC,QAAQC,QAAgBC,cAA+C;AAC9E,SAAKD,SAASA;AACd,SAAKC,eAAeA;AACpB,SAAKJ,WAAWK,KAAI;EACtB;EAESC,KAAKC,SAAwB;AACpC,UAAMC,kBAAkB,KAAKC,aAAaC,IAAIH,QAAQI,QAAQ;AAC9D,QAAI,CAACH,iBAAiB;AACpB,YAAM,IAAII,MAAM,uBAAA;IAClB;AAGAJ,oBAAgBK,OAAOC,MAAMP,OAAAA,EAASQ,MAAM,CAACC,QAAAA;AAC3C,UAAIR,gBAAgBS,QAAQ;AAC1BC,wBAAIH,MAAMC,KAAAA,QAAAA;;;;;;MACZ;IACF,CAAA;EACF;EAESG,aAAmB;EAE5B;EAEA,MACMC,OAAO;AACXC,qCAAU,KAAKxB,oBAAoBC,+BAAeC,QAAM,QAAA;;;;;;;;;AACxD,SAAKF,kBAAkBC,+BAAewB;AAEtC,SAAKC,KAAK,SAAS;MACjBC,SAAS;IACX,CAAA;EACF;EAEA,MACMC,QAAQ;AACZJ,qCAAU,KAAKxB,oBAAoBC,+BAAewB,MAAI,QAAA;;;;;;;;;AAEtD,eAAWI,cAAc,KAAKhC,cAAc;AAC1C,YAAMgC,WAAWP,WAAU;IAC7B;AACA,SAAKzB,aAAaiC,MAAK;AAEvB,SAAK9B,kBAAkBC,+BAAeC;EACxC;EAEA,MAAM6B,gBAAgB;AACpB,UAAM,KAAK5B,WAAW6B,KAAK;MAAEC,SAAS;IAAO,CAAA;EAC/C;EAEA,MACMC,cAAcL,YAA4B;AAC9CL,qCAAU,KAAKlB,QAAM,QAAA;;;;;;;;;AACrBkB,qCAAU,CAAC,KAAK3B,aAAasC,IAAIN,UAAAA,GAAAA,QAAAA;;;;;;;;;AAEjC,UAAMA,WAAWxB,QAAQ;MACvBC,QAAQ,KAAKA;MACb8B,kBAAkB,KAAKC,kBAAkBC,KAAK,IAAI;MAClDC,oBAAoB,KAAKC,oBAAoBF,KAAK,IAAI;IACxD,CAAA;EACF;EAEA,MACMG,iBAAiBZ,YAA4B;AACjDL,qCAAU,KAAK3B,aAAasC,IAAIN,UAAAA,GAAAA,QAAAA;;;;;;;;;AAChC,UAAMA,WAAWP,WAAU;EAC7B;EAEA,MAAMoB,gBAAgBpC,QAAgBqC,QAAiD;AACrF,UAAMC,aAAa,KAAKhC,aAAaC,IAAIP,MAAAA;AACzC,QAAI,CAACsC,YAAY;AACf,aAAO;IACT;AAEA,WAAOA,WAAWA,WAAWF,gBAAgBC,MAAAA;EAC/C;EAEQN,kBAAkBO,YAAkC;AAC1DpB,qCAAU,CAAC,KAAKZ,aAAauB,IAAIS,WAAWtC,MAAM,GAAA,QAAA;;;;;;;;;AAClD,UAAMuC,SAASD,WAAWE,SAASC,UAAS;AAC5C,UAAM/B,SAAS4B,WAAWI,SAASC,UAAS;AAC5C,UAAMtC,kBAAmC;MAAEiC;MAAYC;MAAQ7B;MAAQI,QAAQ;IAAK;AACpF,SAAKR,aAAasC,IAAIN,WAAWtC,QAAkBK,eAAAA;AAEnDwC,mBAAe,YAAA;AACb,UAAI;AACF,eAAO,MAAM;AAEX,gBAAM,EAAEC,MAAMC,MAAK,IAAK,MAAMR,OAAOS,KAAI;AACzC,cAAIF,MAAM;AACR;UACF;AAEA,eAAK1B,KAAK,WAAW2B,KAAAA;QACvB;MACF,SAASlC,KAAK;AACZ,YAAIR,gBAAgBS,QAAQ;AAC1BC,0BAAIH,MAAMC,KAAAA,QAAAA;;;;;;QACZ;MACF;IACF,CAAA;AAEA,SAAKO,KAAK,kBAAkB;MAC1BpB,QAAQsC,WAAWtC;MACnBC,cAAc;;QAEZgD,iBAAiB;MACnB;IACF,CAAA;EACF;EAEQf,oBAAoBI,YAAkC;AAC5D,UAAMY,QAAQ,KAAK5C,aAAaC,IAAI+B,WAAWtC,MAAM;AACrDkB,qCAAUgC,OAAAA,QAAAA;;;;;;;;;AAEVA,UAAMpC,SAAS;AACf,SAAKM,KAAK,qBAAqB;MAAEpB,QAAQsC,WAAWtC;IAAiB,CAAA;AAErE,SAAKkD,MAAMX,OAAOY,OAAM,EAAGvC,MAAM,CAACC,QAAQE,gBAAIH,MAAMC,KAAAA,QAAAA;;;;;;AACpD,SAAKqC,MAAMxC,OAAO0C,MAAK,EAAGxC,MAAM,CAACC,QAAQE,gBAAIH,MAAMC,KAAAA,QAAAA;;;;;;AAEnD,SAAKP,aAAa+C,OAAOf,WAAWtC,MAAM;EAC5C;AACF;;EAnGGsD;GAjCUjE,mBAAAA,WAAAA,QAAAA,IAAAA;;EA2CViE;GA3CUjE,mBAAAA,WAAAA,SAAAA,IAAAA;;EA2DViE;GA3DUjE,mBAAAA,WAAAA,iBAAAA,IAAAA;;EAuEViE;GAvEUjE,mBAAAA,WAAAA,oBAAAA,IAAAA;ACQN,IAAMkE,wBAAN,cAAoCC,yBAAAA;EACzCC,YAA6BC,SAAsC;AACjE,UAAK;SADsBA,UAAAA;EAE7B;EAEA,MAAMC,KAAKC,UAAuD;AAChE,QAAI;AACF,UAAI,KAAKlE,oBAAoBC,gBAAAA,eAAewB,MAAM;AAEhD,eAAO0C;MACT;AACA,aAAO,MAAM,KAAKH,QAAQI,GAAGvD,IAA4BqD,UAAU;QAAE,GAAGG;MAAgB,CAAA;IAC1F,SAASlD,KAAU;AACjB,UAAImD,uBAAuBnD,GAAAA,GAAM;AAC/B,eAAOgD;MACT;AACA,YAAMhD;IACR;EACF;EAEA,MAAMoD,KAAKL,UAAsBM,QAAmC;AAClE,QAAI,KAAKxE,oBAAoBC,gBAAAA,eAAewB,MAAM;AAChD,aAAO0C;IACT;AACA,UAAMM,QAAQ,KAAKT,QAAQI,GAAGK,MAAK;AAEnC,UAAM,KAAKT,QAAQU,WAAWC,aAAa;MAAEC,MAAMV;MAAUO;IAAM,CAAA;AACnEA,UAAMI,IAA4BX,UAAUY,OAAOC,KAAKP,MAAAA,GAAS;MAC/D,GAAGH;IACL,CAAA;AACA,UAAMI,MAAMxD,MAAK;AAEjB,UAAM,KAAK+C,QAAQU,WAAWM,YAAYd,QAAAA;EAC5C;EAEA,MAAMe,OAAOf,UAAqC;AAChD,QAAI,KAAKlE,oBAAoBC,gBAAAA,eAAewB,MAAM;AAChD,aAAO0C;IACT;AACA,UAAM,KAAKH,QAAQI,GAAGc,IAAgBhB,UAAU;MAAE,GAAGG;IAAgB,CAAA;EACvE;EAEA,MAAMc,UAAUC,WAAyC;AACvD,QAAI,KAAKpF,oBAAoBC,gBAAAA,eAAewB,MAAM;AAChD,aAAO,CAAA;IACT;AACA,UAAM4D,SAAkB,CAAA;AACxB,qBAAiB,CAACC,KAAKjC,KAAAA,KAAU,KAAKW,QAAQI,GAAGmB,SAAiC;MAChFC,KAAKJ;MACLK,KAAK;WAAIL;QAAW;;MACpB,GAAGf;IACL,CAAA,GAAI;AACFgB,aAAOK,KAAK;QACVJ;QACAK,MAAMtC;MACR,CAAA;IACF;AACA,WAAOgC;EACT;EAEA,MAAMO,YAAYR,WAAsC;AACtD,QAAI,KAAKpF,oBAAoBC,gBAAAA,eAAewB,MAAM;AAChD,aAAO0C;IACT;AACA,UAAMM,QAAQ,KAAKT,QAAQI,GAAGK,MAAK;AAEnC,qBAAiB,CAACa,GAAAA,KAAQ,KAAKtB,QAAQI,GAAGmB,SAAiC;MACzEC,KAAKJ;MACLK,KAAK;WAAIL;QAAW;;MACpB,GAAGf;IACL,CAAA,GAAI;AACFI,YAAMS,IAAgBI,KAAK;QAAE,GAAGjB;MAAgB,CAAA;IAClD;AACA,UAAMI,MAAMxD,MAAK;EACnB;AACF;AAEA,IAAM4E,aAAgE;EACpEC,QAAQ,CAACR,QACPR,OAAOC,KAAKO,IAAIS,IAAI,CAACC,MAAMA,EAAEC,WAAW,KAAK,KAAA,EAAOA,WAAW,KAAK,KAAA,CAAA,EAAQC,KAAK,GAAA,CAAA;EACnFC,QAAQ,CAACb,QACPR,OAAOC,KAAKO,GAAAA,EACTc,SAAQ,EACRC,MAAM,GAAA,EACNN,IAAI,CAACC,MAAMA,EAAEC,WAAW,OAAO,GAAA,EAAKA,WAAW,OAAO,GAAA,CAAA;EAC3DK,QAAQ;AACV;AAEO,IAAMjC,kBAAkB;EAC7BkC,aAAaV;EACbW,eAAe;AACjB;AAEA,IAAMlC,yBAAyB,CAACnD,QAAsBA,IAAIsF,SAAS;;ACjG5D,IAAMC,0BAAN,cAAsC9G,uBAAAA,eAAAA;EAAtC,cAAA;;AACY+G,SAAAA,SAAuC,oBAAI5G,IAAAA;AAa3CI,SAAAA,aAAa,IAAIC,cAAAA,QAAAA;AAC1BwG,SAAAA,eAAwB;;;;;EAThCC,QAAQ;AAGN,SAAKnF,KAAK,SAAS;MACjBC,SAAS;IACX,CAAA;EACF;;;;;;EAUStB,QAAQC,QAAsB;AACrC,SAAKA,SAASA;AACd,SAAKsG,eAAe;AACpB,SAAKzG,WAAWK,KAAI;EAEtB;EAESC,KAAKC,SAAwB;AACpC,UAAMoG,OAAO,KAAKH,OAAO9F,IAAIH,QAAQI,QAAQ;AAC7CU,0BAAAA,WAAUsF,MAAM,mBAAA;;;;;;;;;AAChBA,SAAKrG,KAAKC,OAAAA;EACZ;EAEA,MAAMkB,QAAQ;AACZ,SAAK+E,OAAOI,QAAQ,CAACD,SAASA,KAAKxF,WAAU,CAAA;AAC7C,SAAKI,KAAK,OAAA;EACZ;EAESJ,aAAmB;EAG5B;EAEA,MAAMS,gBAA+B;AACnC,UAAM,KAAK5B,WAAW6B,KAAK;MAAEC,SAAS;IAAO,CAAA;EAC/C;EAEA+E,SAAS,EAAEC,IAAIC,YAAW,GAA+C;AACvE,UAAM5G,SAAS,KAAK6G,WAAWF,EAAAA;AAE/B,WAAO,IAAIG,6BAAO,CAAC,EAAEC,MAAMzF,MAAK,MAAE;AAChCJ,4BAAAA,WAAU,CAAC,KAAKmF,OAAOxE,IAAI7B,MAAAA,GAAS,2BAAA;;;;;;;;;AACpC,WAAKqG,OAAOzD,IAAI5C,QAAQ;QACtBgH,WAAW;QACX7G,MAAM,CAACC,YAAAA;AACL2G,eAAK;YACHH,aAAaK,4BAAKzB,OAAOpF,OAAAA;UAC3B,CAAA;QACF;QACAY,YAAY,MAAA;AACV,eAAKqF,OAAOhD,OAAOrD,MAAAA;AACnBsB,gBAAAA;AACA,eAAKF,KAAK,qBAAqB;YAC7BpB;UACF,CAAA;QACF;MACF,CAAA;AAEAkB,4BAAAA,WAAU,KAAKoF,cAAY,QAAA;;;;;;;;;AAC3B,WAAKlF,KAAK,kBAAkB;QAC1BnB,cAAc,CAAC;QACfD;MACF,CAAA;IACF,CAAA;EACF;EAEA,MAAMkH,gBAAgB,EAAEP,IAAIC,YAAW,GAAoC;AACzE1F,0BAAAA,WAAU,KAAKoF,cAAY,QAAA;;;;;;;;;AAC3B,UAAMlG,UAAU6G,4BAAKpB,OAAOe,WAAAA;AAC5B,SAAKxF,KAAK,WAAWhB,OAAAA;EACvB;EAEA,MAAM+G,cAAiC;AACrCjG,0BAAAA,WAAU,KAAKoF,cAAY,QAAA;;;;;;;;;AAC3BpF,0BAAAA,WAAU,KAAKlB,QAAQ,oBAAA;;;;;;;;;AACvB,WAAO;MACLA,QAAQ,KAAKA;IACf;EACF;EAEQ6G,WAAWF,IAAoB;AACrC,WAAOA;EACT;AACF;;ACpGO,IAAMS,qBAAN,cAAiC9H,uBAAAA,eAAAA;EAAjC,cAAA;;AACY+H,SAAAA,cAAgD,oBAAI5H,IAAAA;AAC7DI,SAAAA,aAAa,IAAIC,cAAAA,QAAAA;;;;;EAKzByG,QAAQ;AAGN,SAAKnF,KAAK,SAAS;MACjBC,SAAS;IACX,CAAA;EACF;EAEStB,QAAQC,QAAsB;AACrC,SAAKA,SAASA;AACd,SAAKH,WAAWK,KAAI;EACtB;EAESC,KAAKC,SAAwB;AACpC,UAAMkH,aAAalH,QAAQI;AAC3B,UAAM+G,YAAY,KAAKF,YAAY9G,IAAI+G,UAAAA;AACvCpG,0BAAAA,WAAUqG,WAAW,wBAAA;;;;;;;;;AACrBA,cAAUL,gBAAgB;MAAEM,SAASP,uBAAAA,KAAKzB,OAAOpF,OAAAA;IAAS,CAAA,EAAGQ,MAAM,CAACC,QAAQE,YAAAA,IAAIH,MAAMC,KAAAA,QAAAA;;;;;;EACxF;EAESG,aAAmB;EAE5B;EAEAyG,kBAAuC;AACrCvG,0BAAAA,WAAU,KAAKlB,QAAQ,oBAAA;;;;;;;;;AAEvB,QAAI0H;AACJ,UAAMH,YAAY,IAAII,mEACpB;MACE3H,QAAQ,KAAKA;IACf,GACA;MACE4H,oBAAoB,OAAOC,MAAMC,iBAA6B;AAC5D,cAAM,KAAKjI,WAAW6B,KAAI;AAc1BX,wBAAAA,KAAI,sBAAsB;UAAE4F,IAAIkB,KAAKlB;UAAIoB,YAAY,KAAK/H;UAAQ8H,cAAcA,aAAaE,MAAK;QAAG,GAAA;;;;;;AACrG,YAAI,CAAC,KAAKX,YAAYxF,IAAIgG,KAAKlB,EAAE,GAAG;AAClCe,qBAAWG;AAEX,eAAKR,YAAYzE,IAAIiF,KAAKlB,IAAIY,SAAAA;AAE9BxG,0BAAAA,KAAI,kBAAkB;YAAE4F,IAAIkB,KAAKlB;YAAIoB,YAAY,KAAK/H;YAAQ8H,cAAcA,aAAaE,MAAK;UAAG,GAAA;;;;;;AACjG,eAAK5G,KAAK,kBAAkB;;YAE1BnB,cAAc;cACZgI,gBAAgBH,aAAaE,MAAK;YACpC;YACAhI,QAAQ6H,KAAKlB;UACf,CAAA;QACF;MACF;MACAuB,eAAe,OAAO,EAAEV,QAAO,MAAE;AAC/B,YAAI,CAACE,UAAU;AACb;QACF;AACA,cAAMtH,UAAU6G,uBAAAA,KAAKpB,OAAO2B,OAAAA;AAE5B,aAAKpG,KAAK,WAAWhB,OAAAA;MACvB;MACA+H,SAAS,YAAA;AACP,YAAI,CAACT,UAAU;AACb;QACF;AACA,aAAKtG,KAAK,qBAAqB;UAC7BpB,QAAQ0H,SAASf;QACnB,CAAA;AACA,aAAKU,YAAYhE,OAAOqE,SAASf,EAAE;MACrC;IACF,CAAA;AAEF,WAAOY;EACT;AACF;AE/FO,IAAMa,0BAAN,MAAMA;EAKX3E,YAA6B4E,YAAuB;SAAvBA,aAAAA;SAFrBC,SAA8B;EAEe;EAErD,MAAM3E,KAAKqB,KAAkD;AAC3D,QAAI,KAAKsD,WAAW,UAAU;AAC5B,aAAOzE;IACT;AACA,UAAM0E,WAAW,KAAKC,aAAaxD,GAAAA;AACnC,UAAMyD,OAAO,KAAKJ,WAAWK,gBAAgBH,QAAAA;AAC7C,UAAM,EAAEI,KAAI,IAAK,MAAMF,KAAKG,KAAI;AAChC,QAAI,CAACD,QAAQA,SAAS,GAAG;AACvB,aAAO9E;IACT;AACA,UAAMgF,SAAS,MAAMJ,KAAKzF,KAAK,GAAG2F,IAAAA;AAClC,eAAOG,4BAAcD,MAAAA;EACvB;EAEA,MAAM5E,KAAKe,KAAiBK,MAAiC;AAC3D,QAAI,KAAKiD,WAAW,UAAU;AAC5B,aAAOzE;IACT;AACA,UAAM0E,WAAW,KAAKC,aAAaxD,GAAAA;AACnC,UAAMyD,OAAO,KAAKJ,WAAWK,gBAAgBH,QAAAA;AAC7C,UAAME,KAAK9H,MAAM,OAAGoI,4BAAc1D,IAAAA,CAAAA;AAClC,UAAMoD,KAAKO,WAAW3D,KAAK4D,MAAM;AAEjC,UAAMR,KAAKS,QAAK;EAClB;EAEA,MAAMvE,OAAOK,KAAgC;AAC3C,QAAI,KAAKsD,WAAW,UAAU;AAC5B,aAAOzE;IACT;AAEA,UAAM0E,WAAW,KAAKC,aAAaxD,GAAAA;AACnC,UAAMyD,OAAO,KAAKJ,WAAWK,gBAAgBH,QAAAA;AAC7C,UAAME,KAAKU,QAAO;EACpB;EAEA,MAAMtE,UAAUC,WAAyC;AACvD,QAAI,KAAKwD,WAAW,UAAU;AAC5B,aAAO,CAAA;IACT;AACA,UAAMC,WAAW,KAAKC,aAAa1D,SAAAA;AACnC,UAAMsE,UAAU,MAAM,KAAKf,WAAWgB,KAAI;AAC1C,WAAOC,QAAQC,IACbH,QACGI,OAAO,CAACtG,UAAUA,MAAMuG,WAAWlB,QAAAA,CAAAA,EACnC9C,IAAI,OAAOvC,UAAAA;AACV,YAAMuF,OAAO,KAAKJ,WAAWK,gBAAgBxF,KAAAA;AAC7C,YAAM,EAAEyF,KAAI,IAAK,MAAMF,KAAKG,KAAI;AAChC,YAAMC,SAAS,MAAMJ,KAAKzF,KAAK,GAAG2F,IAAAA;AAClC,aAAO;QACL3D,KAAK,KAAK0E,oBAAoBxG,KAAAA;QAC9BmC,UAAMyD,4BAAcD,MAAAA;MACtB;IACF,CAAA,CAAA;EAEN;EAEA,MAAMvD,YAAYR,WAAsC;AACtD,QAAI,KAAKwD,WAAW,UAAU;AAC5B,aAAOzE;IACT;AACA,UAAM0E,WAAW,KAAKC,aAAa1D,SAAAA;AACnC,UAAMsE,UAAU,MAAM,KAAKf,WAAWgB,KAAI;AAC1C,UAAMC,QAAQC,IACZH,QACGI,OAAO,CAACtG,UAAUA,MAAMuG,WAAWlB,QAAAA,CAAAA,EACnC9C,IAAI,OAAOvC,UAAAA;AACV,YAAMuF,OAAO,KAAKJ,WAAWK,gBAAgBxF,KAAAA;AAC7C,YAAMuF,KAAKU,QAAO;IACpB,CAAA,CAAA;EAEN;EAEA,MAAM7H,QAAuB;AAC3B,SAAKgH,SAAS;EAChB;EAEQE,aAAaxD,KAAyB;AAC5C,WAAOA,IAAIS,IAAI,CAACC,MAAMA,EAAEC,WAAW,KAAK,KAAA,EAAOA,WAAW,KAAK,KAAA,CAAA,EAAQC,KAAK,GAAA;EAC9E;EAEQ8D,oBAAoBnB,UAA8B;AACxD,WAAOA,SAASxC,MAAM,GAAA,EAAKN,IAAI,CAACC,MAAMA,EAAEC,WAAW,OAAO,GAAA,EAAKA,WAAW,OAAO,GAAA,CAAA;EACnF;AACF;;ADzFO,IAAMgE,iBAAiB,OAAO,EAAE7F,IAAI8F,UAAS,MAA4C;AAE9F,QAAMC,aAAa,CAAE,MAAM/F,GACxBmB,SAAiC;IAChC,GAAGlB;EACL,CAAA,EACCgD,KAAI;AAEP,MAAI,CAAC8C,YAAY;AACf;EACF;AAEA,QAAMC,oBACJF,UAAUG,SAASC,yCAAYC,MAC3B,IAAIC,gEAAwBN,UAAUtF,MAAM,MAAA,IAC5C,IAAI8D,wBAAwBwB,SAAAA;AAElC,QAAMO,SAAS,MAAML,kBAAkBjF,UAAU,CAAA,CAAE;AACnD,MAAIsF,OAAOlB,WAAW,GAAG;AACvB;EACF;AACA,QAAM9E,QAAQL,GAAGK,MAAK;AACtBpD,cAAAA,IAAI8G,KAAK,uCAAuC;IAAEsC,QAAQA,OAAOlB;EAAO,GAAA;;;;;;AACxE,aAAW,EAAEjE,KAAKK,KAAI,KAAM,MAAMyE,kBAAkBjF,UAAU,CAAA,CAAE,GAAG;AACjEQ,YAAQlB,MAAMI,IAA4BS,KAAKK,MAAM;MAAE,GAAGtB;IAAgB,CAAA;EAC5E;AACA,QAAMI,MAAMxD,MAAK;AACnB;;;;;;;;;;;;ALeO,IAAMyJ,gBAAN,MAAMA;EAsBX3G,YAAY,EAAEmG,WAAW9F,IAAIuG,mBAAkB,GAAyB;AApBvDC,SAAAA,OAAO,IAAIC,uBAAAA;AAGXC,SAAAA,sBAAsB,IAAInL,mBAAAA;8BAaL,IAAIoL,uBAA6CC,sBAAUC,IAAI;AAE9FC,SAAAA,iBAAiB,oBAAIpL,IAAAA;AAG1B,SAAK6I,aAAauB;AAClB,SAAKiB,MAAM/G;AACX,SAAKgH,sBAAsBT;EAC7B;EAEA,MAAMpJ,OAAO;AAEX,SAAKoH,cAAe,MAAMsB,eAAe;MAAE7F,IAAI,KAAK+G;MAAKjB,WAAW,KAAKvB;IAAW,CAAA;AACpF,SAAK0C,WAAW,IAAIxH,sBAAsB;MACxCO,IAAI,KAAK+G;MACTzG,WAAW;QACTC,YAAY,OAAOhC,WAAW,KAAK2I,YAAY3I,MAAAA;QAC/CqC,WAAW,YAAY,KAAKuG,WAAU;MACxC;IACF,CAAA;AACA,UAAM,KAAKF,SAAS9J,OAAI;AACxB,SAAKiK,UAAU,QAAQR,sBAAUS,OAAM,EAAGnD,MAAK,CAAA;AAE/C,SAAKoD,eAAe,IAAIhE,mBAAAA;AACxB,SAAKiE,iBAAiB,IAAIjF,wBAAAA;AAE1B,SAAKkF,QAAQ,IAAIC,2BAAK;MACpBvL,QAAQ,KAAKkL;MACb7J,SAAS;QAAC,KAAKgK;QAAgB,KAAKD;QAAc,KAAKZ;;MACvDgB,SAAS,KAAKT;;;MAIdU,aAAa,OAAOzL,QAAyB0L,eAAwB;AACnE,YAAI1L,OAAOyJ,WAAW,SAAA,GAAY;AAChC,iBAAO;QACT;AAEA,YAAI,CAACiC,YAAY;AACf,iBAAO;QACT;AAEA,cAAMzL,eAAe,KAAK0L,KAAKC,qBAAqB5L,MAAAA;AACpD,YAAKC,cAAsBgD,oBAAoB,sBAAsB;AACnE,iBAAO,KAAKuH,oBAAoBpI,gBAAgBpC,QAAQ;YAAE0L;UAAW,CAAA;QACvE;AAEA,cAAMG,MAAM,KAAKP,MAAMQ,QAAQJ,UAAAA,GAAaK,QAAAA;AAC5C,YAAI,CAACF,KAAK;AACR,gBAAMG,cAAc,KAAKpB,eAAe/I,IAAI,aAAa6J,UAAAA,EAAY;AACrE3K,yBAAAA,KAAI,0BAA0B;YAAEf;YAAQ0L;YAAYM;UAAY,GAAA;;;;;;AAChE,iBAAOA;QACT;AAEA,YAAI;AACF,gBAAMC,WAAWC,mBAAmBL,GAAAA;AACpC,cAAI,CAACI,UAAU;AACblL,2BAAAA,KAAI,8CAA8C;cAAEf;cAAQ0L;YAAW,GAAA;;;;;;AACvE,mBAAO;UACT;AAEA,gBAAMS,oBAAoB,KAAKC,mBAAmB7L,IAAImK,sBAAUjG,KAAKwH,QAAAA,CAAAA;AAGrE,gBAAMI,eAAgBpM,cAAsBgI;AAC5C,cAAI,CAACoE,cAAc;AACjBtL,2BAAAA,KAAI,+CAA+C;cAAEf;cAAQ0L;YAAW,GAAA;;;;;;AACxE,mBAAO;UACT;AACA,gBAAMY,YAAY5B,sBAAUjG,KAAK4H,YAAAA;AAEjC,gBAAME,eAAeJ,mBAAmBtK,IAAIyK,SAAAA,KAAc;AAC1DvL,yBAAAA,KAAI,sBAAsB;YACxByL,WAAW,KAAKtB;YAChBuB,YAAYzM;YACZ0L;YACAY;YACAL;YACAM;UACF,GAAA;;;;;;AACA,iBAAOA;QACT,SAAS1L,KAAK;AACZE,qBAAAA,IAAIH,MAAMC,KAAAA,QAAAA;;;;;;AACV,iBAAO;QACT;MACF;IACF,CAAA;AACA,SAAKwK,eAAe9E,MAAK;AACzB,SAAK6E,aAAa7E,MAAK;AACvB,UAAM,KAAKiE,oBAAoBvJ,KAAI;AAEnC,UAAM,KAAKoK,eAAe5J,cAAa;AACvC,UAAM,KAAK+I,oBAAoB/I,cAAa;EAC9C;EAEA,MAAMH,QAAQ;AACZ,UAAM,KAAKyJ,SAASzJ,QAAK;AACzB,UAAM,KAAK+J,eAAe/J,MAAK;AAC/B,UAAM,KAAKkJ,oBAAoBlJ,MAAK;AACpC,UAAM,KAAKgJ,KAAKoC,QAAO;EACzB;EAEA,IAAIf,OAAa;AACf,WAAO,KAAKL;EACd;EAEA,MAAM1J,cAAcL,YAA4B;AAC9C,UAAM,KAAKiJ,oBAAoB5I,cAAcL,UAAAA;EAC/C;EAEA,MAAMY,iBAAiBZ,YAA4B;AACjD,UAAM,KAAKiJ,oBAAoBrI,iBAAiBZ,UAAAA;EAClD;EAEA,MAAcyJ,YAAY,EAAE1G,MAAMH,MAAK,GAAsB;AAC3D,UAAMwI,SAAS,KAAKrB,MAAMQ,QAAQxH,KAAK,CAAA,CAAE;AACzC,QAAI,CAACqI,QAAQ;AACX;IACF;AACA,UAAMd,MAAMc,OAAOZ,QAAO;AAC1B,QAAI,CAACF,KAAK;AACR;IACF;AAEA,UAAMe,wBAAoBC,2BAAShB,GAAAA;AAEnC,UAAMiB,YAAYC,OAAOC,KAAKnB,IAAIoB,WAAW,CAAC,CAAA;AAC9C,UAAMC,aAAaJ,UAAUrH,IAAI,CAAC0H,aAAaC,yBAAQ5H,OAAO;MAAEkG,YAAYiB,OAAOjB;MAAYyB;IAAS,CAAA,CAAA;AACxG,UAAME,eAAe,IAAI5N,IAAIyN,WAAWzH,IAAI,CAACkB,OAAO;MAACA;MAAIiG;KAAkB,CAAA;AAC3E,SAAK9B,oBAAoBwC,UAAUD,cAAclJ,KAAAA;EACnD;;;;EAKA,MAAc8G,aAAa;AACzB,SAAKH,oBAAoByC,kBAAiB;EAC5C;EAGQC,iBAAiB;AACvB,eAAOC,uBAAU,KAAKnC,MAAMQ,SAAS,CAACa,YAAY;MAChDe,OAAOf,OAAOe;MACdC,QAAQ,CAAC,CAAChB,OAAOZ,QAAO;MACxB6B,OAAOjB,OAAOZ,QAAO,IAAK8B,iBAAAA,KAAUhB,SAASF,OAAOZ,QAAO,CAAA,IAAM;MACjE1G,MACEsH,OAAOZ,QAAO,SACd0B,uBAAUd,OAAOZ,QAAO,GAAI,CAAChJ,OAAOiC,QAAAA;AAClC,YAAI;AACF,kBAAQA,KAAAA;YACN,KAAK;YACL,KAAK;AACH,qBAAOjC;YACT,KAAK;AACH,qBAAOgK,OAAOC,KAAKjK,KAAAA;YACrB;AACE,qBAAO,GAAGA,KAAAA;UACd;QACF,SAASlC,KAAK;AACZ,iBAAO,GAAGA,GAAAA;QACZ;MACF,CAAA;IACJ,EAAA;EACF;EAGQiN,kBAAkB;AACxB,WAAO,KAAKxC,MAAMyC;EACpB;;;;EAKA,MACM7E,MAAM,EAAE8E,OAAM,GAAiC;AAEnD,UAAM1E,QAAQC,IACZyE,QAAQvI,IAAI,OAAO,EAAEmI,OAAOlC,WAAU,MAAE;AACtCxK,2BAAAA,WAAU0M,OAAO,gCAAA;;;;;;;;;AACjB,YAAMjB,SAAS,KAAKhB,KAAKG,QAAQJ,UAAAA,KAA6B,KAAKJ,MAAM2C,KAAKvC,UAAAA;AAC9E,YAAMwC,aAAavB,QAAQiB,KAAAA;IAC7B,CAAA,KAAM,CAAA,CAAE;AAGV,UAAM,KAAKtC,MAAMpC,MAAM8E,QAAQvI,IAAI,CAAC,EAAEiG,WAAU,MAAOA,UAAAA,CAAAA;EACzD;EAEAhF,SAASyH,SAAoD;AAC3D,WAAO,KAAK9C,eAAe3E,SAASyH,OAAAA;EACtC;EAEAjH,gBAAgBiH,SAAyC;AACvD,WAAO,KAAK9C,eAAenE,gBAAgBiH,OAAAA;EAC7C;EAEA,MAAMhH,cAAiC;AACrC,WAAO,KAAKkE,eAAelE,YAAW;EACxC;;;;EAMAM,kBAAuC;AACrC,WAAO,KAAK2D,aAAa3D,gBAAe;EAC1C;EAEA2G,gBAAgBnC,UAAqBK,WAAsB;AACzDvL,mBAAAA,KAAI,mBAAmB;MAAEkL;MAAUK;IAAU,GAAA;;;;;;AAC7C+B,gCAAW,KAAKjC,oBAAoBH,UAAU,MAAM,IAAIqC,uBAAW5D,sBAAUC,IAAI,CAAA,EAAG4D,IAAIjC,SAAAA;EAC1F;AACF;;EAzNGkC,qBAAM3G,KAAI;GAZAuC,cAAAA,WAAAA,WAAAA,MAAAA;;EA6JVoE,qBAAM3G,KAAK;IAAE4G,OAAO;EAAK,CAAA;GA7JfrE,cAAAA,WAAAA,kBAAAA,IAAAA;;EAuLVoE,qBAAM3G,KAAK;IAAE4G,OAAO;EAAK,CAAA;GAvLfrE,cAAAA,WAAAA,mBAAAA,IAAAA;;EA+LVoE,qBAAME,KAAK;IAAEC,uBAAuB;EAAK,CAAA;GA/L/BvE,cAAAA,WAAAA,SAAAA,IAAAA;AAAAA,gBAAAA,cAAAA;EADZoE,qBAAMI,SAAQ;GACFxE,aAAAA;AAuON,IAAM8B,qBAAqB,CAACL,QAAAA;AAEjC,QAAMgD,cAAchD,IAAIiD,QAAQ7C,YAAYJ,IAAIkD;AAChD,MAAIF,eAAe,MAAM;AACvB,WAAO;EACT;AAEA,SAAOG,OAAOH,WAAAA;AAChB;AAEA,IAAMX,eAAe,OAAOvB,QAA6BiB,UAAAA;AACvD,QAAMjB,OAAOsC,UAAS;AACtB,QAAMC,mBAAmB,IAAI1P,IAAIoO,KAAAA;AAEjC,QAAMuB,mBAAMC,KAAuCzC,QAAQ,QAAA,EAAU0C,iBAAiB,MAAA;AAEpF,eAAWC,cAAcJ,iBAAiBK,OAAM,GAAI;AAClD,UAAIC,qBAAqB7C,OAAOZ,QAAO,GAAIuD,UAAAA,GAAa;AACtDJ,yBAAiB7L,OAAOiM,UAAAA;MAC1B;IACF;AAEA,QAAIJ,iBAAiBvG,SAAS,GAAG;AAC/B,aAAO;IACT;AACA,WAAO;EACT,CAAA;AACF;AAEA,IAAM6G,uBAAuB,CAAC3D,KAAeyD,eAAAA;AAC3C,SAAO,CAAC,KAACG,6BAAW5D,GAAAA,EAAK6D,gBAAgBJ,UAAAA;AAC3C;;;;;;;;;;;;AOvRO,IAAMK,8BAAN,MAAMA;EAcXlM,YACmBmM,WACAtE,OACjB;SAFiBsE,YAAAA;SACAtE,QAAAA;SAfXuE,sBAAkD;SAIzCC,yBAAyB,oBAAIrQ,IAAAA;SAK7BsQ,8BAA8B,oBAAIvQ,IAAAA;SAEnCwQ,yBAAyB,IAAIb,cAAAA,MAAAA;EAK1C;EAEHc,gBAAuC;AACrC,WAAO,KAAKJ,uBAAuB,OAC/B;MAAC,KAAKA;SAAwB,IAAIrQ,IAAI,KAAKsQ,uBAAuBP,OAAM,CAAA;QACxE,CAAA;EACN;EAEA,MACaW,uBAAuBC,KAAcC,YAAuC;AACvF,QAAI,KAAKP,uBAAuB,MAAM;AACpC;IACF;AACA,QAAI,CAACO,WAAWC,SAAS;AACvBtP,kBAAAA,IAAIuP,MAAM,mCAAmC;QAAErE,UAAU,KAAK2D;MAAU,GAAA;;;;;;AACxE,WAAKW,qCAAqCJ,GAAAA;IAC5C,OAAO;AACL,YAAMK,oBAAoB,MAAM,KAAKC,eAAeN,KAAKC,WAAWC,OAAO;AAC3E,YAAMxE,MAAM2E,kBAAkBzE,QAAO;AACrC7K,4BAAAA,WAAU2K,KAAAA,QAAAA;;;;;;;;;AACV,UAAIA,IAAIiD,UAAU,MAAM;AACtB,aAAK4B,eAAeF,iBAAAA;MACtB;AACA,WAAKX,sBAAsBW;IAC7B;EACF;EAEOG,mBAAmBC,gBAAmC;AAC3D,UAAM9D,YAAY+D,MAAMC,QAAQF,cAAAA,IAAkBA,iBAAiB;MAACA;;AACpE,QAAIG,gBAAgB;AACpB,UAAMC,aAAgC,CAAC;AACvC,eAAW7D,YAAYL,WAAW;AAChC5L,4BAAAA,WAAU,KAAK2O,qBAAmB,QAAA;;;;;;;;;AAClC,UAAI,KAAKC,uBAAuBjO,IAAIsL,QAAAA,KAAa,KAAK4C,4BAA4BlO,IAAIsL,QAAAA,GAAW;AAC/F;MACF;AACA,YAAM8D,eAAe,KAAKpB,oBAAoB9D,QAAO;AACrD7K,4BAAAA,WAAU+P,cAAAA,QAAAA;;;;;;;;;AACV,YAAMC,eAAeD,aAAaE,SAAS,CAAC,GAAGhE,QAAAA;AAC/C,UAAI+D,eAAe,MAAM;AACvB,aAAKnB,4BAA4BxB,IAAIpB,QAAAA;AACrCpM,oBAAAA,IAAI8G,KAAK,sDAAsD;UAAEsF;QAAS,GAAA;;;;;;MAC5E,OAAO;AACL6D,mBAAW7D,QAAAA,IAAY+D;AACvBH,wBAAgB;MAClB;IACF;AACA,QAAIA,eAAe;AACjB,WAAKK,mBAAmBJ,UAAAA;IAC1B;EACF;EAEOK,qBAAqBF,OAA2B;AACrD,QAAI,CAACA,OAAO;AACV;IACF;AACA,UAAMG,oBAAoBvE,OAAO3D,QAAQ+H,KAAAA,EAAO3H,OAAO,CAAC,CAAC2D,QAAAA,MACvD,KAAK4C,4BAA4BlO,IAAIsL,QAAAA,CAAAA;AAEvC,SAAKiE,mBAAmBrE,OAAOwE,YAAYD,iBAAAA,CAAAA;AAC3CA,sBAAkB7K,QAAQ,CAAC,CAAC0G,QAAAA,MAAc,KAAK4C,4BAA4B1M,OAAO8J,QAAAA,CAAAA;EACpF;EAEOqE,wBAA6C;AAClDtQ,0BAAAA,WAAU,KAAK2O,qBAAmB,QAAA;;;;;;;;;AAClC,WAAO,KAAKA;EACd;EAEO4B,wBAAwBtE,UAAuC;AACpEjM,0BAAAA,WAAU,KAAK2O,qBAAmB,QAAA;;;;;;;;;AAClC,UAAM6B,iBAAiB,KAAKpG,MAAMqG,OAAM;AACxC,SAAKjB,eAAegB,cAAAA;AACpB,SAAKE,wBAAwBF,gBAAgBvE,QAAAA;AAC7C,SAAK0C,oBAAoBgC,OAAO,CAACC,WAAAA;AAC/BA,aAAOX,UAAU,CAAC;AAClBW,aAAOX,MAAMhE,QAAAA,IAAYuE,eAAeK;IAC1C,CAAA;AACA,WAAOL;EACT;EAEOE,wBAAwBjF,QAA6BQ,UAAkB;AAC5E,SAAK2C,uBAAuBlN,IAAIuK,UAAUR,MAAAA;EAC5C;EAEOqF,wBAAkC;AACvC,UAAMC,qBAAqB;SAAI,KAAKnC,uBAAuB9C,KAAI;;AAC/D,SAAK8C,uBAAuBtO,MAAK;AACjC,SAAKqO,sBAAsB;AAC3B,WAAOoC;EACT;EAEQb,mBAAmBD,OAA2B;AACpD,QAAI,CAACA,OAAO;AACV;IACF;AACA,eAAW,CAAChE,UAAU+E,YAAAA,KAAiBnF,OAAO3D,QAAQ+H,KAAAA,GAAQ;AAC5D,YAAMgB,UAAU;QAAEhF;QAAU+E;MAAa;AACzC,YAAME,uBAAuB,KAAKtC,uBAAuBvP,IAAI4M,QAAAA;AAC7D,UAAIiF,wBAAwB,QAAQA,qBAAqBL,QAAQG,cAAc;AAC7EnR,oBAAAA,IAAIsR,KAAK,qEAAqE;UAC5E,GAAGF;UACHG,mBAAmBF,qBAAqBL;QAC1C,GAAA;;;;;;AACA;MACF;AACA,UAAIK,sBAAsBL,QAAQG,cAAc;AAC9CnR,oBAAAA,IAAIsR,KAAK,sCAAsCF,SAAAA;;;;;;AAC/C;MACF;AACA,YAAMxF,SAAS,KAAKrB,MAAM2C,KAAeiE,YAAAA;AACzCnR,kBAAAA,IAAIwR,MAAM,8BAA8BJ,SAAAA;;;;;;AACxC,WAAKrC,uBAAuBlN,IAAIuK,UAAUR,MAAAA;AAC1C,WAAK,KAAK6F,4BAA4B7F,QAAQQ,QAAAA;IAChD;EACF;EAEA,MAAcsD,eAAeN,KAAc4B,KAAa;AACtD,UAAMU,YAAY,KAAKnH,MAAM2C,KAAe8D,GAAAA;AAC5C,WAAO,MAAM;AACX,UAAI;AACF,kBAAMW,+BAAiB,KAAO,iDAAiD,YAAA;AAC7E,oBAAMC,mCAAkBxC,KAAKsC,UAAUxD,UAAS,CAAA;QAClD,CAAA;AACA;MACF,SAASpO,KAAK;AACZ,YAAI,GAAGA,GAAAA,GAAM+R,SAAS,SAAA,GAAY;AAChC7R,sBAAAA,IAAI8G,KAAK,cAAc;YAAElB,IAAI8L,UAAU/G;YAAYgC,OAAO+E,UAAU/E;UAAM,GAAA;;;;;;AAC1E;QACF;AAEA,cAAM7M;MACR;IACF;AAEA,QAAI4R,UAAU/E,UAAU,eAAe;AACrC,YAAM,IAAIjN,MAAM,mCAAA;IAClB;AAEA,WAAOgS;EACT;EAEQlC,qCAAqCJ,KAAc;AACzD,UAAMsC,YAAY,KAAKnH,MAAMqG,OAAM;AACnC,SAAK9B,sBAAsB4C;AAC3BtC,QAAI0C,UAAU,MAAA;AACZJ,gBAAUpP,OAAM;AAChB,WAAKwM,sBAAsB;IAC7B,CAAA;EACF;EAEQa,eAAe/D,QAA6B;AAClDA,WAAOkF,OAAO,CAACC,WAAAA;AACbA,aAAOhD,WAAW;QAAE7C,UAAU,KAAK2D,UAAU5H,MAAK;MAAG;AACrD8J,aAAOhD,OAAO7C,WAAW,KAAK2D,UAAU5H,MAAK;IAC/C,CAAA;EACF;EAEA,MAAcwK,4BAA4B7F,QAA6BQ,UAAkB;AACvF,QAAI;AACF,YAAMR,OAAOd,IAAI;QAAC;OAAQ;AAC1B,YAAMsG,UAAU;QAAEhF;QAAU2F,QAAQnG,OAAOoF;MAAI;AAC/C,UAAI,KAAK/B,uBAAuB+C,cAAa,MAAO,GAAG;AACrDhS,oBAAAA,IAAI8G,KAAK,oDAAoDsK,SAAAA;;;;;;AAC7D;MACF;AACA,YAAMa,kBAAkB,KAAKlD,uBAAuBvP,IAAI4M,QAAAA;AACxD,UAAI6F,iBAAiBjB,QAAQpF,OAAOoF,KAAK;AACvChR,oBAAAA,IAAIsR,KAAK,sEAAsEF,SAAAA;;;;;;AAC/E;MACF;AACA,WAAKnC,uBAAuB5O,KAAK;QAAEuL;QAAQQ;MAAS,CAAA;IACtD,SAAStM,KAAK;AACZ,YAAMoS,qBAAqB,KAAKjD,uBAAuB+C,cAAa,IAAK;AACzEhS,kBAAAA,IAAIsR,KAAK,6BAA6B;QACpClF;QACA+E,cAAcvF,OAAOoF;QACrBmB,cAAcD;QACdpS;MACF,GAAA;;;;;;AACA,UAAIoS,oBAAoB;AACtB,cAAM,KAAKT,4BAA4B7F,QAAQQ,QAAAA;MACjD;IACF;EACF;AACF;;EAhLGqB,gBAAAA,MAAME,KAAK;IAAEC,uBAAuB;EAAK,CAAA;GAzB/BgB,4BAAAA,WAAAA,0BAAAA,IAAAA;AAAAA,8BAAAA,cAAAA;EADZnB,gBAAAA,MAAMI,SAAQ;GACFe,2BAAAA;",
|
|
6
6
|
"names": ["import_async", "import_automerge_repo", "import_context", "import_invariant", "import_log", "import_util", "import_tracing", "EchoNetworkAdapter", "NetworkAdapter", "_replicators", "Set", "Map", "_lifecycleState", "LifecycleState", "CLOSED", "_connected", "Trigger", "connect", "peerId", "peerMetadata", "wake", "send", "message", "connectionEntry", "_connections", "get", "targetId", "Error", "writer", "write", "catch", "err", "isOpen", "log", "disconnect", "open", "invariant", "OPEN", "emit", "network", "close", "replicator", "clear", "whenConnected", "wait", "timeout", "addReplicator", "has", "onConnectionOpen", "_onConnectionOpen", "bind", "onConnectionClosed", "_onConnectionClosed", "removeReplicator", "shouldAdvertize", "params", "connection", "reader", "readable", "getReader", "writable", "getWriter", "set", "queueMicrotask", "done", "value", "read", "dxos_peerSource", "entry", "cancel", "abort", "delete", "synchronized", "LevelDBStorageAdapter", "Resource", "constructor", "_params", "load", "keyArray", "undefined", "db", "encodingOptions", "isLevelDbNotFoundError", "save", "binary", "batch", "callbacks", "beforeSave", "path", "put", "Buffer", "from", "afterSave", "remove", "del", "loadRange", "keyPrefix", "result", "key", "iterator", "gte", "lte", "push", "data", "removeRange", "keyEncoder", "encode", "map", "k", "replaceAll", "join", "decode", "toString", "split", "format", "keyEncoding", "valueEncoding", "code", "LocalHostNetworkAdapter", "_peers", "_isConnected", "ready", "peer", "forEach", "syncRepo", "id", "syncMessage", "_getPeerId", "Stream", "next", "connected", "cbor", "sendSyncMessage", "getHostInfo", "MeshNetworkAdapter", "_extensions", "receiverId", "extension", "payload", "createExtension", "peerInfo", "AutomergeReplicator", "onStartReplication", "info", "remotePeerId", "thisPeerId", "toHex", "dxos_deviceKey", "onSyncMessage", "onClose", "AutomergeStorageAdapter", "_directory", "_state", "filename", "_getFilename", "file", "getOrCreateFile", "size", "stat", "buffer", "bufferToArray", "arrayToBuffer", "truncate", "length", "flush", "destroy", "entries", "list", "Promise", "all", "filter", "startsWith", "_getKeyFromFilename", "levelMigration", "directory", "isNewLevel", "oldStorageAdapter", "type", "StorageType", "IDB", "IndexedDBStorageAdapter", "chunks", "AutomergeHost", "indexMetadataStore", "_ctx", "Context", "_echoNetworkAdapter", "ComplexMap", "PublicKey", "hash", "_requestedDocs", "_db", "_indexMetadataStore", "_storage", "_beforeSave", "_afterSave", "_peerId", "random", "_meshNetwork", "_clientNetwork", "_repo", "Repo", "storage", "sharePolicy", "documentId", "repo", "peerMetadataByPeerId", "doc", "handles", "docSync", "isRequested", "spaceKey", "getSpaceKeyFromDoc", "authorizedDevices", "_authorizedDevices", "deviceKeyHex", "deviceKey", "isAuthorized", "localPeer", "remotePeer", "dispose", "handle", "lastAvailableHash", "getHeads", "objectIds", "Object", "keys", "objects", "encodedIds", "objectId", "idCodec", "idToLastHash", "markDirty", "notifyMarkedDirty", "_automergeDocs", "mapValues", "state", "hasDoc", "heads", "automerge", "_automergePeers", "peers", "states", "find", "waitForHeads", "request", "authorizeDevice", "defaultMap", "ComplexSet", "add", "trace", "depth", "span", "showInBrowserTimeline", "resource", "rawSpaceKey", "access", "experimental_spaceKey", "String", "whenReady", "unavailableHeads", "Event", "wrap", "waitForCondition", "changeHash", "values", "changeIsPresentInDoc", "getBackend", "getChangeByHash", "AutomergeDocumentLoaderImpl", "_spaceKey", "_spaceRootDocHandle", "_objectDocumentHandles", "_objectsPendingDocumentLoad", "onObjectDocumentLoaded", "getAllHandles", "loadSpaceRootDocHandle", "ctx", "spaceState", "rootUrl", "error", "_createContextBoundSpaceRootDocument", "existingDocHandle", "_initDocHandle", "_initDocAccess", "loadObjectDocument", "objectIdOrMany", "Array", "isArray", "hasUrlsToLoad", "urlsToLoad", "spaceRootDoc", "documentUrl", "links", "_loadLinkedObjects", "onObjectLinksUpdated", "linksAwaitingLoad", "fromEntries", "getSpaceRootDocHandle", "createDocumentForObject", "spaceDocHandle", "create", "onObjectBoundToDocument", "change", "newDoc", "url", "clearHandleReferences", "objectsWithHandles", "automergeUrl", "logMeta", "objectDocumentHandle", "warn", "actualDocumentUrl", "debug", "_createObjectOnDocumentLoad", "docHandle", "warnAfterTimeout", "cancelWithContext", "includes", "onDispose", "docUrl", "listenerCount", "objectDocHandle", "shouldRetryLoading", "retryLoading"]
|
|
7
7
|
}
|
|
@@ -2,14 +2,14 @@ import { Repo, type DocumentId } from '@dxos/automerge/automerge-repo';
|
|
|
2
2
|
import { type Stream } from '@dxos/codec-protobuf';
|
|
3
3
|
import { type IndexMetadataStore } from '@dxos/indexing';
|
|
4
4
|
import { PublicKey } from '@dxos/keys';
|
|
5
|
-
import { type
|
|
5
|
+
import { type SublevelDB } from '@dxos/kv-store';
|
|
6
6
|
import { type FlushRequest, type HostInfo, type SyncRepoRequest, type SyncRepoResponse } from '@dxos/protocols/proto/dxos/echo/service';
|
|
7
7
|
import { type Directory } from '@dxos/random-access-storage';
|
|
8
8
|
import { type AutomergeReplicator } from '@dxos/teleport-extension-automerge-replicator';
|
|
9
9
|
import { type EchoReplicator } from './echo-replicator';
|
|
10
10
|
export type { DocumentId };
|
|
11
11
|
export type AutomergeHostParams = {
|
|
12
|
-
db:
|
|
12
|
+
db: SublevelDB;
|
|
13
13
|
/**
|
|
14
14
|
* For migration purposes.
|
|
15
15
|
*/
|
|
@@ -1,10 +1,10 @@
|
|
|
1
1
|
import { type MixedEncoding } from 'level-transcoder';
|
|
2
2
|
import { type StorageAdapterInterface, type Chunk, type StorageKey } from '@dxos/automerge/automerge-repo';
|
|
3
3
|
import { Resource } from '@dxos/context';
|
|
4
|
-
import { type BatchLevel, type
|
|
4
|
+
import { type BatchLevel, type SublevelDB } from '@dxos/kv-store';
|
|
5
5
|
import { type MaybePromise } from '@dxos/util';
|
|
6
6
|
export type LevelDBStorageAdapterParams = {
|
|
7
|
-
db:
|
|
7
|
+
db: SublevelDB;
|
|
8
8
|
callbacks?: StorageCallbacks;
|
|
9
9
|
};
|
|
10
10
|
export type BeforeSaveParams = {
|
|
@@ -1,7 +1,7 @@
|
|
|
1
|
-
import { type
|
|
1
|
+
import { type SublevelDB } from '@dxos/kv-store';
|
|
2
2
|
import { type Directory } from '@dxos/random-access-storage';
|
|
3
3
|
export declare const levelMigration: ({ db, directory }: {
|
|
4
|
-
db:
|
|
4
|
+
db: SublevelDB;
|
|
5
5
|
directory: Directory;
|
|
6
6
|
}) => Promise<void>;
|
|
7
7
|
//# sourceMappingURL=migrations.d.ts.map
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@dxos/echo-pipeline",
|
|
3
|
-
"version": "0.5.3-main.
|
|
3
|
+
"version": "0.5.3-main.f752aaa",
|
|
4
4
|
"description": "ECHO database.",
|
|
5
5
|
"homepage": "https://dxos.org",
|
|
6
6
|
"bugs": "https://github.com/dxos/dxos/issues",
|
|
@@ -39,38 +39,38 @@
|
|
|
39
39
|
"crc-32": "^1.2.2",
|
|
40
40
|
"level": "^8.0.1",
|
|
41
41
|
"level-transcoder": "^1.0.1",
|
|
42
|
-
"@dxos/
|
|
43
|
-
"@dxos/
|
|
44
|
-
"@dxos/codec-protobuf": "0.5.3-main.
|
|
45
|
-
"@dxos/
|
|
46
|
-
"@dxos/
|
|
47
|
-
"@dxos/
|
|
48
|
-
"@dxos/
|
|
49
|
-
"@dxos/
|
|
50
|
-
"@dxos/echo-
|
|
51
|
-
"@dxos/
|
|
52
|
-
"@dxos/
|
|
53
|
-
"@dxos/
|
|
54
|
-
"@dxos/
|
|
55
|
-
"@dxos/
|
|
56
|
-
"@dxos/keys": "0.5.3-main.
|
|
57
|
-
"@dxos/kv-store": "0.5.3-main.
|
|
58
|
-
"@dxos/
|
|
59
|
-
"@dxos/
|
|
60
|
-
"@dxos/
|
|
61
|
-
"@dxos/
|
|
62
|
-
"@dxos/
|
|
63
|
-
"@dxos/
|
|
64
|
-
"@dxos/
|
|
65
|
-
"@dxos/
|
|
66
|
-
"@dxos/teleport-extension-automerge-replicator": "0.5.3-main.
|
|
67
|
-
"@dxos/teleport-extension-gossip": "0.5.3-main.
|
|
68
|
-
"@dxos/teleport-extension-object-sync": "0.5.3-main.
|
|
69
|
-
"@dxos/teleport-extension-replicator": "0.5.3-main.
|
|
70
|
-
"@dxos/
|
|
71
|
-
"@dxos/
|
|
72
|
-
"@dxos/
|
|
73
|
-
"@dxos/
|
|
42
|
+
"@dxos/async": "0.5.3-main.f752aaa",
|
|
43
|
+
"@dxos/automerge": "0.5.3-main.f752aaa",
|
|
44
|
+
"@dxos/codec-protobuf": "0.5.3-main.f752aaa",
|
|
45
|
+
"@dxos/credentials": "0.5.3-main.f752aaa",
|
|
46
|
+
"@dxos/context": "0.5.3-main.f752aaa",
|
|
47
|
+
"@dxos/debug": "0.5.3-main.f752aaa",
|
|
48
|
+
"@dxos/echo-protocol": "0.5.3-main.f752aaa",
|
|
49
|
+
"@dxos/crypto": "0.5.3-main.f752aaa",
|
|
50
|
+
"@dxos/echo-schema": "0.5.3-main.f752aaa",
|
|
51
|
+
"@dxos/hypercore": "0.5.3-main.f752aaa",
|
|
52
|
+
"@dxos/indexing": "0.5.3-main.f752aaa",
|
|
53
|
+
"@dxos/feed-store": "0.5.3-main.f752aaa",
|
|
54
|
+
"@dxos/keyring": "0.5.3-main.f752aaa",
|
|
55
|
+
"@dxos/invariant": "0.5.3-main.f752aaa",
|
|
56
|
+
"@dxos/keys": "0.5.3-main.f752aaa",
|
|
57
|
+
"@dxos/kv-store": "0.5.3-main.f752aaa",
|
|
58
|
+
"@dxos/log": "0.5.3-main.f752aaa",
|
|
59
|
+
"@dxos/messaging": "0.5.3-main.f752aaa",
|
|
60
|
+
"@dxos/network-manager": "0.5.3-main.f752aaa",
|
|
61
|
+
"@dxos/protocols": "0.5.3-main.f752aaa",
|
|
62
|
+
"@dxos/node-std": "0.5.3-main.f752aaa",
|
|
63
|
+
"@dxos/rpc": "0.5.3-main.f752aaa",
|
|
64
|
+
"@dxos/teleport": "0.5.3-main.f752aaa",
|
|
65
|
+
"@dxos/random-access-storage": "0.5.3-main.f752aaa",
|
|
66
|
+
"@dxos/teleport-extension-automerge-replicator": "0.5.3-main.f752aaa",
|
|
67
|
+
"@dxos/teleport-extension-gossip": "0.5.3-main.f752aaa",
|
|
68
|
+
"@dxos/teleport-extension-object-sync": "0.5.3-main.f752aaa",
|
|
69
|
+
"@dxos/teleport-extension-replicator": "0.5.3-main.f752aaa",
|
|
70
|
+
"@dxos/timeframe": "0.5.3-main.f752aaa",
|
|
71
|
+
"@dxos/tracing": "0.5.3-main.f752aaa",
|
|
72
|
+
"@dxos/typings": "0.5.3-main.f752aaa",
|
|
73
|
+
"@dxos/util": "0.5.3-main.f752aaa"
|
|
74
74
|
},
|
|
75
75
|
"devDependencies": {
|
|
76
76
|
"fast-check": "^3.15.1",
|
|
@@ -18,7 +18,7 @@ import { type SpaceDoc } from '@dxos/echo-protocol';
|
|
|
18
18
|
import { type IndexMetadataStore } from '@dxos/indexing';
|
|
19
19
|
import { invariant } from '@dxos/invariant';
|
|
20
20
|
import { PublicKey } from '@dxos/keys';
|
|
21
|
-
import { type
|
|
21
|
+
import { type SublevelDB } from '@dxos/kv-store';
|
|
22
22
|
import { log } from '@dxos/log';
|
|
23
23
|
import { idCodec } from '@dxos/protocols';
|
|
24
24
|
import {
|
|
@@ -43,7 +43,7 @@ import { levelMigration } from './migrations';
|
|
|
43
43
|
export type { DocumentId };
|
|
44
44
|
|
|
45
45
|
export type AutomergeHostParams = {
|
|
46
|
-
db:
|
|
46
|
+
db: SublevelDB;
|
|
47
47
|
/**
|
|
48
48
|
* For migration purposes.
|
|
49
49
|
*/
|
|
@@ -57,7 +57,7 @@ export class AutomergeHost {
|
|
|
57
57
|
private readonly _indexMetadataStore: IndexMetadataStore;
|
|
58
58
|
private readonly _ctx = new Context();
|
|
59
59
|
private readonly _directory?: Directory;
|
|
60
|
-
private readonly _db:
|
|
60
|
+
private readonly _db: SublevelDB;
|
|
61
61
|
private readonly _echoNetworkAdapter = new EchoNetworkAdapter();
|
|
62
62
|
|
|
63
63
|
private _repo!: Repo;
|
|
@@ -6,11 +6,11 @@ import { type MixedEncoding } from 'level-transcoder';
|
|
|
6
6
|
|
|
7
7
|
import { type StorageAdapterInterface, type Chunk, type StorageKey } from '@dxos/automerge/automerge-repo';
|
|
8
8
|
import { LifecycleState, Resource } from '@dxos/context';
|
|
9
|
-
import { type BatchLevel, type
|
|
9
|
+
import { type BatchLevel, type SublevelDB } from '@dxos/kv-store';
|
|
10
10
|
import { type MaybePromise } from '@dxos/util';
|
|
11
11
|
|
|
12
12
|
export type LevelDBStorageAdapterParams = {
|
|
13
|
-
db:
|
|
13
|
+
db: SublevelDB;
|
|
14
14
|
callbacks?: StorageCallbacks;
|
|
15
15
|
};
|
|
16
16
|
|
|
@@ -4,14 +4,14 @@
|
|
|
4
4
|
|
|
5
5
|
import { type StorageKey } from '@dxos/automerge/automerge-repo';
|
|
6
6
|
import { IndexedDBStorageAdapter } from '@dxos/automerge/automerge-repo-storage-indexeddb';
|
|
7
|
-
import { type
|
|
7
|
+
import { type SublevelDB } from '@dxos/kv-store';
|
|
8
8
|
import { log } from '@dxos/log';
|
|
9
9
|
import { StorageType, type Directory } from '@dxos/random-access-storage';
|
|
10
10
|
|
|
11
11
|
import { AutomergeStorageAdapter } from './automerge-storage-adapter';
|
|
12
12
|
import { encodingOptions } from './leveldb-storage-adapter';
|
|
13
13
|
|
|
14
|
-
export const levelMigration = async ({ db, directory }: { db:
|
|
14
|
+
export const levelMigration = async ({ db, directory }: { db: SublevelDB; directory: Directory }) => {
|
|
15
15
|
// Note: Make automigration from previous storage to leveldb here.
|
|
16
16
|
const isNewLevel = !(await db
|
|
17
17
|
.iterator<StorageKey, Uint8Array>({
|