UBERF-8259: Do not store system model into DB (#6716)

Signed-off-by: Andrey Sobolev <haiodo@gmail.com>
This commit is contained in:
Andrey Sobolev 2024-09-25 11:27:45 +07:00 committed by GitHub
parent 0da0d6e80a
commit 6fd8018baf
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
13 changed files with 117 additions and 64 deletions

2
.vscode/launch.json vendored
View File

@ -129,7 +129,7 @@
"MINIO_ACCESS_KEY": "minioadmin", "MINIO_ACCESS_KEY": "minioadmin",
"MINIO_SECRET_KEY": "minioadmin", "MINIO_SECRET_KEY": "minioadmin",
"MINIO_ENDPOINT": "localhost", "MINIO_ENDPOINT": "localhost",
"MODEL_VERSION": "v0.6.286" "MODEL_VERSION": "v0.6.287"
// "INIT_SCRIPT_URL": "https://raw.githubusercontent.com/hcengineering/init/main/script.yaml", // "INIT_SCRIPT_URL": "https://raw.githubusercontent.com/hcengineering/init/main/script.yaml",
// "INIT_WORKSPACE": "onboarding", // "INIT_WORKSPACE": "onboarding",
}, },

View File

@ -1503,7 +1503,7 @@ export function devTool (
.option('-w, --workspace <workspace>', 'Selected workspace only', '') .option('-w, --workspace <workspace>', 'Selected workspace only', '')
.option('-c, --concurrency <concurrency>', 'Number of documents being processed concurrently', '10') .option('-c, --concurrency <concurrency>', 'Number of documents being processed concurrently', '10')
.action(async (cmd: { workspace: string, concurrency: string }) => { .action(async (cmd: { workspace: string, concurrency: string }) => {
const { mongodbUri, dbUrl } = prepareTools() const { mongodbUri, dbUrl, txes } = prepareTools()
await withDatabase(mongodbUri, async (db, client) => { await withDatabase(mongodbUri, async (db, client) => {
await withStorage(mongodbUri, async (adapter) => { await withStorage(mongodbUri, async (adapter) => {
const workspaces = await listWorkspacesPure(db) const workspaces = await listWorkspacesPure(db)
@ -1521,7 +1521,7 @@ export function devTool (
workspaceUrl: workspace.workspaceUrl ?? '' workspaceUrl: workspace.workspaceUrl ?? ''
} }
const { pipeline } = await getServerPipeline(toolCtx, mongodbUri, dbUrl, wsUrl) const { pipeline } = await getServerPipeline(toolCtx, txes, mongodbUri, dbUrl, wsUrl)
await migrateMarkup(toolCtx, adapter, wsId, client, pipeline, parseInt(cmd.concurrency)) await migrateMarkup(toolCtx, adapter, wsId, client, pipeline, parseInt(cmd.concurrency))

View File

@ -15,7 +15,7 @@ import core, {
} from '@hcengineering/core' } from '@hcengineering/core'
import { getMongoClient, getWorkspaceDB } from '@hcengineering/mongo' import { getMongoClient, getWorkspaceDB } from '@hcengineering/mongo'
import { type Pipeline, type StorageAdapter } from '@hcengineering/server-core' import { type Pipeline, type StorageAdapter } from '@hcengineering/server-core'
import { connect, fetchModel } from '@hcengineering/server-tool' import { connect } from '@hcengineering/server-tool'
import { jsonToText, markupToYDoc } from '@hcengineering/text' import { jsonToText, markupToYDoc } from '@hcengineering/text'
import { type Db, type FindCursor, type MongoClient } from 'mongodb' import { type Db, type FindCursor, type MongoClient } from 'mongodb'
@ -123,7 +123,7 @@ export async function migrateMarkup (
pipeline: Pipeline, pipeline: Pipeline,
concurrency: number concurrency: number
): Promise<void> { ): Promise<void> {
const { hierarchy } = await fetchModel(ctx, pipeline) const hierarchy = pipeline.context.hierarchy
const workspaceDb = client.db(workspaceId.name) const workspaceDb = client.db(workspaceId.name)

View File

@ -57,6 +57,7 @@
"dotenv": "~16.0.0", "dotenv": "~16.0.0",
"@hcengineering/backup-service": "^0.6.0", "@hcengineering/backup-service": "^0.6.0",
"@hcengineering/analytics": "^0.6.0", "@hcengineering/analytics": "^0.6.0",
"@hcengineering/analytics-service": "^0.6.0" "@hcengineering/analytics-service": "^0.6.0",
"@hcengineering/model-all": "^0.6.0"
} }
} }

View File

@ -16,12 +16,19 @@
import { Analytics } from '@hcengineering/analytics' import { Analytics } from '@hcengineering/analytics'
import { configureAnalytics, SplitLogger } from '@hcengineering/analytics-service' import { configureAnalytics, SplitLogger } from '@hcengineering/analytics-service'
import { startBackup } from '@hcengineering/backup-service' import { startBackup } from '@hcengineering/backup-service'
import { MeasureMetricsContext, metricsToString, newMetrics } from '@hcengineering/core' import { MeasureMetricsContext, metricsToString, newMetrics, type Tx } from '@hcengineering/core'
import { type PipelineFactory } from '@hcengineering/server-core' import { type PipelineFactory } from '@hcengineering/server-core'
import { createBackupPipeline, getConfig } from '@hcengineering/server-pipeline' import { createBackupPipeline, getConfig } from '@hcengineering/server-pipeline'
import { writeFile } from 'fs/promises' import { writeFile } from 'fs/promises'
import { join } from 'path' import { join } from 'path'
import builder from '@hcengineering/model-all'
const enabled = (process.env.MODEL_ENABLED ?? '*').split(',').map((it) => it.trim())
const disabled = (process.env.MODEL_DISABLED ?? '').split(',').map((it) => it.trim())
const model = JSON.parse(JSON.stringify(builder(enabled, disabled).getTxes())) as Tx[]
const metricsContext = new MeasureMetricsContext( const metricsContext = new MeasureMetricsContext(
'backup', 'backup',
{}, {},
@ -58,7 +65,7 @@ const onClose = (): void => {
startBackup( startBackup(
metricsContext, metricsContext,
(mongoUrl, storageAdapter) => { (mongoUrl, storageAdapter) => {
const factory: PipelineFactory = createBackupPipeline(metricsContext, mongoUrl, { const factory: PipelineFactory = createBackupPipeline(metricsContext, mongoUrl, model, {
externalStorage: storageAdapter, externalStorage: storageAdapter,
usePassedCtx: true usePassedCtx: true
}) })

View File

@ -72,6 +72,7 @@
"@hcengineering/server-telegram": "^0.6.0", "@hcengineering/server-telegram": "^0.6.0",
"@hcengineering/pod-telegram-bot": "^0.6.0", "@hcengineering/pod-telegram-bot": "^0.6.0",
"@hcengineering/server-ai-bot": "^0.6.0", "@hcengineering/server-ai-bot": "^0.6.0",
"@hcengineering/server-ai-bot-resources": "^0.6.0" "@hcengineering/server-ai-bot-resources": "^0.6.0",
"@hcengineering/model-all": "^0.6.0"
} }
} }

View File

@ -14,7 +14,7 @@
// limitations under the License. // limitations under the License.
// //
import { type Branding, type BrandingMap, type WorkspaceIdWithUrl } from '@hcengineering/core' import { type Branding, type BrandingMap, type Tx, type WorkspaceIdWithUrl } from '@hcengineering/core'
import { buildStorageFromConfig, getMetricsContext } from '@hcengineering/server' import { buildStorageFromConfig, getMetricsContext } from '@hcengineering/server'
import { ClientSession, startSessionManager, type ServerFactory, type Session } from '@hcengineering/server' import { ClientSession, startSessionManager, type ServerFactory, type Session } from '@hcengineering/server'
@ -25,6 +25,13 @@ import { serverAiBotId } from '@hcengineering/server-ai-bot'
import { createAIBotAdapter } from '@hcengineering/server-ai-bot-resources' import { createAIBotAdapter } from '@hcengineering/server-ai-bot-resources'
import { createServerPipeline, registerServerPlugins, registerStringLoaders } from '@hcengineering/server-pipeline' import { createServerPipeline, registerServerPlugins, registerStringLoaders } from '@hcengineering/server-pipeline'
import builder from '@hcengineering/model-all'
const enabled = (process.env.MODEL_ENABLED ?? '*').split(',').map((it) => it.trim())
const disabled = (process.env.MODEL_DISABLED ?? '').split(',').map((it) => it.trim())
const model = JSON.parse(JSON.stringify(builder(enabled, disabled).getTxes())) as Tx[]
registerStringLoaders() registerStringLoaders()
/** /**
@ -64,6 +71,7 @@ export function start (
const pipelineFactory = createServerPipeline( const pipelineFactory = createServerPipeline(
metrics, metrics,
dbUrls, dbUrls,
model,
{ ...opt, externalStorage, adapterSecurity: rawDbUrl !== undefined }, { ...opt, externalStorage, adapterSecurity: rawDbUrl !== undefined },
{ {
serviceAdapters: { serviceAdapters: {

View File

@ -14,14 +14,22 @@
// //
import core, { import core, {
type Doc,
type LoadModelResponse, type LoadModelResponse,
type MeasureContext, type MeasureContext,
type Timestamp, type Timestamp,
type Tx, type Tx,
type TxCUD,
DOMAIN_TX DOMAIN_TX
} from '@hcengineering/core' } from '@hcengineering/core'
import { PlatformError, unknownError } from '@hcengineering/platform' import { PlatformError, unknownError } from '@hcengineering/platform'
import type { Middleware, PipelineContext, TxAdapter, TxMiddlewareResult } from '@hcengineering/server-core' import type {
Middleware,
MiddlewareCreator,
PipelineContext,
TxAdapter,
TxMiddlewareResult
} from '@hcengineering/server-core'
import { BaseMiddleware } from '@hcengineering/server-core' import { BaseMiddleware } from '@hcengineering/server-core'
import crypto from 'node:crypto' import crypto from 'node:crypto'
@ -34,20 +42,46 @@ export class ModelMiddleware extends BaseMiddleware implements Middleware {
lastHashResponse!: Promise<LoadModelResponse> lastHashResponse!: Promise<LoadModelResponse>
model!: Tx[] model!: Tx[]
static async create (ctx: MeasureContext, context: PipelineContext, next?: Middleware): Promise<Middleware> { constructor (
const middleware = new ModelMiddleware(context, next) context: PipelineContext,
next: Middleware | undefined,
readonly systemTx: Tx[]
) {
super(context, next)
}
static async doCreate (
ctx: MeasureContext,
context: PipelineContext,
next: Middleware | undefined,
systemTx: Tx[]
): Promise<Middleware> {
const middleware = new ModelMiddleware(context, next, systemTx)
await middleware.init(ctx) await middleware.init(ctx)
return middleware return middleware
} }
static create (tx: Tx[]): MiddlewareCreator {
return (ctx, context, next) => {
return this.doCreate(ctx, context, next, tx)
}
}
async init (ctx: MeasureContext): Promise<void> { async init (ctx: MeasureContext): Promise<void> {
if (this.context.adapterManager == null) { if (this.context.adapterManager == null) {
throw new PlatformError(unknownError('Adapter manager should be configured')) throw new PlatformError(unknownError('Adapter manager should be configured'))
} }
const txAdapter = this.context.adapterManager.getAdapter(DOMAIN_TX, true) as TxAdapter const txAdapter = this.context.adapterManager.getAdapter(DOMAIN_TX, true) as TxAdapter
const isUserTx = (it: Tx): boolean =>
it.modifiedBy !== core.account.System ||
(it as TxCUD<Doc>).objectClass === 'contact:class:Person' ||
(it as TxCUD<Doc>).objectClass === 'contact:class:PersonAccount'
this.model = await ctx.with('get-model', {}, async (ctx) => { this.model = await ctx.with('get-model', {}, async (ctx) => {
const model = await ctx.with('fetch-model', {}, (ctx) => txAdapter.getModel(ctx)) const allUserTxes = await ctx.with('fetch-model', {}, (ctx) => txAdapter.getModel(ctx))
const userTxes = allUserTxes.filter((it) => isUserTx(it))
const model = this.systemTx.concat(userTxes)
for (const tx of model) { for (const tx of model) {
try { try {
this.context.hierarchy.tx(tx) this.context.hierarchy.tx(tx)
@ -55,7 +89,7 @@ export class ModelMiddleware extends BaseMiddleware implements Middleware {
ctx.warn('failed to apply model transaction, skipping', { tx: JSON.stringify(tx), err }) ctx.warn('failed to apply model transaction, skipping', { tx: JSON.stringify(tx), err })
} }
} }
this.context.modelDb.addTxes(ctx, model, false) this.context.modelDb.addTxes(ctx, model, true)
return model return model
}) })

View File

@ -1174,7 +1174,9 @@ abstract class MongoAdapterBase implements DbAdapter {
async clean (ctx: MeasureContext, domain: Domain, docs: Ref<Doc>[]): Promise<void> { async clean (ctx: MeasureContext, domain: Domain, docs: Ref<Doc>[]): Promise<void> {
await ctx.with('clean', {}, async () => { await ctx.with('clean', {}, async () => {
await this.db.collection<Doc>(domain).deleteMany({ _id: { $in: docs } }) if (docs.length > 0) {
await this.db.collection<Doc>(domain).deleteMany({ _id: { $in: docs } })
}
}) })
} }
} }
@ -1693,7 +1695,7 @@ class MongoTxAdapter extends MongoAdapterBase implements TxAdapter {
) )
} }
model.forEach((tx) => (tx.modifiedBy === core.account.System && !isPersonAccount(tx) ? systemTx : userTx).push(tx)) model.forEach((tx) => (tx.modifiedBy === core.account.System && !isPersonAccount(tx) ? systemTx : userTx).push(tx))
return systemTx.concat(userTx) return this.stripHash(systemTx.concat(userTx)) as Tx[]
} }
} }

View File

@ -169,6 +169,7 @@ export class DBCollectionHelper implements DomainHelperOperations {
} }
async init (domain?: Domain): Promise<void> { async init (domain?: Domain): Promise<void> {
// Check and create DB if missin
if (domain === undefined) { if (domain === undefined) {
// Init existing collecfions // Init existing collecfions
for (const c of (await this.db.listCollections({}, { nameOnly: true }).toArray()).map((it) => it.name)) { for (const c of (await this.db.listCollections({}, { nameOnly: true }).toArray()).map((it) => it.name)) {

View File

@ -10,6 +10,7 @@ import {
ModelDb, ModelDb,
type Branding, type Branding,
type MeasureContext, type MeasureContext,
type Tx,
type WorkspaceIdWithUrl type WorkspaceIdWithUrl
} from '@hcengineering/core' } from '@hcengineering/core'
import { createElasticAdapter, createElasticBackupDataAdapter } from '@hcengineering/elastic' import { createElasticAdapter, createElasticBackupDataAdapter } from '@hcengineering/elastic'
@ -100,6 +101,7 @@ export function getTxAdapterFactory (
export function createServerPipeline ( export function createServerPipeline (
metrics: MeasureContext, metrics: MeasureContext,
dbUrls: string, dbUrls: string,
model: Tx[],
opt: { opt: {
fullTextUrl: string fullTextUrl: string
rekoniUrl: string rekoniUrl: string
@ -139,7 +141,7 @@ export function createServerPipeline (
DomainFindMiddleware.create, DomainFindMiddleware.create,
DomainTxMiddleware.create, DomainTxMiddleware.create,
DBAdapterInitMiddleware.create, DBAdapterInitMiddleware.create,
ModelMiddleware.create, ModelMiddleware.create(model),
DBAdapterMiddleware.create(conf), // Configure DB adapters DBAdapterMiddleware.create(conf), // Configure DB adapters
BroadcastMiddleware.create(broadcast) BroadcastMiddleware.create(broadcast)
] ]
@ -164,6 +166,7 @@ export function createServerPipeline (
export function createBackupPipeline ( export function createBackupPipeline (
metrics: MeasureContext, metrics: MeasureContext,
dbUrls: string, dbUrls: string,
systemTx: Tx[],
opt: { opt: {
usePassedCtx?: boolean usePassedCtx?: boolean
adapterSecurity?: boolean adapterSecurity?: boolean
@ -208,7 +211,7 @@ export function createBackupPipeline (
ContextNameMiddleware.create, ContextNameMiddleware.create,
DomainFindMiddleware.create, DomainFindMiddleware.create,
DBAdapterInitMiddleware.create, DBAdapterInitMiddleware.create,
ModelMiddleware.create, ModelMiddleware.create(systemTx),
DBAdapterMiddleware.create(conf) DBAdapterMiddleware.create(conf)
] ]
@ -227,6 +230,7 @@ export function createBackupPipeline (
export async function getServerPipeline ( export async function getServerPipeline (
ctx: MeasureContext, ctx: MeasureContext,
model: Tx[],
mongodbUri: string, mongodbUri: string,
dbUrl: string | undefined, dbUrl: string | undefined,
wsUrl: WorkspaceIdWithUrl wsUrl: WorkspaceIdWithUrl
@ -242,6 +246,7 @@ export async function getServerPipeline (
const pipelineFactory = createServerPipeline( const pipelineFactory = createServerPipeline(
ctx, ctx,
dbUrls, dbUrls,
model,
{ {
externalStorage: storageAdapter, externalStorage: storageAdapter,
fullTextUrl: 'http://localhost:9200', fullTextUrl: 'http://localhost:9200',

View File

@ -36,15 +36,16 @@ import core, {
WorkspaceId, WorkspaceId,
WorkspaceIdWithUrl, WorkspaceIdWithUrl,
type Doc, type Doc,
type Ref,
type TxCUD type TxCUD
} from '@hcengineering/core' } from '@hcengineering/core'
import { consoleModelLogger, MigrateOperation, ModelLogger, tryMigrate } from '@hcengineering/model' import { consoleModelLogger, MigrateOperation, ModelLogger, tryMigrate } from '@hcengineering/model'
import { import {
AggregatorStorageAdapter, AggregatorStorageAdapter,
DbAdapter,
DomainIndexHelperImpl, DomainIndexHelperImpl,
Pipeline, Pipeline,
StorageAdapter StorageAdapter,
type DbAdapter
} from '@hcengineering/server-core' } from '@hcengineering/server-core'
import { connect } from './connect' import { connect } from './connect'
import { InitScript, WorkspaceInitializer } from './initializer' import { InitScript, WorkspaceInitializer } from './initializer'
@ -141,9 +142,16 @@ export async function initModel (
logger.log('transactions deleted.', { workspaceId: workspaceId.name }) logger.log('transactions deleted.', { workspaceId: workspaceId.name })
} }
logger.log('creating model...', workspaceId) logger.log('creating database...', workspaceId)
await adapter.upload(ctx, DOMAIN_TX, txes) await adapter.upload(ctx, DOMAIN_TX, [
logger.log('model transactions inserted.', { count: txes.length }) {
_class: core.class.Tx,
_id: 'first-tx' as Ref<Doc>,
modifiedBy: core.account.System,
modifiedOn: Date.now(),
space: core.space.DerivedTx
}
])
await progress(30) await progress(30)
@ -264,11 +272,14 @@ export async function upgradeModel (
throw Error('Model txes must target only core.space.Model') throw Error('Model txes must target only core.space.Model')
} }
const prevModel = await fetchModel(ctx, pipeline) const newModelRes = await ctx.with('load-model', {}, (ctx) => pipeline.loadModel(ctx, 0))
const newModel = Array.isArray(newModelRes) ? newModelRes : newModelRes.transactions
const { hierarchy, modelDb, model } = await buildModel(ctx, newModel)
const { migrateClient: preMigrateClient } = await prepareMigrationClient( const { migrateClient: preMigrateClient } = await prepareMigrationClient(
pipeline, pipeline,
prevModel.hierarchy, hierarchy,
prevModel.modelDb, modelDb,
logger, logger,
storageAdapter, storageAdapter,
workspaceId workspaceId
@ -305,36 +316,9 @@ export async function upgradeModel (
} }
logger.log('removing model...', { workspaceId: workspaceId.name }) logger.log('removing model...', { workspaceId: workspaceId.name })
await progress(10) await progress(10)
const toRemove = await pipeline.findAll(ctx, core.class.Tx, {
objectSpace: core.space.Model,
modifiedBy: core.account.System,
objectClass: { $nin: [contact.class.PersonAccount, 'contact:class:EmployeeAccount'] }
})
await pipeline.context.lowLevelStorage.clean(
ctx,
DOMAIN_TX,
toRemove.map((p) => p._id)
)
logger.log('transactions deleted.', { workspaceId: workspaceId.name, count: toRemove.length })
logger.log('creating model...', { workspaceId: workspaceId.name })
await pipeline.context.lowLevelStorage.upload(ctx, DOMAIN_TX, txes)
logger.log('model transactions inserted.', { workspaceId: workspaceId.name, count: txes.length }) logger.log('model transactions inserted.', { workspaceId: workspaceId.name, count: txes.length })
await progress(20) await progress(20)
} }
const newModel = [
...txes,
...Array.from(
prevModel.model.filter(
(it) =>
it.modifiedBy !== core.account.System ||
(it as TxCUD<Doc>).objectClass === contact.class.Person ||
(it as TxCUD<Doc>).objectClass === 'contact:class:PersonAccount'
)
)
]
const { hierarchy, modelDb, model } = await fetchModel(ctx, pipeline, newModel)
const { migrateClient, migrateState } = await prepareMigrationClient( const { migrateClient, migrateState } = await prepareMigrationClient(
pipeline, pipeline,
hierarchy, hierarchy,
@ -383,6 +367,21 @@ export async function upgradeModel (
{ {
state: 'indexes-v4', state: 'indexes-v4',
func: upgradeIndexes func: upgradeIndexes
},
{
state: 'delete-model',
func: async (client) => {
const model = await client.find<Tx>(DOMAIN_TX, { objectSpace: core.space.Model })
// Ignore Employee accounts.
const isUserTx = (it: Tx): boolean =>
it.modifiedBy !== core.account.System ||
(it as TxCUD<Doc>).objectClass === 'contact:class:Person' ||
(it as TxCUD<Doc>).objectClass === 'contact:class:PersonAccount'
const toDelete = model.filter((it) => !isUserTx(it)).map((it) => it._id)
await client.deleteMany(DOMAIN_TX, { _id: { $in: toDelete } })
}
} }
]) ])
}) })
@ -463,26 +462,20 @@ async function prepareMigrationClient (
return { migrateClient, migrateState } return { migrateClient, migrateState }
} }
export async function fetchModel ( export async function buildModel (
ctx: MeasureContext, ctx: MeasureContext,
pipeline: Pipeline, model: Tx[]
model?: Tx[]
): Promise<{ hierarchy: Hierarchy, modelDb: ModelDb, model: Tx[] }> { ): Promise<{ hierarchy: Hierarchy, modelDb: ModelDb, model: Tx[] }> {
const hierarchy = new Hierarchy() const hierarchy = new Hierarchy()
const modelDb = new ModelDb(hierarchy) const modelDb = new ModelDb(hierarchy)
if (model === undefined) {
const res = await ctx.with('load-model', {}, (ctx) => pipeline.loadModel(ctx, 0))
model = Array.isArray(res) ? res : res.transactions
}
ctx.withSync('build local model', {}, () => { ctx.withSync('build local model', {}, () => {
for (const tx of model ?? []) { for (const tx of model ?? []) {
try { try {
hierarchy.tx(tx) hierarchy.tx(tx)
} catch (err: any) {} } catch (err: any) {}
} }
modelDb.addTxes(ctx, model as Tx[], false) modelDb.addTxes(ctx, model, false)
}) })
return { hierarchy, modelDb, model: model ?? [] } return { hierarchy, modelDb, model: model ?? [] }
} }

View File

@ -163,6 +163,7 @@ export async function createWorkspace (
const factory: PipelineFactory = createServerPipeline( const factory: PipelineFactory = createServerPipeline(
ctx, ctx,
dbUrls, dbUrls,
txes,
{ {
externalStorage: storageAdapter, externalStorage: storageAdapter,
fullTextUrl: 'http://localhost:9200', fullTextUrl: 'http://localhost:9200',
@ -288,7 +289,7 @@ export async function upgradeWorkspace (
let pipeline: Pipeline | undefined let pipeline: Pipeline | undefined
let storageAdapter: StorageAdapter | undefined let storageAdapter: StorageAdapter | undefined
try { try {
;({ pipeline, storageAdapter } = await getServerPipeline(ctx, mongodbUri, dbUrl, wsUrl)) ;({ pipeline, storageAdapter } = await getServerPipeline(ctx, txes, mongodbUri, dbUrl, wsUrl))
const contextData = new SessionDataImpl( const contextData = new SessionDataImpl(
systemAccountEmail, systemAccountEmail,
'backup', 'backup',