Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 2 additions & 2 deletions server/backup/src/backup.ts
Original file line number Diff line number Diff line change
Expand Up @@ -188,7 +188,7 @@ async function loadDigest (
result.delete(k as Ref<Doc>)
}
} catch (err: any) {
ctx.error('digest is broken, will do full backup for', { domain, err: err.message, snapshot })
ctx.warn('digest is broken, will do full backup for', { domain, err: err.message, snapshot })
}
}
// Stop if stop date is matched and provided
Expand Down Expand Up @@ -377,7 +377,7 @@ async function updateDigest (
} catch (err: any) {
digestToRemove.add(snapshot)
modifiedFiles.push(snapshot)
ctx.error('digest is broken, will do full backup for', { domain, err: err.message, snapshot })
ctx.warn('digest is broken, will do full backup for', { domain, err: err.message, snapshot })
modified = true
}
}
Expand Down
18 changes: 16 additions & 2 deletions server/backup/src/service.ts
Original file line number Diff line number Diff line change
Expand Up @@ -40,9 +40,9 @@ import {
type StorageAdapter
} from '@hcengineering/server-core'
import { generateToken } from '@hcengineering/server-token'
import { clearInterval } from 'node:timers'
import { backup, restore } from '.'
import { createStorageBackupStorage } from './storage'
import { clearInterval } from 'node:timers'
export interface BackupConfig {
AccountsURL: string
Token: string
Expand Down Expand Up @@ -208,6 +208,7 @@ class BackupWorker {
const rateLimiter = new RateLimiter(this.config.Parallel)

const times: number[] = []
const activeWorkspaces = new Set<string>()

const infoTo = setInterval(() => {
const avgTime = times.length > 0 ? Math.round(times.reduce((p, c) => p + c, 0) / times.length) / 1000 : 0
Expand All @@ -218,14 +219,16 @@ class BackupWorker {
index,
Elapsed: (Date.now() - startTime) / 1000,
ETA: Math.round((workspaces.length - processed) * avgTime),
active: rateLimiter.processingQueue.size
activeLen: activeWorkspaces.size,
active: Array.from(activeWorkspaces).join(',')
})
}, 10000)

try {
for (const ws of workspaces) {
await rateLimiter.add(async () => {
try {
activeWorkspaces.add(ws.workspace)
index++
if (this.canceled || Date.now() - startTime > recheckTimeout) {
return // If canceled, we should stop
Expand All @@ -241,6 +244,9 @@ class BackupWorker {
processed++
} catch (err: any) {
ctx.error('Backup failed', { err })
failedWorkspaces.push(ws)
} finally {
activeWorkspaces.delete(ws.workspace)
}
})
}
Expand Down Expand Up @@ -348,6 +354,14 @@ class BackupWorker {
// We need to report update for stats to account service
const token = generateToken(systemAccountEmail, { name: ws.workspace }, { service: 'backup' })
await updateBackupInfo(token, backupInfo)
} else {
rootCtx.error('BACKUP FAILED', {
workspace: ws.workspace,
workspaceUrl: ws.workspaceUrl,
workspaceName: ws.workspaceName,
time: Math.round((Date.now() - st) / 1000)
})
return false
}
} catch (err: any) {
rootCtx.error('\n\nFAILED to BACKUP', { workspace: ws.workspace, err })
Expand Down
9 changes: 5 additions & 4 deletions server/mongo/src/storage.ts
Original file line number Diff line number Diff line change
Expand Up @@ -64,15 +64,16 @@ import core, {
type WorkspaceId
} from '@hcengineering/core'
import {
calcHashHash,
type DbAdapter,
type DbAdapterHandler,
type DomainHelperOperations,
type ServerFindOptions,
type StorageAdapter,
type TxAdapter,
calcHashHash
type TxAdapter
} from '@hcengineering/server-core'
import {
ObjectId,
type AbstractCursor,
type AnyBulkWriteOperation,
type Collection,
Expand Down Expand Up @@ -1123,14 +1124,14 @@ abstract class MongoAdapterBase implements DbAdapter {
const result: DocInfo[] = []
if (d != null) {
result.push({
id: d._id,
id: (d._id as any) instanceof ObjectId ? d._id.toString() : d._id,
hash: this.strimSize((d as any)['%hash%'])
})
}
if (iterator.bufferedCount() > 0) {
result.push(
...iterator.readBufferedDocuments().map((it) => ({
id: it._id,
id: (it._id as any) instanceof ObjectId ? it._id.toString() : it._id,
hash: this.strimSize((it as any)['%hash%'])
}))
)
Expand Down