// Copyright 2020 Signal Messenger, LLC // SPDX-License-Identifier: AGPL-3.0-only import { debounce, isNumber, chunk } from 'lodash'; import pMap from 'p-map'; import Long from 'long'; import { DataReader, DataWriter } from '../sql/Client'; import * as Bytes from '../Bytes'; import { getRandomBytes, deriveStorageItemKey, deriveStorageManifestKey, encryptProfile, decryptProfile, deriveMasterKeyFromGroupV1, deriveStorageServiceKey, } from '../Crypto'; import { mergeAccountRecord, mergeContactRecord, mergeGroupV1Record, mergeGroupV2Record, mergeStoryDistributionListRecord, mergeStickerPackRecord, toAccountRecord, toContactRecord, toGroupV1Record, toGroupV2Record, toStoryDistributionListRecord, toStickerPackRecord, toCallLinkRecord, mergeCallLinkRecord, } from './storageRecordOps'; import type { MergeResultType } from './storageRecordOps'; import { MAX_READ_KEYS } from './storageConstants'; import type { ConversationModel } from '../models/conversations'; import { strictAssert } from '../util/assert'; import { drop } from '../util/drop'; import { dropNull } from '../util/dropNull'; import * as durations from '../util/durations'; import { BackOff } from '../util/BackOff'; import { storageJobQueue } from '../util/JobQueue'; import { sleep } from '../util/sleep'; import { isMoreRecentThan, isOlderThan } from '../util/timestamp'; import { map, filter } from '../util/iterables'; import { ourProfileKeyService } from './ourProfileKey'; import { ConversationTypes, isDirectConversation, typeofConversation, } from '../util/whatTypeOfConversation'; import { SignalService as Proto } from '../protobuf'; import * as log from '../logging/log'; import { singleProtoJobQueue } from '../jobs/singleProtoJobQueue'; import * as Errors from '../types/errors'; import type { ExtendedStorageID, RemoteRecord, UnknownRecord, } from '../types/StorageService.d'; import MessageSender from '../textsecure/SendMessage'; import type { StoryDistributionWithMembersType, StorageServiceFieldsType, StickerPackType, UninstalledStickerPackType, } from '../sql/Interface'; import { MY_STORY_ID } from '../types/Stories'; import { isNotNil } from '../util/isNotNil'; import { isSignalConversation } from '../util/isSignalConversation'; import { redactExtendedStorageID, redactStorageID } from '../util/privacy'; import type { CallLinkRecord } from '../types/CallLink'; import { callLinkFromRecord } from '../util/callLinksRingrtc'; type IManifestRecordIdentifier = Proto.ManifestRecord.IIdentifier; const { getItemById } = DataReader; const { eraseStorageServiceState, flushUpdateConversationBatcher, updateConversation, updateConversations, } = DataWriter; const uploadBucket: Array = []; const validRecordTypes = new Set([ 0, // UNKNOWN 1, // CONTACT 2, // GROUPV1 3, // GROUPV2 4, // ACCOUNT 5, // STORY_DISTRIBUTION_LIST 6, // STICKER_PACK 7, // CALL_LINK ]); const backOff = new BackOff([ durations.SECOND, 5 * durations.SECOND, 30 * durations.SECOND, 2 * durations.MINUTE, 5 * durations.MINUTE, ]); const conflictBackOff = new BackOff([ durations.SECOND, 5 * durations.SECOND, 30 * durations.SECOND, ]); function encryptRecord( storageID: string | undefined, storageRecord: Proto.IStorageRecord ): Proto.StorageItem { const storageItem = new Proto.StorageItem(); const storageKeyBuffer = storageID ? Bytes.fromBase64(storageID) : generateStorageID(); const storageKeyBase64 = window.storage.get('storageKey'); if (!storageKeyBase64) { throw new Error('No storage key'); } const storageKey = Bytes.fromBase64(storageKeyBase64); const storageItemKey = deriveStorageItemKey( storageKey, Bytes.toBase64(storageKeyBuffer) ); const encryptedRecord = encryptProfile( Proto.StorageRecord.encode(storageRecord).finish(), storageItemKey ); storageItem.key = storageKeyBuffer; storageItem.value = encryptedRecord; return storageItem; } function generateStorageID(): Uint8Array { return getRandomBytes(16); } type GeneratedManifestType = { postUploadUpdateFunctions: Array<() => unknown>; recordsByID: Map; insertKeys: Set; deleteKeys: Set; }; async function generateManifest( version: number, previousManifest?: Proto.IManifestRecord, isNewManifest = false ): Promise { log.info( `storageService.upload(${version}): generating manifest ` + `new=${isNewManifest}` ); await window.ConversationController.checkForConflicts(); const ITEM_TYPE = Proto.ManifestRecord.Identifier.Type; const postUploadUpdateFunctions: Array<() => unknown> = []; const insertKeys = new Set(); const deleteKeys = new Set(); const recordsByID = new Map(); function processStorageRecord({ conversation, currentStorageID, currentStorageVersion, identifierType, storageNeedsSync, storageRecord, }: { conversation?: ConversationModel; currentStorageID?: string; currentStorageVersion?: number; identifierType: Proto.ManifestRecord.Identifier.Type; storageNeedsSync: boolean; storageRecord: Proto.IStorageRecord; }) { const currentRedactedID = currentStorageID ? redactStorageID(currentStorageID, currentStorageVersion) : undefined; const isNewItem = isNewManifest || storageNeedsSync || !currentStorageID; const storageID = isNewItem ? Bytes.toBase64(generateStorageID()) : currentStorageID; recordsByID.set(storageID, { itemType: identifierType, storageID, storageRecord, }); // When a client needs to update a given record it should create it // under a new key and delete the existing key. if (isNewItem) { insertKeys.add(storageID); const newRedactedID = redactStorageID(storageID, version, conversation); if (currentStorageID) { log.info( `storageService.upload(${version}): ` + `updating from=${currentRedactedID} ` + `to=${newRedactedID}` ); deleteKeys.add(currentStorageID); } else { log.info( `storageService.upload(${version}): adding key=${newRedactedID}` ); } } return { isNewItem, storageID, }; } const conversations = window.getConversations(); for (let i = 0; i < conversations.length; i += 1) { const conversation = conversations.models[i]; let identifierType; let storageRecord; if (isSignalConversation(conversation.attributes)) { continue; } const conversationType = typeofConversation(conversation.attributes); if (conversationType === ConversationTypes.Me) { storageRecord = new Proto.StorageRecord(); // eslint-disable-next-line no-await-in-loop storageRecord.account = await toAccountRecord(conversation); identifierType = ITEM_TYPE.ACCOUNT; } else if (conversationType === ConversationTypes.Direct) { // Contacts must have UUID if (!conversation.getServiceId()) { continue; } let shouldDrop = false; let dropReason: string | undefined; const validationError = conversation.validate(); if (validationError) { shouldDrop = true; dropReason = `local validation error=${validationError}`; } else if (conversation.isUnregisteredAndStale()) { shouldDrop = true; dropReason = 'unregistered and stale'; } if (shouldDrop) { const droppedID = conversation.get('storageID'); const droppedVersion = conversation.get('storageVersion'); if (!droppedID) { continue; } const recordID = redactStorageID( droppedID, droppedVersion, conversation ); log.warn( `storageService.generateManifest(${version}): ` + `dropping contact=${recordID} ` + `due to ${dropReason}` ); conversation.unset('storageID'); deleteKeys.add(droppedID); continue; } storageRecord = new Proto.StorageRecord(); // eslint-disable-next-line no-await-in-loop storageRecord.contact = await toContactRecord(conversation); identifierType = ITEM_TYPE.CONTACT; } else if (conversationType === ConversationTypes.GroupV2) { storageRecord = new Proto.StorageRecord(); storageRecord.groupV2 = toGroupV2Record(conversation); identifierType = ITEM_TYPE.GROUPV2; } else if (conversationType === ConversationTypes.GroupV1) { storageRecord = new Proto.StorageRecord(); storageRecord.groupV1 = toGroupV1Record(conversation); identifierType = ITEM_TYPE.GROUPV1; } else { log.warn( `storageService.upload(${version}): ` + `unknown conversation=${conversation.idForLogging()}` ); } if (!storageRecord || !identifierType) { continue; } const { isNewItem, storageID } = processStorageRecord({ conversation, currentStorageID: conversation.get('storageID'), currentStorageVersion: conversation.get('storageVersion'), identifierType, storageNeedsSync: Boolean(conversation.get('needsStorageServiceSync')), storageRecord, }); if (isNewItem) { postUploadUpdateFunctions.push(() => { conversation.set({ needsStorageServiceSync: false, storageVersion: version, storageID, }); drop(updateConversation(conversation.attributes)); }); } } const { callLinkDbRecords, storyDistributionLists, installedStickerPacks, uninstalledStickerPacks, } = await getNonConversationRecords(); log.info( `storageService.upload(${version}): ` + `adding storyDistributionLists=${storyDistributionLists.length}` ); for (const storyDistributionList of storyDistributionLists) { const storageRecord = new Proto.StorageRecord(); storageRecord.storyDistributionList = toStoryDistributionListRecord( storyDistributionList ); if ( storyDistributionList.deletedAtTimestamp != null && isOlderThan(storyDistributionList.deletedAtTimestamp, durations.MONTH) ) { const droppedID = storyDistributionList.storageID; const droppedVersion = storyDistributionList.storageVersion; if (!droppedID) { continue; } const recordID = redactStorageID(droppedID, droppedVersion); log.warn( `storageService.generateManifest(${version}): ` + `dropping storyDistributionList=${recordID} ` + `due to expired deleted timestamp=${storyDistributionList.deletedAtTimestamp}` ); deleteKeys.add(droppedID); drop(DataWriter.deleteStoryDistribution(storyDistributionList.id)); continue; } const { isNewItem, storageID } = processStorageRecord({ currentStorageID: storyDistributionList.storageID, currentStorageVersion: storyDistributionList.storageVersion, identifierType: ITEM_TYPE.STORY_DISTRIBUTION_LIST, storageNeedsSync: storyDistributionList.storageNeedsSync, storageRecord, }); if (isNewItem) { postUploadUpdateFunctions.push(() => { void DataWriter.modifyStoryDistribution({ ...storyDistributionList, storageID, storageVersion: version, storageNeedsSync: false, }); }); } } log.info( `storageService.upload(${version}): ` + `adding uninstalled stickerPacks=${uninstalledStickerPacks.length}` ); const uninstalledStickerPackIds = new Set(); uninstalledStickerPacks.forEach(stickerPack => { const storageRecord = new Proto.StorageRecord(); storageRecord.stickerPack = toStickerPackRecord(stickerPack); uninstalledStickerPackIds.add(stickerPack.id); const { isNewItem, storageID } = processStorageRecord({ currentStorageID: stickerPack.storageID, currentStorageVersion: stickerPack.storageVersion, identifierType: ITEM_TYPE.STICKER_PACK, storageNeedsSync: stickerPack.storageNeedsSync, storageRecord, }); if (isNewItem) { postUploadUpdateFunctions.push(() => { void DataWriter.addUninstalledStickerPack({ ...stickerPack, storageID, storageVersion: version, storageNeedsSync: false, }); }); } }); log.info( `storageService.upload(${version}): ` + `adding installed stickerPacks=${installedStickerPacks.length}` ); installedStickerPacks.forEach(stickerPack => { if (uninstalledStickerPackIds.has(stickerPack.id)) { log.error( `storageService.upload(${version}): ` + `sticker pack ${stickerPack.id} is both installed and uninstalled` ); window.reduxActions.stickers.uninstallStickerPack( stickerPack.id, stickerPack.key, { fromSync: true } ); return; } const storageRecord = new Proto.StorageRecord(); storageRecord.stickerPack = toStickerPackRecord(stickerPack); const { isNewItem, storageID } = processStorageRecord({ currentStorageID: stickerPack.storageID, currentStorageVersion: stickerPack.storageVersion, identifierType: ITEM_TYPE.STICKER_PACK, storageNeedsSync: stickerPack.storageNeedsSync, storageRecord, }); if (isNewItem) { postUploadUpdateFunctions.push(() => { void DataWriter.createOrUpdateStickerPack({ ...stickerPack, storageID, storageVersion: version, storageNeedsSync: false, }); }); } }); log.info( `storageService.upload(${version}): ` + `adding callLinks=${callLinkDbRecords.length}` ); for (const callLinkDbRecord of callLinkDbRecords) { const { roomId } = callLinkDbRecord; if (callLinkDbRecord.adminKey == null || callLinkDbRecord.rootKey == null) { log.warn( `storageService.upload(${version}): ` + `call link ${roomId} has empty rootKey` ); continue; } const storageRecord = new Proto.StorageRecord(); storageRecord.callLink = toCallLinkRecord(callLinkDbRecord); const callLink = callLinkFromRecord(callLinkDbRecord); const { isNewItem, storageID } = processStorageRecord({ currentStorageID: callLink.storageID, currentStorageVersion: callLink.storageVersion, identifierType: ITEM_TYPE.CALL_LINK, storageNeedsSync: callLink.storageNeedsSync, storageRecord, }); const storageFields = { storageID, storageVersion: version, storageNeedsSync: false, }; if (isNewItem) { postUploadUpdateFunctions.push(async () => { const freshCallLink = await DataReader.getCallLinkByRoomId(roomId); if (freshCallLink == null) { log.warn( `storageService.upload(${version}): ` + `call link ${roomId} removed locally from DB while we were uploading to storage` ); return; } const callLinkToSave = { ...freshCallLink, ...storageFields }; await DataWriter.updateCallLink(callLinkToSave); window.reduxActions.calling.handleCallLinkUpdateLocal(callLinkToSave); }); } } const unknownRecordsArray: ReadonlyArray = ( window.storage.get('storage-service-unknown-records') || [] ).filter((record: UnknownRecord) => !validRecordTypes.has(record.itemType)); const redactedUnknowns = unknownRecordsArray.map(redactExtendedStorageID); log.info( `storageService.upload(${version}): adding unknown ` + `records=${JSON.stringify(redactedUnknowns)} ` + `count=${redactedUnknowns.length}` ); // When updating the manifest, ensure all "unknown" keys are added to the // new manifest, so we don't inadvertently delete something we don't understand unknownRecordsArray.forEach((record: UnknownRecord) => { recordsByID.set(record.storageID, record); }); const recordsWithErrors: ReadonlyArray = window.storage.get( 'storage-service-error-records', new Array() ); const redactedErrors = recordsWithErrors.map(redactExtendedStorageID); log.info( `storageService.upload(${version}): adding error ` + `records=${JSON.stringify(redactedErrors)} count=${redactedErrors.length}` ); // These records failed to merge in the previous fetchManifest, but we still // need to include them so that the manifest is complete recordsWithErrors.forEach((record: UnknownRecord) => { recordsByID.set(record.storageID, record); }); // Delete keys that we wanted to drop during the processing of the manifest. const storedPendingDeletes = window.storage.get( 'storage-service-pending-deletes', [] ); const redactedPendingDeletes = storedPendingDeletes.map( redactExtendedStorageID ); log.info( `storageService.upload(${version}): ` + `deleting extra keys=${JSON.stringify(redactedPendingDeletes)} ` + `count=${redactedPendingDeletes.length}` ); for (const { storageID } of storedPendingDeletes) { deleteKeys.add(storageID); } // Validate before writing const duplicates = new Set(); const typeDuplicates = new Set(); let hasAccountType = false; for (const [storageID, { itemType }] of recordsByID) { // Ensure there are no duplicate StorageIdentifiers in your manifest // This can be broken down into two parts: // There are no duplicate type+raw pairs // There are no duplicate raw bytes const typeAndID = `${itemType}+${storageID}`; if (duplicates.has(storageID) || typeDuplicates.has(typeAndID)) { log.warn( `storageService.upload(${version}): removing from duplicate item ` + 'from the manifest', redactStorageID(storageID), itemType ); recordsByID.delete(storageID); } duplicates.add(storageID); typeDuplicates.add(typeAndID); // Ensure all deletes are not present in the manifest const hasDeleteKey = deleteKeys.has(storageID); if (hasDeleteKey) { log.warn( `storageService.upload(${version}): removing key which has been deleted`, redactStorageID(storageID), itemType ); recordsByID.delete(storageID); } // Ensure that there is *exactly* one Account type in the manifest if (itemType === ITEM_TYPE.ACCOUNT) { if (hasAccountType) { log.warn( `storageService.upload(${version}): removing duplicate account`, redactStorageID(storageID) ); recordsByID.delete(storageID); } hasAccountType = true; } } duplicates.clear(); typeDuplicates.clear(); const storageKeyDuplicates = new Set(); for (const storageID of insertKeys) { // Ensure there are no duplicate StorageIdentifiers in your list of inserts if (storageKeyDuplicates.has(storageID)) { log.warn( `storageService.upload(${version}): ` + 'removing duplicate identifier from inserts', redactStorageID(storageID) ); insertKeys.delete(storageID); } storageKeyDuplicates.add(storageID); } storageKeyDuplicates.clear(); // If we have a copy of what the current remote manifest is then we run these // additional validations comparing our pending manifest to the remote // manifest: if (previousManifest) { const pendingInserts: Set = new Set(); const pendingDeletes: Set = new Set(); const remoteKeys: Set = new Set(); (previousManifest.keys ?? []).forEach( (identifier: IManifestRecordIdentifier) => { strictAssert(identifier.raw, 'Identifier without raw field'); const storageID = Bytes.toBase64(identifier.raw); remoteKeys.add(storageID); } ); const localKeys: Set = new Set(); for (const storageID of recordsByID.keys()) { localKeys.add(storageID); if (!remoteKeys.has(storageID)) { pendingInserts.add(storageID); } } remoteKeys.forEach(storageID => { if (!localKeys.has(storageID)) { pendingDeletes.add(storageID); } }); // Save pending deletes until we have a confirmed upload await window.storage.put( 'storage-service-pending-deletes', // Note: `deleteKeys` already includes the prev value of // 'storage-service-pending-deletes' Array.from(deleteKeys, storageID => ({ storageID, storageVersion: version, })) ); if (deleteKeys.size !== pendingDeletes.size) { const localDeletes = Array.from(deleteKeys).map(key => redactStorageID(key) ); const remoteDeletes: Array = []; pendingDeletes.forEach(id => remoteDeletes.push(redactStorageID(id))); log.error( `storageService.upload(${version}): delete key sizes do not match`, 'local', localDeletes.join(','), 'remote', remoteDeletes.join(',') ); throw new Error('invalid write delete keys length do not match'); } if (insertKeys.size !== pendingInserts.size) { throw new Error('invalid write insert items length do not match'); } for (const storageID of deleteKeys) { if (!pendingDeletes.has(storageID)) { throw new Error( 'invalid write delete key missing from pending deletes' ); } } for (const storageID of insertKeys) { if (!pendingInserts.has(storageID)) { throw new Error( 'invalid write insert key missing from pending inserts' ); } } } return { postUploadUpdateFunctions, recordsByID, insertKeys, deleteKeys, }; } type EncryptManifestOptionsType = { recordsByID: Map; insertKeys: Set; }; type EncryptedManifestType = { newItems: Set; storageManifest: Proto.IStorageManifest; }; async function encryptManifest( version: number, { recordsByID, insertKeys }: EncryptManifestOptionsType ): Promise { const manifestRecordKeys: Set = new Set(); const newItems: Set = new Set(); for (const [storageID, { itemType, storageRecord }] of recordsByID) { const identifier = new Proto.ManifestRecord.Identifier({ type: itemType, raw: Bytes.fromBase64(storageID), }); manifestRecordKeys.add(identifier); if (insertKeys.has(storageID)) { strictAssert( storageRecord !== undefined, 'Inserted items must have an associated record' ); let storageItem; try { storageItem = encryptRecord(storageID, storageRecord); } catch (err) { log.error( `storageService.upload(${version}): encrypt record failed:`, Errors.toLogFormat(err) ); throw err; } newItems.add(storageItem); } } const manifestRecord = new Proto.ManifestRecord(); manifestRecord.version = Long.fromNumber(version); manifestRecord.sourceDevice = window.storage.user.getDeviceId() ?? 0; manifestRecord.keys = Array.from(manifestRecordKeys); const storageKeyBase64 = window.storage.get('storageKey'); if (!storageKeyBase64) { throw new Error('No storage key'); } const storageKey = Bytes.fromBase64(storageKeyBase64); const storageManifestKey = deriveStorageManifestKey( storageKey, Long.fromNumber(version) ); const encryptedManifest = encryptProfile( Proto.ManifestRecord.encode(manifestRecord).finish(), storageManifestKey ); const storageManifest = new Proto.StorageManifest(); storageManifest.version = manifestRecord.version; storageManifest.value = encryptedManifest; return { newItems, storageManifest, }; } async function uploadManifest( version: number, { postUploadUpdateFunctions, deleteKeys }: GeneratedManifestType, { newItems, storageManifest }: EncryptedManifestType ): Promise { if (!window.textsecure.messaging) { throw new Error('storageService.uploadManifest: We are offline!'); } if (newItems.size === 0 && deleteKeys.size === 0) { log.info(`storageService.upload(${version}): nothing to upload`); return; } const credentials = window.storage.get('storageCredentials'); try { log.info( `storageService.upload(${version}): inserting=${newItems.size} ` + `deleting=${deleteKeys.size}` ); const writeOperation = new Proto.WriteOperation(); writeOperation.manifest = storageManifest; writeOperation.insertItem = Array.from(newItems); writeOperation.deleteKey = Array.from(deleteKeys).map(storageID => Bytes.fromBase64(storageID) ); await window.textsecure.messaging.modifyStorageRecords( Proto.WriteOperation.encode(writeOperation).finish(), { credentials, } ); log.info( `storageService.upload(${version}): upload complete, updating ` + `items=${postUploadUpdateFunctions.length}` ); // update conversations with the new storageID postUploadUpdateFunctions.forEach(fn => fn()); } catch (err) { log.error( `storageService.upload(${version}): failed!`, Errors.toLogFormat(err) ); if (err.code === 409) { if (conflictBackOff.isFull()) { log.error( `storageService.upload(${version}): exceeded maximum consecutive ` + 'conflicts' ); return; } log.info( `storageService.upload(${version}): conflict found with ` + `version=${version}, running sync job ` + `times=${conflictBackOff.getIndex()}` ); throw err; } throw err; } log.info(`storageService.upload(${version}): setting new manifestVersion`); await window.storage.put('manifestVersion', version); conflictBackOff.reset(); backOff.reset(); try { await singleProtoJobQueue.add(MessageSender.getFetchManifestSyncMessage()); } catch (error) { log.error( `storageService.upload(${version}): Failed to queue sync message`, Errors.toLogFormat(error) ); } } async function stopStorageServiceSync(reason: Error) { log.warn('storageService.stopStorageServiceSync', Errors.toLogFormat(reason)); await window.storage.remove('storageKey'); if (backOff.isFull()) { log.warn( 'storageService.stopStorageServiceSync: too many consecutive stops' ); return; } await sleep(backOff.getAndIncrement()); log.info('storageService.stopStorageServiceSync: requesting new keys'); setTimeout(async () => { if (window.ConversationController.areWePrimaryDevice()) { log.warn( 'stopStorageServiceSync: We are primary device; not sending key sync request' ); return; } try { await singleProtoJobQueue.add(MessageSender.getRequestKeySyncMessage()); } catch (error) { log.error( 'storageService.stopStorageServiceSync: Failed to queue sync message', Errors.toLogFormat(error) ); } }); } async function createNewManifest() { log.info('storageService.createNewManifest: creating new manifest'); const version = window.storage.get('manifestVersion', 0); const generatedManifest = await generateManifest(version, undefined, true); const encryptedManifest = await encryptManifest(version, generatedManifest); await uploadManifest( version, { ...generatedManifest, // we have created a new manifest, there should be no keys to delete deleteKeys: new Set(), }, encryptedManifest ); } async function decryptManifest( encryptedManifest: Proto.IStorageManifest ): Promise { const { version, value } = encryptedManifest; const storageKeyBase64 = window.storage.get('storageKey'); if (!storageKeyBase64) { throw new Error('No storage key'); } const storageKey = Bytes.fromBase64(storageKeyBase64); const storageManifestKey = deriveStorageManifestKey( storageKey, dropNull(version) ); strictAssert(value, 'StorageManifest has no value field'); const decryptedManifest = decryptProfile(value, storageManifestKey); return Proto.ManifestRecord.decode(decryptedManifest); } async function fetchManifest( manifestVersion: number ): Promise { log.info('storageService.sync: fetch start'); if (!window.textsecure.messaging) { throw new Error('storageService.sync: we are offline!'); } try { const credentials = await window.textsecure.messaging.getStorageCredentials(); await window.storage.put('storageCredentials', credentials); const manifestBinary = await window.textsecure.messaging.getStorageManifest( { credentials, greaterThanVersion: manifestVersion, } ); const encryptedManifest = Proto.StorageManifest.decode(manifestBinary); try { return decryptManifest(encryptedManifest); } catch (err) { await stopStorageServiceSync(err); } } catch (err) { if (err.code === 204) { log.info('storageService.sync: no newer manifest, ok'); return undefined; } log.error('storageService.sync: failed!', Errors.toLogFormat(err)); if (err.code === 404) { await createNewManifest(); return undefined; } throw err; } return undefined; } type MergeableItemType = { itemType: number; storageID: string; storageRecord: Proto.IStorageRecord; }; type MergedRecordType = UnknownRecord & { hasConflict: boolean; shouldDrop: boolean; hasError: boolean; isUnsupported: boolean; updatedConversations: ReadonlyArray; needProfileFetch: ReadonlyArray; }; async function mergeRecord( storageVersion: number, itemToMerge: MergeableItemType ): Promise { const { itemType, storageID, storageRecord } = itemToMerge; const redactedStorageID = redactExtendedStorageID({ storageID, storageVersion, }); const ITEM_TYPE = Proto.ManifestRecord.Identifier.Type; let mergeResult: MergeResultType = { hasConflict: false, details: [] }; let isUnsupported = false; let hasError = false; let updatedConversations = new Array(); const needProfileFetch = new Array(); try { if (itemType === ITEM_TYPE.UNKNOWN) { log.warn( 'storageService.mergeRecord: Unknown item type', redactedStorageID ); } else if (itemType === ITEM_TYPE.CONTACT && storageRecord.contact) { mergeResult = await mergeContactRecord( storageID, storageVersion, storageRecord.contact ); } else if (itemType === ITEM_TYPE.GROUPV1 && storageRecord.groupV1) { mergeResult = await mergeGroupV1Record( storageID, storageVersion, storageRecord.groupV1 ); } else if (itemType === ITEM_TYPE.GROUPV2 && storageRecord.groupV2) { mergeResult = await mergeGroupV2Record( storageID, storageVersion, storageRecord.groupV2 ); } else if (itemType === ITEM_TYPE.ACCOUNT && storageRecord.account) { mergeResult = await mergeAccountRecord( storageID, storageVersion, storageRecord.account ); } else if ( itemType === ITEM_TYPE.STORY_DISTRIBUTION_LIST && storageRecord.storyDistributionList ) { mergeResult = await mergeStoryDistributionListRecord( storageID, storageVersion, storageRecord.storyDistributionList ); } else if ( itemType === ITEM_TYPE.STICKER_PACK && storageRecord.stickerPack ) { mergeResult = await mergeStickerPackRecord( storageID, storageVersion, storageRecord.stickerPack ); } else if (itemType === ITEM_TYPE.CALL_LINK && storageRecord.callLink) { mergeResult = await mergeCallLinkRecord( storageID, storageVersion, storageRecord.callLink ); } else { isUnsupported = true; log.warn( `storageService.merge(${redactedStorageID}): unknown item type=${itemType}` ); } const redactedID = redactStorageID( storageID, storageVersion, mergeResult.conversation ); const oldID = mergeResult.oldStorageID ? redactStorageID(mergeResult.oldStorageID, mergeResult.oldStorageVersion) : '?'; updatedConversations = [ ...updatedConversations, ...(mergeResult.updatedConversations ?? []), ]; if (mergeResult.needsProfileFetch) { strictAssert(mergeResult.conversation, 'needsProfileFetch, but no convo'); needProfileFetch.push(mergeResult.conversation); } log.info( `storageService.merge(${redactedID}): merged item type=${itemType} ` + `oldID=${oldID} ` + `conflict=${mergeResult.hasConflict} ` + `shouldDrop=${Boolean(mergeResult.shouldDrop)} ` + `details=${JSON.stringify(mergeResult.details)}` ); } catch (err) { hasError = true; const redactedID = redactStorageID(storageID, storageVersion); log.error( `storageService.merge(${redactedID}): error with ` + `item type=${itemType} ` + `details=${Errors.toLogFormat(err)}` ); } return { hasConflict: mergeResult.hasConflict, shouldDrop: Boolean(mergeResult.shouldDrop), hasError, isUnsupported, itemType, storageID, updatedConversations, needProfileFetch, }; } type NonConversationRecordsResultType = Readonly<{ callLinkDbRecords: ReadonlyArray; installedStickerPacks: ReadonlyArray; uninstalledStickerPacks: ReadonlyArray; storyDistributionLists: ReadonlyArray; }>; // TODO: DESKTOP-3929 async function getNonConversationRecords(): Promise { const [ callLinkDbRecords, storyDistributionLists, uninstalledStickerPacks, installedStickerPacks, ] = await Promise.all([ DataReader.getAllCallLinkRecordsWithAdminKey(), DataReader.getAllStoryDistributionsWithMembers(), DataReader.getUninstalledStickerPacks(), DataReader.getInstalledStickerPacks(), ]); return { callLinkDbRecords, storyDistributionLists, uninstalledStickerPacks, installedStickerPacks, }; } async function processManifest( manifest: Proto.IManifestRecord, version: number ): Promise { if (!window.textsecure.messaging) { throw new Error('storageService.processManifest: We are offline!'); } const remoteKeysTypeMap = new Map(); (manifest.keys || []).forEach(({ raw, type }: IManifestRecordIdentifier) => { strictAssert(raw, 'Identifier without raw field'); remoteKeysTypeMap.set(Bytes.toBase64(raw), type); }); const remoteKeys = new Set(remoteKeysTypeMap.keys()); const localVersions = new Map(); let localRecordCount = 0; const conversations = window.getConversations(); conversations.forEach((conversation: ConversationModel) => { const storageID = conversation.get('storageID'); if (storageID) { localVersions.set(storageID, conversation.get('storageVersion')); } }); localRecordCount += conversations.length; { const { callLinkDbRecords, storyDistributionLists, installedStickerPacks, uninstalledStickerPacks, } = await getNonConversationRecords(); const collectLocalKeysFromFields = ({ storageID, storageVersion, }: StorageServiceFieldsType): void => { if (storageID) { localVersions.set(storageID, storageVersion); } }; callLinkDbRecords.forEach(dbRecord => collectLocalKeysFromFields(callLinkFromRecord(dbRecord)) ); localRecordCount += callLinkDbRecords.length; storyDistributionLists.forEach(collectLocalKeysFromFields); localRecordCount += storyDistributionLists.length; uninstalledStickerPacks.forEach(collectLocalKeysFromFields); localRecordCount += uninstalledStickerPacks.length; installedStickerPacks.forEach(collectLocalKeysFromFields); localRecordCount += installedStickerPacks.length; } const unknownRecordsArray: ReadonlyArray = window.storage.get('storage-service-unknown-records') || []; const stillUnknown = unknownRecordsArray.filter((record: UnknownRecord) => { // Do not include any unknown records that we already support if (!validRecordTypes.has(record.itemType)) { localVersions.set(record.storageID, record.storageVersion); return false; } return true; }); const remoteOnlySet = new Set(); for (const key of remoteKeys) { if (!localVersions.has(key)) { remoteOnlySet.add(key); } } const localOnlySet = new Set(); for (const key of localVersions.keys()) { if (!remoteKeys.has(key)) { localOnlySet.add(key); } } const redactedRemoteOnly = Array.from(remoteOnlySet).map(id => redactStorageID(id, version) ); const redactedLocalOnly = Array.from(localOnlySet).map(id => redactStorageID(id, localVersions.get(id)) ); log.info( `storageService.process(${version}): localRecords=${localRecordCount} ` + `localKeys=${localVersions.size} unknownKeys=${stillUnknown.length} ` + `remoteKeys=${remoteKeys.size}` ); log.info( `storageService.process(${version}): ` + `remoteOnlyCount=${remoteOnlySet.size} ` + `remoteOnlyKeys=${JSON.stringify(redactedRemoteOnly)}` ); log.info( `storageService.process(${version}): ` + `localOnlyCount=${localOnlySet.size} ` + `localOnlyKeys=${JSON.stringify(redactedLocalOnly)}` ); const remoteOnlyRecords = new Map(); remoteOnlySet.forEach(storageID => { remoteOnlyRecords.set(storageID, { storageID, itemType: remoteKeysTypeMap.get(storageID), }); }); let conflictCount = 0; if (remoteOnlyRecords.size) { const fetchResult = await fetchRemoteRecords(version, remoteOnlyRecords); conflictCount = await processRemoteRecords(version, fetchResult); } // Post-merge, if our local records contain any storage IDs that were not // present in the remote manifest then we'll need to clear it, generate a // new storageID for that record, and upload. // This might happen if a device pushes a manifest which doesn't contain // the keys that we have in our local database. window.getConversations().forEach((conversation: ConversationModel) => { const storageID = conversation.get('storageID'); if (storageID && !remoteKeys.has(storageID)) { const storageVersion = conversation.get('storageVersion'); const missingKey = redactStorageID( storageID, storageVersion, conversation ); // Remote might have dropped this conversation already, but our value of // `firstUnregisteredAt` is too high for us to drop it. Don't reupload it! if ( isDirectConversation(conversation.attributes) && conversation.isUnregistered() ) { log.info( `storageService.process(${version}): localKey=${missingKey} is ` + 'unregistered and not in remote manifest' ); conversation.setUnregistered({ timestamp: Date.now() - durations.MONTH, fromStorageService: true, // Saving below shouldSave: false, }); } else { log.info( `storageService.process(${version}): localKey=${missingKey} ` + 'was not in remote manifest' ); } conversation.unset('storageID'); conversation.unset('storageVersion'); drop(updateConversation(conversation.attributes)); } }); // Refetch various records post-merge { const { callLinkDbRecords, storyDistributionLists, installedStickerPacks, uninstalledStickerPacks, } = await getNonConversationRecords(); uninstalledStickerPacks.forEach(stickerPack => { const { storageID, storageVersion } = stickerPack; if (!storageID || remoteKeys.has(storageID)) { return; } const missingKey = redactStorageID(storageID, storageVersion); log.info( `storageService.process(${version}): localKey=${missingKey} was not ` + 'in remote manifest' ); void DataWriter.addUninstalledStickerPack({ ...stickerPack, storageID: undefined, storageVersion: undefined, }); }); installedStickerPacks.forEach(stickerPack => { const { storageID, storageVersion } = stickerPack; if (!storageID || remoteKeys.has(storageID)) { return; } const missingKey = redactStorageID(storageID, storageVersion); log.info( `storageService.process(${version}): localKey=${missingKey} was not ` + 'in remote manifest' ); void DataWriter.createOrUpdateStickerPack({ ...stickerPack, storageID: undefined, storageVersion: undefined, }); }); storyDistributionLists.forEach(storyDistributionList => { const { storageID, storageVersion } = storyDistributionList; if (!storageID || remoteKeys.has(storageID)) { return; } const missingKey = redactStorageID(storageID, storageVersion); log.info( `storageService.process(${version}): localKey=${missingKey} was not ` + 'in remote manifest' ); void DataWriter.modifyStoryDistribution({ ...storyDistributionList, storageID: undefined, storageVersion: undefined, }); }); // Check to make sure we have a "My Stories" distribution list set up const myStories = storyDistributionLists.find( ({ id }) => id === MY_STORY_ID ); if (!myStories) { log.info(`storageService.process(${version}): creating my stories`); const storyDistribution: StoryDistributionWithMembersType = { allowsReplies: true, id: MY_STORY_ID, isBlockList: true, members: [], name: MY_STORY_ID, senderKeyInfo: undefined, storageNeedsSync: true, }; await DataWriter.createNewStoryDistribution(storyDistribution); const shouldSave = false; window.reduxActions.storyDistributionLists.createDistributionList( storyDistribution.name, storyDistribution.members, storyDistribution, shouldSave ); conflictCount += 1; } callLinkDbRecords.forEach(callLinkDbRecord => { const { storageID, storageVersion } = callLinkDbRecord; if (!storageID || remoteKeys.has(storageID)) { return; } const missingKey = redactStorageID( storageID, storageVersion || undefined ); log.info( `storageService.process(${version}): localKey=${missingKey} was not ` + 'in remote manifest' ); const callLink = callLinkFromRecord(callLinkDbRecord); drop( DataWriter.updateCallLink({ ...callLink, storageID: undefined, storageVersion: undefined, }) ); }); } log.info( `storageService.process(${version}): conflictCount=${conflictCount}` ); return conflictCount; } export type FetchRemoteRecordsResultType = Readonly<{ missingKeys: Set; decryptedItems: ReadonlyArray; }>; async function fetchRemoteRecords( storageVersion: number, remoteOnlyRecords: Map ): Promise { const storageKeyBase64 = window.storage.get('storageKey'); if (!storageKeyBase64) { throw new Error('No storage key'); } const { messaging } = window.textsecure; if (!messaging) { throw new Error('messaging is not available'); } const storageKey = Bytes.fromBase64(storageKeyBase64); log.info( `storageService.fetchRemoteRecords(${storageVersion}): ` + `fetching remote keys count=${remoteOnlyRecords.size}` ); const credentials = window.storage.get('storageCredentials'); const batches = chunk(Array.from(remoteOnlyRecords.keys()), MAX_READ_KEYS); const storageItems = ( await pMap( batches, async ( batch: ReadonlyArray ): Promise> => { const readOperation = new Proto.ReadOperation(); readOperation.readKey = batch.map(Bytes.fromBase64); const storageItemsBuffer = await messaging.getStorageRecords( Proto.ReadOperation.encode(readOperation).finish(), { credentials, } ); return Proto.StorageItems.decode(storageItemsBuffer).items ?? []; }, { concurrency: 5 } ) ).flat(); const missingKeys = new Set(remoteOnlyRecords.keys()); const decryptedItems = await pMap( storageItems, async ( storageRecordWrapper: Proto.IStorageItem ): Promise => { const { key, value: storageItemCiphertext } = storageRecordWrapper; if (!key || !storageItemCiphertext) { const error = new Error( `storageService.process(${storageVersion}): ` + 'missing key and/or Ciphertext' ); await stopStorageServiceSync(error); throw error; } const base64ItemID = Bytes.toBase64(key); missingKeys.delete(base64ItemID); const storageItemKey = deriveStorageItemKey(storageKey, base64ItemID); let storageItemPlaintext; try { storageItemPlaintext = decryptProfile( storageItemCiphertext, storageItemKey ); } catch (err) { log.error( `storageService.process(${storageVersion}): ` + 'Error decrypting storage item', Errors.toLogFormat(err) ); await stopStorageServiceSync(err); throw err; } const storageRecord = Proto.StorageRecord.decode(storageItemPlaintext); const remoteRecord = remoteOnlyRecords.get(base64ItemID); if (!remoteRecord) { const redactedStorageID = redactExtendedStorageID({ storageID: base64ItemID, storageVersion, }); throw new Error( "Got a remote record that wasn't requested with " + `storageID: ${redactedStorageID}` ); } return { itemType: remoteRecord.itemType, storageID: base64ItemID, storageRecord, }; }, { concurrency: 5 } ); const redactedMissingKeys = Array.from(missingKeys).map(id => redactStorageID(id, storageVersion) ); log.info( `storageService.fetchRemoteRecords(${storageVersion}): missing remote ` + `keys=${JSON.stringify(redactedMissingKeys)} ` + `count=${missingKeys.size}` ); return { decryptedItems, missingKeys }; } async function processRemoteRecords( storageVersion: number, { decryptedItems, missingKeys }: FetchRemoteRecordsResultType ): Promise { const ITEM_TYPE = Proto.ManifestRecord.Identifier.Type; const droppedKeys = new Set(); // Drop all GV1 records for which we have GV2 record in the same manifest const masterKeys = new Map(); for (const { itemType, storageID, storageRecord } of decryptedItems) { if (itemType === ITEM_TYPE.GROUPV2 && storageRecord.groupV2?.masterKey) { masterKeys.set( Bytes.toBase64(storageRecord.groupV2.masterKey), storageID ); } } let accountItem: MergeableItemType | undefined; let prunedStorageItems = decryptedItems.filter(item => { const { itemType, storageID, storageRecord } = item; if (itemType === ITEM_TYPE.ACCOUNT) { if (accountItem !== undefined) { log.warn( `storageService.process(${storageVersion}): duplicate account ` + `record=${redactStorageID(storageID, storageVersion)} ` + `previous=${redactStorageID(accountItem.storageID, storageVersion)}` ); droppedKeys.add(accountItem.storageID); } accountItem = item; return false; } if (itemType !== ITEM_TYPE.GROUPV1 || !storageRecord.groupV1?.id) { return true; } const masterKey = deriveMasterKeyFromGroupV1(storageRecord.groupV1.id); const gv2StorageID = masterKeys.get(Bytes.toBase64(masterKey)); if (!gv2StorageID) { return true; } log.warn( `storageService.process(${storageVersion}): dropping ` + `GV1 record=${redactStorageID(storageID, storageVersion)} ` + `GV2 record=${redactStorageID(gv2StorageID, storageVersion)} ` + 'is in the same manifest' ); droppedKeys.add(storageID); return false; }); // Find remote contact records that: // - Have `remote.pni` and have `remote.serviceE164` // - Match local contact that has `aci`. const splitPNIContacts = new Array(); prunedStorageItems = prunedStorageItems.filter(item => { const { itemType, storageRecord } = item; const { contact } = storageRecord; if (itemType !== ITEM_TYPE.CONTACT || !contact) { return true; } if (!contact.serviceE164 || !contact.pni) { return true; } const localAci = window.ConversationController.get(contact.pni)?.getAci(); if (!localAci) { return true; } splitPNIContacts.push(item); return false; }); try { log.info( `storageService.process(${storageVersion}): ` + `attempting to merge records=${prunedStorageItems.length}` ); if (accountItem !== undefined) { log.info( `storageService.process(${storageVersion}): account ` + `record=${redactStorageID(accountItem.storageID, storageVersion)}` ); } if (splitPNIContacts.length !== 0) { log.info( `storageService.process(${storageVersion}): ` + `split pni contacts=${splitPNIContacts.length}` ); } const mergeWithConcurrency = ( items: ReadonlyArray ): Promise> => { return pMap( items, (item: MergeableItemType) => mergeRecord(storageVersion, item), { concurrency: 32 } ); }; const mergedRecords = [ ...(await mergeWithConcurrency(prunedStorageItems)), // Merge split PNI contacts after processing remote records. If original // e164+ACI+PNI contact is unregistered - it is going to be split so we // have to make that happen first. Otherwise we will ignore ContactRecord // changes on these since there is already a parent "merged" contact. ...(await mergeWithConcurrency(splitPNIContacts)), // Merge Account records last since it contains the pinned conversations // and we need all other records merged first before we can find the pinned // records in our db ...(accountItem ? [await mergeRecord(storageVersion, accountItem)] : []), ]; log.info( `storageService.process(${storageVersion}): ` + `processed records=${mergedRecords.length}` ); const updatedConversations = mergedRecords .map(record => record.updatedConversations) .flat() .map(convo => convo.attributes); await updateConversations(updatedConversations); log.info( `storageService.process(${storageVersion}): ` + `updated conversations=${updatedConversations.length}` ); const needProfileFetch = mergedRecords .map(record => record.needProfileFetch) .flat(); log.info( `storageService.process(${storageVersion}): ` + `kicking off profile fetches=${needProfileFetch.length}` ); // Intentionally not awaiting needProfileFetch.map(convo => drop( convo.getProfiles().catch(() => { /* nothing to do here; logging already happened */ }) ) ); // Collect full map of previously and currently unknown records const unknownRecords: Map = new Map(); const previousUnknownRecords: ReadonlyArray = window.storage.get( 'storage-service-unknown-records', new Array() ); previousUnknownRecords.forEach((record: UnknownRecord) => { unknownRecords.set(record.storageID, record); }); const newRecordsWithErrors: Array = []; let conflictCount = 0; mergedRecords.forEach((mergedRecord: MergedRecordType) => { if (mergedRecord.isUnsupported) { unknownRecords.set(mergedRecord.storageID, { itemType: mergedRecord.itemType, storageID: mergedRecord.storageID, storageVersion, }); } else if (mergedRecord.hasError) { newRecordsWithErrors.push({ itemType: mergedRecord.itemType, storageID: mergedRecord.storageID, storageVersion, }); } if (mergedRecord.hasConflict) { conflictCount += 1; } if (mergedRecord.shouldDrop) { droppedKeys.add(mergedRecord.storageID); } }); const redactedDroppedKeys = Array.from(droppedKeys.values()).map(key => redactStorageID(key, storageVersion) ); log.info( `storageService.process(${storageVersion}): ` + `dropped keys=${JSON.stringify(redactedDroppedKeys)} ` + `count=${redactedDroppedKeys.length}` ); // Filter out all the unknown records we're already supporting const newUnknownRecords = Array.from(unknownRecords.values()).filter( (record: UnknownRecord) => !validRecordTypes.has(record.itemType) ); const redactedNewUnknowns = newUnknownRecords.map(redactExtendedStorageID); log.info( `storageService.process(${storageVersion}): ` + `unknown records=${JSON.stringify(redactedNewUnknowns)} ` + `count=${redactedNewUnknowns.length}` ); await window.storage.put( 'storage-service-unknown-records', newUnknownRecords ); const redactedErrorRecords = newRecordsWithErrors.map( redactExtendedStorageID ); log.info( `storageService.process(${storageVersion}): ` + `error records=${JSON.stringify(redactedErrorRecords)} ` + `count=${redactedErrorRecords.length}` ); // Refresh the list of records that had errors with every push, that way // this list doesn't grow unbounded and we keep the list of storage keys // fresh. await window.storage.put( 'storage-service-error-records', newRecordsWithErrors ); // Store/overwrite keys pending deletion, but use them only when we have to // upload a new manifest to avoid oscillation. const pendingDeletes = [...missingKeys, ...droppedKeys].map(storageID => ({ storageID, storageVersion, })); const redactedPendingDeletes = pendingDeletes.map(redactExtendedStorageID); log.info( `storageService.process(${storageVersion}): ` + `pending deletes=${JSON.stringify(redactedPendingDeletes)} ` + `count=${redactedPendingDeletes.length}` ); await window.storage.put('storage-service-pending-deletes', pendingDeletes); if (conflictCount === 0) { conflictBackOff.reset(); } return conflictCount; } catch (err) { log.error( `storageService.process(${storageVersion}): ` + 'failed to process remote records', Errors.toLogFormat(err) ); } // conflictCount return 0; } async function sync({ ignoreConflicts = false, reason, }: { ignoreConflicts?: boolean; reason: string; }): Promise { if (!window.storage.get('storageKey')) { const masterKeyBase64 = window.storage.get('masterKey'); if (!masterKeyBase64) { log.error( `storageService.sync(${reason}): Cannot start; no storage or master key!` ); return; } const masterKey = Bytes.fromBase64(masterKeyBase64); const storageKeyBase64 = Bytes.toBase64(deriveStorageServiceKey(masterKey)); await window.storage.put('storageKey', storageKeyBase64); log.warn('storageService.sync: fixed storage key'); } log.info( `storageService.sync: starting... ignoreConflicts=${ignoreConflicts}, reason=${reason}` ); let manifest: Proto.ManifestRecord | undefined; try { // If we've previously interacted with storage service, update 'fetchComplete' record const previousFetchComplete = window.storage.get('storageFetchComplete'); const manifestFromStorage = window.storage.get('manifestVersion'); if (!previousFetchComplete && isNumber(manifestFromStorage)) { await window.storage.put('storageFetchComplete', true); } const localManifestVersion = manifestFromStorage || 0; log.info( 'storageService.sync: fetching latest ' + `after version=${localManifestVersion}` ); manifest = await fetchManifest(localManifestVersion); // Guarding against no manifests being returned, everything should be ok if (!manifest) { log.info( `storageService.sync: no updates, version=${localManifestVersion}` ); return undefined; } strictAssert(manifest.version != null, 'Manifest without version'); const version = manifest.version?.toNumber() ?? 0; log.info( `storageService.sync: updating to remoteVersion=${version} ` + `sourceDevice=${manifest.sourceDevice ?? '?'} from ` + `version=${localManifestVersion}` ); const conflictCount = await processManifest(manifest, version); log.info( `storageService.sync: updated to version=${version} ` + `conflicts=${conflictCount}` ); await window.storage.put('manifestVersion', version); const hasConflicts = conflictCount !== 0; if (hasConflicts && !ignoreConflicts) { await upload({ fromSync: true, reason: `sync/${reason}` }); } // We now know that we've successfully completed a storage service fetch await window.storage.put('storageFetchComplete', true); if (window.SignalCI) { window.SignalCI.handleEvent('storageServiceComplete', { manifestVersion: version, }); } } catch (err) { log.error( 'storageService.sync: error processing manifest', Errors.toLogFormat(err) ); } log.info('storageService.sync: complete'); return manifest; } async function upload({ fromSync = false, reason, }: { fromSync?: boolean; reason: string; }): Promise { const logId = `storageService.upload/${reason}`; if (!window.textsecure.messaging) { throw new Error(`${logId}: We are offline!`); } // Rate limit uploads coming from syncing if (fromSync) { uploadBucket.push(Date.now()); if (uploadBucket.length >= 3) { const [firstMostRecentWrite] = uploadBucket; if (isMoreRecentThan(5 * durations.MINUTE, firstMostRecentWrite)) { throw new Error(`${logId}: too many writes too soon.`); } uploadBucket.shift(); } } if (!window.storage.get('storageKey')) { // requesting new keys runs the sync job which will detect the conflict // and re-run the upload job once we're merged and up-to-date. log.info(`${logId}: no storageKey, requesting new keys`); backOff.reset(); if (window.ConversationController.areWePrimaryDevice()) { log.warn(`${logId}: We are primary device; not sending key sync request`); return; } try { await singleProtoJobQueue.add(MessageSender.getRequestKeySyncMessage()); } catch (error) { log.error( `${logId}: Failed to queue sync message`, Errors.toLogFormat(error) ); } return; } let previousManifest: Proto.ManifestRecord | undefined; if (!fromSync) { // Syncing before we upload so that we repair any unknown records and // records with errors as well as ensure that we have the latest up to date // manifest. // We are going to upload after this sync so we can ignore any conflicts // that arise during the sync. const ignoreConflicts = true; previousManifest = await sync({ ignoreConflicts, reason: `upload/${reason}`, }); } const localManifestVersion = window.storage.get('manifestVersion', 0); const version = Number(localManifestVersion) + 1; log.info(`${logId}/${version}: will update to manifest version`); try { const generatedManifest = await generateManifest( version, previousManifest, false ); const encryptedManifest = await encryptManifest(version, generatedManifest); await uploadManifest(version, generatedManifest, encryptedManifest); // Clear pending delete keys after successful upload await window.storage.put('storage-service-pending-deletes', []); } catch (err) { if (err.code === 409) { await sleep(conflictBackOff.getAndIncrement()); log.info(`${logId}: pushing sync on the queue`); // The sync job will check for conflicts and as part of that conflict // check if an item needs sync and doesn't match with the remote record // it'll kick off another upload. setTimeout(runStorageServiceSyncJob); return; } log.error(`${logId}/${version}: error`, Errors.toLogFormat(err)); } } let storageServiceEnabled = false; export function enableStorageService(): void { storageServiceEnabled = true; } export function disableStorageService(): void { storageServiceEnabled = false; } export async function eraseAllStorageServiceState({ keepUnknownFields = false, }: { keepUnknownFields?: boolean } = {}): Promise { log.info('storageService.eraseAllStorageServiceState: starting...'); // First, update high-level storage service metadata await Promise.all([ window.storage.remove('manifestVersion'), keepUnknownFields ? Promise.resolve() : window.storage.remove('storage-service-unknown-records'), window.storage.remove('storageCredentials'), ]); // Then, we make the changes to records in memory: // - Conversations // - Sticker packs // - Uninstalled sticker packs // - Story distribution lists // This call just erases stickers for now. Storage service data is not stored // in memory for Story Distribution Lists. Uninstalled sticker packs are not // kept in memory at all. window.reduxActions.user.eraseStorageServiceState(); // Conversations. These properties are not present in redux. window.getConversations().forEach(conversation => { conversation.unset('storageID'); conversation.unset('needsStorageServiceSync'); conversation.unset('storageUnknownFields'); }); // Then make sure outstanding conversation saves are flushed await flushUpdateConversationBatcher(); // Then make sure that all previously-outstanding database saves are flushed await getItemById('manifestVersion'); // Finally, we update the database directly for all record types: await eraseStorageServiceState(); log.info('storageService.eraseAllStorageServiceState: complete'); } export async function reprocessUnknownFields(): Promise { ourProfileKeyService.blockGetWithPromise( storageJobQueue(async () => { const version = window.storage.get('manifestVersion') ?? 0; log.info(`storageService.reprocessUnknownFields(${version}): starting`); const { recordsByID, insertKeys } = await generateManifest( version, undefined, true ); const newRecords = Array.from( filter( map(recordsByID, ([key, item]): MergeableItemType | undefined => { if (!insertKeys.has(key)) { return undefined; } strictAssert( item.storageRecord !== undefined, 'Inserted records must have storageRecord' ); if (!item.storageRecord.$unknownFields?.length) { return undefined; } return { ...item, storageRecord: Proto.StorageRecord.decode( Proto.StorageRecord.encode(item.storageRecord).finish() ), }; }), isNotNil ) ); const conflictCount = await processRemoteRecords(version, { decryptedItems: newRecords, missingKeys: new Set(), }); log.info( `storageService.reprocessUnknownFields(${version}): done, ` + `conflictCount=${conflictCount}` ); const hasConflicts = conflictCount !== 0; if (hasConflicts) { log.info( `storageService.reprocessUnknownFields(${version}): uploading` ); await upload({ reason: 'reprocessUnknownFields/hasConflicts' }); } }) ); } export const storageServiceUploadJob = debounce( ({ reason }: { reason: string }) => { if (!storageServiceEnabled) { log.info('storageService.storageServiceUploadJob: called before enabled'); return; } void storageJobQueue( async () => { await upload({ reason: `storageServiceUploadJob/${reason}` }); }, `upload v${window.storage.get('manifestVersion')}` ); }, 500 ); export const runStorageServiceSyncJob = debounce( ({ reason }: { reason: string }) => { if (!storageServiceEnabled) { log.info( 'storageService.runStorageServiceSyncJob: called before enabled' ); return; } ourProfileKeyService.blockGetWithPromise( storageJobQueue( async () => { await sync({ reason }); // Notify listeners about sync completion window.Whisper.events.trigger('storageService:syncComplete'); }, `sync v${window.storage.get('manifestVersion')}` ) ); }, 500 ); export const addPendingDelete = (item: ExtendedStorageID): void => { void storageJobQueue( async () => { const storedPendingDeletes = window.storage.get( 'storage-service-pending-deletes', [] ); await window.storage.put('storage-service-pending-deletes', [ ...storedPendingDeletes, item, ]); }, `addPendingDelete(${redactExtendedStorageID(item)})` ); };