You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
822 lines
24 KiB
822 lines
24 KiB
this.workbox = this.workbox || {}; |
|
this.workbox.backgroundSync = (function (exports, WorkboxError_mjs, logger_mjs, assert_mjs, getFriendlyURL_mjs, DBWrapper_mjs) { |
|
'use strict'; |
|
|
|
try { |
|
self['workbox:background-sync:4.3.1'] && _(); |
|
} catch (e) {} // eslint-disable-line |
|
|
|
/* |
|
Copyright 2018 Google LLC |
|
|
|
Use of this source code is governed by an MIT-style |
|
license that can be found in the LICENSE file or at |
|
https://opensource.org/licenses/MIT. |
|
*/ |
|
const DB_VERSION = 3; |
|
const DB_NAME = 'workbox-background-sync'; |
|
const OBJECT_STORE_NAME = 'requests'; |
|
const INDEXED_PROP = 'queueName'; |
|
/** |
|
* A class to manage storing requests from a Queue in IndexedbDB, |
|
* indexed by their queue name for easier access. |
|
* |
|
* @private |
|
*/ |
|
|
|
class QueueStore { |
|
/** |
|
* Associates this instance with a Queue instance, so entries added can be |
|
* identified by their queue name. |
|
* |
|
* @param {string} queueName |
|
* @private |
|
*/ |
|
constructor(queueName) { |
|
this._queueName = queueName; |
|
this._db = new DBWrapper_mjs.DBWrapper(DB_NAME, DB_VERSION, { |
|
onupgradeneeded: this._upgradeDb |
|
}); |
|
} |
|
/** |
|
* Append an entry last in the queue. |
|
* |
|
* @param {Object} entry |
|
* @param {Object} entry.requestData |
|
* @param {number} [entry.timestamp] |
|
* @param {Object} [entry.metadata] |
|
* @private |
|
*/ |
|
|
|
|
|
async pushEntry(entry) { |
|
{ |
|
assert_mjs.assert.isType(entry, 'object', { |
|
moduleName: 'workbox-background-sync', |
|
className: 'QueueStore', |
|
funcName: 'pushEntry', |
|
paramName: 'entry' |
|
}); |
|
assert_mjs.assert.isType(entry.requestData, 'object', { |
|
moduleName: 'workbox-background-sync', |
|
className: 'QueueStore', |
|
funcName: 'pushEntry', |
|
paramName: 'entry.requestData' |
|
}); |
|
} // Don't specify an ID since one is automatically generated. |
|
|
|
|
|
delete entry.id; |
|
entry.queueName = this._queueName; |
|
await this._db.add(OBJECT_STORE_NAME, entry); |
|
} |
|
/** |
|
* Preppend an entry first in the queue. |
|
* |
|
* @param {Object} entry |
|
* @param {Object} entry.requestData |
|
* @param {number} [entry.timestamp] |
|
* @param {Object} [entry.metadata] |
|
* @private |
|
*/ |
|
|
|
|
|
async unshiftEntry(entry) { |
|
{ |
|
assert_mjs.assert.isType(entry, 'object', { |
|
moduleName: 'workbox-background-sync', |
|
className: 'QueueStore', |
|
funcName: 'unshiftEntry', |
|
paramName: 'entry' |
|
}); |
|
assert_mjs.assert.isType(entry.requestData, 'object', { |
|
moduleName: 'workbox-background-sync', |
|
className: 'QueueStore', |
|
funcName: 'unshiftEntry', |
|
paramName: 'entry.requestData' |
|
}); |
|
} |
|
|
|
const [firstEntry] = await this._db.getAllMatching(OBJECT_STORE_NAME, { |
|
count: 1 |
|
}); |
|
|
|
if (firstEntry) { |
|
// Pick an ID one less than the lowest ID in the object store. |
|
entry.id = firstEntry.id - 1; |
|
} else { |
|
// Otherwise let the auto-incrementor assign the ID. |
|
delete entry.id; |
|
} |
|
|
|
entry.queueName = this._queueName; |
|
await this._db.add(OBJECT_STORE_NAME, entry); |
|
} |
|
/** |
|
* Removes and returns the last entry in the queue matching the `queueName`. |
|
* |
|
* @return {Promise<Object>} |
|
* @private |
|
*/ |
|
|
|
|
|
async popEntry() { |
|
return this._removeEntry({ |
|
direction: 'prev' |
|
}); |
|
} |
|
/** |
|
* Removes and returns the first entry in the queue matching the `queueName`. |
|
* |
|
* @return {Promise<Object>} |
|
* @private |
|
*/ |
|
|
|
|
|
async shiftEntry() { |
|
return this._removeEntry({ |
|
direction: 'next' |
|
}); |
|
} |
|
/** |
|
* Returns all entries in the store matching the `queueName`. |
|
* |
|
* @param {Object} options See workbox.backgroundSync.Queue~getAll} |
|
* @return {Promise<Array<Object>>} |
|
* @private |
|
*/ |
|
|
|
|
|
async getAll() { |
|
return await this._db.getAllMatching(OBJECT_STORE_NAME, { |
|
index: INDEXED_PROP, |
|
query: IDBKeyRange.only(this._queueName) |
|
}); |
|
} |
|
/** |
|
* Deletes the entry for the given ID. |
|
* |
|
* WARNING: this method does not ensure the deleted enry belongs to this |
|
* queue (i.e. matches the `queueName`). But this limitation is acceptable |
|
* as this class is not publicly exposed. An additional check would make |
|
* this method slower than it needs to be. |
|
* |
|
* @private |
|
* @param {number} id |
|
*/ |
|
|
|
|
|
async deleteEntry(id) { |
|
await this._db.delete(OBJECT_STORE_NAME, id); |
|
} |
|
/** |
|
* Removes and returns the first or last entry in the queue (based on the |
|
* `direction` argument) matching the `queueName`. |
|
* |
|
* @return {Promise<Object>} |
|
* @private |
|
*/ |
|
|
|
|
|
async _removeEntry({ |
|
direction |
|
}) { |
|
const [entry] = await this._db.getAllMatching(OBJECT_STORE_NAME, { |
|
direction, |
|
index: INDEXED_PROP, |
|
query: IDBKeyRange.only(this._queueName), |
|
count: 1 |
|
}); |
|
|
|
if (entry) { |
|
await this.deleteEntry(entry.id); |
|
return entry; |
|
} |
|
} |
|
/** |
|
* Upgrades the database given an `upgradeneeded` event. |
|
* |
|
* @param {Event} event |
|
* @private |
|
*/ |
|
|
|
|
|
_upgradeDb(event) { |
|
const db = event.target.result; |
|
|
|
if (event.oldVersion > 0 && event.oldVersion < DB_VERSION) { |
|
if (db.objectStoreNames.contains(OBJECT_STORE_NAME)) { |
|
db.deleteObjectStore(OBJECT_STORE_NAME); |
|
} |
|
} |
|
|
|
const objStore = db.createObjectStore(OBJECT_STORE_NAME, { |
|
autoIncrement: true, |
|
keyPath: 'id' |
|
}); |
|
objStore.createIndex(INDEXED_PROP, INDEXED_PROP, { |
|
unique: false |
|
}); |
|
} |
|
|
|
} |
|
|
|
/* |
|
Copyright 2018 Google LLC |
|
|
|
Use of this source code is governed by an MIT-style |
|
license that can be found in the LICENSE file or at |
|
https://opensource.org/licenses/MIT. |
|
*/ |
|
const serializableProperties = ['method', 'referrer', 'referrerPolicy', 'mode', 'credentials', 'cache', 'redirect', 'integrity', 'keepalive']; |
|
/** |
|
* A class to make it easier to serialize and de-serialize requests so they |
|
* can be stored in IndexedDB. |
|
* |
|
* @private |
|
*/ |
|
|
|
class StorableRequest { |
|
/** |
|
* Converts a Request object to a plain object that can be structured |
|
* cloned or JSON-stringified. |
|
* |
|
* @param {Request} request |
|
* @return {Promise<StorableRequest>} |
|
* |
|
* @private |
|
*/ |
|
static async fromRequest(request) { |
|
const requestData = { |
|
url: request.url, |
|
headers: {} |
|
}; // Set the body if present. |
|
|
|
if (request.method !== 'GET') { |
|
// Use ArrayBuffer to support non-text request bodies. |
|
// NOTE: we can't use Blobs becuse Safari doesn't support storing |
|
// Blobs in IndexedDB in some cases: |
|
// https://github.com/dfahlander/Dexie.js/issues/618#issuecomment-398348457 |
|
requestData.body = await request.clone().arrayBuffer(); |
|
} // Convert the headers from an iterable to an object. |
|
|
|
|
|
for (const [key, value] of request.headers.entries()) { |
|
requestData.headers[key] = value; |
|
} // Add all other serializable request properties |
|
|
|
|
|
for (const prop of serializableProperties) { |
|
if (request[prop] !== undefined) { |
|
requestData[prop] = request[prop]; |
|
} |
|
} |
|
|
|
return new StorableRequest(requestData); |
|
} |
|
/** |
|
* Accepts an object of request data that can be used to construct a |
|
* `Request` but can also be stored in IndexedDB. |
|
* |
|
* @param {Object} requestData An object of request data that includes the |
|
* `url` plus any relevant properties of |
|
* [requestInit]{@link https://fetch.spec.whatwg.org/#requestinit}. |
|
* @private |
|
*/ |
|
|
|
|
|
constructor(requestData) { |
|
{ |
|
assert_mjs.assert.isType(requestData, 'object', { |
|
moduleName: 'workbox-background-sync', |
|
className: 'StorableRequest', |
|
funcName: 'constructor', |
|
paramName: 'requestData' |
|
}); |
|
assert_mjs.assert.isType(requestData.url, 'string', { |
|
moduleName: 'workbox-background-sync', |
|
className: 'StorableRequest', |
|
funcName: 'constructor', |
|
paramName: 'requestData.url' |
|
}); |
|
} // If the request's mode is `navigate`, convert it to `same-origin` since |
|
// navigation requests can't be constructed via script. |
|
|
|
|
|
if (requestData.mode === 'navigate') { |
|
requestData.mode = 'same-origin'; |
|
} |
|
|
|
this._requestData = requestData; |
|
} |
|
/** |
|
* Returns a deep clone of the instances `_requestData` object. |
|
* |
|
* @return {Object} |
|
* |
|
* @private |
|
*/ |
|
|
|
|
|
toObject() { |
|
const requestData = Object.assign({}, this._requestData); |
|
requestData.headers = Object.assign({}, this._requestData.headers); |
|
|
|
if (requestData.body) { |
|
requestData.body = requestData.body.slice(0); |
|
} |
|
|
|
return requestData; |
|
} |
|
/** |
|
* Converts this instance to a Request. |
|
* |
|
* @return {Request} |
|
* |
|
* @private |
|
*/ |
|
|
|
|
|
toRequest() { |
|
return new Request(this._requestData.url, this._requestData); |
|
} |
|
/** |
|
* Creates and returns a deep clone of the instance. |
|
* |
|
* @return {StorableRequest} |
|
* |
|
* @private |
|
*/ |
|
|
|
|
|
clone() { |
|
return new StorableRequest(this.toObject()); |
|
} |
|
|
|
} |
|
|
|
/* |
|
Copyright 2018 Google LLC |
|
|
|
Use of this source code is governed by an MIT-style |
|
license that can be found in the LICENSE file or at |
|
https://opensource.org/licenses/MIT. |
|
*/ |
|
const TAG_PREFIX = 'workbox-background-sync'; |
|
const MAX_RETENTION_TIME = 60 * 24 * 7; // 7 days in minutes |
|
|
|
const queueNames = new Set(); |
|
/** |
|
* A class to manage storing failed requests in IndexedDB and retrying them |
|
* later. All parts of the storing and replaying process are observable via |
|
* callbacks. |
|
* |
|
* @memberof workbox.backgroundSync |
|
*/ |
|
|
|
class Queue { |
|
/** |
|
* Creates an instance of Queue with the given options |
|
* |
|
* @param {string} name The unique name for this queue. This name must be |
|
* unique as it's used to register sync events and store requests |
|
* in IndexedDB specific to this instance. An error will be thrown if |
|
* a duplicate name is detected. |
|
* @param {Object} [options] |
|
* @param {Function} [options.onSync] A function that gets invoked whenever |
|
* the 'sync' event fires. The function is invoked with an object |
|
* containing the `queue` property (referencing this instance), and you |
|
* can use the callback to customize the replay behavior of the queue. |
|
* When not set the `replayRequests()` method is called. |
|
* Note: if the replay fails after a sync event, make sure you throw an |
|
* error, so the browser knows to retry the sync event later. |
|
* @param {number} [options.maxRetentionTime=7 days] The amount of time (in |
|
* minutes) a request may be retried. After this amount of time has |
|
* passed, the request will be deleted from the queue. |
|
*/ |
|
constructor(name, { |
|
onSync, |
|
maxRetentionTime |
|
} = {}) { |
|
// Ensure the store name is not already being used |
|
if (queueNames.has(name)) { |
|
throw new WorkboxError_mjs.WorkboxError('duplicate-queue-name', { |
|
name |
|
}); |
|
} else { |
|
queueNames.add(name); |
|
} |
|
|
|
this._name = name; |
|
this._onSync = onSync || this.replayRequests; |
|
this._maxRetentionTime = maxRetentionTime || MAX_RETENTION_TIME; |
|
this._queueStore = new QueueStore(this._name); |
|
|
|
this._addSyncListener(); |
|
} |
|
/** |
|
* @return {string} |
|
*/ |
|
|
|
|
|
get name() { |
|
return this._name; |
|
} |
|
/** |
|
* Stores the passed request in IndexedDB (with its timestamp and any |
|
* metadata) at the end of the queue. |
|
* |
|
* @param {Object} entry |
|
* @param {Request} entry.request The request to store in the queue. |
|
* @param {Object} [entry.metadata] Any metadata you want associated with the |
|
* stored request. When requests are replayed you'll have access to this |
|
* metadata object in case you need to modify the request beforehand. |
|
* @param {number} [entry.timestamp] The timestamp (Epoch time in |
|
* milliseconds) when the request was first added to the queue. This is |
|
* used along with `maxRetentionTime` to remove outdated requests. In |
|
* general you don't need to set this value, as it's automatically set |
|
* for you (defaulting to `Date.now()`), but you can update it if you |
|
* don't want particular requests to expire. |
|
*/ |
|
|
|
|
|
async pushRequest(entry) { |
|
{ |
|
assert_mjs.assert.isType(entry, 'object', { |
|
moduleName: 'workbox-background-sync', |
|
className: 'Queue', |
|
funcName: 'pushRequest', |
|
paramName: 'entry' |
|
}); |
|
assert_mjs.assert.isInstance(entry.request, Request, { |
|
moduleName: 'workbox-background-sync', |
|
className: 'Queue', |
|
funcName: 'pushRequest', |
|
paramName: 'entry.request' |
|
}); |
|
} |
|
|
|
await this._addRequest(entry, 'push'); |
|
} |
|
/** |
|
* Stores the passed request in IndexedDB (with its timestamp and any |
|
* metadata) at the beginning of the queue. |
|
* |
|
* @param {Object} entry |
|
* @param {Request} entry.request The request to store in the queue. |
|
* @param {Object} [entry.metadata] Any metadata you want associated with the |
|
* stored request. When requests are replayed you'll have access to this |
|
* metadata object in case you need to modify the request beforehand. |
|
* @param {number} [entry.timestamp] The timestamp (Epoch time in |
|
* milliseconds) when the request was first added to the queue. This is |
|
* used along with `maxRetentionTime` to remove outdated requests. In |
|
* general you don't need to set this value, as it's automatically set |
|
* for you (defaulting to `Date.now()`), but you can update it if you |
|
* don't want particular requests to expire. |
|
*/ |
|
|
|
|
|
async unshiftRequest(entry) { |
|
{ |
|
assert_mjs.assert.isType(entry, 'object', { |
|
moduleName: 'workbox-background-sync', |
|
className: 'Queue', |
|
funcName: 'unshiftRequest', |
|
paramName: 'entry' |
|
}); |
|
assert_mjs.assert.isInstance(entry.request, Request, { |
|
moduleName: 'workbox-background-sync', |
|
className: 'Queue', |
|
funcName: 'unshiftRequest', |
|
paramName: 'entry.request' |
|
}); |
|
} |
|
|
|
await this._addRequest(entry, 'unshift'); |
|
} |
|
/** |
|
* Removes and returns the last request in the queue (along with its |
|
* timestamp and any metadata). The returned object takes the form: |
|
* `{request, timestamp, metadata}`. |
|
* |
|
* @return {Promise<Object>} |
|
*/ |
|
|
|
|
|
async popRequest() { |
|
return this._removeRequest('pop'); |
|
} |
|
/** |
|
* Removes and returns the first request in the queue (along with its |
|
* timestamp and any metadata). The returned object takes the form: |
|
* `{request, timestamp, metadata}`. |
|
* |
|
* @return {Promise<Object>} |
|
*/ |
|
|
|
|
|
async shiftRequest() { |
|
return this._removeRequest('shift'); |
|
} |
|
/** |
|
* Returns all the entries that have not expired (per `maxRetentionTime`). |
|
* Any expired entries are removed from the queue. |
|
* |
|
* @return {Promise<Array<Object>>} |
|
*/ |
|
|
|
|
|
async getAll() { |
|
const allEntries = await this._queueStore.getAll(); |
|
const now = Date.now(); |
|
const unexpiredEntries = []; |
|
|
|
for (const entry of allEntries) { |
|
// Ignore requests older than maxRetentionTime. Call this function |
|
// recursively until an unexpired request is found. |
|
const maxRetentionTimeInMs = this._maxRetentionTime * 60 * 1000; |
|
|
|
if (now - entry.timestamp > maxRetentionTimeInMs) { |
|
await this._queueStore.deleteEntry(entry.id); |
|
} else { |
|
unexpiredEntries.push(convertEntry(entry)); |
|
} |
|
} |
|
|
|
return unexpiredEntries; |
|
} |
|
/** |
|
* Adds the entry to the QueueStore and registers for a sync event. |
|
* |
|
* @param {Object} entry |
|
* @param {Request} entry.request |
|
* @param {Object} [entry.metadata] |
|
* @param {number} [entry.timestamp=Date.now()] |
|
* @param {string} operation ('push' or 'unshift') |
|
* @private |
|
*/ |
|
|
|
|
|
async _addRequest({ |
|
request, |
|
metadata, |
|
timestamp = Date.now() |
|
}, operation) { |
|
const storableRequest = await StorableRequest.fromRequest(request.clone()); |
|
const entry = { |
|
requestData: storableRequest.toObject(), |
|
timestamp |
|
}; // Only include metadata if it's present. |
|
|
|
if (metadata) { |
|
entry.metadata = metadata; |
|
} |
|
|
|
await this._queueStore[`${operation}Entry`](entry); |
|
|
|
{ |
|
logger_mjs.logger.log(`Request for '${getFriendlyURL_mjs.getFriendlyURL(request.url)}' has ` + `been added to background sync queue '${this._name}'.`); |
|
} // Don't register for a sync if we're in the middle of a sync. Instead, |
|
// we wait until the sync is complete and call register if |
|
// `this._requestsAddedDuringSync` is true. |
|
|
|
|
|
if (this._syncInProgress) { |
|
this._requestsAddedDuringSync = true; |
|
} else { |
|
await this.registerSync(); |
|
} |
|
} |
|
/** |
|
* Removes and returns the first or last (depending on `operation`) entry |
|
* from the QueueStore that's not older than the `maxRetentionTime`. |
|
* |
|
* @param {string} operation ('pop' or 'shift') |
|
* @return {Object|undefined} |
|
* @private |
|
*/ |
|
|
|
|
|
async _removeRequest(operation) { |
|
const now = Date.now(); |
|
const entry = await this._queueStore[`${operation}Entry`](); |
|
|
|
if (entry) { |
|
// Ignore requests older than maxRetentionTime. Call this function |
|
// recursively until an unexpired request is found. |
|
const maxRetentionTimeInMs = this._maxRetentionTime * 60 * 1000; |
|
|
|
if (now - entry.timestamp > maxRetentionTimeInMs) { |
|
return this._removeRequest(operation); |
|
} |
|
|
|
return convertEntry(entry); |
|
} |
|
} |
|
/** |
|
* Loops through each request in the queue and attempts to re-fetch it. |
|
* If any request fails to re-fetch, it's put back in the same position in |
|
* the queue (which registers a retry for the next sync event). |
|
*/ |
|
|
|
|
|
async replayRequests() { |
|
let entry; |
|
|
|
while (entry = await this.shiftRequest()) { |
|
try { |
|
await fetch(entry.request.clone()); |
|
|
|
{ |
|
logger_mjs.logger.log(`Request for '${getFriendlyURL_mjs.getFriendlyURL(entry.request.url)}'` + `has been replayed in queue '${this._name}'`); |
|
} |
|
} catch (error) { |
|
await this.unshiftRequest(entry); |
|
|
|
{ |
|
logger_mjs.logger.log(`Request for '${getFriendlyURL_mjs.getFriendlyURL(entry.request.url)}'` + `failed to replay, putting it back in queue '${this._name}'`); |
|
} |
|
|
|
throw new WorkboxError_mjs.WorkboxError('queue-replay-failed', { |
|
name: this._name |
|
}); |
|
} |
|
} |
|
|
|
{ |
|
logger_mjs.logger.log(`All requests in queue '${this.name}' have successfully ` + `replayed; the queue is now empty!`); |
|
} |
|
} |
|
/** |
|
* Registers a sync event with a tag unique to this instance. |
|
*/ |
|
|
|
|
|
async registerSync() { |
|
if ('sync' in registration) { |
|
try { |
|
await registration.sync.register(`${TAG_PREFIX}:${this._name}`); |
|
} catch (err) { |
|
// This means the registration failed for some reason, possibly due to |
|
// the user disabling it. |
|
{ |
|
logger_mjs.logger.warn(`Unable to register sync event for '${this._name}'.`, err); |
|
} |
|
} |
|
} |
|
} |
|
/** |
|
* In sync-supporting browsers, this adds a listener for the sync event. |
|
* In non-sync-supporting browsers, this will retry the queue on service |
|
* worker startup. |
|
* |
|
* @private |
|
*/ |
|
|
|
|
|
_addSyncListener() { |
|
if ('sync' in registration) { |
|
self.addEventListener('sync', event => { |
|
if (event.tag === `${TAG_PREFIX}:${this._name}`) { |
|
{ |
|
logger_mjs.logger.log(`Background sync for tag '${event.tag}'` + `has been received`); |
|
} |
|
|
|
const syncComplete = async () => { |
|
this._syncInProgress = true; |
|
let syncError; |
|
|
|
try { |
|
await this._onSync({ |
|
queue: this |
|
}); |
|
} catch (error) { |
|
syncError = error; // Rethrow the error. Note: the logic in the finally clause |
|
// will run before this gets rethrown. |
|
|
|
throw syncError; |
|
} finally { |
|
// New items may have been added to the queue during the sync, |
|
// so we need to register for a new sync if that's happened... |
|
// Unless there was an error during the sync, in which |
|
// case the browser will automatically retry later, as long |
|
// as `event.lastChance` is not true. |
|
if (this._requestsAddedDuringSync && !(syncError && !event.lastChance)) { |
|
await this.registerSync(); |
|
} |
|
|
|
this._syncInProgress = false; |
|
this._requestsAddedDuringSync = false; |
|
} |
|
}; |
|
|
|
event.waitUntil(syncComplete()); |
|
} |
|
}); |
|
} else { |
|
{ |
|
logger_mjs.logger.log(`Background sync replaying without background sync event`); |
|
} // If the browser doesn't support background sync, retry |
|
// every time the service worker starts up as a fallback. |
|
|
|
|
|
this._onSync({ |
|
queue: this |
|
}); |
|
} |
|
} |
|
/** |
|
* Returns the set of queue names. This is primarily used to reset the list |
|
* of queue names in tests. |
|
* |
|
* @return {Set} |
|
* |
|
* @private |
|
*/ |
|
|
|
|
|
static get _queueNames() { |
|
return queueNames; |
|
} |
|
|
|
} |
|
/** |
|
* Converts a QueueStore entry into the format exposed by Queue. This entails |
|
* converting the request data into a real request and omitting the `id` and |
|
* `queueName` properties. |
|
* |
|
* @param {Object} queueStoreEntry |
|
* @return {Object} |
|
* @private |
|
*/ |
|
|
|
|
|
const convertEntry = queueStoreEntry => { |
|
const queueEntry = { |
|
request: new StorableRequest(queueStoreEntry.requestData).toRequest(), |
|
timestamp: queueStoreEntry.timestamp |
|
}; |
|
|
|
if (queueStoreEntry.metadata) { |
|
queueEntry.metadata = queueStoreEntry.metadata; |
|
} |
|
|
|
return queueEntry; |
|
}; |
|
|
|
/* |
|
Copyright 2018 Google LLC |
|
|
|
Use of this source code is governed by an MIT-style |
|
license that can be found in the LICENSE file or at |
|
https://opensource.org/licenses/MIT. |
|
*/ |
|
/** |
|
* A class implementing the `fetchDidFail` lifecycle callback. This makes it |
|
* easier to add failed requests to a background sync Queue. |
|
* |
|
* @memberof workbox.backgroundSync |
|
*/ |
|
|
|
class Plugin { |
|
/** |
|
* @param {...*} queueArgs Args to forward to the composed Queue instance. |
|
* See the [Queue]{@link workbox.backgroundSync.Queue} documentation for |
|
* parameter details. |
|
*/ |
|
constructor(...queueArgs) { |
|
this._queue = new Queue(...queueArgs); |
|
this.fetchDidFail = this.fetchDidFail.bind(this); |
|
} |
|
/** |
|
* @param {Object} options |
|
* @param {Request} options.request |
|
* @private |
|
*/ |
|
|
|
|
|
async fetchDidFail({ |
|
request |
|
}) { |
|
await this._queue.pushRequest({ |
|
request |
|
}); |
|
} |
|
|
|
} |
|
|
|
/* |
|
Copyright 2018 Google LLC |
|
|
|
Use of this source code is governed by an MIT-style |
|
license that can be found in the LICENSE file or at |
|
https://opensource.org/licenses/MIT. |
|
*/ |
|
|
|
exports.Queue = Queue; |
|
exports.Plugin = Plugin; |
|
|
|
return exports; |
|
|
|
}({}, workbox.core._private, workbox.core._private, workbox.core._private, workbox.core._private, workbox.core._private)); |
|
//# sourceMappingURL=workbox-background-sync.dev.js.map
|
|
|