2022-11-23 13:33:44 +03:00
const logging = require ( '@tryghost/logging' ) ;
const ObjectID = require ( 'bson-objectid' ) . default ;
const errors = require ( '@tryghost/errors' ) ;
const tpl = require ( '@tryghost/tpl' ) ;
2023-05-02 23:43:47 +03:00
const EmailBodyCache = require ( './EmailBodyCache' ) ;
2022-11-23 13:33:44 +03:00
const messages = {
2023-01-20 20:36:33 +03:00
emailErrorPartialFailure : 'An error occurred, and your newsletter was only partially sent. Please retry sending the remaining emails.' ,
emailError : 'An unexpected error occurred, please retry sending your newsletter.'
2022-11-23 13:33:44 +03:00
} ;
2023-01-24 20:02:10 +03:00
const MAX _SENDING _CONCURRENCY = 2 ;
2022-11-23 13:33:44 +03:00
/ * *
2023-05-02 23:43:47 +03:00
* @ typedef { import ( './SendingService' ) } SendingService
* @ typedef { import ( './EmailSegmenter' ) } EmailSegmenter
* @ typedef { import ( './EmailRenderer' ) } EmailRenderer
* @ typedef { import ( './EmailRenderer' ) . MemberLike } MemberLike
2022-11-23 13:33:44 +03:00
* @ typedef { object } JobsService
* @ typedef { object } Email
* @ typedef { object } Newsletter
* @ typedef { object } Post
* @ typedef { object } EmailBatch
* /
class BatchSendingService {
# emailRenderer ;
# sendingService ;
# emailSegmenter ;
# jobsService ;
# models ;
# db ;
2023-01-25 16:57:10 +03:00
# sentry ;
2023-05-30 12:24:14 +03:00
# debugStorageFilePath ;
2022-11-23 13:33:44 +03:00
2023-02-02 16:12:54 +03:00
// Retry database queries happening before sending the email
# BEFORE _RETRY _CONFIG = { maxRetries : 10 , maxTime : 10 * 60 * 1000 , sleep : 2000 } ;
# AFTER _RETRY _CONFIG = { maxRetries : 20 , maxTime : 30 * 60 * 1000 , sleep : 2000 } ;
2023-02-21 18:22:26 +03:00
# MAILGUN _API _RETRY _CONFIG = { sleep : 10 * 1000 , maxRetries : 6 } ;
2023-02-02 16:12:54 +03:00
2022-11-23 13:33:44 +03:00
/ * *
2022-11-30 13:51:58 +03:00
* @ param { Object } dependencies
2022-11-23 13:33:44 +03:00
* @ param { EmailRenderer } dependencies . emailRenderer
* @ param { SendingService } dependencies . sendingService
* @ param { JobsService } dependencies . jobsService
* @ param { EmailSegmenter } dependencies . emailSegmenter
* @ param { object } dependencies . models
* @ param { object } dependencies . models . EmailRecipient
* @ param { EmailBatch } dependencies . models . EmailBatch
* @ param { Email } dependencies . models . Email
* @ param { object } dependencies . models . Member
* @ param { object } dependencies . db
2023-01-25 16:57:10 +03:00
* @ param { object } [ dependencies . sentry ]
2023-02-02 16:12:54 +03:00
* @ param { object } [ dependencies . BEFORE _RETRY _CONFIG ]
* @ param { object } [ dependencies . AFTER _RETRY _CONFIG ]
2023-05-19 13:57:24 +03:00
* @ param { object } [ dependencies . MAILGUN _API _RETRY _CONFIG ]
2023-05-30 12:24:14 +03:00
* @ param { string } [ dependencies . debugStorageFilePath ]
2022-11-23 13:33:44 +03:00
* /
constructor ( {
emailRenderer ,
sendingService ,
jobsService ,
emailSegmenter ,
models ,
2023-01-25 16:57:10 +03:00
db ,
2023-02-02 16:12:54 +03:00
sentry ,
BEFORE _RETRY _CONFIG ,
2023-02-21 17:41:00 +03:00
AFTER _RETRY _CONFIG ,
2023-05-30 12:24:14 +03:00
MAILGUN _API _RETRY _CONFIG ,
debugStorageFilePath
2022-11-23 13:33:44 +03:00
} ) {
this . # emailRenderer = emailRenderer ;
this . # sendingService = sendingService ;
this . # jobsService = jobsService ;
this . # emailSegmenter = emailSegmenter ;
this . # models = models ;
this . # db = db ;
2023-01-25 16:57:10 +03:00
this . # sentry = sentry ;
2023-05-30 12:24:14 +03:00
this . # debugStorageFilePath = debugStorageFilePath ;
2023-02-02 16:12:54 +03:00
if ( BEFORE _RETRY _CONFIG ) {
this . # BEFORE _RETRY _CONFIG = BEFORE _RETRY _CONFIG ;
2023-03-07 13:40:55 +03:00
} else {
2023-03-09 14:32:22 +03:00
if ( process . env . NODE _ENV . startsWith ( 'test' ) || process . env . NODE _ENV === 'development' ) {
2023-03-07 13:40:55 +03:00
this . # BEFORE _RETRY _CONFIG = { maxRetries : 0 } ;
}
2023-02-02 16:12:54 +03:00
}
if ( AFTER _RETRY _CONFIG ) {
this . # AFTER _RETRY _CONFIG = AFTER _RETRY _CONFIG ;
2023-03-07 13:40:55 +03:00
} else {
2023-03-09 14:32:22 +03:00
if ( process . env . NODE _ENV . startsWith ( 'test' ) || process . env . NODE _ENV === 'development' ) {
2023-03-07 13:40:55 +03:00
this . # AFTER _RETRY _CONFIG = { maxRetries : 0 } ;
}
2023-02-02 16:12:54 +03:00
}
2023-03-07 13:40:55 +03:00
2023-02-21 17:41:00 +03:00
if ( MAILGUN _API _RETRY _CONFIG ) {
this . # MAILGUN _API _RETRY _CONFIG = MAILGUN _API _RETRY _CONFIG ;
} else {
2023-03-09 14:32:22 +03:00
if ( process . env . NODE _ENV . startsWith ( 'test' ) || process . env . NODE _ENV === 'development' ) {
2023-02-21 17:41:00 +03:00
this . # MAILGUN _API _RETRY _CONFIG = { maxRetries : 0 } ;
}
}
2023-02-02 16:12:54 +03:00
}
# getBeforeRetryConfig ( email ) {
if ( email . _retryCutOffTime ) {
return { ... this . # BEFORE _RETRY _CONFIG , stopAfterDate : email . _retryCutOffTime } ;
}
return this . # BEFORE _RETRY _CONFIG ;
2022-11-23 13:33:44 +03:00
}
/ * *
* Schedules a background job that sends the email in the background if it is pending or failed .
2022-11-30 13:51:58 +03:00
* @ param { Email } email
2022-11-23 13:33:44 +03:00
* @ returns { void }
* /
scheduleEmail ( email ) {
return this . # jobsService . addJob ( {
2022-12-01 15:43:49 +03:00
name : 'batch-sending-service-job' ,
2022-11-23 13:33:44 +03:00
job : this . emailJob . bind ( this ) ,
data : { emailId : email . id } ,
offloaded : false
} ) ;
}
/ * *
* @ private
* @ param { { emailId : string } } data Data passed from the job service . We only need the emailId because we need to refetch the email anyway to make sure the status is right and 'locked' .
* /
async emailJob ( { emailId } ) {
logging . info ( ` Starting email job for email ${ emailId } ` ) ;
2023-02-02 16:12:54 +03:00
// We'll stop all automatic DB retries after this date
const retryCutOffTime = new Date ( Date . now ( ) + this . # BEFORE _RETRY _CONFIG . maxTime ) ;
2022-11-23 13:33:44 +03:00
// Check if email is 'pending' only + change status to submitting in one transaction.
// This allows us to have a lock around the email job that makes sure an email can only have one active job.
2023-02-02 16:12:54 +03:00
let email = await this . retryDb (
async ( ) => {
return await this . updateStatusLock ( this . # models . Email , emailId , 'submitting' , [ 'pending' , 'failed' ] ) ;
} ,
{ ... this . # BEFORE _RETRY _CONFIG , description : ` updateStatusLock email ${ emailId } -> submitting ` }
) ;
2022-11-23 13:33:44 +03:00
if ( ! email ) {
logging . error ( ` Tried sending email that is not pending or failed ${ emailId } ` ) ;
return ;
}
2023-02-02 16:12:54 +03:00
// Save a strict cutoff time for retries
email . _retryCutOffTime = retryCutOffTime ;
2022-11-23 13:33:44 +03:00
try {
await this . sendEmail ( email ) ;
2023-02-02 16:12:54 +03:00
await this . retryDb ( async ( ) => {
await email . save ( {
status : 'submitted' ,
submitted _at : new Date ( ) ,
error : null
} , { patch : true , autoRefresh : false } ) ;
} , { ... this . # AFTER _RETRY _CONFIG , description : ` email ${ emailId } -> submitted ` } ) ;
2022-11-23 13:33:44 +03:00
} catch ( e ) {
2023-01-25 16:57:10 +03:00
const ghostError = new errors . EmailError ( {
err : e ,
code : 'BULK_EMAIL_SEND_FAILED' ,
message : ` Error sending email ${ email . id } `
} ) ;
logging . error ( ghostError ) ;
if ( this . # sentry ) {
// Log the original error to Sentry
this . # sentry . captureException ( e ) ;
}
2022-11-23 13:33:44 +03:00
2023-02-02 16:12:54 +03:00
// Store error and status in email model
await this . retryDb ( async ( ) => {
await email . save ( {
status : 'failed' ,
error : e . message || 'Something went wrong while sending the email'
} , { patch : true , autoRefresh : false } ) ;
} , { ... this . # AFTER _RETRY _CONFIG , description : ` email ${ emailId } -> failed ` } ) ;
2022-11-23 13:33:44 +03:00
}
}
/ * *
* @ private
2022-11-30 13:51:58 +03:00
* @ param { Email } email
2022-11-23 13:33:44 +03:00
* @ throws { errors . EmailError } If one of the batches fails
* /
async sendEmail ( email ) {
logging . info ( ` Sending email ${ email . id } ` ) ;
// Load required relations
2023-02-02 16:12:54 +03:00
const newsletter = await this . retryDb ( async ( ) => {
return await email . getLazyRelation ( 'newsletter' , { require : true } ) ;
} , { ... this . # getBeforeRetryConfig ( email ) , description : ` getLazyRelation newsletter for email ${ email . id } ` } ) ;
const post = await this . retryDb ( async ( ) => {
return await email . getLazyRelation ( 'post' , { require : true , withRelated : [ 'posts_meta' , 'authors' ] } ) ;
} , { ... this . # getBeforeRetryConfig ( email ) , description : ` getLazyRelation post for email ${ email . id } ` } ) ;
let batches = await this . retryDb ( async ( ) => {
return await this . getBatches ( email ) ;
} , { ... this . # getBeforeRetryConfig ( email ) , description : ` getBatches for email ${ email . id } ` } ) ;
2022-11-23 13:33:44 +03:00
if ( batches . length === 0 ) {
batches = await this . createBatches ( { email , newsletter , post } ) ;
}
await this . sendBatches ( { email , batches , post , newsletter } ) ;
}
/ * *
* @ private
* @ param { Email } email
* @ returns { Promise < EmailBatch [ ] > }
* /
async getBatches ( email ) {
logging . info ( ` Getting batches for email ${ email . id } ` ) ;
2023-11-13 14:00:20 +03:00
return await this . # models . EmailBatch . findAll ( { filter : 'email_id:\'' + email . id + '\'' } ) ;
2022-11-23 13:33:44 +03:00
}
/ * *
* @ private
* @ param { { email : Email , newsletter : Newsletter , post : Post } } data
* @ returns { Promise < EmailBatch [ ] > }
* /
async createBatches ( { email , post , newsletter } ) {
logging . info ( ` Creating batches for email ${ email . id } ` ) ;
2023-07-21 01:48:48 +03:00
const segments = await this . # emailRenderer . getSegments ( post ) ;
2022-11-23 13:33:44 +03:00
const batches = [ ] ;
2022-12-01 15:43:49 +03:00
const BATCH _SIZE = this . # sendingService . getMaximumRecipients ( ) ;
2022-11-23 13:33:44 +03:00
let totalCount = 0 ;
for ( const segment of segments ) {
logging . info ( ` Creating batches for email ${ email . id } segment ${ segment } ` ) ;
const segmentFilter = this . # emailSegmenter . getMemberFilterForSegment ( newsletter , email . get ( 'recipient_filter' ) , segment ) ;
// Avoiding Bookshelf for performance reasons
let members ;
2023-01-04 13:22:12 +03:00
// Start with the id of the email, which is an objectId. We'll only fetch members that are created before the email. This is a special property of ObjectIds.
// Note: we use ID and not created_at, because imported members could set a created_at in the future or past and avoid limit checking.
let lastId = email . id ;
2022-11-23 13:33:44 +03:00
while ( ! members || lastId ) {
logging . info ( ` Fetching members batch for email ${ email . id } segment ${ segment } , lastId: ${ lastId } ` ) ;
2023-11-10 04:24:56 +03:00
const filter = segmentFilter + ` +id:<' ${ lastId } ' ` ;
2022-12-01 15:43:49 +03:00
members = await this . # models . Member . getFilteredCollectionQuery ( { filter } )
. orderByRaw ( 'id DESC' )
. select ( 'members.id' , 'members.uuid' , 'members.email' , 'members.name' ) . limit ( BATCH _SIZE + 1 ) ;
2022-11-23 13:33:44 +03:00
if ( members . length > 0 ) {
totalCount += Math . min ( members . length , BATCH _SIZE ) ;
2023-02-02 16:12:54 +03:00
const batch = await this . retryDb (
async ( ) => {
return await this . createBatch ( email , segment , members . slice ( 0 , BATCH _SIZE ) ) ;
} ,
{ ... this . # getBeforeRetryConfig ( email ) , description : ` createBatch email ${ email . id } segment ${ segment } ` }
) ;
2022-11-23 13:33:44 +03:00
batches . push ( batch ) ;
}
2023-01-04 13:22:12 +03:00
if ( members . length > BATCH _SIZE ) {
lastId = members [ members . length - 2 ] . id ;
} else {
break ;
}
2022-11-23 13:33:44 +03:00
}
}
logging . info ( ` Created ${ batches . length } batches for email ${ email . id } with ${ totalCount } recipients ` ) ;
if ( email . get ( 'email_count' ) !== totalCount ) {
2023-01-10 16:58:50 +03:00
logging . error ( ` Email ${ email . id } has wrong stored email_count ${ email . get ( 'email_count' ) } , did expect ${ totalCount } . Updating the model. ` ) ;
2022-11-23 13:33:44 +03:00
2023-01-04 13:22:12 +03:00
// We update the email model because this might happen in rare cases where the initial member count changed (e.g. deleted members)
// between creating the email and sending it
2022-11-23 13:33:44 +03:00
await email . save ( {
email _count : totalCount
2022-12-05 14:09:30 +03:00
} , { patch : true , require : false , autoRefresh : false } ) ;
2022-11-23 13:33:44 +03:00
}
return batches ;
}
/ * *
* @ private
* @ param { Email } email
2023-05-02 23:43:47 +03:00
* @ param { import ( './EmailRenderer' ) . Segment } segment
2022-11-23 13:33:44 +03:00
* @ param { object [ ] } members
* @ returns { Promise < EmailBatch > }
* /
async createBatch ( email , segment , members , options ) {
if ( ! options || ! options . transacting ) {
return this . # models . EmailBatch . transaction ( async ( transacting ) => {
return this . createBatch ( email , segment , members , { transacting } ) ;
} ) ;
}
logging . info ( ` Creating batch for email ${ email . id } segment ${ segment } with ${ members . length } members ` ) ;
const batch = await this . # models . EmailBatch . add ( {
email _id : email . id ,
member _segment : segment ,
status : 'pending'
} , options ) ;
const recipientData = [ ] ;
members . forEach ( ( memberRow ) => {
if ( ! memberRow . id || ! memberRow . uuid || ! memberRow . email ) {
logging . warn ( ` Member row not included as email recipient due to missing data - id: ${ memberRow . id } , uuid: ${ memberRow . uuid } , email: ${ memberRow . email } ` ) ;
return ;
}
recipientData . push ( {
id : ObjectID ( ) . toHexString ( ) ,
email _id : email . id ,
member _id : memberRow . id ,
batch _id : batch . id ,
member _uuid : memberRow . uuid ,
member _email : memberRow . email ,
member _name : memberRow . name
} ) ;
} ) ;
const insertQuery = this . # db . knex ( 'email_recipients' ) . insert ( recipientData ) ;
if ( options . transacting ) {
insertQuery . transacting ( options . transacting ) ;
}
logging . info ( ` Inserting ${ recipientData . length } recipients for email ${ email . id } batch ${ batch . id } ` ) ;
await insertQuery ;
return batch ;
}
async sendBatches ( { email , batches , post , newsletter } ) {
logging . info ( ` Sending ${ batches . length } batches for email ${ email . id } ` ) ;
2023-02-07 13:01:49 +03:00
// Reuse same HTML body if we send an email to the same segment
const emailBodyCache = new EmailBodyCache ( ) ;
2022-11-23 13:33:44 +03:00
// Loop batches and send them via the EmailProvider
let succeededCount = 0 ;
2022-12-02 17:30:02 +03:00
const queue = batches . slice ( ) ;
// Bind this
let runNext ;
runNext = async ( ) => {
const batch = queue . shift ( ) ;
if ( batch ) {
2023-02-07 13:01:49 +03:00
if ( await this . sendBatch ( { email , batch , post , newsletter , emailBodyCache } ) ) {
2022-12-02 17:30:02 +03:00
succeededCount += 1 ;
}
await runNext ( ) ;
2022-11-23 13:33:44 +03:00
}
2022-12-02 17:30:02 +03:00
} ;
2023-01-24 20:02:10 +03:00
// Run maximum MAX_SENDING_CONCURRENCY at the same time
await Promise . all ( new Array ( MAX _SENDING _CONCURRENCY ) . fill ( 0 ) . map ( ( ) => runNext ( ) ) ) ;
2022-11-23 13:33:44 +03:00
if ( succeededCount < batches . length ) {
if ( succeededCount > 0 ) {
throw new errors . EmailError ( {
message : tpl ( messages . emailErrorPartialFailure )
} ) ;
}
throw new errors . EmailError ( {
message : tpl ( messages . emailError )
} ) ;
}
}
/ * *
2022-11-30 13:51:58 +03:00
*
* @ param { { email : Email , batch : EmailBatch , post : Post , newsletter : Newsletter } } data
2022-11-23 13:33:44 +03:00
* @ returns { Promise < boolean > } True when succeeded , false when failed with an error
* /
2023-02-07 13:01:49 +03:00
async sendBatch ( { email , batch : originalBatch , post , newsletter , emailBodyCache } ) {
2022-12-01 15:43:49 +03:00
logging . info ( ` Sending batch ${ originalBatch . id } for email ${ email . id } ` ) ;
2022-11-23 13:33:44 +03:00
// Check the status of the email batch in a 'for update' transaction
2023-02-02 16:12:54 +03:00
const batch = await this . retryDb (
async ( ) => {
return await this . updateStatusLock ( this . # models . EmailBatch , originalBatch . id , 'submitting' , [ 'pending' , 'failed' ] ) ;
} ,
{ ... this . # getBeforeRetryConfig ( email ) , description : ` updateStatusLock batch ${ originalBatch . id } -> submitting ` }
) ;
2022-11-23 13:33:44 +03:00
if ( ! batch ) {
2022-12-01 15:43:49 +03:00
logging . error ( ` Tried sending email batch that is not pending or failed ${ originalBatch . id } ` ) ;
2022-11-23 13:33:44 +03:00
return true ;
}
let succeeded = false ;
try {
2023-05-19 13:57:24 +03:00
let members = await this . retryDb (
2023-02-02 16:12:54 +03:00
async ( ) => {
const m = await this . getBatchMembers ( batch . id ) ;
// If we receive 0 rows, there is a possibility that we switched to a secondary database and have replication lag
// So we throw an error and we retry
if ( m . length === 0 ) {
throw new errors . EmailError ( {
message : ` No members found for batch ${ batch . id } , possible replication lag `
} ) ;
}
return m ;
} ,
{ ... this . # getBeforeRetryConfig ( email ) , description : ` getBatchMembers batch ${ originalBatch . id } ` }
) ;
2023-02-21 17:41:00 +03:00
const response = await this . retryDb ( async ( ) => {
return await this . # sendingService . send ( {
emailId : email . id ,
post ,
newsletter ,
segment : batch . get ( 'member_segment' ) ,
members
} , {
openTrackingEnabled : ! ! email . get ( 'track_opens' ) ,
clickTrackingEnabled : ! ! email . get ( 'track_clicks' ) ,
emailBodyCache
} ) ;
} , { ... this . # MAILGUN _API _RETRY _CONFIG , description : ` Sending email batch ${ originalBatch . id } ` } ) ;
2022-11-23 13:33:44 +03:00
succeeded = true ;
2023-02-02 16:12:54 +03:00
await this . retryDb (
async ( ) => {
await batch . save ( {
status : 'submitted' ,
provider _id : response . id ,
// reset error fields when sending succeeds
error _status _code : null ,
error _message : null ,
error _data : null
} , { patch : true , require : false , autoRefresh : false } ) ;
} ,
{ ... this . # AFTER _RETRY _CONFIG , description : ` save batch ${ originalBatch . id } -> submitted ` }
) ;
2022-11-23 13:33:44 +03:00
} catch ( err ) {
2023-02-21 17:41:00 +03:00
if ( err . code && err . code === 'BULK_EMAIL_SEND_FAILED' ) {
logging . error ( err ) ;
if ( this . # sentry ) {
// Log the original error to Sentry
this . # sentry . captureException ( err ) ;
}
} else {
2023-01-25 16:57:10 +03:00
const ghostError = new errors . EmailError ( {
err ,
code : 'BULK_EMAIL_SEND_FAILED' ,
message : ` Error sending email batch ${ batch . id } ` ,
context : err . message
} ) ;
logging . error ( ghostError ) ;
if ( this . # sentry ) {
// Log the original error to Sentry
this . # sentry . captureException ( err ) ;
}
}
2022-11-23 13:33:44 +03:00
2023-02-02 16:12:54 +03:00
if ( ! succeeded ) {
// We check succeeded because a Rare edge case where the batch was send, but we failed to set status to submitted, then we don't want to set it to failed
await this . retryDb (
async ( ) => {
await batch . save ( {
status : 'failed' ,
error _status _code : err . statusCode ? ? null ,
error _message : err . message ,
error _data : err . errorDetails ? ? null
} , { patch : true , require : false , autoRefresh : false } ) ;
} ,
{ ... this . # AFTER _RETRY _CONFIG , description : ` save batch ${ originalBatch . id } -> failed ` }
) ;
}
2022-11-23 13:33:44 +03:00
}
// Mark as processed, even when failed
2023-02-02 16:12:54 +03:00
await this . retryDb (
async ( ) => {
await this . # models . EmailRecipient
. where ( { batch _id : batch . id } )
. save ( { processed _at : new Date ( ) } , { patch : true , require : false , autoRefresh : false } ) ;
} ,
{ ... this . # AFTER _RETRY _CONFIG , description : ` save EmailRecipients ${ originalBatch . id } processed_at ` }
) ;
2022-11-23 13:33:44 +03:00
return succeeded ;
}
/ * *
* We don ' t want to pass EmailRecipient models to the sendingService .
* So we transform them into the MemberLike interface .
2023-05-19 13:57:24 +03:00
* That keeps the sending service nicely separated so it isn ' t dependent on the batch sending data structure .
2022-11-23 13:33:44 +03:00
* @ returns { Promise < MemberLike [ ] > }
* /
async getBatchMembers ( batchId ) {
2023-11-13 14:00:20 +03:00
let models = await this . # models . EmailRecipient . findAll ( { filter : ` batch_id:' ${ batchId } ' ` , withRelated : [ 'member' , 'member.stripeSubscriptions' , 'member.products' ] } ) ;
2023-06-09 09:50:53 +03:00
const BATCH _SIZE = this . # sendingService . getMaximumRecipients ( ) ;
if ( models . length > BATCH _SIZE ) {
2023-11-20 16:50:07 +03:00
// @NOTE: filtering by batch_id is our best effort to "correct" returned data
logging . warn ( ` Email batch ${ batchId } has ${ models . length } members, which exceeds the maximum of ${ BATCH _SIZE } members per batch. Filtering by batch_id: ${ batchId } ` ) ;
models = models . filter ( m => m . get ( 'batch_id' ) === batchId ) ;
if ( models . length > BATCH _SIZE ) {
// @NOTE this is a best effort logic to still try sending an email batch
// even if it exceeds the maximum recipients limit of the sending service.
// In theory this should never happen, but being extra safe to make sure
// the email delivery still happens.
logging . error ( ` Email batch ${ batchId } has ${ models . length } members, which exceeds the maximum of ${ BATCH _SIZE } . Truncating to ${ BATCH _SIZE } ` ) ;
models = models . slice ( 0 , BATCH _SIZE ) ;
}
2023-06-09 09:50:53 +03:00
}
2023-05-19 13:57:24 +03:00
2023-06-09 09:50:53 +03:00
return models . map ( ( model ) => {
2023-03-22 13:52:41 +03:00
// Map subscriptions
2023-03-28 13:26:57 +03:00
const subscriptions = model . related ( 'member' ) . related ( 'stripeSubscriptions' ) . toJSON ( ) ;
const tiers = model . related ( 'member' ) . related ( 'products' ) . toJSON ( ) ;
2023-03-22 13:52:41 +03:00
2022-11-23 13:33:44 +03:00
return {
id : model . get ( 'member_id' ) ,
uuid : model . get ( 'member_uuid' ) ,
email : model . get ( 'member_email' ) ,
2023-03-15 19:08:57 +03:00
name : model . get ( 'member_name' ) ,
2023-03-22 13:52:41 +03:00
createdAt : model . related ( 'member' ) ? . get ( 'created_at' ) ? ? null ,
status : model . related ( 'member' ) ? . get ( 'status' ) ? ? 'free' ,
subscriptions ,
tiers
2022-11-23 13:33:44 +03:00
} ;
} ) ;
}
/ * *
* @ private
* Update the status of an email or emailBatch to a given status , but first check if their current status is 'pending' or 'failed' .
* @ param { object } Model Bookshelf model constructor
* @ param { string } id id of the model
* @ param { string } status set the status of the model to this value
* @ param { string [ ] } allowedStatuses Check if the models current status is one of these values
* @ returns { Promise < object | undefined > } The updated model . Undefined if the model didn ' t pass the status check .
* /
async updateStatusLock ( Model , id , status , allowedStatuses ) {
let model ;
await Model . transaction ( async ( transacting ) => {
model = await Model . findOne ( { id } , { require : true , transacting , forUpdate : true } ) ;
if ( ! allowedStatuses . includes ( model . get ( 'status' ) ) ) {
model = undefined ;
return ;
}
await model . save ( {
status
2022-12-05 14:09:30 +03:00
} , { patch : true , transacting , autoRefresh : false } ) ;
2022-11-23 13:33:44 +03:00
} ) ;
return model ;
}
2023-02-02 16:12:54 +03:00
/ * *
* @ private
* Retry a function until it doesn ' t throw an error or the max retries / max time are reached .
* @ template T
* @ param { ( ) => Promise < T > } func
* @ param { object } options
* @ param { string } options . description Used for logging
* @ param { number } options . sleep time between each retry ( ms ) , will get multiplied by the number of retries
* @ param { number } options . maxRetries note : retries , not tries . So 0 means maximum 1 try , 1 means maximum 2 tries , etc .
* @ param { number } [ options . retryCount ] ( internal ) Amount of retries already done . 0 intially .
* @ param { number } [ options . maxTime ] ( ms )
* @ param { Date } [ options . stopAfterDate ]
* @ returns { Promise < T > }
* /
async retryDb ( func , options ) {
if ( options . maxTime !== undefined ) {
const stopAfterDate = new Date ( Date . now ( ) + options . maxTime ) ;
if ( ! options . stopAfterDate || stopAfterDate < options . stopAfterDate ) {
options = { ... options , stopAfterDate } ;
}
}
try {
return await func ( ) ;
} catch ( e ) {
const retryCount = ( options . retryCount ? ? 0 ) ;
2023-02-21 17:41:00 +03:00
const sleep = ( options . sleep ? ? 0 ) ;
2023-02-02 16:12:54 +03:00
if ( retryCount >= options . maxRetries || ( options . stopAfterDate && ( new Date ( Date . now ( ) + sleep ) ) > options . stopAfterDate ) ) {
2023-02-21 17:41:00 +03:00
if ( retryCount > 0 ) {
const ghostError = new errors . EmailError ( {
err : e ,
code : 'BULK_EMAIL_DB_RETRY' ,
message : ` [BULK_EMAIL_DB_RETRY] ${ options . description } - Stopped retrying ` ,
context : e . message
} ) ;
logging . error ( ghostError ) ;
}
2023-02-02 16:12:54 +03:00
throw e ;
}
const ghostError = new errors . EmailError ( {
err : e ,
code : 'BULK_EMAIL_DB_RETRY' ,
message : ` [BULK_EMAIL_DB_RETRY] ${ options . description } - After ${ retryCount } retries ` ,
context : e . message
} ) ;
logging . error ( ghostError ) ;
if ( sleep ) {
await new Promise ( ( resolve ) => {
setTimeout ( resolve , sleep ) ;
} ) ;
}
2023-02-21 17:41:00 +03:00
return await this . retryDb ( func , { ... options , retryCount : retryCount + 1 , sleep : sleep * 2 } ) ;
2023-02-02 16:12:54 +03:00
}
}
2022-11-23 13:33:44 +03:00
}
module . exports = BatchSendingService ;