The Secret Lives of Failed Amazon SQS Messages

How to redrive your SQS messages to success

Danielle Heberling

A common pattern in serverless architecture is to have a queue before a function. This is great because you can create a second queue for all of the messages that failed in the function execution (or, if we want to put it in terms that don’t sound like we’re aggressively shaming them, we can classify them as having “encountered an error at some point”). This second queue is known as a “dead letter queue” or DLQ for short.

If messages arrive in the DLQ, you can analyze what might have caused the error by reviewing any relevant logs, making changes to your stack, and running a redrive function to retry those messages. DLQs are useful for debugging your serverless application because they let you isolate problem-messages to determine why their processing isn’t working.

A redrive function is how we get those failed messages out of the DLQ and back into the original pipeline to retry the operation. I wanted to set up this type of architecture in a new stack using Stackery and decided to take a look around the internet for inspiration on different patterns and ideas on how to go about it.

I found lots of information about dead letter queues, but no examples that demoed a redrive function to retry those failed messages. This post is meant to serve as a nice example for folks like me who want to do this on AWS and haven’t found many examples.

The setup

Happy path

Let’s start out with our happy path flow. I’ve set up a stack that consists of an SNS topic > SQS queue > Lambda function. In Stackery it looks something like this:

Failed SQS messages.png

Retry path

In the event that something goes wrong during the function execution, I set up a second queue; this will serve as my DLQ for failed messages to be delivered. I then added a redrive function to allow me to retry those failed messages through the original pipeline. Here’s what that set up looks like:

I added a RedrivePolicy to the original SQS queue to point those failed messages to my DLQ by adding some CloudFormation in my template.yaml:


    Type: AWS::SQS::Queue


      QueueName: !Sub ${AWS::StackName}-Queue


        deadLetterTargetArn: !GetAtt DeadLetterQueue.Arn

        maxReceiveCount: 3

By default, the queue’s message retention period is 4 days. I wanted to give myself some extra time to review the messages in the DLQ, so I upped the MessageRetentionPeriod on my DLQ to the max allowed 14 days like this in my template.yaml.

This is the sci-fi beauty of my profession: creating more time is possible! As many before me have discussed on the floor of a dorm room, time is a malleable concept— and at least in software engineering, it has a practical purpose like gaining more days to review DLQ messages. Note that the values are in seconds:


    Type: AWS::SQS::Queue


      QueueName: !Sub ${AWS::StackName}-DeadLetterQueue

      MessageRetentionPeriod: 1209600

Redrive Function Content

Now that the base architecture is all set up, let’s take a look inside the redrive function. This is how I decided to set mine up, but keep in mind that there are many different ways you can do this:

Here’s the high-level steps of what our retry function needs to do:

  1. Receive messages from the DLQ

  2. If no messages are received exit the function

  3. If there are received messages, loop through each message > send it to the original queue

  4. Delete the message from the DLQ on a successful send

  5. Repeat

Here’s what the code looks like written in nodeJS:

const aws = require('aws-sdk');

const sqs = new aws.SQS();

exports.handler = async event => {

  // Log the event argument for debugging and for use in local development.

  console.log(JSON.stringify(event, undefined, 2));

  while (true) {

    try {

      // Use long polling to avoid empty message responses

      const receiveParams = {

        QueueUrl: process.env.DLQ_URL,

        MaxNumberOfMessages: 10,

        WaitTimeSeconds: 1


      // Get messages from the DLQ

      // Continue looping until no more messages are left

      const DLQMessages = await sqs.receiveMessage(receiveParams).promise();

      if (!DLQMessages.Messages || DLQMessages.Messages.length === 0) {

        console.log(`NO MESSAGES FOUND IN ${process.env.DLQ_URL}`);

        // Exit the loop since there aren't any messages left



      console.log(`RECEIVED ${DLQMessages.Messages.length} MESSAGES`);

      for (const message of DLQMessages.Messages) {

        // Send message to original queue

        const outboundMessage = {

          MessageBody: message.Body,

          QueueUrl: process.env.QUEUE_URL


        console.log(`SENDING: ${JSON.stringify(outboundMessage, null, 2)}`);

        await sqs.sendMessage(outboundMessage).promise();

        console.log('SEND MESSAGE SUCCEEDED');

        // Delete message from DLQ

        const deleteParams = {

          QueueUrl: process.env.DLQ_URL,

          ReceiptHandle: message.ReceiptHandle


        console.log(`DELETING: ${JSON.stringify(deleteParams, null, 2)}`);

        await sqs.deleteMessage(deleteParams).promise();

        console.log('DELETE MESSAGE SUCCEEDED');


    } catch (err) {

      console.log(`AN ERROR OCCURRED: ${err.message}`);

      console.log(JSON.stringify(err, null, 2));

      throw err;




Things to keep in mind when setting up a redrive function.

Receive message step

  • The default MaxNumberOfMessages to return is 1, so you’ll most likely want to up that number depending on your use case. I just went to the max allowed which is 10. 
  • If you hypothetically have 10 messages in your DLQ and do receiveMessage on that queue with a MaxNumberOfMessages set to 10, you might possibly not receive all of those messages in a single receiveMessage batch. It’s important to have looping to ensure that you receive all of the messages when they are available.
  • Enabling long polling can help reduce the cost of SQS by reducing the number of empty responses. I decided to go this route by adding WaitTimeSeconds to my receiveMessage params.

Delete message step

  • Ensure the send message was successful before deleting it. I decided to wrap the entire function in a try, catch…so if there’s an error the function will stop processing.

As you can see, it does require a bit of additional setup, but creating dead letter queues and a redrive function to re-run failed messages can be extremely helpful in your application’s life cycle. You can go to to see an example of a simple stack that uses a DLQ and a redrive function in action.

For a solid and succinct description of the benefits of a DLQ (alongside lots of other info about them,) check out AWS’ documentation here.

Want to make it easier to access function logs? Stackery makes this easier for troubleshooting benefits. Don’t miss our function resource overview on the docs site.

Get this in your inbox


Join uscoffee graphic



JAN 27 | 12 PM PST


Learn More

Quick links to get familiar with Stackery

Related posts

Simple Authentication with AWS Cognito
Using StackerySimple Authentication with AWS Cognito

Overview of Cognito

Creating Cognito User Pools with CloudFormation
Using StackeryCreating Cognito User Pools with CloudFormation

Using Cognito in AWS

Curious about Stackery and its capabilities?

Learn more