Serverless DependsOn Plugin
If you have a Serverless application that executes AWS Lambdas inside a VPC, then chances are you have encountered this error:
Your request has been throttled by EC2, please make sure you have enough API rate limit.
EC2 Error Code: RequestLimitExceeded. EC2 Error Message: Request limit exceeded.
One solution to this error is to make all your lambdas dependent on each other in a chain. This will make CloudFormation deploy your lambdas sequentially and prevent the RequestLimitExceeded error. For more information on this error see: https://github.com/serverless/serverless/issues/3339
The drawback to this solution is that you have to manually customize the Resources
section of your serverless.yml
and override the lambda definitions to set the CloudFormation DependsOn attribute. This plugin will do that automatically.
Setup
Install via npm:
npm install serverless-dependson-plugin --save-dev
plugins
section of your serverless.yml
:
Update the plugins: - serverless-dependson-plugin
Deploy without errors!
Options
Disable the plugin
Sometimes you may be able to deploy your serverless application without receiving the RequestLimitExceeded error. To test deploying your application without throttling, you can temporarily disable the plugin by:
Passing the following command line option to serverless:
--dependson-plugin=[disabled|off|false]
Adding a dependsOn
section to the custom
section of your serverless.yml
:
custom: dependsOn: enabled: false
Deployment performance
Because your lambdas will now be deployed sequentially, your stack deployment time will drastically increase. You can try to improve your deployment time by letting the plugin build multiple lambda DependsOn "chains". This will let CloudFormation do some parallelization of the lambda deployments.
You can configure this by adding a dependsOn
section to the custom
section of your serverless.yml
:
custom: dependsOn: chains: <an integer greater than 1>
If the value of the chains
parameter is not an integer or is less than 1, it will default to 1 without failing the deployment.
Enabling this option may still trigger the RequestLimitExceeded error. Amazon does not disclose what will trip their rate limiter, so you may need to experiment with this option to get the best deployment performance without hitting the request limit.