- 新しい順
- 投票が多い順
- コメントが多い順
Hello.
I have never used "scipy", but what size will it be when layered?
Lambda layers have a limit of 250MB, so if they are larger than this, they cannot be used.
https://docs.aws.amazon.com/lambda/latest/dg/gettingstarted-limits.html
I think one of the workarounds for layer size is to use a container image as the runtime.
I think you can use this with somewhat larger layers.
https://docs.aws.amazon.com/lambda/latest/dg/images-create.html#images-types
Also, if there is a memory limit during execution, I think there will be no problem if you increase it from the Lambda settings.
https://docs.aws.amazon.com/lambda/latest/dg/configuration-function-common.html#configuration-memory-console
As per documention, deployment package limit is 250 MB
which is about 262,144,000 bytes . This quota applies to all the files you upload, including layers and custom runtimes. Try removing unnecessary packages,
If you are unable to reduce package size, you can perhaps explore changing from Lambda Python runtime to Lambda container image instead. The code package size limit is 10 GB
(maximum uncompressed image size, including all layers)
If need to process large files or load many large dependencies in AWS Lambda, you can do it by putting them on EFS volume which can be mounted to Lambda.