In this article, we are going to tackle all the drawbacks from my previous article like model load time, and dependency size, and usage.

We are going to build the same “Serverless BERT powered Question-Answering API” as last time. But instead of using compressing techniques to fit our Python dependencies into our AWS Lambda function, we are using a tool called efsync. I built efsync to automatically upload dependencies to an AWS EFS filesystem and then mount them into our AWS Lambda function. This allows us to include our machine learning model into our function without the need to load it from S3.

#deep-learning #bert #aws #machine-learning

New Serverless Bert with Huggingface, AWS Lambda, and AWS EFS
2.75 GEEK