Am looking for suggestions from experts to solution a use case where we do data migration to dynamo db . The number we are expecting here is 10 to 15 million records . All these records will be coming in chunks over multiple files to s3 and am expecting it to be 30plus files. When file arrived in s3 , it triggers lambda which gonna process the records and push it to DynamoDB .
The lambda function is all set to process records in each file . But the major concern I have here is
What if 30 to 40 files comes one after another , how can I ensure , there is enough lambda instance to receive the trigger and process the files . Should I be setting enough RC ?
What will happen if I set RC to 20 and files that arrived are more than 20 ?
What is best practice to make sure all files are processed, and processing parallely would be best option as the numbers we expect is huge .
Looking for your suggestions.