Hi,
I have a requirement where I need to process about 400K records which are in an excel. Of the 400K records, there could be about 75K records (based on some eligibility criteria) for which, I need to go to a web portal and obtain the required information and update the details in the portal. For rest of the records which did not satisfy eligibility criteria, no updates are necessary. There are two ways I can accomplish this task.
1) Using Work Queues
Approach:
a) read entire 400K records into a collection and load them to a work queue
b) use multi bot architecture to process the records and update the details in excel.
2) Using Single bot:
Approach:
a) Read each record from the excel and identify if the record needs to be updated.
b) if yes, get the details from the web portal and update the details in the excel.
c) repeat above two steps until end of the excel file.
I almost certainly need to develop the bot using first design approach (multi bot architecture) unless there is some thing that cannot be achieved using first approach. In this regard, below are the questions I have -
a) is there any limitation on the size of the collection I can use in Blue prism. would blue prism be able to read 400K records into a collection?
b) I know work queues can handle large amounts of data but would data size of 400k records be of any challenge for work queue to handle?
c) are there any pitfalls I need to look out for in this design?
Any suggestions/comments/feedback on the approaches mentioned above are welcome. If you have any other thoughts on any new approach, please let me know. Your inputs would be invaluable for me in designing the bot.