Hi,
I have a datable containing 100 000 rows.
I need to iterate this datatable and do some operation for each row.
Obviously I cannot do that on one service (I will have 100% timeout problem) and I cannot increase the timeout value.
The solution I am thinking of:
1-Take rows slice by slice. (50 rows every time).
2-When the 50 iterations are done , I trigger the next 50 rows (subscription on a property).
3-I finish with I reach the count of all rows in the datatable.
What do you think about this solution ?
Is there a better alternative ?
How to get the 50 rows I need to process since I don't have an id that increment automatically (so I cannot see how to manage the query to get the right 50 rows I need to process)
Thank you for your help
Solved! Go to Solution.
Hi @AliBenBelgacem,
I suggest (if you have the CSV-Parser)
This will atleast reduce the time for 'GetDataTabelEntries' you do on the initial step.
Just a though from top of head
Thanks,
Hi @AliBenBelgacem can you take a look at the below reference, and let us know your output?
https://community.ptc.com/t5/ThingWorx-Developers/Iterate-Datatable/m-p/646167#M42203
Hi @AliBenBelgacem,
I suggest (if you have the CSV-Parser)
This will atleast reduce the time for 'GetDataTabelEntries' you do on the initial step.
Just a though from top of head
Thanks,
If you found either of the responses helpful, please mark the appropriate one as the Accepted Solution for the benefit of others in the community.
Regards.
--Sharon