Looking for users who have had experience implementing a spatial temporal data store into their environment. Specially looking for the impacts on CPU/RAM & the disk space big data takes up.
I was searching for some information in regards to how much disk space should we request for a Spatial Temporal Data Store?
Was specifically looking to use location tracking to automatically capture where users have been at lets say 1 min intervals.
Lets say with the following variables:
- a work day is 60min x 8 hours = 480 points per user
- 50 users
- 52 weeks x 5 days = 260 days
Gives me around 6.3 million points in a year.
Does anybody know how much disk space this would approximate take in a year?
We do have a point based feature class in a file geodatabase with 18 million points with 30 attributes and it is 5GB in size. Would it be a good estimate to say that the location tracking will take up approx 2GB a year?
CPU & RAM
I have found that the minimum memory requirement is 16GB to start.
Lets say there is 50 users hitting the environment, that is an additional memory requirement.
Can't see any information on whether it has an impact on CPU but I was assuming it would.
Has anybody had any experience on the impact on CPU and memory?