In order to ensure that our customer database queries remains performant, GitClear uses different levels of caching based on the age of the data being stored. When querying the API, if you request data that is outside the caching limits, we can only approximate the value of data that occurred in intervals shorter than the cache record size.


Record size

Cache limit

Effective date range

Daily

31

One month

Weekly

52

One year

Monthly

36

Three years

Yearly

20

20 years


If you need extremely granular data outside these limits, we recommend utilizing the Restful API (specifically the Reports API or the Audits API) to cache the segments that are important to you, to the extent that you want historical granularity to extend longer than the GitClear defaults.


linkRationale, Potential for a New Tier?

Since GitClear instruments more than 100 different segments, on a per-committer, per-team, per-organization, per-repo, and per-segment (e.g., for each type of line operation GitClear analyzes) basis, the data that GitClear generates and stores is very exponential in its accumulation (think "fold a piece of paper 42 times to reach the moon"). Several terabytes of database storage (which must be zipped & backed up to redundant locations on a daily basis) are necessary to maintain the current tiers of granularity.


If customers are interested in storing granular data over a longer time range than currently available, we may add another tier of service that extends the data preservation window. Please email hello@gitclear.com if you would pay a surcharge per month to gain access to more granular data over a longer time window (vs querying the API to store your data). We could effectively double the length of the data preservation tiers for an extra $5/month/dev. If we get enough interest, we will add an "Extra Data Preservation" option our roadmap.