Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

index.cache.json files way bigger than common indexes #1873

Closed
vlantier opened this issue Dec 11, 2019 · 3 comments · Fixed by #1952
Closed

index.cache.json files way bigger than common indexes #1873

vlantier opened this issue Dec 11, 2019 · 3 comments · Fixed by #1952
Labels

Comments

@vlantier
Copy link

Thanos, Prometheus and Golang version used:
Thanos: 0.9.0
Go: 1.13.1
Prometheus: 2.14

Object Storage Provider: GCS

What happened:
Our index.cache.json files for a couple of days are being created greater than the normal index files. This is expected?

Comparing with 5 days agor we have index files of 1G and 300/400MB cache.json. No we have index with 1.8G and caches with 1.7.

I think that the big files are responsaible for the actual overload in thanos-store that we have, because only when we configure min-size and max-size for the correspondent days that we have the bigger cache they crash.

The compactor is running without errors.

What you expected to happen:
The previous behavior: Cache files way small than index files

Anything else we need to know:

@bwplotka
Copy link
Member

Yes, the reason behind that is that index.cache.json is unoptimized e.g strings are not interned. With huge number of labels and long strings this might be the case, but your case looks really extreme.

We are working on this to resolve your issue: #1839

@vlantier
Copy link
Author

I see.
Thanks for the responde @bwplotka, I'll be watching this issue..

With the new Prometheus 2.14 we saw isome Highest Cardinality Metric Names and Highest Cardinality Labels who may impact in this matter. Those metrics are from home-made exporters and we are working on it.

But something is odd because those metrics are there for a long time already. May you have some tips about where to look?

@stale
Copy link

stale bot commented Jan 11, 2020

This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions.

@stale stale bot added the stale label Jan 11, 2020
@stale stale bot closed this as completed Jan 18, 2020
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

Successfully merging a pull request may close this issue.

2 participants