Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Answered
Hi Everyone, Is There A Way To Increase The Cache Size Of Each Clearml Task? I'M Running An Experiment And Many Artifacts Are Downloaded. My Dataloader Fails To Load Some Of The File Since They Are Missing, Although They Were Downloaded. I Guess There Is

Hi everyone,
Is there a way to increase the cache size of each ClearML task?
I'm running an experiment and many artifacts are downloaded. My dataloader fails to load some of the file since they are missing, although they were downloaded. I guess there is a limit for caching artifacts for each task, so some the downloaded files are deleted after exceeding the limit.
Thank you in advance.

Here is the error:
FileNotFoundError: [Errno 2] No such file or directory: '/home/developer/.clearml/cache/storage_manager/global/94fc7259e0247a7781982a95a8a9fc5a.xxx_xxx_xx_x_features.npz'
before reaching the data loader, I checked it and the file was available under this path.

  
  
Posted one year ago
Votes Newest

Answers 12


Are you suggesting the conf file did not set the default size? It sounds like a bug, can you verify?

  
  
Posted one year ago

Hi ScaryKoala63
Sure, add the following to your clearml.conf:
sdk.storage.cache.default_cache_manager_size = 400I think you are correct, it seems like for some reason you hit the cache limit, and a previous entry was deleted

  
  
Posted one year ago

Thanks 🙂

  
  
Posted one year ago

Yes, it happens on the same file.
I try to provide a code snippet. No, it's a single docker container.

  
  
Posted one year ago

And when retrieve just this file? is it working ?
(Maybe for some reason the file is corrupted) ?

  
  
Posted one year ago

I resolved it by adding this line:
StorageManager.set_cache_file_limit(10000, cache_context='global')

  
  
Posted one year ago

I increased the cache size a lot but still I get the same error.

  
  
Posted one year ago

Is this consistent on the same file? can you provide a code snippet to reproduce (or understand the flow) ?
Could it be two machines are accessing the same cache folder ?

  
  
Posted one year ago

ScaryKoala63
When it fails what's the number of files you have in:
/home/developer/.clearml/cache/storage_manager/global/ ?

  
  
Posted one year ago

AgitatedDove14 The number is 1001 and I set the default_cache_manager_size to 20000.

  
  
Posted one year ago

ScaryKoala63 nice!!!!!

  
  
Posted one year ago

Yes, I tried updating clearml.conf file but didn't help. finally adding the line I mentioned with cache_contex='global' resolved the issue.

  
  
Posted one year ago
610 Views
12 Answers
one year ago
one year ago
Tags
Similar posts