Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Answered
Is There A Way To Save The Models Completely On The Clearml Server? It Seems That Clearml Server Does Not Store The Models Or Artifacts Itself, But They Are Stored Somewhere Else (E.G., Aws S3-Bucket) Or On My Local Machine And Clearml Server Is Only Sto

Is there a way to save the models completely on the ClearML server?

It seems that ClearML Server does not store the models or artifacts itself, but they are stored somewhere else (e.g., AWS S3-bucket) or on my local machine and ClearML Server is only storing configuration parameters and previews (e.g., when the artifact is a pandas dataframe). Is that right?

  
  
Posted 2 years ago
Votes Newest

Answers 45


UnevenDolphin73 : From which URL is your most recent screenshot?

  
  
Posted 2 years ago

It should store it on the fileserver, perhaps you're missing a configuration option somewhere?

  
  
Posted 2 years ago

It is documented at None ... super deep in the code. If you don't know that output_uri in TASK's (!) init is relevant, you would never know...

  
  
Posted 2 years ago

Heh, well, John wrote that in the first reply in this thread 🙂
And in Task.init main documentation page (nowhere near the code), it says the following -
image

  
  
Posted 2 years ago

FWIW It’s also listed in other places VivaciousBadger56 , e.g. None says:

In order to make sure we also automatically upload the model snapshot (instead of saving its local path), we need to pass a storage location for the model files to be uploaded to.
For example, upload all snapshots to an S3 bucket…

  
  
Posted 2 years ago

Do you mean "exactly" as in "you finally got it" or in the sense of "yes, that was easy to miss"?

  
  
Posted 2 years ago

VivaciousBadger56 I'm not sure I'm following you - is the issue not being able to upload to the ClearML server or to load the downloaded file?

  
  
Posted 2 years ago

We'll try to add referenced to that in other places as well 👍

  
  
Posted 2 years ago

FWIW, we prefer to set it in the agent’s configuration file, then it’s all automatic

  
  
Posted 2 years ago

Yes, you're correct, I misread the exception.
Maybe it hasn't completed uploading? At least for Datasets one needs to explicitly wait IIRC

  
  
Posted 2 years ago

"Messy" is putting it nicely.

  
  
Posted 2 years ago

By the way, output_uri is also documented as part of the Task.init() docstring ( None )

  
  
Posted 2 years ago

UnevenDolphin73 : Thanks, but it does not mention the File Storage of "ClearML Hosted Server".

  
  
Posted 2 years ago

From the one you sent - None

  
  
Posted 2 years ago

The documentation is messy, I’ve complained about it the in the past too 🙈

  
  
Posted 2 years ago

Hi all, sorry for not being so responsive today 🙏

  
  
Posted 2 years ago

I can only say I’ve found ClearML to be very helpful, even given the documentation issue.
I think they’ve been working on upgrading it for a while, hopefully something new comes out soon.
Maybe AgitatedDove14 has further info 🙂

  
  
Posted 2 years ago

I have already been trying to contribute (have three pull requests), but honestly I feel it is a bit weird, that I need to update a documentation about something I do not understand, while I actually try to evaluate if ClearML is the right tool for our company...

  
  
Posted 2 years ago

We're certainly working hard on improving the documentation (and I do apologize for the frustrating experience)

  
  
Posted 2 years ago

UnevenDolphin73 : I do not get this impression, because during update_weights I get the message

2023-02-21 13:54:49,185 - clearml.model - INFO - No output storage destination defined, registering local model C:\Users..._Demodaten_FF_2023-02-21_13-53-51.624362.model

  
  
Posted 2 years ago

SuccessfulKoala55 : I referenced this conversation in the issue None

  
  
Posted 2 years ago

missing a configuration option

Which one, where? Any idea? I did not set output_uri - do I have to do that?

I am refering to

  
  
Posted 2 years ago

Heh, good VivaciousBadger56 😁
I was just repeating what CostlyOstrich36 suggested, credits to him

  
  
Posted 2 years ago

🙂

  
  
Posted 2 years ago

UnevenDolphin73 : How do you figure? In the past, my colleagues and I just shared the .zip file via email / MS Teams and it worked. So I don't think so.

  
  
Posted 2 years ago

But we do use S3

  
  
Posted 2 years ago

Exactly 🙂

  
  
Posted 2 years ago

SuccessfulKoala55 Also, I think that - in this case, but also in other cases - the issue is not just the documentation, but also the design of the SDK.

  
  
Posted 2 years ago

SuccessfulKoala55 I think I might have made a mistake earlier - but not in the code I posted before. Now, I have the following situation:

  • In my training Python process on my notebook I train the custom made model and put it on my harddrive as a zip file. Then I run the code
output_model = OutputModel(task=task, config_dict={...}, name=f"...")
output_model.update_weights(weights_filename=r"C:\path\to\mymodel.zip", is_package=True)
  1. I delete the "C:\path\to\mymodel.zip", because it would not be available on my colleagues' computers.

  2. In a second process, the model-inference process, I run

mymodel = task.models['output'][-1]
mymodel = mymodel.get_local_copy(extract_archive=True, raise_on_error=True)

and get the error

ValueError: Could not retrieve a local copy of model weights 8ad4db1561474c43b0747f7e69d241a6, failed downloading

I do not have an aws S3 instance or something like that. This is why I would like to store my mymodel.zip file directly on the ClearML Hosted Service. The model is around 2MB large.

How should I proceed?

  
  
Posted 2 years ago
95K Views
45 Answers
2 years ago
2 years ago
Tags