Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Unanswered
Hello Everyone! I'M Uploading A Dataset To The Server. During Dataset.Finalize() I Get The Following Error. Could You Please Advise What Could Be The Problem, Where Should I Start Looking?


@<1523701070390366208:profile|CostlyOstrich36> Then when I try to get the dataset I get the following error

Failed getting object size: RetryError('HTTPSConnectionPool(host='files.clearml.dbrain.io', port=443): Max retries exceeded with url: /Labeled%20datasets/.datasets/printed%20multilang%20crops/printed%20multilang%20crops.1e76fd4ad77f4d2790e4acf1c8241c59/artifacts/state/state.json (Caused by ResponseError('too many 503 error responses'))')
Could not download 
 , err: HTTPSConnectionPool(host='files.clearml.dbrain.io', port=443): Max retries exceeded with url: /Labeled%20datasets/.datasets/printed%20multilang%20crops/printed%20multilang%20crops.1e76fd4ad77f4d2790e4acf1c8241c59/artifacts/state/state.json (Caused by ResponseError('too many 502 error responses')) 
---------------------------------------------------------------------------
ValueError                                Traceback (most recent call last)
Cell In[34], line 1
----> 1 dataset1 = Dataset.get(dataset_id='1e76fd4ad77f4d2790e4acf1c8241c59')

File ~/miniconda3/lib/python3.9/site-packages/clearml/datasets/dataset.py:1731, in Dataset.get(cls, dataset_id, dataset_project, dataset_name, dataset_tags, only_completed, only_published, include_archived, auto_create, writable_copy, dataset_version, alias, overridable, shallow_search, **kwargs)
   1727     instance = Dataset.create(
   1728         dataset_name=dataset_name, dataset_project=dataset_project, dataset_tags=dataset_tags
   1729     )
   1730     return finish_dataset_get(instance, instance._id)
-> 1731 instance = get_instance(dataset_id)
   1732 # Now we have the requested dataset, but if we want a mutable copy instead, we create a new dataset with the
   1733 # current one as its parent. So one can add files to it and finalize as a new version.
   1734 if writable_copy:

File ~/miniconda3/lib/python3.9/site-packages/clearml/datasets/dataset.py:1643, in Dataset.get.<locals>.get_instance(dataset_id_)
   1635     local_state_file = StorageManager.get_local_copy(
   1636         remote_url=task.artifacts[cls.__state_entry_name].url,
   1637         cache_context=cls.__cache_context,
   (...)
   1640         force_download=force_download,
   1641     )
   1642     if not local_state_file:
-> 1643         raise ValueError("Could not load Dataset id={} state".format(task.id))
   1644 else:
   1645     # we could not find the serialized state, start empty
   1646     local_state_file = {}

ValueError: Could not load Dataset id=1e76fd4ad77f4d2790e4acf1c8241c59 state
  
  
Posted one year ago
167 Views
0 Answers
one year ago
one year ago