Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Answered
Hi Everyone, I'M Using Torch.Distributed For Training On 2 Gpus. It Works, But Each Gpu Creates A New (Duplicated) Task, And I Prefer To Have Only One Clearml Experiment Running. I Looked Here

Hi everyone,
I'm using torch.distributed for training on 2 GPUs. It works, but each GPU creates a new (duplicated) task, and I prefer to have only one ClearML experiment running. I looked here None , but was not able to solve this issue.
Any thoughts? Thanks!

  
  
Posted 10 months ago
Votes Newest

Answers 2


Hi @<1578918167965601792:profile|DistinctBeetle43> ! This is currently not possible. A different task will be created for each instance

  
  
Posted 10 months ago

Thanks!

  
  
Posted 10 months ago
566 Views
2 Answers
10 months ago
10 months ago
Tags