Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
MysteriousSeahorse54
Moderator
1 Question, 9 Answers
  Active since 10 January 2023
  Last activity one year ago

Reputation

0

Badges 1

9 × Eureka!
0 Votes
15 Answers
586 Views
0 Votes 15 Answers 586 Views
Hi all, I hope someone can help. I am using ClearML agents with docker containers to train RL models with stable baselines 3 on an on-premise server. I am ha...
one year ago
0 Hi All, I Hope Someone Can Help. I Am Using Clearml Agents With Docker Containers To Train Rl Models With Stable Baselines 3 On An On-Premise Server. I Am Having Issues With Saving And Loading Models. If I Don'T Specify

Hi AnxiousSeal95 , the models are saved both with a weights and biases call back and through stable baselines 3 model.save. Yes that makes sense to me that the files local to the docker container can't be downloaded. But yes when setting output_uri to true no models appear in the UI at all which seems strange

one year ago
0 Hi All, I Hope Someone Can Help. I Am Using Clearml Agents With Docker Containers To Train Rl Models With Stable Baselines 3 On An On-Premise Server. I Am Having Issues With Saving And Loading Models. If I Don'T Specify

Manual logging has the same behavior. When the output destination is not set the model artifacts are saved but can't be downloaded. They are saved to the docker in which they ran and not the fileserver. When the output uri is set the artifacts don't appear at all.

one year ago
0 Hi All, I Hope Someone Can Help. I Am Using Clearml Agents With Docker Containers To Train Rl Models With Stable Baselines 3 On An On-Premise Server. I Am Having Issues With Saving And Loading Models. If I Don'T Specify

Sure, here is a snippet.
` run = wandb.init(project="rsTest",sync_tensorboard=True)

add tensorboard logging to the model

model = PPO('MlpPolicy', env, verbose=1, tensorboard_log=f"runs/{run.id}",
learning_rate=args.learning_rate,
batch_size=args.batch_size,
n_steps=args.n_steps,
n_epochs=args.n_epochs,
device='cpu')

create wandb callback

wandb_callback = WandbCallback(model_save_freq=1000,
model...

one year ago