Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
SmallDeer34
Moderator
21 Questions, 155 Answers
  Active since 10 January 2023
  Last activity 8 months ago

Reputation

0

Badges 1

132 × Eureka!
0 Votes
6 Answers
1K Views
0 Votes 6 Answers 1K Views
Currently trying to figure out how to extend clearML's automagical reporting to JoeyNMT. https://github.com/joeynmt/joeynmt/blob/master/joey_demo.ipynb is a ...
3 years ago
0 Votes
3 Answers
1K Views
0 Votes 3 Answers 1K Views
So, I'm trying to do a several-step process, but it needs to run on a GPU queue in ClearML. How would I do that? Specifically, here's what I'm trying to do, ...
3 years ago
0 Votes
18 Answers
1K Views
0 Votes 18 Answers 1K Views
Is there any way to: within the UI, select and compare the scalars for more than 10 experiments? I'd like to do something like: select these 10 run in such a...
3 years ago
0 Votes
3 Answers
1K Views
0 Votes 3 Answers 1K Views
OK, we've got a GPU Queue setup on one of our local machines. I managed to run a script on it, which was intended to download a clearML dataset stored in s3....
3 years ago
0 Votes
28 Answers
1K Views
0 Votes 28 Answers 1K Views
3 years ago
0 Votes
7 Answers
977 Views
0 Votes 7 Answers 977 Views
Question about https://allegro.ai/clearml/docs/rst/references/clearml_python_ref/task_module/task_task.html#clearml.task.Task.upload_artifact : Let's say I g...
3 years ago
0 Votes
9 Answers
1K Views
0 Votes 9 Answers 1K Views
3 years ago
0 Votes
30 Answers
1K Views
0 Votes 30 Answers 1K Views
3 years ago
0 Votes
30 Answers
986 Views
0 Votes 30 Answers 986 Views
Hello! Getting credential errors when attempting to pip install transformers from git repo, on a GPU Queue. fatal: unable to write credential store: Device o...
3 years ago
0 Votes
4 Answers
1K Views
0 Votes 4 Answers 1K Views
OK, next question, I've got some training args that I'd like to manually upload and have them show up in the attached place, under Configuration. It is a Hug...
3 years ago
0 Votes
8 Answers
1K Views
0 Votes 8 Answers 1K Views
So I'm in a Colab notebook, and after running my Trainer(), how do I upload my test metrics to ClearML? ClearML caught these metrics and uploaded them: train...
3 years ago
0 Votes
18 Answers
1K Views
0 Votes 18 Answers 1K Views
Second: is there a way to take internally tracked training runs and publish them publicly, e.g. for a research paper? "Appendix A: training runs can be found...
2 years ago
0 Votes
13 Answers
993 Views
0 Votes 13 Answers 993 Views
How, if at all, should we cite ClearML in a research paper? Would you like us to? How about a footnote/acknowledgement?
3 years ago
0 Votes
21 Answers
1K Views
0 Votes 21 Answers 1K Views
2 years ago
0 Votes
0 Answers
1K Views
0 Votes 0 Answers 1K Views
So, this is something I've noticed, this line always seems to crash my Colab Notebooks: Task.current_task().completed()
3 years ago
0 Votes
7 Answers
1K Views
0 Votes 7 Answers 1K Views
Is there any way to get just one dataset folder of a Dataset? e.g. only "train" or only "dev"?
3 years ago
0 Votes
0 Answers
1K Views
0 Votes 0 Answers 1K Views
Here's the original Colab notebook. It can import torch without error: https://colab.research.google.com/github/huggingface/blog/blob/master/notebooks/01_how...
3 years ago
0 Votes
13 Answers
1K Views
0 Votes 13 Answers 1K Views
Hello, there's a particular metric (perplexity) I'd like to track, but clearML didn't seem to catch it. Specifically, this "Evaluation" section of run_mlm.py...
3 years ago
0 Votes
10 Answers
1K Views
0 Votes 10 Answers 1K Views
So, I did a slew of pretrainings, then finetuned those pretrained models. Is there a way to go backwards from the finetuning Task ID to the pretraining Task ...
3 years ago
0 Votes
3 Answers
977 Views
0 Votes 3 Answers 977 Views
Question: has anyone done anything with Ray or RLLib, and ClearML? Would ClearML be able to integrate with those out of the box? https://medium.com/distribut...
3 years ago
0 Votes
10 Answers
1K Views
0 Votes 10 Answers 1K Views
Hello! I'm just starting out with ClearML, and I seem to be having some sort of conflict between clearml and torch , at least in Colab In this guide ( https:...
3 years ago
0 Hi, There Is Small Bug In The Web Ui When Comparing Two Experiments Scalars: If The Two Tasks Have The Same Name, Then Clicking On The “Maximize Graph” Button On One Scalar Series To Get The Bigger View On That Scalar Series, Then The Color Of Both Series

I've been trying to do things like "color these five experiments one color, color these other five a different color", but then once I maximize the thing the colors all change

3 years ago
0 Currently Trying To Figure Out How To Extend Clearml'S Automagical Reporting To Joeynmt.

It seems to create a folder and put things into it, I was hoping to just observe the tensorboard folder

3 years ago
3 years ago
0 This Will Close It

It's not a big deal because it happens after I'm done with everything, I can just reset the Colab runtime and start over

3 years ago
3 years ago
0 This Will Close It

Yup, I just wanted to mark it completed, honestly. But then when I run it, Colab crashes.

3 years ago
0 So I'M In A Colab Notebook, And After Running My Trainer(), How Do I Upload My Test Metrics To Clearml? Clearml Caught These Metrics And Uploaded Them:

This seems to work:

` from clearml import Logger
for test_metric in posttrain_metrics:
print(test_metric, posttrain_metrics[test_metric])

#report_scalar(title, series, value, iteration)
Logger.current_logger().report_scalar("test", test_metric, posttrain_metrics[test_metric], 0) `

3 years ago
3 years ago
0 Is There Any Way To: Within The Ui, Select And Compare The Scalars For More Than 10 Experiments? I'D Like To Do Something Like:

And the reason is, because I have a bunch of "runs" with the same settings, and I want to compare broadly across several settings. So if I select "a bunch" with setting A I can see a general pattern when compared with setting B.

I've got 7-10 runs per setting, and about 7 or 8 settings

3 years ago
0 My Nth Question For The Day

seconded, I'm curious about this also.

3 years ago
0 I Must Compliment The Python Level And Documentation Level In The Source Code Is Superb, I Love Reading It

Ah, makes sense! Have you considered adding a "this is the old website! Click here to get to the new one!" banner, kinda like on docs for python2 functions? https://docs.python.org/2.7/library/string.html

3 years ago
0 So, I'M Trying To Do A Several-Step Process, But It Needs To Run On A Gpu Queue In Clearml. How Would I Do That? Specifically, Here'S What I'M Trying To Do, Is It Possible?

Gave it a try, it seems our GPU Queue doesn't have the S3 creds set up correctly. Making a separate thread about that

3 years ago
0 Ok, Next Question, I'Ve Got Some Training Args That I'D Like To Manually Upload And Have Them Show Up In The Attached Place, Under Configuration. It Is A Huggingface Trainingarguments Object, Which Has A To_Dict() And To_Json Function

So for example:
` {'output_dir': 'shiba_ner_trainer', 'overwrite_output_dir': False, 'do_train': True, 'do_eval': True, 'do_predict': True, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 16, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'learning_rate': 0.0004, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam...

3 years ago
0 Ok, Next Question, I'Ve Got Some Training Args That I'D Like To Manually Upload And Have Them Show Up In The Attached Place, Under Configuration. It Is A Huggingface Trainingarguments Object, Which Has A To_Dict() And To_Json Function

OK, I guess
` training_args_dict = training_args.to_dict()

Task.current_task().set_parameters_as_dict(training_args_dict) `works, but how to change the name from "General"?

3 years ago
0 My Nth Question For The Day

AgitatedDove14 I'm making some progress on this. I've currently got the situation that my training run saved all of these files, and Task.get_task(param['TaskA']).models['output''][-1] gets me just one of them, training_args.bin . Then -2 gets me another, rng_state.pth

If I just get Task.get_task(param['TaskA']).models['output'] , I end up getting a huge list of, like, ` [<clearml.model.Model object at 0x7fec2841c880>, <clearml.model.Model object at 0x7fec2841...

3 years ago
Show more results compactanswers