Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
FrothyDog40
Moderator
0 Questions, 60 Answers
  Active since 10 January 2023
  Last activity 10 months ago

Reputation

0
0 Hey, I Tried To Use

RotundHedgehog76 Have you tried clearml-data add --files .
? (Probably best to try on a smaller subset first)

one year ago
0 Hey, I Hope This Is The Right Place To Ask. We'Re A Small Data Science Team That Wants To Log Everything About Our Ml Models. Looking Around On The Internet, Mostly Mlflow Is Being Recommended, But Occasionally The Name Trains Pop-Ups. According To You,

DefeatedCrab47 For the most part, mlflow can serve basic ML models using scikit-learn. In contrast, Trains was designed with more general purpose ML/DL workflows in mind, for which there's no "generic" way to serve models as different scenarios can use different input encoding, models results would be represented in a variety of forms, etc.
Consider also, that creating an HTTP endpoint for model inference is quite a breeze: there are multiple examples of Flask on top of any DL/ML framework w...

3 years ago
0 Hey, I Hope This Is The Right Place To Ask. We'Re A Small Data Science Team That Wants To Log Everything About Our Ml Models. Looking Around On The Internet, Mostly Mlflow Is Being Recommended, But Occasionally The Name Trains Pop-Ups. According To You,

DefeatedCrab47 Happy you're finding Trains useful 🙂

but it definitely has it's advantages if TRAINS would support it (early stage Data Science infrastructure).

No doubt, and I definitely see such usable example in the cards for Trains' upcoming versions...

3 years ago
0 The Links To Pytorch Lightning Are Broken In The

DefeatedCrab47 Thanks for pointing it out.
We'll get in touch with the PyTorch Lightning team to better understand the code restructure they're effecting (see https://github.com/PyTorchLightning/pytorch-lightning/pull/2384 ).
In the mean time, you can look at the prior version: https://github.com/PyTorchLightning/pytorch-lightning/blob/0.8.1/pytorch_lightning/loggers/trains.py

3 years ago
0 When Uploading An Artifact, Can I List It In Some Grouping (Like With Parameters, Having E.G.

UnevenDolphin73 I think it'd be easier to track as a separate one.

one year ago
0 Hi, Is There A Way To Log Sklearn Metrices (Like Accuracy/Precision) In A Tabular Way Rather Than Plot ?

HappyDove3 Notice that in https://github.com/allegroai/clearml/issues/400 the goal is to see a table plot in the UI scalars tab for a specific experiment (with additional discussions on how these will be addressed when comparing experiments).
Note that once you take the approach you suggested of logging your metrics single values, you can configure your experiment comparison scalars view to show single values instead of the time-series graph which I think will provide you with the matrix c...

one year ago
0 Hi! I Just Spent Several Hours Of My Colleague'S And My Working Time Figuring Out Why I Cannot Enqueue An Experiment To A Certain Queue In The Web Ui. Turns Out I Just Needed To Scroll Down The List Of Queues, There Were Some Hidden Ones Below. This Is No

MelancholyElk85 Thanks for calling this to attention. What do you think would have made it easier for you to notice the available extended list content?
I would assume that a "type to match" option would also have helped?
Appreciate if you could https://github.com/allegroai/clearml/issues/new/choose so this can be pushed forward.

one year ago
0 When Uploading An Artifact, Can I List It In Some Grouping (Like With Parameters, Having E.G.

UnevenDolphin73 Well... not right now... Currently the ClearML UI only partitions internal artifact types.
That said, having user-defined artifact groups sure sounds worth looking into - Care to https://github.com/allegroai/clearml/issues/new/choose ?

one year ago
0 In Your

Hi DefeatedCrab47 ,
The examples folder has just been restructured: Find the example here:
https://github.com/allegroai/trains/blob/master/examples/services/hyper-parameter-optimization/hyper_parameter_optimizer.py

3 years ago
0 Ui Suggestion #2:

DepressedChimpanzee34 Have you noticed the "Show n experiments selected" button on the bottom bar? This effectively toggles your view between whatever is currently sorted/filtered and the current item selection.
To address the scenario you describe: Switch to "Show selected experiments", remove the redundant items, and switch back to the original view: "Show all experiments"

Thoughts?

2 years ago
0 Ui Suggestion #2:

DepressedChimpanzee34 Apologies for missing your previous comment.
Totally agree that the global selection indicator should maintain its 'clear selection' behaviour even if some/all of the selection is off-screen.

2 years ago
0 Hi Again

SuperiorPanda77 Would love to get a PR if you're up for it

one year ago
0 Hello, I Want To Update Allegro. I Found The Link:

UnsightlySeagull42 The upgrade process is slightly different depending on the environment in which you've deployed your ClearML server (e.g. for a https://allegro.ai/clearml/docs/docs/deploying_clearml/clearml_server_linux_mac.html#upgrading ).
Note the document you are referring to only applies once when you're moving from the older pre-0.16 versions in which case DB migration is required.
If your server is more up to date (0.16 and newer) you should be OK with the link above.

2 years ago
0 More Stuff I Think People Would Find Useful For When Comparing Experiments: Comparing

WittyOwl57 No worries 🙂 happens to the best!

2 years ago
0 Hi All, Is There A Way To Filter A Experiments In A Hyperparameter Sweep Based On A Given Range Of A Parameter/Metric In The Ui (Similar To

ExcitedFish86 You can https://clear.ml/docs/latest/docs/webapp/webapp_exp_table#adding-metrics-and--or-hyperparameters to include any parameter/metric column that helps your analysis (and subsequently filter the table on those columns).
There's not yet the equivalent of a parameter importance visualization, though such insight visualizations are definitely in our sights.
Sure appreciate if you can https://github.com/allegroai/clearml/issues/new on the subject :)

2 years ago
0 More Stuff I Think People Would Find Useful For When Comparing Experiments: Comparing

WittyOwl57 The UI shows repo and package detailed comparison under the "Details"/"Execution" (See sample screenshot), whereas auto-logged environment variables are shown under the "HyperParameters" comparison tab.
What do you find missing beyond those?

2 years ago
0 Can We Somehow Add A Column "Duration" To The List Of Experiments ?

TightElk12 This makes a lot of sense - should make it into one of the coming releases

3 years ago
0 Can I Run An Autoscaler Listening To A Regular Queue (I.E. Combine Autoscaling With On Premise Machines)? Or Does It Run The Danger Of Creating An Ec2 Instance While An On-Premise Machine Takes The Job?

UnevenDolphin73 Am I missing anything in rephrasing your use case to "Have a single autoscaler service multiple queues" (where the autoscaler resource configuration is, in essence, the pool you mention)?

one year ago
0 Ui Suggestion:

DepressedChimpanzee34 Experience has shown that some mechanisms for mitigating large sets impact on browser performance are required.
Your 2nd suggestion for adding an in-app search tool for such sections seems to be completely in line with ClearML's behaviour in other UI sections (e.g. console logs) - It'd be great if you can https://github.com/allegroai/clearml/issues/new/choose

2 years ago
0 Hi Team, Wondering If There Is A Better Way To Do Table Reporting? Currently Am Using The Report_Table() Function But The Display Doesn'T Support Custom Column Width, New Line, Etc. Thanks.

@<1628927672681762816:profile|GreasyKitten62> When you have specific display considerations, you can implement them through report_table's 'extra_layout' and 'extra_data' parameters

27 days ago
0 Hi, I Logged My Params To Clearml Like This:

HappyDove3 you can get some more insight on the different configuration methods and how to use theme https://clear.ml/docs/latest/docs/fundamentals/hyperparameters

one year ago
0 Hi, I Have A Question Regarding The

OutrageousSheep60 You can see https://github.com/allegroai/clearml/issues/724 a discussion on the topic.
TL;DR:
Currently the containing project is available in the UI as a tooltip to the dataset name An alternate "Project view" to the datasets page is in the works

one year ago
0 More Stuff I Think People Would Find Useful For When Comparing Experiments: Comparing

WittyOwl57 Is that information available for you on each of the compared experiments when you view them individually?

2 years ago
0 Hi, In The Clearml Web Ui, Is It Possible To Edit The Description Of A Project For Which The Description Was Initially Empty? When There Is A Description There Is An "Edit" Button On The Overview Tab. When There Is No Description There Is No "Edit" Button

GreasyPenguin14 That's an annoying bug indeed - Thanks for spotting it. If you need to circumvent it before a fix comes out in one of the near releases, you can programatically use the https://clear.ml/docs/latest/docs/references/api/endpoints#post-projectsupdate e.g.
from clearml.backend_api.session.client import APIClient client = APIClient() client.projects.update(project='<project ID>', description='My new description')
Note you can get your project's ID either from the webapp URL...

2 years ago
Show more results compactanswers