Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
LudicrousParrot69
Moderator
7 Questions, 37 Answers
  Active since 10 January 2023
  Last activity one year ago

Reputation

0

Badges 1

32 × Eureka!
0 Votes
3 Answers
902 Views
0 Votes 3 Answers 902 Views
3 years ago
0 Votes
5 Answers
1K Views
0 Votes 5 Answers 1K Views
Hey all! Ive gone through the doco and not found anything at the moment, but does ClearML have model versioning and staging (similar to mlflow). GrumpyPengui...
3 years ago
0 Votes
10 Answers
888 Views
0 Votes 10 Answers 888 Views
3 years ago
0 Votes
13 Answers
1K Views
0 Votes 13 Answers 1K Views
3 years ago
0 Votes
8 Answers
1K Views
0 Votes 8 Answers 1K Views
Hi all, playing around with HP optimisation, and I notice in the HyperParameterOptimizer class itself, the objective_metric_sign is set to min by default. Ca...
3 years ago
0 Votes
10 Answers
917 Views
0 Votes 10 Answers 917 Views
3 years ago
0 Votes
23 Answers
1K Views
0 Votes 23 Answers 1K Views
Hi team, me again! Im curious if someone can explain to me better how task and optimisers integrate with each other. In the example hyperparameter optimisati...
3 years ago
0 Hey All! Ive Gone Through The Doco And Not Found Anything At The Moment, But Does Clearml Have Model Versioning And Staging (Similar To Mlflow).

Yeah its trying to plan down the line into model deployment. Whilst its easy to save out a keras SavedModel or similar and have that artifact uploaded into the store, just wanted to check if there was a more generic solution. I could just create a Python class and serialise that out such that it has a standard interface, but good to check. So for example, some artifact representing an arbitrary math function. For better context, the idea is to make deploying any artifact we upload using clear...

3 years ago
0 Hi, I Try To Report A Plot To Clearml In This Way:

So Im guessing either return_fig is False when you ran this, or plt.cdf is having issues (which it can). Id always recommend explicitly creating the figure instead of using the plt shortcuts.

fix, ax = plt.subplots()
ax.plot(โ€ฆ)

etc

3 years ago
0 Hi, I Try To Report A Plot To Clearml In This Way:

Yes, this is expected, because you are trying to log a matplotlib figure as if it was a plotly figure. Try using report_matplotlib_figure

3 years ago
0 Hi, I Try To Report A Plot To Clearml In This Way:

Can you print out fig to verify it at all? The logger should call to_plotly_json on the figure, so Im guessing its None

3 years ago
0 Hi, I Try To Report A Plot To Clearml In This Way:

I meant the object instance itself so we can see what class it is and if its initialised. Also that looks like a matplotlib plot, not a plotly plot which you logging call indicates

3 years ago
0 Hi, I Try To Report A Plot To Clearml In This Way:

Can you throw a screenshot of the code where fig is defined?

3 years ago
0 Hi All, Playing Around With Hp Optimisation, And I Notice In The Hyperparameteroptimizer Class Itself, The

Okay, that makes sense then. Whats still got me scratching my head is the examples printing out the WOOT WOOT for breaking the record, despite it clearly not. Hmm.

3 years ago
0 On A Similar Note, The Million Autogenerated Experiments When Doing Tuning Swamp Out Everything Else In The Experiments And Models Tabs. Is There A Current Solution To Hide Autogenerated Runs, Give Them Specific Tags, Etc, Or Is This Not Yet Possible? Sor

+1 for autoarchiving. Right now the interface feels incredibly clunky to use once the number of HPO trials starts to increase. I currently have a demo project and have different algos to make predictions (a simple keras model, a RF, etc). Ideally Iโ€™d want to see the HPO execution just once with all the trials underneath it, or just the top (few) models. At the moment, I have pages and pages of models, 99% of them I dont care about. Is it possible to archive models and set tags in the code rat...

3 years ago
0 Hey All! Ive Gone Through The Doco And Not Found Anything At The Moment, But Does Clearml Have Model Versioning And Staging (Similar To Mlflow).

Ah fantastic, thanks! Another one for me - is there support for custom python models at all? For example, dummy models that simply return the output of an equation run over the dataframe after transforming some of the input columns. Something similar to mlflows custom pyfunc that allows a standard way of interfacing with custom models as you do with keras/sklearn/pytorch models

3 years ago
0 Potential Feature Request: Having The Parallel Coordinates Plot Available From The Hp Parent Task. Right Now, If I Want To See The Parallel Coord Plot (Shown Below), I Have To Manually Select All Trials In A Hpo Run > Compare > Hyperparameters > Parallel

How so? The view can be used to just initialise the plot above like I had done all the steps by hand, but any flexibility (add in other scalars/parameters) Id expect would still be viable options in the UI

3 years ago
0 Can I Make A Super Small Fr Or See If This Already Exists. I Want To Ensure/Add A Tag On A Run, But There Is No Add_Tag. Set_Tags Allows Duplication, Which Isnt Something I Think Is Useful With Tags (And Cant Be Done On The Ui I Believe). Currently, I Si

How odd, its not in my task.py file!

Let me checkโ€ฆ

Ah. The code inspector is jumping to backend_interface.task.py, not clearml.task.py, how embarrassing for me!

3 years ago
0 Can I Make A Super Small Fr Or See If This Already Exists. I Want To Ensure/Add A Tag On A Run, But There Is No Add_Tag. Set_Tags Allows Duplication, Which Isnt Something I Think Is Useful With Tags (And Cant Be Done On The Ui I Believe). Currently, I Si

Oh btw the set_system_tags we talked about previously to auto-archive by adding the archived tag is throwing me an error.

for t in top_exp: t.set_tags(t.get_tags() + ["trial"]) t.set_system_tags(t.get_system_tags() + ["archived"])With error: ValueError: Task object can only be updated if created or in_progress

I think its because the _edit call happens after the set_task_property call?

3 years ago
0 Hi Team, Me Again! Im Curious If Someone Can Explain To Me Better How Task And Optimisers Integrate With Each Other. In The Example Hyperparameter Optimisation, There Is Both A Task Initialised With

Fantastic. Essentially the example provide just prints out ids to the log file, and Im trying to play around with better things to do so that the top models and similar are saved out in some way I can access without manually reading a log file. Maybe reporting a scalar thats a string which has the task id for the top model? Unsure the best way, hence why I was trying to access the optimiser itself which would naturally contain that info

3 years ago
0 Hi Team, Me Again! Im Curious If Someone Can Explain To Me Better How Task And Optimisers Integrate With Each Other. In The Example Hyperparameter Optimisation, There Is Both A Task Initialised With

Thanks Martin, this is super useful. Using the get_top_experiments would be great, but do I actually have access to the controller (an_optimizer) from the Task object itself? I dont see anything like an_optimizer = task.connect(an_optimizer) which seems to be the normal way of connecthing things up?

3 years ago
0 Hi Team, Me Again! Im Curious If Someone Can Explain To Me Better How Task And Optimisers Integrate With Each Other. In The Example Hyperparameter Optimisation, There Is Both A Task Initialised With

Hey AgitatedDove14 , another question if I can! Im trying to access this information from the API so I can put it as an artifact as well. Currently this is quite a few lines of code using get_top_experiments and get_last_scalar_metrics()[โ€œevaluateโ€][โ€œmaeโ€][โ€œlastโ€], again I feel like Im missing something as I assume theres a far simpler way of getting data displayed so easily in the UI ๐Ÿ™‚

3 years ago
0 Hi Team, Me Again! Im Curious If Someone Can Explain To Me Better How Task And Optimisers Integrate With Each Other. In The Example Hyperparameter Optimisation, There Is Both A Task Initialised With

Awesome. If I end up convincing the team to use ClearML, Ill probably have a ton of small requests to streamline the automation side of orchestration, is the team open to PRs from external people?

3 years ago
0 Hi Team, Me Again! Im Curious If Someone Can Explain To Me Better How Task And Optimisers Integrate With Each Other. In The Example Hyperparameter Optimisation, There Is Both A Task Initialised With

But ideally yes, the HPO should have a df artifact summarising the HPO itself so I can try and make use of the information properly ๐Ÿ™‚

3 years ago
0 Hi Team, Me Again! Im Curious If Someone Can Explain To Me Better How Task And Optimisers Integrate With Each Other. In The Example Hyperparameter Optimisation, There Is Both A Task Initialised With

Ah I see, it does print out the top experiments, you jus thave to make sure the metric and what not agrees. If I was looking to just attach some basic information to the task (after its been rerun, instead of printing it to the log), would the best option be to use the Logger to try and attach it, or set parameters, set comment, or is there a general way to set some metadata that is intended to be used in that capacity.

3 years ago
0 Hi Team, Me Again! Im Curious If Someone Can Explain To Me Better How Task And Optimisers Integrate With Each Other. In The Example Hyperparameter Optimisation, There Is Both A Task Initialised With

Im trying to do it at the end of the optimisation. The same place in the example where you print the ids to the log. Im just hoping theres a way to get said table simply rather than going through a bunch of api calls to construct it myself

3 years ago
0 Hi Team, Me Again! Im Curious If Someone Can Explain To Me Better How Task And Optimisers Integrate With Each Other. In The Example Hyperparameter Optimisation, There Is Both A Task Initialised With

Yup, thats how Ive been doing it now. Will happily update to a simpler method call whenever one gets made. Trying to make use of the HPO is a big thing Im trying to sell the team on, as its what sets ClearML apart from MLFlow or neptune - useful task orchestration and cloning ๐Ÿ™‚

3 years ago
Show more results compactanswers