Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
AgitatedDove14
Moderator
49 Questions, 8122 Answers
  Active since 10 January 2023
  Last activity one year ago

Reputation

0

Badges 1

25 × Eureka!
0 Clearml Pipelines Can Be Build From Tasks, Functions, And Decorated Functions, According To The Examples In

I am writing quite a bit of documentation on the topic of pipelines. I am happy to share the article here, once my questions are answered and we can make a pull request for the official documentation out of it.

Amazing please share once done, I will make sure we merge it into the docs!

Does this mean that within component or add_function_step I cannot use any code of my current directories code base, only code from external packages that are imported - unless I add my code with ...

2 years ago
0 How Can I Run A New Version Of A Pipeline, Wait For It To Finish And Then Check Its Completion/Failure Status? I Want To Kick Off The Pipeline And Then Check Completion
  • At its simplest, this could just mean checking that all of the steps and the pipeline itself have completed successfully (by checking their โ€œTask statusโ€).If a pipeline step ends with "failed" status in the pipeline execution function an exception will be raised, if the exception is not caught, the pipeline itself will also fail

run

pipeline_script.py

which contains the pipeline code as decorators.

So in theory the following should actually work.
Let's assume you ...

2 years ago
0 Hey Guys Trying To Save A Model Via The Outputmodel.Update_Weights Function I Get The Following Error:

of that makes sense, basically here is what you should do:

Task.init(... output_uri='
')
output_model.update_weights(register_uri=model_path)

It will automatically create a unique target folder / file under None to store your model
(btw: passing the register_uri basically sais: "I already uploaded the model there, just store the link" - i.e. does Not upload the model)

2 years ago
0 Hey Guys Trying To Save A Model Via The Outputmodel.Update_Weights Function I Get The Following Error:

no, i just commented it and it worked fine

Yeah, we should add a comment saying "optional" because it looks as if you need to have it there if you are using Azure.

2 years ago
0 Hey Guys Trying To Save A Model Via The Outputmodel.Update_Weights Function I Get The Following Error:

ok so i accidentally (probably with luck) noticed the max_connection: 2 in the azure.storage config.

NICE!!!! ๐ŸŽŠ
But wait where is that set?
None
Should we change the default or add a comment ?

2 years ago
0 Hi We Just Got The Aws Autoscaler To Create A New Instance When You Enqueue A Task To The Relevant Queue. However, For Some Reason The Task Itself Is Never Run, It Stays In The Pending State. When Looking At The Worker Details, It Says "No Queues Curren

When looking at the worker details, it says "No queues currently assigned to this worker"

Yes, I think we should have better information there, the "AWS service" is not directly pulling jobs from any specific queue, hence nothing there. It is "listening" to queues and launching machines, those machines will be listening to the queue. I wonder if it is just easier to also make sure it is listed as "assigned" to those queues . wdyt?

2 years ago
0 Hi We Just Got The Aws Autoscaler To Create A New Instance When You Enqueue A Task To The Relevant Queue. However, For Some Reason The Task Itself Is Never Run, It Stays In The Pending State. When Looking At The Worker Details, It Says "No Queues Curren

Hi @<1551376687504035840:profile|StraightSealion9>

AWS Autoscaler to create a new instance when you enqueue a task to the relevant queue.

Does that mean that you were able to enqueue a Task and have it launch on the remote EC2 machine ?

2 years ago
0 Hi, How Would I Retrieve The Running Time (Using Python Sdk) For Any Clearml Task? Thank You.

Hi @<1547028052028952576:profile|ExuberantBat52>

task = Task.get_task(...)
print(task.data)

wdyt?

2 years ago
0 Hey, Do Hyperdatasets Offer The Same Features With Tabular Data? Almost All Examples On The Docs Are On Image Datasets

basically @<1554638166823014400:profile|ExuberantBat24> you can think of hyper-datasets as a "feature-store for unstructured data"

2 years ago
0 Hello, I Have Two Experiments Having The Same Plot With The Same X Values. When I Compare These Two Experiments, The Plots Are Drawn Next To Each Other (See Figure), But I Would Appreciate To See The Y-Values Of The Experiments Just In One Plot. The Plot

where the ui merges the plots just as we want and I was wondering if there is some simple way to do it in the case of all plots.

we can do it for scalars (this is trivial)
We can merge specific plots when they are simple, I think basic histograms.
But for any generic plots we fear the merge will just fail, and this is why it defaults to side by side.

how can I combine two plots in the ui as you mentioned?

The easiest solution is to use, "report_scatter2d", these are specific pl...

3 years ago
0 Hi All

Hi @<1546303293918023680:profile|MiniatureRobin9> could it be the pipeline logic is created via the clrarml-task CLI? If this is the case, I think this is an edge case we should fix. Basically it creates a Task instead of pipeline, which in.essence only effects the UI. To solve it, just run the pipeline locally, notice that by default when you start it, it will actually stop the local run and relaunch itself on an agent.
Also, could you open a GitHub issue so we add a flag for it?

2 years ago
0 Thank You For Your Help So Far. I Have A Question About Trains Authentication And Privacy When Deploying On K8S. I Want Integrate Building A Trains-Server Into Our Iac. Now That I Got A Server To Work With An Agent Deployment Im Thinking About Authorizati

Hi ColossalAnt7
Following on SuccessfulKoala55 answer

I saw that there is a config file where you can specify specific users and passwords, but it currently requires

  1. mount the configuration file (the one holding the user/pass) into the pod from a persistent volume .

I think the k8s way to do this would be to use mounted config maps and secrets.

You can use ConfigMaps to make sure the routing is always correct, then add a load-balancer (a.k.a a fixed IP) for the users a...

4 years ago
0 Hi, Is There A Way To Move Existing Pipelines Between Projects? Use Case: I Want To Utilize The New Pipeline Grouping Option And View Them By Project Name. Thing Is, I Accidently Named A Few Of Them Wrongly And It Messes Up The Grouping. Of Course I Can

Is there a way to move existing pipelines between projects?

You should be able to, go to your settings page and turn on "show hidden folders"
Then go to your project, you should see " .pipeline " sub project there, right click it and move it to another folder.

2 years ago
0 Hi All

Thank you! ๐Ÿคฉ

2 years ago
0 If I Create A Task Using Task.Create And Then In A Separate Piece Of Code I Want To Report To It (By Using

I don't know whether you have access to the backend,

Creepy , no I do not ๐Ÿ™‚

I can't make anything appear in the console part of the ui

clearml_task.logger.report_text("some text") should work

3 years ago
0 Hi! Is It Possible To Change A Port On Which Clearml Server Is Launched?

Hi @<1544853721739956224:profile|QuizzicalFox36>
Sure just change the ports on the docker compose

2 years ago
0 I Am Trying To Use

yes its the JWT issue

4 years ago
0 Hi All! Does Anyone Know A Solution To My Issue With Deploying Models Saved On Azure On The Clearml-Serving Docker Container?

how specifically I map my clearml.conf to the containers?

You add:

    volumes:
      - $HOME/clearml.conf:/root/clearml.conf

to here:
None
and here:
None

2 years ago
0 Hi There, Any Plan/Benefit To Support Virtualenv= 20 ?

JitteryCoyote63 virtualenv v20 is supported, pip v21 needs the latest trains/trains-agent RC,

5 years ago
0 What Would Be The Best Way To Approach This Flow?

Because by definition the Task already exists

2 years ago
0 Hi Community, Using Clearml And Loving It So Far! I Had A Question Around The Models That Clearml Automatically Picks Up From The Frameworks And Saves Them. Is There A Way That I Can Control Which Models Actually Get Saved Though? I Would Like To Save Onl

Hi @<1523701132025663488:profile|SlimyElephant79>

I would like to save only the last & best checkpoints and not all of them if possible.

Basically it will mimic the local file system, so if you overwrite the local files it will overwrite the remote model.
You can also disable auto logging, and manually upload the models
In Task.init pass auto_connect_frameworks False for the specific framework
see:
[None](https://clear.ml/docs/latest/docs/clearml_sdk/task_sdk/#automatic-lo...

2 years ago
0 [Task Gets Interrupted / Aborted / Reset When In Offline Mode] For Local Testing, We Have Added A

BTW, this one seems to work ....
` from time import sleep
from clearml import Task

Task.set_offline(True)
task = Task.init(project_name="debug", task_name="offline test")

print("starting")

for i in range(300):
print(f"{i}")
sleep(1)

print("done") `

2 years ago
0 How Do I Create Sub Projects With The New Version 1.0?

does this work for multiple levels?

Yep ๐Ÿ˜„

4 years ago
0 Another Question: Is It Possible To Read The Dependencies Manually From A Conda Environment.Yml? It Seems Like Clearml Is Not Able To Fetch The Dependencies Correctly When

ReassuredTiger98 oh wow I did not realize you actually call importlib to import your libraries (any reason not to call import ?)
And yes, I think we will miss it as the package analysis is actually static text analysts of the code

4 years ago
0 How Can I Ensure Tasks In A Pipeline Have The Same Environment As The Pipeline Itself? It Seems A Bit Counter-Intuitive That The Pipeline (Executed Remotely) Captures The Local Environment, But The Tasks (Executed Remotely) Do Not Use That Same Environmen

If you use this one for example, will the component have pandas as part of the requirement
None

def step_two(...):
    import pandas as pd
    # do stuff

If so (and it should), what's the difference, where is "internal.repo " different from pandas ?

2 years ago
Show more results compactanswers