Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
WackyRabbit7
Moderator
73 Questions, 550 Answers
  Active since 10 January 2023
  Last activity one year ago

Reputation

0

Badges 1

533 × Eureka!
0 I'M Running

anyway, my ultimate goal is to create templates for other tasks... Is that possible in any other way through the CLI?

4 years ago
0 Why Would The Mongo4 Migration Scripts (For Clearml-Server 1.2) Try To Chown 1000:1000?

CostlyOstrich36 so why 1000:1000? My user and group are not that and so do all the otehr files I have under /opt/clearml

3 years ago
0 Question About The Auto Scaling Service Under

I doubled checked the credentials in the configurations, and they have full EC2 access

4 years ago
0 Question About The Auto Scaling Service Under

and I'm using the same credentials

4 years ago
0 Question About The Auto Scaling Service Under

Hi FriendlySquid61 I did all the changes you said

4 years ago
0 Moreover, When I Go To The Queue Page, I See The Queue Is Empty, But When I'M On The Queued Task'S Page I Can See It Is Enqueued To Right Right Queue... So The Task Says It Is In The Queue, But The Queue Says It Is Empty

` # define pipeline
pipe = clearml.PipelineController(
name=TASK_NAME,
project=PROJECT_NAME,
version='0.0.1',
add_pipeline_tags=False,
)
pipe.set_default_execution_queue('default')

Adding steps

pipe.add_step(name=f'{start_date_train}_{end_date_train}_choose_best',
base_task_project=CHOOSE_PROJECT_NAME,
base_task_name=CHOOSE_TASK_NAME,
parameter_override=params_override,
...

3 years ago
0 Hi, I Have A Question Regarding Trains And Git. If I Make Changes To A File And Run It Using Trains (Without Pushing The Changes To Github), What Version Of The Script Is Running? It Seems That Every Change I Make Need To Be Added, Committed And Pushed Be

I think you are talking about separate problems - the "WARNING DIFF IS TOO LARGE" is only a UI issue, that you can't see hte diff in the UI - correct me if I'm wrong with this

Maria seems to be saying that the execution FAILS when she has uncomitted changes, which is not the expected behavior - am I right maria?

5 years ago
0 Hi, I Have A Question Regarding Trains And Git. If I Make Changes To A File And Run It Using Trains (Without Pushing The Changes To Github), What Version Of The Script Is Running? It Seems That Every Change I Make Need To Be Added, Committed And Pushed Be

(I'm working with maria)

essentially, what maria says is when she has a script with uncomitted changes, when executing remotely, the script that actually runs on the remote machine is without the uncomitted changes

e.g.:
Her git status is clean, she makes some changes to script.py and executes it remotely. What gets executed remotely is the original script.py and not the modified version she has locally

5 years ago
0 Question About The Usage Of Trains Agents. In Our Company We Have 3 Hpc Servers, Two Of Them Have Multiple Gpus, One Is Cpu Only. I Saw In The Docs The Multiple Agents Can Be Run Separately Assigning Gpus In Whatever Manner You Want. My Questions Are 1

So regarding 1, I'm not really sure what is the difference

When running in docker mode what is different the the regular mode? No where in the instructions is nvidia docker a prerequisite, so how exacly will tasks on GPU get executed?

I feel I don't underatand enough of the mechanism to (1) understand the difference between docker mode and not and (2) what is the use casr for each

5 years ago
0 I Have A Data Scientist Constantly Having The Same Problem. If She Did Not Push His Latest Changes To Git, She Gets The Following Error

What do you mean by submodules?

She did not push, I told her she does not have to push before executing as trains figures out the diffs.
When she pushes - it works

5 years ago
0 In Pipelinev2, Is It Possible To Register Artifacts To The Pipeline Task? I See There Is A Private Variable

and then how would I register the final artifact to the pipelien? AgitatedDove14 ⬆

4 years ago
4 years ago
0 I Have To Say I'M Totally Confused By The Pipeline I Want To Execute The Pipeline On My Local Computer. I Followed

btw my site packages is false - should it be true? You pasted that but I'm not sure what it should be, in the paste is false but you are asking about true

4 years ago
0 I Have To Say I'M Totally Confused By The Pipeline I Want To Execute The Pipeline On My Local Computer. I Followed

it seems that only the packages that are on the script are getting installed

4 years ago
0 Using

I am noticing that the files are saved locally, is there any chance that the files are over-written during the run or get deleted at some point and then replaced?
Yes they are local - I don't think there is a possibility they are getting overwritten... But that depends on how clearml names them. I showed you the code that saves the artifacts, but this code runs multiple times from a given template with different values - essentially it creates like 10 times the same task with different param...

4 years ago
0 Question About The Configuration Format - I'D Like To Parse It Within My Python Code So I'Ll Be Able To Access Things Like

SuccessfulKoala55 this actually doesn't work

` apiserver_conf = ConfigFactory.parse_file(API_SERVER_CONF_PATH)

POINT 1

conf_content = HOCONConverter.to_hocon(config=ConfigFactory.from_dict(apiserver_conf.as_plain_ordered_dict()),
compact=False,
level=0, indent=2)

apiserver_conf['auth']['fixed_users']['users'].append(
ConfigFactory.from_dict({'username': username, 'password': password, 'name': name}))

##...

5 years ago
0 Question About Out-Of-Python Dependencies. For Example Nltk Requires

Maybe even a dedicated argument specifically for apt-get packages, since it is very common to need stuff like that

5 years ago
0 I Have To Say I'M Totally Confused By The Pipeline I Want To Execute The Pipeline On My Local Computer. I Followed

` # Python 3.8.10 (default, Jun 2 2021, 10:49:15) [GCC 9.4.0]
clearml == 1.0.5
hyperopt == 0.2.5
matplotlib == 3.4.3
numpy == 1.21.2
pandas == 1.3.2
plotly == 5.3.0
python_dateutil == 2.8.2
scikit_learn == 0.24.2
statsmodels == 0.12.2
tqdm == 4.62.2

Detailed import analysis

**************************

IMPORT PACKAGE clearml

tasks/data_projection.py: 9

tasks/hp_optimization.py: 6

tasks/hpo_n_best_evaluation.py: 6

tasks/pipelines/monthly_predictions.py: 4

IMPORT PACKAGE hypero...

4 years ago
0 I'M Running

SuccessfulKoala55 seems like you got it spot on, it contains the entire repo, but no .git directory

So what can we do about it? All I want is to create templates for some tasks, so I can later execute them through a Pipelinecontroller

4 years ago
0 Following The Instructions For Local Trains Deployment On Linux (

a machine that had previous installation, but I deleted the /opt/trains directory beforehand

4 years ago
0 I Have A Production Inference Pipeline Which I Want To Continuously Test On My Github To Make Sure It Doesn'T Break As We Move Forward. The Ideal Scenario For Me Is To Use

Gotcha, didn't think of an external server as Service Containers are part of Github's offering, I'll consider that

4 years ago
Show more results compactanswers