Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
AgitatedDove14
Moderator
49 Questions, 8126 Answers
  Active since 10 January 2023
  Last activity one year ago

Reputation

0

Badges 1

25 × Eureka!
0 Votes
0 Answers
2K Views
0 Votes 0 Answers 2K Views
3 years ago
0 Votes
0 Answers
2K Views
0 Votes 0 Answers 2K Views
Is it a one time thing? or recurring?
5 years ago
0 Votes
0 Answers
2K Views
0 Votes 0 Answers 2K Views
https://m.facebook.com/story.php?story_fbid=2484620658505570&id=1620822758218702&refid=52&tn=-R
5 years ago
0 Votes
0 Answers
2K Views
0 Votes 0 Answers 2K Views
https://allegro.ai/docs
5 years ago
0 Votes
3 Answers
2K Views
0 Votes 3 Answers 2K Views
Hi
Hi , v0.15 is out, πŸŽ‰ πŸš€ Your feedback had a major influence on the features we added πŸ™‚ thank you! A selected list of features: Column resizing / ordering /...
5 years ago
0 Votes
0 Answers
2K Views
0 Votes 0 Answers 2K Views
5 years ago
0 Votes
3 Answers
2K Views
0 Votes 3 Answers 2K Views
This will close it Task.current_task().close()I think we should rename completed() because it just marks the Task as completed on the backend but does not ac...
4 years ago
0 Votes
0 Answers
2K Views
0 Votes 0 Answers 2K Views
5 years ago
0 Votes
0 Answers
2K Views
0 Votes 0 Answers 2K Views
New releases: pip install trains==0.13.3https://github.com/allegroai/trains/releases/tag/0.13.3 pip install trains-agent==0.13.2https://github.com/allegroai/...
5 years ago
0 Votes
0 Answers
2K Views
0 Votes 0 Answers 2K Views
Finally
5 years ago
0 Votes
6 Answers
2K Views
0 Votes 6 Answers 2K Views
Hi
Hi ! ClearML Server + SDK v1.9.0 is out! πŸŽ‰ πŸš€ 🎊 Happy Holidays and Happy New Year! ❇️ πŸŽ‡ πŸŽ„
2 years ago
0 Votes
7 Answers
1K Views
0 Votes 7 Answers 1K Views
Thank you all for taking the time to answer our survey (If you haven't already, we urge you to do so ). Your feedback has a major impact on what we build, do...
5 years ago
0 Votes
0 Answers
2K Views
0 Votes 0 Answers 2K Views
4 years ago
0 Votes
0 Answers
2K Views
0 Votes 0 Answers 2K Views
Gals, Guys & :robot_face: If you want to get some inspiration on building DL Continuous Integration pipelines, I suggest this post (obviously built on top of...
5 years ago
0 Votes
1 Answers
2K Views
0 Votes 1 Answers 2K Views
This is usually due to enterprise level issued https certificates not part of the local installation (basically any python generated SSL request will fail)
5 years ago
0 Votes
0 Answers
2K Views
0 Votes 0 Answers 2K Views
Hi Guys/Gals, If you want to checkout the latest RC we have 0.15.0rc0 out : pip install trains==0.15.0rc0 pip install trains-agent==0.15.0rc0Many of the impr...
5 years ago
0 Votes
0 Answers
2K Views
0 Votes 0 Answers 2K Views
🎊 🍾 Happy new year ! πŸŽ† πŸŽ‡ We wanted to thank you all for the great feedback, contribution and general support you guys give us. It is truly fulfilling to ...
4 years ago
0 Votes
0 Answers
2K Views
0 Votes 0 Answers 2K Views
Lol, I wonder what the adblock rule was ;)
5 years ago
0 Votes
0 Answers
2K Views
0 Votes 0 Answers 2K Views
New RC for trains-agent is out pip install trains-agent==0.13.2rc1
5 years ago
Show more results questions
0 Hey All. Quick Question About The

TenseOstrich47 every agent instance has its own venv copy. Obviously every new experiment will remove the old venv and create a new one. Make sense?

4 years ago
0 Hi, I Am Trying To Upload A Plot To An Existing Task Using The

What's the matplotlib version ? and python version?

4 years ago
4 years ago
0 Hi, I'M Trying To Clone And Queue Experiments For Running Them On My Workers. I Am Able To Successfully Clone And Queue The Task, But Seems Like The Task Does Not Pass The Correct Parameters To My Python Script On The Worker. We Use Hydra For Configuring

JumpyPig73 Do you see all the configurations under the Args section in the "Configuration" Tab ?
(Maybe I'm wrong and the latest RC does Not include the python-fire support)

3 years ago
0 Hi, The Following Does Not Seem To Work

SmarmySeaurchin8 regrading (2)
I'm not sure the current visualization supports it. I mean we can put "{}", but that would imply you can edit it, which then we have to support, possible but weird, and this is why:
task.connect({'a':{},'b': {'nested': 'value}}will become
'a' = '{}'
'b/nested' = 'value'
But then if you edit to:
'a' = '{'nested': 'value'}'
'b/nested' = 'value'
you have two different ways of presenting the same type of structure...

4 years ago
0 Is There An Easy Way To Add A Link To One Of The Tasks Panels? (As An Artifact, Configuration, Info, Etc)? Edit: And Follow Up Regarding The Dataset. As Discussed Somewhere Previously, The Datasets Are Now Automatically Moved To A Hidden "Sub-Project" Pr

Why is it using an OutputModel and an InputModel?

So calling OutputModel will create the new Model entity and upload the data, InputModel will store it as required input Model.
Basically on the Task you have input & output section, when you clone the Task you are copying the input section into the newly created Task, and the assumption is that when you execute it, your code will create the output section.
Here when you clone the Task you will be clone the reference to the InputModel (i...

3 years ago
0 Hello! Does Anyone Know How To Do

try Hydra/trainer.params.batch_size
hydra separates nesting with "."

2 years ago
4 years ago
0 Quick Question.. Is Redis Used As Permanent Data Storage Or Just Cache? Would There Be Any Problems If It Is Restarted And Comes Up Clean?

Hi DisgustedDove53

Is redis used as permanent data storage or just cache?

Mostly cache (Ithink)

Would there be any problems if it is restarted and comes up clean?

Pretty sure it should be fine, why do you ask ?

4 years ago
0 Hey! I Would Like To Connect To Same Task From Multiple Consumer And Upload Debug Image. Is It Possibile? It Seems Like I Can Connect To The Task. Get The Logger But Nothing Is Uploaded.

Logger.current_logger()Will return the logger for the "main" Task.
The "Main" task is the task of this process, a singleton for the process.
All other instances create Task object. you can have multiple Task objects and log different things to them, but you can only have a single "main" Task (the one created with Task.init).
All the auto-magic stuff is logged automatically to the "main" task.
Make sense ?

5 years ago
4 years ago
0 Hello. It'D Be Really Helpful If Someone Could Let Me Know Why I Keep Getting "Misconfigurationexception('No Supported Gpu Backend Found!')" Error. I Am Using "Task.Execute_Remotely(Queue_Name="Default", Exit_Process=True)". Once It Gets Queued, I Clone I

Hi @<1715175986749771776:profile|FuzzySeaanemone21>

and then run "clearml-agent daemon --gpus 0 --queue gcp-l4" to start the worker.

I'm assuming the docker service cannot spin a container with GPU access, usually this means you are missing the nvidia docker runtime component

one year ago
0 Hello There! I Was Trying To Update The Url For Debug Samples After Migration Of The Server To A New Domain And Was Following The Steps From Here:

Hi @<1684010629741940736:profile|NonsensicalSparrow35>
So sorry I missed this thread πŸ™
Basically your issue is the load balancer that prevents the post command, you can change that, just add to any clearml.conf the following line:

api.http.default_method: "put"
one year ago
0 Hi Guys, With The New Venv Caching Available In Clearml, I Have The Following Problem: I Force My Pip Requirements To Be:

Since my deps are listed in the dependencies of my setup.py, I don't want clearml to list the dependencies of the current environment

Make sense πŸ™‚
Okay let me check regrading the "." in the venv cache.

4 years ago
0 Does Clearml Have The Ability To Run A Single Experiment Across Multiple Nodes/Gpus In A K8 Cluster?

Actually this is by default for any multi node training framework torch DDP / openmpi etc.

3 years ago
0 Hi Guys. Say That We Train A Model With 10 Epoch, And Suddenly Interruption Occur On Epoch 5. How Can We Continue The By Using Clearml?

Hi @<1546665666675740672:profile|AttractiveFrog67>

  • Make sure you stored the model's checkpoint (either pass output_uri=True in Task.init or manually upload)
  • When you call Task.init pass " continue_last_task=True "
  • Now you can do last_checkpoint=task.models["output"][-1].get_local_copy() and all you need is to load last_checkpoint
2 years ago
0 When I Pass Invalid Key To

it fails but with COMPLETED status

Which Task is marked "completed" the pipeline Task or the Step ?

4 years ago
0 If I Have A Dataset And I Process It And I Want The Processed Data As Another Dataset, Is Parent The Right Approach?

LOL AlertBlackbird30 had a PR and pulled it πŸ™‚
Major release due next week after that we will put a a roadmap on the main GitHub page.
Anything specific you have in mind ?

4 years ago
0 Hi! Trying To Run The Following Very Basic Code. The First Few Parts Works As They Should:

2021-07-11 19:17:32,822 - clearml.Task - INFO - Waiting to finish uploads

I'm assuming a very large uncommitted changes πŸ™‚

4 years ago
0 Hi, I Am Creating Pipeline From Function With Dynamically Created Steps, Eg. If I Pass Pipeline Param Tune_Optime='Recall,Precision', My Pipeline Is Creating 2 Tasks/Steps - Each For Trained Model. Everything Is Working Really Nice, When I Start Pipeline

Ad1. yes, think this is kind of bug. Using _task to get pipeline input values is a little bit ugly

Good point, let;s fix it πŸ™‚

new pipeline is built from scratch (all steps etc), but by clicking "NEW RUN" in GUI it just reuse existing pipeline. Is it correct?

Oh I think I understand what happens, the way the pipeline logic is built, is that the "DAG" is created the first time the code runs, then when you re-run the pipeline step it serializes the DAG from the Task/backend.
Th...

3 years ago
0 Hi Guys, Following Up On This

Hi JitteryCoyote63
The new pipeline is almost ready for release (0.16.2),
It actually contains this exact scenario support.
Check out the example, and let me know if it fits what you are looking for:
https://github.com/allegroai/trains/blob/master/examples/pipeline/pipeline_controller.py

5 years ago
0 Another Question, I Have Written A Code That Includes A Task Scheduler That Calls A Function. That Function Watches A Folder And If There Are Sufficient Images, It Creates And Publishes The Dataset, After Which It Clears The Folder. Problem, For Some Rea

VexedCat68

a Dataset is published, that activates a Dataset trigger. So if every day I publish one dataset, I activate a Dataset Trigger that day once it's published.

From this description it sounds like you created a trigger cycle, am I missing something ?
Basically you can break the cycle by saying, trigger only on New Dataset with a specific Tag (or create the auto dataset in a different project/sub-project).
This will stop your automatic dataset creation from triggering the "orig...

3 years ago
0 I Use

Hi SteadyFox10

I'll use your version instead and put any comment if I find something.

Feel free to join the discussion πŸ™‚ https://github.com/pytorch/ignite/issues/892

Thansk for theΒ 

ouput_uri

Β can I put in theΒ 

~/trains.conf

Β file ?

Sure you can πŸ™‚
https://github.com/allegroai/trains/blob/master/docs/trains.conf#L152
You can add it in the trains-agent machine's conf file, or/and on your development machine. Notice that once you run ...

5 years ago
Show more results compactanswers