Reputation
Badges 1
25 × Eureka!UI for some anomalous file,
Notice the metrics are not files/artifacts, just scalars/plots/console
Another question, do you have the argparse with type=str ?
Right, if this is the case, then just use 'title/name 001' it should be enough (I think this is how TB separates title/series or metric/variant )
Are you trying to upload an artifact post execution ?
ThickDove42 you need the latest cleaml-agent RC for the docker setup script (next version due next week)pip install clearml-agent==0.17.3rc0
Okay this is very close to what the agent is building:
Could you start a new conda env,
then install cudatoolkit=11.1
then run:
conda env update -p <conda_env_path_here> --file the_env_yaml.yml
Okay that makes sense.best_diabetes_detection is different from your example curl -X POST " None " notice best_mage_diabetes_detection` ?
So that means your home folder is always mapped to ~/ on any machine you ssh to ?
This is definitely a but, in the super class it should have the same condition (the issue is checking if you are trying to change the "main" task)
Thanks ApprehensiveFox95
I'll make sure we push a fix π
When you have a bit of experience, please suggest a path forward, it will be great to integrate
I'm really for adding an interface, but I was not able to locate a simple integration option with basically anything, Wdyt ?
so that you can get the latest artifacts of that experiment
what do you mean by " the latest artifacts "? do you have multiple artifacts on the same Task or s it the latest Task holding a specific artifact?
Weβd be using https in production
Nice π
@<1687653458951278592:profile|StrangeStork48> , I was reading this thread trying to understand what exactly is the security concern/fear here, and I'm not sure I fully understand. Any chance you can elaborate ?
We actually added a specific call to stop the local execution and continue remotely , see it here: https://github.com/allegroai/trains/blob/master/trains/task.py#L2409
Hi SweetGiraffe8
could you try with the latest RCpip install 0.17.5rc2
@<1523703080200179712:profile|NastySeahorse61> so glad you managed to solve it π π
have a CI/CD (e.g Github Actions) thats update my βproductionβ pipeline on ClearML UI,
I think this is the easiest way, basically the CI/CD launches a pipeline (which under the hood is another type of Task), by querying the latest "Published" pipeline that is also Not archived, then cloning+pushing it to execution queue.
In the UI when you want to "upgrade" the production pipeline you just right click "Publish" on the pipeline you want to launch. Another way is to do the same with Tags...
If Task.init() is called in an already running task, donβt reset auto_connect_frameworks? (if i am understanding the behaviour right)
Hmm we might need to somehow store the state of it ...
Option to disable these in the clearml.conf
I think this will be to general, as this is code specific , no?
ChubbyLouse32 could it be the configuration file is not passed to the agent machine itself ?
(were you able to run anything against this internal server? I mean to connect to it from code, clearml/cleamrl-agent) ?
I'm kind of at a point where I don't know a lot of what to even search for.
we feel you π , yes there still isn't a very good source of information on where to get started...
This is because the entire field is constantly changing and evolving, and one solution will usually only apply to specific use case...
I would start with the mlops community slack channel, and youtube talks (specifically those by companies describe how they built their own internal infrastructure, i...
NastySeahorse61 it might that the frequency it tests the metric storage is only once a day (or maybe half a day), let me see if I can ask around
(just making sure you can still login to the platform?)
Hmm I think everything is generated inside the c++ library code, and python is just an external interface. That means there is no was to collect the metrics as they are created (i.e. inside the c++ code), which means the only was to collect them is to actively analyze/read the tfrecord created by catboost π
Is there a python code that does that (reads the tfrecords it creates) ?
OddAlligator72 FYI you can also import / export an entire Task (basically allowing you to create it from scratch/json, even without calling Task.create)Task.import_task(...) Task.export_task(...)
I might have found it, tqdm is sending{ 1b 5b 41 } unicode arrow up?https://github.com/horovod/horovod/issues/2367
I suggest a bump in the GitHub issue