Reputation
Badges 1
662 × Eureka!Note that it would succeed if e.g. run with pytest -s
i.e.ERROR Fetching experiments failed. Reason: Backend timeout (600s)ERROR Fetching experiments failed. Reason: Invalid project ID
Much much appreciated ๐
(in the current version, that is, weโd very much like to use them obviously :D)
The new task is not running inside a new subprocess. Our platform trains several models, and we'd like each of them to be tracked in their own Task . When running locally, this is "out of the box", as we can init and close before and after each model.
When running remotely, one cannot close the main task (since it is what orchestrates everything), and so this workaround was needed.
I believe that a Pipeline should have the system tags ( pipeline , maybe hidden ), even if it created in a running Task .
Looks good! Why is it using an OutputModel and an InputModel?
Ultimately we're trying to avoid docker in AWS autoscaler (virtualization on top of virtualization seems redundant), and instead we maintain an AMI for a faster boot sequence.
We had no issues when we used pip , but now when trying to work with poetry all these issues came up.
The way I understand poetry to work, is that it is expected there is one system-wide installation that is used for virtual environment creation and manipulation. So at least it may be desired that the ...
Why not give ClearML read-only access credentials to the repository?
I guess the big question is how can I transfer local environment variables to a new Task
Not necessarily on the same branch, no
Ah! Makes sense. Thanks!
I thought this follows from our previous discussion SuccessfulKoala55 , where this is a built-in feature of pyhocon?
It does not ๐
We started discussing it here - https://clearml.slack.com/archives/CTK20V944/p1640955599257500?thread_ts=1640867211.238900&cid=CTK20V944
You suggested this solution - https://clearml.slack.com/archives/CTK20V944/p1640973263261400?thread_ts=1640867211.238900&cid=CTK20V944
And I eventually found this solution to work - https://clearml.slack.com/archives/CTK20V944/p1641034236266500?thread_ts=1640867211.238900&cid=CTK20V944
Yes; I tried running it both outside venv and inside a venv. No idea why it uses 2.7?
You can use logger.report_scalar and pass a single value.
Happens with the latest version indeed.
I canโt share our code, but the gist of it is:
pipe = PipelineController(name=..., project=..., version=...)
pipe.add_function_step(...) # Many calls
pipe.set_default_execution_queue(...)
pipe.start(queue=..., wait=True)
I can see the task in the UI, it is not archived, and that's pretty much the snippet, but in full I do e.g.
So basically what I'm looking for and what I have now is something like the following:
(Local) I have a well-defined aws_autoscaler.yaml that is used to run the AWS autoscaler. That same autoscaler is also run with CLEARML_CONFIG_FILE=.... (Remotely) The autoscaler launches, listens to the predefined queue, and is able to launch instances as needed. I would run a remote execution task object that's appended to the autoscaler queue. The autoscaler picks it up, launches a new instanc...
Am I making sense ?
No, not really. I don't see how task.connect_configuration interacts with our existing CLI? Additionally, the documentation for task.connect_configuration say the second argument is the name of a file, not the path to it? So something is off
Itโs just that for the packages argument, ClearML says:
If not provided, packages are automatically added based on the imports used inside the wrapped function.
Soโฆ ๐ค
I guess it depends on what you'd like to configure.
Since we let the user choose parents, component name, etc - we cannot use the decorators. We also infer required packages at runtime (the autodetection based on import statements fails with a non-trivial namespace) and need to set that to all components, so the decorators do not work for us.
e.g. a separate structured user guide with common tips, usability, best practices - https://pandas.pydata.org/pandas-docs/stable/user_guide/index.html
vs the doc, where each function is its own page, e.g.
https://pandas.pydata.org/pandas-docs/stable/reference/api/pandas.read_csv.html
I'll have yet another look at both the latest agent RC and at the docker-compose, thanks!
There was no "default" services agent btw, just the queue, I had to launch an agent myself (not sure if it's relevant)