i had a misconception that the conf comes from the machine triggering the pipeline
Sorry, this one :)
Yes... I think that this might be a bit much automagic even for clearml đ
Yes it worked đ
I loaded my entire clearml.conf in the âextra confâ part of the auto scaler, that worked
PricklyRaven28 at the beginning of the log, the clearml agent should print the configuration, do you have api.fileserver as the S3 bucket?
thatâs what i started with, doesnât work in pipelines
CostlyOstrich36 This is for a step in the pipeline
In the UI check under the execution tab in the experiment view then scroll to the bottom - You will have a field called "OUTPUT" what is in there? Select an experiment that is giving you trouble?
tried your suggestion, still got to file serverâŚ
PricklyRaven28 did you set the iam role support in the conf?
https://github.com/allegroai/clearml/blob/0397f2b41e41325db2a191070e01b218251bc8b2/docs/clearml.conf#L86
Try with sdk.development.default_output_uri
as well
that does happen when you create a normal local task, that's why i was confused
The parts that are not passed in both cases are the configurations from the conf file. Only the environment is passed (e.g. git python packages etc) , . For example if you have storage credentials in your conf file , they are not passed to a remote agent, instead the credentials from the remote agent are used when it runs the task.
make sense?
but it makes sense, because the agent in that case is local
that does happen when you create a normal local task, thatâs why i was confused
Is the entire pipeline running on the autoscaler?
i had a misconception that the conf comes from the machine triggering the pipeline
AgitatedDove14 So it looks like it started to do something, but now itâs missing parts of the configuration
Missing key and secret for S3 storage access
(iâm using boto credential chain, which is off by defaultâŚ)
why isnât the config being passed to the inner step properly ?
If possible, i would like all together prevent the fileserver and write everything to S3 (without needing every user to change their config)
There is no current way to "globally" change the default files server (I think this is part of the enterprise version, alongside vault etc.).
What you can do is use an OS environment to override the conf file:CLEARML_FILES_HOST="
"
PricklyRaven28 wdyt?
regarding what AgitatedDove14 suggested, iâll try tomorrow and update
when i did this with a normal task it worked wonderfully, with pipeline it didnât
using api.files_server? not default_output ?
and the agent is outputting sdk.development.default_output_uri =
although itâs different in both the original config, and the agent extra config