
Reputation
Badges 1
212 × Eureka!If you look lower, it is there '/home/npuser/.clearml/venvs-builds/3.7/task_repository/commons-imagery-models-py'
For instance, if I wanted the default queue and gpu queue that I create, how do I do that?
Can you fix this or should I open a PR. I'm blocked by this.
It seems like the clearml python sdk might have issues when a subprocess is opened?
{"asctime": "2022-09-28 18:45:55,353", "levelname": "INFO", "name": "root", "module": "ldc_train_end_to_end", "threadName": "MainThread", "message": "Training classifier with command:\npython -m sfi.imagery.models.bbox_predictorv2.train ./sfi/imagery/models/training/train_config.json", "filename": "ldc_train_end_to_end.py", "funcName": "train_model"} File "/usr/lib64/python3.7/site.py", line 177 file=sys.stderr) ^ SyntaxError: invalid syntax
"additionalInfo": { "inBytes": "438", "localPort": "9134", "outBytes": "401", "unusual": "80", "value": "{\"inBytes\":\"438\",\"localPort\":\"9134\",\"outBytes\":\"401\",\"unusual\":\"80\"}", "type": "default" },
Could I simply just reference the files by name and pass in a string such as ~/.clearml/my_file.json
Also I'd like to create the queues pragmatically, is that possible?
Yes I will try that
You could change infrastructure or hosting, and now your data is associated with the wrong URL
How would I do similar with a new queue
The task pod (experiment) started reaching out to an IP associated with malicious activity. The IP was associated with 1000+ domain names. The activity was identified in AWS guard duty with a high severity level.
So this is an additional config file with enterprise? Is this new config file deployable via helm charts?
Also, how do I associate that new queue with a worker?
Just curious, if https://github.com/allegroai/clearml-helm-charts/blob/19a6785a03b780c2d22da1e79bcd69ac9ffcd839/charts/clearml-agent/values.yaml#L50 is a value I can set, where is it used? It would be great if it overrides the Dataset.get embedded url parsed from my clearml conf file
that is the containerinit logs from k8glueagent
I'm getting import errors when I do, from project.utils import tool
all from the working directory of project
When I exec into the pod, it says I need sudo, but wondering if extra_docker_shell_script
is executed as sudo already?
I don't know how to get past this? My k8 pods shouldn't need to reach out to the public file server URL.
So I'd create the queue in the UI, then update the helm yaml as above, and install? How would I add a 3rd queue?
I got the EFS volume mounted. Curious what advantage it would be to use the StorageManager
It seems like https://github.com/allegroai/clearml-helm-charts/blob/main/charts/clearml-agent/values.yaml#L72-L80 doesn't actually do anything as the values set here aren't applied in the agent template
Yes, as an example: My task starts up and checks the mounted EFS volume for x data, if x data does not exist there, it then pulls x data from S3.