Reputation
Badges 1
212 × Eureka!That is the problem, the if
condition is not evaluating to True
Also how do I provide the k8 glue agent permissions to spin up/down ec2 nodes?
For instance, quotes are used
Are there any work arounds to this issue? Our team is evaluating this product to potentially buy enterprise license. If we can't fetch data this is a problem.
For example, in my agent helm yaml, I have
` queue: default
podTemplate:
nodeSelector:
purpose: gpu-nvidia-t4-c8-m32-g1-od `
When I deployed the webserver, I changed the value https://github.com/allegroai/clearml-helm-charts/blob/main/charts/clearml/values.yaml#L36 to be the public file server URL. Then in the UI, I copied the blob from the settings/API keys. Which had the public URLs. After that I did my data uploads which worked fine as they used public URLs. The problem is due to tight security on this k8 cluster, the k8 pod cannot reach the public file server url which is associated with the dataset.
SuccessfulKoala55 Darn, so I can only scale vertically?
It will then parse the above information from my local workstation?
These are the logs from the fileserver pod
ok yes, this is the problem
I think if I use the local service URL this problem is fixed
basically, can I do local installs vs supplying a requirements.txt
Also what is the purpose of the aws
block in the clearml.conf? Where are those values used?
curl --insecure -sw %{http_code}
` -o /dev/null │
│ init-k8s-glue waiting for apiserver ...
yes, I see in the UI how to create a new queue. How do I associate that queue with a nodeSelector though?
Does the file on the EFS get downloaded to the k8 pod local volume?
SuccessfulKoala55 It looks like it should eval to True?
and removed the duplicate Task.init()
I'm getting import errors when I do, from project.utils import tool
all from the working directory of project
Then it tries to curl the files API and gets a 405