Reputation
Badges 1
8 × Eureka!@<1523701070390366208:profile|CostlyOstrich36> This is an example of timeout I am talking about.
No, there is no issue in connectivity. Logs are fine. You can see in git issue I provided its the same error. I am getting error but it's getting completed as well.
I have mounted my s3 bucket at the location /opt/clearml/data/fileserver/ but I can see my data is not being stored in s3 but its storing in ebs. How so? It should only be stored in s3. Is there mounting issue in my solution?
I can't configure my local machine for IAM Role.
Ok so if we can't send the data to s3 through server, can I mount the s3 bucket as file system on place where dataset is uploaded on server will it solve this problem?
Yes, I resolved it. So basically we need to give docker containers extra privileges. add --privileged for single container and in case of docker-compose.yml and privileged: true.
@<1523701205467926528:profile|AgitatedDove14> I was able to mount but if I am running the experiment and uploading data it is going to s3 as well as ebs. Why so it should only go to s3
Ohh, btw how did you approach the problem of mounting ebs/efs where clearml-data and everything will be stored
Hi @<1594863216222015488:profile|ConvincingGrasshopper20> @<1541954607595393024:profile|BattyCrocodile47> , I am facing the same problem of Elasticsearch container, How did you resolve this. I am mounting EFS instead of EBS.