Hi @<1534496192929468416:profile|EagerGiraffe33> , what if you try to put a specific version of pytorch you've tested on your remote environment in the requirements section of the cloned task?
Also, is there a reason you don't want to work with the default ports provided in the docker-compose.yml ?
@<1547028079333871616:profile|IdealElephant83> , what are you trying to do during the code execution?
Is it a self hosted server?
Hi @<1556450111259676672:profile|PlainSeaurchin97> , I think what you're looking for is the output_uri parameter in Task.init()
Did you try the workaround provided in https://clearml.slack.com/archives/CTK20V944/p1664887550256279 by AgitatedDove14 ?
EcstaticGoat95 , thanks a lot! Will take a look 🙂
JitteryCoyote63 , heya, yes it is :)
You can save the entire folder as an artifact.
Hi @<1523707996645888000:profile|GrievingTurkey78> , did you configure the agent to install requirements from requirements.txt ? You can get more information using the --debug flag when running clearml-agent
Hi SuperiorCockroach75 , can you please add the full log + how to reproduce this?
Regarding the queue error - you need to create a queue called 'CPU Queue'
I have no idea, but considering that the version for http://app.clear.ml was updated recently (last week from what I noticed) I'd be guessing that the self hosted server should be right around the corner 😉
I'll clarify - on the server you have two parts - the clearml folders where all the mongo/elastic/redis data sits and you have the dockers. So, downgrading would mean using previous version dockers. However if you don't have a backup of your data I don't suggest you do this since data might become corrupt (mismatching elastic versions is bad for elastic)
What are the packages that are detected when you run the experiment locally?
Dataset.get only fetches the dataset object, it doesn't try accessing files yet. What else are you doing in your code that reproduces your issue?
@<1664079296102141952:profile|DangerousStarfish38> , can you provide logs please?
What is the combination of --storage and configuration that worked in the end?
The problem was that the plot I created myself
How was the plot created? Can you give me a small snippet to try and play around with?
@<1787653555927126016:profile|SoggyDuck67> , can you try setting the binary to 3.11 instead of 3.10?
Hi DrabCockroach54 , I think you are looking for these endpoints:
https://clear.ml/docs/latest/docs/references/api/workers#post-workersget_activity_report
https://clear.ml/docs/latest/docs/references/api/workers#post-workersget_stats
Regarding the api token - you can use the key/secret pair that you generated through the ClearML UI
CluelessElephant89 , I've added screenshots. Tell me if those help 🙂
Meaning that you should configure your host as follows host: "somehost.com:9000"
Hi @<1576381444509405184:profile|ManiacalLizard2> , please note the failure error:docker: Error response from daemon: pull access denied for new_docker, repository does not exist or may require 'docker login': denied: requested access to the resource is denied.
I think you need to first login to the repository
Hi @<1572032849320611840:profile|HurtRaccoon43> are you referring to Dima's request?
As you mentioned it seems that the authentication you're using is not created as part of the boto object. Is there a specific reason you don't want to use access/secret pairs for authentication?
Hi @<1719524641879363584:profile|ThankfulClams64> , what do you mean regarding ClearML GPU Compute? Do you mean the Genesis autoscaler?
Hi ScrawnyLion96 ,
I think it handles some data like worker stats. It's required for the server to run. What do you mean by the redis getting fuller and fuller?
Hi @<1797438038670839808:profile|PanickyDolphin50> , can you please elaborate? What is this accelerate functionality?
I'm not personally familiar, but I'm sure searching for docker images with the python version you need will yield the required results 🙂
JitteryCoyote63 , you can also double click an experiment to get the context menu 🙂