Reputation
Badges 1
103 × Eureka!SmugDolphin23 Where can I check the lates RC? I was not able to find it in the clearml github repo
we reinstalled the clearml-agent$clearml-agent --version CLEARML-AGENT version 1.2.3
running top | grep clearml
we can see the agent running
running clearml-agent list
we can see 2 workers
before running clearml-agent daemon --stop
We updated the clearml.conf and updated the worker_id
and worker_name
with the relevant name/id that we can see from clearml-agent list
and we get
` Could not find a running clearml-agent instance with worker_name=<clearml_worker_na...
Hi @<1523701205467926528:profile|AgitatedDove14>
I'm having a similar issue.
Also notice the cleaml-agent will not change the entry point of the docker meaning if the entry point does not end with plain bash, it will not actually run anything
Not sure I understand how to run a docker_bash_setup_script
and then run a python script - Do you have an example? I could not find one.
Here is our CLI command
clearml-task --name <TASK NAME> \
--project <PRJ NAME> \
--repo git@gi...
ClearML key/secret provided to the agent
When is this provided? Is this during the build
?
Looking in the repo I was not able to see an example - only reference to https://github.com/allegroai/clearml/blob/b9b0a506f35a414f6a9c2da7748f3ec3445b7d2d/docs/clearml.conf#L13 - I just need to add company.id
or user.id
in the credential dict?
This also may help with the configuration for GCS
https://clearml.slack.com/archives/CTK20V944/p1635957916292500?thread_ts=1635781244.237800&cid=CTK20V944
I saw https://clear.ml/docs/latest/docs/references/sdk/dataset/#verify_dataset_hash - but I don't think it is the correct one. the https://pandas.pydata.org/pandas-docs/stable/reference/api/pandas.DataFrame.shape.html property
Distributor ID: Ubuntu
Description: Ubuntu 20.04.4 LTS
Release: 20.04Codename: focal
Strange
I ranclearml-agent daemon --stop
and after 10 min I ranclearml-agent list
and I still see a worker
Hi,
You may want to consider to do the visualizing while creating the Datasets - see https://github.com/thepycoder/asteroid_example/blob/main/get_data.py#L34 logging the head()
of the dataframe
will do
A work around that worked for me is to explicitly complete the task, seems like the flush
has some bug
task = Task.get_task('...')
task.close()
task.mark_completed()
ds.is_final()
True
upgrading to 1.12.1 didn't help
I think the issue is that when I create the dataset
- i used
use_current_task=True,
If I change it to
use_current_task=False,
then it finalizes
Hi AnxiousSeal95 ,
Is there an estimate when the above feature will be available?
shape -> tuple([int],[int])
I decided to use
._task.upload_artifact(name='metadata', artifact_object=metadata)
where metadata is a dict
metadata = {**metadata, **{"name":f"{Path(file_tmp_path).name}", "shape": f"{df.shape}"}}
not sure I understand
runningclearml-agent list
I get
`
workers:
- company:
id: d1bd92...1e52b
name: clearml
id: clearml-server-...wdh:0
ip: x.x.x.x
... `
updated the clearml.conf
with empty worker_id/name ran
clearml-agent daemon --stop
top | grep clearmKilled the pidsran
clearml-agent list
still both of the workers are listed
Well it seems that we have similar https://github.com/allegroai/clearml-agent/issues/86
currently we are just creating a new worker and on a separate queue
from the example -
since the `mp_hander`` runs
cmd = [sys.executable, sys.argv[0],
'--counter', str(counter - 1),
'--num_workers', str(args.num_workers),
'--use-subprocess' if args.subprocess else '--no-subprocess']
p = subprocess.Popen(cmd, cwd=os.getcwd())
can I run another subprocess
in the mp_worker
?
exactly - (that is how I used it in my initial code) - but if you have to convert it back to the original data type then something is broken...
yes - the pre_installations.sh
runs and completes - but the pytorch/main.py
file doesn't run.
so the Task completes successfully but without running the script
@<1523701205467926528:profile|AgitatedDove14> -
I'm getting the following error when running the following code within the mp_worker
command = ["ffmpeg","-i",f"{url}","-vcodec","libx264", "output.mp4"]
subprocess.run(command, stderr=subprocess.STDOUT)
TypeError: fork_exec() takes exactly 21 arguments (17 given)
Any suggestions?