we are cleaning, but there is a major problem
When deleting a task from web UI, nothing is deleted elsewhere
Debug images are not deleted, models are not deleted. And I suspect that scalars and logs are not deleted too
Im not sure why is that so
@<1523701070390366208:profile|CostlyOstrich36> Updated webserver and the problem still persists
This is the new stack:
WebApp: 1.15.1-478 • Server: 1.14.1-451 • API: 2.28
notice, we didnt update API (we had running experiments)
@<1523703436166565888:profile|DeterminedCrab71> Thanks for responding
It was unclear to me that I need to set 443 also everywhere in clearml.conf
Setting s3 host urls with 443 in clearml.conf and also in web UI made it work
Im now almost at the finish line. The last thing that would be great is to fix archived task deletion.
For some reason i have error of missing S3 keys in clearml docker compose logs, the folder / files are not deleted in S3 bucket.
You can see how storage_credentials.co...
Here are my clearml versions and elastisearch taking up 50GB
elastisearch also takes like 15GB of ram
How can I do that?
I need to save the original hash, otherwise I lose all trackability to about 2k experiments
But there are stil some wierd issues, i cannot see the files uploaded in bucket
@<1523701070390366208:profile|CostlyOstrich36> Hello, im still unable to understand how to fix this
I solved the problem.
I had to add tensorboard loggger and pass it to pytorch_lightning trainer logger=logger
Is that normal?
It is also possible to just make a copy of all the database files and move them to another server
i also think that if my package manager is set to uv, then it should only use uv and ignore pip at all
Getting errors in elastisearch when deleting tasks, get retunred "cant delete experiment"
I do notice another strange thing
Agent-services is down because It has no API key to clearm
ok, slight update. It seems like artifacts are uploading now to bucket. Maybe my folder explorer used old cache or something.
However, reported images are uploaded to fileserver instead of s3
here is the script im using to test things. Thanks
@<1523701435869433856:profile|SmugDolphin23> Any ideas how to fix this?
will it be appended in clearml?
"s3" is part of domain to the host
Yes, but does add_external_files makes chunked zips as add_files do?
It looks like im moving forward
Setting url in clearml.conf without "s3" as suggested works (But I dont add port ther, not sure if it breaks something, we dont have a port)
host: " our-host.com "
Then in test_task.py
task: clearml.Task = clearml.Task.init(
project_name="project",
task_name="task",
output_uri=" None ",
)
I think connection is created
What im getting now is bucket error, i suppose I have to specify it so...
good morning, I tried the script you provided and Im getting somewhere
I was on 1.7 version and now im on latest 1.11
Cant get screenshow yet (copying data), will add later.
What worries me is that config and agent folders are empty. I can reconfigure all agents, no problems.
But where is info about projects stored?
I purged all docker images and it still doesnt seem right
I see no side panel and it doesnt ask for login name