
Reputation
Badges 1
103 × Eureka!good questions 🙂
they are plots. they have unique titles. i’m using the auto-logging mechanism—so set up the task, then plt.show()
no more than 114 plots are shown in the plots tab.
should api.credentials.access_key
be the same as the access_key
in clearml.conf
?
looking in the web app, under the “App Credentials” section, it lists those credentials as “used” when I attempted the curl
command.
yep, that’s what i’m seeing, they’re all PNGs in that folder.
yes, i see no more than 114 plots in the list on the left side in full screen mode—just checked and the behavior exists on safari and chrome
hmmm, i’m not creating the task in __main__
, i wonder if that’s why
i appreciate your help today. it can’t be very fun working on a sunday. i hope you get some relax time away from the computer today, and look forward to hearing more when you are working.
i did want to point out, though, that when manually reporting, it looks like the plots don’t get “cleared” properly: https://demoapp.demo.clear.ml/projects/52eb5c9d938244daaa6fa460edce5e22/experiments/78fa65250e0544d7b50425a82dde75f5/info-output/metrics/plots?columns=selected&columns=type&columns=name&colu...
but, the call used to start the script was python -m module.name --args
great news! thank you! when there’s a new release, i need to docker-compose build && docker-compose up
to get the latest?
🤔
Media is uploaded to a preconfigured bucket (see setup_upload()) with a key (filename) describing the task ID, title, series and iteration.
❯ cat ~/clearml.conf | grep git_user
git_user: "aaaaaaaaaaaaa"
❯ cat ~/clearml.conf | grep -A 2 vcs_cache
vcs_cache: {
enabled: false,
path: ~/.clearml/vcs-cache
of course, SERVER_IP_ADDRESS
is the actual IP address of the server, AND i made sure that CLEARML_HOST_IP
was set correctly before issuing the docker-compose
command
thanks for doing that and thanks for your work on the project 🙂
hi SubstantialElk6 , not sure if you were successful on this but i struggled with it as well, and it looks like the information is not in the linked document anymore.
in the end i realized that i needed to download apiserver.conf
from the clearml-server
repo ( https://github.com/allegroai/clearml-server/blob/master/apiserver/config/default/apiserver.conf ) and then add a user/pass for myself (starting at line 82).
actually its missing imports from the second level too
is there a limit to the search depth for this?
i’ve got a training script that imports local package files and those items import other local package files. ex:
train.py
from local_package.callbacks import Callbacks
local_package/callbacks.py
from local_package.analysis import Analysis
local_package/analysis.py
import pandas as pd
the original task only lists the following as installed packages:
clearml == 1.9.1rc0
pytorch_lightning == 1.8.6
torchvisi...
if i have code that’s just in a git repo but is not installed in any way, it runs fine if i invoke the entrypoint in a shell. but clearml will not find dependencies in secondary imports (as described above) if the agent someone just clones the repo but does not install the python package in some way.
okay, so my problem is actually that using a “local” package is not supported—ie i need to pip install the code i’m running and that must correctly specify its dependencies
the VCS cache was empty before that run. then, even with the VCS cache being disabled in the config, there was a new lock file and directory after running.
2023-05-06 12:05:49,168 - clearml.Task - WARNING - ### TASK STOPPED - USER ABORTED - STATUS CHANGED ###
lol.
this changes the status in the UI to “aborted”.
not ideal, but if the answer is “for this to work, tasks must be run by an agent” i accept it
so now i have
git_pass: "[NEW KEY]"
enable_git_ask_pass: false
in my clearml.conf file
i’ve just verified that they’re all writen to /opt/clearml/data/fileserver/[PROJECT_NAME]/[DESCRIPTION]/metrics