Badges 151 × Eureka!
Sorry to ping you @<1523701087100473344:profile|SuccessfulKoala55> , can you offer any ideas to the two questions from my reply (about the correct web app cloud access and the correct way to specify a blob storage in the
clearml.conf file? Thanks 🙏
Hi @<1523701087100473344:profile|SuccessfulKoala55> ,
thanks for the pointers.
I didn't know that the plot data is stored in elasticsearch. Good to know. It relates to the rest of my questions in that I want to understand where everything is saved, all the parts of my experiments. The plots are actually the most important part, since I have direct access to the artifacts I save (like, say, models) but not to the plot data which helps me compare and rank experiments. I mention tensorboard be...
This is how the links to the artifacts looks like (the part I blurred out is is the last part of the secret, which is working fine since the task was able to upload those correctly to storage, I can check that):
I apologize 😳 . Turns out I have all those things. Again, my apologies 😞 .
Not sure how to check that tbh. Does this help:
root@aea5d96a8ed3:/usr/agent# clearml-agent --version CLEARML-AGENT version 1.0.0Would be nice to display this info maybe somewhere inhere:
OK I won't edit the db 😄 . Thanks for the suggestion, we'll use that!
` radu on vm-aimbrain-01 in volt on rg/dev [$] is 📦 v7.0.1 via 🐍 v3.8.5 via C volt
✦2 ❯ git status
On branch rg/dev
nothing to commit, working tree clean
radu on vm-aimbrain-01 in volt on rg/dev [$] is 📦 v7.0.1 via 🐍 v3.8.5 via C volt
✦ ❯ du -sh .
35M . `
OK I can confirm that those display ok for me too. My problem is, they don't for my experiments, which is I care about..
Yes, I find myself trying to select "points" on the overview tab. And I find myself wanting to see more interesting info in the tootip.
Version: 1.1.1-135 • 1.1.1 • 2.14
New to lightning too, but I'm suspecting that since your args don't mention a specific logger, the pl trainer will instantiate the default one. Excerpt from the trainer docstring:
logger: Logger (or iterable collection of loggers) for experiment tracking. A ``True`` value uses the default ``TensorBoardLogger``. ``False`` will disable logging. If multiple loggers are provided and thesave_dir` property of that logger is not set, local files (check...
I'm not sure how to access the developer tools 😳
ah, I thought you meant something clearml specific.
AgitatedDove14 Yes! That would be exactly what I want (i.e.
get_configuration_as_dict .) Alas, no such thing exists in 1.4.1. Is that supposed to come in a next version?
Hi SweetBadger76 , thanks, I think I've made it work. The main point of confusion was between dealing with different type of
Task objects (i.e.
clearml.backend_api.services.v2_13.tasks.Task returned by
get_all , which don't have any of those methods).
set_parameters didn't just work as expected, I had to flatten the dicts myself (which clearml apparently does on its own when I call
set_parameters on a new task.
Thank you all. 🙏
True, the Hyperparameters tab does show the environment (sorry, my bad). The repo information, or the uncommitted changes don't show up for me 😞 . What version did you use?
The problem appears to be related to the lack of an "origin" remote. However, I think the configuration is perfectly legitimate, so maybe there's a better alternative than relying on the behaviour of
git ls-remote --get-url .
Unfortunately it still happens 😞 :
` Epoch 51: 100%|███████████████████████████████████████████████████████████| 361/361 [02:52<00:00, 2.10it/s, loss=0.169, v_num=9-29]
2021-09-17 09:58:22,253 - clearml.Task - INFO - Waiting for repository detection and full package requirement analysis
2021-09-17 10:03:22,254 - clearml.Task - INFO - Repository and package analysis timed out (300.0 sec), giving up
2021-09-17 10:03:22,313 - clearml.Task - WARNING - Failed auto-det...
The template appears to be
<alias> <url> <fetch|push> .
.git/config file has sections for each remote too. Example:
[remote "github"] url = firstname.lastname@example.org:biocatchltd/volt.git fetch = +refs/heads/:refs/remotes/github/Would be nice to report which remote the checked out branch actually tracks.
you think that simply adding origin will fix this - I really don't mind doing that
I'll let you know asap
I'm been overly optimistic it seems (based on the hints from this issue: https://stackoverflow.com/questions/59455268/how-to-disable-progress-bar-in-pytorch-lightning ). It really looked like it worked. But I was mislead by the fact that the issue has a peculiar pattern to it: it doesn't appear at the beginning of each epoch. There seem to be a couple of extra cr / flushes at the beginning and then a (puzzling) extra "Validation" log line seems to trigger the continuous scrolling. Anyway, my...
'scikit' worked nicely, thanks again
Must be something else foul at play here..
This is adapted from one of the methods in their
from tqdm import tqdm bar = tqdm( desc="Training", initial=1, position=1, disable=False, leave=False, dynamic_ncols=True, file=sys.stderr, smoothing=0) with bar: for i in range(10): time.sleep(0.1) bar.update() print('done')In the console this works as expected, but in a jupyter notebook this produces a scrolling log (because of the position=1 argument, which happens whenever the bar is not th...
That works fine:
1631895370729 vm-aimbrain-01 info ClearML Task: created new task id=cfed3ea8512d4d9f858d085bd79e62e8 2021-09-17 16:16:10,744 - clearml.Task - INFO - No repository found, storing script code instead ClearML results page: `
1631895370892 vm-aimbrain-01 info start
1631895370896 vm-aimbrain-01 error 0%| | 0/100 [00:00<?, ?it/s]
1631895471026 vm-aimbrain-01 error 100%|████...
Take a guess 😂 (it's "origin")