
Reputation
Badges 1
53 × Eureka!Hi AgitatedDove14
this is how our calls look like:
` from pytorch_lightning.loggers import TensorBoardLogger
logger = TensorBoardLogger(save_dir=".", name="debug plotting", 1)
logger.experiment.add_histogram(f"A", data[data.by == 0])
logger.experiment.add_histogram(f"B", data[data.by == 1]) `the result of which is shown in my post above.
This is some test data, and how we'd like things to look:
` def make_data(size: int=10000, n: int=5) -> pd.DataFrame:
x = np.abs(np.random.normal(siz...
Hi Martin, to expand on my previous comments: the template for _Driver
already exists; I'm suggesting to make it public. Consequently, StorageHelper
should accept a driver
parameter to __init__
, defaulting to None
. Only when its value is not provided by the user should the library go out of its way to do the right thing and check all the known storage providers, fetch credentials, what not - stuff that will not work for most users, most of the time (even if you ...
ah nice, I'll try auto_connect_frameworks
(probably with {'joblib': False}
? - we don't use scikit-learn)
'scikit' worked nicely, thanks again
I'm been overly optimistic it seems (based on the hints from this issue: https://stackoverflow.com/questions/59455268/how-to-disable-progress-bar-in-pytorch-lightning ). It really looked like it worked. But I was mislead by the fact that the issue has a peculiar pattern to it: it doesn't appear at the beginning of each epoch. There seem to be a couple of extra cr / flushes at the beginning and then a (puzzling) extra "Validation" log line seems to trigger the continuous scrolling. Anyway, my...
I don't control tqdm, (otherwise I would have already gone for Stef's suggestion) - pytorch-lightning does in this particular script 😞 .
Hi Martin, it is a tqdm parameter (the default ProgressBar
in pytorch lightning is unfortunately relying on tqdm). This is from the tqdm docs:dynamic_ncols : bool, optional If set, constantly alters
ncolsand
nrows` to the
environment (allowing for window resizes) [default: False].
nrows : int, optional
The screen height. If specified, hides nested bars outside this
bound. If unspecified, attempts to use environment...
That works fine:1631895370729 vm-aimbrain-01 info ClearML Task: created new task id=cfed3ea8512d4d9f858d085bd79e62e8 2021-09-17 16:16:10,744 - clearml.Task - INFO - No repository found, storing script code instead ClearML results page:
`
1631895370892 vm-aimbrain-01 info start
1631895370896 vm-aimbrain-01 error 0%| | 0/100 [00:00<?, ?it/s]
1631895471026 vm-aimbrain-01 error 100%|████...
I found out that the lightning trainer has a progress_bar_refresh_rate
argument (default set to 1) which produces the spamming logs. If I set that to 10, I get 1/10th of the spam (but a janky progress bar in the console). I could set it to 0 to disable it, but that's not really a fix. What I'd really want is the same behaviour in the console (one smooth progress bar) and one line per epoch in the logs; high hopes, right? 😊
In case anyone is interested, the minimum effort workaround I found is to edit pytorch_lightning/callbacks/progress.py
and change all occurrences of dynamic_ncols=True
to dynamic_ncols=False
in the calls to tqdm
. One could of course implement a custom callback inheriting from their ProgressBar
class.
Sorry, I meant the "origin" part. The warning is no more.
We can't really know (possibly ever 🙂 ), but if the bug happens again I'll be sure to report it here.
Hi AgitatedDove14 , I deleted everything in /opt/clearml as per the docs. Should I delete anything else?
The problem appears to be related to the lack of an "origin" remote. However, I think the configuration is perfectly legitimate, so maybe there's a better alternative than relying on the behaviour of git ls-remote --get-url
.
you think that simply adding origin will fix this - I really don't mind doing that
Thanks AppetizingMouse58 . I managed to fix it by removing docker completely and reinstalling it.
I actually put all the commands in a script. The failure mode is exactly the same. I have no idea what to do next.
` #!/bin/bash
clearml_root=$1
if [[ $# -gt 0 ]]; then
echo Using "$1" as root
else
echo No root argument was provided, using /datadrive1
clearml_root=/datadrive1
fi
clearml="$clearml_root/clearml"
rm -R "$clearml"
mkdir -p "$clearml"/data/elastic_7
mkdir -p "$clearml"/data/mongo_4/db
mkdir -p "$clearml"/data/mongo_4/configdb
mkdir -p "$clearml"/data/redis
mkdir -p "$cl...
I'll let you know asap
Hi SuccessfulKoala55 ,yes, I am running as sudo. It's not my first time setting this up either. Just that on this vm (which is a particularly large instance) I get this issue which has me stumped.
` radu on vm-aimbrain-01 in volt on rg/dev [$] is 📦 v7.0.1 via 🐍 v3.8.5 via C volt
✦2 ❯ git status
On branch rg/dev
nothing to commit, working tree clean
radu on vm-aimbrain-01 in volt on rg/dev [$] is 📦 v7.0.1 via 🐍 v3.8.5 via C volt
✦ ❯ du -sh .
35M . `
I didn't add that to the script since the effect is persistent (i.e. it only needs to be done once, right?) In any case, I checked that multiple times and it was as expected.
Take a guess 😂 (it's "origin")
✦2 ❯ git remote show
github
The template appears to be <alias> <url> <fetch|push>
.
The .git/config
file has sections for each remote too. Example:[remote "github"] url = git@github.com:biocatchltd/volt.git fetch = +refs/heads/:refs/remotes/github/
Would be nice to report which remote the checked out branch actually tracks.