Reputation
Badges 1
53 × Eureka!Hi SweetBadger76 , thanks, I think I've made it work. The main point of confusion was between dealing with different type of Task objects (i.e. clearml.backend_api.services.v2_13.tasks.Task returned by get_all , which don't have any of those methods).
Interestingly, set_parameters didn't just work as expected, I had to flatten the dicts myself (which clearml apparently does on its own when I call set_parameters on a new task.
Thank you all. 🙏
True, the Hyperparameters tab does show the environment (sorry, my bad). The repo information, or the uncommitted changes don't show up for me 😞 . What version did you use?
there is - it's called "github"
Try updating to 1.1.0?
The problem appears to be related to the lack of an "origin" remote. However, I think the configuration is perfectly legitimate, so maybe there's a better alternative than relying on the behaviour of git ls-remote --get-url .
` # Development mode worker
worker {
# Status report period in seconds
report_period_sec: 2
# ping to the server - check connectivity
ping_period_sec: 30
# Log all stdout & stderr
log_stdout: true
# Carriage return (\r) support. If zero (0) \r treated as \n and flushed to backend
# Carriage return flush support in seconds, flush consecutive line feeds (\r) every X (default: 10) s...
Hi again. After looking into the matter a little bit, I realise I'd have liked having the option of using a StoreManager ABC which I would implement myself using whatever storage provider I happen to use and whatever package versions happened to support it. To put it differently, instead of you implementing managers for gcs, azure, aws, etc, it would be a much nicer alternative (for me, and I suspect eventually for you too) for clearml's store manager to wrap whatever object the user pr...
` radu on vm-aimbrain-01 in volt on rg/dev [$] is 📦 v7.0.1 via 🐍 v3.8.5 via C volt
✦2 ❯ git status
On branch rg/dev
nothing to commit, working tree clean
radu on vm-aimbrain-01 in volt on rg/dev [$] is 📦 v7.0.1 via 🐍 v3.8.5 via C volt
✦ ❯ du -sh .
35M . `
I don't control tqdm, (otherwise I would have already gone for Stef's suggestion) - pytorch-lightning does in this particular script 😞 .
✦2 ❯ git remote showgithub
We can't really know (possibly ever 🙂 ), but if the bug happens again I'll be sure to report it here.
you think that simply adding origin will fix this - I really don't mind doing that
Hi AgitatedDove14 , I deleted everything in /opt/clearml as per the docs. Should I delete anything else?
If we decide go forward with clearml we'll probably do just that 🙂
Hi Martin, it is a tqdm parameter (the default ProgressBar in pytorch lightning is unfortunately relying on tqdm). This is from the tqdm docs:dynamic_ncols : bool, optional If set, constantly altersncolsandnrows` to the
environment (allowing for window resizes) [default: False].
nrows : int, optional
The screen height. If specified, hides nested bars outside this
bound. If unspecified, attempts to use environment...
Yeah, I experienced the same issue. Training stopps / freezes at the end of the 10th, or 15th epoch. Using pytorch_lightning as well.
Hi AgitatedDove14
this is how our calls look like:
` from pytorch_lightning.loggers import TensorBoardLogger
logger = TensorBoardLogger(save_dir=".", name="debug plotting", 1)
logger.experiment.add_histogram(f"A", data[data.by == 0])
logger.experiment.add_histogram(f"B", data[data.by == 1]) `the result of which is shown in my post above.
This is some test data, and how we'd like things to look:
` def make_data(size: int=10000, n: int=5) -> pd.DataFrame:
x = np.abs(np.random.normal(siz...
AgitatedDove14 Yes! That would be exactly what I want (i.e. get_configuration_as_dict .) Alas, no such thing exists in 1.4.1. Is that supposed to come in a next version?
CostlyOstrich36
` {"meta":{"id":"3cceedbbc65d480096ebb02b5aba5902","trx":"3cceedbbc65d480096ebb02b5aba5902","endpoint":{"name":"tasks.get_configurations","requested_version":"2.17","actual_version":"1.0"},"result_code":200,"result_subcode":0,"result_msg":"OK","error_stack":"","error_data":{}},"data":{"configurations"...
Hi @<1523701087100473344:profile|SuccessfulKoala55> ,
thanks for the pointers.
I didn't know that the plot data is stored in elasticsearch. Good to know. It relates to the rest of my questions in that I want to understand where everything is saved, all the parts of my experiments. The plots are actually the most important part, since I have direct access to the artifacts I save (like, say, models) but not to the plot data which helps me compare and rank experiments. I mention tensorboard be...
Sorry to ping you @<1523701087100473344:profile|SuccessfulKoala55> , can you offer any ideas to the two questions from my reply (about the correct web app cloud access and the correct way to specify a blob storage in the clearml.conf file? Thanks 🙏
The template appears to be <alias> <url> <fetch|push> .
The .git/config file has sections for each remote too. Example:[remote "github"] url = git@github.com:biocatchltd/volt.git fetch = +refs/heads/:refs/remotes/github/Would be nice to report which remote the checked out branch actually tracks.
This is how a configuration item looks like:<tasks.ConfigurationItem: { "name": "filter", "value": "inference = [{\n type = \"StreamFilter\"\n params {\n context = \"full\"\n op = \"or\"\n lower_bounds {\n key = 16\n mouse = 32\n }\n }\n }]\ntrain {\n users {\n op = \"and\"\n lower_bounds {\n min_sessions = 32\n }\n }\n}", "type": "dictionary" }>The value is a string that prints pretty but I'm not sure how to p...
Hi Martin, to expand on my previous comments: the template for _Driver already exists; I'm suggesting to make it public. Consequently, StorageHelper should accept a driver parameter to __init__ , defaulting to None . Only when its value is not provided by the user should the library go out of its way to do the right thing and check all the known storage providers, fetch credentials, what not - stuff that will not work for most users, most of the time (even if you ...
I'll let you know asap
