hey OutrageousSheep60
what about the process ? there must be one clearml-agent process that runs somwhere, and that is why it can continue reporting to the server
Hey Igor
I am not the expert about this topic. I have someone who better knows the topic that is coming back to you straight after his meeting. 🙂
can you please provide the apiserver log and the elasticsearch log?
Hey TartSeagull57
We have released a version that fixes the bug. It is a RC but it is stable. Version number is 1.4.2rc1
hey
"when cloning an experiment via the WebUI, shouldn't the cloned experiment have the original experiment as a parent? It seems to be empty"
you are right, i think there is a bug here. We will release a fix asap 🙂
hey WickedElephant66 TenderCoyote78
I'm working on a solution, just hold on, I update you asap
It depend if you have added the parameter monitor_models=['mymodel'] in the PipelineDecorator.component. If not, the model will not be logged. If you have, then torch.save will save your model to a local path, that will be logged into the main pipeline task
JuicyFox94
can you try again after having upgraded to 3.6.2 ?
hey Maximilian,
which version of clearml are you using ?
Last (very) little thing : could you please open a Github issue for this irrelevant warning 🙏 ? It makes sense to register on GH those bugs, because our code and releases are hosted there.
Thank you !
http://github.com/allegroai/clearml/issues
Just to keep you updated, as promised 🙂
we have found the bug and will release a fix asap. for that too i will keep you updated 🙂
i am not sure i get you here.
when pip installing clearml-agent, it doesnt fire any agent. the procedure is that after having installed the package, if there isnt any config file, you do clearml-agent init
and you enter the credentials, which are stored in clearml.conf. If there is a conf file, you simply edit it and manually enter the credentials. so i dont understand what you mean by "remove it"
btw can you screenshot your clearml-agent list and UI please ?
Of course. Here it is
https://github.com/allegroai/clearml/issues/684
I'll keep you updated
Hi EnormousWorm79
The Pycharm testrunner wraps the script into a local cript, and thats what you are getting.
(jb pytest runner). Because it is local, you lose the source info
Let me check if I have a workaround or solution for you. I keep you updated
Hi,
We are going to try to reproduce this issue and will update you asap
Hey Atalya 🙂
Thanks for your feedback. This is indeed a good feature to think asbout.
So far there is no other ordering than the alphabetical. Could you please create a feature request on github ?
Thanks
Hi Max
you can configure a clearml agent to pull your docker image from ECR and run the experiment into it. Is that answering your question ?
btw here is the content of the imported file:
import
torch
from
torchvision
import
datasets, transforms
import
os
MY_GLOBAL_VAR = 32
def my_dataloder
():
return
torch.utils.data.DataLoader(
datasets.MNIST(os.path.join('./', 'data'), train=True, download=True,
transform=transforms.Compose([
transforms.ToTensor()
` ...
I check that
Concerning the snippet example, here is the link :
https://github.com/allegroai/clearml/issues/682
Hi Alon
This is indeed a known bug, we are currently working on a fix.
Hi BeefyHippopotamus73
did you managed to get rid of your issue ?
Hey UnevenDolphin73
Is there any particular reason why not to create the dataset ? I mean, you need to use it in different tasks, so it could make sense to create it , for it to exist on its own, and then to use it at will in any task, by simply retrieving its id (using Dataset.get)
Makes sense ?
hi TenderCoyote78
can you please give some more precision about what you intend to achieve ? I am afraid not to well understand your question
can you share with me an example or part from your code ? I might miss something in wht you intend to achieve