Reputation
Badges 1
371 × Eureka!wrong image. lemme upload the correct one.
It seems that is the case. Thank you for all your help guys.
CostlyOstrich36
I've finally gotten the triton engine to run. I'll be going through nvidia triton docs to find how to make an inference request. If you have an example inference request, I'll appreciate if you can share it with me.
I basically go to the model from the experiment first, then when in the model, I'm trying to download it but can't. I've screenshotted the situation.
I'll read the 3 examples now. Am I right to assume that I should drop Pipeline_Controller.py
Still unsure between finalize and publish? Since upload should upload the data to the server
Basically, right now when I save the model, it just goes in draft mode. What I want to do is that save the model only if it is better than the previous one, and once saved, publish it and have a name and tags that I want to add.
My main query is do I wait for it to be a sufficient batch size or do I just send each image as soon as it comes to train
Thus I wanted to pass the model id from the prior step to the next one.
'dataset' is the name of my Dataset Object
they're also enqueued
I actually just asked about this in another thread. Here's the link. Asking about the usage of the upload_artifact
Ok since its my first time working with pipelines, I wanted to ask. Does the pipeline controller run endlessly or does it run from start to end with me telling it when to start based on a trigger?
Like there are files in a specific folder on Machine A. A script on Machine A, creates a Dataset, adds files located in that folder, and publishes it. Now can you look at that dataset on the server machine? Not from the ClearML interface but inside normal directories, like in /opt/clearml etc. this directory mentioned is just an example.
I'll test it with the updated one.
Can you take a look here?
https://clearml.slack.com/archives/CTK20V944/p1637914660103300
This is where I've mentioned the anonymous task spawn issue. I kind of want to understand what's causing the problem, if it is a problem etc
Basically the environment/container the agent is running in needs to have specific cuda installed. Is that correct CostlyOstrich36 ?
Thanks for the help.
I would normally like for it to install any requirements needed on its own.
Considering I don't think the function itself requires Venv to run normally but in this case it says it can't find venv
In this case, it would wait until any change occurred in the project mentioned here. Even if a task started, and not a dataset change or publish or anything, it would trigger.
You could be right, I just had a couple of packages with this issue so I just removed the version requirement for now. Another issue that might be the case, might be that I'm on ubuntu some of the packages might've been for windows thus the different versions not existing
do I just post the issue on the main clearml repo?
Let me try to be a bit more clear.
If I have a training task in which I'm getting multiple ClearML Datasets from multiple ClearML IDs. I get local copies, train the model, save the model, and delete the local copy in that script.
Does ClearML keep track of which data versions were gotten and used from ClearML Data?
Anyway I restarted the triton serving engine.
