Reputation
Badges 1
195 × Eureka!CostlyOstrich36 ,I went through tasks and session and couldn't find an equivalent ping
SuccessfulKoala55 , thanks I was looking for a way to do it programatically.. solved now, thanksrequest = url+ '/v2.14/debug.ping' r = requests.post(request) serverIsResponsive = r.json()['meta']['result_code'] == 200
is there a chance you help me with the specific POST call for debug? I was trying to implement it using requests package but I got errors.. didn't work for me.. I believe it something trivial
response:{"meta":{"id":"33e9e80e94ee4384b45962acafcd2af5","trx":"33e9e80e94ee4384b45962acafcd2af5","endpoint":{"name":"","requested_version":1.0,"actual_version":null},"result_code":400,"result_subcode":0,"result_msg":"Invalid request path /v2.14/debug/ping","error_stack":null,"error_data":{}},"data":{}}
AgitatedDove14 ,yes my own server
is there an available reference to such a post request? I was trying some variations and it didn't work for me
AgitatedDove14 , missed your message, python
thanks AgitatedDove14 ! this is what I was looking for
and then second part is to check if it is up and alive
this way I can avoid the heavy computation I describe above for each individual trial
Thanks Martin! I'll test it in the following days, I'll keep you updated!
The difference is that I want a single persistent machine, with a single persistent python script that can pull execute and report multiple tasks
for my specific use case I don't mind if they keep the order, I just want them to be at the top
Bonus 2: having a way to push experiments up in the queue
I am familiar with the above..
I am talking about an overview.. I can count the workers manually but when I have a 100 workers its too much
CostlyOstrich36 thanks, is there an example for using the post\get in a pythonic way to access the mentioned debug.ping ?
no programmatic python options? could be nice..
so a different behavior between a string and a string in a tuple is by design? I find it confusing, I guess this is the YAML convention?
https://colab.research.google.com/drive/1w5lQGxsblnLGlhJEDH_b0aIiUvjGeLjy?usp=sharing
Another option is to pull Tasks from a dedicated queue and use the LocalClearMLJob to spwan them
This sounds like it can work. we are talking about something like:
` #<Machine 1>
#Init Optimizer with some dedicated queue
<Machine 2>
heavy one time Common Initialization
while True:
# sample queue
# enqueue with LocalClearMLJob
# Execute Something
# report results
<Machine i>
heavy one time Common Initialization
while True:
# sample same queue
# enqueue wi...
SuccessfulKoala55 can you refer me to how to increase the number of API server processes?
AgitatedDove14 if we look at the host machine we can see a single python process that is actually busy
AgitatedDove14 , seem to work significantly better! thanks!
It seems to me that the source of the mismatch is the str(tuple())
FrothyDog40 Hi, there is a sort of a UI bug related to the mentioned above..
if you choose a single experiment then use filter or sorting such that it is not displayed anymore you can't deselect it.
you have this - sign in the top selection box (select all \ deselect all) but it doesn't do anything.
the workaround is to select an additional experiment and then when the multiple experiment bar pops up choose show experiments selected and deselect them and go back.
its inconvenient, both in th...
do you happen to have some reference snippet to create such a callback? I am only familiar with:from clearml.binding.frameworks import WeightsFileHandler
Thanks AgitatedDove14 , we have no issue with "html stuck in cache" I'm not sure what are you referring to..
am I supposed to change the WeightsFileHandler inplace?