![Profile picture](https://clearml-web-assets.s3.amazonaws.com/scoold/avatars/DepressedChimpanzee34.png)
Reputation
Badges 1
195 × Eureka!that is the heaviest part for me
thanks AgitatedDove14 ! this is what I was looking for
I have some custom configuration that describes the experiment I am running..
Some dictionary.. nothing to do with frameworks
client has the following attributes:['auth', 'events', 'models', 'projects', 'queues', 'session', 'tasks', 'workers']
I get it with the simplest config if I define it as a dataclass, using the example you share as a basis:
config_files/cfg.py
` from hydra.core.config_store import ConfigStore
from dataclasses import dataclass
@dataclass
class MasterConfig:
test: str = 'test'
cs = ConfigStore.instance()
cs.store(name="config", node=MasterConfig) and for the main I had to make some small changes to connect my local server (I'm sharing them just in case): hydra_example.py:
# ClearML - Hydra Example
im...
I found a super weird fix for this error.. no idea why it works but maybe it can help with debugging it..try: task = Task.init(project_name="examples", task_name="hydra configuration", reuse_last_task_id=False) except: task = Task.init(project_name="examples", task_name="hydra configuration", reuse_last_task_id=True)
its the module where the cfg is defined, like in the example you shared config_files/cfg.py
neither Task nor task seem to have this attribute 🤔
cool thanks! local is quite confusing in this context.. but works 🙂
CostlyOstrich36 thanks, is there an example for using the post\get in a pythonic way to access the mentioned debug.ping ?
I tried also with the app website..
CostlyOstrich36 , I am not sure what you mean, but if you refer to the name of the category in the configuration, it doesn't matter.. the names are arbitrary for this example..
AgitatedDove14 ,yes my own server
is there an available reference to such a post request? I was trying some variations and it didn't work for me
response:{"meta":{"id":"33e9e80e94ee4384b45962acafcd2af5","trx":"33e9e80e94ee4384b45962acafcd2af5","endpoint":{"name":"","requested_version":1.0,"actual_version":null},"result_code":400,"result_subcode":0,"result_msg":"Invalid request path /v2.14/debug/ping","error_stack":null,"error_data":{}},"data":{}}
SuccessfulKoala55 , thanks I was looking for a way to do it programatically.. solved now, thanksrequest = url+ '/v2.14/debug.ping' r = requests.post(request) serverIsResponsive = r.json()['meta']['result_code'] == 200
CostlyOstrich36 ,I went through tasks and session and couldn't find an equivalent ping
is there a chance you help me with the specific POST call for debug? I was trying to implement it using requests package but I got errors.. didn't work for me.. I believe it something trivial
I am not familiar with this.. thats why I'm struggling.. this is what I tried:import requests import os r = requests.post(url+ '/v2.14/debug/ping') r.json()
response:
` {'meta': {'id': 'ebd6bdeaa95c4c3397009c71d9444040',
'trx': 'ebd6bdeaa95c4c3397009c71d9444040',
'endpoint': {'name': '', 'requested_version': 1.0, 'actual_version': None},
'result_code': 400,
'result_subcode': 0,
'result_msg': 'Invalid request path /v2.14/debug/ping',
'error_stack': None,
'error_data': {}},
'...
worked like a charm
its only these two files.. nothing else
something like in the example I shared<Machine 1> #Init Optimizer <Machine 2> **heavy one time Common Initialization** while True: #sample Optimizer # init task # Execute Something # report results <Machine i> **heavy one time Common Initialization** while True: #sample **same** Optimizer # init task # Execute Something # report results
The difference is that I want a single persistent machine, with a single persistent python script that can pull execute and report multiple tasks
Lets say I inherit from the Optimizer (you mean HyperParameterOptimizer class? or SearchStrategy?), implement a custom logic for experiment creation logic,
what does it actually exposes? creating an experiment means defining a task, enqueue it and then? I am trying to think what you meant I can put in the logic such that I get the desired effect
I am getting the following error when I try the above:
` /opt/conda/envs/torch_38/lib/python3.8/site-packages/clearml/backend_api/session/client/client.py in new_func(self, *args, **kwargs)
374 @wrap
375 def new_func(self, *args, **kwargs):
--> 376 return Response(self.session.send(request_cls(*args, **kwargs)))
377
378 new_func.name = new_func.qualname = action
TypeError: init() missing 1 required positional argument: 'items' `
I am familiar with the above..
I am talking about an overview.. I can count the workers manually but when I have a 100 workers its too much
it happens on task init