Reputation
Badges 1
76 × Eureka!projects.get_all_ex API call
` {"meta":{"id":"17c6e609ace54bf8bfdf3113c39fd470","trx":"17c6e609ace54bf8bfdf3113c39fd470","endpoint":{"name":"projects.get_all_ex","requested_version":"2.20","actual_version":"1.0"},"result_code":200,"result_subcode":0,"result_msg":"OK","error_stack":"","error_data":{}},"data":{"projects":[{"id":"75d04598197a445ebef533814022c58d","company":{"id":"d1bd92a3b039400cbafc60a7a5b1e52b"},"user":{"id":"a174c4e36b0446a7b3b5dd1ff5261962"},"name":"ex-1","basename":"ex-1","description":"","created":"20...
yeah, we cannot do anything to that [undefined]. i got this when i click that.
hi @<1576381444509405184:profile|ManiacalLizard2> thanks for the answer, i will try that!
it seems if i access with my dns cannot see
and if access with ip address can see
i see, it solved right now using default_output_uri, Thanks!
i need custom output_uri for some function because split dataset and model artifacs.
Hi @<1523701070390366208:profile|CostlyOstrich36> , thanks for response, sorry for late replay,
this is my configuration in yaml, i facing difficulty when there is params in list. somehow, form to display bunch list not easy to see. do you have suggestion? Thanks!
download-data:
dataset_train:
-
-
-
dataset_test:
-
-
-
train:
data:
batch: 4
input_size: 224
split:
t...
i feel pain to make it as form if so much varible want to changes.
i see okay thanks
also i found, i cannot delete project, even the project is show empty experiment.
the current my solution is upload my config to s3, and the pipeline will download it and read it when execute. but its decrase flexiblity.
sorry, but can clearml pipeline do this scenario?
Hi @<1523701087100473344:profile|SuccessfulKoala55> , Thanks for your response.
I'm not entirely sure about the use of CLEARML_ENV
since I haven't interacted with it before. Could you guide me on what I should set as its value?
Previously, the system was running smoothly. However, I've run into some issues after making certain configuration changes to modify the server permissions. Specifically, I'm curious if these changes might have influenced the agent's permission to access certain...
Hi CostlyOstrich36 ,
nope, i mean my server does not have pip/conda. so i will go for docker/container, is that possible if i install clearml-agent inside python:3.10 container?
my case more like there is a task/process that running but somehow its takes too long to completed. it can be because connection issue forgot to put connection timeout, a problem connection database, etc that makes status still running, but its traped in a situation like that.
so i want to force shutdown a task to failed if that happen
Thanks for response.
from clearml import Task
from clearml.automation import TaskScheduler
from datetime import timedelta, datetime
def my_task():
task = Task.init(...)
# do somthinge
print("do something")
# sleep 10
condition = True
if condition:
# i want to trigger run another task by
# set some config in task, but execute tomorrow/sometime
# not directly run at the time.
# here i use
task_id = task.id
task.cl...
Hi AppetizingMouse58 , this is from My Work View
` # Payload
{"meta":{"id":"4ff606b50768402495674f4a2b37bdf4","trx":"4ff606b50768402495674f4a2b37bdf4","endpoint":{"name":"projects.get_all_ex","requested_version":"2.20","actual_version":"1.0"},"result_code":200,"result_subcode":0,"result_msg":"OK","error_stack":"","error_data":{}},"data":{"projects":[{"id":"75d04598197a445ebef533814022c58d","company":{"id":"d1bd92a3b039400cbafc60a7a5b1e52b"},"user":{"id":"a174c4e36b0446a7b3b5dd1ff5261962"},...
it seems i forgot using clearml-agent init.
i follow this way: https://clear.ml/docs/latest/docs/guides/ide/google_colab/
i see thanks for the answer, i will read that reference.
Hi @<1523701205467926528:profile|AgitatedDove14> , Thanks for rresponse!
this my simple code to test scheduler
import datetime
from clearml.automation import TaskScheduler
def test_make():
print('test running', datetime.datetime.now())
if __name__ == '__main__':
task_scheduler = TaskScheduler(
sync_frequency_minutes=30,
force_create_task_name='controller_feedback',
force_create_task_project='Automation/Controller',
)
print('\n[utc_timestamp]...
hmm i want to make custom function that need credential registed on clearml.conf, like aws s3.
is clearml-agent have clearml.conf? where the path for that? i just test it, running using clearml-agent, but not found /root/clearml.conf
i am using dictionary, more convinient for me and can categorize each params.
Thanks, hope that feature will ready soon!
AgitatedDove14 , yes. i tried 3 times, and it always happen.
Hi SmugDolphin23 , i have try 1.8.4rc1, and yeah its working! Thanks!
Hi @<1523701087100473344:profile|SuccessfulKoala55> ,
We have successfully created a sample for the migration. Here are the changes:
- URL for MongoDB from
s3://
toazure://
- Elasticsearch as you suggested
However, our main focus is that most of our production fetches models from ClearML, which are configured withs3://
URLs.
There is an issue/bug in the UI when downloading via Azure. Here are the details: None .
Thanks guys, i will try to learn that first. i will updates when executing these ideas. @<1523701087100473344:profile|SuccessfulKoala55> @<1590514584836378624:profile|AmiableSeaturtle81> 🙌
yes, so far i know, if we want to upload dataset on clearml, we need provide local_path to data, then clearml will upload to the platform.
my data not on local, but s3 bucket.
is there a way to point s3 url ? my currently workflow is download my data from s3 bucket to local, then upload to clearml.
`
payload
{"id":["75d04598197a445ebef533814022c58d"],"include_stats":true,"check_own_contents":true,"active_users":["a174c4e36b0446a7b3b5dd1ff5261962"],"search_hidden":true}
Response
{"meta":{"id":"8d18a89599db4d899ab40959a39970b3","trx":"8d18a89599db4d899ab40959a39970b3","endpoint":{"name":"projects.get_all_ex","requested_version":"2.20","actual_version":"1.0"},"result_code":200,"result_subcode":0,"result_msg":"OK","error_stack":"","error_data":{}},"data":{"projects":[{"id":"75d0459819...
alright, will try thanks!