yeah, it seems i will try option 2 that you mentioned.
i upload my full my configuration, and when done using google colab my configuration will gone too.
hi i have similar case, but can we scheduled new task here?
def trigger_task_func(task_id):
print("trigger running...")
try:
previous_task = Task.get_task(task_id=task_id)
print(previous_task.artifacts)
try:
fp = previous_task.artifacts['latest_condition'].get_local_copy()
params = open_json(fp)
last_index = params.get('last_index')
day_n = params.get('iteration')
print("Success Fetching", param...
it seems i forgot using clearml-agent init.
i follow this way: https://clear.ml/docs/latest/docs/guides/ide/google_colab/
projects.get_all_ex API call
` {"meta":{"id":"369cbeafb3ec4b36ae7c07e781be5941","trx":"369cbeafb3ec4b36ae7c07e781be5941","endpoint":{"name":"projects.get_all_ex","requested_version":"2.20","actual_version":"1.0"},"result_code":200,"result_subcode":0,"result_msg":"OK","error_stack":"","error_data":{}},"data":{"projects":[{"company":{"id":"d1bd92a3b039400cbafc60a7a5b1e52b"},"created":"2022-09-26T01:33:28.216000+00:00","id":"5664bd88f48d4868a80805f6d824d5a9","name":"ex-1/sub-ex-1","user":{"id":"a174c4e36b0446a7b3b5dd1ff5261...
alright, will try thanks!
Hi SmugDolphin23 , i have try 1.8.4rc1, and yeah its working! Thanks!
Thanks for response.
from clearml import Task
from clearml.automation import TaskScheduler
from datetime import timedelta, datetime
def my_task():
task = Task.init(...)
# do somthinge
print("do something")
# sleep 10
condition = True
if condition:
# i want to trigger run another task by
# set some config in task, but execute tomorrow/sometime
# not directly run at the time.
# here i use
task_id = task.id
task.cl...
hi @<1523701087100473344:profile|SuccessfulKoala55> , it solved! thanks for information CLEARML_ENV ! I just accidently write environment varible CLEARML_ENV on every clearml-agent.conf. π
yup,
example just need choosing between SGD, Adam, AdamW on optimizer field
Hi AgitatedDove14 ,
right now i can delete, i set configuration to see hidden project/file on UI, and it show task related (e.g pipeline, dataset) on that project. but yeah, every time i make project with subproject inside, there is [undefined] there. i create project using code and UI, the result is same.
this is my docker-compose conf. i change all port 8XXX to 7XXX nad also change /opt/cleaml to /mnt/hdd_2/clearml.
` version: "3.6"
services:
apiserver:
command:
- apiserver
...
i see okay thanks
from src.net import Classifier
ModuleNotFoundError: No module named 'src'
hmm yeah i think, this is not possible to share a whole script here.
Hi @<1523701070390366208:profile|CostlyOstrich36>
i attach for complete log
here my structure:
.
βββ app
β βββ backend
β βββ frontend
βββ assets
β βββ demo-app-sample.png
β βββ workflow.png
βββ config
β βββ clearml.conf
β βββ list_models.py
β βββ list_optimizer.py
β βββ __pycache__
β βββ train_config.py
βββ docker
β βββ Dockerfile
β βββ Dockerfile.app
β βββ requirements.prod.txt
β βββ requirements.train.txt
β βββ requirements.txt
βββ lightning_logs
βββ Mak...
I see, yeah my alternative solution right now is just to show the list of options outside on ClearML UI.
Iβm running the agent in βpipβ mode. I need to fetch certain secret values, which would be best done using Python code. If I incorporate it into the script (repository), others could deduce the path to retrieve the environment or secret value. Storing the environment variables in the clearml.config isnβt very flexible either.
i running this on 2.35 am, but the job not launching after 2.40 am
AgitatedDove14 , yes. i tried 3 times, and it always happen.
Hi @<1523701070390366208:profile|CostlyOstrich36> , thanks for response, sorry for late replay,
this is my configuration in yaml, i facing difficulty when there is params in list. somehow, form to display bunch list not easy to see. do you have suggestion? Thanks!
download-data:
dataset_train:
-
-
-
dataset_test:
-
-
-
train:
data:
batch: 4
input_size: 224
split:
t...
` {"meta":{"id":"17c6e609ace54bf8bfdf3113c39fd470","trx":"17c6e609ace54bf8bfdf3113c39fd470","endpoint":{"name":"projects.get_all_ex","requested_version":"2.20","actual_version":"1.0"},"result_code":200,"result_subcode":0,"result_msg":"OK","error_stack":"","error_data":{}},"data":{"projects":[{"id":"75d04598197a445ebef533814022c58d","company":{"id":"d1bd92a3b039400cbafc60a7a5b1e52b"},"user":{"id":"a174c4e36b0446a7b3b5dd1ff5261962"},"name":"ex-1","basename":"ex-1","description":"","created":"20...
https://github.com/mert-kurttutan/torchview
maybe can try this one, and can send to logger clearml at the end.
the current my solution is upload my config to s3, and the pipeline will download it and read it when execute. but its decrase flexiblity.
Thanks! i just prove it can run in next day, but not for the same day. i hope can run in same day too.
Syncing scheduler
Waiting for next run, sleeping for 5.13 minutes, until next sync.
Launching job: ScheduleJob(name='fetch feedback', base_task_id='', base_function=<function test_make at 0x7f91fd123d90>, queue=None, target_project='Automation/testing', single_instance=False, task_parameters={}, task_overrides={}, clone_task=True, _executed_instances=None, execution_limit_hours=None, r...
you can spesificly use Task.add_requirements and pointing to path requirement.txt
