Hi RoughTiger69
A. Yes makes total sense . Basically you can use Task.export Task.import to do achieve this process (notice we assume the dataset artifacts links are available on both, usually this is the case)
B. The easiest way would be to use Process , then one subprocess is exporting from dev , where the credentials and configuration is passed with os environment. The another subprocess imports it to the prod server (again with os environment pointing to the prod server). Make sense?
RoughTiger69 Hi 🙂
Regarding your questions:
Moving certain tasks/datasets from server to server would require a data merge. This process basically requires merging the databases (mongodb, elasticsearch, files etc.). I think it's something that can be done in the paid version as a service but not in the free. I think if you'd like to 'promote' tasks to production you can either work on a special project for that OR upload the models to S3 and then re-run the experiment and point it to the model Currently there is option to pass tasks between different servers. The agent can connect to any server depending on configuration but not pass data between them. I think you'd have to register the pipeline in prod "from scratch" for now