Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Answered
Hey,

Hey,
I'm trying to build a clearml pipeline using decorators. My training script includes two modules(.py files) that's stored locally. This however causes an error with clearml as I suppose it creates separate environments for each function. Is there any solution to include local .py files? (Apart from explicitly creating a module)

  
  
Posted 2 years ago
Votes Newest

Answers 31


can you share with me an example or part from your code ? I might miss something in wht you intend to achieve

  
  
Posted 2 years ago

btw here is the content of the imported file:

import torch
from torchvision import datasets, transforms
import os
MY_GLOBAL_VAR = 32

def my_dataloder ():
return torch.utils.data.DataLoader(
datasets.MNIST(os.path.join('./', 'data'), train=True, download=True,
transform=transforms.Compose([
transforms.ToTensor()
])),
batch_size=32, shuffle=True)

  
  
Posted 2 years ago

With pleasures ! Hope that will help

  
  
Posted 2 years ago

No, it is supposed to have its status updated automatically. We may have a bug. Can you share some example code with me, so that i could try to figure out what is happening here ?

  
  
Posted 2 years ago

Yep, the pipeline finishes but the status is still at running . Do we need to close a logger that we use for scalers or anything?

  
  
Posted 2 years ago

Hey We figured a temporary solution - by importing the modules and reloading the contents of the artefact by pickle. It still gives us a warning, though training works now. Do send an update if you find a better solution

  
  
Posted 2 years ago

How do we close pipelinedecorators?
It is showing running even after pipeline was completed

  
  
Posted 2 years ago

Though as per your docs the add_requirements is for a requirements .txt

  
  
Posted 2 years ago

stuff is a package that has my local modules - I've added it to my path by sys.path.insert, though here it isn't able to unpickle

  
  
Posted 2 years ago

Thanks a lot David!

  
  
Posted 2 years ago

can you share the logs please ?

  
  
Posted 2 years ago

Not local .py files

  
  
Posted 2 years ago

have you tried to add the requirements using Task.add_requirements( local_packages ) in your main file ?

  
  
Posted 2 years ago

I'm facing the same issue, is there any solution to this?

  
  
Posted 2 years ago

image

  
  
Posted 2 years ago

However, I use this to create an instance of a dataloader(torch) this is fed into my next stage in the pipeline - though I import the local modules and add the folders to the path it is unable to unpickle the artifact

  
  
Posted 2 years ago

Nop e

  
  
Posted 2 years ago

Sure, in a moment

  
  
Posted 2 years ago

Umm I suppose that won't work - this package consists of .py scripts that I use for a set of configs and Utils for my model.

  
  
Posted 2 years ago

Super sorry for being a bit late!

  
  
Posted 2 years ago

How would you structure PyTorch pipelines in clearml? Especially dealing with image data

  
  
Posted 2 years ago

Hey so I was able to get the local .py files imported by adding the folder to my path sys .path

  
  
Posted 2 years ago

TenderCoyote78
the status should normally be automatically updated . Do all the steps finish successfully ? And also the pipeline ?

  
  
Posted 2 years ago

I tried it - it works for a library that you can install, not for something local I suppose

  
  
Posted 2 years ago

you can also specify a package, with or without specifying its version
https://clear.ml/docs/latest/docs/references/sdk/task#taskadd_requirements

  
  
Posted 2 years ago

Thank you so much for being active!

  
  
Posted 2 years ago

Thanks a lot David!

  
  
Posted 2 years ago

Is there a way to store the return values after each pipeline stage in a format other than pickle?

  
  
Posted 2 years ago

Here's the code, we're trying to make a pipeline using PyTorch so the first step has the dataset that ’ s created using ‘stuff’ - a local folder that serves as a package for my code. The issue seems to be in the unpicking stage in the train function.

  
  
Posted 2 years ago

hey WickedElephant66 TenderCoyote78
I'm working on a solution, just hold on, I update you asap

  
  
Posted 2 years ago
11K Views
31 Answers
2 years ago
3 months ago
Tags