Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Answered
Hi, I Try To Execute Pipeline With Pipelinecontroller And Define It Like This: Pipe = Pipelinecontroller(

Hi, i try to execute pipeline with PipelineController and define it like this:
pipe = PipelineController(
name =config["clear_ml"]["pipeline_name"],
project =config["clear_ml"]["project_name"],
version ="0.1",
add_pipeline_tags =False,
repo =config["clear_ml"]["repo"],
docker =config["clear_ml"]["docker"],
packages =config["clear_ml"]["packages"]
)
And define tasks like this:
pipe.add_function_step(
name ="load_data",
function =load_data,
function_kwargs ={"config": config["data"]},
)
but tasks of this pipeline dont inherit docker and packages, why? I want to build or pull one docker and env for all pipeline steps only once, so ow can i do it?

  
  
Posted one month ago
Votes Newest

Answers 5


I also tried to build my own image and uploaded it to harbor, but the tasks didn't use the packages from it.
FROM nvidia/cuda:12.4.0-devel-ubuntu22.04

ENV TZ=Europe/Moscow

RUN ln -snf /usr/share/zoneinfo/$TZ /etc/localtime && echo $TZ > /etc/timezone

WORKDIR /workspace

RUN apt-get update && apt-get install -y
software-properties-common
curl &&
add-apt-repository ppa:deadsnakes/ppa -y &&
apt-get update && apt-get install -y
python3.11
python3.11-venv
python3.11-dev &&
rm -rf /var/lib/apt/lists/*

RUN curl -sS None | python3.11

RUN update-alternatives --install /usr/bin/python3 python3 /usr/bin/python3.11 1

RUN python3.11 -m venv /opt/venv
ENV VIRTUAL_ENV=/opt/venv
ENV PATH="$VIRTUAL_ENV/bin:$PATH"

COPY . .

RUN pip install --upgrade pip
RUN pip install torch --index-url None
RUN pip install clearml --index-url None

RUN pip install -r requirements.txt --index-url None

  
  
Posted one month ago

yes thanks , but if I do this, the packages will be installed for each step again, is it possible to use a single venv?

Notice that the venv is Cached on the clearml-agent host machine (if this is k8s glue, make sure to setup the Cache as a PV to achieve the same)
This means there is no need to worry about that and this is stable.
That said, if you have an existing VENV inside the container, just add docker_args="-e CLEARML_AGENT_SKIP_PIP_VENV_INSTALL =/path/to/bin/python"
See here: None

BTW: notice that in your example you are probably installing to the system path not the venv path, and to your point specifically, add ENV=CLEARML_AGENT_SKIP_PIP_VENV_INSTALL=$VIRTUAL_ENV

Lastly if you want your agent to Alwoays use the activated Venv inside the container set:

docker_use_activated_venv=true

None

  
  
Posted one month ago

Hi ImmenseMole52

but tasks of this pipeline dont inherit docker and packages, why? I want to build or pull one docker and env for all pipeline steps only once, so ow can i do it?

you have to specify the docker image for the pipeline Tasks, by default it will not assume it is the same as the pipeline controller, basically just pass:

pipe.add_function_step(
        name="load_data",
        function=load_data,
        function_kwargs={"config": config["data"]},
        docker="MY CONTAINER HERE"
    )

make sense?

  
  
Posted one month ago

Thanks!

  
  
Posted one month ago

AgitatedDove14 yes thanks , but if I do this, the packages will be installed for each step again, is it possible to use a single venv? I use pytorch/pytorch:2.6.0-cuda12.4-cudnn9-runtime and install some extra packages

  
  
Posted one month ago