Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
AgitatedDove14
Moderator
48 Questions, 8051 Answers
  Active since 10 January 2023
  Last activity 8 months ago

Reputation

0

Badges 1

25 × Eureka!
0 Regarding The New Version 1.1.2, I Have Noticed Type Hints Are Now Included In The Script Generated By

Looks great, let me see if I can understand what's missing, because it should have worked ...

3 years ago
0 I Originally Posted In

Hi LackadaisicalOtter14

However, whenever we spin up a session,Β 

Β always gets run and overwrites our configs

what do you mean by that?
The what config are being overwritten? (generally speaking, it just add the OS environment it needs to for the setup process)

2 years ago
0 Hi I Wanted To Use Method Task.Reset() Or Task.Delete() However None Of That Seems To Be Able To Delete

I want to be able to delete only the logs since they are taking a lot of space in my case.

I see... I do not think this is possible 😞
You can disable the auto logging though ... pass auto_connect_streams=False to Task.init

one year ago
0 Hi Everyone, I Have Questions Related To Clearml-Serving.

It said the command --aux-config got invalid input

This seems like an interface bug.. let me see if we can fix that πŸ™‚

BTW: this seems like a triton LSTM configuration issue, we might want to move the discussion to the Triton server issue, wdyt?

Definitely!

Could you start an issue https://github.com/triton-inference-server/server/issues , and I'll jump join the conversation?

. Is there any reference about integrating kafka data streaming directly to clearml-serving...

2 years ago
4 years ago
0 Hi Everyone, I Have Questions Related To Clearml-Serving.

I have timeseries dataset with dimension 1,60,1 which the first dimension is number of data, the second one is timestep

I think it should be --input-size 1 60 ` if the last dimension is the batch size?
(BTW: this goes directly to Triton configuration, it is the information Triton needs in order to run the model itself)

2 years ago
0 Hi Everyone, I Have Questions Related To Clearml-Serving.

Hmm, how does your preprocessing code looks like?

2 years ago
2 years ago
0 Hey, We Were Trying To Run An Experiment On Clearml Using Its Python-Sdk. When I Run An Experiment Using

Right, you need to pass "repo" and direct it to the repository path
(BTW, what's the cleaml version)

3 years ago
0 Hi Everyone, I Have Questions Related To Clearml-Serving.

If there is new issue will let you know in the new thread

Thanks! I would really like to understand what is the correct configuration

2 years ago
0 Hi Everyone, I Have Questions Related To Clearml-Serving.

the trend step artifact used to keep track the time of the data so we know the expected trend of the input data. For example, on the first data which is trend_step = 1 the trend value is 10, then if the trend_step = 10 (the tenth data) our regressor will predict the trend value of the selected trend_step. this method is still in research to make it more efficient so it doesn't need to upload artifact every request

Make sense! I would suggest you add a GitHub issue with feature request ...

2 years ago
0 Hi Everyone, I Have Questions Related To Clearml-Serving.

EmbarrassedPeacock82 are you using keras/pytorch etc for serving (i.e. Triton) ?

2 years ago
0 Unrelated Problem (Or Is It?) The Clearml'S Built In Cleanup Service Fails

Very odd, I still can't reproduce. This is just the cleanup service running without anything else ?
What's the clearml version it is using ?

2 years ago
0 Hi

Hi IrritableJellyfish76
If you are running a code that uses clearml from kubeflow, you have out of the box integration between the two, what am I missing?

2 years ago
0 Hi

No sure I follow, you mean to launch it on the kubernretes cluster from the ClearML UI?
(like the clearml-k8s-glue ?)

2 years ago
0 Getting This Error At

Hi TrickySheep9
Hmm I think you are correct, exit remotely will not work inside a jupyter notebook because it will not be able to close it.

I was just revising workflows that might be similar, wdyt?
https://clearml.slack.com/archives/CTK20V944/p1620506210463400?thread_ts=1614234125.066600&cid=CTK20V944

3 years ago
0 Getting This Error At

I just called exit(0) in a notebooke and it closed it (the kernel) no exception

3 years ago
0 Getting This Error At

This is a Sagemaker notebook instances

Yes I think this is the issue

3 years ago
0 Getting This Error At

Jupyter server v 6.0.3

3 years ago
0 Getting This Error At

You cannot call exit(0) and kill the kernel from the SageMake notebook

3 years ago
0 Getting This Error At

Just a bit of background, the execute)remotely will kill the current process (after the Task is synced) and enqueue the Task that was created for remote execution. What seems to fail is actually killing the current process. You can just pass exit_process=False

3 years ago
0 Getting This Error At

No worries πŸ™‚

3 years ago
0 Hello Everyone! A Question Regarding Uploading Model Weights As Artifacts. I Use

Hi @<1684010629741940736:profile|NonsensicalSparrow35>

however for the remote file it always creates the name with the following pattern:

{filename_prefix}checkpoint{n}.pt

..

Is this the main issue?
Notice that the model name (i.e. the entry on the Task itself) is not directly connected with the stored file name on the target file server (or S3)

8 months ago
0 Hi There, I Have A Pipeline That Query Data From A Neo4J Database. When I Run It Using

Hi IrritableGiraffe81
PipelineDecorator.debug_pipeline() runs everything as regular python functions, but "PipelineDecorator.run_locally()" is actually sumulating all the steps on the same local machine (so that it is easier to debug the "real" pipeline running on multiple machines)
What I think is happening is that the casting of the arguments passed to the component fail.
Basically the type hints are currently ignored (we are working on using them for casting in the next version)
but righ...

2 years ago
0 Hello, Community. I Hope You Are All Doing Well. I'M Seeking Information Regarding A Specific Problem, Specially In The Field Of Computer Vision. Typically, An App In The Field Of Computer Vision Will Have Multiple Models, Each With Its Own Preprocessing,

You mean to add these two to the model when deploying?

    β”‚   β”œβ”€β”€ model_NVIDIA_GeForce_RTX_3080.plan
    β”‚   └── model_Tesla_T4.plan

Notice the preprocess.py is Not running on the GPU instance, it is running on a CPU instance (technically not the same machine)

10 months ago
0 So, I Have Just Started Using Clearml For Local Data And Experiment Tracking And Its Been Super Helpful. Now That I Am Moving Towards Deploying And Serving The Models Using Clearml-Serving And Triton. I Have Done Some Basic Experimenting With The Provided

Hi RipeAnt6

What would be the best way to add another model from another project say C to the same triton server serving the previous model?

You can add multiple call to cleaml-serving , each one with a new endpoint and a new project/model to watch, then when you launch it it will setup all endpoints on a single Triton server (the model optimization loading is taken care by Triton anyhow)

3 years ago
Show more results compactanswers