Examples: query, "exact match", wildcard*, wild?ard, wild*rd
Fuzzy search: cake~ (finds cakes, bake)
Term boost: "red velvet"^4, chocolate^2
Field grouping: tags:(+work -"fun-stuff")
Escaping: Escape characters +-&|!(){}[]^"~*?:\ with \, e.g. \+
Range search: properties.timestamp:[1587729413488 TO *] (inclusive), properties.title:{A TO Z}(excluding A and Z)
Combinations: chocolate AND vanilla, chocolate OR vanilla, (chocolate OR vanilla) NOT "vanilla pudding"
Field search: properties.title:"The Title" AND text
Profile picture
SuccessfulRaven86
Moderator
16 Questions, 63 Answers
  Active since 12 April 2023
  Last activity one year ago

Reputation

0

Badges 1

62 × Eureka!
0 Votes
5 Answers
1K Views
0 Votes 5 Answers 1K Views
Hello, I have the same issue as this github issue: None I tried setting up my AWS autoscaler conf file with the following params: sdk.development.store_uncom...
one year ago
0 Votes
8 Answers
1K Views
0 Votes 8 Answers 1K Views
one year ago
0 Votes
11 Answers
1K Views
0 Votes 11 Answers 1K Views
Hello channel, Two other related questions: - ClearML is supposed to automatically detect GIT repo directly. It works when I run a python script but it does ...
one year ago
0 Votes
1 Answers
1K Views
0 Votes 1 Answers 1K Views
Hi, Can someone give more information about what an API call means? Our team has been charged for 10 Millions API calls, but we struggle to understand where ...
one year ago
0 Votes
5 Answers
1K Views
0 Votes 5 Answers 1K Views
Hello, I am trying to modify my clearml-agent running on a AWS autoscaler (From ClearML applications). I want to be able to clone my repo (working), and inst...
one year ago
0 Votes
3 Answers
836 Views
0 Votes 3 Answers 836 Views
one year ago
0 Votes
0 Answers
1K Views
0 Votes 0 Answers 1K Views
Hey channel, Clearml-serving question Is it good practice to save a .zip file as model, and unzip it in the custom endpoint for usage?
one year ago
0 Votes
3 Answers
821 Views
0 Votes 3 Answers 821 Views
Hello! I have a small question regarding storage data retrieval with ClearML 😉 Context: My team uploads thousands of data samples for training as one ClearM...
one year ago
0 Votes
4 Answers
1K Views
0 Votes 4 Answers 1K Views
one year ago
0 Votes
4 Answers
1K Views
0 Votes 4 Answers 1K Views
one year ago
0 Votes
3 Answers
1K Views
0 Votes 3 Answers 1K Views
Hi channel, I am using K8s clearml-serving helm chart and noticed a small issue. The current implementation of ...ingress.yaml resource does not contain the ...
one year ago
0 Votes
12 Answers
1K Views
0 Votes 12 Answers 1K Views
Can we use the simple docker-compose.yml file for clearml serving on a huggingface model (not processed to tensorrt)?
one year ago
0 Votes
7 Answers
1K Views
0 Votes 7 Answers 1K Views
one year ago
0 Votes
40 Answers
50K Views
0 Votes 40 Answers 50K Views
Hello channel, I am struggling a lot on an issue linked to ClearMl agent and AWS Autoscaler . This issue is very problematic and urgent, please help me out! ...
one year ago
0 Votes
3 Answers
1K Views
0 Votes 3 Answers 1K Views
Hello, Question about the time of upload: Is it faster or exactly the same to upload 1 file of 1Gb compared to 10 files of 100 Mb?
one year ago
0 Votes
1 Answers
1K Views
0 Votes 1 Answers 1K Views
one year ago
0 Hello Channel, Two Other Related Questions:

@<1523701205467926528:profile|AgitatedDove14> If you have any other insights, pls do not hesitate! Thanks a lot

one year ago
0 Hello Channel, I Am Struggling A Lot On An Issue Linked To

Is it a bug inside the AWS autoscaler??

one year ago
0 Hello Channel, Two Other Related Questions:

No problem. I guess this might be a small visualisation bug, but I really have the impression that these workers still pick up tasks, which is strange. I should test again to be sure.

one year ago
0 Hello, I Am Trying To Modify My Clearml-Agent Running On A Aws Autoscaler (From Clearml Applications). I Want To Be Able To Clone My Repo (Working), And Install My Poetry Dependencies From

@<1523701070390366208:profile|CostlyOstrich36> The base docker image of the AWS autoscaler is nvidia/cuda:10.2-runtime-ubuntu18.04 . According to me, the python version is not set inside the image, but I am might be wrong and it could be the problem indeed... ?

one year ago
0 Hello Channel, I Have A Question Regarding Clearml Serving In Production. I Have Different Environments, And Different Models Each Of Them Linked To A Use Case. I Would Like To Spin Up One Kubernetes Cluster (From Triton Gpu Docker Compose) Taking Into

Thanks ! So regarding question2, it means that I can spin up a K8s cluster with triton enabled, and by specifiying the type of model while creating the endpoint, it will use or not the triton engine.
Linked to that, Is the triton engine expecting the tensorrt format or is it just an improvement step compared to other model weights ?

Finally, last question ( I swear 😛 ) : How is the serving on Kubernetes flow supposed to look like? Is it something like that:

  • Create en...
one year ago
0 Hello, I Am Trying To Modify My Clearml-Agent Running On A Aws Autoscaler (From Clearml Applications). I Want To Be Able To Clone My Repo (Working), And Install My Poetry Dependencies From

Thank you for the quick replies!

I might do it the wrong way but the above snippet of code is the additional clearml.conf file I add to the AWS autoscaler. Should I add a complete clearml.conf file to it?

That is a good question @<1537605940121964544:profile|EnthusiasticShrimp49> ! I am not sure the image has python 3.9. I tried to check it but did not find the answer. I am using the following AMI: AWS Deep Learning AMI (Ubuntu 18.04) with Support by Terracloudx (Nvidia deep learni...

one year ago
0 Hello Channel, I Am Struggling A Lot On An Issue Linked To

I guess it makes no sense because of the steps a clearml-agent works...
I also thought about going to pip mode but not all packages are detected from our poetry.lock file unfortunately so cannot do that.

one year ago
0 Hello Channel, I Am Struggling A Lot On An Issue Linked To

This is really extremely hard to debug. I am thinking to create another repo and iterate on the packages to hopefully find the problem, but it will take ages.

one year ago
0 Hello Channel, I Am Struggling A Lot On An Issue Linked To

I am literrally trying with 1 package and python and it fails. I tried with python 3.8 3.9 and 3.9.16. and it always fail --> not linked to python version. What is the problem then? I am wondering if there is not an intrinsic bug

one year ago
0 Hello Channel, I Am Struggling A Lot On An Issue Linked To

I literrally connected to it at runtime, and ran poetry install -n and it worked

one year ago
0 Hello everyone, *Context:* I am currently facing a headache-inducing issue regarding the integration of flash attention V2 for LLM training. I am running a python script locally, that then runs remotely. Without the integration of flash attention, the co

Hi @<1523701087100473344:profile|SuccessfulKoala55> , the EC2 instance is spinned-up from the AWS autoscaler provided by ClearML. I use this following docker image: nvidia/cuda:11.8.0-devel-ubuntu20.0

So the EC2 instance runs a docker container

one year ago
0 Hello Channel, I Am Struggling A Lot On An Issue Linked To

but I still had time to go inside the container, export the PATH variables for my poetry and python versions, and run the poetry install command there

one year ago
one year ago
0 Hi, I Have A Small Question Regarding K8S Clearml-Serving Behavior. I Have In My Cluster One Gpu Of 16Gb Ram, And Another One Of 24 Gb Ram. I Have A Llm Model Fitting The 24Gb But Not The 16Gb Gpu. When I Call The Endpoint, How Will I Know To Which Gpu I

Hey @<1523701205467926528:profile|AgitatedDove14> , thank you for your input
Could you clarify what you mean by clearml-serving session?

Are you refering to the servingTaskId ?

one year ago
0 Can We Use The Simple Docker-Compose.Yml File For Clearml Serving On A Huggingface Model (Not Processed To Tensorrt)?

I basically would like to know if we can serve the model without tensorrt format which is highly efficient but more complicated to get.

one year ago
0 Can We Use The Simple Docker-Compose.Yml File For Clearml Serving On A Huggingface Model (Not Processed To Tensorrt)?

Prerequisites, PyTorch models require Triton engine support, please use docker-compose-triton.yml / docker-compose-triton-gpu.yml or if running on Kubernetes, the matching helm chart.

one year ago
0 Can We Use The Simple Docker-Compose.Yml File For Clearml Serving On A Huggingface Model (Not Processed To Tensorrt)?

I would like to know if it is possible to run any pytorch model on the basic docker compose file ? Without triton?

one year ago
0 Can We Use The Simple Docker-Compose.Yml File For Clearml Serving On A Huggingface Model (Not Processed To Tensorrt)?

Sorry to come back to this! Regarding the Kubernetes Serving helm chart, I can see horyzontal scaling of docker containers. What about vertical scaling? Is it implemented? More specifically, where is defined the SKU of the VMs in use?

one year ago
0 Hey, Can Clearml Uploads Data To Private Blob Storage Azure? I Have Authorization Errors. Is It Due To The Fact That I Do Not Have The Required Permission On The Storage Account (Could Be Possible) Or The Fact That The Storage Account Is Set As Private A

Great, and can we specify an environment variable of ClearML that directly updates the clearml.conf file regarding the azure config or do something similar. I do not want to ask every engineer of my team to modify its clearml.conf file? @<1523701070390366208:profile|CostlyOstrich36> Thanks

one year ago
0 Can We Use The Simple Docker-Compose.Yml File For Clearml Serving On A Huggingface Model (Not Processed To Tensorrt)?

In production, we should use the clearml-helm-charts right? Docker-compose in the clearml-serving is more for local testing

one year ago
Show more results compactanswers