
Reputation
Badges 1
25 × Eureka!The thing is, even on the community server, not all the datasets have automatic previews. So for the same code/dataset, some of the runs have previews and some of them don't.
As you can see, it eventually manages to reach the apiserver
however, it still says that access key was not provided and that the service will not be started. I get the same behaviour whether I set the flag to 0 or 1.
Yes, I am using a virtualenv that has pandas and clearml installed.
The above output is on the clearml community server
Thank you for your reply SuccessfulKoala55 . 😄
It is currently set to 1, so I am assuming setting it to 0 would mute the errors from logging?
The current behaviour is, if I keep it set to 1 the services agent would automatically shutdown if the access key is not configured. Assuming I set it to 0, then the agent services should not shutdown anymore, right?
@<1523701070390366208:profile|CostlyOstrich36> I am facing the same issue:
{"meta":{"id":"90841d05dfb1431a8d9dfc6bfdb39f9e","trx":"90841d05dfb1431a8d9dfc6bfdb39f9e","endpoint":{"name":"events.debug_images","requested_version":"2.23","actual_version":"2.7"},"result_code":200,"result_subcode":0,"result_msg":"OK","error_stack":"","error_data":{}},"data":{"metrics":[]}}
Let me rerun it, so that I can capture it. I am currently running it on AWS Fargate, so I have the logs for that.
I have been using them for a while, they've always had autopreviews.
I am using the latest version clearml server
and I am using version 1.9.1
for the sdk.
Here is the code that I am currently using:
if __name__ == "__main__":
# create clearml data processing task
dataset = Dataset.create(
dataset_name="palmer_penguins",
dataset_project="palmer penguins",
dataset_tags=["raw"]
)
dataset_path = "data/raw/penguins.csv"
# add the downloaded files to the current dataset
dataset.add_files(path=dataset_pa...
Thanks @<1523701205467926528:profile|AgitatedDove14>
Which would make sense because of the name SHUTDOWN_IF_NO_ACCESS_KEY
. The thing is, when I tried setting it to 0, it still shutdown.
I was able to resolve the issue. I am currently using clearml on wsl2 and my machine is connected to a vpn that allows me to connect on to the clearml instance hosted on AWS. You were right it was a network issue, I was able to resolve it by modifying my /etc/resolv.conf
file.
Right so I figured out why it was calling it multiple times. Everytime a dataset is serialiazed, it calls the _serialize()
function inside of clearml/datasets/dataset.py
file, the _serialize()
method calls self.get(parent_dataset_id)
which is the same get()
method. This means that the user will always be prompted with the log, even if they are not "getting" a dataset. So anytime a user creates, uploads, finalizes a dataset, they will be prompted with the message...
Not exactly, the dataset gets called in the script using Dataset.get() and the second dataset is an output dataset using Dataset.create().. Which means that dataset_1 is a parent dataset of dataset_2.
Hi AgitatedDove14 ,
I am planning to use terraform to retrieve the secrets from AWS, after I retrieve the user list from the secrets manager, I am going to pass them as Environment variables.
The reason I am passing them as environment variables is that, I couldn't find a way to automatically upload files to AWS EFS from Terraform. Since the config file needs to be mounted as an EFS volume to the ECS task definition.
I was able to make the web authentication work while passing the followi...
SuccessfulKoala55 That seemed to do the trick, thanks for your help! 😄
Just waiting for the changes to be completed
I would like to see it used in a clear example as it was intended to be used before giving my opinion on it, if that makes sense
Thank you so much for your reply, will give that a shot!
I am currently running the scripts on WSL ubuntu
So what's the point of the alias? It's not very clear.. Even after specifying an alias I am still getting the following message: Dataset.get() did not specify alias. Dataset information will not be automatically logged in ClearML Server
Thanks for the reply. I was trying out this feature on a dummy example. I used the following commanddataset = Dataset.get( dataset_project="palmer penguins", dataset_name="raw palmer penguins", alias="my_test_alias_name", overridable=True)
That was the only time I called the get()
command. I still got the message that I should specify the alias. I can try and do a bit of debugging to see why it gets called multiple times.
My server is hosted on AWS Fargate
I'm actually trying that as we speak 😛
Thanks for the reply anyways 😄
Above is the response for the events.debug_images