Turns out the bucket
param expected was expecting the bucket name without the s3://
protocol specification, but now that this issue is fixed i still have the same incorrect region specified error
,
task = Task.init( project_name='XXXX', task_name=f'Training-{training_uuid}', task_type=Task.TaskTypes.training, output_uri=f's3://{constants.CLEARML_BUCKET}' ) task.setup_aws_upload( bucket=constants.CLEARML_BUCKET, region=constants.AWS_REGION, )
Despite having overriden the region with the .setup_aws_upload()
method
@<1523701087100473344:profile|SuccessfulKoala55> I had already bumped boto3 to its latest version and all the files I added to the dataset were pickle binary files
ClearML package version used: 1.9.1
ClearML Server: SaaS - Pro Tier
FierceHamster54 in all reports I've seen concerning a similar error, it seems the problem was when trying to copy a folder from/to a bucket without specifying the recursive option (some were with the aws cli, but I suspect this is similar). Since the setup_aws_upload()
attempts to upload a file, I suspect there's some confusion and it tries somehow to upload a folder, but does that as if it was a file... the actual implementation is providing boto3 with a Key and a Body, which should be treated as a file (although the key has no suffix) - perhaps upgrading boto will help?
Okay, turns out the output_uri
in the constuctor was overriding the .setup_aws_upload()
and not the other way arround