I did notice that the last 24 hours I dropped quite a bit, so my theory that the 140K might have some spillover from previous day might have been correct. Last 24 hours went from 1.24M to 1.32M, so about half as much as the day before, with the same training running.
well, in my case, if I am trying to make sure I do not go over the allotted usage, it matters, as I am already hitting the ceiling and I have no idea what is pushing this volume of data
FYI, I did not even know to look into this until I logged in and saw that I was being throttled because I had hit my monthly limit with API calls ( on my very first use of your platform ), and my last dozen or so epochs were just not even logged ( also a bummer ). I only had that one model in training, and thought there was no way I sent over a million API requests, so had to figure out where those were coming from, and tracked it down to that STDOUT, and was like ... wait, what?!?! Found that console tab, which I did not even use before, and saw that screenshot I posted, and was like ... well, there's your problem, ha ha
Welp, it's been a day with the new settings, and stats went up 140K for API calls
... going to check again tomorrow to see if any of that was spill over from yesterday
140K calls a day, how often are you sending scalars ? how long is it running? how many experiments are running ?
Hmm if this is case, you can add some prints in here:
None
the service/action will tell you what you are sending
wdyt?
Just wish I could actually see somewhere what is being sent over API so I could know where to focus my efforts to refine this kind of stuff 😉
Maybe ClearML is using  tensorboard  in ways that I can fine tune? I saw there was a manual way if you were not using  tensorboard  to send over data, but the videos I saw from your team used this solution when demoing YOLOv8 on YouTube ( there were a few collab videos your team did with theirs, so I just followed their instructions ).  But my gut is telling me that might be the issue for the remaining data being sent over that I have no insight into.
@<1572395184505753600:profile|GleamingSeagull15>  see " Can I control what ClearML automatically logs? " in  None  (specifically the  auto_connect_frameworks argument to  Task.init() )
Since it's literally something we have to pay for ( which I signed up to do ) I would love to know what drives this cost
It was at 1.1M when I shut it down yesterday, and today it's at 1.24M
@<1523701087100473344:profile|SuccessfulKoala55> You are my hero !!! This is EXACTLY what I needed !!!
each epoch runs about 55 minutes, and that screenshot I posted earlier kind of show the logs for the rest of the info being output, if you wanted to check that out None
(Not sure it actually has that information)
Is there a place in ClearML that shows Platform Usage? Like, what's actually taking up the API calls?
Scary to think how common that might be, could be interesting way to optimize your platform, detect excessive console logging and prompt user to confirm continued usage ( or link to docs on how to disable if they want to stop it )
If you do not have a lot of workers, that I would guess console outputs
But I will try to set the reduce the number of log reports first
I appreciate your help @<1523701205467926528:profile|AgitatedDove14> 🙂
well from 2 to 30sec is a factor of 15, I think this is a good start 🙂
