Unanswered
Hello Everyone! I'M Encountering An Issue When Trying To Deploy An Endpoint For A Large-Sized Model Or Get Inference On A Large Dataset (Both Exceeding ~100Mb). It Seems That They Can Only Be Downloaded Up To About 100Mb. Is There A Way To Increase A Time
Oh...
None
try to add to your config file:
sdk.http.timeout.total = 300
142 Views
0
Answers
one year ago
one year ago