Skip to main content

I’m using the python API to download a batch of images and I get the following error:


DownloadFailedException: Failed to download from:



https://services-uswest2.sentinel-hub.com/api/v1/process

with HTTPError:

500 Server Error: Internal Server Error for url: https://services-uswest2.sentinel-hub.com/api/v1/process

Server response: “{“error”:{“status”:500,“reason”:“Internal Server Error”,“message”:“Illegal request to s3://usgs-landsat/collection02/level-2/standard/tm/2007/190/031/LT05_L2SP_190031_20070314_20200830_02_T2/LT05_L2SP_190031_20070314_20200830_02_T2_SR_B1.TIF. HTTP Status: ‘403’”,“code”:“RENDERER_EXCEPTION”}}”



before calling the API I’m doing a search to see what the catalog has. and this date exists in the catalogue:



datetime.datetime(2007, 3, 14, 9, 42, 9, 464000, tzinfo=tzutc()),



a few other datapoints also have an issue. for example:



DownloadFailedException: Failed to download from:
https://services-uswest2.sentinel-hub.com/api/v1/process

with HTTPError:

500 Server Error: Internal Server Error for url: https://services-uswest2.sentinel-hub.com/api/v1/process

Server response: “{“error”:{“status”:500,“reason”:“Internal Server Error”,“message”:“Illegal request to s3://usgs-landsat/collection02/level-2/standard/tm/2007/191/031/LT05_L2SP_191031_20070711_20200830_02_T1/LT05_L2SP_191031_20070711_20200830_02_T1_SR_B1.TIF. HTTP Status: ‘403’”,“code”:“RENDERER_EXCEPTION”}}”



I’m going over my list from 2005 to 2011 and removing the bad connections.

in this coordinates (a very small area):



<13.110,41.8185,13.1195,41.8275]



my code works for other dates. I don’t mind to miss those few images, but was wondering if the problem is on my end or the server/catalogue?

update:

the full list of “faulty” dates:

filtereddates_remove=o datetime.datetime(2007, 3, 14, 9, 42, 9, 464000, tzinfo=tzutc()),

datetime.datetime(2007, 7, 11, 9, 47, 5, 138000, tzinfo=tzutc()),

datetime.datetime(2007, 8, 12, 9, 46, 46, 151000, tzinfo=tzutc()),

datetime.datetime(2008, 8, 30, 9, 38, 53, 645000, tzinfo=tzutc()),]


once those 4 are removes, I can extract the data for the entire duration.


Hi @dataTfS ,


Thank you for reporting. We’re investigating and will get back to you soon.


Best Regards


Hi @dataTfS ,


It turns out that the error came from the data provider. We’ve reported the issue to USGS.


Best Regards


oh nice.

thanks for the update


Reply