Hi everyone,
I’ve been working on a project that involves heavy SPARQL querying against the DBpedia endpoint for entity extraction, but I’ve hit a bit of a wall regarding consistency. While my queries usually run fine, I’ve noticed that I’m getting frequent Virtuoso S1T00 Error SR171: Transaction timed out messages during specific times of day.
Lately, I’ve been running some local automation and testing my app’s resilience using a roblox exector in a separate container to simulate high-load background activity. Interestingly, it seems that when my local machine is under heavy script execution stress, the HTTP 500 errors and timeouts from the DBpedia endpoint become much more frequent, even for relatively simple SELECT queries that usually resolve in under 500ms.
Has anyone else noticed their connection to the public SPARQL endpoint becoming more fragile when there are other intensive background executors active on the same network or device? I’m trying to determine if the issue is client-side socket throttling or if I should be implementing a more aggressive retry-with-backoff logic in my Python scripts to handle these timeouts.
Is there a recommended setting for the SPARQLWrapper timeout or perhaps a way to better isolate my network requests from other system-wide script execution to ensure the DBpedia responses don’t get dropped? I’d love to hear how others manage stable data extraction while multitasking with high-resource tools.