Add exponential backoff for retrying
This commit is contained in:
parent
f47895fd46
commit
b67ab06b55
@ -3,6 +3,7 @@ from urllib.parse import urlparse
|
|||||||
from time import strptime, mktime
|
from time import strptime, mktime
|
||||||
import dateutil.parser
|
import dateutil.parser
|
||||||
from pyquery import PyQuery as pq
|
from pyquery import PyQuery as pq
|
||||||
|
from retrying import retry
|
||||||
|
|
||||||
time_format = "%b %d %y %I:%M %p"
|
time_format = "%b %d %y %I:%M %p"
|
||||||
|
|
||||||
@ -18,6 +19,7 @@ def scrape (url):
|
|||||||
elif (not path) or path == "/":
|
elif (not path) or path == "/":
|
||||||
return scrape_index(url)
|
return scrape_index(url)
|
||||||
|
|
||||||
|
@retry(wait_exponential_multiplier=1000, wait_exponential_max=10000)
|
||||||
def scrape_index (url):
|
def scrape_index (url):
|
||||||
print("Scraping forum index from url: {}".format(url))
|
print("Scraping forum index from url: {}".format(url))
|
||||||
urlparts = urlparse(url)
|
urlparts = urlparse(url)
|
||||||
@ -35,6 +37,7 @@ def scrape_index (url):
|
|||||||
|
|
||||||
return forum
|
return forum
|
||||||
|
|
||||||
|
@retry(wait_exponential_multiplier=1000, wait_exponential_max=10000)
|
||||||
def scrape_board (url):
|
def scrape_board (url):
|
||||||
print("Scraping board from url: {}".format(url))
|
print("Scraping board from url: {}".format(url))
|
||||||
urlparts = urlparse(url)
|
urlparts = urlparse(url)
|
||||||
@ -57,6 +60,7 @@ def scrape_board (url):
|
|||||||
|
|
||||||
return board
|
return board
|
||||||
|
|
||||||
|
@retry(wait_exponential_multiplier=1000, wait_exponential_max=10000)
|
||||||
def scrape_thread (url):
|
def scrape_thread (url):
|
||||||
print("Scraping thread from url: {}".format(url))
|
print("Scraping thread from url: {}".format(url))
|
||||||
urlparts = urlparse(url)
|
urlparts = urlparse(url)
|
||||||
|
Loading…
x
Reference in New Issue
Block a user