Delete wrapper.py
This commit is contained in:
parent
ea3913477c
commit
0c7be24235
@ -1,81 +0,0 @@
|
|||||||
# -*- coding: utf-8 -*-
|
|
||||||
|
|
||||||
from datetime import datetime
|
|
||||||
from urllib.request import Request, urlopen
|
|
||||||
import urllib.error
|
|
||||||
|
|
||||||
class TooManyArchivingRequestsError(Exception):
|
|
||||||
"""
|
|
||||||
Error when a single url reqeusted for archiving too many times in a short timespam.
|
|
||||||
Wayback machine doesn't supports archivng any url too many times in a short period of time.
|
|
||||||
"""
|
|
||||||
|
|
||||||
class ArchivingNotAllowed(Exception):
|
|
||||||
"""
|
|
||||||
Files like robots.txt are set to deny robot archiving.
|
|
||||||
Wayback machine respects these file, will not archive.
|
|
||||||
"""
|
|
||||||
|
|
||||||
class PageNotSavedError(Exception):
|
|
||||||
"""
|
|
||||||
Files like robots.txt are set to deny robot archiving.
|
|
||||||
Wayback machine respects these file, will not archive.
|
|
||||||
"""
|
|
||||||
|
|
||||||
class InvalidUrlError(Exception):
|
|
||||||
"""
|
|
||||||
Files like robots.txt are set to deny robot archiving.
|
|
||||||
Wayback machine respects these file, will not archive.
|
|
||||||
"""
|
|
||||||
|
|
||||||
def save(url,UA="pywayback python module"):
|
|
||||||
base_save_url = "https://web.archive.org/save/"
|
|
||||||
request_url = base_save_url + url
|
|
||||||
hdr = { 'User-Agent' : '%s' % UA }
|
|
||||||
req = Request(request_url, headers=hdr)
|
|
||||||
if "." not in url:
|
|
||||||
raise InvalidUrlError("'%s' is not a vaild url." % url)
|
|
||||||
try:
|
|
||||||
response = urlopen(req) #nosec
|
|
||||||
except urllib.error.HTTPError as e:
|
|
||||||
if e.code == 502:
|
|
||||||
raise PageNotSavedError(e)
|
|
||||||
elif e.code == 429:
|
|
||||||
raise TooManyArchivingRequestsError(e)
|
|
||||||
|
|
||||||
header = response.headers
|
|
||||||
if "exclusion.robots.policy" in str(header):
|
|
||||||
raise ArchivingNotAllowed("Can not archive %s. Disabled by site owner." % (url))
|
|
||||||
archive_id = header['Content-Location']
|
|
||||||
archived_url = "https://web.archive.org" + archive_id
|
|
||||||
return archived_url
|
|
||||||
|
|
||||||
def near(
|
|
||||||
url,
|
|
||||||
year=datetime.utcnow().strftime('%Y'),
|
|
||||||
month=datetime.utcnow().strftime('%m'),
|
|
||||||
day=datetime.utcnow().strftime('%d'),
|
|
||||||
hour=datetime.utcnow().strftime('%H'),
|
|
||||||
minute=datetime.utcnow().strftime('%M'),
|
|
||||||
UA="pywayback python module",
|
|
||||||
):
|
|
||||||
timestamp = str(year)+str(month)+str(day)+str(hour)+str(minute)
|
|
||||||
request_url = "https://archive.org/wayback/available?url=%s×tamp=%s" % (str(url), str(timestamp))
|
|
||||||
hdr = { 'User-Agent' : '%s' % UA }
|
|
||||||
req = Request(request_url, headers=hdr)
|
|
||||||
response = urlopen(req) #nosec
|
|
||||||
encoding = response.info().get_content_charset('utf8')
|
|
||||||
import json
|
|
||||||
data = json.loads(response.read().decode(encoding))
|
|
||||||
print(data)
|
|
||||||
if not data["archived_snapshots"]:
|
|
||||||
raise PageNotSavedError("'%s' was not archived." % url)
|
|
||||||
|
|
||||||
archive_url = (data["archived_snapshots"]["closest"]["url"])
|
|
||||||
return archive_url
|
|
||||||
|
|
||||||
def oldest(url,UA="pywayback python module"):
|
|
||||||
return near(url,year=1995,UA=UA)
|
|
||||||
|
|
||||||
def newest(url,UA="pywayback python module"):
|
|
||||||
return near(url,UA=UA)
|
|
Loading…
Reference in New Issue
Block a user