waybackpy/waybackpy/wrapper.py

83 lines
2.6 KiB
Python
Raw Normal View History

2020-05-04 04:35:52 +02:00
# -*- coding: utf-8 -*-
from datetime import datetime
from urllib.request import Request, urlopen
import urllib.error
default_UA = "waybackpy python package"
class TooManyArchivingRequestsError(Exception):
"""
Error when a single url reqeusted for archiving too many times in a short timespam.
Wayback machine doesn't supports archivng any url too many times in a short period of time.
"""
class ArchivingNotAllowed(Exception):
"""
Files like robots.txt are set to deny robot archiving.
Wayback machine respects these file, will not archive.
"""
class PageNotSavedError(Exception):
"""
Files like robots.txt are set to deny robot archiving.
Wayback machine respects these file, will not archive.
"""
class InvalidUrlError(Exception):
"""
2020-05-04 04:43:54 +02:00
When the inpute url doesn't have '.' in it.
2020-05-04 04:35:52 +02:00
"""
def save(url,UA=default_UA):
base_save_url = "https://web.archive.org/save/"
request_url = base_save_url + url
hdr = { 'User-Agent' : '%s' % UA }
req = Request(request_url, headers=hdr)
if "." not in url:
raise InvalidUrlError("'%s' is not a vaild url." % url)
try:
response = urlopen(req) #nosec
except urllib.error.HTTPError as e:
if e.code == 502:
raise PageNotSavedError(e)
elif e.code == 429:
raise TooManyArchivingRequestsError(e)
header = response.headers
if "exclusion.robots.policy" in str(header):
raise ArchivingNotAllowed("Can not archive %s. Disabled by site owner." % (url))
archive_id = header['Content-Location']
archived_url = "https://web.archive.org" + archive_id
return archived_url
def near(
url,
year=datetime.utcnow().strftime('%Y'),
month=datetime.utcnow().strftime('%m'),
day=datetime.utcnow().strftime('%d'),
hour=datetime.utcnow().strftime('%H'),
minute=datetime.utcnow().strftime('%M'),
UA=default_UA,
):
timestamp = str(year)+str(month)+str(day)+str(hour)+str(minute)
request_url = "https://archive.org/wayback/available?url=%s&timestamp=%s" % (str(url), str(timestamp))
hdr = { 'User-Agent' : '%s' % UA }
req = Request(request_url, headers=hdr)
response = urlopen(req) #nosec
encoding = response.info().get_content_charset('utf8')
import json
data = json.loads(response.read().decode(encoding))
print(data)
if not data["archived_snapshots"]:
raise PageNotSavedError("'%s' was not archived." % url)
archive_url = (data["archived_snapshots"]["closest"]["url"])
return archive_url
def oldest(url,UA=default_UA):
return near(url,year=1995,UA=UA)
def newest(url,UA=default_UA):
return near(url,UA=UA)