mirror of
https://github.com/internetarchive/brozzler.git
synced 2025-08-07 22:12:15 -04:00
Merge branch 'pageInterstitialShown' into qa
This commit is contained in:
commit
60cfd684b2
7 changed files with 67 additions and 13 deletions
|
@ -58,6 +58,28 @@ class ReachedLimit(Exception):
|
||||||
def __str__(self):
|
def __str__(self):
|
||||||
return self.__repr__()
|
return self.__repr__()
|
||||||
|
|
||||||
|
class PageInterstitialShown(Exception):
|
||||||
|
def __init__(self, http_error=None, warcprox_meta=None, http_payload=None):
|
||||||
|
import json
|
||||||
|
if http_error:
|
||||||
|
if "warcprox-meta" in http_error.headers:
|
||||||
|
self.warcprox_meta = json.loads(
|
||||||
|
http_error.headers["warcprox-meta"])
|
||||||
|
else:
|
||||||
|
self.warcprox_meta = None
|
||||||
|
self.http_payload = http_error.read()
|
||||||
|
elif warcprox_meta:
|
||||||
|
self.warcprox_meta = warcprox_meta
|
||||||
|
self.http_payload = http_payload
|
||||||
|
|
||||||
|
def __repr__(self):
|
||||||
|
return "PageInterstitialShown(warcprox_meta=%r,http_payload=%r)" % (
|
||||||
|
self.warcprox_meta if hasattr(self, 'warcprox_meta') else None,
|
||||||
|
self.http_payload if hasattr(self, 'http_payload') else None)
|
||||||
|
|
||||||
|
def __str__(self):
|
||||||
|
return self.__repr__()
|
||||||
|
|
||||||
# monkey-patch log levels TRACE and NOTICE
|
# monkey-patch log levels TRACE and NOTICE
|
||||||
logging.TRACE = (logging.NOTSET + logging.DEBUG) // 2
|
logging.TRACE = (logging.NOTSET + logging.DEBUG) // 2
|
||||||
def _logger_trace(self, msg, *args, **kwargs):
|
def _logger_trace(self, msg, *args, **kwargs):
|
||||||
|
|
|
@ -241,10 +241,12 @@ class WebsockReceiverThread(threading.Thread):
|
||||||
if self.on_request:
|
if self.on_request:
|
||||||
self.on_request(message)
|
self.on_request(message)
|
||||||
elif message['method'] == 'Page.interstitialShown':
|
elif message['method'] == 'Page.interstitialShown':
|
||||||
# for AITFIVE-1529: handle http auth
|
# AITFIVE-1529: handle http auth
|
||||||
# for now, we should consider killing the browser when we receive Page.interstitialShown and
|
# we should kill the browser when we receive Page.interstitialShown and
|
||||||
# consider the page finished—-first we should figure out when else that event might happen
|
# consider the page finished, until this is fixed:
|
||||||
self.logger.info('Page.interstitialShown received')
|
# https://bugs.chromium.org/p/chromium/issues/detail?id=764505
|
||||||
|
self.logger.info('Page.interstialShown (likely unsupported http auth request)')
|
||||||
|
brozzler.thread_raise(self.calling_thread, brozzler.PageInterstitialShown)
|
||||||
elif message['method'] == 'Inspector.targetCrashed':
|
elif message['method'] == 'Inspector.targetCrashed':
|
||||||
self.logger.error(
|
self.logger.error(
|
||||||
'''chrome tab went "aw snap" or "he's dead jim"!''')
|
'''chrome tab went "aw snap" or "he's dead jim"!''')
|
||||||
|
|
|
@ -199,6 +199,8 @@ def brozzle_page(argv=None):
|
||||||
logging.info('outlinks: \n\t%s', '\n\t'.join(sorted(outlinks)))
|
logging.info('outlinks: \n\t%s', '\n\t'.join(sorted(outlinks)))
|
||||||
except brozzler.ReachedLimit as e:
|
except brozzler.ReachedLimit as e:
|
||||||
logging.error('reached limit %s', e)
|
logging.error('reached limit %s', e)
|
||||||
|
except brozzler.PageInterstitialShown as e:
|
||||||
|
logging.error('page interstitial shown %s', e)
|
||||||
finally:
|
finally:
|
||||||
browser.stop()
|
browser.stop()
|
||||||
|
|
||||||
|
|
|
@ -209,9 +209,13 @@ class BrozzlerWorker:
|
||||||
|
|
||||||
if self._needs_browsing(page, ydl_fetches):
|
if self._needs_browsing(page, ydl_fetches):
|
||||||
self.logger.info('needs browsing: %s', page)
|
self.logger.info('needs browsing: %s', page)
|
||||||
|
try:
|
||||||
outlinks = self._browse_page(browser, site, page, on_screenshot,
|
outlinks = self._browse_page(browser, site, page, on_screenshot,
|
||||||
on_request)
|
on_request)
|
||||||
return outlinks
|
return outlinks
|
||||||
|
except brozzler.PageInterstitialShown:
|
||||||
|
self.logger.info('page interstitial shown (http auth): %s', page)
|
||||||
|
return []
|
||||||
else:
|
else:
|
||||||
if not self._already_fetched(page, ydl_fetches):
|
if not self._already_fetched(page, ydl_fetches):
|
||||||
self.logger.info('needs fetch: %s', page)
|
self.logger.info('needs fetch: %s', page)
|
||||||
|
|
|
@ -28,6 +28,9 @@ import os
|
||||||
import json
|
import json
|
||||||
import doublethink
|
import doublethink
|
||||||
import datetime
|
import datetime
|
||||||
|
import threading
|
||||||
|
|
||||||
|
global_ydl_lock = threading.Lock()
|
||||||
|
|
||||||
_orig_webpage_read_content = youtube_dl.extractor.generic.GenericIE._webpage_read_content
|
_orig_webpage_read_content = youtube_dl.extractor.generic.GenericIE._webpage_read_content
|
||||||
def _webpage_read_content(self, *args, **kwargs):
|
def _webpage_read_content(self, *args, **kwargs):
|
||||||
|
@ -162,10 +165,12 @@ def _build_youtube_dl(worker, destdir, site):
|
||||||
with open(ctx['filename'], 'rb') as f:
|
with open(ctx['filename'], 'rb') as f:
|
||||||
# include content-length header to avoid chunked
|
# include content-length header to avoid chunked
|
||||||
# transfer, which warcprox currently rejects
|
# transfer, which warcprox currently rejects
|
||||||
|
extra_headers = dict(site.extra_headers())
|
||||||
|
extra_headers['content-length'] = size
|
||||||
request, response = worker._warcprox_write_record(
|
request, response = worker._warcprox_write_record(
|
||||||
warcprox_address=worker._proxy_for(site), url=url,
|
warcprox_address=worker._proxy_for(site), url=url,
|
||||||
warc_type='resource', content_type=mimetype, payload=f,
|
warc_type='resource', content_type=mimetype, payload=f,
|
||||||
extra_headers={'content-length': size})
|
extra_headers=extra_headers)
|
||||||
# consulted by _remember_videos()
|
# consulted by _remember_videos()
|
||||||
self.stitch_ups.append({
|
self.stitch_ups.append({
|
||||||
'url': url,
|
'url': url,
|
||||||
|
@ -182,8 +187,14 @@ def _build_youtube_dl(worker, destdir, site):
|
||||||
if worker._using_warcprox(site):
|
if worker._using_warcprox(site):
|
||||||
self._push_stitched_up_vid_to_warcprox(site, info_dict, ctx)
|
self._push_stitched_up_vid_to_warcprox(site, info_dict, ctx)
|
||||||
|
|
||||||
|
# lock this section to prevent race condition between threads that
|
||||||
|
# want to monkey patch _finish_frag_download() at the same time
|
||||||
|
with global_ydl_lock:
|
||||||
|
try:
|
||||||
youtube_dl.downloader.fragment.FragmentFD._finish_frag_download = _finish_frag_download
|
youtube_dl.downloader.fragment.FragmentFD._finish_frag_download = _finish_frag_download
|
||||||
return super().process_info(info_dict)
|
return super().process_info(info_dict)
|
||||||
|
finally:
|
||||||
|
youtube_dl.downloader.fragment.FragmentFD._finish_frag_download = _orig__finish_frag_download
|
||||||
|
|
||||||
def maybe_heartbeat_site_last_claimed(*args, **kwargs):
|
def maybe_heartbeat_site_last_claimed(*args, **kwargs):
|
||||||
# in case youtube-dl takes a long time, heartbeat site.last_claimed
|
# in case youtube-dl takes a long time, heartbeat site.last_claimed
|
||||||
|
|
2
setup.py
2
setup.py
|
@ -32,7 +32,7 @@ def find_package_data(package):
|
||||||
|
|
||||||
setuptools.setup(
|
setuptools.setup(
|
||||||
name='brozzler',
|
name='brozzler',
|
||||||
version='1.5.dev303',
|
version='1.5.dev304',
|
||||||
description='Distributed web crawling with browsers',
|
description='Distributed web crawling with browsers',
|
||||||
url='https://github.com/internetarchive/brozzler',
|
url='https://github.com/internetarchive/brozzler',
|
||||||
author='Noah Levitt',
|
author='Noah Levitt',
|
||||||
|
|
|
@ -801,7 +801,10 @@ def test_ydl_stitching(httpd):
|
||||||
rr = doublethink.Rethinker('localhost', db='brozzler')
|
rr = doublethink.Rethinker('localhost', db='brozzler')
|
||||||
frontier = brozzler.RethinkDbFrontier(rr)
|
frontier = brozzler.RethinkDbFrontier(rr)
|
||||||
site = brozzler.Site(rr, {
|
site = brozzler.Site(rr, {
|
||||||
'seed': 'http://localhost:%s/site10/' % httpd.server_port})
|
'seed': 'http://localhost:%s/site10/' % httpd.server_port,
|
||||||
|
'warcprox_meta': {
|
||||||
|
'warc-prefix': 'test_ydl_stitching',
|
||||||
|
'captures-table-extra-fields': {'test_id':test_id}}})
|
||||||
brozzler.new_site(frontier, site)
|
brozzler.new_site(frontier, site)
|
||||||
|
|
||||||
# the site should be brozzled fairly quickly
|
# the site should be brozzled fairly quickly
|
||||||
|
@ -816,11 +819,21 @@ def test_ydl_stitching(httpd):
|
||||||
assert len(pages) == 1
|
assert len(pages) == 1
|
||||||
page = pages[0]
|
page = pages[0]
|
||||||
assert len(page.videos) == 6
|
assert len(page.videos) == 6
|
||||||
|
stitched_url = 'youtube-dl:00001:http://localhost:%s/site10/' % httpd.server_port
|
||||||
assert {
|
assert {
|
||||||
'blame': 'youtube-dl',
|
'blame': 'youtube-dl',
|
||||||
'content-length': 267900,
|
'content-length': 267900,
|
||||||
'content-type': 'video/mp4',
|
'content-type': 'video/mp4',
|
||||||
'response_code': 204,
|
'response_code': 204,
|
||||||
'url': 'youtube-dl:00001:http://localhost:%s/site10/' % httpd.server_port,
|
'url': stitched_url,
|
||||||
} in page.videos
|
} in page.videos
|
||||||
|
|
||||||
|
time.sleep(2) # in case warcprox hasn't finished processing urls
|
||||||
|
# take a look at the captures table
|
||||||
|
captures = list(rr.table('captures').filter({'test_id':test_id}).run())
|
||||||
|
l = [c for c in captures if c['url'] == stitched_url]
|
||||||
|
assert len(l) == 1
|
||||||
|
c = l[0]
|
||||||
|
assert c['filename'].startswith('test_ydl_stitching')
|
||||||
|
assert c['content_type'] == 'video/mp4'
|
||||||
|
assert c['http_method'] == 'WARCPROX_WRITE_RECORD'
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue