ArchiveBox/archivebox/pkgs/abx-plugin-singlefile/abx_plugin_singlefile/singlefile.py

104 lines
3.4 KiB
Python

__package__ = 'abx_plugin_singlefile'
import json
from pathlib import Path
from typing import Optional
from archivebox.index.schema import Link, ArchiveResult, ArchiveError
from archivebox.misc.system import run, chmod_file
from archivebox.misc.util import enforce_types, is_static_file, dedupe
from archivebox.misc.logging_util import TimedProgress
from abx_plugin_chrome.config import CHROME_CONFIG
from abx_plugin_chrome.binaries import CHROME_BINARY
from .config import SINGLEFILE_CONFIG
from .binaries import SINGLEFILE_BINARY
def get_output_path():
return 'singlefile.html'
@enforce_types
def should_save_singlefile(link: Link, out_dir: Optional[Path]=None, overwrite: Optional[bool]=False) -> bool:
if is_static_file(link.url):
return False
out_dir = out_dir or Path(link.link_dir)
if not overwrite and (out_dir / get_output_path()).exists():
return False
return CHROME_CONFIG.USE_CHROME and SINGLEFILE_CONFIG.SAVE_SINGLEFILE
@enforce_types
def save_singlefile(link: Link, out_dir: Optional[Path]=None, timeout: int=60) -> ArchiveResult:
"""download full site using single-file"""
CHROME_BIN = CHROME_BINARY.load()
assert CHROME_BIN.abspath and CHROME_BIN.version
SINGLEFILE_BIN = SINGLEFILE_BINARY.load()
assert SINGLEFILE_BIN.abspath and SINGLEFILE_BIN.version
out_dir = out_dir or Path(link.link_dir)
output = get_output_path()
browser_args = CHROME_CONFIG.chrome_args(CHROME_TIMEOUT=0)
# SingleFile CLI Docs: https://github.com/gildas-lormeau/SingleFile/tree/master/cli
options = [
'--browser-executable-path={}'.format(CHROME_BIN.abspath),
*(["--browser-cookies-file={}".format(SINGLEFILE_CONFIG.SINGLEFILE_COOKIES_FILE)] if SINGLEFILE_CONFIG.SINGLEFILE_COOKIES_FILE else []),
'--browser-args={}'.format(json.dumps(browser_args)),
*SINGLEFILE_CONFIG.SINGLEFILE_EXTRA_ARGS,
]
cmd = [
str(SINGLEFILE_BIN.abspath),
*dedupe(options),
link.url,
output,
]
status = 'succeeded'
timer = TimedProgress(timeout, prefix=' ')
result = None
try:
result = run(cmd, cwd=str(out_dir), timeout=timeout, text=True, capture_output=True)
# parse out number of files downloaded from last line of stderr:
# "Downloaded: 76 files, 4.0M in 1.6s (2.52 MB/s)"
output_tail = [
line.strip()
for line in (result.stdout + result.stderr).rsplit('\n', 5)[-5:]
if line.strip()
]
hints = (
'Got single-file response code: {}.'.format(result.returncode),
*output_tail,
)
# Check for common failure cases
if (result.returncode > 0) or not (out_dir / output).is_file():
raise ArchiveError(f'SingleFile was not able to archive the page (status={result.returncode})', hints)
chmod_file(output, cwd=str(out_dir))
except (Exception, OSError) as err:
status = 'failed'
# TODO: Make this prettier. This is necessary to run the command (escape JSON internal quotes).
cmd[2] = cmd[2].replace('"', "\\\"")
if result:
err.hints = (result.stdout + result.stderr).split('\n')
output = err
finally:
timer.end()
return ArchiveResult(
cmd=cmd,
pwd=str(out_dir),
cmd_version=str(SINGLEFILE_BIN.version),
output=output,
status=status,
**timer.stats,
)