Update tests to require proper checksums

This commit is contained in:
David Dalcino
2022-03-06 17:04:48 -08:00
parent f979d80899
commit be23b62549
4 changed files with 30 additions and 24 deletions

View File

@@ -447,6 +447,7 @@ def test_archives_weird_module_7z_name(
expect_archives: Set[str],
):
monkeypatch.setattr("aqt.archives.getUrl", lambda *args: xml)
monkeypatch.setattr("aqt.archives.get_hash", lambda *args, **kwargs: hashlib.sha256(bytes(xml, "utf-8")).hexdigest())
qt_archives = make_archives_fn(subarchives, modules, is_include_base)
archives = {pkg.archive for pkg in qt_archives.archives}

View File

@@ -331,3 +331,12 @@ def test_helper_getUrl_conn_error(monkeypatch):
getUrl(url, timeout)
assert e.type == ArchiveConnectionError
assert expect_re.match(format(e.value))
def test_helper_getUrl_checksum_error(monkeypatch):
mocked_get, mocked_session_get = mock_get_redirect(0)
monkeypatch.setattr(requests, "get", mocked_get)
monkeypatch.setattr(requests.Session, "get", mocked_session_get)
with pytest.raises(ArchiveChecksumError) as e:
getUrl("some_url", timeout=(5, 5), expected_hash=b"AAAAAAAAAAA")
assert e.type == ArchiveChecksumError

View File

@@ -1,3 +1,4 @@
import hashlib
import logging
import os
import re
@@ -134,11 +135,14 @@ def make_mock_geturl_download_archive(
for _arc in archives:
assert _arc.filename_7z.endswith(".7z")
xml = "<Updates>\n{}\n</Updates>".format("\n".join([archive.xml_package_update() for archive in archives]))
def mock_getUrl(url: str, *args) -> str:
if url.endswith(updates_url):
return "<Updates>\n{}\n</Updates>".format("\n".join([archive.xml_package_update() for archive in archives]))
return xml
elif url.endswith(".sha256"):
return "" # Skip the checksum
filename = url.split("/")[-1][: -len(".sha256")]
return f"{hashlib.sha256(bytes(xml, 'utf-8')).hexdigest()} {filename}"
assert False
def mock_download_archive(url: str, out: str, *args):
@@ -707,13 +711,16 @@ def test_install(
),
)
def test_install_nonexistent_archives(monkeypatch, capsys, cmd, xml_file: Optional[str], expected):
xml = (Path(__file__).parent / "data" / xml_file).read_text("utf-8") if xml_file else ""
def mock_get_url(url, *args, **kwargs):
if not xml_file:
raise ArchiveDownloadError(f"Failed to retrieve file at {url}\nServer response code: 404, reason: Not Found")
return (Path(__file__).parent / "data" / xml_file).read_text("utf-8")
return xml
monkeypatch.setattr("aqt.archives.getUrl", mock_get_url)
monkeypatch.setattr("aqt.helper.getUrl", mock_get_url)
monkeypatch.setattr("aqt.archives.get_hash", lambda *args, **kwargs: hashlib.sha256(bytes(xml, "utf-8")).hexdigest())
monkeypatch.setattr("aqt.metadata.get_hash", lambda *args, **kwargs: hashlib.sha256(bytes(xml, "utf-8")).hexdigest())
monkeypatch.setattr("aqt.metadata.getUrl", mock_get_url)
cli = Cli()

View File

@@ -1,3 +1,4 @@
import hashlib
import json
import os
import re
@@ -953,6 +954,7 @@ def test_list_tool_cli(monkeypatch, capsys, host: str, target: str, tool_name: s
def test_fetch_http_ok(monkeypatch):
monkeypatch.setattr("aqt.metadata.get_hash", lambda *args, **kwargs: hashlib.sha256(b"some_html_content").hexdigest())
monkeypatch.setattr("aqt.metadata.getUrl", lambda **kwargs: "some_html_content")
assert MetadataFactory.fetch_http("some_url") == "some_html_content"
@@ -966,6 +968,7 @@ def test_fetch_http_failover(monkeypatch):
raise ArchiveDownloadError()
return "some_html_content"
monkeypatch.setattr("aqt.metadata.get_hash", lambda *args, **kwargs: hashlib.sha256(b"some_html_content").hexdigest())
monkeypatch.setattr("aqt.metadata.getUrl", _mock)
# Require that the first attempt failed, but the second did not
@@ -973,33 +976,19 @@ def test_fetch_http_failover(monkeypatch):
assert len(urls_requested) == 2
def test_fetch_http_download_error(monkeypatch):
@pytest.mark.parametrize("exception_on_error", (ArchiveDownloadError, ArchiveConnectionError))
def test_fetch_http_download_error(monkeypatch, exception_on_error):
urls_requested = set()
def _mock(url, **kwargs):
urls_requested.add(url)
raise ArchiveDownloadError()
raise exception_on_error()
monkeypatch.setattr("aqt.metadata.get_hash", lambda *args, **kwargs: hashlib.sha256(b"some_html_content").hexdigest())
monkeypatch.setattr("aqt.metadata.getUrl", _mock)
with pytest.raises(ArchiveDownloadError) as e:
with pytest.raises(exception_on_error) as e:
MetadataFactory.fetch_http("some_url")
assert e.type == ArchiveDownloadError
# Require that a fallback url was tried
assert len(urls_requested) == 2
def test_fetch_http_conn_error(monkeypatch):
urls_requested = set()
def _mock(url, **kwargs):
urls_requested.add(url)
raise ArchiveConnectionError()
monkeypatch.setattr("aqt.metadata.getUrl", _mock)
with pytest.raises(ArchiveConnectionError) as e:
MetadataFactory.fetch_http("some_url")
assert e.type == ArchiveConnectionError
assert e.type == exception_on_error
# Require that a fallback url was tried
assert len(urls_requested) == 2