New tests, Cloner.sync() without tests, but recursive

Signed-off-by: Václav Valíček <valicek1994@gmail.com>
This commit is contained in:
Václav Valíček 2022-08-02 15:55:22 +02:00
parent 4cf4c3ff80
commit 8bc6a7b578
Signed by: valicek
GPG Key ID: FF05BDCA0C73BB31
5 changed files with 138 additions and 15 deletions

View File

@ -0,0 +1,9 @@
from .checksum import gen_repo_hashed_name
from .cloner_config import ClonerConfig, ClonerConfigParser
from .config_file_not_found_error import ConfigFileNotFoundError
from .default_cloner_config import DefaultClonerConfig
from .dir_not_found_error import DirNotFoundError
from .disk_stored_list import DiskStoredList
from .repo_tool import RepoTool
from .repo_dir_structure import RepoDirStructure
from .cloner import Cloner

View File

@ -1,8 +1,7 @@
from repo_cloner.lib.cloner_config import ClonerConfig
from repo_cloner.lib.repo_dir_structure import RepoDirStructure
from repo_cloner.lib.dir_not_found_error import DirNotFoundError
from repo_cloner.lib.repo_tool import RepoTool
from repo_cloner.lib.checksum import gen_repo_hashed_name
from repo_cloner.lib import gen_repo_hashed_name
from repo_cloner.lib import DirNotFoundError
from repo_cloner.lib import ClonerConfig, DiskStoredList, RepoDirStructure, RepoTool
from pathlib import Path
from typing import Optional
from time import time
@ -40,7 +39,6 @@ class Cloner:
log.info("Submodule cache dir does not exist! -> creating")
Path(self.__submodule_cache).mkdir(parents = True)
def check_interval(self):
log.debug(f"Checking interval for {self._config.cloner_project_name}")
# get interval
@ -82,13 +80,98 @@ class Cloner:
def __main_repo_path(self) -> str:
return self._repo_path_by_url(self._config.cloner_repo_url)
@classmethod
def check_submodules_repo(cls, repo_tool: RepoTool, cache_file: str, submodule_list: DiskStoredList,
scan_depth: Optional[int]):
base = os.path.basename(repo_tool.path)
log.debug(f"Loading submodule cache for repo {base}")
repo_commits = DiskStoredList(cache_file)
log.debug(f"Loaded {len(repo_commits)} commits")
# list fetched repo
log.debug(f"Scanning repo {base} for new submodules")
new_commits = repo_tool.list_commits(scan_depth)
# discover new submodules in new commits
for commit in new_commits:
log.debug(f"Discovering submodules in {commit.hexsha}")
if commit.hexsha in repo_commits:
log.debug(f"Cached commit... Okay")
continue
discovered = repo_tool.list_submodules(commit)
if discovered:
for submodule in discovered:
if submodule not in submodule_list:
log.warning(f"Found new submodule: {submodule}")
submodule_list.append(submodule)
repo_commits.append(commit.hexsha)
return submodule_list
def sync(self) -> bool:
if not self.__opened:
self._repo = RepoTool(self.__main_repo_path)
if not self._repo.initialized:
return False
# determine recursive behavior
if not self._config.cloner_submodules:
return self._repo.fetch()
# recursive now
if not self._repo.fetch():
log.critical(f"Repo fetch failed for {self._config.cloner_project_name}")
return False
log.debug(f"Loading submodules.cache")
submodules = DiskStoredList(os.path.join(self.__submodule_cache, "submodules.cache"))
log.debug(f"Loaded submodules.cache - {len(submodules)} items")
path = gen_repo_hashed_name(self._config.cloner_repo_url)
log.debug(f"Main repo hashed name to load: {path}")
# recursion limit?
scan_depth = self._config.cloner_submodule_depth
log.debug(f"Scan depth is {scan_depth} commits")
if scan_depth == 0:
log.debug(f"Repository scan depth is not limited! -> setting scan_depth to none")
scan_depth = None
submodules = Cloner.check_submodules_repo(
self._repo, os.path.join(self.__submodule_cache, path), submodules, scan_depth)
everything_succeed: bool = True
everything_checked: bool = False
fetched_repos = set()
while not everything_checked:
# recursively scan and clone repositories
everything_checked = True
# for every url in list
# list() is needed - Runtime Error for set() changed during iteration
for url in list(submodules):
if url not in fetched_repos:
everything_checked = False
# generate new path
directory = os.path.dirname(self.__main_repo_path)
submodule_cloner = RepoTool(os.path.join(directory, gen_repo_hashed_name(url)))
# clone or checkout?
if not submodule_cloner.initialized:
log.info(f"New uninitialized submodule found: {url}. Cloning...")
checked: bool = submodule_cloner.clone(url)
else:
checked: bool = submodule_cloner.fetch()
# mark cloned even if failed afterwards - while loop stuck solution
fetched_repos.add(url)
if not checked:
log.critical(f"Clone/fetch of submodule: {url} failed")
everything_succeed = False
continue
submodules = Cloner.check_submodules_repo(
submodule_cloner,
os.path.join(self.__submodule_cache, gen_repo_hashed_name(url)),
submodules, scan_depth)
return everything_succeed
def perform_check(self):
log.info(f"Started check for {self._config.cloner_project_name}, url: {self._config.cloner_repo_url}")
if self.check_interval():

View File

@ -184,7 +184,7 @@ class RepoTool:
@__check_initialized
def fetch(self) -> bool:
log.info("Fetching repo state")
log.info("Fetching repo :)")
if not len(self._repo.remotes):
log.warning(f"Repo: {self._path} does not contain any remotes!")
return False
@ -214,6 +214,10 @@ class RepoTool:
log.debug(f"Repo fingerprint is {x}")
return x
@__check_initialized
def list_commits(self, max_depth: Optional[int] = None):
return self._repo.iter_commits(all = True, max_count = max_depth)
@__check_initialized
def list_submodules(self, commit: str = "HEAD") -> Union[list, bool]:
commit = self._repo.commit(commit)
@ -241,7 +245,7 @@ class RepoTool:
def list_submodules_history(self, limit_of_commits: Optional[int] = None) -> Union[list, bool]:
log.info(f"Listing repository submodule history")
iterator = self._repo.iter_commits(all = True, max_count = limit_of_commits)
iterator = self.list_commits(limit_of_commits)
submodules = set()
counter: int = 0
last_status = time.time()

View File

@ -1,12 +1,10 @@
import git
import pytest
from repo_cloner.lib.checksum import gen_repo_hashed_name
import repo_cloner.lib.dir_not_found_error
from repo_cloner.lib import gen_repo_hashed_name, DirNotFoundError, Cloner
from cloner_test_fixtures import *
from repo_cloner.lib.cloner import Cloner
from pathlib import Path
import logging
import unittest.mock
from unittest.mock import patch, PropertyMock
def mock_time() -> float:
@ -35,7 +33,7 @@ class MockDirStruct:
@property
def cache_dir_exists(self):
if self.raise_cache_exists:
raise repo_cloner.lib.dir_not_found_error.DirNotFoundError("mock_dir")
raise DirNotFoundError("mock_dir")
return True
@ -266,7 +264,6 @@ def test_clone_initialized(tmp_path, path_repo_base, caplog):
def test_clone_recursive(tmp_path, path_repo_base, caplog):
from unittest.mock import patch, PropertyMock
mock = MockDirStruct(tmp_path)
mock.config.cloner_repo_url = "https://repo"
mock.config.cloner_submodules = True

View File

@ -458,6 +458,36 @@ def test_fingerprint(support_data_path: Path, repo: str, hash):
assert rt.repo_fingerprint() == hash
def test_list_commits_unopened(tmp_path):
rt = RepoTool(tmp_path)
assert not rt.list_commits()
def test_list_commits(cloned_base_repo_obj):
assert cloned_base_repo_obj.list_commits(3)
count = 0
commits = set()
for commit in cloned_base_repo_obj.list_commits(20):
count += 1
commits.add(commit.hexsha)
assert count == 20
assert commits == {
'93e97409040f6e4ba5cebc4ddca679a2d3203efa', 'd8cd3257e05dc4a3d0061b5320ef5efd0b45d840',
'f916de4ff4c65b93250716354917488ea480f592', 'd6b1a8d59185437ea2fd0055700b4656707ba2d2',
'10c0963c6ded9e41bcb82cf524661d00383d07b4', '06bdaca2c1bbb0b599fdace8698214f82c5850d6',
'95d31440dbcaf38e596e6b712d0d288215ce2a56', '86dab14dfdb87738fe7ff3613222dffb6dd2f796',
'558f12143c6d6d60690acb291cf64a863e12c2d0', '30c00bf3850ec01341c7961a251dee2c4d4cb771',
'7bc1195651dd8bbc62bcc52edf6d07923b2d4f1a', '2cd8137eeb96e2d23169d5728e9071b0ed663677',
'786c4412069526051cbbce6a87e76a63aa64f1e7', '143637c4fc05837d238c0ca67543a4494c07cc85',
'fa0c7b9ac17a342b51fb5414317b54a5dfd2074b', '77f77c3e8bca3f5b43b6f55a88d619b30f6a6ae0',
'b3615798ace53b983251a57593b328e707a8f419', 'e0c7e2a72579e24657c05e875201011d2b48bf94',
'c090c23662933aa5fbbc6385bfae17e316e56a84', '4f5e3068cce0bc562a1ac7719915a86a6d01283e'
}
assert 339 == sum(1 for commit in cloned_base_repo_obj.list_commits())
def test_list_submodules_no_submodules(cloned_base_repo_obj):
assert cloned_base_repo_obj.list_submodules() == []
assert cloned_base_repo_obj.discovered_submodules_commits == ["e0c7e2a72579e24657c05e875201011d2b48bf94"]