From 5d4ac14044cba2c7ac9b0ed425528999fa759a72 Mon Sep 17 00:00:00 2001 From: Mylloon Date: Tue, 24 Aug 2021 13:00:28 +0200 Subject: [PATCH] better comments --- main.py | 22 +++++++++++----------- 1 file changed, 11 insertions(+), 11 deletions(-) diff --git a/main.py b/main.py index 5a8998c..b9ecab9 100644 --- a/main.py +++ b/main.py @@ -5,7 +5,7 @@ from cloudscraper import CloudScraper, create_scraper from re import findall, sub class Scraper: - def __init__(self, pseudo, password, app, debug = False): + def __init__(self, pseudo: str, password: str, app: str, debug: bool = False): self.debug = debug self.url = "https://forum.mobilism.org" self.requested_app = app @@ -32,7 +32,7 @@ class Scraper: return session - def search(self, session) -> tuple[list[dict], list[dict]]: + def search(self, session: CloudScraper) -> tuple[list[dict], list[dict]]: """Do the research.""" if self.debug: print("Going to search page and check connection...", end = " ") reponse = session.get(f"{self.url}/search.php", params = {"keywords": self.requested_app, "sr": "topics", "sf": "titleonly"}) # fetch results page @@ -80,7 +80,7 @@ class Scraper: return elements - def getInfos(self, session, elements: list) -> list: + def getInfos(self, session: CloudScraper, elements: list) -> list: """Go to the first n pages and get a lot of infos""" page = 3 if self.debug: print(f"Going to the {page} first pages...", end = " ") @@ -118,14 +118,14 @@ class Scraper: _downloadLinks = findall(r"Download Instructions: ?
(.*|[\s\S]*)", elements[i])[0] except: _downloadLinks = None - _downloadLinks = sub(r"\n|||\">(\S*)", "", _downloadLinks) - _downloadLinks = sub(r"
\n?", "\n", _downloadLinks) - _downloadLinks = sub(r"Mirrors(?!:)|Mirror(?!s)(?!:)", "Mirror:", _downloadLinks) + _downloadLinks = sub(r"\n|||\">(\S*)", "", _downloadLinks) # remove html garbage + _downloadLinks = sub(r"
\n?", "\n", _downloadLinks) # convert newline html to \n + _downloadLinks = sub(r"Mirrors(?!:)|Mirror(?!s)(?!:)", "Mirror:", _downloadLinks) # add ":" elements[i] = {"changelogs": _changelogs, "downloadLinks": _downloadLinks} return elements - def prettyPrint(self, topics: tuple[list[dict], list[dict]]): + def prettyPrint(self, topics: tuple[list[dict], list[dict]]) -> list: """Show a pretty message with all the specialized infos""" topics, topicsInfos = topics print("\n") @@ -147,15 +147,15 @@ class Scraper: return result - def work(self) -> str: + def work(self) -> list: """Call all the others methods.""" return self.prettyPrint(self.search(self.connect())) - def save(self, elements): + def save(self, elements: list) -> None: """Save all the results parsed to a CSV file.""" taille = len(elements) if taille == 0: - print("Aucun élément n'a été trouvé avec la recherche.") + print("No elements were found with the search.") return filename = "results.csv" with open(filename, "w") as f: @@ -167,7 +167,7 @@ class Scraper: if element != "linkParams": f.write(";".join(str(e) for e in list(element.values())[:-1])) f.write("\n") - print(f"{taille} éléments ont étés enrengistés dans le fichier {filename}.") + print(f"{taille} elements have been registered in the {filename} file.") if __name__ == "__main__": argv = argv[1:]