diff --git a/main.py b/main.py
index 5a8998c..b9ecab9 100644
--- a/main.py
+++ b/main.py
@@ -5,7 +5,7 @@ from cloudscraper import CloudScraper, create_scraper
from re import findall, sub
class Scraper:
- def __init__(self, pseudo, password, app, debug = False):
+ def __init__(self, pseudo: str, password: str, app: str, debug: bool = False):
self.debug = debug
self.url = "https://forum.mobilism.org"
self.requested_app = app
@@ -32,7 +32,7 @@ class Scraper:
return session
- def search(self, session) -> tuple[list[dict], list[dict]]:
+ def search(self, session: CloudScraper) -> tuple[list[dict], list[dict]]:
"""Do the research."""
if self.debug: print("Going to search page and check connection...", end = " ")
reponse = session.get(f"{self.url}/search.php", params = {"keywords": self.requested_app, "sr": "topics", "sf": "titleonly"}) # fetch results page
@@ -80,7 +80,7 @@ class Scraper:
return elements
- def getInfos(self, session, elements: list) -> list:
+ def getInfos(self, session: CloudScraper, elements: list) -> list:
"""Go to the first n pages and get a lot of infos"""
page = 3
if self.debug: print(f"Going to the {page} first pages...", end = " ")
@@ -118,14 +118,14 @@ class Scraper:
_downloadLinks = findall(r"Download Instructions: ?
(.*|[\s\S]*)", elements[i])[0]
except:
_downloadLinks = None
- _downloadLinks = sub(r"\n|||\">(\S*)", "", _downloadLinks)
- _downloadLinks = sub(r"
\n?", "\n", _downloadLinks)
- _downloadLinks = sub(r"Mirrors(?!:)|Mirror(?!s)(?!:)", "Mirror:", _downloadLinks)
+ _downloadLinks = sub(r"\n|||\">(\S*)", "", _downloadLinks) # remove html garbage
+ _downloadLinks = sub(r"
\n?", "\n", _downloadLinks) # convert newline html to \n
+ _downloadLinks = sub(r"Mirrors(?!:)|Mirror(?!s)(?!:)", "Mirror:", _downloadLinks) # add ":"
elements[i] = {"changelogs": _changelogs, "downloadLinks": _downloadLinks}
return elements
- def prettyPrint(self, topics: tuple[list[dict], list[dict]]):
+ def prettyPrint(self, topics: tuple[list[dict], list[dict]]) -> list:
"""Show a pretty message with all the specialized infos"""
topics, topicsInfos = topics
print("\n")
@@ -147,15 +147,15 @@ class Scraper:
return result
- def work(self) -> str:
+ def work(self) -> list:
"""Call all the others methods."""
return self.prettyPrint(self.search(self.connect()))
- def save(self, elements):
+ def save(self, elements: list) -> None:
"""Save all the results parsed to a CSV file."""
taille = len(elements)
if taille == 0:
- print("Aucun élément n'a été trouvé avec la recherche.")
+ print("No elements were found with the search.")
return
filename = "results.csv"
with open(filename, "w") as f:
@@ -167,7 +167,7 @@ class Scraper:
if element != "linkParams":
f.write(";".join(str(e) for e in list(element.values())[:-1]))
f.write("\n")
- print(f"{taille} éléments ont étés enrengistés dans le fichier {filename}.")
+ print(f"{taille} elements have been registered in the {filename} file.")
if __name__ == "__main__":
argv = argv[1:]