BETTERZON-59: Adding crawler basic framework (#29)

This commit is contained in:
Patrick 2021-04-14 21:51:36 +02:00 committed by GitHub
parent f5fd1825d7
commit 610808ad03
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
2 changed files with 79 additions and 1 deletions

78
Crawler/crawler.py Normal file
View File

@ -0,0 +1,78 @@
import sql
def crawl(product_ids: [int]) -> dict:
"""
Crawls the given list of products and saves the results to sql
:param products: The list of product IDs to fetch
:return: A dict with the following fields:
total_crawls: number of total crawl tries (products * vendors per product)
successful_crawls: number of successful products
products_with_problems: list of products that have not been crawled successfully
"""
total_crawls = 0
successful_crawls = 0
products_with_problems = []
# Iterate over every product that has to be crawled
for product_id in product_ids:
# Get all links for this product
product_links = sql.getProductLinksForProduct(product_id)
crawled_data = []
# Iterate over every link / vendor
for product_vendor_info in product_links:
total_crawls += 1
# Call the appropriate vendor crawling function and append the result to the list of crawled data
if product_vendor_info['vendor_id'] == 1:
# Amazon
crawled_data.append(__crawl_amazon__(product_vendor_info))
elif product_vendor_info['vendor_id'] == 2:
# Apple
crawled_data.append(__crawl_apple__(product_vendor_info))
elif product_vendor_info['vendor_id'] == 3:
# Media Markt
crawled_data.append(__crawl_mediamarkt__(product_vendor_info))
else:
products_with_problems.append(product_vendor_info)
continue
successful_crawls += 1
# Insert data to SQL
sql.insertData(crawled_data)
return {
'total_crawls': total_crawls,
'successful_crawls': successful_crawls,
'products_with_problems': products_with_problems
}
def __crawl_amazon__(product_info: dict) -> tuple:
"""
Crawls the price for the given product from amazon
:param product_info: A dict with product info containing product_id, vendor_id, url
:return: A tuple with the crawled data, containing (product_id, vendor_id, price_in_cents)
"""
return (product_info['product_id'], product_info['vendor_id'], 123)
def __crawl_apple__(product_info: dict) -> tuple:
"""
Crawls the price for the given product from apple
:param product_info: A dict with product info containing product_id, vendor_id, url
:return: A tuple with the crawled data, containing (product_id, vendor_id, price_in_cents)
"""
return (product_info['product_id'], product_info['vendor_id'], 123)
def __crawl_mediamarkt__(product_info: dict) -> tuple:
"""
Crawls the price for the given product from media markt
:param product_info: A dict with product info containing product_id, vendor_id, url
:return: A tuple with the crawled data, containing (product_id, vendor_id, price_in_cents)
"""
pass

View File

@ -43,7 +43,7 @@ def getProductsForVendor(vendor_id: int) -> [{}]:
return products return products
def getProductLinksForProduct(product_id: int) -> [{}]: def getProductLinksForProduct(product_id: int) -> [dict]:
""" """
Queries all the product links for the given product Queries all the product links for the given product
:param product_id: The product to query data for :param product_id: The product to query data for