amazon metadata
This commit is contained in:
parent
6a37c7ca9d
commit
62e7aca0fb
98
cps/metadata_provider/amazon.py
Normal file
98
cps/metadata_provider/amazon.py
Normal file
|
@ -0,0 +1,98 @@
|
|||
import concurrent.futures
|
||||
import requests
|
||||
from bs4 import BeautifulSoup as BS # requirement
|
||||
import lxml #requirement for better speed
|
||||
import cchardet #optional for better speed
|
||||
from cps.services.Metadata import Metadata
|
||||
#from time import time
|
||||
from operator import itemgetter
|
||||
class Amazon(Metadata):
|
||||
__name__ = "Amazon"
|
||||
__id__ = "amazon"
|
||||
headers = {'upgrade-insecure-requests': '1',
|
||||
'user-agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/97.0.4692.71 Safari/537.36',
|
||||
'accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,image/avif,image/webp,image/apng,*/*;q=0.8,application/signed-exchange;v=b3;q=0.9',
|
||||
'sec-gpc': '1',
|
||||
'sec-fetch-site': 'none',
|
||||
'sec-fetch-mode': 'navigate',
|
||||
'sec-fetch-user': '?1',
|
||||
'sec-fetch-dest': 'document',
|
||||
'accept-encoding': 'gzip, deflate, br',
|
||||
'accept-language': 'en-US,en;q=0.9'}
|
||||
session = requests.Session()
|
||||
session.headers=headers
|
||||
|
||||
def search(self, query: str, generic_cover=""):
|
||||
#timer=time()
|
||||
def inner(link,index)->[dict,int]:
|
||||
with self.session as session:
|
||||
r = session.get(f"https://www.amazon.com/{link}")
|
||||
r.raise_for_status()
|
||||
long_soup = BS(r.text, "lxml") #~4sec :/
|
||||
soup2 = long_soup.find("div", attrs={"cel_widget_id": "dpx-books-ppd_csm_instrumentation_wrapper"})
|
||||
if soup2 is None:
|
||||
return
|
||||
try:
|
||||
v = dict()
|
||||
v['source'] = {
|
||||
"id": self.__id__,
|
||||
"description": "Amazon Books",
|
||||
"link": "https://amazon.com/"
|
||||
}
|
||||
v['url'] = f"https://www.amazon.com/{link}"
|
||||
|
||||
#the more searches the slower, these are too hard to find in reasonable time or might not even exist
|
||||
v['publisher'] = "" # very unreliable
|
||||
v['publishedDate'] = "" # very unreliable
|
||||
v['id'] = None # ?
|
||||
v['tags'] = [] # dont exist on amazon
|
||||
|
||||
|
||||
try:
|
||||
v['description'] = "\n".join(
|
||||
soup2.find("div", attrs={"data-feature-name": "bookDescription"}).stripped_strings).replace(
|
||||
"\xa0"," ")[:-9].strip().strip("\n")
|
||||
except (AttributeError, TypeError):
|
||||
return None # if there is no description it is not a book and therefore should be ignored
|
||||
try:
|
||||
v['title'] = soup2.find("span", attrs={"id": "productTitle"}).text
|
||||
except (AttributeError, TypeError):
|
||||
v['title'] = ""
|
||||
try:
|
||||
v['authors'] = [next(
|
||||
filter(lambda i: i != " " and i != "\n" and not i.startswith("{"),
|
||||
x.findAll(text=True))).strip()
|
||||
for x in soup2.findAll("span", attrs={"class": "author"})]
|
||||
except (AttributeError, TypeError, StopIteration):
|
||||
v['authors'] = ""
|
||||
try:
|
||||
v['rating'] = int(
|
||||
soup2.find("span", class_="a-icon-alt").text.split(" ")[0].split(".")[
|
||||
0]) # first number in string
|
||||
except (AttributeError, ValueError):
|
||||
v['rating'] = 0
|
||||
try:
|
||||
v['cover'] = soup2.find("img", attrs={"class": "a-dynamic-image frontImage"})["src"]
|
||||
except (AttributeError, TypeError):
|
||||
v['cover'] = ""
|
||||
return v,index
|
||||
except:
|
||||
return
|
||||
|
||||
val = list()
|
||||
if self.active:
|
||||
results = self.session.get(
|
||||
f"https://www.amazon.com/s?k={query.replace(' ', '+')}&i=digital-text&sprefix={query.replace(' ', '+')}%2Cdigital-text&ref=nb_sb_noss",
|
||||
headers=self.headers)
|
||||
results.raise_for_status()
|
||||
soup = BS(results.text, 'html.parser')
|
||||
links_list = [next(filter(lambda i: "digital-text" in i["href"], x.findAll("a")))["href"] for x in
|
||||
soup.findAll("div", attrs={"data-component-type": "s-search-result"})]
|
||||
with concurrent.futures.ThreadPoolExecutor(max_workers=5) as executor:
|
||||
fut = {executor.submit(inner, link, index) for index, link in enumerate(links_list[:5])}
|
||||
val=list(map(lambda x : x.result() ,concurrent.futures.as_completed(fut)))
|
||||
#print(val)
|
||||
#print(f"time was {time()-timer}")
|
||||
result=list(filter(lambda x: x, val))
|
||||
return [x[0] for x in sorted(result,key=itemgetter(1))] #sort by amazons listing order for best relevance
|
||||
|
Loading…
Reference in New Issue
Block a user