import json import random import re import typing import bs4 import fake_headers import requests import pgbotlib.dbstuff class ApiWrapper: FAILED = 'я обосрался :<' GIF_REGEX = { 'part': re.compile(r'(?<=\).*(?=\<\/center\>)'), 'gif': re.compile(r'(?<=src=").*(?="\s)') } SEARCH_TOKENS = ['botname', '!find'] def __init__(self, tokens: dict, db_conn: pgbotlib.dbstuff.DBConn) -> None: self.tokens = tokens self.db_conn = db_conn self.nonw = re.compile(r'\W') self.headers = fake_headers.Headers(headers=True) # this is the entry point for the api calls # if you add another api, make sure there is a match here # this could have used match - case statement, but python 3.9 def call(self, api: str, data: typing.Union[str, None], message: str) -> str: match api: case 'img_url': return self.format_img(data) case 'gif': return self.get_gif() case 'kmp': return self.get_kmp() case 'fga': return self.get_fga() case 'fakenews': return self.get_fakenews() case 'anek': return self.get_anek() case 'y_search': return self.y_search(message) case _: return self.FAILED def __sanitize_search(self, message: str) -> str: """Removes one of each of the search tokens from the query so that "bot find" phrase does not poison the search query It's not guaranteed it will delete the first match though, and I see no point in implementing that""" keywords = self.nonw.sub(' ', message) for token_spec in self.tokens: if token_spec[0] not in self.SEARCH_TOKENS: continue for regex in token_spec[1]: sub_spec = regex.subn('', keywords, count=1) if sub_spec[1]: keywords = sub_spec[0] break return keywords def y_search(self, message: str) -> str: """Pretty much copy & paste from the original bot I have no fucking clue how this black magic works""" query = self.__sanitize_search(message) request = requests.get('https://yandex.ru/images/search', timeout=30, params={'text': query, 'nomisspell': 1, 'noreask': 1, 'isize': 'medium'}, headers=self.headers.generate()) parser = bs4.BeautifulSoup(request.text, 'html.parser') items_tag = parser.find('div', {'role': 'main'}) items_full = json.loads(items_tag.find('div')['data-state']) items = items_full['initialState']['serpList']['items']['entities'] images = [] for item in items.values(): images.append(item.get('origUrl')) if not images: return self.FAILED result = random.choice(images) return f'[url]({result})' def get_gif(self) -> str: resp = requests.get("http://xdgif.ru/random/", timeout=30) part = self.GIF_REGEX['part'].search(resp.text).group(0) gif = self.GIF_REGEX['gif'].search(part).group(0) return gif @staticmethod def get_kmp() -> str: request = requests.get("https://killpls.me/random/", timeout=30) parser = bs4.BeautifulSoup(request.text, features="html.parser") result = parser.find("div", attrs={ "style": "margin:0.5em 0;line-height:1.785em"}) return result.text.strip() @staticmethod def get_fga() -> str: request = requests.get("http://fucking-great-advice.ru/api/random", timeout=30) return json.loads(request.text)["text"] @staticmethod def get_fakenews() -> str: request = requests.get("http://news.olegmakarenko.ru/news", timeout=30) parser = bs4.BeautifulSoup(request.text, features="html.parser") news = [item.text.strip() for item in parser.find_all( "span", attrs={"class": "headlinetext"})] return random.choice(news) @staticmethod def get_anek() -> str: request = requests.get("http://rzhunemogu.ru/Rand.aspx?CType=11", timeout=30) result = request.text.split('')[1].split('')[0] return result.strip() @staticmethod def format_img(data: str) -> str: return f'[url]({data})'