Хочу добавить туда еще сайт betodds24.
import requests from bs4 import BeautifulSoup import pandas as pd import re import cloudscraper # Default dict that will be used to store everything predicts = {} # Simple class just to simplify class Game: def __init__(self): self.name = '' self.predict = '' # Running every function that we have for scraping the predicts from each website # To add more websites, you will need to add the function name on the function array def main(): functions = [forebet, predictz, windrawwin, soccervista, prosoccer, vitibet, footystats] for func in functions: func() def forebet(): global predicts # https://www.forebet.com/en/football-tips-and-predictions-for-today predicts['forebet'] = [] url = "https://www.forebet.com/en/football-tips-and-predictions-for-today" page = requests.get(url, headers={"user-agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/92.0.4515.159 Safari/537.36"}) soup = BeautifulSoup(page.content, "html.parser") games = soup.find_all(class_="rcnt tr_1") for game in games: name = game.find("meta", {"itemprop":"name"}) if name is None: continue game_class = Game() game_class.name = name.attrs['content'] game_class.predict = game.find('span', {"class": "forepr"}).text predicts['forebet'].append({'game': game_class.name, 'predict': game_class.predict}) def predictz(): global predicts # https://www.predictz.com/predictions predicts['predictz'] = [] scraper = cloudscraper.create_scraper() page = scraper.get("https://www.predictz.com/predictions") soup = BeautifulSoup(page.text, "html.parser") games = soup.find_all(class_='ptcnt') for game in games: if game.find("div", {"class": "ptmobh"}) is None: continue home = game.find("div", {"class": "ptmobh"}).text away = game.find("div", {"class": "ptmoba"}).text if home == '' or away == '': continue game_class = Game() game_class.name = home + " vs " + away predict_text = game.find("div", {"class": "ptpredboxsml"}).text game_class.predict = '1' if 'Home' in predict_text else '2' if 'Away' in predict_text else 'X' predicts['predictz'].append({'game': game_class.name, 'predict': game_class.predict}) def windrawwin(): global predicts # https://www.windrawwin.com/predictions/today predicts['windrawwin'] = [] scraper = cloudscraper.create_scraper() page = scraper.get("https://www.windrawwin.com/predictions/today") soup = BeautifulSoup(page.content, "html.parser") games = soup.find_all(class_='wttr') for game in games: teams = game.find_all("div", {"class": "wtmoblnk"}) game_class = Game() game_class.name = teams[0].text + " vs " + teams[1].text predict_text = game.find("div", {"class": "wtprd"}).text # Home 2-0 game_class.predict = '1' if 'Home' in predict_text else '2' if 'Away' in predict_text else 'X' predicts['windrawwin'].append({'game': game_class.name, 'predict': game_class.predict}) def soccervista(): global predicts # https://www.newsoccervista.com/ predicts['soccervista'] = [] url = "https://www.newsoccervista.com" page = requests.get(url, headers={"user-agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/92.0.4515.159 Safari/537.36"}) soup = BeautifulSoup(page.content, "html.parser") games = soup.find_all(class_='twom') for game in games: home = game.find("div", {"class": "hometeam"}).text away = game.find("div", {"class": "awayteam"}).text if home == '' or away == '': continue game_class = Game() game_class.name = home + " vs " + away game_class.predict = game.find("strong").text predicts['soccervista'].append({'game': game_class.name, 'predict': game_class.predict}) def prosoccer(): global predicts # https://www.prosoccer.gr/en/football/predictions/ predicts['prosoccer'] = [] url = "https://www.prosoccer.gr/en/football/predictions" page = requests.get(url, headers={"user-agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/92.0.4515.159 Safari/537.36"}) soup = BeautifulSoup(page.content, "html.parser") games = soup.find_all('tr') for game in games: try: game_name = game.find("td", {"class": "mio"}).text.lower() except: continue if game_name is None: continue game_class = Game() game_class.name = game_name.split('-')[0][:-1] + ' vs ' + game_name.split('-')[1][1:] predict = game.find("span", {"class": "sctip"}).text[1:] if '-' in predict: predict = predict.split('-')[0] game_class.predict = predict predicts['prosoccer'].append({'game': game_class.name, 'predict': game_class.predict}) def vitibet(): global predicts # https://www.vitibet.com/index.php?clanek=quicktips&sekce=fotbal&lang=en predicts['vitibet'] = [] url = "https://www.vitibet.com/index.php?clanek=quicktips&sekce=fotbal&lang=en" page = requests.get(url, headers={"user-agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/92.0.4515.159 Safari/537.36"}) soup = BeautifulSoup(page.content, "html.parser") games = soup.find_all('tr', class_=None) for game in games: try: game_name = game.find_all("td", {"class": "standardbunka"}) except: continue if game_name is None or game_name == []: continue game_class = Game() game_class.name = game_name[1].text + ' vs ' + game_name[2].text regex = re.compile('barvapodtipek.*') game_class.predict = game.find("td", {"class": regex}).text.replace('0', 'X') predicts['vitibet'].append({'game': game_class.name, 'predict': game_class.predict}) def footystats(): global predicts # https://footystats.org/predictions/ predicts['footystats'] = [] url = "https://footystats.org/predictions/" page = requests.get(url, headers={"user-agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/92.0.4515.159 Safari/537.36"}) soup = BeautifulSoup(page.content, "html.parser") games = soup.find_all(class_='betHeaderTitle') for game in games: predict = game.find("span", {"class": "market"}).text.lower() game.find('span', class_="market").decompose() game_name = game.text.strip() if game_name == 'See More Football Predictions': continue game_class = Game() game_class.name = game_name game_class.predict = '1' if 'home win' in predict else '2' if 'away win' in predict else 'X' if 'draw' in predict else predict predicts['footystats'].append({'game': game_class.name, 'predict': game_class.predict}) main() # Storing all the array names to filter the games and the predicts to_filter = list(predicts.keys()) # Creating an empty array to store all the games from every website predicts['games'] = [''] # Group all the games from every website and tried to ignore existing ones for arr in to_filter: for to_add_games in predicts[arr]: found = False for game in predicts['games']: game_teams = to_add_games['game'].split(' vs ') home_name = game_teams[0] away_name = game_teams[1] if home_name.lower() in game.lower() or away_name.lower() in game.lower(): found = True if found == False: predicts['games'].append(to_add_games['game']) # Match the predicts with the games from the websites for arr in to_filter: predicts['predicts_' + arr] = [] for game in predicts['games']: found = False for game_to_filter in predicts[arr]: game_teams = game_to_filter['game'].split(' vs ') home_name = game_teams[0] away_name = game_teams[1] if home_name.lower() in game.lower() or away_name.lower() in game.lower(): predicts['predicts_' + arr].append(game_to_filter['predict']) found = True break if found == False: predicts['predicts_' + arr].append('') # Creating the xlsx with the games and the predicts of every website and for each game df = pd.DataFrame({ 'Games': predicts['games'], 'Forebet': predicts['predicts_forebet'], 'PredictZ': predicts['predicts_predictz'], 'WinDrawWin': predicts['predicts_windrawwin'], 'SoccerVista': predicts['predicts_soccervista'], 'ProSoccer': predicts['predicts_prosoccer'], 'Vitibet': predicts['predicts_vitibet'], 'Footystats': predicts['predicts_footystats'] }) writer = pd.ExcelWriter('predicts.xlsx', engine='xlsxwriter') df.to_excel(writer, sheet_name='Predicts') # Simple stylings writer.sheets['Predicts'].set_column('B:B', 50) writer.sheets['Predicts'].set_column('E:E', 15) writer.save()
def betodds24(): global predicts # https://www.betodds24.com/ predicts['betodds24'] = [] scraper = cloudscraper.create_scraper() page = scraper.get("https://www.betodds24.com") soup = BeautifulSoup(page.content, "html.parser") games = soup.find_all(class_="tr") for game in games: home = game.find("tr").text avay = game.find("tr").text if home == '' or away == '': continue game_class = Game() game_class.name = home + " vs " + away game_class.predict = game.find("th").text predicts['betodds24'].append({'game': game_class.name, 'predict': game_class.predict})