TopDownHockey-Scraper 2.1.11__tar.gz → 2.1.13__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of TopDownHockey-Scraper might be problematic. Click here for more details.

@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: TopDownHockey_Scraper
3
- Version: 2.1.11
3
+ Version: 2.1.13
4
4
  Summary: The TopDownHockey Scraper
5
5
  Home-page: https://github.com/TopDownHockey/TopDownHockey_Scraper
6
6
  Author: Patrick Bacon
@@ -1,6 +1,6 @@
1
1
  [metadata]
2
2
  name = TopDownHockey_Scraper
3
- version = 2.1.11
3
+ version = 2.1.13
4
4
  author = Patrick Bacon
5
5
  author_email = patrick.s.bacon@gmail.com
6
6
  description = A package built for scraping hockey data from EliteProspects, the NHL's HTML/API reports, and ESPN's XML reports.
@@ -9,7 +9,7 @@ with open("README.md", "r", encoding="utf-8") as fh:
9
9
 
10
10
  setup(
11
11
  name="TopDownHockey_Scraper", # Replace with your own username
12
- version="2.1.11",
12
+ version="2.1.13",
13
13
  author="Patrick Bacon",
14
14
  author_email="patrick.s.bacon@gmail.com",
15
15
  description="The TopDownHockey Scraper",
@@ -433,6 +433,8 @@ def scrape_html_roster(season, game_id):
433
433
  (np.where(roster_df['Name']== "JACOB MIDDLETON" , "JAKE MIDDLETON",
434
434
  roster_df['Name']))))))
435
435
 
436
+ roster_df['Name'] = roster_df['Name'].str.normalize('NFKD').str.encode('ascii', errors='ignore').str.decode('utf-8').str.upper()
437
+
436
438
  return roster_df
437
439
 
438
440
  def scrape_html_shifts(season, game_id):
@@ -1097,6 +1099,8 @@ def scrape_html_shifts(season, game_id):
1097
1099
  (np.where(all_shifts['name']== 'ALEX BARRÉ-BOULET', 'ALEX BARRE-BOULET',
1098
1100
  (np.where(all_shifts['name']== "JOSH DUNNE" , "JOSHUA DUNNE",all_shifts['name'])))))))))))))))))))))))))))))))))))))))))))
1099
1101
  )))))))))))))))))))))))))))))))))))))
1102
+
1103
+ all_shits['name'] = all_shifts['name'].str.normalize('NFKD').str.encode('ascii', errors='ignore').str.decode('utf-8').str.upper()
1100
1104
 
1101
1105
 
1102
1106
  all_shifts = all_shifts.assign(end_time = np.where(pd.to_datetime(all_shifts.start_time).dt.time > pd.to_datetime(all_shifts.end_time).dt.time, '20:00', all_shifts.end_time),
@@ -1400,6 +1404,8 @@ def scrape_api_events(game_id, drop_description = True, shift_to_espn = False):
1400
1404
 
1401
1405
  # 21-22 CHANGES
1402
1406
 
1407
+ api_events['ep1_name'] = api_events['ep1_name'].str.normalize('NFKD').str.encode('ascii', errors='ignore').str.decode('utf-8').str.upper()
1408
+
1403
1409
  api_events = api_events.assign(ep1_name = np.where(api_events.ep1_name=='ALEX BARRÉ-BOULET', 'ALEX BARRE_BOULET', api_events.ep1_name))
1404
1410
 
1405
1411
  if drop_description == True:
@@ -1888,6 +1894,8 @@ def scrape_espn_events(espn_game_id, drop_description = True):
1888
1894
 
1889
1895
  espn_events['event_player_1'] = espn_events['event_player_1'].str.strip()
1890
1896
 
1897
+ espn_events['event_player_1'] = espn_events['event_player_1'].str.normalize('NFKD').str.encode('ascii', errors='ignore').str.decode('utf-8').str.upper()
1898
+
1891
1899
  #espn_events = espn_events.assign(event_player_1 = np.where(
1892
1900
  #espn_events.event_player_1=='ALEX BURROWS', 'ALEXANDRE BURROWS', espn_events.event_player_1))
1893
1901
 
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: TopDownHockey-Scraper
3
- Version: 2.1.11
3
+ Version: 2.1.13
4
4
  Summary: The TopDownHockey Scraper
5
5
  Home-page: https://github.com/TopDownHockey/TopDownHockey_Scraper
6
6
  Author: Patrick Bacon