uk_bin_collection 0.118.0__py3-none-any.whl → 0.120.0__py3-none-any.whl

Sign up to get free protection for your applications and to get access to all the features.
Files changed (19) hide show
  1. uk_bin_collection/tests/input.json +70 -2
  2. uk_bin_collection/uk_bin_collection/councils/AberdeenCityCouncil.py +122 -0
  3. uk_bin_collection/uk_bin_collection/councils/BaberghDistrictCouncil.py +3 -1
  4. uk_bin_collection/uk_bin_collection/councils/BraintreeDistrictCouncil.py +70 -0
  5. uk_bin_collection/uk_bin_collection/councils/BurnleyBoroughCouncil.py +88 -0
  6. uk_bin_collection/uk_bin_collection/councils/CumberlandAllerdaleCouncil.py +93 -0
  7. uk_bin_collection/uk_bin_collection/councils/EdinburghCityCouncil.py +98 -0
  8. uk_bin_collection/uk_bin_collection/councils/EnvironmentFirst.py +14 -0
  9. uk_bin_collection/uk_bin_collection/councils/ExeterCityCouncil.py +52 -0
  10. uk_bin_collection/uk_bin_collection/councils/HartlepoolBoroughCouncil.py +83 -0
  11. uk_bin_collection/uk_bin_collection/councils/LondonBoroughHavering.py +75 -0
  12. uk_bin_collection/uk_bin_collection/councils/MidSuffolkDistrictCouncil.py +3 -1
  13. uk_bin_collection/uk_bin_collection/councils/NewcastleUnderLymeCouncil.py +66 -0
  14. uk_bin_collection/uk_bin_collection/councils/NorthHertfordshireDistrictCouncil.py +93 -0
  15. {uk_bin_collection-0.118.0.dist-info → uk_bin_collection-0.120.0.dist-info}/METADATA +1 -1
  16. {uk_bin_collection-0.118.0.dist-info → uk_bin_collection-0.120.0.dist-info}/RECORD +19 -9
  17. {uk_bin_collection-0.118.0.dist-info → uk_bin_collection-0.120.0.dist-info}/LICENSE +0 -0
  18. {uk_bin_collection-0.118.0.dist-info → uk_bin_collection-0.120.0.dist-info}/WHEEL +0 -0
  19. {uk_bin_collection-0.118.0.dist-info → uk_bin_collection-0.120.0.dist-info}/entry_points.txt +0 -0
@@ -6,6 +6,12 @@
6
6
  "wiki_name": "Aberdeenshire Council",
7
7
  "wiki_note": "You will need to use [FindMyAddress](https://www.findmyaddress.co.uk/search) to find the UPRN."
8
8
  },
9
+ "AberdeenCityCouncil": {
10
+ "url": "https://www.aberdeencity.gov.uk",
11
+ "uprn": "9051156186",
12
+ "wiki_name": "Aberdeen City Council",
13
+ "wiki_note": "You will need to use [FindMyAddress](https://www.findmyaddress.co.uk/search) to find the UPRN."
14
+ },
9
15
  "AdurAndWorthingCouncils": {
10
16
  "url": "https://www.adur-worthing.gov.uk/bin-day/?brlu-selected-address=100061878829",
11
17
  "wiki_command_url_override": "https://www.adur-worthing.gov.uk/bin-day/?brlu-selected-address=XXXXXXXX",
@@ -82,9 +88,10 @@
82
88
  "skip_get_url": true,
83
89
  "house_number": "Monday",
84
90
  "postcode": "Week 1",
91
+ "uprn": "Tuesday",
85
92
  "url": "https://www.babergh.gov.uk",
86
93
  "wiki_name": "Babergh District Council",
87
- "wiki_note": "Use the House Number field to pass the DAY of the week for your collections. Monday/Tuesday/Wednesday/Thursday/Friday. [OPTIONAL] Use the 'postcode' field to pass the WEEK for your garden collection. [Week 1/Week 2]"
94
+ "wiki_note": "Use the House Number field to pass the DAY of the week for your NORMAL collections. [Monday/Tuesday/Wednesday/Thursday/Friday]. [OPTIONAL] Use the 'postcode' field to pass the WEEK for your garden collection. [Week 1/Week 2]. [OPTIONAL] Use the 'uprn' field to pass the DAY for your garden collection. [Monday/Tuesday/Wednesday/Thursday/Friday]"
88
95
  },
89
96
  "BCPCouncil": {
90
97
  "skip_get_url": true,
@@ -213,6 +220,14 @@
213
220
  "wiki_name": "Bradford MDC",
214
221
  "wiki_note": "To get the UPRN, you will need to use [FindMyAddress](https://www.findmyaddress.co.uk/search). Postcode isn't parsed by this script, but you can pass it in double quotes."
215
222
  },
223
+ "BraintreeDistrictCouncil": {
224
+ "postcode": "CO5 9BD",
225
+ "skip_get_url": true,
226
+ "uprn": "10006930172",
227
+ "url": "https://www.braintree.gov.uk/",
228
+ "wiki_name": "Braintree District Council",
229
+ "wiki_note": "Provide your UPRN and postcode. Use [FindMyAddress](https://www.findmyaddress.co.uk/search) to find your UPRN."
230
+ },
216
231
  "BrecklandCouncil": {
217
232
  "url": "https://www.breckland.gov.uk",
218
233
  "wiki_command_url_override": "https://www.breckland.gov.uk",
@@ -276,6 +291,12 @@
276
291
  "wiki_name": "Buckinghamshire Council (Chiltern, South Bucks, Wycombe)",
277
292
  "wiki_note": "Pass the house name/number and postcode in their respective arguments, both wrapped in quotes."
278
293
  },
294
+ "BurnleyBoroughCouncil": {
295
+ "uprn": "100010347165",
296
+ "url": "https://www.burnley.gov.uk",
297
+ "wiki_name": "Burnley Borough Council",
298
+ "wiki_note": "Pass the UPRN. You can find it using [FindMyAddress](https://www.findmyaddress.co.uk/search)."
299
+ },
279
300
  "BuryCouncil": {
280
301
  "house_number": "3",
281
302
  "postcode": "M26 3XY",
@@ -448,6 +469,13 @@
448
469
  "wiki_name": "Croydon Council",
449
470
  "wiki_note": "Pass the house number and postcode in their respective parameters."
450
471
  },
472
+ "CumberlandAllerdaleCouncil": {
473
+ "house_number": "2",
474
+ "postcode": "CA13 0DE",
475
+ "url": "https://www.allerdale.gov.uk",
476
+ "wiki_name": "Cumberland Council - Allerdale District",
477
+ "wiki_note": "Pass the house number and postcode in their respective parameters."
478
+ },
451
479
  "DacorumBoroughCouncil": {
452
480
  "house_number": "13",
453
481
  "postcode": "HP3 9JY",
@@ -597,6 +625,14 @@
597
625
  "wiki_name": "Eastleigh Borough Council",
598
626
  "wiki_note": "Pass the UPRN. You can find it using [FindMyAddress](https://www.findmyaddress.co.uk/search)."
599
627
  },
628
+ "EdinburghCityCouncil": {
629
+ "skip_get_url": true,
630
+ "house_number": "Tuesday",
631
+ "postcode": "Week 1",
632
+ "url": "https://www.edinburgh.gov.uk",
633
+ "wiki_name": "Edinburgh City Council",
634
+ "wiki_note": "Use the House Number field to pass the DAY of the week for your collections. Monday/Tuesday/Wednesday/Thursday/Friday. Use the 'postcode' field to pass the WEEK for your collection. [Week 1/Week 2]"
635
+ },
600
636
  "ElmbridgeBoroughCouncil": {
601
637
  "url": "https://www.elmbridge.gov.uk",
602
638
  "wiki_command_url_override": "https://www.elmbridge.gov.uk",
@@ -632,6 +668,12 @@
632
668
  "wiki_name": "Erewash Borough Council",
633
669
  "wiki_note": "Pass the UPRN. You can find it using [FindMyAddress](https://www.findmyaddress.co.uk/search)."
634
670
  },
671
+ "ExeterCityCouncil": {
672
+ "uprn": "100040212270",
673
+ "url": "https://www.exeter.gov.uk",
674
+ "wiki_name": "Exeter City Council",
675
+ "wiki_note": "Pass the UPRN. You can find it using [FindMyAddress](https://www.findmyaddress.co.uk/search)."
676
+ },
635
677
  "FalkirkCouncil": {
636
678
  "url": "https://www.falkirk.gov.uk",
637
679
  "wiki_command_url_override": "https://www.falkirk.gov.uk",
@@ -776,6 +818,12 @@
776
818
  "wiki_name": "Harrogate Borough Council",
777
819
  "wiki_note": "Pass the UPRN, which can be found at [this site](https://secure.harrogate.gov.uk/inmyarea). URL doesn't need to be passed."
778
820
  },
821
+ "HartlepoolBoroughCouncil": {
822
+ "url": "https://www.hartlepool.gov.uk",
823
+ "uprn": "100110019551",
824
+ "wiki_name": "Hartlepool Borough Council",
825
+ "wiki_note": "You will need to use [FindMyAddress](https://www.findmyaddress.co.uk/search) to find your UPRN."
826
+ },
779
827
  "HertsmereBoroughCouncil": {
780
828
  "house_number": "1",
781
829
  "postcode": "WD7 9HZ",
@@ -916,6 +964,12 @@
916
964
  "wiki_name": "London Borough Harrow",
917
965
  "wiki_note": "You will need to use [FindMyAddress](https://www.findmyaddress.co.uk/search) to find the UPRN."
918
966
  },
967
+ "LondonBoroughHavering": {
968
+ "url": "https://www.havering.gov.uk",
969
+ "uprn": "100021380730",
970
+ "wiki_name": "London Borough Havering",
971
+ "wiki_note": "Pass the UPRN. You can find it using [FindMyAddress](https://www.findmyaddress.co.uk/search)."
972
+ },
919
973
  "LondonBoroughHounslow": {
920
974
  "skip_get_url": true,
921
975
  "uprn": "100021577765",
@@ -1013,9 +1067,10 @@
1013
1067
  "skip_get_url": true,
1014
1068
  "house_number": "Monday",
1015
1069
  "postcode": "Week 2",
1070
+ "uprn": "Monday",
1016
1071
  "url": "https://www.midsuffolk.gov.uk",
1017
1072
  "wiki_name": "Mid Suffolk District Council",
1018
- "wiki_note": "Use the House Number field to pass the DAY of the week for your collections. Monday/Tuesday/Wednesday/Thursday/Friday. [OPTIONAL] Use the 'postcode' field to pass the WEEK for your garden collection. [Week 1/Week 2]"
1073
+ "wiki_note": "Use the House Number field to pass the DAY of the week for your NORMAL collections. [Monday/Tuesday/Wednesday/Thursday/Friday]. [OPTIONAL] Use the 'postcode' field to pass the WEEK for your garden collection. [Week 1/Week 2]. [OPTIONAL] Use the 'uprn' field to pass the DAY for your garden collection. [Monday/Tuesday/Wednesday/Thursday/Friday]"
1019
1074
  },
1020
1075
  "MidSussexDistrictCouncil": {
1021
1076
  "house_number": "OAKLANDS, OAKLANDS ROAD RH16 1SS",
@@ -1070,6 +1125,12 @@
1070
1125
  "wiki_name": "Newcastle City Council",
1071
1126
  "wiki_note": "Replace XXXXXXXX with your UPRN."
1072
1127
  },
1128
+ "NewcastleUnderLymeCouncil": {
1129
+ "url": "https://www.newcastle-staffs.gov.uk",
1130
+ "uprn": "100031725433",
1131
+ "wiki_name": "Newcastle Under Lyme Council",
1132
+ "wiki_note": "You will need to use [FindMyAddress](https://www.findmyaddress.co.uk/search) to find your UPRN."
1133
+ },
1073
1134
  "NewhamCouncil": {
1074
1135
  "skip_get_url": true,
1075
1136
  "url": "https://bincollection.newham.gov.uk/Details/Index/000046029461",
@@ -1108,6 +1169,13 @@
1108
1169
  "wiki_name": "North East Lincolnshire Council",
1109
1170
  "wiki_note": "Replace XXXXXXXX with your UPRN."
1110
1171
  },
1172
+ "NorthHertfordshireDistrictCouncil": {
1173
+ "house_number": "2",
1174
+ "postcode": "SG6 4BJ",
1175
+ "url": "https://www.north-herts.gov.uk",
1176
+ "wiki_name": "North Hertfordshire District Council",
1177
+ "wiki_note": "Pass the house number and postcode in their respective parameters."
1178
+ },
1111
1179
  "NorthKestevenDistrictCouncil": {
1112
1180
  "url": "https://www.n-kesteven.org.uk/bins/display?uprn=100030869513",
1113
1181
  "wiki_command_url_override": "https://www.n-kesteven.org.uk/bins/display?uprn=XXXXXXXX",
@@ -0,0 +1,122 @@
1
+ import time
2
+
3
+ import requests
4
+
5
+ from uk_bin_collection.uk_bin_collection.common import *
6
+ from uk_bin_collection.uk_bin_collection.get_bin_data import AbstractGetBinDataClass
7
+
8
+
9
+ # import the wonderful Beautiful Soup and the URL grabber
10
+ class CouncilClass(AbstractGetBinDataClass):
11
+ """
12
+ Concrete classes have to implement all abstract operations of the
13
+ base class. They can also override some operations with a default
14
+ implementation.
15
+ """
16
+
17
+ def parse_data(self, page: str, **kwargs) -> dict:
18
+
19
+ user_uprn = kwargs.get("uprn")
20
+ check_uprn(user_uprn)
21
+ bindata = {"bins": []}
22
+
23
+ SESSION_URL = "https://integration.aberdeencity.gov.uk/authapi/isauthenticated?uri=https%253A%252F%252Fintegration.aberdeencity.gov.uk%252Fservice%252Fbin_collection_calendar___view&hostname=integration.aberdeencity.gov.uk&withCredentials=true"
24
+
25
+ API_URL = "https://integration.aberdeencity.gov.uk/apibroker/runLookup"
26
+
27
+ headers = {
28
+ "Content-Type": "application/json",
29
+ "Accept": "application/json",
30
+ "User-Agent": "Mozilla/5.0",
31
+ "X-Requested-With": "XMLHttpRequest",
32
+ "Referer": "https://integration.aberdeencity.gov.uk/fillform/?iframe_id=fillform-frame-1&db_id=",
33
+ }
34
+ s = requests.session()
35
+ r = s.get(SESSION_URL)
36
+ r.raise_for_status()
37
+ session_data = r.json()
38
+ sid = session_data["auth-session"]
39
+ params = {
40
+ "id": "583c08ffc47fe",
41
+ "repeat_against": "",
42
+ "noRetry": "true",
43
+ "getOnlyTokens": "undefined",
44
+ "log_id": "",
45
+ "app_name": "AF-Renderer::Self",
46
+ # unix_timestamp
47
+ "_": str(int(time.time() * 1000)),
48
+ "sid": sid,
49
+ }
50
+
51
+ r = s.post(API_URL, headers=headers, params=params)
52
+ r.raise_for_status()
53
+
54
+ data = r.json()
55
+ rows_data = data["integration"]["transformed"]["rows_data"]["0"]
56
+ if not isinstance(rows_data, dict):
57
+ raise ValueError("Invalid data returned from API")
58
+ token = rows_data["token"]
59
+
60
+ data = {
61
+ "formValues": {
62
+ "Section 1": {
63
+ "nauprn": {
64
+ "value": user_uprn,
65
+ },
66
+ "token": {
67
+ "value": token,
68
+ },
69
+ "mindate": {
70
+ "value": datetime.now().strftime("%Y-%m-%d"),
71
+ },
72
+ "maxdate": {
73
+ "value": (datetime.now() + timedelta(days=30)).strftime(
74
+ "%Y-%m-%d"
75
+ ),
76
+ },
77
+ },
78
+ },
79
+ }
80
+
81
+ params = {
82
+ "id": "5a3141caf4016",
83
+ "repeat_against": "",
84
+ "noRetry": "true",
85
+ "getOnlyTokens": "undefined",
86
+ "log_id": "",
87
+ "app_name": "AF-Renderer::Self",
88
+ # unix_timestamp
89
+ "_": str(int(time.time() * 1000)),
90
+ "sid": sid,
91
+ }
92
+
93
+ r = s.post(API_URL, json=data, headers=headers, params=params)
94
+ r.raise_for_status()
95
+
96
+ data = r.json()
97
+ rows_data = data["integration"]["transformed"]["rows_data"]["0"]
98
+ if not isinstance(rows_data, dict):
99
+ raise ValueError("Invalid data returned from API")
100
+
101
+ date_pattern = re.compile(r"^(.*?)(Date\d+)$")
102
+ count_pattern = re.compile(r"^Count(.*)$")
103
+ for key, value in rows_data.items():
104
+ date_match = date_pattern.match(key)
105
+ # Match count keys
106
+ count_match = count_pattern.match(key)
107
+ if count_match:
108
+ continue
109
+
110
+ # Match date keys
111
+ date_match = date_pattern.match(key)
112
+ if date_match:
113
+ bin_type = date_match.group(1)
114
+ dict_data = {
115
+ "type": bin_type,
116
+ "collectionDate": datetime.strptime(value, "%A %d %B %Y").strftime(
117
+ date_format
118
+ ),
119
+ }
120
+ bindata["bins"].append(dict_data)
121
+
122
+ return bindata
@@ -24,6 +24,7 @@ class CouncilClass(AbstractGetBinDataClass):
24
24
 
25
25
  collection_day = kwargs.get("paon")
26
26
  garden_collection_week = kwargs.get("postcode")
27
+ garden_collection_day = kwargs.get("uprn")
27
28
  bindata = {"bins": []}
28
29
 
29
30
  days_of_week = [
@@ -42,6 +43,7 @@ class CouncilClass(AbstractGetBinDataClass):
42
43
  recyclingstartDate = datetime(2024, 11, 11)
43
44
 
44
45
  offset_days = days_of_week.index(collection_day)
46
+ offset_days_garden = days_of_week.index(garden_collection_day)
45
47
  if garden_collection_week:
46
48
  garden_collection = garden_week.index(garden_collection_week)
47
49
 
@@ -155,7 +157,7 @@ class CouncilClass(AbstractGetBinDataClass):
155
157
 
156
158
  collection_date = (
157
159
  datetime.strptime(gardenDate, "%d/%m/%Y")
158
- + timedelta(days=offset_days)
160
+ + timedelta(days=offset_days_garden)
159
161
  ).strftime("%d/%m/%Y")
160
162
 
161
163
  garden_holiday = next(
@@ -0,0 +1,70 @@
1
+ import time
2
+
3
+ import requests
4
+ from bs4 import BeautifulSoup
5
+
6
+ from uk_bin_collection.uk_bin_collection.common import *
7
+ from uk_bin_collection.uk_bin_collection.get_bin_data import AbstractGetBinDataClass
8
+
9
+
10
+ # import the wonderful Beautiful Soup and the URL grabber
11
+ class CouncilClass(AbstractGetBinDataClass):
12
+ """
13
+ Concrete classes have to implement all abstract operations of the
14
+ base class. They can also override some operations with a default
15
+ implementation.
16
+ """
17
+
18
+ def parse_data(self, page: str, **kwargs) -> dict:
19
+
20
+ user_postcode = kwargs.get("postcode")
21
+ user_uprn = kwargs.get("uprn")
22
+ check_postcode(user_postcode)
23
+ check_uprn(user_uprn)
24
+ bindata = {"bins": []}
25
+
26
+ URI = "https://www.braintree.gov.uk/xfp/form/554"
27
+
28
+ response = requests.get(URI)
29
+ soup = BeautifulSoup(response.content, "html.parser")
30
+ token = (soup.find("input", {"name": "__token"})).get("value")
31
+
32
+ headers = {
33
+ "Content-Type": "application/x-www-form-urlencoded",
34
+ "User-Agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/130.0.0.0 Safari/537.36",
35
+ "Referer": "https://www.braintree.gov.uk/xfp/form/554",
36
+ }
37
+
38
+ form_data = {
39
+ "__token": token,
40
+ "page": "5730",
41
+ "locale": "en_GB",
42
+ "qe15dda0155d237d1ea161004d1839e3369ed4831_0_0": user_postcode,
43
+ "qe15dda0155d237d1ea161004d1839e3369ed4831_1_0": user_uprn,
44
+ "next": "Next",
45
+ }
46
+ collection_lookup = requests.post(URI, data=form_data, headers=headers)
47
+ collection_lookup.raise_for_status()
48
+ for results in BeautifulSoup(collection_lookup.text, "html.parser").find_all(
49
+ "div", class_="date_display"
50
+ ):
51
+ collection_info = results.text.strip().split("\n")
52
+ collection_type = collection_info[0].strip()
53
+
54
+ # Skip if no collection date is found
55
+ if len(collection_info) < 2:
56
+ continue
57
+
58
+ collection_date = collection_info[1].strip()
59
+
60
+ dict_data = {
61
+ "type": collection_type,
62
+ "collectionDate": collection_date,
63
+ }
64
+ bindata["bins"].append(dict_data)
65
+
66
+ bindata["bins"].sort(
67
+ key=lambda x: datetime.strptime(x.get("collectionDate"), "%d/%m/%Y")
68
+ )
69
+
70
+ return bindata
@@ -0,0 +1,88 @@
1
+ import time
2
+
3
+ import requests
4
+
5
+ from uk_bin_collection.uk_bin_collection.common import *
6
+ from uk_bin_collection.uk_bin_collection.get_bin_data import AbstractGetBinDataClass
7
+
8
+
9
+ # import the wonderful Beautiful Soup and the URL grabber
10
+ class CouncilClass(AbstractGetBinDataClass):
11
+ """
12
+ Concrete classes have to implement all abstract operations of the
13
+ base class. They can also override some operations with a default
14
+ implementation.
15
+ """
16
+
17
+ def parse_data(self, page: str, **kwargs) -> dict:
18
+
19
+ user_uprn = kwargs.get("uprn")
20
+ check_uprn(user_uprn)
21
+ bindata = {"bins": []}
22
+
23
+ SESSION_URL = "https://your.burnley.gov.uk/authapi/isauthenticated?uri=https%253A%252F%252Fyour.burnley.gov.uk%252Fen%252FAchieveForms%252F%253Fform_uri%253Dsandbox-publish%253A%252F%252FAF-Process-b41dcd03-9a98-41be-93ba-6c172ba9f80c%252FAF-Stage-edb97458-fc4d-4316-b6e0-85598ec7fce8%252Fdefinition.json%2526redirectlink%253D%25252Fen%2526cancelRedirectLink%253D%25252Fen%2526consentMessage%253Dyes&hostname=your.burnley.gov.uk&withCredentials=true"
24
+
25
+ API_URL = "https://your.burnley.gov.uk/apibroker/runLookup"
26
+
27
+ headers = {
28
+ "Content-Type": "application/json",
29
+ "Accept": "application/json",
30
+ "User-Agent": "Mozilla/5.0",
31
+ "X-Requested-With": "XMLHttpRequest",
32
+ "Referer": "https://your.burnley.gov.uk/fillform/?iframe_id=fillform-frame-1&db_id=",
33
+ }
34
+ s = requests.session()
35
+ r = s.get(SESSION_URL)
36
+ r.raise_for_status()
37
+ session_data = r.json()
38
+ sid = session_data["auth-session"]
39
+
40
+ data = {
41
+ "formValues": {
42
+ "Section 1": {
43
+ "case_uprn1": {
44
+ "value": user_uprn,
45
+ }
46
+ },
47
+ },
48
+ }
49
+
50
+ params = {
51
+ "id": "607fe757df87c",
52
+ "repeat_against": "",
53
+ "noRetry": "false",
54
+ "getOnlyTokens": "undefined",
55
+ "log_id": "",
56
+ "app_name": "AF-Renderer::Self",
57
+ # unix_timestamp
58
+ "_": str(int(time.time() * 1000)),
59
+ "sid": sid,
60
+ }
61
+
62
+ r = s.post(API_URL, json=data, headers=headers, params=params)
63
+ r.raise_for_status()
64
+
65
+ data = r.json()
66
+ rows_data = data["integration"]["transformed"]["rows_data"]
67
+ if not isinstance(rows_data, dict):
68
+ raise ValueError("Invalid data returned from API")
69
+
70
+ current_year = (datetime.now()).year
71
+ for key, value in rows_data.items():
72
+ bin_type = value["display"].split(" - ")[0]
73
+ collection_date = datetime.strptime(
74
+ value["display"].split(" - ")[1], "%A %d %B"
75
+ )
76
+
77
+ if collection_date.month == 1:
78
+ collection_date = collection_date.replace(year=current_year + 1)
79
+ else:
80
+ collection_date = collection_date.replace(year=current_year)
81
+
82
+ dict_data = {
83
+ "type": bin_type,
84
+ "collectionDate": collection_date.strftime(date_format),
85
+ }
86
+ bindata["bins"].append(dict_data)
87
+
88
+ return bindata
@@ -0,0 +1,93 @@
1
+ import requests
2
+ from bs4 import BeautifulSoup
3
+
4
+ from uk_bin_collection.uk_bin_collection.common import *
5
+ from uk_bin_collection.uk_bin_collection.get_bin_data import AbstractGetBinDataClass
6
+
7
+
8
+ # import the wonderful Beautiful Soup and the URL grabber
9
+ class CouncilClass(AbstractGetBinDataClass):
10
+ """
11
+ Concrete classes have to implement all abstract operations of the
12
+ base class. They can also override some operations with a default
13
+ implementation.
14
+ """
15
+
16
+ def parse_data(self, page: str, **kwargs) -> dict:
17
+
18
+ user_postcode = kwargs.get("postcode")
19
+ user_paon = kwargs.get("paon")
20
+ check_postcode(user_postcode)
21
+ check_paon(user_paon)
22
+ bindata = {"bins": []}
23
+
24
+ URI = "https://abc-wrp.whitespacews.com/"
25
+
26
+ session = requests.Session()
27
+
28
+ # get link from first page as has some kind of unique hash
29
+ r = session.get(
30
+ URI,
31
+ )
32
+ r.raise_for_status()
33
+ soup = BeautifulSoup(r.text, features="html.parser")
34
+
35
+ alink = soup.find("a", text="View My Collections")
36
+
37
+ if alink is None:
38
+ raise Exception("Initial page did not load correctly")
39
+
40
+ # greplace 'seq' query string to skip next step
41
+ nextpageurl = alink["href"].replace("seq=1", "seq=2")
42
+
43
+ data = {
44
+ "address_name_number": user_paon,
45
+ "address_postcode": user_postcode,
46
+ }
47
+
48
+ # get list of addresses
49
+ r = session.post(nextpageurl, data)
50
+ r.raise_for_status()
51
+
52
+ soup = BeautifulSoup(r.text, features="html.parser")
53
+
54
+ # get first address (if you don't enter enough argument values this won't find the right address)
55
+ alink = soup.find("div", id="property_list").find("a")
56
+
57
+ if alink is None:
58
+ raise Exception("Address not found")
59
+
60
+ nextpageurl = URI + alink["href"]
61
+
62
+ # get collection page
63
+ r = session.get(
64
+ nextpageurl,
65
+ )
66
+ r.raise_for_status()
67
+ soup = BeautifulSoup(r.text, features="html.parser")
68
+
69
+ if soup.find("span", id="waste-hint"):
70
+ raise Exception("No scheduled services at this address")
71
+
72
+ u1s = soup.find("section", id="scheduled-collections").find_all("u1")
73
+
74
+ for u1 in u1s:
75
+ lis = u1.find_all("li", recursive=False)
76
+
77
+ date = lis[1].text.replace("\n", "")
78
+ bin_type = lis[2].text.replace("\n", "")
79
+
80
+ dict_data = {
81
+ "type": bin_type,
82
+ "collectionDate": datetime.strptime(
83
+ date,
84
+ "%d/%m/%Y",
85
+ ).strftime(date_format),
86
+ }
87
+ bindata["bins"].append(dict_data)
88
+
89
+ bindata["bins"].sort(
90
+ key=lambda x: datetime.strptime(x.get("collectionDate"), date_format)
91
+ )
92
+
93
+ return bindata
@@ -0,0 +1,98 @@
1
+ import re
2
+ import time
3
+
4
+ import requests
5
+ from bs4 import BeautifulSoup
6
+ from selenium.webdriver.common.by import By
7
+ from selenium.webdriver.support import expected_conditions as EC
8
+ from selenium.webdriver.support.ui import Select
9
+ from selenium.webdriver.support.wait import WebDriverWait
10
+
11
+ from uk_bin_collection.uk_bin_collection.common import *
12
+ from uk_bin_collection.uk_bin_collection.get_bin_data import AbstractGetBinDataClass
13
+
14
+
15
+ # import the wonderful Beautiful Soup and the URL grabber
16
+ class CouncilClass(AbstractGetBinDataClass):
17
+ """
18
+ Concrete classes have to implement all abstract operations of the
19
+ base class. They can also override some operations with a default
20
+ implementation.
21
+ """
22
+
23
+ def parse_data(self, page: str, **kwargs) -> dict:
24
+
25
+ collection_day = kwargs.get("paon")
26
+ collection_week = kwargs.get("postcode")
27
+ bindata = {"bins": []}
28
+
29
+ days_of_week = [
30
+ "Monday",
31
+ "Tuesday",
32
+ "Wednesday",
33
+ "Thursday",
34
+ "Friday",
35
+ "Saturday",
36
+ "Sunday",
37
+ ]
38
+
39
+ collection_weeks = ["Week 1", "Week 2"]
40
+ collection_week = collection_weeks.index(collection_week)
41
+
42
+ offset_days = days_of_week.index(collection_day)
43
+
44
+ if collection_week == 0:
45
+ recyclingstartDate = datetime(2024, 11, 4)
46
+ glassstartDate = datetime(2024, 11, 4)
47
+ refusestartDate = datetime(2024, 11, 11)
48
+ elif collection_week == 1:
49
+ recyclingstartDate = datetime(2024, 11, 11)
50
+ glassstartDate = datetime(2024, 11, 11)
51
+ refusestartDate = datetime(2024, 11, 4)
52
+
53
+ refuse_dates = get_dates_every_x_days(refusestartDate, 14, 28)
54
+ glass_dates = get_dates_every_x_days(glassstartDate, 14, 28)
55
+ recycling_dates = get_dates_every_x_days(recyclingstartDate, 14, 28)
56
+
57
+ for refuseDate in refuse_dates:
58
+
59
+ collection_date = (
60
+ datetime.strptime(refuseDate, "%d/%m/%Y") + timedelta(days=offset_days)
61
+ ).strftime("%d/%m/%Y")
62
+
63
+ dict_data = {
64
+ "type": "Grey Bin",
65
+ "collectionDate": collection_date,
66
+ }
67
+ bindata["bins"].append(dict_data)
68
+
69
+ for recyclingDate in recycling_dates:
70
+
71
+ collection_date = (
72
+ datetime.strptime(recyclingDate, "%d/%m/%Y")
73
+ + timedelta(days=offset_days)
74
+ ).strftime("%d/%m/%Y")
75
+
76
+ dict_data = {
77
+ "type": "Green Bin",
78
+ "collectionDate": collection_date,
79
+ }
80
+ bindata["bins"].append(dict_data)
81
+
82
+ for glassDate in glass_dates:
83
+
84
+ collection_date = (
85
+ datetime.strptime(glassDate, "%d/%m/%Y") + timedelta(days=offset_days)
86
+ ).strftime("%d/%m/%Y")
87
+
88
+ dict_data = {
89
+ "type": "Glass Box",
90
+ "collectionDate": collection_date,
91
+ }
92
+ bindata["bins"].append(dict_data)
93
+
94
+ bindata["bins"].sort(
95
+ key=lambda x: datetime.strptime(x.get("collectionDate"), "%d/%m/%Y")
96
+ )
97
+
98
+ return bindata
@@ -1,4 +1,5 @@
1
1
  from bs4 import BeautifulSoup
2
+
2
3
  from uk_bin_collection.uk_bin_collection.common import *
3
4
  from uk_bin_collection.uk_bin_collection.get_bin_data import AbstractGetBinDataClass
4
5
 
@@ -45,4 +46,17 @@ class CouncilClass(AbstractGetBinDataClass):
45
46
  }
46
47
  data["bins"].append(dict_data)
47
48
 
49
+ if len(page_text) > 5:
50
+ garden_day = datetime.strptime(
51
+ remove_ordinal_indicator_from_date_string(
52
+ page_text[6].find_next("strong").text
53
+ ),
54
+ "%d %B %Y",
55
+ ).strftime(date_format)
56
+ dict_data = {
57
+ "type": "Garden",
58
+ "collectionDate": garden_day,
59
+ }
60
+ data["bins"].append(dict_data)
61
+
48
62
  return data
@@ -0,0 +1,52 @@
1
+ import time
2
+
3
+ import requests
4
+ from bs4 import BeautifulSoup
5
+
6
+ from uk_bin_collection.uk_bin_collection.common import *
7
+ from uk_bin_collection.uk_bin_collection.get_bin_data import AbstractGetBinDataClass
8
+
9
+
10
+ # import the wonderful Beautiful Soup and the URL grabber
11
+ class CouncilClass(AbstractGetBinDataClass):
12
+ """
13
+ Concrete classes have to implement all abstract operations of the
14
+ base class. They can also override some operations with a default
15
+ implementation.
16
+ """
17
+
18
+ def parse_data(self, page: str, **kwargs) -> dict:
19
+
20
+ user_uprn = kwargs.get("uprn")
21
+ check_uprn(user_uprn)
22
+ bindata = {"bins": []}
23
+
24
+ URI = f"https://exeter.gov.uk/repositories/hidden-pages/address-finder/?qsource=UPRN&qtype=bins&term={user_uprn}"
25
+
26
+ response = requests.get(URI)
27
+ response.raise_for_status()
28
+
29
+ data = response.json()
30
+
31
+ soup = BeautifulSoup(data[0]["Results"], "html.parser")
32
+ soup.prettify()
33
+
34
+ # Extract bin schedule
35
+ for section in soup.find_all("h2"):
36
+ bin_type = section.text.strip()
37
+ collection_date = section.find_next("h3").text.strip()
38
+
39
+ dict_data = {
40
+ "type": bin_type,
41
+ "collectionDate": datetime.strptime(
42
+ remove_ordinal_indicator_from_date_string(collection_date),
43
+ "%A, %d %B %Y",
44
+ ).strftime(date_format),
45
+ }
46
+ bindata["bins"].append(dict_data)
47
+
48
+ bindata["bins"].sort(
49
+ key=lambda x: datetime.strptime(x.get("collectionDate"), date_format)
50
+ )
51
+
52
+ return bindata
@@ -0,0 +1,83 @@
1
+ import time
2
+
3
+ import requests
4
+ from bs4 import BeautifulSoup
5
+
6
+ from uk_bin_collection.uk_bin_collection.common import *
7
+ from uk_bin_collection.uk_bin_collection.get_bin_data import AbstractGetBinDataClass
8
+
9
+
10
+ # import the wonderful Beautiful Soup and the URL grabber
11
+ class CouncilClass(AbstractGetBinDataClass):
12
+ """
13
+ Concrete classes have to implement all abstract operations of the
14
+ base class. They can also override some operations with a default
15
+ implementation.
16
+ """
17
+
18
+ def parse_data(self, page: str, **kwargs) -> dict:
19
+
20
+ user_uprn = kwargs.get("uprn")
21
+ check_uprn(user_uprn)
22
+ bindata = {"bins": []}
23
+
24
+ SESSION_URL = "https://online.hartlepool.gov.uk/authapi/isauthenticated?uri=https%253A%252F%252Fonline.hartlepool.gov.uk%252Fservice%252FRefuse_and_recycling___check_bin_day&hostname=online.hartlepool.gov.uk&withCredentials=true"
25
+
26
+ API_URL = "https://online.hartlepool.gov.uk/apibroker/runLookup"
27
+
28
+ headers = {
29
+ "Content-Type": "application/json",
30
+ "Accept": "application/json",
31
+ "User-Agent": "Mozilla/5.0",
32
+ "X-Requested-With": "XMLHttpRequest",
33
+ "Referer": "https://online.hartlepool.gov.uk/fillform/?iframe_id=fillform-frame-1&db_id=",
34
+ }
35
+ s = requests.session()
36
+ r = s.get(SESSION_URL)
37
+ r.raise_for_status()
38
+ session_data = r.json()
39
+ sid = session_data["auth-session"]
40
+ params = {
41
+ "id": "5ec67e019ffdd",
42
+ "repeat_against": "",
43
+ "noRetry": "true",
44
+ "getOnlyTokens": "undefined",
45
+ "log_id": "",
46
+ "app_name": "AF-Renderer::Self",
47
+ # unix_timestamp
48
+ "_": str(int(time.time() * 1000)),
49
+ "sid": sid,
50
+ }
51
+
52
+ data = {
53
+ "formValues": {
54
+ "Section 1": {
55
+ "collectionLocationUPRN": {
56
+ "value": user_uprn,
57
+ },
58
+ },
59
+ },
60
+ }
61
+
62
+ r = s.post(API_URL, json=data, headers=headers, params=params)
63
+ r.raise_for_status()
64
+
65
+ data = r.json()
66
+ rows_data = data["integration"]["transformed"]["rows_data"]["0"]
67
+ if not isinstance(rows_data, dict):
68
+ raise ValueError("Invalid data returned from API")
69
+
70
+ soup = BeautifulSoup(rows_data["HTMLCollectionDatesText"], "html.parser")
71
+
72
+ # Find all div elements containing the bin schedule
73
+ for div in soup.find_all("div"):
74
+ # Extract bin type and date from the span tag
75
+ text = div.find("span").text.strip()
76
+ bin_type, date = text.split(" ", 1)
77
+ dict_data = {
78
+ "type": bin_type,
79
+ "collectionDate": date,
80
+ }
81
+ bindata["bins"].append(dict_data)
82
+
83
+ return bindata
@@ -0,0 +1,75 @@
1
+ import time
2
+
3
+ import requests
4
+ from bs4 import BeautifulSoup
5
+
6
+ from uk_bin_collection.uk_bin_collection.common import *
7
+ from uk_bin_collection.uk_bin_collection.get_bin_data import AbstractGetBinDataClass
8
+
9
+
10
+ # import the wonderful Beautiful Soup and the URL grabber
11
+ class CouncilClass(AbstractGetBinDataClass):
12
+ """
13
+ Concrete classes have to implement all abstract operations of the
14
+ base class. They can also override some operations with a default
15
+ implementation.
16
+ """
17
+
18
+ def parse_data(self, page: str, **kwargs) -> dict:
19
+
20
+ user_uprn = kwargs.get("uprn")
21
+ check_uprn(user_uprn)
22
+ bindata = {"bins": []}
23
+
24
+ URI = "https://lbhapiprod.azure-api.net"
25
+ endpoint = f"{URI}/whitespace/GetCollectionByUprnAndDate"
26
+ subscription_key = "2ea6a75f9ea34bb58d299a0c9f84e72e"
27
+
28
+ # Get today's date in 'YYYY-MM-DD' format
29
+ collection_date = datetime.now().strftime("%Y-%m-%d")
30
+
31
+ # Define the request headers
32
+ headers = {
33
+ "Content-Type": "application/json",
34
+ "Ocp-Apim-Subscription-Key": subscription_key,
35
+ }
36
+
37
+ # Define the request body
38
+ data = {
39
+ "getCollectionByUprnAndDate": {
40
+ "getCollectionByUprnAndDateInput": {
41
+ "uprn": user_uprn,
42
+ "nextCollectionFromDate": collection_date,
43
+ }
44
+ }
45
+ }
46
+ # Make the POST request
47
+ response = requests.post(endpoint, headers=headers, data=json.dumps(data))
48
+ response.raise_for_status() # Raise an exception for HTTP errors
49
+
50
+ # Parse the JSON response
51
+ response_data = response.json()
52
+
53
+ collections = (
54
+ response_data.get("getCollectionByUprnAndDateResponse", {})
55
+ .get("getCollectionByUprnAndDateResult", {})
56
+ .get("Collections", [])
57
+ )
58
+
59
+ for collection in collections:
60
+ bin_type = collection["service"]
61
+ collection_date = collection["date"]
62
+
63
+ dict_data = {
64
+ "type": bin_type,
65
+ "collectionDate": datetime.strptime(
66
+ collection_date,
67
+ "%d/%m/%Y %H:%M:%S",
68
+ ).strftime(date_format),
69
+ }
70
+ bindata["bins"].append(dict_data)
71
+ bindata["bins"].sort(
72
+ key=lambda x: datetime.strptime(x.get("collectionDate"), date_format)
73
+ )
74
+
75
+ return bindata
@@ -24,6 +24,7 @@ class CouncilClass(AbstractGetBinDataClass):
24
24
 
25
25
  collection_day = kwargs.get("paon")
26
26
  garden_collection_week = kwargs.get("postcode")
27
+ garden_collection_day = kwargs.get("uprn")
27
28
  bindata = {"bins": []}
28
29
 
29
30
  days_of_week = [
@@ -42,6 +43,7 @@ class CouncilClass(AbstractGetBinDataClass):
42
43
  recyclingstartDate = datetime(2024, 11, 4)
43
44
 
44
45
  offset_days = days_of_week.index(collection_day)
46
+ offset_days_garden = days_of_week.index(garden_collection_day)
45
47
  if garden_collection_week:
46
48
  garden_collection = garden_week.index(garden_collection_week)
47
49
 
@@ -155,7 +157,7 @@ class CouncilClass(AbstractGetBinDataClass):
155
157
 
156
158
  collection_date = (
157
159
  datetime.strptime(gardenDate, "%d/%m/%Y")
158
- + timedelta(days=offset_days)
160
+ + timedelta(days=offset_days_garden)
159
161
  ).strftime("%d/%m/%Y")
160
162
 
161
163
  garden_holiday = next(
@@ -0,0 +1,66 @@
1
+ import requests
2
+ from bs4 import BeautifulSoup
3
+ from dateutil.relativedelta import relativedelta
4
+
5
+ from uk_bin_collection.uk_bin_collection.common import *
6
+ from uk_bin_collection.uk_bin_collection.get_bin_data import AbstractGetBinDataClass
7
+
8
+
9
+ # import the wonderful Beautiful Soup and the URL grabber
10
+ class CouncilClass(AbstractGetBinDataClass):
11
+ """
12
+ Concrete classes have to implement all abstract operations of the
13
+ base class. They can also override some operations with a default
14
+ implementation.
15
+ """
16
+
17
+ def parse_data(self, page: str, **kwargs) -> dict:
18
+
19
+ user_uprn = kwargs.get("uprn")
20
+ check_uprn(user_uprn)
21
+ bindata = {"bins": []}
22
+
23
+ URI = f"https://www.newcastle-staffs.gov.uk/homepage/97/check-your-bin-day?uprn={user_uprn}"
24
+
25
+ # Make the GET request
26
+ response = requests.get(URI)
27
+ response.raise_for_status()
28
+ soup = BeautifulSoup(response.text, features="html.parser")
29
+ soup.prettify()
30
+
31
+ # Find the table
32
+ table = soup.find("table", {"class": "data-table"})
33
+
34
+ if table:
35
+ rows = table.find("tbody").find_all("tr")
36
+ for row in rows:
37
+ date = datetime.strptime(
38
+ (
39
+ row.find_all("td")[0]
40
+ .get_text(strip=True)
41
+ .replace("Date:", "")
42
+ .strip()
43
+ ),
44
+ "%A %d %B",
45
+ ).replace(year=datetime.now().year)
46
+ if datetime.now().month > 10 and date.month < 3:
47
+ date = date + relativedelta(years=1)
48
+ bin_types = (
49
+ row.find_all("td")[1]
50
+ .text.replace("Collection Type:", "")
51
+ .splitlines()
52
+ )
53
+ for bin_type in bin_types:
54
+ bin_type = bin_type.strip()
55
+ if bin_type:
56
+ dict_data = {
57
+ "type": bin_type.strip(),
58
+ "collectionDate": date.strftime("%d/%m/%Y"),
59
+ }
60
+ bindata["bins"].append(dict_data)
61
+
62
+ bindata["bins"].sort(
63
+ key=lambda x: datetime.strptime(x.get("collectionDate"), "%d/%m/%Y")
64
+ )
65
+
66
+ return bindata
@@ -0,0 +1,93 @@
1
+ import requests
2
+ from bs4 import BeautifulSoup
3
+
4
+ from uk_bin_collection.uk_bin_collection.common import *
5
+ from uk_bin_collection.uk_bin_collection.get_bin_data import AbstractGetBinDataClass
6
+
7
+
8
+ # import the wonderful Beautiful Soup and the URL grabber
9
+ class CouncilClass(AbstractGetBinDataClass):
10
+ """
11
+ Concrete classes have to implement all abstract operations of the
12
+ base class. They can also override some operations with a default
13
+ implementation.
14
+ """
15
+
16
+ def parse_data(self, page: str, **kwargs) -> dict:
17
+
18
+ user_postcode = kwargs.get("postcode")
19
+ user_paon = kwargs.get("paon")
20
+ check_postcode(user_postcode)
21
+ check_paon(user_paon)
22
+ bindata = {"bins": []}
23
+
24
+ URI = "https://uhtn-wrp.whitespacews.com/"
25
+
26
+ session = requests.Session()
27
+
28
+ # get link from first page as has some kind of unique hash
29
+ r = session.get(
30
+ URI,
31
+ )
32
+ r.raise_for_status()
33
+ soup = BeautifulSoup(r.text, features="html.parser")
34
+
35
+ alink = soup.find("a", text="Find my bin collection day")
36
+
37
+ if alink is None:
38
+ raise Exception("Initial page did not load correctly")
39
+
40
+ # greplace 'seq' query string to skip next step
41
+ nextpageurl = alink["href"].replace("seq=1", "seq=2")
42
+
43
+ data = {
44
+ "address_name_number": user_paon,
45
+ "address_postcode": user_postcode,
46
+ }
47
+
48
+ # get list of addresses
49
+ r = session.post(nextpageurl, data)
50
+ r.raise_for_status()
51
+
52
+ soup = BeautifulSoup(r.text, features="html.parser")
53
+
54
+ # get first address (if you don't enter enough argument values this won't find the right address)
55
+ alink = soup.find("div", id="property_list").find("a")
56
+
57
+ if alink is None:
58
+ raise Exception("Address not found")
59
+
60
+ nextpageurl = URI + alink["href"]
61
+
62
+ # get collection page
63
+ r = session.get(
64
+ nextpageurl,
65
+ )
66
+ r.raise_for_status()
67
+ soup = BeautifulSoup(r.text, features="html.parser")
68
+
69
+ if soup.find("span", id="waste-hint"):
70
+ raise Exception("No scheduled services at this address")
71
+
72
+ u1s = soup.find("section", id="scheduled-collections").find_all("u1")
73
+
74
+ for u1 in u1s:
75
+ lis = u1.find_all("li", recursive=False)
76
+
77
+ date = lis[1].text.replace("\n", "")
78
+ bin_type = lis[2].text.replace("\n", "")
79
+
80
+ dict_data = {
81
+ "type": bin_type,
82
+ "collectionDate": datetime.strptime(
83
+ date,
84
+ "%d/%m/%Y",
85
+ ).strftime(date_format),
86
+ }
87
+ bindata["bins"].append(dict_data)
88
+
89
+ bindata["bins"].sort(
90
+ key=lambda x: datetime.strptime(x.get("collectionDate"), date_format)
91
+ )
92
+
93
+ return bindata
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: uk_bin_collection
3
- Version: 0.118.0
3
+ Version: 0.120.0
4
4
  Summary: Python Lib to collect UK Bin Data
5
5
  Author: Robert Bradley
6
6
  Author-email: robbrad182@gmail.com
@@ -2,7 +2,7 @@ uk_bin_collection/README.rst,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,
2
2
  uk_bin_collection/tests/council_feature_input_parity.py,sha256=DO6Mk4ImYgM5ZCZ-cutwz5RoYYWZRLYx2tr6zIs_9Rc,3843
3
3
  uk_bin_collection/tests/features/environment.py,sha256=VQZjJdJI_kZn08M0j5cUgvKT4k3iTw8icJge1DGOkoA,127
4
4
  uk_bin_collection/tests/features/validate_council_outputs.feature,sha256=SJK-Vc737hrf03tssxxbeg_JIvAH-ddB8f6gU1LTbuQ,251
5
- uk_bin_collection/tests/input.json,sha256=d54M0Eun5WcoFXATl2jFSaNkMv3QYAZFl20ujKzW9Co,103837
5
+ uk_bin_collection/tests/input.json,sha256=icnJmApeU711a_vyDoloeQv9ept1fjonba0zGYEbB_c,107279
6
6
  uk_bin_collection/tests/output.schema,sha256=ZwKQBwYyTDEM4G2hJwfLUVM-5v1vKRvRK9W9SS1sd18,1086
7
7
  uk_bin_collection/tests/step_defs/step_helpers/file_handler.py,sha256=Ygzi4V0S1MIHqbdstUlIqtRIwnynvhu4UtpweJ6-5N8,1474
8
8
  uk_bin_collection/tests/step_defs/test_validate_council.py,sha256=VZ0a81sioJULD7syAYHjvK_-nT_Rd36tUyzPetSA0gk,3475
@@ -12,6 +12,7 @@ uk_bin_collection/tests/test_conftest.py,sha256=qI_zgGjNOnwE9gmZUiuirL1SYz3TFw5y
12
12
  uk_bin_collection/tests/test_get_data.py,sha256=sFJz_Fd6o-1r2gdmzY52JGwVi0Of_mDzvYSoc7a3RUw,7239
13
13
  uk_bin_collection/uk_bin_collection/collect_data.py,sha256=dB7wWXsJX4fm5bIf84lexkvHIcO54CZ3JPxqmS-60YY,4654
14
14
  uk_bin_collection/uk_bin_collection/common.py,sha256=fJG9ruqsCYOaYm-fzRb_l5kTeeB7i9k7qphWt3t7kks,10107
15
+ uk_bin_collection/uk_bin_collection/councils/AberdeenCityCouncil.py,sha256=Je8VwVLK9KnYl9vqf2gWJ7ZYDgUq3A7caDiIzk5Xof8,4194
15
16
  uk_bin_collection/uk_bin_collection/councils/AberdeenshireCouncil.py,sha256=aO1CSdyqa8oAD0fB79y1Q9bikAWCP_JFa7CsyTa2j9s,1655
16
17
  uk_bin_collection/uk_bin_collection/councils/AdurAndWorthingCouncils.py,sha256=ppbrmm-MzB1wOulK--CU_0j4P-djNf3ozMhHnmQFqLo,1511
17
18
  uk_bin_collection/uk_bin_collection/councils/AntrimAndNewtonabbeyCouncil.py,sha256=Hp5pteaC5RjL5ZqPZ564S9WQ6ZTKLMO6Dl_fxip2TUc,1653
@@ -23,7 +24,7 @@ uk_bin_collection/uk_bin_collection/councils/AshfieldDistrictCouncil.py,sha256=2
23
24
  uk_bin_collection/uk_bin_collection/councils/AshfordBoroughCouncil.py,sha256=yC-8UMQHSbvze43PJ2_F4Z3cu7M7cynKTojipBJU7Ug,4307
24
25
  uk_bin_collection/uk_bin_collection/councils/AylesburyValeCouncil.py,sha256=LouqjspEMt1TkOGqWHs2zkxwOETIy3n7p64uKIlAgUg,2401
25
26
  uk_bin_collection/uk_bin_collection/councils/BCPCouncil.py,sha256=W7QBx6Mgso8RYosuXsaYo3GGNAu-tiyBSmuYxr1JSOU,1707
26
- uk_bin_collection/uk_bin_collection/councils/BaberghDistrictCouncil.py,sha256=aPONTzKXtPy7aBHFTRvhKTJ13qXqNicSquNRVAYV2d8,6200
27
+ uk_bin_collection/uk_bin_collection/councils/BaberghDistrictCouncil.py,sha256=1eXdST58xFRMdYl8AGNG_EwyQeLa31WSWUe882hQ2ec,6329
27
28
  uk_bin_collection/uk_bin_collection/councils/BarnetCouncil.py,sha256=Sd4-pbv0QZsR7soxvXYqsfdOUIqZqS6notyoZthG77s,9182
28
29
  uk_bin_collection/uk_bin_collection/councils/BarnsleyMBCouncil.py,sha256=RKuH8HzGc3Q0WtLg-g_xVMn9hUYqdENgfcvvR4Bx5PI,4763
29
30
  uk_bin_collection/uk_bin_collection/councils/BasildonCouncil.py,sha256=NymPmq5pud0PJ8ePcc2r1SKED4EHQ0EY2l71O-Metxc,3313
@@ -39,6 +40,7 @@ uk_bin_collection/uk_bin_collection/councils/BlackburnCouncil.py,sha256=jHbCK8sL
39
40
  uk_bin_collection/uk_bin_collection/councils/BoltonCouncil.py,sha256=WI68r8jB0IHPUT4CgmZMtng899AAMFTxkyTdPg9yLF8,4117
40
41
  uk_bin_collection/uk_bin_collection/councils/BracknellForestCouncil.py,sha256=Llo1rULaAZ8rChVYZqXFFLo7CN6vbT0ULUJD6ActouY,9015
41
42
  uk_bin_collection/uk_bin_collection/councils/BradfordMDC.py,sha256=BEWS2c62cOsf26jqn1AkNUvVmc5AlUADYLaQuPn9RY4,5456
43
+ uk_bin_collection/uk_bin_collection/councils/BraintreeDistrictCouncil.py,sha256=2vYHilpI8mSwC2Ykdr1gxYAN3excDWqF6AwtGbkwbTw,2441
42
44
  uk_bin_collection/uk_bin_collection/councils/BrecklandCouncil.py,sha256=PX6A_pDvaN109aSNWmEhm88GFKfkClIkmbwGURWvsks,1744
43
45
  uk_bin_collection/uk_bin_collection/councils/BrightonandHoveCityCouncil.py,sha256=k6qt4cds-Ejd97Z-__pw2BYvGVbFdc9SUfF73PPrTNA,5823
44
46
  uk_bin_collection/uk_bin_collection/councils/BristolCityCouncil.py,sha256=kJmmDJz_kQ45DHmG7ocrUpNJonEn0kuXYEDQyZaf9ks,5576
@@ -47,6 +49,7 @@ uk_bin_collection/uk_bin_collection/councils/BromsgroveDistrictCouncil.py,sha256
47
49
  uk_bin_collection/uk_bin_collection/councils/BroxbourneCouncil.py,sha256=JC6Qqou1Rj4awn2VP3iuvwFpYayKDTt2_JNuNitjSoY,2393
48
50
  uk_bin_collection/uk_bin_collection/councils/BroxtoweBoroughCouncil.py,sha256=-Facq-ToQkcWUePpKBwq90LZUFxgUSydNL2sYaLX4yw,4473
49
51
  uk_bin_collection/uk_bin_collection/councils/BuckinghamshireCouncil.py,sha256=_ELVUM5VLp1nwDxRpvpsp6n8SzLJvp_UyMp-i_MXYuo,4383
52
+ uk_bin_collection/uk_bin_collection/councils/BurnleyBoroughCouncil.py,sha256=GJf1OPvUVj3vqsR3KjG0DFHZrSBu4ogIz_MJeVV8tNA,3192
50
53
  uk_bin_collection/uk_bin_collection/councils/BuryCouncil.py,sha256=H7wAxO1nfxkewVoRRolumq8bBJG04siE3jieFH3RGpQ,2632
51
54
  uk_bin_collection/uk_bin_collection/councils/CalderdaleCouncil.py,sha256=OJZcHYlvZDzmBpjjPPm3J8CRK9Twc49vRj7O9c5fyQ4,4971
52
55
  uk_bin_collection/uk_bin_collection/councils/CannockChaseDistrictCouncil.py,sha256=ZamevXN8_Q8mRZOTESWtkb8jVyDXkTczcmhXMAVVSkM,2276
@@ -69,6 +72,7 @@ uk_bin_collection/uk_bin_collection/councils/CotswoldDistrictCouncil.py,sha256=K
69
72
  uk_bin_collection/uk_bin_collection/councils/CoventryCityCouncil.py,sha256=kfAvA2e4MlO0W9YT70U_mW9gxVPrmr0BOGzV99Tw2Bg,2012
70
73
  uk_bin_collection/uk_bin_collection/councils/CrawleyBoroughCouncil.py,sha256=Oaj5INA3zNjtzBRsfLvRTIxZzcd4E4bJfVF1ULWlrL4,4322
71
74
  uk_bin_collection/uk_bin_collection/councils/CroydonCouncil.py,sha256=Vxh5ICoaXTAvx0nDOq_95XQ4He9sQKcLdI5keV2uxM4,11384
75
+ uk_bin_collection/uk_bin_collection/councils/CumberlandAllerdaleCouncil.py,sha256=bPOmkyzNnHrOtUprbouHdOsgpu7WilUADcaccWTCPFI,2839
72
76
  uk_bin_collection/uk_bin_collection/councils/DacorumBoroughCouncil.py,sha256=Tm_6pvBPj-6qStbe6-02LXaoCOlnnDvVXAAocGVvf_E,3970
73
77
  uk_bin_collection/uk_bin_collection/councils/DartfordBoroughCouncil.py,sha256=SPirUUoweMwX5Txtsr0ocdcFtKxCQ9LhzTTJN20tM4w,1550
74
78
  uk_bin_collection/uk_bin_collection/councils/DerbyCityCouncil.py,sha256=M8FGLhZn9wdRCq1W6z_yqJQqeba3EKyba3vhM22MzB4,1883
@@ -88,11 +92,13 @@ uk_bin_collection/uk_bin_collection/councils/EastRenfrewshireCouncil.py,sha256=5
88
92
  uk_bin_collection/uk_bin_collection/councils/EastRidingCouncil.py,sha256=oL-NqriLVy_NChGASNh8qTqeakLn4iP_XzoMC6VlPGM,5216
89
93
  uk_bin_collection/uk_bin_collection/councils/EastSuffolkCouncil.py,sha256=qQ0oOfGd0sWcczse_B22YoeL9uj3og8v3UJLt_Sx29c,4353
90
94
  uk_bin_collection/uk_bin_collection/councils/EastleighBoroughCouncil.py,sha256=V4Vso4DvawFiezKlmXbTlJEK9Sjhz9nA8WeYjwtO2e4,2310
95
+ uk_bin_collection/uk_bin_collection/councils/EdinburghCityCouncil.py,sha256=YRjNgevnCxfaAIU8BV9dkqG17NiT6S-hp7l-1rdLVgQ,3150
91
96
  uk_bin_collection/uk_bin_collection/councils/ElmbridgeBoroughCouncil.py,sha256=TgBOaReHWBbm0avV7HqRf0x7cxDe9cacTUcP9TFFprs,3005
92
97
  uk_bin_collection/uk_bin_collection/councils/EnfieldCouncil.py,sha256=2yR5p-kdApOm1gHiynNECP0jQDvaYHOiT6MAQJAvunE,6144
93
- uk_bin_collection/uk_bin_collection/councils/EnvironmentFirst.py,sha256=_9QJYDHpdnYK5R6znvZk1w0F9GnPnI8G4b6I_p26h4U,1695
98
+ uk_bin_collection/uk_bin_collection/councils/EnvironmentFirst.py,sha256=R0Y94nEn2G4JgzJ3PKkUXP6a0fPpjaunpRxbfMLKT4s,2135
94
99
  uk_bin_collection/uk_bin_collection/councils/EppingForestDistrictCouncil.py,sha256=cKFllQ4zt6MGkwiz_HedZvw3iL1kRMLA6Ct2spUE5og,2085
95
100
  uk_bin_collection/uk_bin_collection/councils/ErewashBoroughCouncil.py,sha256=QTQA6NjZtTL2baDeerIQW1SQpawwu6kGDMGdVvYQRRo,2501
101
+ uk_bin_collection/uk_bin_collection/councils/ExeterCityCouncil.py,sha256=FPNyBuQFYGFECZfCemgeHr27bXJFVvadinw43h-G8qs,1623
96
102
  uk_bin_collection/uk_bin_collection/councils/FalkirkCouncil.py,sha256=C3OA9PEhBsCYPzwsSdqVi_SbF8uiB186i2XfHWKd3VI,1694
97
103
  uk_bin_collection/uk_bin_collection/councils/FarehamBoroughCouncil.py,sha256=25QxeN5q3ad1Wwexs2d-B7ooH0ru6pOUx58413FOTY4,2352
98
104
  uk_bin_collection/uk_bin_collection/councils/FenlandDistrictCouncil.py,sha256=sFrnKzIE2tIcz0YrC6A9HcevzgNdf6E6_HLGMWDKtGw,2513
@@ -109,6 +115,7 @@ uk_bin_collection/uk_bin_collection/councils/HaltonBoroughCouncil.py,sha256=gq_C
109
115
  uk_bin_collection/uk_bin_collection/councils/HarboroughDistrictCouncil.py,sha256=uAbCgfrqkIkEKUyLVE8l72s5tzbfMFsw775i0nVRAyc,1934
110
116
  uk_bin_collection/uk_bin_collection/councils/HaringeyCouncil.py,sha256=t_6AkAu4wrv8Q0WlDhWh_82I0djl5tk531Pzs-SjWzg,2647
111
117
  uk_bin_collection/uk_bin_collection/councils/HarrogateBoroughCouncil.py,sha256=_g3fP5Nq-OUjgNrfRf4UEyFKzq0x8QK-4enh5RP1efA,2050
118
+ uk_bin_collection/uk_bin_collection/councils/HartlepoolBoroughCouncil.py,sha256=MUT1A24iZShT2p55rXEvgYwGUuw3W05Z4ZQAveehv-s,2842
112
119
  uk_bin_collection/uk_bin_collection/councils/HertsmereBoroughCouncil.py,sha256=-ThSG6NIJP_wf2GmGL7SAvxbOujdhanZ8ECP4VSQCBs,5415
113
120
  uk_bin_collection/uk_bin_collection/councils/HighPeakCouncil.py,sha256=oqF8M0lcT3KsrG6W6I6JJX07E6Sc_-_sr7MybfIMab8,4626
114
121
  uk_bin_collection/uk_bin_collection/councils/HighlandCouncil.py,sha256=GNxDU65QuZHV5va2IrKtcJ6TQoDdwmV03JvkVqOauP4,3291
@@ -127,6 +134,7 @@ uk_bin_collection/uk_bin_collection/councils/LisburnCastlereaghCityCouncil.py,sh
127
134
  uk_bin_collection/uk_bin_collection/councils/LiverpoolCityCouncil.py,sha256=n17OqZrCGrPrnxGUfHc-RGkb4oJ9Bx6uUWiLdzxfQlY,2587
128
135
  uk_bin_collection/uk_bin_collection/councils/LondonBoroughEaling.py,sha256=QDx2Izr-6hUSFMi4UWqsgo3p6U8aRZ9d_Cu9cBSp2rY,1653
129
136
  uk_bin_collection/uk_bin_collection/councils/LondonBoroughHarrow.py,sha256=kzKwbjDBCbJXF6JdvTZWdXK0fpZxuH1CQF9GGyVMAus,1408
137
+ uk_bin_collection/uk_bin_collection/councils/LondonBoroughHavering.py,sha256=6DNX6IShdoEG4FjoyrzrY_HLEu9R2Bgl6PN0bSCbhow,2436
130
138
  uk_bin_collection/uk_bin_collection/councils/LondonBoroughHounslow.py,sha256=UOeiOxGMvVMm2UFaqjmQpm7vxzqJNSSN8LM9lAUjs2c,3021
131
139
  uk_bin_collection/uk_bin_collection/councils/LondonBoroughLambeth.py,sha256=r9D5lHe5kIRStCd5lRIax16yhb4KTFzzfYEFv1bacWw,2009
132
140
  uk_bin_collection/uk_bin_collection/councils/LondonBoroughRedbridge.py,sha256=A_6Sis5hsF53Th04KeadHRasGbpAm6aoaWJ6X8eC4Y8,6604
@@ -139,7 +147,7 @@ uk_bin_collection/uk_bin_collection/councils/MansfieldDistrictCouncil.py,sha256=
139
147
  uk_bin_collection/uk_bin_collection/councils/MertonCouncil.py,sha256=3Y2Un4xXo1sCcMsudynODSzocV_mMofWkX2JqONDb5o,1997
140
148
  uk_bin_collection/uk_bin_collection/councils/MidAndEastAntrimBoroughCouncil.py,sha256=oOWwU5FSgGej2Mv7FQ66N-EzS5nZgmGsd0WnfLWUc1I,5238
141
149
  uk_bin_collection/uk_bin_collection/councils/MidDevonCouncil.py,sha256=RjBZ7R3_Pax9p1d2DCygqryjV1RP4BYvqb-rT_KyOEg,3322
142
- uk_bin_collection/uk_bin_collection/councils/MidSuffolkDistrictCouncil.py,sha256=b1E47oGUmBfLHzoIT74t556kO-SuopelRprf_ywUsvM,6200
150
+ uk_bin_collection/uk_bin_collection/councils/MidSuffolkDistrictCouncil.py,sha256=h6M-v5jVYe7OlQ47Vf-0pEgECZLOOacK3_XE6zbpsM4,6329
143
151
  uk_bin_collection/uk_bin_collection/councils/MidSussexDistrictCouncil.py,sha256=AZgC9wmDLEjUOtIFvf0ehF5LHturXTH4DkE3ioPSVBA,6254
144
152
  uk_bin_collection/uk_bin_collection/councils/MidlothianCouncil.py,sha256=mM5-itJDNhjsT5UEjSFfWppmfmPFSns4u_1QblewuFU,5605
145
153
  uk_bin_collection/uk_bin_collection/councils/MiltonKeynesCityCouncil.py,sha256=7e2pGBLCw24pNItHeI9jkxQ3rEOZ4WC4zVlbvKYGdXE,2600
@@ -148,11 +156,13 @@ uk_bin_collection/uk_bin_collection/councils/NeathPortTalbotCouncil.py,sha256=yc
148
156
  uk_bin_collection/uk_bin_collection/councils/NewForestCouncil.py,sha256=ylTn9KmWITtaO9_Z8kJCN2w2ALfhrfGt3SeJ78lgw7M,5391
149
157
  uk_bin_collection/uk_bin_collection/councils/NewarkAndSherwoodDC.py,sha256=lAleYfCGUWCKOi7Ye_cjgfpI3pWwTcFctlYmh0hjebM,2140
150
158
  uk_bin_collection/uk_bin_collection/councils/NewcastleCityCouncil.py,sha256=eJMX10CG9QO7FRhHSmUDL-jO_44qoK3_1ztNTAXhkbw,2085
159
+ uk_bin_collection/uk_bin_collection/councils/NewcastleUnderLymeCouncil.py,sha256=t150wPtF2eEFYNTVj2EnhXt6JJJRiEX8976bh9JBXlg,2328
151
160
  uk_bin_collection/uk_bin_collection/councils/NewhamCouncil.py,sha256=klFqr2Km1BuQ4wERmjbXZzOvN1N0jnVmhTGRSVzVOaU,2086
152
161
  uk_bin_collection/uk_bin_collection/councils/NewportCityCouncil.py,sha256=dAcl5P97bttl2xCPvxof1a18kmqOrMmiElgtn3Ej7zs,8480
153
162
  uk_bin_collection/uk_bin_collection/councils/NorthAyrshireCouncil.py,sha256=o8zv40Wt19d51mrN5lsgLMCKMokMPmI1cMHBNT5yAho,1976
154
163
  uk_bin_collection/uk_bin_collection/councils/NorthEastDerbyshireDistrictCouncil.py,sha256=bps_uCFAeUHQla7AE5lZOv6sUkUz1fb6zduZdAcYMuw,4651
155
164
  uk_bin_collection/uk_bin_collection/councils/NorthEastLincs.py,sha256=fYf438VZIaOaqPSwdTTWVjFTdrI0jGfFsxVzOc-QdkA,1817
165
+ uk_bin_collection/uk_bin_collection/councils/NorthHertfordshireDistrictCouncil.py,sha256=dFgvZqVKEVEP0zSPeh2s9xIWSCGbhYHpXn2U6Nk0HXM,2847
156
166
  uk_bin_collection/uk_bin_collection/councils/NorthKestevenDistrictCouncil.py,sha256=vYOCerJXr9LTP6F2wm4vpYNYbQaWNZ6yfHEQ33N_hTw,1681
157
167
  uk_bin_collection/uk_bin_collection/councils/NorthLanarkshireCouncil.py,sha256=npK1V8D3SLNTSSKkfEpEPvVgXDFyhH_tAsuGogsVKQY,1763
158
168
  uk_bin_collection/uk_bin_collection/councils/NorthLincolnshireCouncil.py,sha256=kCT82SU-etoGBehoIbbzGw_XS2YHllTw7lCCbdQj7UE,2467
@@ -265,8 +275,8 @@ uk_bin_collection/uk_bin_collection/councils/YorkCouncil.py,sha256=I2kBYMlsD4bId
265
275
  uk_bin_collection/uk_bin_collection/councils/council_class_template/councilclasstemplate.py,sha256=EQWRhZ2pEejlvm0fPyOTsOHKvUZmPnxEYO_OWRGKTjs,1158
266
276
  uk_bin_collection/uk_bin_collection/create_new_council.py,sha256=m-IhmWmeWQlFsTZC4OxuFvtw5ZtB8EAJHxJTH4O59lQ,1536
267
277
  uk_bin_collection/uk_bin_collection/get_bin_data.py,sha256=YvmHfZqanwrJ8ToGch34x-L-7yPe31nB_x77_Mgl_vo,4545
268
- uk_bin_collection-0.118.0.dist-info/LICENSE,sha256=vABBUOzcrgfaTKpzeo-si9YVEun6juDkndqA8RKdKGs,1071
269
- uk_bin_collection-0.118.0.dist-info/METADATA,sha256=qfGDaPkQ6hAS_gVAFVs10LUi2EzwBJtfvFK5c7m-rh0,17574
270
- uk_bin_collection-0.118.0.dist-info/WHEEL,sha256=Nq82e9rUAnEjt98J6MlVmMCZb-t9cYE2Ir1kpBmnWfs,88
271
- uk_bin_collection-0.118.0.dist-info/entry_points.txt,sha256=36WCSGMWSc916S3Hi1ZkazzDKHaJ6CD-4fCEFm5MIao,90
272
- uk_bin_collection-0.118.0.dist-info/RECORD,,
278
+ uk_bin_collection-0.120.0.dist-info/LICENSE,sha256=vABBUOzcrgfaTKpzeo-si9YVEun6juDkndqA8RKdKGs,1071
279
+ uk_bin_collection-0.120.0.dist-info/METADATA,sha256=wvXpQZc70f0TI2_UTRc4RDel0O-sFh7-yO_vEVdVe_o,17574
280
+ uk_bin_collection-0.120.0.dist-info/WHEEL,sha256=Nq82e9rUAnEjt98J6MlVmMCZb-t9cYE2Ir1kpBmnWfs,88
281
+ uk_bin_collection-0.120.0.dist-info/entry_points.txt,sha256=36WCSGMWSc916S3Hi1ZkazzDKHaJ6CD-4fCEFm5MIao,90
282
+ uk_bin_collection-0.120.0.dist-info/RECORD,,