divergent-beamsearch 0.2.0__tar.gz → 0.2.1__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,12 +1,11 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: divergent-beamsearch
3
- Version: 0.2.0
3
+ Version: 0.2.1
4
4
  Summary: A variant of the beam search algorithm that focuses on finding answers that maximize the probability of generating an answer before diverging into another subject.
5
5
  License-File: LICENCE
6
6
  Requires-Python: >=3.11
7
- Requires-Dist: multi-choices-parser>=0.9.61
7
+ Requires-Dist: multi-choices-parser>=0.9.72
8
8
  Requires-Dist: torch>=2.0.0
9
- Requires-Dist: transformers>=4.47.1
10
9
  Description-Content-Type: text/markdown
11
10
 
12
11
  # Divergent Beam Search
@@ -1,18 +1,18 @@
1
1
  [project]
2
2
  name = "divergent-beamsearch"
3
- version = "0.2.0"
3
+ version = "0.2.1"
4
4
  description = "A variant of the beam search algorithm that focuses on finding answers that maximize the probability of generating an answer before diverging into another subject."
5
5
  readme = "README.md"
6
6
  requires-python = ">=3.11"
7
7
  dependencies = [
8
- "multi-choices-parser>=0.9.61",
8
+ "multi-choices-parser>=0.9.72",
9
9
  "torch>=2.0.0",
10
- "transformers>=4.47.1",
11
10
  ]
12
11
 
13
12
  [dependency-groups]
14
13
  dev = [
15
14
  "pytest>=8.3.4",
15
+ "transformers>=4.47.1"
16
16
  ]
17
17
 
18
18
  [build-system]
@@ -1,6 +1,9 @@
1
1
  import math
2
2
  import torch
3
- from transformers import GPT2LMHeadModel
3
+ try:
4
+ from transformers import GPT2LMHeadModel
5
+ except ImportError:
6
+ pass
4
7
  from multi_choices_parser import DEFAULT_END_SYMB
5
8
 
6
9
 
@@ -35,7 +38,7 @@ def apply_mask_tokens(pred : torch.Tensor, parsers_tokens):
35
38
  return pred[~pred.isinf().all(dim=-1)]
36
39
 
37
40
 
38
- def batched_inference_logits(model : GPT2LMHeadModel, input_ids : torch.Tensor,
41
+ def batched_inference_logits(model : "GPT2LMHeadModel", input_ids : torch.Tensor,
39
42
  attention_mask : torch.Tensor | None = None, batch_size : int = 32,
40
43
  to_cpu=False) -> torch.Tensor:
41
44
  logits = []
@@ -96,7 +99,7 @@ def pad_to_same_size(tensors : list[torch.Tensor], padding_value : int) -> torch
96
99
  return torch.cat(padded_tensors, dim=0)
97
100
 
98
101
  @torch.no_grad()
99
- def divergent_beamsearch(input_ids : torch.Tensor, model : GPT2LMHeadModel, beam_size : int,
102
+ def divergent_beamsearch(input_ids : torch.Tensor, model : "GPT2LMHeadModel", beam_size : int,
100
103
  max_length : int, parser : Parser, pad_token_id : int, batch_size=32,
101
104
  num_solutions = None, end_symb=DEFAULT_END_SYMB, optimize_gpu_mem=True) -> tuple[torch.Tensor, torch.Tensor]:
102
105
  assert input_ids.shape[0] == 1, "Batch size must be 1"
@@ -180,7 +183,7 @@ def set_slice_row(x : torch.Tensor, slices : torch.IntTensor, value) -> torch.Te
180
183
  x[i].index_fill_(0, indices[i], 0)
181
184
 
182
185
  @torch.no_grad()
183
- def divergent_logprob(input_ids : torch.Tensor, attention_mask : torch.Tensor | None, model : GPT2LMHeadModel,
186
+ def divergent_logprob(input_ids : torch.Tensor, attention_mask : torch.Tensor | None, model : "GPT2LMHeadModel",
184
187
  parsers : Parser | list[Parser] | None, batch_size=32,
185
188
  start : int | torch.IntTensor = None, end_symb=DEFAULT_END_SYMB, optimize_gpu_mem=True) -> torch.FloatTensor:
186
189
  if start is None:
@@ -79,11 +79,13 @@ def test_divergent_beamsearch(model_and_tokenizer, device, end_symb):
79
79
  end_symb=end_symb
80
80
  )
81
81
  true_solutions = torch.nn.utils.rnn.pad_sequence([torch.tensor(ans) for ans in tokenized_answers], batch_first=True, padding_value=pad_token_id)
82
- assert (solutions == true_solutions).all(), "Beam search did not return the expected solutions"
82
+
83
83
  assert torch.isclose(scores[0], logprob_paris_diverge), "Beam search did not return the expected score"
84
84
  assert torch.isclose(scores[1], logprob_madrid), "Beam search did not return the expected score"
85
85
  assert torch.isclose(scores[2], logprob_paris_hilton), "Beam search did not return the expected score"
86
86
  assert torch.isclose(scores[3], logprob_garbage), "Beam search did not return the expected score"
87
+ assert (solutions == true_solutions).all(), "Beam search did not return the expected solutions"
88
+
87
89
 
88
90
 
89
91
  @pytest.mark.parametrize("device", ['cpu', 'cuda'])
@@ -73,7 +73,7 @@ wheels = [
73
73
 
74
74
  [[package]]
75
75
  name = "divergent-beamsearch"
76
- version = "0.1.5"
76
+ version = "0.2.0"
77
77
  source = { editable = "." }
78
78
  dependencies = [
79
79
  { name = "multi-choices-parser" },
@@ -88,7 +88,7 @@ dev = [
88
88
 
89
89
  [package.metadata]
90
90
  requires-dist = [
91
- { name = "multi-choices-parser", specifier = ">=0.9.61" },
91
+ { name = "multi-choices-parser", specifier = ">=0.9.72" },
92
92
  { name = "torch", specifier = ">=2.0.0" },
93
93
  { name = "transformers", specifier = ">=4.47.1" },
94
94
  ]
@@ -221,11 +221,18 @@ wheels = [
221
221
 
222
222
  [[package]]
223
223
  name = "multi-choices-parser"
224
- version = "0.9.61"
225
- source = { registry = "https://pypi.org/simple" }
226
- sdist = { url = "https://files.pythonhosted.org/packages/e2/17/90a6125a2145c03e39c3e7f78f65121eb14dedb9de8b40aee3c8a24a709b/multi_choices_parser-0.9.61.tar.gz", hash = "sha256:be932cac4aeabe9ee057c6d4592ea4325a0a92e52758d77a9e08bafa2cd23294", size = 7889 }
227
- wheels = [
228
- { url = "https://files.pythonhosted.org/packages/3c/4f/c5a514a510779202ff37505220edfba9154ceff31958ed71fa1878781af9/multi_choices_parser-0.9.61-py3-none-any.whl", hash = "sha256:36bc367bceb66bbfb1bea26d9a38aa9cd10273b54cef331dd7c69da582fb9c2a", size = 6870 },
224
+ version = "0.9.72"
225
+ source = { registry = "https://pypi.org/simple" }
226
+ wheels = [
227
+ { url = "https://files.pythonhosted.org/packages/ab/14/7a99908c455ed355563c1a59c3953fd2e1e0b8bd3699f616adf44f31c019/multi_choices_parser-0.9.72-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:7554b9928b663952d50dad2be070b33eac12a7cf0a5d0237ca273f075e598d09", size = 99165 },
228
+ { url = "https://files.pythonhosted.org/packages/aa/3c/fb13affb1061050fb0f2988d1fdd0f37943e17abf1644ac681d6cda45615/multi_choices_parser-0.9.72-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:4700eab86215bd4f0da9fcea0650e5336bc22d62a77625b2d3d1b1a83081b0d4", size = 139545 },
229
+ { url = "https://files.pythonhosted.org/packages/47/c0/5b47daed1dd6cff64c602cdcefda285eacfdf71f43d0452ed2f68e17ae9e/multi_choices_parser-0.9.72-cp311-cp311-win_amd64.whl", hash = "sha256:65725c593363b8c207748478ca966e5fc0288118b95c2e6b7cc338003417a185", size = 105544 },
230
+ { url = "https://files.pythonhosted.org/packages/51/98/10331d2da4c0c036720f1cd41a60f33cf35a4ac2aad963dd58e486d97ccb/multi_choices_parser-0.9.72-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:277ed1e6e4c7044313281caa5c20cb09eab518f27271001afea748793acc26a9", size = 99360 },
231
+ { url = "https://files.pythonhosted.org/packages/98/6a/5c90c3b19013aa02b40ed6ef193213bfb4ad92e4500c8e1009e712c0d6db/multi_choices_parser-0.9.72-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:441ef6dd784c9d7fbf0effe66f2e910ad308604749924e0dccec79fd24cfdf2e", size = 138117 },
232
+ { url = "https://files.pythonhosted.org/packages/07/13/4c601c9336b7a83e762937c2d75823964a9a9773903cba2696ec59107dbf/multi_choices_parser-0.9.72-cp312-cp312-win_amd64.whl", hash = "sha256:28ac8cea47639b434fc88e143f5d38a0bd5ab4ce9a040a036e532896185d672b", size = 105658 },
233
+ { url = "https://files.pythonhosted.org/packages/db/9d/8ea1f8a87282da07b0d5044c682566a68eced933bb675c8936400bb72a54/multi_choices_parser-0.9.72-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:3dfd87c968a6e5618a0cacc3d29fa244cf427208829eeda82802fd60250ea1b3", size = 99398 },
234
+ { url = "https://files.pythonhosted.org/packages/8b/8d/f0a244c59e13e4591e5be9f0793a22e3cde6b631801f9473e96fe44c76bb/multi_choices_parser-0.9.72-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:c99dae2ba228c9362648ed66bd9790df07a71176f22938595487605b489a8dc8", size = 138403 },
235
+ { url = "https://files.pythonhosted.org/packages/f3/f2/73e929b894fe379be5e1b8d373c9acf66c5b34da3f886edbe22ff8725593/multi_choices_parser-0.9.72-cp313-cp313-win_amd64.whl", hash = "sha256:9576300f71ba688f799832e8b86b3cb24ea74cde29aa4e70ac63ec7545e32790", size = 105658 },
229
236
  ]
230
237
 
231
238
  [[package]]