code-graph-context 2.12.0 → 2.12.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/cli/cli.js CHANGED
@@ -264,7 +264,7 @@ const preDownloadModel = async (sidecarDir, python, modelName) => {
264
264
  return new Promise((resolve) => {
265
265
  const proc = spawnProcess(python, [
266
266
  '-c',
267
- `from sentence_transformers import SentenceTransformer; m = SentenceTransformer("${modelName}"); print(f"dims:{len(m.encode(['test'])[0])}")`,
267
+ `from sentence_transformers import SentenceTransformer; m = SentenceTransformer("${modelName}", trust_remote_code=True); print(f"dims:{len(m.encode(['test'])[0])}")`,
268
268
  ], { cwd: sidecarDir, stdio: ['pipe', 'pipe', 'pipe'], timeout: 30_000 });
269
269
  let stdout = '';
270
270
  proc.stdout?.on('data', (d) => (stdout += d.toString()));
@@ -286,7 +286,7 @@ const preDownloadModel = async (sidecarDir, python, modelName) => {
286
286
  const downloaded = await new Promise((resolve) => {
287
287
  const proc = spawnProcess(python, [
288
288
  '-c',
289
- `from sentence_transformers import SentenceTransformer; print("downloading..."); m = SentenceTransformer("${modelName}"); print(f"done dims:{len(m.encode(['test'])[0])}")`,
289
+ `from sentence_transformers import SentenceTransformer; print("downloading..."); m = SentenceTransformer("${modelName}", trust_remote_code=True); print(f"done dims:{len(m.encode(['test'])[0])}")`,
290
290
  ], { cwd: sidecarDir, stdio: ['pipe', 'pipe', 'inherit'] });
291
291
  let stdout = '';
292
292
  proc.stdout?.on('data', (d) => {
@@ -302,7 +302,7 @@ const preDownloadModel = async (sidecarDir, python, modelName) => {
302
302
  }
303
303
  else {
304
304
  log(sym.warn, 'Model download failed — it will retry on first use');
305
- console.log(` ${c.dim}You can download manually: ${python} -c "from sentence_transformers import SentenceTransformer; SentenceTransformer('${modelName}')"${c.reset}`);
305
+ console.log(` ${c.dim}You can download manually: ${python} -c "from sentence_transformers import SentenceTransformer; SentenceTransformer('${modelName}', trust_remote_code=True)"${c.reset}`);
306
306
  }
307
307
  };
308
308
  /**
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "code-graph-context",
3
- "version": "2.12.0",
3
+ "version": "2.12.2",
4
4
  "description": "MCP server that builds code graphs to provide rich context to LLMs",
5
5
  "type": "module",
6
6
  "homepage": "https://github.com/drewdrewH/code-graph-context#readme",
@@ -58,10 +58,10 @@ def load_model():
58
58
 
59
59
  use_half = os.environ.get("EMBEDDING_HALF_PRECISION", "").lower() == "true"
60
60
  if use_half:
61
- model = SentenceTransformer(model_name, device=device, model_kwargs={"torch_dtype": "float16"})
61
+ model = SentenceTransformer(model_name, device=device, trust_remote_code=True, model_kwargs={"torch_dtype": "float16"})
62
62
  logger.info(f"Model loaded in float16 (half precision)")
63
63
  else:
64
- model = SentenceTransformer(model_name, device=device)
64
+ model = SentenceTransformer(model_name, device=device, trust_remote_code=True)
65
65
  logger.info(f"Model loaded in float32 (full precision)")
66
66
  logger.info(f"Running warmup...")
67
67