@mariozechner/pi-ai 0.29.0 → 0.30.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
|
@@ -6489,7 +6489,7 @@ export declare const MODELS: {
|
|
|
6489
6489
|
contextWindow: number;
|
|
6490
6490
|
maxTokens: number;
|
|
6491
6491
|
};
|
|
6492
|
-
readonly "meta-llama/llama-3.1-
|
|
6492
|
+
readonly "meta-llama/llama-3.1-8b-instruct": {
|
|
6493
6493
|
id: string;
|
|
6494
6494
|
name: string;
|
|
6495
6495
|
api: "openai-completions";
|
|
@@ -6506,7 +6506,7 @@ export declare const MODELS: {
|
|
|
6506
6506
|
contextWindow: number;
|
|
6507
6507
|
maxTokens: number;
|
|
6508
6508
|
};
|
|
6509
|
-
readonly "meta-llama/llama-3.1-
|
|
6509
|
+
readonly "meta-llama/llama-3.1-405b-instruct": {
|
|
6510
6510
|
id: string;
|
|
6511
6511
|
name: string;
|
|
6512
6512
|
api: "openai-completions";
|
|
@@ -6523,7 +6523,7 @@ export declare const MODELS: {
|
|
|
6523
6523
|
contextWindow: number;
|
|
6524
6524
|
maxTokens: number;
|
|
6525
6525
|
};
|
|
6526
|
-
readonly "meta-llama/llama-3.1-
|
|
6526
|
+
readonly "meta-llama/llama-3.1-70b-instruct": {
|
|
6527
6527
|
id: string;
|
|
6528
6528
|
name: string;
|
|
6529
6529
|
api: "openai-completions";
|
|
@@ -6676,7 +6676,7 @@ export declare const MODELS: {
|
|
|
6676
6676
|
contextWindow: number;
|
|
6677
6677
|
maxTokens: number;
|
|
6678
6678
|
};
|
|
6679
|
-
readonly "openai/gpt-4o": {
|
|
6679
|
+
readonly "openai/gpt-4o-2024-05-13": {
|
|
6680
6680
|
id: string;
|
|
6681
6681
|
name: string;
|
|
6682
6682
|
api: "openai-completions";
|
|
@@ -6693,7 +6693,7 @@ export declare const MODELS: {
|
|
|
6693
6693
|
contextWindow: number;
|
|
6694
6694
|
maxTokens: number;
|
|
6695
6695
|
};
|
|
6696
|
-
readonly "openai/gpt-4o
|
|
6696
|
+
readonly "openai/gpt-4o": {
|
|
6697
6697
|
id: string;
|
|
6698
6698
|
name: string;
|
|
6699
6699
|
api: "openai-completions";
|
|
@@ -6710,7 +6710,7 @@ export declare const MODELS: {
|
|
|
6710
6710
|
contextWindow: number;
|
|
6711
6711
|
maxTokens: number;
|
|
6712
6712
|
};
|
|
6713
|
-
readonly "openai/gpt-4o
|
|
6713
|
+
readonly "openai/gpt-4o:extended": {
|
|
6714
6714
|
id: string;
|
|
6715
6715
|
name: string;
|
|
6716
6716
|
api: "openai-completions";
|
|
@@ -6846,7 +6846,7 @@ export declare const MODELS: {
|
|
|
6846
6846
|
contextWindow: number;
|
|
6847
6847
|
maxTokens: number;
|
|
6848
6848
|
};
|
|
6849
|
-
readonly "openai/gpt-
|
|
6849
|
+
readonly "openai/gpt-3.5-turbo-0613": {
|
|
6850
6850
|
id: string;
|
|
6851
6851
|
name: string;
|
|
6852
6852
|
api: "openai-completions";
|
|
@@ -6863,7 +6863,7 @@ export declare const MODELS: {
|
|
|
6863
6863
|
contextWindow: number;
|
|
6864
6864
|
maxTokens: number;
|
|
6865
6865
|
};
|
|
6866
|
-
readonly "openai/gpt-
|
|
6866
|
+
readonly "openai/gpt-4-turbo-preview": {
|
|
6867
6867
|
id: string;
|
|
6868
6868
|
name: string;
|
|
6869
6869
|
api: "openai-completions";
|