@adaline/google 0.4.0 → 0.6.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.d.mts +15 -4
- package/dist/index.d.ts +15 -4
- package/dist/index.js +116 -115
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +10 -10
- package/dist/index.mjs.map +1 -1
- package/package.json +5 -5
package/dist/index.mjs
CHANGED
|
@@ -2,16 +2,16 @@ import { z as z$1 } from 'zod';
|
|
|
2
2
|
import { RangeConfigItem, CHAT_CONFIG, SelectStringConfigItem, ObjectSchemaConfigItem, ChatModelSchema, EmbeddingModelSchema, MultiStringConfigItem, EMBEDDING_CONFIG, ProviderError, urlWithoutTrailingSlash, InvalidModelRequestError, removeUndefinedEntries, InvalidMessagesError, InvalidConfigError, InvalidToolsError, ModelResponseError, InvalidEmbeddingRequestsError } from '@adaline/provider';
|
|
3
3
|
import { SystemRoleLiteral, UserRoleLiteral, AssistantRoleLiteral, ToolRoleLiteral, TextModalityLiteral, ImageModalityLiteral, ToolCallModalityLiteral, ToolResponseModalityLiteral, EmbeddingTextModalityLiteral, Config, Base64ImageContentTypeLiteral, Message, Tool, createTextContent, createToolCallContent, EmbeddingRequests, FloatEmbeddingLiteral, createPartialTextMessage, createPartialToolCallMessage } from '@adaline/types';
|
|
4
4
|
|
|
5
|
-
var
|
|
6
|
-
[${this.chatModelLiterals().join(", ")}]`)});let t=this.chatModelFactories[o].model,
|
|
7
|
-
[${this.embeddingModelLiterals().join(", ")}]`)});let t=this.embeddingModelFactories[o].model,
|
|
8
|
-
available keys : [${Object.keys(this.modelSchema.config.def).join(", ")}]`)})});let m=Object.keys(
|
|
9
|
-
one of [${C.choices.join(", ")}]`)})}}return G(G({generation_config:m},p?{tool_config:p}:{}),
|
|
10
|
-
available modalities : [${this.modelSchema.modalities.join(", ")}]`)})});}),o.forEach(
|
|
11
|
-
available roles : [${Object.keys(this.modelSchema.roles).join(", ")}]`)})});let t={parts:[]},
|
|
12
|
-
available roles : [${Object.keys(this.modelSchema.roles).join(", ")}]`)})}}),s[0].role!==this.modelSchema.roles[UserRoleLiteral])throw new InvalidMessagesError({info:`Invalid message 'role' for model : ${this.modelName}`,cause:new Error(`model : '${this.modelName}' requires first message to be from user`)});let r=n=>n===this.modelSchema.roles[UserRoleLiteral]||n===this.modelSchema.roles[ToolRoleLiteral]?[this.modelSchema.roles[AssistantRoleLiteral]]:[this.modelSchema.roles[UserRoleLiteral],this.modelSchema.roles[ToolRoleLiteral]];for(let n=1;n<s.length;n++)if(!r(s[n-1].role).includes(s[n].role))throw new InvalidMessagesError({info:`Invalid message format for model : ${this.modelName}`,cause:new Error(`model : '${this.modelName}' cannot have message with role : '${s[n].role}' after message with role : '${s[n-1].role}'`)});if(s[s.length-1].role!==this.modelSchema.roles[UserRoleLiteral])throw new InvalidMessagesError({info:`Invalid message format for model : ${this.modelName}`,cause:new Error(`model : '${this.modelName}' requires last message to be from user`)});return G({contents:s},t.parts.length>0?{system_instruction:t}:{})}transformTools(e){if(!this.modelSchema.modalities.includes(ToolCallModalityLiteral))throw new InvalidToolsError({info:`Invalid tool 'modality' for model : ${this.modelName}`,cause:new Error(`model : '${this.modelName}' does not support tool modality : '${ToolCallModalityLiteral}'`)});return !e||e&&e.length===0?{tools:[]}:{tools:e.map(s=>{let r=Tool().safeParse(s);if(!r.success)throw new InvalidToolsError({info:"Invalid tools",cause:r.error});return r.data}).map(s=>({function_declarations:[{name:s.definition.schema.name,description:s.definition.schema.description,parameters:s.definition.schema.parameters}]}))}}getCompleteChatUrl(e,o,t){return R(this,null,function*(){return new Promise(s=>{s(this.completeChatUrl);})})}getCompleteChatHeaders(e,o,t){return R(this,null,function*(){return new Promise(s=>{s(this.getDefaultHeaders());})})}getCompleteChatData(e,o,t){return R(this,null,function*(){let s=this.transformConfig(e,o,t),r=this.transformMessages(o);if(r.messages&&r.messages.length===0)throw new InvalidMessagesError({info:"Messages are required",cause:new Error("Messages are required")});let n=t?this.transformTools(t):{};return new Promise(m=>{m(G(G(G(G({},this.getDefaultParams()),s),r),n));})})}transformCompleteChatResponse(e){let o=wo.safeParse(e);if(o.success){if(o.data.candidates.length===0)throw new ModelResponseError({info:"Invalid response from model",cause:new Error(`No choices in response : ${JSON.stringify(o.data)}`)});let t=o.data,s=[],r,n=t.candidates[0].content;if(n){let p=n.parts.map((f,C)=>{if("text"in f&&f.text!==void 0)return createTextContent(f.text);if("functionCall"in f&&f.functionCall!==void 0)return createToolCallContent(C,`${f.functionCall.name}_${C}`,f.functionCall.name,JSON.stringify(f.functionCall.args))});return s.push({role:AssistantRoleLiteral,content:p}),t.usageMetadata&&(r={promptTokens:t.usageMetadata.promptTokenCount,totalTokens:t.usageMetadata.totalTokenCount,completionTokens:t.usageMetadata.candidatesTokenCount||0}),{messages:s,usage:r,logProbs:void 0}}let m=t.candidates[0].safetyRatings;if(m&&m.length>0&&m.forEach(p=>{if(p.blocked)throw new ModelResponseError({info:`Blocked content for category: ${p.category} with probability: ${p.probability}`,cause:new Error(`Blocked content for category: ${p.category} with probability: ${p.probability}`)})}),t.candidates[0].finishReason==="SAFETY")throw new ModelResponseError({info:"Blocked content, model response finished with safety reason",cause:new Error("Blocked content, model response finished with safety reason")})}throw new ModelResponseError({info:"Invalid response from model",cause:o.error})}getStreamChatUrl(e,o,t){return R(this,null,function*(){return new Promise(s=>{s(this.streamChatUrl);})})}getStreamChatHeaders(e,o,t){return R(this,null,function*(){return new Promise(s=>{s(this.getDefaultHeaders());})})}getStreamChatData(e,o,t){return R(this,null,function*(){let s=this.transformConfig(e,o,t),r=this.transformMessages(o);if(r.messages&&r.messages.length===0)throw new InvalidMessagesError({info:"Messages are required",cause:new Error("Messages are required")});let n=t?this.transformTools(t):{};return new Promise(m=>{m(G(G(G(G({},this.getDefaultParams()),s),r),n));})})}transformStreamChatResponseChunk(e,o){return bo(this,null,function*(){let t=(o+e).split(",\r").filter(s=>s.trim()!=="");for(let s of t){let r=s;if(r=r.replace(/\n/g,""),r.startsWith("["))r=r.slice(1);else if(r.endsWith("]")){if(r==="]")return;r=r.slice(0,-1);}let n;try{n=JSON.parse(r);}catch(l){if(l instanceof SyntaxError){o=r;continue}else throw l}o="";let m=Lo.safeParse(n);if(m.success){let l={partialMessages:[]},p=m.data;if(p.candidates.length>0){let f=p.candidates[0].content;f&&"parts"in f&&f.parts.length>0&&f.parts.forEach((C,M)=>{if("text"in C&&C.text!==void 0&&l.partialMessages.push(createPartialTextMessage(AssistantRoleLiteral,C.text)),"functionCall"in C&&C.functionCall!==void 0){let g=C.functionCall;l.partialMessages.push(createPartialToolCallMessage(AssistantRoleLiteral,M,`${g.name}_${M}`,g.name,JSON.stringify(g.args)));}});}p.usageMetadata&&(l.usage={promptTokens:p.usageMetadata.promptTokenCount,completionTokens:p.usageMetadata.candidatesTokenCount,totalTokens:p.usageMetadata.totalTokenCount}),yield {partialResponse:l,buffer:o};}else throw new ModelResponseError({info:"Invalid response from model",cause:m.error})}yield {partialResponse:{partialMessages:[]},buffer:o};})}};var P=class extends T{transformMessages(e){let o=super.transformMessages(e);if(o.systemInstruction){let t={role:this.modelSchema.roles[UserRoleLiteral],parts:o.systemInstruction.parts};o.contents.unshift(t),delete o.systemInstruction;}return o}};var Fe="gemini-pro",Wt="A model for scaling across a wide range of tasks Optimized for natural language tasks, multi-turn text and code chat, and code generation",$e=ChatModelSchema(u,I).parse({name:Fe,description:Wt,maxInputTokens:30720,maxOutputTokens:2048,roles:y,modalities:L,config:{def:c.base(1,.9,2048,4,1).def,schema:c.base(1,.9,2048,4,1).schema}}),$o=h,me=class extends P{constructor(e){super($e,e);}};var ze="gemini-pro-vision",Qt="An image understanding model to handle a broad range of applications",je=ChatModelSchema(u,re).parse({name:ze,description:Qt,maxInputTokens:12288,maxOutputTokens:4096,roles:y,modalities:ae,config:{def:c.c1(1,.4,4096,4,1,32).def,schema:c.c1(1,.4,4096,4,1,32).schema}}),zo=h,de=class extends P{constructor(e){super(je,e);}};var Be="gemini-1.0-pro",en="Google's predecessor to Gemini 1.5 Pro, a model for scaling across a wide range of tasks Optimized for natural language tasks, multi-turn text and code chat, and code generation",De=ChatModelSchema(u,I).parse({name:Be,description:en,maxInputTokens:30720,maxOutputTokens:2048,roles:y,modalities:L,config:{def:c.base(1,.9,2048,4,1).def,schema:c.base(1,.9,2048,4,1).schema}}),jo=h,pe=class extends P{constructor(e){super(De,e);}};var Ue="gemini-1.0-pro-001",tn="Google's predecessor to Gemini 1.5 Pro, a model for scaling across a wide range of tasks Optimized for natural language tasks, multi-turn text and code chat, and code generation",Ae=ChatModelSchema(u,I).parse({name:Ue,description:tn,maxInputTokens:30720,maxOutputTokens:2048,roles:y,modalities:L,config:{def:c.base(1,.9,2048,4,1).def,schema:c.base(1,.9,2048,4,1).schema}}),Bo=h,ce=class extends P{constructor(e){super(Ae,e);}};var Ve="gemini-1.0-pro-latest",sn="Google's latest multimodal model with great performance for high-frequency tasks. Optimized for natural language tasks, multi-turn text and code chat, and code generation",He=ChatModelSchema(u,I).parse({name:Ve,description:sn,maxInputTokens:30720,maxOutputTokens:2048,roles:y,modalities:L,config:{def:c.base(1,.9,2048,4,1).def,schema:c.base(1,.9,2048,4,1).schema}}),Do=h,fe=class extends P{constructor(e){super(He,e);}};var Ke="gemini-1.0-pro-vision",rn="Google's predecessor to Gemini 1.5 Pro, an image understanding model to handle a broad range of applications",Ye=ChatModelSchema(u,re).parse({name:Ke,description:rn,maxInputTokens:12288,maxOutputTokens:4096,roles:y,modalities:ae,config:{def:c.c1(1,.4,4096,4,1,32).def,schema:c.c1(1,.4,4096,4,1,32).schema}}),Uo=h,he=class extends P{constructor(e){super(Ye,e);}};var Je="gemini-1.5-flash",mn="Google's fastest, most cost-efficient multimodal model with great performance for high-frequency tasks. Optimized for fast and versatile performance across a diverse variety of tasks",We=ChatModelSchema(u,E).parse({name:Je,description:mn,maxInputTokens:1e6,maxOutputTokens:8192,roles:y,modalities:_,config:{def:c.c1(2,1,8192,4,.95,64).def,schema:c.c1(2,1,8192,4,.95,64).schema}}),Ao=h,ge=class extends T{constructor(e){super(We,e);}};var Xe="gemini-1.5-flash-001",pn="Google's fastest, most cost-efficient multimodal model with great performance for high-frequency tasks. Optimized for fast and versatile performance across a diverse variety of tasks",Qe=ChatModelSchema(u,E).parse({name:Xe,description:pn,maxInputTokens:1e6,maxOutputTokens:8192,roles:y,modalities:_,config:{def:c.c1(2,1,8192,4,.95,64).def,schema:c.c1(2,1,8192,4,.95,64).schema}}),Vo=h,ue=class extends T{constructor(e){super(Qe,e);}};var Ze="gemini-1.5-flash-002",fn="Google's fastest, most cost-efficient multimodal model with great performance for high-frequency tasks. Optimized for fast and versatile performance across a diverse variety of tasks",eo=ChatModelSchema(u,E).parse({name:Ze,description:fn,maxInputTokens:1e6,maxOutputTokens:8192,roles:y,modalities:_,config:{def:c.c1(2,1,8192,4,.95,40).def,schema:c.c1(2,1,8192,4,.95,40).schema}}),Ho=h,ye=class extends T{constructor(e){super(eo,e);}};var oo="gemini-1.5-flash-latest",gn="Google's latest multimodal model with great performance for high-frequency tasks. Optimized for fast and versatile performance across a diverse variety of tasks",to=ChatModelSchema(u,E).parse({name:oo,description:gn,maxInputTokens:1e6,maxOutputTokens:8192,roles:y,modalities:_,config:{def:c.c1(2,1,8192,4,.95,64).def,schema:c.c1(2,1,8192,4,.95,64).schema}}),Ko=h,Ce=class extends T{constructor(e){super(to,e);}};var no="gemini-1.5-pro",yn="Google's best performing multimodal model with features for a wide variety of reasoning tasks. Optimized for complex reasoning tasks requiring more intelligence",so=ChatModelSchema(u,E).parse({name:no,description:yn,maxInputTokens:2e6,maxOutputTokens:8192,roles:y,modalities:_,config:{def:c.c1(2,1,8192,4,.95,64).def,schema:c.c1(2,1,8192,4,.95,64).schema}}),Yo=h,Te=class extends T{constructor(e){super(so,e);}};var io="gemini-1.5-pro-001",Tn="Google's best performing multimodal model with features for a wide variety of reasoning tasks. Optimized for complex reasoning tasks requiring more intelligence",ao=ChatModelSchema(u,E).parse({name:io,description:Tn,maxInputTokens:2e6,maxOutputTokens:8192,roles:y,modalities:_,config:{def:c.c1(2,1,8192,4,.95,64).def,schema:c.c1(2,1,8192,4,.95,64).schema}}),Jo=h,Ge=class extends T{constructor(e){super(ao,e);}};var ro="gemini-1.5-pro-002",Mn="Google's best performing multimodal model with features for a wide variety of reasoning tasks. Optimized for complex reasoning tasks requiring more intelligence",lo=ChatModelSchema(u,E).parse({name:ro,description:Mn,maxInputTokens:2e6,maxOutputTokens:8192,roles:y,modalities:_,config:{def:c.c1(2,1,8192,4,.95,40).def,schema:c.c1(2,1,8192,4,.95,40).schema}}),Wo=h,Me=class extends T{constructor(e){super(lo,e);}};var mo="gemini-1.5-pro-latest",_n="Google's best performing multimodal model with features for a wide variety of reasoning tasks. Optimized for complex reasoning tasks requiring more intelligence",po=ChatModelSchema(u,E).parse({name:mo,description:_n,maxInputTokens:2e6,maxOutputTokens:8192,roles:y,modalities:_,config:{def:c.c1(2,1,8192,4,.95,64).def,schema:c.c1(2,1,8192,4,.95,64).schema}}),Xo=h,be=class extends T{constructor(e){super(po,e);}};var Re=[EmbeddingTextModalityLiteral],xe=z$1.enum([EmbeddingTextModalityLiteral]);var it=z$1.object({embeddings:z$1.array(z$1.object({values:z$1.array(z$1.number())}))});var On=z$1.object({model:z$1.string().min(1),content:z$1.object({parts:z$1.array(z$1.object({text:z$1.string().min(1)})).min(1)})}),at=z$1.object({model:z$1.string().min(1).optional(),requests:z$1.array(On).min(1),outputDimensionality:z$1.number().int().min(1).optional()});var ee=z$1.object({modelName:z$1.string(),apiKey:z$1.string(),baseUrl:z$1.string().url(),getEmbeddingsUrl:z$1.string().url().optional()}),A=class{constructor(e,o){this.version="v1";let t=ee.parse(o);this.modelSchema=e,this.modelName=t.modelName,this.apiKey=t.apiKey,this.baseUrl=urlWithoutTrailingSlash(t.baseUrl),this.getEmbeddingsUrl=urlWithoutTrailingSlash(t.getEmbeddingsUrl||`${this.baseUrl}/models/${this.modelName}:batchEmbedContents?key=${this.apiKey}`);}getDefaultBaseUrl(){return this.baseUrl}getDefaultHeaders(){return {"Content-Type":"application/json",source:"adaline.ai"}}getDefaultParams(){return {model:this.modelName}}getRetryDelay(e){return {shouldRetry:!1,delayMs:0}}getTokenCount(e){return e.requests.reduce((o,t)=>o+t.length,0)}transformModelRequest(e){let o=at.safeParse(e);if(!o.success)throw new InvalidModelRequestError({info:"Invalid model request",cause:o.error});let t=o.data,s=t.model,r={outputDimensionality:t.outputDimensionality},n=Config().parse(removeUndefinedEntries(r)),m={modality:EmbeddingTextModalityLiteral,requests:t.requests.reduce((l,p)=>(l.push(...p.content.parts.map(f=>f.text)),l),[])};return {modelName:s,config:n,embeddingRequests:m}}transformConfig(e,o){let t=this.modelSchema.config.schema.safeParse(e);if(!t.success)throw new InvalidConfigError({info:`Invalid config for model : '${this.modelName}'`,cause:t.error});let s=t.data;return Object.keys(s).forEach(n=>{if(!this.modelSchema.config.def[n])throw new InvalidConfigError({info:`Invalid config for model : '${this.modelName}'`,cause:new Error(`Invalid config key : '${n}',
|
|
13
|
-
available keys : [${Object.keys(this.modelSchema.config.def).join(", ")}]`)})}),Object.keys(
|
|
5
|
+
var lt=Object.defineProperty;var Co=Object.getOwnPropertySymbols;var mt=Object.prototype.hasOwnProperty,dt=Object.prototype.propertyIsEnumerable;var pt=(i,e)=>(e=Symbol[i])?e:Symbol.for("Symbol."+i);var To=(i,e,o)=>e in i?lt(i,e,{enumerable:!0,configurable:!0,writable:!0,value:o}):i[e]=o,G=(i,e)=>{for(var o in e||(e={}))mt.call(e,o)&&To(i,o,e[o]);if(Co)for(var o of Co(e))dt.call(e,o)&&To(i,o,e[o]);return i};var R=(i,e,o)=>new Promise((t,n)=>{var l=r=>{try{m(o.next(r));}catch(p){n(p);}},s=r=>{try{m(o.throw(r));}catch(p){n(p);}},m=r=>r.done?t(r.value):Promise.resolve(r.value).then(l,s);m((o=o.apply(i,e)).next());}),ct=function(i,e){this[0]=i,this[1]=e;},Mo=(i,e,o)=>{var t=(s,m,r,p)=>{try{var f=o[s](m),C=(m=f.value)instanceof ct,M=f.done;Promise.resolve(C?m[0]:m).then(g=>C?t(s==="return"?s:"next",m[1]?{done:g.done,value:g.value}:g,r,p):r({value:g,done:M})).catch(g=>t("throw",g,r,p));}catch(g){p(g);}},n=s=>l[s]=m=>new Promise((r,p)=>t(s,m,r,p)),l={};return o=o.apply(i,e),l[pt("asyncIterator")]=()=>l,n("next"),n("throw"),n("return"),l};var v=(i,e)=>RangeConfigItem({param:"temperature",title:CHAT_CONFIG.TEMPERATURE.title,description:CHAT_CONFIG.TEMPERATURE.description,min:0,max:i,step:.01,default:e}),N=i=>RangeConfigItem({param:"maxOutputTokens",title:CHAT_CONFIG.MAX_TOKENS.title,description:CHAT_CONFIG.MAX_TOKENS.description,min:0,max:i,step:1,default:0}),F=i=>MultiStringConfigItem({param:"stopSequences",title:CHAT_CONFIG.STOP(i).title,description:CHAT_CONFIG.STOP(i).description,max:i}),q=i=>RangeConfigItem({param:"topP",title:CHAT_CONFIG.TOP_P.title,description:CHAT_CONFIG.TOP_P.description,min:0,max:1,step:.01,default:i}),xe=i=>RangeConfigItem({param:"topK",title:CHAT_CONFIG.TOP_K.title,description:CHAT_CONFIG.TOP_K.description,min:1,max:40,step:1,default:i}),we=RangeConfigItem({param:"frequencyPenalty",title:CHAT_CONFIG.FREQUENCY_PENALTY.title,description:CHAT_CONFIG.FREQUENCY_PENALTY.description,min:-2,max:2,step:.01,default:0}),Ie=RangeConfigItem({param:"presencePenalty",title:CHAT_CONFIG.PRESENCE_PENALTY.title,description:CHAT_CONFIG.PRESENCE_PENALTY.description,min:-2,max:2,step:.01,default:0}),Le=RangeConfigItem({param:"seed",title:CHAT_CONFIG.SEED.title,description:CHAT_CONFIG.SEED.description,min:0,max:1e6,step:1,default:0}),$=SelectStringConfigItem({param:"toolChoice",title:"Tool choice",description:"Controls which (if any) tool is called by the model. 'none' means the model will not call a function. 'auto' means the model can pick between generating a message or calling a tool.",default:"auto",choices:["auto","any","none"]}),z=ObjectSchemaConfigItem({param:"safetySettings",title:"Safety settings",description:"The safety rating contains the category of harm and the harm probability level in that category for a piece of content.",objectSchema:z$1.array(z$1.object({threshold:z$1.enum(["HARM_BLOCK_THRESHOLD_UNSPECIFIED","BLOCK_LOW_AND_ABOVE","BLOCK_MEDIUM_AND_ABOVE","BLOCK_ONLY_HIGH","BLOCK_NONE","OFF"]),category:z$1.enum(["HARM_CATEGORY_UNSPECIFIED","HARM_CATEGORY_HARASSMENT","HARM_CATEGORY_HATE_SPEECH","HARM_CATEGORY_SEXUALLY_EXPLICIT","HARM_CATEGORY_DANGEROUS_CONTENT","HARM_CATEGORY_CIVIC_INTEGRITY"])}))});var Go=(i,e,o,t,n)=>z$1.object({temperature:v(i,e).schema,maxTokens:N(o).schema,stop:F(t).schema,topP:q(n).schema,toolChoice:$.schema,safetySettings:z.schema}),bo=(i,e,o,t,n)=>({temperature:v(i,e).def,maxTokens:N(o).def,stop:F(t).def,topP:q(n).def,toolChoice:$.def,safetySettings:z.def});var _o=(i,e,o,t,n,l)=>z$1.object({temperature:v(i,e).schema,maxTokens:N(o).schema,stop:F(t).schema,topP:q(n).schema,topK:xe(l).schema,frequencyPenalty:we.schema,presencePenalty:Ie.schema,seed:Le.schema.transform(s=>s===0?void 0:s),toolChoice:$.schema,safetySettings:z.schema}),Eo=(i,e,o,t,n,l)=>({temperature:v(i,e).def,maxTokens:N(o).def,stop:F(t).def,topP:q(n).def,topK:xe(l).def,frequencyPenalty:we.def,presencePenalty:Ie.def,seed:Le.def,toolChoice:$.def,safetySettings:z.def});var ke=i=>RangeConfigItem({param:"outputDimensionality",title:EMBEDDING_CONFIG.DIMENSIONS.title,description:EMBEDDING_CONFIG.DIMENSIONS.description,min:1,max:i,step:1,default:i});var Po=i=>z$1.object({dimensions:ke(i).schema}),So=i=>({dimensions:ke(i).def});var c={base:(i,e,o,t,n)=>({def:bo(i,e,o,t,n),schema:Go(i,e,o,t,n)}),c1:(i,e,o,t,n,l)=>({def:Eo(i,e,o,t,n,l),schema:_o(i,e,o,t,n,l)})},j={base:i=>({def:So(i),schema:Po(i)})};var u=z$1.enum([SystemRoleLiteral,UserRoleLiteral,AssistantRoleLiteral,ToolRoleLiteral]),Et="model",Ot="function",y={system:UserRoleLiteral,user:UserRoleLiteral,assistant:Et,tool:Ot};var _=[TextModalityLiteral,ImageModalityLiteral,ToolCallModalityLiteral,ToolResponseModalityLiteral],E=z$1.enum([TextModalityLiteral,ImageModalityLiteral,ToolCallModalityLiteral,ToolResponseModalityLiteral]),on=[TextModalityLiteral],tn=z$1.enum([TextModalityLiteral]),ae=[TextModalityLiteral,ImageModalityLiteral],re=z$1.enum([TextModalityLiteral,ImageModalityLiteral]),I=[TextModalityLiteral,ToolCallModalityLiteral,ToolResponseModalityLiteral],L=z$1.enum([TextModalityLiteral,ToolCallModalityLiteral,ToolResponseModalityLiteral]);var Pt=z$1.object({text:z$1.string()}),St=z$1.object({functionCall:z$1.object({name:z$1.string(),args:z$1.record(z$1.any())})}),Ro=z$1.object({candidates:z$1.array(z$1.object({content:z$1.object({role:z$1.string(),parts:z$1.array(z$1.union([Pt,St]))}).optional(),finishReason:z$1.string(),index:z$1.number().optional(),safetyRatings:z$1.optional(z$1.array(z$1.object({category:z$1.string(),probability:z$1.string(),blocked:z$1.boolean().optional()})))})),promptFeedback:z$1.optional(z$1.object({safetyRatings:z$1.optional(z$1.array(z$1.object({category:z$1.string(),probability:z$1.string()})))})),usageMetadata:z$1.object({promptTokenCount:z$1.number(),cachedContentTokenCount:z$1.number().optional(),candidatesTokenCount:z$1.number().optional(),totalTokenCount:z$1.number()}).optional()}),Rt=z$1.object({text:z$1.string()}),xt=z$1.object({functionCall:z$1.object({name:z$1.string(),args:z$1.record(z$1.any())})}),xo=z$1.object({candidates:z$1.array(z$1.object({content:z$1.object({role:z$1.string(),parts:z$1.array(z$1.union([Rt,xt]))}).optional(),finishReason:z$1.string().optional(),index:z$1.number().optional(),safetyRatings:z$1.optional(z$1.array(z$1.object({category:z$1.string(),probability:z$1.string(),blocked:z$1.boolean().optional()})))})),promptFeedback:z$1.optional(z$1.object({safetyRatings:z$1.optional(z$1.array(z$1.object({category:z$1.string(),probability:z$1.string()})))})),usageMetadata:z$1.object({promptTokenCount:z$1.number(),cachedContentTokenCount:z$1.number().optional(),candidatesTokenCount:z$1.number(),totalTokenCount:z$1.number()}).optional()});var vo=z$1.object({text:z$1.string().min(1)}),wt=z$1.object({inline_data:z$1.object({mime_type:z$1.string().min(1),data:z$1.string().base64()})}),It=z$1.object({function_call:z$1.object({name:z$1.string().min(1),args:z$1.record(z$1.string().min(1))})}),Lt=z$1.object({function_response:z$1.object({name:z$1.string().min(1),response:z$1.record(z$1.string().min(1))})}),kt=z$1.object({role:z$1.enum(["user","model","function"]),parts:z$1.array(z$1.union([vo,wt,It,Lt]))}),wo=z$1.object({parts:z$1.array(vo)}),vt=z$1.object({name:z$1.string().min(1),description:z$1.string().min(1),parameters:z$1.any()}),Io=z$1.object({function_calling_config:z$1.object({mode:z$1.enum(["ANY","AUTO","NONE"]),allowed_function_names:z$1.array(z$1.string()).optional()})}),Lo=z$1.object({stopSequences:z$1.array(z$1.string()).optional(),maxOutputTokens:z$1.number().optional(),temperature:z$1.number().optional(),topP:z$1.number().optional(),topK:z$1.number().optional(),presencePenalty:z$1.number().optional(),frequencyPenalty:z$1.number().optional(),seed:z$1.number().optional()}),ko=z$1.object({category:z$1.enum(["HARM_CATEGORY_HARASSMENT","HARM_CATEGORY_HATE_SPEECH","HARM_CATEGORY_SEXUALLY_EXPLICIT","HARM_CATEGORY_DANGEROUS_CONTENT","HARM_CATEGORY_CIVIC_INTEGRITY"]),threshold:z$1.enum(["HARM_BLOCK_THRESHOLD_UNSPECIFIED","BLOCK_LOW_AND_ABOVE","BLOCK_MEDIUM_AND_ABOVE","BLOCK_ONLY_HIGH","BLOCK_NONE","OFF"])}),No=z$1.object({model:z$1.string().min(1).optional(),contents:z$1.array(kt),systemInstruction:wo.optional(),system_instruction:wo.optional(),generationConfig:Lo.optional(),generation_config:Lo.optional(),safetySettings:z$1.array(ko).optional(),safety_settings:z$1.array(ko).optional(),tools:z$1.object({function_declarations:z$1.array(vt)}).optional(),toolConfig:Io.optional(),tool_config:Io.optional()});var Nt="google",Y=class{constructor(){this.version="v1";this.name=Nt;this.chatModelFactories={[Ne]:{model:le,modelOptions:qo,modelSchema:Fe},[qe]:{model:me,modelOptions:$o,modelSchema:$e},[Je]:{model:ge,modelOptions:Ao,modelSchema:We},[Xe]:{model:ue,modelOptions:Vo,modelSchema:Qe},[Ze]:{model:ye,modelOptions:Ho,modelSchema:eo},[Ke]:{model:he,modelOptions:Uo,modelSchema:Ye},[so]:{model:Te,modelOptions:Yo,modelSchema:no},[io]:{model:Me,modelOptions:Jo,modelSchema:ao},[ro]:{model:Ge,modelOptions:Wo,modelSchema:lo},[oo]:{model:Ce,modelOptions:Ko,modelSchema:to},[ze]:{model:de,modelOptions:zo,modelSchema:je},[Ve]:{model:fe,modelOptions:Do,modelSchema:He},[Ue]:{model:ce,modelOptions:Bo,modelSchema:Ae},[Be]:{model:pe,modelOptions:jo,modelSchema:De}};this.embeddingModelFactories={[mo]:{model:be,modelOptions:Xo,modelSchema:po},[co]:{model:_e,modelOptions:Qo,modelSchema:fo}};}chatModelLiterals(){return Object.keys(this.chatModelFactories)}chatModelSchemas(){return Object.keys(this.chatModelFactories).reduce((e,o)=>(e[o]=this.chatModelFactories[o].modelSchema,e),{})}chatModel(e){let o=e.modelName;if(!(o in this.chatModelFactories))throw new ProviderError({info:`Google chat model: ${o} not found`,cause:new Error(`Google chat model: ${o} not found, available chat models:
|
|
6
|
+
[${this.chatModelLiterals().join(", ")}]`)});let t=this.chatModelFactories[o].model,n=this.chatModelFactories[o].modelOptions.parse(e);return new t(n)}embeddingModelLiterals(){return Object.keys(this.embeddingModelFactories)}embeddingModelSchemas(){return Object.keys(this.embeddingModelFactories).reduce((e,o)=>(e[o]=this.embeddingModelFactories[o].modelSchema,e),{})}embeddingModel(e){let o=e.modelName;if(!(o in this.embeddingModelFactories))throw new ProviderError({info:`Google embedding model: ${o} not found`,cause:new Error(`Google embedding model: ${o} not found, available embedding models:
|
|
7
|
+
[${this.embeddingModelLiterals().join(", ")}]`)});let t=this.embeddingModelFactories[o].model,n=this.embeddingModelFactories[o].modelOptions.parse(e);return new t(n)}};Y.baseUrl="https://generativelanguage.googleapis.com/v1beta";var h=z$1.object({modelName:z$1.string(),apiKey:z$1.string(),baseUrl:z$1.string().url().optional(),completeChatUrl:z$1.string().url().optional(),streamChatUrl:z$1.string().url().optional()}),T=class{constructor(e,o){this.version="v1";var n;let t=h.parse(o);this.modelSchema=e,this.modelName=t.modelName,this.apiKey=t.apiKey,this.baseUrl=urlWithoutTrailingSlash((n=t.baseUrl)!=null?n:Y.baseUrl),this.completeChatUrl=urlWithoutTrailingSlash(t.completeChatUrl||`${this.baseUrl}/models/${this.modelName}:generateContent?key=${this.apiKey}`),this.streamChatUrl=urlWithoutTrailingSlash(t.streamChatUrl||`${this.baseUrl}/models/${this.modelName}:streamGenerateContent?key=${this.apiKey}`);}getDefaultBaseUrl(){return this.baseUrl}getDefaultHeaders(){return {"Content-Type":"application/json"}}getDefaultParams(){return {}}getRetryDelay(e){return {shouldRetry:!1,delayMs:0}}getTokenCount(e){return e.reduce((o,t)=>o+t.content.map(n=>n.modality==="text"?n.value:"").join(" ").length,0)}transformModelRequest(e){let o=No.safeParse(e);if(!o.success)throw new InvalidModelRequestError({info:"Invalid model request",cause:o.error});let t=o.data,n=t.model;if(t.system_instruction&&t.systemInstruction)throw new InvalidModelRequestError({info:`Invalid model request for model : '${this.modelName}'`,cause:new Error("'system_instruction' and 'systemInstruction' are not allowed at the same time")});if(t.generation_config&&t.generationConfig)throw new InvalidModelRequestError({info:`Invalid model request for model : '${this.modelName}'`,cause:new Error("'generation_config' and 'generationConfig' are not allowed at the same time")});if(t.tool_config&&t.toolConfig)throw new InvalidModelRequestError({info:`Invalid model request for model : '${this.modelName}'`,cause:new Error("'tool_config' and 'toolConfig' are not allowed at the same time")});let l=t.system_instruction||t.systemInstruction,s=t.generation_config||t.generationConfig,m=t.safety_settings||t.safetySettings,r=t.tool_config||t.toolConfig;if(r&&(!t.tools||t.tools.function_declarations.length===0))throw new InvalidModelRequestError({info:`Invalid model request for model : '${this.modelName}'`,cause:new Error("'tools' are required when 'tool_choice' is specified")});let p={};r&&(r.function_calling_config.mode==="ANY"&&r.function_calling_config.allowed_function_names&&r.function_calling_config.allowed_function_names.length===1?p.toolChoice=r.function_calling_config.allowed_function_names[0]:p.toolChoice=r.function_calling_config.mode.toLowerCase()),p.seed=s==null?void 0:s.seed,p.maxTokens=s==null?void 0:s.maxOutputTokens,p.temperature=s==null?void 0:s.temperature,p.topP=s==null?void 0:s.topP,p.presencePenalty=s==null?void 0:s.presencePenalty,p.frequencyPenalty=s==null?void 0:s.frequencyPenalty,p.stop=s==null?void 0:s.stopSequences,p.safetySettings=m;let f=Config().parse(removeUndefinedEntries(p)),C=[];l&&l.parts.forEach(g=>{C.push({role:SystemRoleLiteral,content:[{modality:TextModalityLiteral,value:g.text}]});}),t.contents.forEach(g=>{let V=g.role;switch(V){case"user":{let H=g.parts.map(S=>"text"in S?{modality:TextModalityLiteral,value:S.text}:{modality:ImageModalityLiteral,detail:"auto",value:{type:Base64ImageContentTypeLiteral,base64:S.inline_data.data,media_type:S.inline_data.mime_type.split("/")[1]}});C.push({role:V,content:H});}break;case"model":{let H=g.parts.map((S,K)=>"text"in S?{modality:TextModalityLiteral,value:S.text}:{modality:ToolCallModalityLiteral,id:K.toString(),index:K,name:S.function_call.name,arguments:JSON.stringify(S.function_call.args)});C.push({role:AssistantRoleLiteral,content:H});}break;case"function":{let H=g.parts.map((S,K)=>({modality:ToolResponseModalityLiteral,id:K.toString(),index:K,name:S.function_response.name,data:JSON.stringify(S.function_response.response)}));C.push({role:ToolRoleLiteral,content:H});}break;default:throw new InvalidMessagesError({info:`Invalid message 'role' for model : ${this.modelName}`,cause:new Error(`role : '${g.role}' is not supported for model : ${this.modelName}`)})}});let M=[];return t.tools&&t.tools.function_declarations.forEach(g=>{M.push({type:"function",definition:{schema:{name:g.name,description:g.description,parameters:g.parameters}}});}),{modelName:n,config:f,messages:C,tools:M.length>0?M:void 0}}transformConfig(e,o,t){let n=e.toolChoice;delete e.toolChoice;let l=this.modelSchema.config.schema.safeParse(e);if(!l.success)throw new InvalidConfigError({info:`Invalid config for model : '${this.modelName}'`,cause:l.error});let s=l.data;Object.keys(s).forEach(f=>{if(!(f in this.modelSchema.config.def))throw new InvalidConfigError({info:`Invalid config for model : '${this.modelName}'`,cause:new Error(`Invalid config key : '${f}',
|
|
8
|
+
available keys : [${Object.keys(this.modelSchema.config.def).join(", ")}]`)})});let m=Object.keys(s).reduce((f,C)=>{let M=this.modelSchema.config.def[C],g=M.param,V=s[C];return g==="maxOutputTokens"&&M.type==="range"&&V===0?f[g]=M.max:f[g]=V,f},{}),r=m.safetySettings;delete m.safetySettings;let p;if(n!==void 0){let f=n;if(!t||t&&t.length===0)throw new InvalidConfigError({info:`Invalid config for model : '${this.modelName}'`,cause:new Error("'tools' are required when 'toolChoice' is specified")});if(t&&t.length>0){let C=this.modelSchema.config.def.toolChoice;if(C.choices.includes(f))f==="any"?p={function_calling_config:{mode:"ANY",allowed_function_names:t.map(M=>M.definition.schema.name)}}:p={function_calling_config:{mode:f.toUpperCase()}};else if(t.map(M=>M.definition.schema.name).includes(f))p={function_calling_config:{mode:"ANY",allowed_function_names:[f]}};else throw new InvalidConfigError({info:`Invalid config for model : '${this.modelName}'`,cause:new Error(`toolChoice : '${f}' is not part of provided 'tools' names or
|
|
9
|
+
one of [${C.choices.join(", ")}]`)})}}return G(G({generation_config:m},p?{tool_config:p}:{}),r?{safety_settings:r}:{})}transformMessages(e){if(!e||e&&e.length===0)return {messages:[]};let o=e.map(s=>{let m=Message().safeParse(s);if(!m.success)throw new InvalidMessagesError({info:"Invalid messages",cause:m.error});return m.data});o.forEach(s=>{s.content.forEach(m=>{if(!this.modelSchema.modalities.includes(m.modality))throw new InvalidMessagesError({info:`Invalid message content for model : '${this.modelName}'`,cause:new Error(`model : '${this.modelName}' does not support modality : '${m.modality}',
|
|
10
|
+
available modalities : [${this.modelSchema.modalities.join(", ")}]`)})});}),o.forEach(s=>{if(!Object.keys(this.modelSchema.roles).includes(s.role))throw new InvalidMessagesError({info:`Invalid message content for model : '${this.modelName}'`,cause:new Error(`model : '${this.modelName}' does not support role : '${s.role}',
|
|
11
|
+
available roles : [${Object.keys(this.modelSchema.roles).join(", ")}]`)})});let t={parts:[]},n=[];if(o.forEach(s=>{switch(s.role){case SystemRoleLiteral:s.content.forEach(m=>{if(m.modality===TextModalityLiteral)t.parts.push({text:m.value});else throw new InvalidMessagesError({info:`Invalid message 'role' and 'modality' combination for model : ${this.modelName}`,cause:new Error(`role : '${s.role}' cannot have content with modality : '${m.modality}'`)})});break;case AssistantRoleLiteral:{let m=[];s.content.forEach(r=>{if(r.modality===TextModalityLiteral)m.push({text:r.value});else if(r.modality===ToolCallModalityLiteral)m.push({function_call:{name:r.name,args:JSON.parse(r.arguments)}});else throw new InvalidMessagesError({info:`Invalid message 'role' and 'modality' combination for model : ${this.modelName}`,cause:new Error(`role : '${s.role}' cannot have content with modality : '${r.modality}'`)})}),n.push({role:this.modelSchema.roles[s.role],parts:m});}break;case UserRoleLiteral:{let m=[];s.content.forEach(r=>{if(r.modality===TextModalityLiteral)m.push({text:r.value});else if(r.modality===ImageModalityLiteral){if(r.value.type==="base64")m.push({inline_data:{mime_type:r.value.media_type,data:r.value.base64}});else if(r.value.type==="url")throw new InvalidMessagesError({info:`Invalid message 'modality' for model : ${this.modelName}`,cause:new Error(`model: '${this.modelName}' does not support image content type: '${r.value.type}'`)})}else throw new InvalidMessagesError({info:`Invalid message 'role' and 'modality' combination for model : ${this.modelName}`,cause:new Error(`role : '${s.role}' cannot have content with modality : '${r.modality}'`)})}),n.push({role:this.modelSchema.roles[s.role],parts:m});}break;case ToolRoleLiteral:{let m=[];s.content.forEach(r=>{if(r.modality===ToolResponseModalityLiteral)m.push({function_response:{name:r.name,response:JSON.parse(r.data)}});else throw new InvalidMessagesError({info:`Invalid message 'role' and 'modality' combination for model : ${this.modelName}`,cause:new Error(`role : '${s.role}' cannot have content with modality : '${r.modality}'`)})}),n.push({role:this.modelSchema.roles[s.role],parts:m});}break;default:throw new InvalidMessagesError({info:`Invalid message 'role' for model : ${this.modelName}`,cause:new Error(`role : '${s.role}' is not supported,
|
|
12
|
+
available roles : [${Object.keys(this.modelSchema.roles).join(", ")}]`)})}}),n[0].role!==this.modelSchema.roles[UserRoleLiteral])throw new InvalidMessagesError({info:`Invalid message 'role' for model : ${this.modelName}`,cause:new Error(`model : '${this.modelName}' requires first message to be from user`)});let l=s=>s===this.modelSchema.roles[UserRoleLiteral]||s===this.modelSchema.roles[ToolRoleLiteral]?[this.modelSchema.roles[AssistantRoleLiteral]]:[this.modelSchema.roles[UserRoleLiteral],this.modelSchema.roles[ToolRoleLiteral]];for(let s=1;s<n.length;s++)if(!l(n[s-1].role).includes(n[s].role))throw new InvalidMessagesError({info:`Invalid message format for model : ${this.modelName}`,cause:new Error(`model : '${this.modelName}' cannot have message with role : '${n[s].role}' after message with role : '${n[s-1].role}'`)});if(n[n.length-1].role!==this.modelSchema.roles[UserRoleLiteral])throw new InvalidMessagesError({info:`Invalid message format for model : ${this.modelName}`,cause:new Error(`model : '${this.modelName}' requires last message to be from user`)});return G({contents:n},t.parts.length>0?{system_instruction:t}:{})}transformTools(e){if(!this.modelSchema.modalities.includes(ToolCallModalityLiteral))throw new InvalidToolsError({info:`Invalid tool 'modality' for model : ${this.modelName}`,cause:new Error(`model : '${this.modelName}' does not support tool modality : '${ToolCallModalityLiteral}'`)});return !e||e&&e.length===0?{tools:[]}:{tools:e.map(n=>{let l=Tool().safeParse(n);if(!l.success)throw new InvalidToolsError({info:"Invalid tools",cause:l.error});return l.data}).map(n=>({function_declarations:[{name:n.definition.schema.name,description:n.definition.schema.description,parameters:n.definition.schema.parameters}]}))}}getCompleteChatUrl(e,o,t){return R(this,null,function*(){return new Promise(n=>{n(this.completeChatUrl);})})}getCompleteChatHeaders(e,o,t){return R(this,null,function*(){return new Promise(n=>{n(this.getDefaultHeaders());})})}getCompleteChatData(e,o,t){return R(this,null,function*(){let n=this.transformConfig(e,o,t),l=this.transformMessages(o);if(l.messages&&l.messages.length===0)throw new InvalidMessagesError({info:"Messages are required",cause:new Error("Messages are required")});let s=t?this.transformTools(t):{};return new Promise(m=>{m(G(G(G(G({},this.getDefaultParams()),n),l),s));})})}transformCompleteChatResponse(e){let o=Ro.safeParse(e);if(o.success){if(o.data.candidates.length===0)throw new ModelResponseError({info:"Invalid response from model",cause:new Error(`No choices in response : ${JSON.stringify(o.data)}`)});let t=o.data,n=[],l,s=t.candidates[0].content;if(s){let p=s.parts.map((f,C)=>{if("text"in f&&f.text!==void 0)return createTextContent(f.text);if("functionCall"in f&&f.functionCall!==void 0)return createToolCallContent(C,`${f.functionCall.name}_${C}`,f.functionCall.name,JSON.stringify(f.functionCall.args))});return n.push({role:AssistantRoleLiteral,content:p}),t.usageMetadata&&(l={promptTokens:t.usageMetadata.promptTokenCount,totalTokens:t.usageMetadata.totalTokenCount,completionTokens:t.usageMetadata.candidatesTokenCount||0}),{messages:n,usage:l,logProbs:void 0}}let m=t.candidates[0].safetyRatings;if(m&&m.length>0&&m.forEach(p=>{if(p.blocked)throw new ModelResponseError({info:`Blocked content for category: ${p.category} with probability: ${p.probability}`,cause:new Error(`Blocked content for category: ${p.category} with probability: ${p.probability}`)})}),t.candidates[0].finishReason==="SAFETY")throw new ModelResponseError({info:"Blocked content, model response finished with safety reason",cause:new Error("Blocked content, model response finished with safety reason")})}throw new ModelResponseError({info:"Invalid response from model",cause:o.error})}getStreamChatUrl(e,o,t){return R(this,null,function*(){return new Promise(n=>{n(this.streamChatUrl);})})}getStreamChatHeaders(e,o,t){return R(this,null,function*(){return new Promise(n=>{n(this.getDefaultHeaders());})})}getStreamChatData(e,o,t){return R(this,null,function*(){let n=this.transformConfig(e,o,t),l=this.transformMessages(o);if(l.messages&&l.messages.length===0)throw new InvalidMessagesError({info:"Messages are required",cause:new Error("Messages are required")});let s=t?this.transformTools(t):{};return new Promise(m=>{m(G(G(G(G({},this.getDefaultParams()),n),l),s));})})}transformStreamChatResponseChunk(e,o){return Mo(this,null,function*(){let t=(o+e).split(",\r").filter(n=>n.trim()!=="");for(let n of t){let l=n;if(l=l.replace(/\n/g,""),l.startsWith("["))l=l.slice(1);else if(l.endsWith("]")){if(l==="]")return;l=l.slice(0,-1);}let s;try{s=JSON.parse(l);}catch(r){if(r instanceof SyntaxError){o=l;continue}else throw r}o="";let m=xo.safeParse(s);if(m.success){let r={partialMessages:[]},p=m.data;if(p.candidates.length>0){let f=p.candidates[0].content;f&&"parts"in f&&f.parts.length>0&&f.parts.forEach((C,M)=>{if("text"in C&&C.text!==void 0&&r.partialMessages.push(createPartialTextMessage(AssistantRoleLiteral,C.text)),"functionCall"in C&&C.functionCall!==void 0){let g=C.functionCall;r.partialMessages.push(createPartialToolCallMessage(AssistantRoleLiteral,M,`${g.name}_${M}`,g.name,JSON.stringify(g.args)));}});}p.usageMetadata&&(r.usage={promptTokens:p.usageMetadata.promptTokenCount,completionTokens:p.usageMetadata.candidatesTokenCount,totalTokens:p.usageMetadata.totalTokenCount}),yield {partialResponse:r,buffer:o};}else throw new ModelResponseError({info:"Invalid response from model",cause:m.error})}yield {partialResponse:{partialMessages:[]},buffer:o};})}};var P=class extends T{transformMessages(e){let o=super.transformMessages(e);if(o.systemInstruction){let t={role:this.modelSchema.roles[UserRoleLiteral],parts:o.systemInstruction.parts};o.contents.unshift(t),delete o.systemInstruction;}return o}};var Ne="gemini-pro",Kt="A model for scaling across a wide range of tasks Optimized for natural language tasks, multi-turn text and code chat, and code generation",Fe=ChatModelSchema(u,L).parse({name:Ne,description:Kt,maxInputTokens:30720,maxOutputTokens:2048,roles:y,modalities:I,config:{def:c.base(1,.9,2048,4,1).def,schema:c.base(1,.9,2048,4,1).schema}}),qo=h,le=class extends P{constructor(e){super(Fe,e);}};var qe="gemini-pro-vision",Jt="An image understanding model to handle a broad range of applications",$e=ChatModelSchema(u,re).parse({name:qe,description:Jt,maxInputTokens:12288,maxOutputTokens:4096,roles:y,modalities:ae,config:{def:c.c1(1,.4,4096,4,1,32).def,schema:c.c1(1,.4,4096,4,1,32).schema}}),$o=h,me=class extends P{constructor(e){super($e,e);}};var ze="gemini-1.0-pro",Xt="Google's predecessor to Gemini 1.5 Pro, a model for scaling across a wide range of tasks Optimized for natural language tasks, multi-turn text and code chat, and code generation",je=ChatModelSchema(u,L).parse({name:ze,description:Xt,maxInputTokens:30720,maxOutputTokens:2048,roles:y,modalities:I,config:{def:c.base(1,.9,2048,4,1).def,schema:c.base(1,.9,2048,4,1).schema}}),zo=h,de=class extends P{constructor(e){super(je,e);}};var Be="gemini-1.0-pro-001",Zt="Google's predecessor to Gemini 1.5 Pro, a model for scaling across a wide range of tasks Optimized for natural language tasks, multi-turn text and code chat, and code generation",De=ChatModelSchema(u,L).parse({name:Be,description:Zt,maxInputTokens:30720,maxOutputTokens:2048,roles:y,modalities:I,config:{def:c.base(1,.9,2048,4,1).def,schema:c.base(1,.9,2048,4,1).schema}}),jo=h,pe=class extends P{constructor(e){super(De,e);}};var Ue="gemini-1.0-pro-latest",os="Google's latest multimodal model with great performance for high-frequency tasks. Optimized for natural language tasks, multi-turn text and code chat, and code generation",Ae=ChatModelSchema(u,L).parse({name:Ue,description:os,maxInputTokens:30720,maxOutputTokens:2048,roles:y,modalities:I,config:{def:c.base(1,.9,2048,4,1).def,schema:c.base(1,.9,2048,4,1).schema}}),Bo=h,ce=class extends P{constructor(e){super(Ae,e);}};var Ve="gemini-1.0-pro-vision",ss="Google's predecessor to Gemini 1.5 Pro, an image understanding model to handle a broad range of applications",He=ChatModelSchema(u,re).parse({name:Ve,description:ss,maxInputTokens:12288,maxOutputTokens:4096,roles:y,modalities:ae,config:{def:c.c1(1,.4,4096,4,1,32).def,schema:c.c1(1,.4,4096,4,1,32).schema}}),Do=h,fe=class extends P{constructor(e){super(He,e);}};var Ke="gemini-1.5-flash",is="Google's fastest, most cost-efficient multimodal model with great performance for high-frequency tasks. Optimized for fast and versatile performance across a diverse variety of tasks",Ye=ChatModelSchema(u,E).parse({name:Ke,description:is,maxInputTokens:1e6,maxOutputTokens:8192,roles:y,modalities:_,config:{def:c.c1(2,1,8192,4,.95,64).def,schema:c.c1(2,1,8192,4,.95,64).schema}}),Uo=h,he=class extends T{constructor(e){super(Ye,e);}};var Je="gemini-1.5-flash-001",rs="Google's fastest, most cost-efficient multimodal model with great performance for high-frequency tasks. Optimized for fast and versatile performance across a diverse variety of tasks",We=ChatModelSchema(u,E).parse({name:Je,description:rs,maxInputTokens:1e6,maxOutputTokens:8192,roles:y,modalities:_,config:{def:c.c1(2,1,8192,4,.95,64).def,schema:c.c1(2,1,8192,4,.95,64).schema}}),Ao=h,ge=class extends T{constructor(e){super(We,e);}};var Xe="gemini-1.5-flash-002",ms="Google's fastest, most cost-efficient multimodal model with great performance for high-frequency tasks. Optimized for fast and versatile performance across a diverse variety of tasks",Qe=ChatModelSchema(u,E).parse({name:Xe,description:ms,maxInputTokens:1e6,maxOutputTokens:8192,roles:y,modalities:_,config:{def:c.c1(2,1,8192,4,.95,40).def,schema:c.c1(2,1,8192,4,.95,40).schema}}),Vo=h,ue=class extends T{constructor(e){super(Qe,e);}};var Ze="gemini-1.5-flash-latest",ps="Google's latest multimodal model with great performance for high-frequency tasks. Optimized for fast and versatile performance across a diverse variety of tasks",eo=ChatModelSchema(u,E).parse({name:Ze,description:ps,maxInputTokens:1e6,maxOutputTokens:8192,roles:y,modalities:_,config:{def:c.c1(2,1,8192,4,.95,64).def,schema:c.c1(2,1,8192,4,.95,64).schema}}),Ho=h,ye=class extends T{constructor(e){super(eo,e);}};var oo="gemini-1.5-pro",fs="Google's best performing multimodal model with features for a wide variety of reasoning tasks. Optimized for complex reasoning tasks requiring more intelligence",to=ChatModelSchema(u,E).parse({name:oo,description:fs,maxInputTokens:2e6,maxOutputTokens:8192,roles:y,modalities:_,config:{def:c.c1(2,1,8192,4,.95,64).def,schema:c.c1(2,1,8192,4,.95,64).schema}}),Ko=h,Ce=class extends T{constructor(e){super(to,e);}};var so="gemini-1.5-pro-001",gs="Google's best performing multimodal model with features for a wide variety of reasoning tasks. Optimized for complex reasoning tasks requiring more intelligence",no=ChatModelSchema(u,E).parse({name:so,description:gs,maxInputTokens:2e6,maxOutputTokens:8192,roles:y,modalities:_,config:{def:c.c1(2,1,8192,4,.95,64).def,schema:c.c1(2,1,8192,4,.95,64).schema}}),Yo=h,Te=class extends T{constructor(e){super(no,e);}};var io="gemini-1.5-pro-002",ys="Google's best performing multimodal model with features for a wide variety of reasoning tasks. Optimized for complex reasoning tasks requiring more intelligence",ao=ChatModelSchema(u,E).parse({name:io,description:ys,maxInputTokens:2e6,maxOutputTokens:8192,roles:y,modalities:_,config:{def:c.c1(2,1,8192,4,.95,40).def,schema:c.c1(2,1,8192,4,.95,40).schema}}),Jo=h,Me=class extends T{constructor(e){super(ao,e);}};var ro="gemini-1.5-pro-latest",Ts="Google's best performing multimodal model with features for a wide variety of reasoning tasks. Optimized for complex reasoning tasks requiring more intelligence",lo=ChatModelSchema(u,E).parse({name:ro,description:Ts,maxInputTokens:2e6,maxOutputTokens:8192,roles:y,modalities:_,config:{def:c.c1(2,1,8192,4,.95,64).def,schema:c.c1(2,1,8192,4,.95,64).schema}}),Wo=h,Ge=class extends T{constructor(e){super(lo,e);}};var Se=[EmbeddingTextModalityLiteral],Re=z$1.enum([EmbeddingTextModalityLiteral]);var nt=z$1.object({embeddings:z$1.array(z$1.object({values:z$1.array(z$1.number())}))});var Gs=z$1.object({model:z$1.string().min(1),content:z$1.object({parts:z$1.array(z$1.object({text:z$1.string().min(1)})).min(1)})}),it=z$1.object({model:z$1.string().min(1).optional(),requests:z$1.array(Gs).min(1),outputDimensionality:z$1.number().int().min(1).optional()});var ee=z$1.object({modelName:z$1.string(),apiKey:z$1.string(),baseUrl:z$1.string().url(),getEmbeddingsUrl:z$1.string().url().optional()}),A=class{constructor(e,o){this.version="v1";let t=ee.parse(o);this.modelSchema=e,this.modelName=t.modelName,this.apiKey=t.apiKey,this.baseUrl=urlWithoutTrailingSlash(t.baseUrl),this.getEmbeddingsUrl=urlWithoutTrailingSlash(t.getEmbeddingsUrl||`${this.baseUrl}/models/${this.modelName}:batchEmbedContents?key=${this.apiKey}`);}getDefaultBaseUrl(){return this.baseUrl}getDefaultHeaders(){return {"Content-Type":"application/json"}}getDefaultParams(){return {model:this.modelName}}getRetryDelay(e){return {shouldRetry:!1,delayMs:0}}getTokenCount(e){return e.requests.reduce((o,t)=>o+t.length,0)}transformModelRequest(e){let o=it.safeParse(e);if(!o.success)throw new InvalidModelRequestError({info:"Invalid model request",cause:o.error});let t=o.data,n=t.model,l={outputDimensionality:t.outputDimensionality},s=Config().parse(removeUndefinedEntries(l)),m={modality:EmbeddingTextModalityLiteral,requests:t.requests.reduce((r,p)=>(r.push(...p.content.parts.map(f=>f.text)),r),[])};return {modelName:n,config:s,embeddingRequests:m}}transformConfig(e,o){let t=this.modelSchema.config.schema.safeParse(e);if(!t.success)throw new InvalidConfigError({info:`Invalid config for model : '${this.modelName}'`,cause:t.error});let n=t.data;return Object.keys(n).forEach(s=>{if(!this.modelSchema.config.def[s])throw new InvalidConfigError({info:`Invalid config for model : '${this.modelName}'`,cause:new Error(`Invalid config key : '${s}',
|
|
13
|
+
available keys : [${Object.keys(this.modelSchema.config.def).join(", ")}]`)})}),Object.keys(n).reduce((s,m)=>{let p=this.modelSchema.config.def[m].param,f=n[m];return s[p]=f,s},{})}transformEmbeddingRequests(e){let o=EmbeddingRequests().safeParse(e);if(!o.success)throw new InvalidEmbeddingRequestsError({info:"Invalid embedding requests",cause:o.error});if(e.modality!==EmbeddingTextModalityLiteral)throw new InvalidEmbeddingRequestsError({info:`Invalid embedding requests for model : '${this.modelName}'`,cause:new Error(`Only '${EmbeddingTextModalityLiteral}' modality is supported for model : '${this.modelName}'`)});return {requests:o.data.requests.map(n=>({model:`models/${this.modelName}`,content:{parts:[{text:n}]}}))}}getGetEmbeddingsUrl(e,o){return R(this,null,function*(){return new Promise(t=>{t(this.getEmbeddingsUrl);})})}getGetEmbeddingsHeaders(e,o){return R(this,null,function*(){return new Promise(t=>{t(this.getDefaultHeaders());})})}getGetEmbeddingsData(e,o){return R(this,null,function*(){return new Promise(t=>{let n=this.transformConfig(e),l=this.transformEmbeddingRequests(o);if(o.requests.length===0)throw new InvalidEmbeddingRequestsError({info:`Invalid embedding requests for model : '${this.modelName}'`,cause:new Error("requests cannot be empty")});n.outputDimensionality&&(l.requests.forEach(s=>{s.outputDimensionality=n.outputDimensionality;}),delete n.outputDimensionality),t(G(G(G({},this.getDefaultParams()),n),l));})})}transformGetEmbeddingsResponse(e){let o=nt.safeParse(e);if(o.success){let n=o.data.embeddings.map((l,s)=>({index:s,embedding:l.values}));return {encodingFormat:FloatEmbeddingLiteral,embeddings:n}}throw new ModelResponseError({info:"Invalid response from model",cause:o.error})}};var mo="text-embedding-001",xs="text-embedding-001",po=EmbeddingModelSchema(Re).parse({name:mo,description:xs,modalities:Se,maxInputTokens:2048,maxOutputTokens:768,config:{def:j.base(768).def,schema:j.base(768).schema}}),Xo=ee,be=class extends A{constructor(e){super(po,e);}};var co="text-embedding-004",Is="text-embedding-004",fo=EmbeddingModelSchema(Re).parse({name:co,description:Is,modalities:Se,maxInputTokens:2048,maxOutputTokens:768,config:{def:j.base(768).def,schema:j.base(768).schema}}),Qo=ee,_e=class extends A{constructor(e){super(fo,e);}};
|
|
14
14
|
|
|
15
|
-
export { T as BaseChatModel, P as BaseChatModelGemini1, h as BaseChatModelOptions, A as BaseEmbeddingModel, ee as BaseEmbeddingModelOptions,
|
|
15
|
+
export { T as BaseChatModel, P as BaseChatModelGemini1, h as BaseChatModelOptions, A as BaseEmbeddingModel, ee as BaseEmbeddingModelOptions, bo as ChatModelBaseConfigDef, Go as ChatModelBaseConfigSchema, Eo as ChatModelC1ConfigDef, _o as ChatModelC1ConfigSchema, So as EmbeddingModelBaseConfigDef, Po as EmbeddingModelBaseConfigSchema, de as Gemini1_0Pro, ce as Gemini1_0ProLatest, Ue as Gemini1_0ProLatestLiteral, Bo as Gemini1_0ProLatestOptions, Ae as Gemini1_0ProLatestSchema, ze as Gemini1_0ProLiteral, zo as Gemini1_0ProOptions, je as Gemini1_0ProSchema, fe as Gemini1_0ProVision, Ve as Gemini1_0ProVisionLiteral, Do as Gemini1_0ProVisionOptions, He as Gemini1_0ProVisionSchema, pe as Gemini1_0Pro_001, Be as Gemini1_0Pro_001Literal, jo as Gemini1_0Pro_001Options, De as Gemini1_0Pro_001Schema, he as Gemini1_5Flash, ge as Gemini1_5Flash001, Je as Gemini1_5Flash001Literal, Ao as Gemini1_5Flash001Options, We as Gemini1_5Flash001Schema, ue as Gemini1_5Flash002, Xe as Gemini1_5Flash002Literal, Vo as Gemini1_5Flash002Options, Qe as Gemini1_5Flash002Schema, ye as Gemini1_5FlashLatest, Ze as Gemini1_5FlashLatestLiteral, Ho as Gemini1_5FlashLatestOptions, eo as Gemini1_5FlashLatestSchema, Ke as Gemini1_5FlashLiteral, Uo as Gemini1_5FlashOptions, Ye as Gemini1_5FlashSchema, Ce as Gemini1_5Pro, Te as Gemini1_5Pro001, so as Gemini1_5Pro001Literal, Yo as Gemini1_5Pro001Options, no as Gemini1_5Pro001Schema, Me as Gemini1_5Pro002, io as Gemini1_5Pro002Literal, Jo as Gemini1_5Pro002Options, ao as Gemini1_5Pro002Schema, Ge as Gemini1_5ProLatest, ro as Gemini1_5ProLatestLiteral, Wo as Gemini1_5ProLatestOptions, lo as Gemini1_5ProLatestSchema, oo as Gemini1_5ProLiteral, Ko as Gemini1_5ProOptions, to as Gemini1_5ProSchema, le as GeminiPro, Ne as GeminiProLiteral, qo as GeminiProOptions, Fe as GeminiProSchema, me as GeminiProVision, qe as GeminiProVisionLiteral, $o as GeminiProVisionOptions, $e as GeminiProVisionSchema, Y as Google, Et as GoogleChatAssistantRoleLiteral, kt as GoogleChatContent, It as GoogleChatContentPartFunctionCall, Lt as GoogleChatContentPartFunctionResponse, wt as GoogleChatContentPartInlineData, vo as GoogleChatContentPartText, Lo as GoogleChatGenerationConfig, c as GoogleChatModelConfigs, _ as GoogleChatModelModalities, E as GoogleChatModelModalitiesEnum, u as GoogleChatModelRoles, y as GoogleChatModelRolesMap, on as GoogleChatModelTextModalities, tn as GoogleChatModelTextModalitiesEnum, I as GoogleChatModelTextToolModalities, L as GoogleChatModelTextToolModalitiesEnum, ae as GoogleChatModelTextVisionModalities, re as GoogleChatModelTextVisionModalitiesEnum, No as GoogleChatRequest, ko as GoogleChatSafetySettings, wo as GoogleChatSystemInstruction, vt as GoogleChatTool, Io as GoogleChatToolConfig, Ot as GoogleChatToolRoleLiteral, Ro as GoogleCompleteChatResponse, Pt as GoogleCompleteChatTextResponse, St as GoogleCompleteChatToolResponse, j as GoogleEmbeddingModelConfigs, Se as GoogleEmbeddingModelModalities, Re as GoogleEmbeddingModelModalitiesEnum, it as GoogleEmbeddingRequest, Gs as GoogleEmbeddingRequestInput, nt as GoogleGetEmbeddingsResponse, xo as GoogleStreamChatResponse, Rt as GoogleStreamChatTextResponse, xt as GoogleStreamChatToolResponse, Nt as ProviderLiteral, be as Text_Embedding_001, mo as Text_Embedding_001Literal, Xo as Text_Embedding_001Options, po as Text_Embedding_001Schema, _e as Text_Embedding_004, co as Text_Embedding_004Literal, Qo as Text_Embedding_004Options, fo as Text_Embedding_004Schema, ke as dimensions, we as frequencyPenalty, N as maxTokens, Ie as presencePenalty, z as safetySettings, Le as seed, F as stop, v as temperature, $ as toolChoice, xe as topK, q as topP };
|
|
16
16
|
//# sourceMappingURL=index.mjs.map
|
|
17
17
|
//# sourceMappingURL=index.mjs.map
|