ide-assi 0.298.0 → 0.300.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -193608,7 +193608,7 @@ class IdeAi
193608
193608
  #invoke = async (path, params) => {
193609
193609
  const prompt = await IdeUtils.generatePrompt(path, params);
193610
193610
 
193611
- console.log(prompt);
193611
+ //console.log(prompt);
193612
193612
 
193613
193613
  try {
193614
193614
  const response = await this.#model.invoke([
@@ -193912,6 +193912,7 @@ class IdeAi
193912
193912
  if (apply.javascript) {
193913
193913
  await this.#generateTmplFile("/prompts/meta/U.BuildReactJsx.txt", "react.jsx", {
193914
193914
  userPrompt: userPrompt,
193915
+ mybatis: where.mybatis,
193915
193916
  originSrc: src.javascript,
193916
193917
  menuUrl: where.menu.url,
193917
193918
  menuName: where.menu.name,
@@ -193951,7 +193952,7 @@ class IdeAi
193951
193952
 
193952
193953
 
193953
193954
 
193954
- static generateWhereCause = async (xmlPath, queryId, apiKey) => {
193955
+ static generateWhereCause = async (xmlPath, queryId, userPrompt, apiKey) => {
193955
193956
 
193956
193957
  const invoke = async (path, params) => {
193957
193958
  const prompt = await IdeUtils.generatePrompt(path, params);
@@ -193604,7 +193604,7 @@ class IdeAi
193604
193604
  #invoke = async (path, params) => {
193605
193605
  const prompt = await IdeUtils.generatePrompt(path, params);
193606
193606
 
193607
- console.log(prompt);
193607
+ //console.log(prompt);
193608
193608
 
193609
193609
  try {
193610
193610
  const response = await this.#model.invoke([
@@ -193908,6 +193908,7 @@ class IdeAi
193908
193908
  if (apply.javascript) {
193909
193909
  await this.#generateTmplFile("/prompts/meta/U.BuildReactJsx.txt", "react.jsx", {
193910
193910
  userPrompt: userPrompt,
193911
+ mybatis: where.mybatis,
193911
193912
  originSrc: src.javascript,
193912
193913
  menuUrl: where.menu.url,
193913
193914
  menuName: where.menu.name,
@@ -193947,7 +193948,7 @@ class IdeAi
193947
193948
 
193948
193949
 
193949
193950
 
193950
- static generateWhereCause = async (xmlPath, queryId, apiKey) => {
193951
+ static generateWhereCause = async (xmlPath, queryId, userPrompt, apiKey) => {
193951
193952
 
193952
193953
  const invoke = async (path, params) => {
193953
193954
  const prompt = await IdeUtils.generatePrompt(path, params);
@@ -5,7 +5,6 @@ import { HumanMessage, SystemMessage } from '@langchain/core/messages';
5
5
  import { ChatGoogleGenerativeAI } from "@langchain/google-genai";
6
6
  import { Ollama } from "@langchain/ollama";
7
7
  import { ChatOpenAI } from '@langchain/openai';
8
- import { PromptTemplate } from "@langchain/core/prompts";
9
8
 
10
9
  export class IdeAi
11
10
  {
@@ -189,7 +188,7 @@ export class IdeAi
189
188
  #invoke = async (path, params) => {
190
189
  const prompt = await IdeUtils.generatePrompt(path, params);
191
190
 
192
- console.log(prompt);
191
+ //console.log(prompt);
193
192
 
194
193
  try {
195
194
  const response = await this.#model.invoke([
@@ -495,6 +494,7 @@ export class IdeAi
495
494
  if (apply.javascript) {
496
495
  jsSrc = await this.#generateTmplFile("/prompts/meta/U.BuildReactJsx.txt", "react.jsx", {
497
496
  userPrompt: userPrompt,
497
+ mybatis: where.mybatis,
498
498
  originSrc: src.javascript,
499
499
  menuUrl: where.menu.url,
500
500
  menuName: where.menu.name,
@@ -536,7 +536,7 @@ export class IdeAi
536
536
 
537
537
 
538
538
 
539
- static generateWhereCause = async (xmlPath, queryId, apiKey) => {
539
+ static generateWhereCause = async (xmlPath, queryId, userPrompt, apiKey) => {
540
540
 
541
541
  const invoke = async (path, params) => {
542
542
  const prompt = await IdeUtils.generatePrompt(path, params);
package/package.json CHANGED
@@ -1,7 +1,7 @@
1
1
  {
2
2
  "name": "ide-assi",
3
3
  "type": "module",
4
- "version": "0.298.0",
4
+ "version": "0.300.0",
5
5
  "main": "dist/index.js",
6
6
  "types": "dist/index.d.ts",
7
7
  "exports": {
@@ -5,7 +5,6 @@ import { HumanMessage, SystemMessage } from '@langchain/core/messages';
5
5
  import { ChatGoogleGenerativeAI } from "@langchain/google-genai";
6
6
  import { Ollama } from "@langchain/ollama";
7
7
  import { ChatOpenAI } from '@langchain/openai';
8
- import { PromptTemplate } from "@langchain/core/prompts";
9
8
 
10
9
  export class IdeAi
11
10
  {
@@ -189,7 +188,7 @@ export class IdeAi
189
188
  #invoke = async (path, params) => {
190
189
  const prompt = await IdeUtils.generatePrompt(path, params);
191
190
 
192
- console.log(prompt);
191
+ //console.log(prompt);
193
192
 
194
193
  try {
195
194
  const response = await this.#model.invoke([
@@ -495,6 +494,7 @@ export class IdeAi
495
494
  if (apply.javascript) {
496
495
  jsSrc = await this.#generateTmplFile("/prompts/meta/U.BuildReactJsx.txt", "react.jsx", {
497
496
  userPrompt: userPrompt,
497
+ mybatis: where.mybatis,
498
498
  originSrc: src.javascript,
499
499
  menuUrl: where.menu.url,
500
500
  menuName: where.menu.name,
@@ -536,7 +536,7 @@ export class IdeAi
536
536
 
537
537
 
538
538
 
539
- static generateWhereCause = async (xmlPath, queryId, apiKey) => {
539
+ static generateWhereCause = async (xmlPath, queryId, userPrompt, apiKey) => {
540
540
 
541
541
  const invoke = async (path, params) => {
542
542
  const prompt = await IdeUtils.generatePrompt(path, params);