@synsci/cli-darwin-x64 1.1.49
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/bin/skills/accelerate/SKILL.md +332 -0
- package/bin/skills/accelerate/references/custom-plugins.md +453 -0
- package/bin/skills/accelerate/references/megatron-integration.md +489 -0
- package/bin/skills/accelerate/references/performance.md +525 -0
- package/bin/skills/audiocraft/SKILL.md +564 -0
- package/bin/skills/audiocraft/references/advanced-usage.md +666 -0
- package/bin/skills/audiocraft/references/troubleshooting.md +504 -0
- package/bin/skills/autogpt/SKILL.md +403 -0
- package/bin/skills/autogpt/references/advanced-usage.md +535 -0
- package/bin/skills/autogpt/references/troubleshooting.md +420 -0
- package/bin/skills/awq/SKILL.md +310 -0
- package/bin/skills/awq/references/advanced-usage.md +324 -0
- package/bin/skills/awq/references/troubleshooting.md +344 -0
- package/bin/skills/axolotl/SKILL.md +158 -0
- package/bin/skills/axolotl/references/api.md +5548 -0
- package/bin/skills/axolotl/references/dataset-formats.md +1029 -0
- package/bin/skills/axolotl/references/index.md +15 -0
- package/bin/skills/axolotl/references/other.md +3563 -0
- package/bin/skills/bigcode-evaluation-harness/SKILL.md +405 -0
- package/bin/skills/bigcode-evaluation-harness/references/benchmarks.md +393 -0
- package/bin/skills/bigcode-evaluation-harness/references/custom-tasks.md +424 -0
- package/bin/skills/bigcode-evaluation-harness/references/issues.md +394 -0
- package/bin/skills/bitsandbytes/SKILL.md +411 -0
- package/bin/skills/bitsandbytes/references/memory-optimization.md +521 -0
- package/bin/skills/bitsandbytes/references/qlora-training.md +521 -0
- package/bin/skills/bitsandbytes/references/quantization-formats.md +447 -0
- package/bin/skills/blip-2/SKILL.md +564 -0
- package/bin/skills/blip-2/references/advanced-usage.md +680 -0
- package/bin/skills/blip-2/references/troubleshooting.md +526 -0
- package/bin/skills/chroma/SKILL.md +406 -0
- package/bin/skills/chroma/references/integration.md +38 -0
- package/bin/skills/clip/SKILL.md +253 -0
- package/bin/skills/clip/references/applications.md +207 -0
- package/bin/skills/constitutional-ai/SKILL.md +290 -0
- package/bin/skills/crewai/SKILL.md +498 -0
- package/bin/skills/crewai/references/flows.md +438 -0
- package/bin/skills/crewai/references/tools.md +429 -0
- package/bin/skills/crewai/references/troubleshooting.md +480 -0
- package/bin/skills/deepspeed/SKILL.md +141 -0
- package/bin/skills/deepspeed/references/08.md +17 -0
- package/bin/skills/deepspeed/references/09.md +173 -0
- package/bin/skills/deepspeed/references/2020.md +378 -0
- package/bin/skills/deepspeed/references/2023.md +279 -0
- package/bin/skills/deepspeed/references/assets.md +179 -0
- package/bin/skills/deepspeed/references/index.md +35 -0
- package/bin/skills/deepspeed/references/mii.md +118 -0
- package/bin/skills/deepspeed/references/other.md +1191 -0
- package/bin/skills/deepspeed/references/tutorials.md +6554 -0
- package/bin/skills/dspy/SKILL.md +590 -0
- package/bin/skills/dspy/references/examples.md +663 -0
- package/bin/skills/dspy/references/modules.md +475 -0
- package/bin/skills/dspy/references/optimizers.md +566 -0
- package/bin/skills/faiss/SKILL.md +221 -0
- package/bin/skills/faiss/references/index_types.md +280 -0
- package/bin/skills/flash-attention/SKILL.md +367 -0
- package/bin/skills/flash-attention/references/benchmarks.md +215 -0
- package/bin/skills/flash-attention/references/transformers-integration.md +293 -0
- package/bin/skills/gguf/SKILL.md +427 -0
- package/bin/skills/gguf/references/advanced-usage.md +504 -0
- package/bin/skills/gguf/references/troubleshooting.md +442 -0
- package/bin/skills/gptq/SKILL.md +450 -0
- package/bin/skills/gptq/references/calibration.md +337 -0
- package/bin/skills/gptq/references/integration.md +129 -0
- package/bin/skills/gptq/references/troubleshooting.md +95 -0
- package/bin/skills/grpo-rl-training/README.md +97 -0
- package/bin/skills/grpo-rl-training/SKILL.md +572 -0
- package/bin/skills/grpo-rl-training/examples/reward_functions_library.py +393 -0
- package/bin/skills/grpo-rl-training/templates/basic_grpo_training.py +228 -0
- package/bin/skills/guidance/SKILL.md +572 -0
- package/bin/skills/guidance/references/backends.md +554 -0
- package/bin/skills/guidance/references/constraints.md +674 -0
- package/bin/skills/guidance/references/examples.md +767 -0
- package/bin/skills/hqq/SKILL.md +445 -0
- package/bin/skills/hqq/references/advanced-usage.md +528 -0
- package/bin/skills/hqq/references/troubleshooting.md +503 -0
- package/bin/skills/hugging-face-cli/SKILL.md +191 -0
- package/bin/skills/hugging-face-cli/references/commands.md +954 -0
- package/bin/skills/hugging-face-cli/references/examples.md +374 -0
- package/bin/skills/hugging-face-datasets/SKILL.md +547 -0
- package/bin/skills/hugging-face-datasets/examples/diverse_training_examples.json +239 -0
- package/bin/skills/hugging-face-datasets/examples/system_prompt_template.txt +196 -0
- package/bin/skills/hugging-face-datasets/examples/training_examples.json +176 -0
- package/bin/skills/hugging-face-datasets/scripts/dataset_manager.py +522 -0
- package/bin/skills/hugging-face-datasets/scripts/sql_manager.py +844 -0
- package/bin/skills/hugging-face-datasets/templates/chat.json +55 -0
- package/bin/skills/hugging-face-datasets/templates/classification.json +62 -0
- package/bin/skills/hugging-face-datasets/templates/completion.json +51 -0
- package/bin/skills/hugging-face-datasets/templates/custom.json +75 -0
- package/bin/skills/hugging-face-datasets/templates/qa.json +54 -0
- package/bin/skills/hugging-face-datasets/templates/tabular.json +81 -0
- package/bin/skills/hugging-face-evaluation/SKILL.md +656 -0
- package/bin/skills/hugging-face-evaluation/examples/USAGE_EXAMPLES.md +382 -0
- package/bin/skills/hugging-face-evaluation/examples/artificial_analysis_to_hub.py +141 -0
- package/bin/skills/hugging-face-evaluation/examples/example_readme_tables.md +135 -0
- package/bin/skills/hugging-face-evaluation/examples/metric_mapping.json +50 -0
- package/bin/skills/hugging-face-evaluation/requirements.txt +20 -0
- package/bin/skills/hugging-face-evaluation/scripts/evaluation_manager.py +1374 -0
- package/bin/skills/hugging-face-evaluation/scripts/inspect_eval_uv.py +104 -0
- package/bin/skills/hugging-face-evaluation/scripts/inspect_vllm_uv.py +317 -0
- package/bin/skills/hugging-face-evaluation/scripts/lighteval_vllm_uv.py +303 -0
- package/bin/skills/hugging-face-evaluation/scripts/run_eval_job.py +98 -0
- package/bin/skills/hugging-face-evaluation/scripts/run_vllm_eval_job.py +331 -0
- package/bin/skills/hugging-face-evaluation/scripts/test_extraction.py +206 -0
- package/bin/skills/hugging-face-jobs/SKILL.md +1041 -0
- package/bin/skills/hugging-face-jobs/index.html +216 -0
- package/bin/skills/hugging-face-jobs/references/hardware_guide.md +336 -0
- package/bin/skills/hugging-face-jobs/references/hub_saving.md +352 -0
- package/bin/skills/hugging-face-jobs/references/token_usage.md +546 -0
- package/bin/skills/hugging-face-jobs/references/troubleshooting.md +475 -0
- package/bin/skills/hugging-face-jobs/scripts/cot-self-instruct.py +718 -0
- package/bin/skills/hugging-face-jobs/scripts/finepdfs-stats.py +546 -0
- package/bin/skills/hugging-face-jobs/scripts/generate-responses.py +587 -0
- package/bin/skills/hugging-face-model-trainer/SKILL.md +711 -0
- package/bin/skills/hugging-face-model-trainer/references/gguf_conversion.md +296 -0
- package/bin/skills/hugging-face-model-trainer/references/hardware_guide.md +283 -0
- package/bin/skills/hugging-face-model-trainer/references/hub_saving.md +364 -0
- package/bin/skills/hugging-face-model-trainer/references/reliability_principles.md +371 -0
- package/bin/skills/hugging-face-model-trainer/references/trackio_guide.md +189 -0
- package/bin/skills/hugging-face-model-trainer/references/training_methods.md +150 -0
- package/bin/skills/hugging-face-model-trainer/references/training_patterns.md +203 -0
- package/bin/skills/hugging-face-model-trainer/references/troubleshooting.md +282 -0
- package/bin/skills/hugging-face-model-trainer/scripts/convert_to_gguf.py +424 -0
- package/bin/skills/hugging-face-model-trainer/scripts/dataset_inspector.py +417 -0
- package/bin/skills/hugging-face-model-trainer/scripts/estimate_cost.py +150 -0
- package/bin/skills/hugging-face-model-trainer/scripts/train_dpo_example.py +106 -0
- package/bin/skills/hugging-face-model-trainer/scripts/train_grpo_example.py +89 -0
- package/bin/skills/hugging-face-model-trainer/scripts/train_sft_example.py +122 -0
- package/bin/skills/hugging-face-paper-publisher/SKILL.md +627 -0
- package/bin/skills/hugging-face-paper-publisher/examples/example_usage.md +327 -0
- package/bin/skills/hugging-face-paper-publisher/references/quick_reference.md +216 -0
- package/bin/skills/hugging-face-paper-publisher/scripts/paper_manager.py +508 -0
- package/bin/skills/hugging-face-paper-publisher/templates/arxiv.md +299 -0
- package/bin/skills/hugging-face-paper-publisher/templates/ml-report.md +358 -0
- package/bin/skills/hugging-face-paper-publisher/templates/modern.md +319 -0
- package/bin/skills/hugging-face-paper-publisher/templates/standard.md +201 -0
- package/bin/skills/hugging-face-tool-builder/SKILL.md +115 -0
- package/bin/skills/hugging-face-tool-builder/references/baseline_hf_api.py +57 -0
- package/bin/skills/hugging-face-tool-builder/references/baseline_hf_api.sh +40 -0
- package/bin/skills/hugging-face-tool-builder/references/baseline_hf_api.tsx +57 -0
- package/bin/skills/hugging-face-tool-builder/references/find_models_by_paper.sh +230 -0
- package/bin/skills/hugging-face-tool-builder/references/hf_enrich_models.sh +96 -0
- package/bin/skills/hugging-face-tool-builder/references/hf_model_card_frontmatter.sh +188 -0
- package/bin/skills/hugging-face-tool-builder/references/hf_model_papers_auth.sh +171 -0
- package/bin/skills/hugging-face-trackio/SKILL.md +65 -0
- package/bin/skills/hugging-face-trackio/references/logging_metrics.md +206 -0
- package/bin/skills/hugging-face-trackio/references/retrieving_metrics.md +223 -0
- package/bin/skills/huggingface-tokenizers/SKILL.md +516 -0
- package/bin/skills/huggingface-tokenizers/references/algorithms.md +653 -0
- package/bin/skills/huggingface-tokenizers/references/integration.md +637 -0
- package/bin/skills/huggingface-tokenizers/references/pipeline.md +723 -0
- package/bin/skills/huggingface-tokenizers/references/training.md +565 -0
- package/bin/skills/instructor/SKILL.md +740 -0
- package/bin/skills/instructor/references/examples.md +107 -0
- package/bin/skills/instructor/references/providers.md +70 -0
- package/bin/skills/instructor/references/validation.md +606 -0
- package/bin/skills/knowledge-distillation/SKILL.md +458 -0
- package/bin/skills/knowledge-distillation/references/minillm.md +334 -0
- package/bin/skills/lambda-labs/SKILL.md +545 -0
- package/bin/skills/lambda-labs/references/advanced-usage.md +611 -0
- package/bin/skills/lambda-labs/references/troubleshooting.md +530 -0
- package/bin/skills/langchain/SKILL.md +480 -0
- package/bin/skills/langchain/references/agents.md +499 -0
- package/bin/skills/langchain/references/integration.md +562 -0
- package/bin/skills/langchain/references/rag.md +600 -0
- package/bin/skills/langsmith/SKILL.md +422 -0
- package/bin/skills/langsmith/references/advanced-usage.md +548 -0
- package/bin/skills/langsmith/references/troubleshooting.md +537 -0
- package/bin/skills/litgpt/SKILL.md +469 -0
- package/bin/skills/litgpt/references/custom-models.md +568 -0
- package/bin/skills/litgpt/references/distributed-training.md +451 -0
- package/bin/skills/litgpt/references/supported-models.md +336 -0
- package/bin/skills/litgpt/references/training-recipes.md +619 -0
- package/bin/skills/llama-cpp/SKILL.md +258 -0
- package/bin/skills/llama-cpp/references/optimization.md +89 -0
- package/bin/skills/llama-cpp/references/quantization.md +213 -0
- package/bin/skills/llama-cpp/references/server.md +125 -0
- package/bin/skills/llama-factory/SKILL.md +80 -0
- package/bin/skills/llama-factory/references/_images.md +23 -0
- package/bin/skills/llama-factory/references/advanced.md +1055 -0
- package/bin/skills/llama-factory/references/getting_started.md +349 -0
- package/bin/skills/llama-factory/references/index.md +19 -0
- package/bin/skills/llama-factory/references/other.md +31 -0
- package/bin/skills/llamaguard/SKILL.md +337 -0
- package/bin/skills/llamaindex/SKILL.md +569 -0
- package/bin/skills/llamaindex/references/agents.md +83 -0
- package/bin/skills/llamaindex/references/data_connectors.md +108 -0
- package/bin/skills/llamaindex/references/query_engines.md +406 -0
- package/bin/skills/llava/SKILL.md +304 -0
- package/bin/skills/llava/references/training.md +197 -0
- package/bin/skills/lm-evaluation-harness/SKILL.md +490 -0
- package/bin/skills/lm-evaluation-harness/references/api-evaluation.md +490 -0
- package/bin/skills/lm-evaluation-harness/references/benchmark-guide.md +488 -0
- package/bin/skills/lm-evaluation-harness/references/custom-tasks.md +602 -0
- package/bin/skills/lm-evaluation-harness/references/distributed-eval.md +519 -0
- package/bin/skills/long-context/SKILL.md +536 -0
- package/bin/skills/long-context/references/extension_methods.md +468 -0
- package/bin/skills/long-context/references/fine_tuning.md +611 -0
- package/bin/skills/long-context/references/rope.md +402 -0
- package/bin/skills/mamba/SKILL.md +260 -0
- package/bin/skills/mamba/references/architecture-details.md +206 -0
- package/bin/skills/mamba/references/benchmarks.md +255 -0
- package/bin/skills/mamba/references/training-guide.md +388 -0
- package/bin/skills/megatron-core/SKILL.md +366 -0
- package/bin/skills/megatron-core/references/benchmarks.md +249 -0
- package/bin/skills/megatron-core/references/parallelism-guide.md +404 -0
- package/bin/skills/megatron-core/references/production-examples.md +473 -0
- package/bin/skills/megatron-core/references/training-recipes.md +547 -0
- package/bin/skills/miles/SKILL.md +315 -0
- package/bin/skills/miles/references/api-reference.md +141 -0
- package/bin/skills/miles/references/troubleshooting.md +352 -0
- package/bin/skills/mlflow/SKILL.md +704 -0
- package/bin/skills/mlflow/references/deployment.md +744 -0
- package/bin/skills/mlflow/references/model-registry.md +770 -0
- package/bin/skills/mlflow/references/tracking.md +680 -0
- package/bin/skills/modal/SKILL.md +341 -0
- package/bin/skills/modal/references/advanced-usage.md +503 -0
- package/bin/skills/modal/references/troubleshooting.md +494 -0
- package/bin/skills/model-merging/SKILL.md +539 -0
- package/bin/skills/model-merging/references/evaluation.md +462 -0
- package/bin/skills/model-merging/references/examples.md +428 -0
- package/bin/skills/model-merging/references/methods.md +352 -0
- package/bin/skills/model-pruning/SKILL.md +495 -0
- package/bin/skills/model-pruning/references/wanda.md +347 -0
- package/bin/skills/moe-training/SKILL.md +526 -0
- package/bin/skills/moe-training/references/architectures.md +432 -0
- package/bin/skills/moe-training/references/inference.md +348 -0
- package/bin/skills/moe-training/references/training.md +425 -0
- package/bin/skills/nanogpt/SKILL.md +290 -0
- package/bin/skills/nanogpt/references/architecture.md +382 -0
- package/bin/skills/nanogpt/references/data.md +476 -0
- package/bin/skills/nanogpt/references/training.md +564 -0
- package/bin/skills/nemo-curator/SKILL.md +383 -0
- package/bin/skills/nemo-curator/references/deduplication.md +87 -0
- package/bin/skills/nemo-curator/references/filtering.md +102 -0
- package/bin/skills/nemo-evaluator/SKILL.md +494 -0
- package/bin/skills/nemo-evaluator/references/adapter-system.md +340 -0
- package/bin/skills/nemo-evaluator/references/configuration.md +447 -0
- package/bin/skills/nemo-evaluator/references/custom-benchmarks.md +315 -0
- package/bin/skills/nemo-evaluator/references/execution-backends.md +361 -0
- package/bin/skills/nemo-guardrails/SKILL.md +297 -0
- package/bin/skills/nnsight/SKILL.md +436 -0
- package/bin/skills/nnsight/references/README.md +78 -0
- package/bin/skills/nnsight/references/api.md +344 -0
- package/bin/skills/nnsight/references/tutorials.md +300 -0
- package/bin/skills/openrlhf/SKILL.md +249 -0
- package/bin/skills/openrlhf/references/algorithm-comparison.md +404 -0
- package/bin/skills/openrlhf/references/custom-rewards.md +530 -0
- package/bin/skills/openrlhf/references/hybrid-engine.md +287 -0
- package/bin/skills/openrlhf/references/multi-node-training.md +454 -0
- package/bin/skills/outlines/SKILL.md +652 -0
- package/bin/skills/outlines/references/backends.md +615 -0
- package/bin/skills/outlines/references/examples.md +773 -0
- package/bin/skills/outlines/references/json_generation.md +652 -0
- package/bin/skills/peft/SKILL.md +431 -0
- package/bin/skills/peft/references/advanced-usage.md +514 -0
- package/bin/skills/peft/references/troubleshooting.md +480 -0
- package/bin/skills/phoenix/SKILL.md +475 -0
- package/bin/skills/phoenix/references/advanced-usage.md +619 -0
- package/bin/skills/phoenix/references/troubleshooting.md +538 -0
- package/bin/skills/pinecone/SKILL.md +358 -0
- package/bin/skills/pinecone/references/deployment.md +181 -0
- package/bin/skills/pytorch-fsdp/SKILL.md +126 -0
- package/bin/skills/pytorch-fsdp/references/index.md +7 -0
- package/bin/skills/pytorch-fsdp/references/other.md +4249 -0
- package/bin/skills/pytorch-lightning/SKILL.md +346 -0
- package/bin/skills/pytorch-lightning/references/callbacks.md +436 -0
- package/bin/skills/pytorch-lightning/references/distributed.md +490 -0
- package/bin/skills/pytorch-lightning/references/hyperparameter-tuning.md +556 -0
- package/bin/skills/pyvene/SKILL.md +473 -0
- package/bin/skills/pyvene/references/README.md +73 -0
- package/bin/skills/pyvene/references/api.md +383 -0
- package/bin/skills/pyvene/references/tutorials.md +376 -0
- package/bin/skills/qdrant/SKILL.md +493 -0
- package/bin/skills/qdrant/references/advanced-usage.md +648 -0
- package/bin/skills/qdrant/references/troubleshooting.md +631 -0
- package/bin/skills/ray-data/SKILL.md +326 -0
- package/bin/skills/ray-data/references/integration.md +82 -0
- package/bin/skills/ray-data/references/transformations.md +83 -0
- package/bin/skills/ray-train/SKILL.md +406 -0
- package/bin/skills/ray-train/references/multi-node.md +628 -0
- package/bin/skills/rwkv/SKILL.md +260 -0
- package/bin/skills/rwkv/references/architecture-details.md +344 -0
- package/bin/skills/rwkv/references/rwkv7.md +386 -0
- package/bin/skills/rwkv/references/state-management.md +369 -0
- package/bin/skills/saelens/SKILL.md +386 -0
- package/bin/skills/saelens/references/README.md +70 -0
- package/bin/skills/saelens/references/api.md +333 -0
- package/bin/skills/saelens/references/tutorials.md +318 -0
- package/bin/skills/segment-anything/SKILL.md +500 -0
- package/bin/skills/segment-anything/references/advanced-usage.md +589 -0
- package/bin/skills/segment-anything/references/troubleshooting.md +484 -0
- package/bin/skills/sentence-transformers/SKILL.md +255 -0
- package/bin/skills/sentence-transformers/references/models.md +123 -0
- package/bin/skills/sentencepiece/SKILL.md +235 -0
- package/bin/skills/sentencepiece/references/algorithms.md +200 -0
- package/bin/skills/sentencepiece/references/training.md +304 -0
- package/bin/skills/sglang/SKILL.md +442 -0
- package/bin/skills/sglang/references/deployment.md +490 -0
- package/bin/skills/sglang/references/radix-attention.md +413 -0
- package/bin/skills/sglang/references/structured-generation.md +541 -0
- package/bin/skills/simpo/SKILL.md +219 -0
- package/bin/skills/simpo/references/datasets.md +478 -0
- package/bin/skills/simpo/references/hyperparameters.md +452 -0
- package/bin/skills/simpo/references/loss-functions.md +350 -0
- package/bin/skills/skypilot/SKILL.md +509 -0
- package/bin/skills/skypilot/references/advanced-usage.md +491 -0
- package/bin/skills/skypilot/references/troubleshooting.md +570 -0
- package/bin/skills/slime/SKILL.md +464 -0
- package/bin/skills/slime/references/api-reference.md +392 -0
- package/bin/skills/slime/references/troubleshooting.md +386 -0
- package/bin/skills/speculative-decoding/SKILL.md +467 -0
- package/bin/skills/speculative-decoding/references/lookahead.md +309 -0
- package/bin/skills/speculative-decoding/references/medusa.md +350 -0
- package/bin/skills/stable-diffusion/SKILL.md +519 -0
- package/bin/skills/stable-diffusion/references/advanced-usage.md +716 -0
- package/bin/skills/stable-diffusion/references/troubleshooting.md +555 -0
- package/bin/skills/tensorboard/SKILL.md +629 -0
- package/bin/skills/tensorboard/references/integrations.md +638 -0
- package/bin/skills/tensorboard/references/profiling.md +545 -0
- package/bin/skills/tensorboard/references/visualization.md +620 -0
- package/bin/skills/tensorrt-llm/SKILL.md +187 -0
- package/bin/skills/tensorrt-llm/references/multi-gpu.md +298 -0
- package/bin/skills/tensorrt-llm/references/optimization.md +242 -0
- package/bin/skills/tensorrt-llm/references/serving.md +470 -0
- package/bin/skills/tinker/SKILL.md +362 -0
- package/bin/skills/tinker/references/api-reference.md +168 -0
- package/bin/skills/tinker/references/getting-started.md +157 -0
- package/bin/skills/tinker/references/loss-functions.md +163 -0
- package/bin/skills/tinker/references/models-and-lora.md +139 -0
- package/bin/skills/tinker/references/recipes.md +280 -0
- package/bin/skills/tinker/references/reinforcement-learning.md +212 -0
- package/bin/skills/tinker/references/rendering.md +243 -0
- package/bin/skills/tinker/references/supervised-learning.md +232 -0
- package/bin/skills/tinker-training-cost/SKILL.md +187 -0
- package/bin/skills/tinker-training-cost/scripts/calculate_cost.py +123 -0
- package/bin/skills/torchforge/SKILL.md +433 -0
- package/bin/skills/torchforge/references/api-reference.md +327 -0
- package/bin/skills/torchforge/references/troubleshooting.md +409 -0
- package/bin/skills/torchtitan/SKILL.md +358 -0
- package/bin/skills/torchtitan/references/checkpoint.md +181 -0
- package/bin/skills/torchtitan/references/custom-models.md +258 -0
- package/bin/skills/torchtitan/references/float8.md +133 -0
- package/bin/skills/torchtitan/references/fsdp.md +126 -0
- package/bin/skills/transformer-lens/SKILL.md +346 -0
- package/bin/skills/transformer-lens/references/README.md +54 -0
- package/bin/skills/transformer-lens/references/api.md +362 -0
- package/bin/skills/transformer-lens/references/tutorials.md +339 -0
- package/bin/skills/trl-fine-tuning/SKILL.md +455 -0
- package/bin/skills/trl-fine-tuning/references/dpo-variants.md +227 -0
- package/bin/skills/trl-fine-tuning/references/online-rl.md +82 -0
- package/bin/skills/trl-fine-tuning/references/reward-modeling.md +122 -0
- package/bin/skills/trl-fine-tuning/references/sft-training.md +168 -0
- package/bin/skills/unsloth/SKILL.md +80 -0
- package/bin/skills/unsloth/references/index.md +7 -0
- package/bin/skills/unsloth/references/llms-full.md +16799 -0
- package/bin/skills/unsloth/references/llms-txt.md +12044 -0
- package/bin/skills/unsloth/references/llms.md +82 -0
- package/bin/skills/verl/SKILL.md +391 -0
- package/bin/skills/verl/references/api-reference.md +301 -0
- package/bin/skills/verl/references/troubleshooting.md +391 -0
- package/bin/skills/vllm/SKILL.md +364 -0
- package/bin/skills/vllm/references/optimization.md +226 -0
- package/bin/skills/vllm/references/quantization.md +284 -0
- package/bin/skills/vllm/references/server-deployment.md +255 -0
- package/bin/skills/vllm/references/troubleshooting.md +447 -0
- package/bin/skills/weights-and-biases/SKILL.md +590 -0
- package/bin/skills/weights-and-biases/references/artifacts.md +584 -0
- package/bin/skills/weights-and-biases/references/integrations.md +700 -0
- package/bin/skills/weights-and-biases/references/sweeps.md +847 -0
- package/bin/skills/whisper/SKILL.md +317 -0
- package/bin/skills/whisper/references/languages.md +189 -0
- package/bin/synsc +0 -0
- package/package.json +10 -0
|
@@ -0,0 +1,108 @@
|
|
|
1
|
+
# LlamaIndex Data Connectors Guide
|
|
2
|
+
|
|
3
|
+
300+ data connectors via LlamaHub.
|
|
4
|
+
|
|
5
|
+
## Built-in loaders
|
|
6
|
+
|
|
7
|
+
### SimpleDirectoryReader
|
|
8
|
+
|
|
9
|
+
```python
|
|
10
|
+
from llama_index.core import SimpleDirectoryReader
|
|
11
|
+
|
|
12
|
+
# Load all files
|
|
13
|
+
documents = SimpleDirectoryReader("./data").load_data()
|
|
14
|
+
|
|
15
|
+
# Filter by extension
|
|
16
|
+
documents = SimpleDirectoryReader(
|
|
17
|
+
"./data",
|
|
18
|
+
required_exts=[".pdf", ".docx", ".txt"]
|
|
19
|
+
).load_data()
|
|
20
|
+
|
|
21
|
+
# Recursive
|
|
22
|
+
documents = SimpleDirectoryReader("./data", recursive=True).load_data()
|
|
23
|
+
```
|
|
24
|
+
|
|
25
|
+
### Web pages
|
|
26
|
+
|
|
27
|
+
```python
|
|
28
|
+
from llama_index.readers.web import SimpleWebPageReader, BeautifulSoupWebReader
|
|
29
|
+
|
|
30
|
+
# Simple loader
|
|
31
|
+
reader = SimpleWebPageReader()
|
|
32
|
+
documents = reader.load_data(["https://example.com"])
|
|
33
|
+
|
|
34
|
+
# Advanced (BeautifulSoup)
|
|
35
|
+
reader = BeautifulSoupWebReader()
|
|
36
|
+
documents = reader.load_data(urls=[
|
|
37
|
+
"https://docs.python.org",
|
|
38
|
+
"https://numpy.org"
|
|
39
|
+
])
|
|
40
|
+
```
|
|
41
|
+
|
|
42
|
+
### PDF
|
|
43
|
+
|
|
44
|
+
```python
|
|
45
|
+
from llama_index.readers.file import PDFReader
|
|
46
|
+
|
|
47
|
+
reader = PDFReader()
|
|
48
|
+
documents = reader.load_data("paper.pdf")
|
|
49
|
+
```
|
|
50
|
+
|
|
51
|
+
### GitHub
|
|
52
|
+
|
|
53
|
+
```python
|
|
54
|
+
from llama_index.readers.github import GithubRepositoryReader
|
|
55
|
+
|
|
56
|
+
reader = GithubRepositoryReader(
|
|
57
|
+
owner="facebook",
|
|
58
|
+
repo="react",
|
|
59
|
+
filter_file_extensions=[".js", ".jsx"],
|
|
60
|
+
verbose=True
|
|
61
|
+
)
|
|
62
|
+
|
|
63
|
+
documents = reader.load_data(branch="main")
|
|
64
|
+
```
|
|
65
|
+
|
|
66
|
+
## LlamaHub connectors
|
|
67
|
+
|
|
68
|
+
Visit https://llamahub.ai for 300+ connectors:
|
|
69
|
+
- Notion, Google Docs, Confluence
|
|
70
|
+
- Slack, Discord, Twitter
|
|
71
|
+
- PostgreSQL, MongoDB, MySQL
|
|
72
|
+
- S3, GCS, Azure Blob
|
|
73
|
+
- Stripe, Shopify, Salesforce
|
|
74
|
+
|
|
75
|
+
### Install from LlamaHub
|
|
76
|
+
|
|
77
|
+
```bash
|
|
78
|
+
pip install llama-index-readers-notion
|
|
79
|
+
```
|
|
80
|
+
|
|
81
|
+
```python
|
|
82
|
+
from llama_index.readers.notion import NotionPageReader
|
|
83
|
+
|
|
84
|
+
reader = NotionPageReader(integration_token="your-token")
|
|
85
|
+
documents = reader.load_data(page_ids=["page-id"])
|
|
86
|
+
```
|
|
87
|
+
|
|
88
|
+
## Custom loader
|
|
89
|
+
|
|
90
|
+
```python
|
|
91
|
+
from llama_index.core.readers.base import BaseReader
|
|
92
|
+
from llama_index.core import Document
|
|
93
|
+
|
|
94
|
+
class CustomReader(BaseReader):
|
|
95
|
+
def load_data(self, file_path: str):
|
|
96
|
+
# Your custom loading logic
|
|
97
|
+
with open(file_path) as f:
|
|
98
|
+
text = f.read()
|
|
99
|
+
return [Document(text=text, metadata={"source": file_path})]
|
|
100
|
+
|
|
101
|
+
reader = CustomReader()
|
|
102
|
+
documents = reader.load_data("data.txt")
|
|
103
|
+
```
|
|
104
|
+
|
|
105
|
+
## Resources
|
|
106
|
+
|
|
107
|
+
- **LlamaHub**: https://llamahub.ai
|
|
108
|
+
- **Data Connectors Docs**: https://developers.llamaindex.ai/python/framework/modules/data_connectors/
|
|
@@ -0,0 +1,406 @@
|
|
|
1
|
+
# LlamaIndex Query Engines Guide
|
|
2
|
+
|
|
3
|
+
Complete guide to query engines, modes, and customization.
|
|
4
|
+
|
|
5
|
+
## What are query engines?
|
|
6
|
+
|
|
7
|
+
Query engines power the retrieval and response generation in LlamaIndex:
|
|
8
|
+
1. Retrieve relevant chunks from index
|
|
9
|
+
2. Generate response using LLM + context
|
|
10
|
+
3. Return answer (optionally with sources)
|
|
11
|
+
|
|
12
|
+
## Basic query engine
|
|
13
|
+
|
|
14
|
+
```python
|
|
15
|
+
from llama_index.core import VectorStoreIndex
|
|
16
|
+
|
|
17
|
+
index = VectorStoreIndex.from_documents(documents)
|
|
18
|
+
|
|
19
|
+
# Default query engine
|
|
20
|
+
query_engine = index.as_query_engine()
|
|
21
|
+
response = query_engine.query("What is the main topic?")
|
|
22
|
+
print(response)
|
|
23
|
+
```
|
|
24
|
+
|
|
25
|
+
## Response modes
|
|
26
|
+
|
|
27
|
+
### 1. Compact (default) - Best for most cases
|
|
28
|
+
|
|
29
|
+
```python
|
|
30
|
+
query_engine = index.as_query_engine(
|
|
31
|
+
response_mode="compact"
|
|
32
|
+
)
|
|
33
|
+
|
|
34
|
+
# Combines chunks that fit in context window
|
|
35
|
+
response = query_engine.query("Explain quantum computing")
|
|
36
|
+
```
|
|
37
|
+
|
|
38
|
+
### 2. Tree summarize - Hierarchical summarization
|
|
39
|
+
|
|
40
|
+
```python
|
|
41
|
+
query_engine = index.as_query_engine(
|
|
42
|
+
response_mode="tree_summarize"
|
|
43
|
+
)
|
|
44
|
+
|
|
45
|
+
# Builds summary tree from chunks
|
|
46
|
+
# Best for: Summarization tasks, many retrieved chunks
|
|
47
|
+
response = query_engine.query("Summarize all the key findings")
|
|
48
|
+
```
|
|
49
|
+
|
|
50
|
+
### 3. Simple summarize - Concatenate and summarize
|
|
51
|
+
|
|
52
|
+
```python
|
|
53
|
+
query_engine = index.as_query_engine(
|
|
54
|
+
response_mode="simple_summarize"
|
|
55
|
+
)
|
|
56
|
+
|
|
57
|
+
# Concatenates all chunks, then summarizes
|
|
58
|
+
# Fast but may lose context if too many chunks
|
|
59
|
+
```
|
|
60
|
+
|
|
61
|
+
### 4. Refine - Iterative refinement
|
|
62
|
+
|
|
63
|
+
```python
|
|
64
|
+
query_engine = index.as_query_engine(
|
|
65
|
+
response_mode="refine"
|
|
66
|
+
)
|
|
67
|
+
|
|
68
|
+
# Refines answer iteratively across chunks
|
|
69
|
+
# Most thorough, slowest
|
|
70
|
+
# Best for: Complex questions requiring synthesis
|
|
71
|
+
```
|
|
72
|
+
|
|
73
|
+
### 5. No text - Return nodes only
|
|
74
|
+
|
|
75
|
+
```python
|
|
76
|
+
query_engine = index.as_query_engine(
|
|
77
|
+
response_mode="no_text"
|
|
78
|
+
)
|
|
79
|
+
|
|
80
|
+
# Returns retrieved nodes without LLM response
|
|
81
|
+
# Useful for: Debugging retrieval, custom processing
|
|
82
|
+
response = query_engine.query("machine learning")
|
|
83
|
+
for node in response.source_nodes:
|
|
84
|
+
print(node.text)
|
|
85
|
+
```
|
|
86
|
+
|
|
87
|
+
## Configuration options
|
|
88
|
+
|
|
89
|
+
### Similarity top-k
|
|
90
|
+
|
|
91
|
+
```python
|
|
92
|
+
# Return top 3 most similar chunks
|
|
93
|
+
query_engine = index.as_query_engine(
|
|
94
|
+
similarity_top_k=3 # Default: 2
|
|
95
|
+
)
|
|
96
|
+
```
|
|
97
|
+
|
|
98
|
+
### Streaming
|
|
99
|
+
|
|
100
|
+
```python
|
|
101
|
+
# Stream response tokens
|
|
102
|
+
query_engine = index.as_query_engine(streaming=True)
|
|
103
|
+
|
|
104
|
+
response = query_engine.query("Explain neural networks")
|
|
105
|
+
for text in response.response_gen:
|
|
106
|
+
print(text, end="", flush=True)
|
|
107
|
+
```
|
|
108
|
+
|
|
109
|
+
### Verbose mode
|
|
110
|
+
|
|
111
|
+
```python
|
|
112
|
+
# Show retrieval and generation process
|
|
113
|
+
query_engine = index.as_query_engine(verbose=True)
|
|
114
|
+
|
|
115
|
+
response = query_engine.query("What is Python?")
|
|
116
|
+
# Prints: Retrieved chunks, prompts, LLM calls
|
|
117
|
+
```
|
|
118
|
+
|
|
119
|
+
## Custom prompts
|
|
120
|
+
|
|
121
|
+
### Text QA template
|
|
122
|
+
|
|
123
|
+
```python
|
|
124
|
+
from llama_index.core import PromptTemplate
|
|
125
|
+
|
|
126
|
+
qa_prompt = PromptTemplate(
|
|
127
|
+
"Context information is below.\n"
|
|
128
|
+
"---------------------\n"
|
|
129
|
+
"{context_str}\n"
|
|
130
|
+
"---------------------\n"
|
|
131
|
+
"Given the context, answer: {query_str}\n"
|
|
132
|
+
"If the context doesn't contain the answer, say 'I don't know'.\n"
|
|
133
|
+
"Answer: "
|
|
134
|
+
)
|
|
135
|
+
|
|
136
|
+
query_engine = index.as_query_engine(text_qa_template=qa_prompt)
|
|
137
|
+
```
|
|
138
|
+
|
|
139
|
+
### Refine template
|
|
140
|
+
|
|
141
|
+
```python
|
|
142
|
+
refine_prompt = PromptTemplate(
|
|
143
|
+
"The original query is: {query_str}\n"
|
|
144
|
+
"We have an existing answer: {existing_answer}\n"
|
|
145
|
+
"We have new context: {context_msg}\n"
|
|
146
|
+
"Refine the answer based on new context. "
|
|
147
|
+
"If context isn't useful, return original answer.\n"
|
|
148
|
+
"Refined Answer: "
|
|
149
|
+
)
|
|
150
|
+
|
|
151
|
+
query_engine = index.as_query_engine(
|
|
152
|
+
response_mode="refine",
|
|
153
|
+
refine_template=refine_prompt
|
|
154
|
+
)
|
|
155
|
+
```
|
|
156
|
+
|
|
157
|
+
## Node postprocessors
|
|
158
|
+
|
|
159
|
+
### Metadata filtering
|
|
160
|
+
|
|
161
|
+
```python
|
|
162
|
+
from llama_index.core.postprocessor import MetadataReplacementPostProcessor
|
|
163
|
+
|
|
164
|
+
postprocessor = MetadataReplacementPostProcessor(
|
|
165
|
+
target_metadata_key="window" # Replace node content with window
|
|
166
|
+
)
|
|
167
|
+
|
|
168
|
+
query_engine = index.as_query_engine(
|
|
169
|
+
node_postprocessors=[postprocessor]
|
|
170
|
+
)
|
|
171
|
+
```
|
|
172
|
+
|
|
173
|
+
### Similarity cutoff
|
|
174
|
+
|
|
175
|
+
```python
|
|
176
|
+
from llama_index.core.postprocessor import SimilarityPostprocessor
|
|
177
|
+
|
|
178
|
+
# Filter nodes below similarity threshold
|
|
179
|
+
postprocessor = SimilarityPostprocessor(similarity_cutoff=0.7)
|
|
180
|
+
|
|
181
|
+
query_engine = index.as_query_engine(
|
|
182
|
+
node_postprocessors=[postprocessor]
|
|
183
|
+
)
|
|
184
|
+
```
|
|
185
|
+
|
|
186
|
+
### Reranking
|
|
187
|
+
|
|
188
|
+
```python
|
|
189
|
+
from llama_index.core.postprocessor import SentenceTransformerRerank
|
|
190
|
+
|
|
191
|
+
# Rerank retrieved nodes
|
|
192
|
+
reranker = SentenceTransformerRerank(
|
|
193
|
+
model="cross-encoder/ms-marco-MiniLM-L-2-v2",
|
|
194
|
+
top_n=3
|
|
195
|
+
)
|
|
196
|
+
|
|
197
|
+
query_engine = index.as_query_engine(
|
|
198
|
+
node_postprocessors=[reranker],
|
|
199
|
+
similarity_top_k=10 # Retrieve 10, rerank to 3
|
|
200
|
+
)
|
|
201
|
+
```
|
|
202
|
+
|
|
203
|
+
## Advanced query engines
|
|
204
|
+
|
|
205
|
+
### Sub-question query engine
|
|
206
|
+
|
|
207
|
+
```python
|
|
208
|
+
from llama_index.core.query_engine import SubQuestionQueryEngine
|
|
209
|
+
from llama_index.core.tools import QueryEngineTool
|
|
210
|
+
|
|
211
|
+
# Multiple indices for different topics
|
|
212
|
+
python_index = VectorStoreIndex.from_documents(python_docs)
|
|
213
|
+
numpy_index = VectorStoreIndex.from_documents(numpy_docs)
|
|
214
|
+
|
|
215
|
+
# Create tools
|
|
216
|
+
python_tool = QueryEngineTool.from_defaults(
|
|
217
|
+
query_engine=python_index.as_query_engine(),
|
|
218
|
+
description="Useful for Python programming questions"
|
|
219
|
+
)
|
|
220
|
+
numpy_tool = QueryEngineTool.from_defaults(
|
|
221
|
+
query_engine=numpy_index.as_query_engine(),
|
|
222
|
+
description="Useful for NumPy array questions"
|
|
223
|
+
)
|
|
224
|
+
|
|
225
|
+
# Sub-question engine decomposes complex queries
|
|
226
|
+
query_engine = SubQuestionQueryEngine.from_defaults(
|
|
227
|
+
query_engine_tools=[python_tool, numpy_tool]
|
|
228
|
+
)
|
|
229
|
+
|
|
230
|
+
# "How do I create numpy arrays in Python?" becomes:
|
|
231
|
+
# 1. Query numpy_tool about array creation
|
|
232
|
+
# 2. Query python_tool about syntax
|
|
233
|
+
# 3. Synthesize answers
|
|
234
|
+
response = query_engine.query("How do I create numpy arrays in Python?")
|
|
235
|
+
```
|
|
236
|
+
|
|
237
|
+
### Router query engine
|
|
238
|
+
|
|
239
|
+
```python
|
|
240
|
+
from llama_index.core.query_engine import RouterQueryEngine
|
|
241
|
+
from llama_index.core.selectors import LLMSingleSelector
|
|
242
|
+
|
|
243
|
+
# Route to appropriate index based on query
|
|
244
|
+
selector = LLMSingleSelector.from_defaults()
|
|
245
|
+
|
|
246
|
+
query_engine = RouterQueryEngine(
|
|
247
|
+
selector=selector,
|
|
248
|
+
query_engine_tools=[python_tool, numpy_tool]
|
|
249
|
+
)
|
|
250
|
+
|
|
251
|
+
# Automatically routes to correct index
|
|
252
|
+
response = query_engine.query("What is Python?") # Routes to python_tool
|
|
253
|
+
response = query_engine.query("NumPy broadcasting?") # Routes to numpy_tool
|
|
254
|
+
```
|
|
255
|
+
|
|
256
|
+
### Transform query engine
|
|
257
|
+
|
|
258
|
+
```python
|
|
259
|
+
from llama_index.core.query_engine import TransformQueryEngine
|
|
260
|
+
from llama_index.core.query_transforms import HyDEQueryTransform
|
|
261
|
+
|
|
262
|
+
# HyDE: Generate hypothetical document before retrieval
|
|
263
|
+
hyde_transform = HyDEQueryTransform(include_original=True)
|
|
264
|
+
|
|
265
|
+
query_engine = TransformQueryEngine(
|
|
266
|
+
query_engine=base_query_engine,
|
|
267
|
+
query_transform=hyde_transform
|
|
268
|
+
)
|
|
269
|
+
|
|
270
|
+
# Improves retrieval quality
|
|
271
|
+
response = query_engine.query("What are the benefits of Python?")
|
|
272
|
+
```
|
|
273
|
+
|
|
274
|
+
## Chat engine (conversational)
|
|
275
|
+
|
|
276
|
+
### Basic chat engine
|
|
277
|
+
|
|
278
|
+
```python
|
|
279
|
+
# Chat engine with memory
|
|
280
|
+
chat_engine = index.as_chat_engine(
|
|
281
|
+
chat_mode="condense_plus_context"
|
|
282
|
+
)
|
|
283
|
+
|
|
284
|
+
# Multi-turn conversation
|
|
285
|
+
response1 = chat_engine.chat("What is Python?")
|
|
286
|
+
response2 = chat_engine.chat("What are its main features?") # Remembers context
|
|
287
|
+
response3 = chat_engine.chat("Can you give examples?")
|
|
288
|
+
```
|
|
289
|
+
|
|
290
|
+
### Chat modes
|
|
291
|
+
|
|
292
|
+
```python
|
|
293
|
+
# 1. condense_plus_context (recommended)
|
|
294
|
+
chat_engine = index.as_chat_engine(chat_mode="condense_plus_context")
|
|
295
|
+
# Condenses chat history + retrieves relevant context
|
|
296
|
+
|
|
297
|
+
# 2. context - Simple RAG
|
|
298
|
+
chat_engine = index.as_chat_engine(chat_mode="context")
|
|
299
|
+
# Retrieves context for each query
|
|
300
|
+
|
|
301
|
+
# 3. react - Agent-based
|
|
302
|
+
chat_engine = index.as_chat_engine(chat_mode="react")
|
|
303
|
+
# Uses ReAct agent pattern with tools
|
|
304
|
+
|
|
305
|
+
# 4. best - Automatically selects best mode
|
|
306
|
+
chat_engine = index.as_chat_engine(chat_mode="best")
|
|
307
|
+
```
|
|
308
|
+
|
|
309
|
+
### Reset conversation
|
|
310
|
+
|
|
311
|
+
```python
|
|
312
|
+
# Clear chat history
|
|
313
|
+
chat_engine.reset()
|
|
314
|
+
|
|
315
|
+
# Start new conversation
|
|
316
|
+
response = chat_engine.chat("New topic: what is machine learning?")
|
|
317
|
+
```
|
|
318
|
+
|
|
319
|
+
## Structured output
|
|
320
|
+
|
|
321
|
+
### Pydantic models
|
|
322
|
+
|
|
323
|
+
```python
|
|
324
|
+
from pydantic import BaseModel
|
|
325
|
+
from llama_index.core.output_parsers import PydanticOutputParser
|
|
326
|
+
|
|
327
|
+
class Summary(BaseModel):
|
|
328
|
+
title: str
|
|
329
|
+
main_points: list[str]
|
|
330
|
+
category: str
|
|
331
|
+
|
|
332
|
+
output_parser = PydanticOutputParser(output_cls=Summary)
|
|
333
|
+
|
|
334
|
+
query_engine = index.as_query_engine(
|
|
335
|
+
output_parser=output_parser
|
|
336
|
+
)
|
|
337
|
+
|
|
338
|
+
response = query_engine.query("Summarize the document")
|
|
339
|
+
# response is a Pydantic model
|
|
340
|
+
print(response.title, response.main_points)
|
|
341
|
+
```
|
|
342
|
+
|
|
343
|
+
## Source tracking
|
|
344
|
+
|
|
345
|
+
### Get source nodes
|
|
346
|
+
|
|
347
|
+
```python
|
|
348
|
+
query_engine = index.as_query_engine()
|
|
349
|
+
|
|
350
|
+
response = query_engine.query("What is Python?")
|
|
351
|
+
|
|
352
|
+
# Access source nodes
|
|
353
|
+
for node in response.source_nodes:
|
|
354
|
+
print(f"Text: {node.text}")
|
|
355
|
+
print(f"Score: {node.score}")
|
|
356
|
+
print(f"Metadata: {node.metadata}")
|
|
357
|
+
```
|
|
358
|
+
|
|
359
|
+
## Best practices
|
|
360
|
+
|
|
361
|
+
1. **Use compact mode for most cases** - Good balance
|
|
362
|
+
2. **Set similarity_top_k appropriately** - 2-5 usually optimal
|
|
363
|
+
3. **Enable streaming for long responses** - Better UX
|
|
364
|
+
4. **Add postprocessors for quality** - Reranking improves results
|
|
365
|
+
5. **Use chat engine for conversations** - Built-in memory
|
|
366
|
+
6. **Track source nodes** - Cite sources to users
|
|
367
|
+
7. **Custom prompts for domain** - Better responses
|
|
368
|
+
8. **Test different response modes** - Pick best for use case
|
|
369
|
+
9. **Monitor token usage** - Retrieval + generation costs
|
|
370
|
+
10. **Cache query engines** - Don't recreate each time
|
|
371
|
+
|
|
372
|
+
## Performance tips
|
|
373
|
+
|
|
374
|
+
### Caching
|
|
375
|
+
|
|
376
|
+
```python
|
|
377
|
+
from llama_index.core.storage.chat_store import SimpleChatStore
|
|
378
|
+
|
|
379
|
+
# Cache chat history
|
|
380
|
+
chat_store = SimpleChatStore()
|
|
381
|
+
chat_engine = index.as_chat_engine(
|
|
382
|
+
chat_mode="condense_plus_context",
|
|
383
|
+
chat_store=chat_store
|
|
384
|
+
)
|
|
385
|
+
```
|
|
386
|
+
|
|
387
|
+
### Async queries
|
|
388
|
+
|
|
389
|
+
```python
|
|
390
|
+
import asyncio
|
|
391
|
+
|
|
392
|
+
# Async query for concurrent requests
|
|
393
|
+
response = await query_engine.aquery("What is Python?")
|
|
394
|
+
|
|
395
|
+
# Multiple concurrent queries
|
|
396
|
+
responses = await asyncio.gather(
|
|
397
|
+
query_engine.aquery("What is Python?"),
|
|
398
|
+
query_engine.aquery("What is Java?")
|
|
399
|
+
)
|
|
400
|
+
```
|
|
401
|
+
|
|
402
|
+
## Resources
|
|
403
|
+
|
|
404
|
+
- **Query Engines Docs**: https://developers.llamaindex.ai/python/framework/modules/querying/
|
|
405
|
+
- **Response Modes**: https://developers.llamaindex.ai/python/framework/modules/querying/response_modes/
|
|
406
|
+
- **Chat Engines**: https://developers.llamaindex.ai/python/framework/modules/chat/
|