sglang 0.4.3__tar.gz → 0.4.3.post2__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {sglang-0.4.3/sglang.egg-info → sglang-0.4.3.post2}/PKG-INFO +3 -3
- {sglang-0.4.3 → sglang-0.4.3.post2}/pyproject.toml +3 -3
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/lang/backend/openai.py +5 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/lang/chat_template.py +22 -7
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/lang/ir.py +1 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/configs/__init__.py +6 -3
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/configs/model_config.py +2 -0
- sglang-0.4.3.post2/sglang/srt/configs/qwen2_5_vl_config.py +1003 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/entrypoints/engine.py +17 -2
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/hf_transformers_utils.py +2 -3
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/attention/flashinfer_backend.py +101 -30
- sglang-0.4.3.post2/sglang/srt/layers/moe/fused_moe_triton/configs/E=256,N=128,device_name=NVIDIA_L20Y,dtype=fp8_w8a8,block_shape=[128, 128].json +146 -0
- sglang-0.4.3.post2/sglang/srt/layers/quantization/configs/N=1536,K=1536,device_name=NVIDIA_L20Y,dtype=fp8_w8a8,block_shape=[128, 128].json +26 -0
- sglang-0.4.3.post2/sglang/srt/layers/quantization/configs/N=1536,K=7168,device_name=NVIDIA_L20Y,dtype=fp8_w8a8,block_shape=[128, 128].json +26 -0
- sglang-0.4.3.post2/sglang/srt/layers/quantization/configs/N=2048,K=512,device_name=NVIDIA_L20Y,dtype=fp8_w8a8,block_shape=[128, 128].json +26 -0
- sglang-0.4.3.post2/sglang/srt/layers/quantization/configs/N=2304,K=7168,device_name=NVIDIA_L20Y,dtype=fp8_w8a8,block_shape=[128, 128].json +26 -0
- sglang-0.4.3.post2/sglang/srt/layers/quantization/configs/N=24576,K=7168,device_name=NVIDIA_L20Y,dtype=fp8_w8a8,block_shape=[128, 128].json +26 -0
- sglang-0.4.3.post2/sglang/srt/layers/quantization/configs/N=256,K=7168,device_name=NVIDIA_L20Y,dtype=fp8_w8a8,block_shape=[128, 128].json +26 -0
- sglang-0.4.3.post2/sglang/srt/layers/quantization/configs/N=32768,K=512,device_name=NVIDIA_L20Y,dtype=fp8_w8a8,block_shape=[128, 128].json +26 -0
- sglang-0.4.3.post2/sglang/srt/layers/quantization/configs/N=576,K=7168,device_name=NVIDIA_L20Y,dtype=fp8_w8a8,block_shape=[128, 128].json +26 -0
- sglang-0.4.3.post2/sglang/srt/layers/quantization/configs/N=7168,K=1024,device_name=NVIDIA_L20Y,dtype=fp8_w8a8,block_shape=[128, 128].json +26 -0
- sglang-0.4.3.post2/sglang/srt/layers/quantization/configs/N=7168,K=1152,device_name=NVIDIA_L20Y,dtype=fp8_w8a8,block_shape=[128, 128].json +26 -0
- sglang-0.4.3.post2/sglang/srt/layers/quantization/configs/N=7168,K=128,device_name=NVIDIA_L20Y,dtype=fp8_w8a8,block_shape=[128, 128].json +26 -0
- sglang-0.4.3.post2/sglang/srt/layers/quantization/configs/N=7168,K=16384,device_name=NVIDIA_L20Y,dtype=fp8_w8a8,block_shape=[128, 128].json +26 -0
- sglang-0.4.3.post2/sglang/srt/layers/quantization/configs/N=7168,K=18432,device_name=NVIDIA_L20Y,dtype=fp8_w8a8,block_shape=[128, 128].json +26 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/managers/image_processor.py +217 -122
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/managers/schedule_batch.py +1 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/model_executor/forward_batch_info.py +4 -1
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/model_executor/model_runner.py +1 -0
- sglang-0.4.3.post2/sglang/srt/models/deepseek_nextn.py +295 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/deepseek_v2.py +9 -3
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/llava.py +2 -1
- sglang-0.4.3.post2/sglang/srt/models/qwen2_5_vl.py +722 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/qwen2_vl.py +2 -1
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/openai_api/adapter.py +17 -3
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/server_args.py +6 -3
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/speculative/eagle_worker.py +7 -2
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/speculative/spec_info.py +11 -1
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/utils.py +99 -19
- sglang-0.4.3.post2/sglang/version.py +1 -0
- {sglang-0.4.3 → sglang-0.4.3.post2/sglang.egg-info}/PKG-INFO +3 -3
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang.egg-info/SOURCES.txt +17 -1
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang.egg-info/requires.txt +2 -2
- sglang-0.4.3/sglang/srt/configs/qwen2vl.py +0 -130
- sglang-0.4.3/sglang/version.py +0 -1
- {sglang-0.4.3 → sglang-0.4.3.post2}/LICENSE +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/README.md +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/setup.cfg +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/__init__.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/api.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/bench_latency.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/bench_offline_throughput.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/bench_one_batch.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/bench_one_batch_server.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/bench_serving.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/check_env.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/global_config.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/lang/__init__.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/lang/backend/__init__.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/lang/backend/anthropic.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/lang/backend/base_backend.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/lang/backend/litellm.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/lang/backend/runtime_endpoint.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/lang/backend/vertexai.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/lang/choices.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/lang/compiler.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/lang/interpreter.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/lang/tracer.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/launch_server.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/llama3_eval.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/_custom_ops.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/aio_rwlock.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/configs/chatglm.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/configs/dbrx.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/configs/device_config.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/configs/exaone.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/configs/load_config.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/constrained/base_grammar_backend.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/constrained/outlines_backend.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/constrained/outlines_jump_forward.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/constrained/xgrammar_backend.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/conversation.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/custom_op.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/distributed/__init__.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/distributed/communication_op.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/distributed/device_communicators/cuda_wrapper.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/distributed/device_communicators/custom_all_reduce.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/distributed/device_communicators/custom_all_reduce_utils.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/distributed/device_communicators/hpu_communicator.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/distributed/device_communicators/pynccl.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/distributed/device_communicators/pynccl_wrapper.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/distributed/device_communicators/shm_broadcast.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/distributed/device_communicators/xpu_communicator.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/distributed/parallel_state.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/distributed/utils.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/entrypoints/http_server.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/function_call_parser.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/activation.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/attention/__init__.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/attention/double_sparsity_backend.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/attention/torch_native_backend.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/attention/triton_backend.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/attention/triton_ops/decode_attention.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/attention/triton_ops/double_sparsity_attention.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/attention/triton_ops/extend_attention.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/attention/triton_ops/prefill_attention.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/attention/vision.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/dp_attention.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/layernorm.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/linear.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/logits_processor.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/ep_moe/__init__.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/ep_moe/kernels.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/ep_moe/layer.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_native.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/__init__.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=1,N=14336,device_name=NVIDIA_A100-SXM4-80GB,dtype=int8_w8a16.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=1,N=14336,device_name=NVIDIA_A100-SXM4-80GB.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=1,N=1792,device_name=NVIDIA_A100-SXM4-80GB,dtype=int8_w8a16.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=1,N=1792,device_name=NVIDIA_A100-SXM4-80GB.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=1,N=3072,device_name=NVIDIA_A100-SXM4-80GB,dtype=int8_w8a16.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=1,N=3072,device_name=NVIDIA_H100_80GB_HBM3,dtype=int8_w8a16.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=1,N=3072,device_name=NVIDIA_H100_80GB_HBM3.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=1,N=3584,device_name=NVIDIA_A100-SXM4-80GB,dtype=int8_w8a16.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=1,N=3584,device_name=NVIDIA_A100-SXM4-80GB.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=1,N=7168,device_name=NVIDIA_A100-SXM4-80GB,dtype=int8_w8a16.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=1,N=7168,device_name=NVIDIA_A100-SXM4-80GB.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=16,N=1344,device_name=NVIDIA_A100-SXM4-40GB.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=16,N=1344,device_name=NVIDIA_A100-SXM4-80GB.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=16,N=1344,device_name=NVIDIA_H100_80GB_HBM3.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=16,N=14336,device_name=NVIDIA_A100-SXM4-80GB,dtype=int8_w8a16.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=16,N=14336,device_name=NVIDIA_A100-SXM4-80GB.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=16,N=1792,device_name=NVIDIA_A100-SXM4-80GB,dtype=int8_w8a16.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=16,N=1792,device_name=NVIDIA_A100-SXM4-80GB.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=16,N=2688,device_name=NVIDIA_A100-SXM4-80GB.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=16,N=2688,device_name=NVIDIA_H100_80GB_HBM3.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=16,N=3072,device_name=NVIDIA_A100-SXM4-80GB,dtype=int8_w8a16.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=16,N=3072,device_name=NVIDIA_H100_80GB_HBM3,dtype=int8_w8a16.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=16,N=3200,device_name=NVIDIA_H100_80GB_HBM3,dtype=fp8_w8a8.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=16,N=3584,device_name=NVIDIA_A100-SXM4-80GB,dtype=int8_w8a16.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=16,N=3584,device_name=NVIDIA_A100-SXM4-80GB.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=16,N=6400,device_name=NVIDIA_H100_80GB_HBM3,dtype=fp8_w8a8.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=16,N=7168,device_name=NVIDIA_A100-SXM4-80GB,dtype=int8_w8a16.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=16,N=7168,device_name=NVIDIA_A100-SXM4-80GB.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=16,N=7168,device_name=NVIDIA_H100_80GB_HBM3,dtype=int8_w8a16.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=16,N=800,device_name=NVIDIA_H100_80GB_HBM3,dtype=fp8_w8a8.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=256,N=128,device_name=NVIDIA_H100_80GB_HBM3,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=256,N=256,device_name=AMD_Instinct_MI300X,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=256,N=256,device_name=AMD_Instinct_MI325X,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=256,N=256,device_name=AMD_Radeon_Graphics,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=256,N=256,device_name=NVIDIA_B200,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=256,N=256,device_name=NVIDIA_H200,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=64,N=1280,device_name=NVIDIA_A100-SXM4-80GB.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=64,N=1280,device_name=NVIDIA_A800-SXM4-80GB.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=64,N=1280,device_name=NVIDIA_H100_80GB_HBM3,dtype=fp8_w8a8.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=64,N=1280,device_name=NVIDIA_H100_80GB_HBM3.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=64,N=1280,device_name=NVIDIA_H200,dtype=fp8_w8a8.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=64,N=1280,device_name=NVIDIA_H200.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=64,N=2560,device_name=NVIDIA_H100_80GB_HBM3,dtype=fp8_w8a8.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=64,N=2560,device_name=NVIDIA_H200,dtype=fp8_w8a8.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=64,N=2560,device_name=NVIDIA_H200.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=64,N=320,device_name=NVIDIA_H100_80GB_HBM3,dtype=fp8_w8a8.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=64,N=320,device_name=NVIDIA_H100_80GB_HBM3.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=64,N=320,device_name=NVIDIA_H200,dtype=fp8_w8a8.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=64,N=320,device_name=NVIDIA_H200.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=64,N=640,device_name=NVIDIA_A100-SXM4-80GB.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=64,N=640,device_name=NVIDIA_A800-SXM4-80GB.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=64,N=640,device_name=NVIDIA_GeForce_RTX_4090,dtype=fp8_w8a8.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=64,N=640,device_name=NVIDIA_H100_80GB_HBM3,dtype=fp8_w8a8.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=64,N=640,device_name=NVIDIA_H100_80GB_HBM3.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=64,N=640,device_name=NVIDIA_H200,dtype=fp8_w8a8.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=64,N=640,device_name=NVIDIA_H200.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=14336,device_name=AMD_Instinct_MI300X.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=14336,device_name=AMD_Instinct_MI325X.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=14336,device_name=AMD_Radeon_Graphics.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=14336,device_name=NVIDIA_H100_80GB_HBM3,dtype=fp8_w8a8.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=14336,device_name=NVIDIA_H200,dtype=fp8_w8a8.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=14336,device_name=NVIDIA_H200.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=1792,device_name=AMD_Instinct_MI300X.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=1792,device_name=AMD_Instinct_MI325X.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=1792,device_name=AMD_Radeon_Graphics.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=1792,device_name=NVIDIA_A100-SXM4-40GB.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=1792,device_name=NVIDIA_A100-SXM4-80GB.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=1792,device_name=NVIDIA_H100_80GB_HBM3.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=1792,device_name=NVIDIA_H200,dtype=fp8_w8a8.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=1792,device_name=NVIDIA_H200.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=2048,device_name=NVIDIA_A100-SXM4-80GB.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=2048,device_name=NVIDIA_H100_80GB_HBM3,dtype=fp8_w8a8.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=2048,device_name=NVIDIA_H100_80GB_HBM3.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=2048,device_name=NVIDIA_H200,dtype=fp8_w8a8.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=2048,device_name=NVIDIA_H200.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=3584,device_name=AMD_Instinct_MI300X.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=3584,device_name=AMD_Instinct_MI325X.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=3584,device_name=AMD_Radeon_Graphics.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=3584,device_name=NVIDIA_A100-SXM4-40GB.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=3584,device_name=NVIDIA_A100-SXM4-80GB.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=3584,device_name=NVIDIA_GeForce_RTX_4090,dtype=fp8_w8a8.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=3584,device_name=NVIDIA_H100_80GB_HBM3,dtype=fp8_w8a8.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=3584,device_name=NVIDIA_H100_80GB_HBM3.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=3584,device_name=NVIDIA_H200,dtype=fp8_w8a8.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=3584,device_name=NVIDIA_H200.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=3584,device_name=NVIDIA_L40S.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=4096,device_name=AMD_Instinct_MI300X,dtype=fp8_w8a8.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=4096,device_name=AMD_Instinct_MI325X,dtype=fp8_w8a8.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=4096,device_name=AMD_Radeon_Graphics,dtype=fp8_w8a8.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=4096,device_name=NVIDIA_A100-SXM4-80GB.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=4096,device_name=NVIDIA_H100_80GB_HBM3,dtype=fp8_w8a8.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=4096,device_name=NVIDIA_H100_80GB_HBM3.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=4096,device_name=NVIDIA_H200,dtype=fp8_w8a8.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=4096,device_name=NVIDIA_H200.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=7168,device_name=AMD_Instinct_MI300X.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=7168,device_name=AMD_Instinct_MI325X.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=7168,device_name=AMD_Radeon_Graphics.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=7168,device_name=NVIDIA_A100-SXM4-80GB.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=7168,device_name=NVIDIA_H100_80GB_HBM3,dtype=fp8_w8a8.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=7168,device_name=NVIDIA_H100_80GB_HBM3.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=7168,device_name=NVIDIA_H200,dtype=fp8_w8a8.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=7168,device_name=NVIDIA_H200.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=8192,device_name=AMD_Instinct_MI300X,dtype=fp8_w8a8.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=8192,device_name=AMD_Instinct_MI325X,dtype=fp8_w8a8.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=8192,device_name=AMD_Radeon_Graphics,dtype=fp8_w8a8.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=8192,device_name=NVIDIA_H100_80GB_HBM3,dtype=fp8_w8a8.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/configs/E=8,N=8192,device_name=NVIDIA_H200,dtype=fp8_w8a8.json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/fused_moe.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/fused_moe_triton/layer.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/moe/topk.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/parameter.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/pooler.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/__init__.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/base_config.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=1536,K=1536,device_name=NVIDIA_H100_80GB_HBM3,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=1536,K=1536,device_name=NVIDIA_H20,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=1536,K=7168,device_name=AMD_Instinct_MI300X,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=1536,K=7168,device_name=AMD_Instinct_MI325X,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=1536,K=7168,device_name=AMD_Radeon_Graphics,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=1536,K=7168,device_name=NVIDIA_H100_80GB_HBM3,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=1536,K=7168,device_name=NVIDIA_H20,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=1536,K=7168,device_name=NVIDIA_H200,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=2048,K=512,device_name=NVIDIA_H100_80GB_HBM3,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=2048,K=512,device_name=NVIDIA_H20,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=2048,K=512,device_name=NVIDIA_H200,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=2304,K=7168,device_name=NVIDIA_H100_80GB_HBM3,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=2304,K=7168,device_name=NVIDIA_H20,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=2304,K=7168,device_name=NVIDIA_H200,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=24576,K=7168,device_name=NVIDIA_B200,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=24576,K=7168,device_name=NVIDIA_H100_80GB_HBM3,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=24576,K=7168,device_name=NVIDIA_H20,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=24576,K=7168,device_name=NVIDIA_H200,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=256,K=7168,device_name=NVIDIA_H100_80GB_HBM3,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=256,K=7168,device_name=NVIDIA_H20,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=3072,K=1536,device_name=AMD_Instinct_MI300X,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=3072,K=1536,device_name=AMD_Instinct_MI325X,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=3072,K=1536,device_name=AMD_Radeon_Graphics,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=3072,K=1536,device_name=NVIDIA_B200,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=3072,K=1536,device_name=NVIDIA_H200,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=3072,K=7168,device_name=NVIDIA_B200,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=3072,K=7168,device_name=NVIDIA_H100_80GB_HBM3,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=3072,K=7168,device_name=NVIDIA_H200,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=32768,K=512,device_name=NVIDIA_B200,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=32768,K=512,device_name=NVIDIA_H100_80GB_HBM3,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=32768,K=512,device_name=NVIDIA_H20,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=32768,K=512,device_name=NVIDIA_H200,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=36864,K=7168,device_name=NVIDIA_H100_80GB_HBM3,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=36864,K=7168,device_name=NVIDIA_H200,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=4096,K=512,device_name=AMD_Instinct_MI300X,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=4096,K=512,device_name=AMD_Instinct_MI325X,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=4096,K=512,device_name=AMD_Radeon_Graphics,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=4096,K=512,device_name=NVIDIA_B200,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=4096,K=512,device_name=NVIDIA_H100_80GB_HBM3,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=4096,K=512,device_name=NVIDIA_H200,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=4608,K=7168,device_name=AMD_Instinct_MI300X,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=4608,K=7168,device_name=AMD_Instinct_MI325X,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=4608,K=7168,device_name=AMD_Radeon_Graphics,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=4608,K=7168,device_name=NVIDIA_B200,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=4608,K=7168,device_name=NVIDIA_H100_80GB_HBM3,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=4608,K=7168,device_name=NVIDIA_H200,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=512,K=7168,device_name=AMD_Instinct_MI300X,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=512,K=7168,device_name=AMD_Instinct_MI325X,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=512,K=7168,device_name=AMD_Radeon_Graphics,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=512,K=7168,device_name=NVIDIA_B200,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=512,K=7168,device_name=NVIDIA_H200,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=576,K=7168,device_name=AMD_Instinct_MI300X,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=576,K=7168,device_name=AMD_Instinct_MI325X,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=576,K=7168,device_name=AMD_Radeon_Graphics,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=576,K=7168,device_name=NVIDIA_B200,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=576,K=7168,device_name=NVIDIA_H100_80GB_HBM3,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=576,K=7168,device_name=NVIDIA_H20,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=576,K=7168,device_name=NVIDIA_H200,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=7168,K=1024,device_name=NVIDIA_H100_80GB_HBM3,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=7168,K=1024,device_name=NVIDIA_H20,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=7168,K=1024,device_name=NVIDIA_H200,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=7168,K=1152,device_name=NVIDIA_H100_80GB_HBM3,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=7168,K=1152,device_name=NVIDIA_H20,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=7168,K=1152,device_name=NVIDIA_H200,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=7168,K=128,device_name=NVIDIA_H100_80GB_HBM3,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=7168,K=128,device_name=NVIDIA_H20,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=7168,K=16384,device_name=NVIDIA_B200,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=7168,K=16384,device_name=NVIDIA_H100_80GB_HBM3,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=7168,K=16384,device_name=NVIDIA_H20,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=7168,K=16384,device_name=NVIDIA_H200,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=7168,K=18432,device_name=NVIDIA_B200,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=7168,K=18432,device_name=NVIDIA_H100_80GB_HBM3,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=7168,K=18432,device_name=NVIDIA_H20,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=7168,K=18432,device_name=NVIDIA_H200,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=7168,K=2048,device_name=AMD_Instinct_MI300X,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=7168,K=2048,device_name=AMD_Instinct_MI325X,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=7168,K=2048,device_name=AMD_Radeon_Graphics,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=7168,K=2048,device_name=NVIDIA_B200,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=7168,K=2048,device_name=NVIDIA_H100_80GB_HBM3,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=7168,K=2048,device_name=NVIDIA_H200,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=7168,K=2304,device_name=AMD_Instinct_MI300X,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=7168,K=2304,device_name=AMD_Instinct_MI325X,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=7168,K=2304,device_name=AMD_Radeon_Graphics,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=7168,K=2304,device_name=NVIDIA_B200,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=7168,K=2304,device_name=NVIDIA_H100_80GB_HBM3,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=7168,K=2304,device_name=NVIDIA_H200,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=7168,K=256,device_name=AMD_Instinct_MI300X,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=7168,K=256,device_name=AMD_Instinct_MI325X,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=7168,K=256,device_name=AMD_Radeon_Graphics,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=7168,K=256,device_name=NVIDIA_B200,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/configs/N=7168,K=256,device_name=NVIDIA_H200,dtype=fp8_w8a8,block_shape=[128, 128].json +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/fp8.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/fp8_kernel.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/fp8_utils.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/int8_kernel.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/modelopt_quant.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/quantization/w8a8_int8.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/radix_attention.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/rotary_embedding.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/sampler.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/torchao_utils.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/layers/vocab_parallel_embedding.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/lora/backend/__init__.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/lora/backend/base_backend.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/lora/backend/flashinfer_backend.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/lora/backend/triton_backend.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/lora/layers.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/lora/lora.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/lora/lora_config.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/lora/lora_manager.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/lora/mem_pool.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/lora/triton_ops/__init__.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/lora/triton_ops/gate_up_lora_b.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/lora/triton_ops/qkv_lora_b.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/lora/triton_ops/sgemm_lora_a.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/lora/triton_ops/sgemm_lora_b.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/lora/utils.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/managers/cache_controller.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/managers/configure_logging.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/managers/data_parallel_controller.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/managers/detokenizer_manager.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/managers/io_struct.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/managers/schedule_policy.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/managers/scheduler.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/managers/session_controller.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/managers/tokenizer_manager.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/managers/tp_worker.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/managers/tp_worker_overlap_thread.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/managers/utils.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/mem_cache/base_prefix_cache.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/mem_cache/chunk_cache.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/mem_cache/flush_cache.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/mem_cache/memory_pool.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/mem_cache/radix_cache.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/metrics/collector.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/metrics/func_timer.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/mm_utils.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/model_executor/cuda_graph_runner.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/model_loader/__init__.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/model_loader/loader.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/model_loader/utils.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/model_loader/weight_utils.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/model_parallel.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/baichuan.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/chatglm.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/commandr.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/dbrx.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/deepseek.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/exaone.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/gemma.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/gemma2.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/gemma2_reward.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/gpt2.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/gpt_bigcode.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/granite.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/grok.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/internlm2.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/internlm2_reward.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/llama.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/llama_classification.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/llama_eagle.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/llama_embedding.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/llama_reward.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/llavavid.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/minicpm.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/minicpm3.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/minicpmv.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/mistral.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/mixtral.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/mixtral_quant.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/mllama.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/olmo.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/olmo2.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/olmoe.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/phi3_small.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/qwen.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/qwen2.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/qwen2_eagle.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/qwen2_moe.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/registry.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/stablelm.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/torch_native_llama.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/xverse.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/xverse_moe.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/models/yivl.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/openai_api/protocol.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/sampling/custom_logit_processor.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/sampling/penaltylib/__init__.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/sampling/penaltylib/orchestrator.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/sampling/penaltylib/penalizers/frequency_penalty.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/sampling/penaltylib/penalizers/min_new_tokens.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/sampling/penaltylib/penalizers/presence_penalty.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/sampling/penaltylib/penalizers/repetition_penalty.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/sampling/sampling_batch_info.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/sampling/sampling_params.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/server.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/speculative/build_eagle_tree.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/speculative/eagle_draft_cuda_graph_runner.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/speculative/eagle_utils.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/torch_memory_saver_adapter.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/srt/utils.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/test/few_shot_gsm8k.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/test/few_shot_gsm8k_engine.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/test/run_eval.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/test/runners.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/test/simple_eval_common.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/test/simple_eval_gpqa.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/test/simple_eval_humaneval.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/test/simple_eval_math.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/test/simple_eval_mgsm.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/test/simple_eval_mmlu.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/test/srt/sampling/penaltylib/utils.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/test/test_activation.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/test/test_block_fp8.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/test/test_layernorm.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/test/test_programs.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang/test/test_utils.py +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang.egg-info/dependency_links.txt +0 -0
- {sglang-0.4.3 → sglang-0.4.3.post2}/sglang.egg-info/top_level.txt +0 -0
@@ -1,6 +1,6 @@
|
|
1
1
|
Metadata-Version: 2.2
|
2
2
|
Name: sglang
|
3
|
-
Version: 0.4.3
|
3
|
+
Version: 0.4.3.post2
|
4
4
|
Summary: SGLang is yet another fast serving framework for large language models and vision language models.
|
5
5
|
License: Apache License
|
6
6
|
Version 2.0, January 2004
|
@@ -235,7 +235,7 @@ Requires-Dist: pyzmq>=25.1.2; extra == "runtime-common"
|
|
235
235
|
Requires-Dist: torchao>=0.7.0; extra == "runtime-common"
|
236
236
|
Requires-Dist: uvicorn; extra == "runtime-common"
|
237
237
|
Requires-Dist: uvloop; extra == "runtime-common"
|
238
|
-
Requires-Dist: xgrammar
|
238
|
+
Requires-Dist: xgrammar==0.1.10; extra == "runtime-common"
|
239
239
|
Requires-Dist: ninja; extra == "runtime-common"
|
240
240
|
Provides-Extra: srt
|
241
241
|
Requires-Dist: sglang[runtime_common]; extra == "srt"
|
@@ -243,7 +243,7 @@ Requires-Dist: cuda-python; extra == "srt"
|
|
243
243
|
Requires-Dist: sgl-kernel>=0.0.3.post6; extra == "srt"
|
244
244
|
Requires-Dist: torch; extra == "srt"
|
245
245
|
Requires-Dist: vllm<=0.7.2,>=0.6.4.post1; extra == "srt"
|
246
|
-
Requires-Dist: flashinfer_python>=0.2.1.
|
246
|
+
Requires-Dist: flashinfer_python>=0.2.1.post2; extra == "srt"
|
247
247
|
Requires-Dist: outlines<=0.1.11,>=0.0.44; extra == "srt"
|
248
248
|
Provides-Extra: srt-hip
|
249
249
|
Requires-Dist: sglang[runtime_common]; extra == "srt-hip"
|
@@ -4,7 +4,7 @@ build-backend = "setuptools.build_meta"
|
|
4
4
|
|
5
5
|
[project]
|
6
6
|
name = "sglang"
|
7
|
-
version = "0.4.3"
|
7
|
+
version = "0.4.3.post2"
|
8
8
|
description = "SGLang is yet another fast serving framework for large language models and vision language models."
|
9
9
|
readme = "README.md"
|
10
10
|
requires-python = ">=3.8"
|
@@ -21,12 +21,12 @@ runtime_common = [
|
|
21
21
|
"hf_transfer", "huggingface_hub", "interegular", "modelscope",
|
22
22
|
"orjson", "packaging", "pillow", "prometheus-client>=0.20.0",
|
23
23
|
"psutil", "pydantic", "python-multipart", "pyzmq>=25.1.2",
|
24
|
-
"torchao>=0.7.0", "uvicorn", "uvloop", "xgrammar
|
24
|
+
"torchao>=0.7.0", "uvicorn", "uvloop", "xgrammar==0.1.10", "ninja"
|
25
25
|
]
|
26
26
|
srt = [
|
27
27
|
"sglang[runtime_common]", "cuda-python",
|
28
28
|
"sgl-kernel>=0.0.3.post6", "torch", "vllm>=0.6.4.post1,<=0.7.2",
|
29
|
-
"flashinfer_python>=0.2.1.
|
29
|
+
"flashinfer_python>=0.2.1.post2",
|
30
30
|
"outlines>=0.0.44,<=0.1.11",
|
31
31
|
]
|
32
32
|
|
@@ -161,6 +161,10 @@ class OpenAI(BaseBackend):
|
|
161
161
|
prompt = s.text_
|
162
162
|
|
163
163
|
kwargs = sampling_params.to_openai_kwargs()
|
164
|
+
if self.model_name.startswith("o1") or self.model_name.startswith("o3"):
|
165
|
+
kwargs.pop("max_tokens", None)
|
166
|
+
else:
|
167
|
+
kwargs.pop("max_completion_tokens", None)
|
164
168
|
comp = openai_completion(
|
165
169
|
client=self.client,
|
166
170
|
token_usage=self.token_usage,
|
@@ -175,6 +179,7 @@ class OpenAI(BaseBackend):
|
|
175
179
|
), "constrained type not supported on chat model"
|
176
180
|
kwargs = sampling_params.to_openai_kwargs()
|
177
181
|
kwargs.pop("stop")
|
182
|
+
|
178
183
|
comp = openai_completion(
|
179
184
|
client=self.client,
|
180
185
|
token_usage=self.token_usage,
|
@@ -353,7 +353,6 @@ register_chat_template(
|
|
353
353
|
)
|
354
354
|
)
|
355
355
|
|
356
|
-
|
357
356
|
register_chat_template(
|
358
357
|
ChatTemplate(
|
359
358
|
name="deepseek-v3",
|
@@ -428,12 +427,15 @@ def match_chat_ml(model_path: str):
|
|
428
427
|
if "tinyllama" in model_path:
|
429
428
|
return get_chat_template("chatml")
|
430
429
|
# Now the suffix for qwen2 chat model is "instruct"
|
431
|
-
if
|
432
|
-
"
|
433
|
-
|
434
|
-
|
435
|
-
|
436
|
-
|
430
|
+
if "qwen" in model_path and "vl" in model_path:
|
431
|
+
return get_chat_template("qwen2-vl")
|
432
|
+
if "qwen" in model_path:
|
433
|
+
if "vl" in model_path:
|
434
|
+
return get_chat_template("qwen2-vl")
|
435
|
+
if ("chat" in model_path or "instruct" in model_path) and (
|
436
|
+
"llava" not in model_path
|
437
|
+
):
|
438
|
+
return get_chat_template("qwen")
|
437
439
|
if (
|
438
440
|
"llava-v1.6-34b" in model_path
|
439
441
|
or "llava-v1.6-yi-34b" in model_path
|
@@ -443,6 +445,12 @@ def match_chat_ml(model_path: str):
|
|
443
445
|
return get_chat_template("chatml-llava")
|
444
446
|
|
445
447
|
|
448
|
+
@register_chat_template_matching_function
|
449
|
+
def match_chat_minicpm(model_path: str):
|
450
|
+
if "minicpm" in model_path:
|
451
|
+
return get_chat_template("minicpmv")
|
452
|
+
|
453
|
+
|
446
454
|
@register_chat_template_matching_function
|
447
455
|
def match_chat_yi(model_path: str):
|
448
456
|
model_path = model_path.lower()
|
@@ -459,6 +467,13 @@ def match_gemma_it(model_path: str):
|
|
459
467
|
return get_chat_template("gemma-it")
|
460
468
|
|
461
469
|
|
470
|
+
@register_chat_template_matching_function
|
471
|
+
def match_openbmb_minicpm(model_path: str):
|
472
|
+
model_path = model_path.lower()
|
473
|
+
if "minicpm" in model_path:
|
474
|
+
return get_chat_template("minicpmv")
|
475
|
+
|
476
|
+
|
462
477
|
@register_chat_template_matching_function
|
463
478
|
def match_c4ai_command_r(model_path: str):
|
464
479
|
model_path = model_path.lower()
|
@@ -63,6 +63,7 @@ class SglSamplingParams:
|
|
63
63
|
warnings.warn("Regular expression is not supported in the OpenAI backend.")
|
64
64
|
return {
|
65
65
|
"max_tokens": self.max_new_tokens,
|
66
|
+
"max_completion_tokens": self.max_new_tokens,
|
66
67
|
"stop": self.stop or None,
|
67
68
|
"temperature": self.temperature,
|
68
69
|
"top_p": self.top_p,
|
@@ -1,12 +1,15 @@
|
|
1
1
|
from sglang.srt.configs.chatglm import ChatGLMConfig
|
2
2
|
from sglang.srt.configs.dbrx import DbrxConfig
|
3
3
|
from sglang.srt.configs.exaone import ExaoneConfig
|
4
|
-
from sglang.srt.configs.
|
4
|
+
from sglang.srt.configs.qwen2_5_vl_config import (
|
5
|
+
Qwen2_5_VLConfig,
|
6
|
+
Qwen2_5_VLVisionConfig,
|
7
|
+
)
|
5
8
|
|
6
9
|
__all__ = [
|
7
10
|
"ExaoneConfig",
|
8
|
-
"Qwen2VLConfig",
|
9
|
-
"Qwen2VLVisionConfig",
|
10
11
|
"ChatGLMConfig",
|
11
12
|
"DbrxConfig",
|
13
|
+
"Qwen2_5_VLConfig",
|
14
|
+
"Qwen2_5_VLVisionConfig",
|
12
15
|
]
|
@@ -98,6 +98,7 @@ class ModelConfig:
|
|
98
98
|
if (
|
99
99
|
"DeepseekV2ForCausalLM" in self.hf_config.architectures
|
100
100
|
or "DeepseekV3ForCausalLM" in self.hf_config.architectures
|
101
|
+
or "DeepseekV3ForCausalLMNextN" in self.hf_config.architectures
|
101
102
|
):
|
102
103
|
self.head_dim = 256
|
103
104
|
self.attention_arch = AttentionArch.MLA
|
@@ -402,6 +403,7 @@ def is_multimodal_model(model_architectures: List[str]):
|
|
402
403
|
or "LlavaVidForCausalLM" in model_architectures
|
403
404
|
or "MllamaForConditionalGeneration" in model_architectures
|
404
405
|
or "Qwen2VLForConditionalGeneration" in model_architectures
|
406
|
+
or "Qwen2_5_VLForConditionalGeneration" in model_architectures
|
405
407
|
or "MiniCPMV" in model_architectures
|
406
408
|
):
|
407
409
|
return True
|