|
9 | 9 | /vllm/model_executor/layers/quantization @mgoin @robertgshaw2-redhat @tlrmchlsmth @yewentao256 @pavanimajety |
10 | 10 | /vllm/model_executor/layers/mamba @tdoublep |
11 | 11 | /vllm/model_executor/model_loader @22quinn |
12 | | -/vllm/multimodal @DarkLight1337 @ywang96 @NickLucche |
| 12 | +/vllm/multimodal @DarkLight1337 @ywang96 @NickLucche @tjtanaa |
13 | 13 | /vllm/vllm_flash_attn @LucasWilkinson |
14 | 14 | /vllm/lora @jeejeelee |
15 | 15 | /vllm/reasoning @aarnphm @chaunceyjiang |
@@ -105,11 +105,21 @@ mkdocs.yaml @hmellor |
105 | 105 | /vllm/attention/ops/triton_unified_attention.py @tdoublep |
106 | 106 |
|
107 | 107 | # ROCm related: specify owner with write access to notify AMD folks for careful code review |
108 | | -/docker/Dockerfile.rocm* @gshtras |
109 | | -/vllm/v1/attention/backends/rocm*.py @gshtras |
110 | | -/vllm/v1/attention/backends/mla/rocm*.py @gshtras |
111 | | -/vllm/attention/ops/rocm*.py @gshtras |
112 | | -/vllm/model_executor/layers/fused_moe/rocm*.py @gshtras |
| 108 | +/vllm/**/*rocm* @tjtanaa |
| 109 | +/docker/Dockerfile.rocm* @gshtras @tjtanaa |
| 110 | +/vllm/v1/attention/backends/rocm*.py @gshtras @tjtanaa |
| 111 | +/vllm/v1/attention/backends/mla/rocm*.py @gshtras @tjtanaa |
| 112 | +/vllm/attention/ops/rocm*.py @gshtras @tjtanaa |
| 113 | +/vllm/model_executor/layers/fused_moe/rocm*.py @gshtras @tjtanaa |
| 114 | +/csrc/rocm @gshtras @tjtanaa |
| 115 | +/requirements/*rocm* @tjtanaa |
| 116 | +/tests/**/*rocm* @tjtanaa |
| 117 | +/docs/**/*rocm* @tjtanaa |
| 118 | +/vllm/**/*quark* @tjtanaa |
| 119 | +/tests/**/*quark* @tjtanaa |
| 120 | +/docs/**/*quark* @tjtanaa |
| 121 | +/vllm/**/*aiter* @tjtanaa |
| 122 | +/tests/**/*aiter* @tjtanaa |
113 | 123 |
|
114 | 124 | # TPU |
115 | 125 | /vllm/v1/worker/tpu* @NickLucche |
|
0 commit comments