@@ -393,13 +393,14 @@ steps:
393
393
commands :
394
394
- pytest -v -s benchmarks/
395
395
396
- - label : Quantization Test # 33min
396
+ - label : Quantization Test
397
397
torch_nightly : true
398
398
source_file_dependencies :
399
399
- csrc/
400
400
- vllm/model_executor/layers/quantization
401
401
- tests/quantization
402
- command : VLLM_TEST_FORCE_LOAD_FORMAT=auto pytest -v -s quantization
402
+ commands :
403
+ - VLLM_TEST_FORCE_LOAD_FORMAT=auto pytest -v -s quantization
403
404
404
405
- label : LM Eval Small Models # 53min
405
406
working_dir : " /vllm-workspace/.buildkite/lm-eval-harness"
@@ -446,82 +447,78 @@ steps:
446
447
commands :
447
448
- pytest -v -s models/test_transformers.py
448
449
- pytest -v -s models/test_registry.py
450
+ - pytest -v -s models/test_utils.py
451
+ - pytest -v -s models/test_vision.py
449
452
# V1 Test: https://github.com/vllm-project/vllm/issues/14531
450
453
- VLLM_USE_V1=0 pytest -v -s models/test_initialization.py -k 'not llama4 and not plamo2'
451
454
- VLLM_USE_V1=0 pytest -v -s models/test_initialization.py -k 'llama4'
452
455
- VLLM_USE_V1=0 pytest -v -s models/test_initialization.py -k 'plamo2'
453
456
454
- - label : Language Models Test (Standard) # 32min
457
+ - label : Language Models Test (Standard)
455
458
# mirror_hardwares: [amd]
456
459
source_file_dependencies :
457
460
- vllm/
458
- - tests/models/decoder_only/language
459
- - tests/models/embedding/language
460
- - tests/models/encoder_decoder/language
461
+ - tests/models/language
461
462
commands :
462
463
# Install causal-conv1d for plamo2 models here, as it is not compatible with pip-compile.
463
464
- pip install 'git+https://github.com/Dao-AILab/causal-conv1d@v1.5.0.post8'
464
- - pytest -v -s models/decoder_only/language -m 'core_model or quant_model'
465
- - pytest -v -s models/embedding/language -m core_model
465
+ - pytest -v -s models/language -m core_model
466
466
467
- - label : Language Models Test (Extended) # 1h10min
467
+ - label : Language Models Test (Extended)
468
468
optional : true
469
469
source_file_dependencies :
470
470
- vllm/
471
- - tests/models/decoder_only/language
472
- - tests/models/embedding/language
473
- - tests/models/encoder_decoder/language
471
+ - tests/models/language
474
472
commands :
475
473
# Install causal-conv1d for plamo2 models here, as it is not compatible with pip-compile.
476
- - pip install causal-conv1d
477
- - pytest -v -s models/decoder_only/language -m 'not core_model and not quant_model'
478
- - pytest -v -s models/embedding/language -m 'not core_model'
474
+ - pip install 'git+https://github.com/Dao-AILab/causal-conv1d@v1.5.0.post8'
475
+ - pytest -v -s models/language -m 'not core_model'
479
476
480
- - label : Multi-Modal Models Test (Standard) # 40min
477
+ - label : Multi-Modal Models Test (Standard)
481
478
# mirror_hardwares: [amd]
482
479
source_file_dependencies :
483
480
- vllm/
484
- - tests/models/decoder_only/audio_language
485
- - tests/models/decoder_only/vision_language
486
- - tests/models/embedding/vision_language
487
- - tests/models/encoder_decoder/audio_language
488
- - tests/models/encoder_decoder/vision_language
481
+ - tests/models/multimodal
489
482
commands :
490
483
- pip install git+https://github.com/TIGER-AI-Lab/Mantis.git
491
- - pytest -v -s models/multimodal
492
- - pytest -v -s models/decoder_only/audio_language -m 'core_model or quant_model'
493
- - pytest -v -s models/decoder_only/vision_language -m 'core_model or quant_model'
494
- - pytest -v -s models/embedding/vision_language -m core_model
495
- - pytest -v -s models/encoder_decoder/audio_language -m core_model
496
- - pytest -v -s models/encoder_decoder/language -m core_model
497
- - pytest -v -s models/encoder_decoder/vision_language -m core_model
498
- - pytest -v -s models/decoder_only/vision_language/test_interleaved.py
499
-
500
- - label : Multi-Modal Models Test (Extended) 1 # 48m
484
+ - pytest -v -s models/multimodal/processing
485
+ - pytest -v -s --ignore models/multimodal/generation/test_whisper.py models/multimodal -m core_model
486
+ - cd .. && pytest -v -s tests/models/multimodal/generation/test_whisper.py -m core_model # Otherwise, mp_method="spawn" doesn't work
487
+
488
+ - label : Multi-Modal Models Test (Extended) 1
501
489
optional : true
502
490
source_file_dependencies :
503
491
- vllm/
504
- - tests/models/decoder_only/audio_language
505
- - tests/models/decoder_only/vision_language
506
- - tests/models/embedding/vision_language
507
- - tests/models/encoder_decoder/vision_language
492
+ - tests/models/multimodal
508
493
commands :
509
494
- pip install git+https://github.com/TIGER-AI-Lab/Mantis.git
510
- - pytest -v -s models/decoder_only/audio_language -m 'not core_model and not quant_model'
511
- - pytest -v -s models/decoder_only/vision_language/test_models.py -m 'split(group=0) and not core_model and not quant_model'
512
- - pytest -v -s --ignore models/decoder_only/vision_language/test_models.py models/decoder_only/vision_language -m 'not core_model and not quant_model'
513
- - pytest -v -s models/embedding/vision_language -m 'not core_model'
514
- - pytest -v -s models/encoder_decoder/language -m 'not core_model'
515
- - pytest -v -s models/encoder_decoder/vision_language -m 'not core_model'
516
-
517
- - label : Multi-Modal Models Test (Extended) 2 # 38m
495
+ - pytest -v -s --ignore models/multimodal/generation/test_common.py --ignore models/multimodal/processing models/multimodal -m 'not core_model'
496
+
497
+ - label : Multi-Modal Models Test (Extended) 2
518
498
optional : true
519
499
source_file_dependencies :
520
500
- vllm/
521
- - tests/models/decoder_only/vision_language
501
+ - tests/models/multimodal
522
502
commands :
523
503
- pip install git+https://github.com/TIGER-AI-Lab/Mantis.git
524
- - pytest -v -s models/decoder_only/vision_language/test_models.py -m 'split(group=1) and not core_model and not quant_model'
504
+ - pytest -v -s models/multimodal/generation/test_common.py -m 'split(group=0) and not core_model'
505
+
506
+ - label : Multi-Modal Models Test (Extended) 3
507
+ optional : true
508
+ source_file_dependencies :
509
+ - vllm/
510
+ - tests/models/multimodal
511
+ commands :
512
+ - pip install git+https://github.com/TIGER-AI-Lab/Mantis.git
513
+ - pytest -v -s models/multimodal/generation/test_common.py -m 'split(group=1) and not core_model'
514
+
515
+ - label : Quantized Models Test
516
+ # mirror_hardwares: [amd]
517
+ source_file_dependencies :
518
+ - vllm/model_executor/layers/quantization
519
+ - tests/models/quantization
520
+ commands :
521
+ - pytest -v -s models/quantization
525
522
526
523
# This test is used only in PR development phase to test individual models and should never run on main
527
524
- label : Custom Models Test
@@ -591,9 +588,8 @@ steps:
591
588
- TARGET_TEST_SUITE=L4 pytest basic_correctness/ -v -s -m 'distributed(num_gpus=2)'
592
589
# Avoid importing model tests that cause CUDA reinitialization error
593
590
- pytest models/test_transformers.py -v -s -m 'distributed(num_gpus=2)'
594
- - pytest models/encoder_decoder/language/test_bart.py -v -s -m 'distributed(num_gpus=2)'
595
- - pytest models/encoder_decoder/vision_language/test_broadcast.py -v -s -m 'distributed(num_gpus=2)'
596
- - pytest models/decoder_only/vision_language/test_models.py -v -s -m 'distributed(num_gpus=2)'
591
+ - pytest models/language -v -s -m 'distributed(num_gpus=2)'
592
+ - pytest models/multimodal -v -s -m 'distributed(num_gpus=2)'
597
593
# test sequence parallel
598
594
- pytest -v -s distributed/test_sequence_parallel.py
599
595
# this test fails consistently.
0 commit comments