@@ -486,6 +486,56 @@ jobs:
486
486
with :
487
487
name : torch_cuda_${{ matrix.config.backend }}_reports
488
488
path : reports
489
+
490
+ run_nightly_pipeline_level_quantization_tests :
491
+ name : Torch quantization nightly tests
492
+ strategy :
493
+ fail-fast : false
494
+ max-parallel : 2
495
+ runs-on :
496
+ group : aws-g6e-xlarge-plus
497
+ container :
498
+ image : diffusers/diffusers-pytorch-cuda
499
+ options : --shm-size "20gb" --ipc host --gpus 0
500
+ steps :
501
+ - name : Checkout diffusers
502
+ uses : actions/checkout@v3
503
+ with :
504
+ fetch-depth : 2
505
+ - name : NVIDIA-SMI
506
+ run : nvidia-smi
507
+ - name : Install dependencies
508
+ run : |
509
+ python -m venv /opt/venv && export PATH="/opt/venv/bin:$PATH"
510
+ python -m uv pip install -e [quality,test]
511
+ python -m uv pip install -U bitsandbytes optimum_quanto
512
+ python -m uv pip install pytest-reportlog
513
+ - name : Environment
514
+ run : |
515
+ python utils/print_env.py
516
+ - name : Pipeline-level quantization tests on GPU
517
+ env :
518
+ HF_TOKEN : ${{ secrets.DIFFUSERS_HF_HUB_READ_TOKEN }}
519
+ # https://pytorch.org/docs/stable/notes/randomness.html#avoiding-nondeterministic-algorithms
520
+ CUBLAS_WORKSPACE_CONFIG : :16:8
521
+ BIG_GPU_MEMORY : 40
522
+ run : |
523
+ python -m pytest -n 1 --max-worker-restart=0 --dist=loadfile \
524
+ --make-reports=tests_pipeline_level_quant_torch_cuda \
525
+ --report-log=tests_pipeline_level_quant_torch_cuda.log \
526
+ tests/quantization/test_pipeline_level_quantization.py
527
+ - name : Failure short reports
528
+ if : ${{ failure() }}
529
+ run : |
530
+ cat reports/tests_pipeline_level_quant_torch_cuda_stats.txt
531
+ cat reports/tests_pipeline_level_quant_torch_cuda_failures_short.txt
532
+ - name : Test suite reports artifacts
533
+ if : ${{ always() }}
534
+ uses : actions/upload-artifact@v4
535
+ with :
536
+ name : torch_cuda_pipeline_level_quant_reports
537
+ path : reports
538
+
489
539
generate_consolidated_report :
490
540
name : Generate Consolidated Test Report
491
541
needs : [
@@ -494,6 +544,7 @@ jobs:
494
544
run_big_gpu_torch_tests,
495
545
run_nightly_onnx_tests,
496
546
run_nightly_quantization_tests,
547
+ run_nightly_pipeline_level_quantization_tests,
497
548
run_flax_tpu_tests
498
549
]
499
550
if : always()
@@ -542,60 +593,6 @@ jobs:
542
593
name : consolidated_test_report
543
594
path : ${{ env.CONSOLIDATED_REPORT_PATH }}
544
595
545
- run_nightly_pipeline_level_quantization_tests :
546
- name : Torch quantization nightly tests
547
- strategy :
548
- fail-fast : false
549
- max-parallel : 2
550
- runs-on :
551
- group : aws-g6e-xlarge-plus
552
- container :
553
- image : diffusers/diffusers-pytorch-cuda
554
- options : --shm-size "20gb" --ipc host --gpus 0
555
- steps :
556
- - name : Checkout diffusers
557
- uses : actions/checkout@v3
558
- with :
559
- fetch-depth : 2
560
- - name : NVIDIA-SMI
561
- run : nvidia-smi
562
- - name : Install dependencies
563
- run : |
564
- python -m venv /opt/venv && export PATH="/opt/venv/bin:$PATH"
565
- python -m uv pip install -e [quality,test]
566
- python -m uv pip install -U bitsandbytes optimum_quanto
567
- python -m uv pip install pytest-reportlog
568
- - name : Environment
569
- run : |
570
- python utils/print_env.py
571
- - name : Pipeline-level quantization tests on GPU
572
- env :
573
- HF_TOKEN : ${{ secrets.DIFFUSERS_HF_HUB_READ_TOKEN }}
574
- # https://pytorch.org/docs/stable/notes/randomness.html#avoiding-nondeterministic-algorithms
575
- CUBLAS_WORKSPACE_CONFIG : :16:8
576
- BIG_GPU_MEMORY : 40
577
- run : |
578
- python -m pytest -n 1 --max-worker-restart=0 --dist=loadfile \
579
- --make-reports=tests_pipeline_level_quant_torch_cuda \
580
- --report-log=tests_pipeline_level_quant_torch_cuda.log \
581
- tests/quantization/test_pipeline_level_quantization.py
582
- - name : Failure short reports
583
- if : ${{ failure() }}
584
- run : |
585
- cat reports/tests_pipeline_level_quant_torch_cuda_stats.txt
586
- cat reports/tests_pipeline_level_quant_torch_cuda_failures_short.txt
587
- - name : Test suite reports artifacts
588
- if : ${{ always() }}
589
- uses : actions/upload-artifact@v4
590
- with :
591
- name : torch_cuda_pipeline_level_quant_reports
592
- path : reports
593
- - name : Generate Report and Notify Channel
594
- if : always()
595
- run : |
596
- pip install slack_sdk tabulate
597
- python utils/log_reports.py >> $GITHUB_STEP_SUMMARY
598
-
599
596
# M1 runner currently not well supported
600
597
# TODO: (Dhruv) add these back when we setup better testing for Apple Silicon
601
598
# run_nightly_tests_apple_m1:
0 commit comments