Skip to content

Support H100 training with FP8 in Trainer and Deepspeed #25333

@michaelroyzen

Description

@michaelroyzen

Feature request

Support H100 training with FP8 in Trainer and Deepspeed

Motivation

FP8 should be much faster than FP16 on supported Hopper hardware. Particularly with Deepspeed integration @stas00

Your contribution

Happy to help in any way that I can.

Metadata

Metadata

Assignees

No one assigned

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions