Skip to content

implement weight_norm on mps  #104513

@nwoodall

Description

@nwoodall

🚀 The feature, motivation and pitch

For some models the best normalization method is using weight_norm. This is especially true when training for a hardware implementation for processing high resolution images. It's not possible to use batch_norm because the batch size is one, group/layer normalization adds calculations and latency to the hardware implementation.

Alternatives

Allowing fallback on the CPU during training with the obvious training time penalty

Additional context

No response

cc @kulinseth @albanD @malfet @DenisVieriu97 @razarmehr @abhudev

Metadata

Metadata

Assignees

No one assigned

    Labels

    module: mpsRelated to Apple Metal Performance Shaders frameworktriagedThis issue has been looked at a team member, and triaged and prioritized into an appropriate module

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions