Skip to content

[REQUEST] - CUDA support #26

@ryanznie

Description

@ryanznie

Feature

Support for FP16, INT8 models on the CPU / MPS

Reason

Faster inference. ONNX currently does not support FP16 on CPU or MPS: Issue

Metadata

Metadata

Assignees

No one assigned

    Labels

    enhancementNew feature or request

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions