Skip to content

[Devices] Offer support for hardware-accelerated inference in Firecracker #1179

Open
@raduweiss

Description

@raduweiss

Doing hardware-accelerated inference in a serverless environment is compelling use case.

However, adding straight up GPU passthrough means that microVM can't oversubscribe memory, and we need to add PCI emulation to Firecracker, which comes with a lot of extra complexity/attack surface.

The first step here will be to research the options and alternatives (e.g., GPU passthrough, or something else), and figure out the path forward.

Related issues: #849, #776.

Metadata

Metadata

Assignees

No one assigned

    Labels

    Status: ParkedIndicates that an issues or pull request will be revisited laterType: EnhancementIndicates new feature requests

    Type

    No type

    Projects

    Status

    Researching

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions