Skip to content

[Feature] Enable remote inference endpoints for examples #1973

@joshuayao

Description

@joshuayao

Priority

P1-Stopper

OS type

Ubuntu

Hardware type

Xeon-GNR

Running nodes

Single Node

Description

Story for users

  1. for Xeon users, users can try bigger models on remote endpoint with Gaudi while bigger models might not work well on Xeon
  2. for Xeon users, if users want to try on Gaudi to understand the difference between Xeon and Gaudi, they can use remote endpoint
  3. for Gaudi users, we assume that users have local Gaudi access to run GenAIExamples.
  4. target Denvr and IBM for remote endpoints in the GenAIExamples

Enable remote inference endpoints for

Models are not supported in current public remote endpoint. ON HOLD for now.

  • VideoQnA : (?)
  • MulitmodalQnA : Alex
  • VisualQnA: Alex

Metadata

Metadata

Labels

Projects

Status

In progress

Relationships

None yet

Development

No branches or pull requests

Issue actions