Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open question: should we enable the user to explicitly control whether and what hardware accelerator is used? #27

Open
yuhonglin opened this issue Mar 10, 2022 · 0 comments

Comments

@yuhonglin
Copy link
Contributor

The ML hardware accelerator is still developing quickly and it is hard to group them all into one single category (not like GPU which has become comparatively stable). Should we expose some interface to let the client control whether and what such SoC will be used for inference?

The pros is that a ML model may or may not be suitable for a particular SoC so a user can explicitly control this for optimal performance. But the cons are big:

  1. This may introduce strong privacy issue (i.e. the client needs to know which SoCs are available?)
  2. This is rather complex, not sure whether web developers can correctly handle this.
  3. The API will be much more complex.

So I personally prefer not to expose such interface. Instead, besides "CPU" and "GPU", maybe we can provide a third type of device preference "Auto", for which, we will use some heuristic to transparently apply the ML SoCs when it is available and appropriate.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant