-
Notifications
You must be signed in to change notification settings - Fork 0
Open
Description
Background
There was a fantastic idea from a Redditor about making EnergeticAI run well in Cloudflare Workers, so you can have super fast inference at the edge, without the need to distribute your model weights:
Goal
This task is to add support for Deno Deploy to EnergeticAI.
Approach
Given that Deno Deploy has even more restrictive bundle limits than AWS Lambda, I suspect the way to do this would be to distribute sharded model weights in Deno KV, and then fetch from that in parallel on function invocation. On paper at least KV values should be colocated with the functions enough that this should be fast.
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
No labels