Server/deployment considerations for Bumblebee?

what processor constraints does this stack have?
GPU?

is there a deployment README.md somewhere?

2 Likes

Phoenix app examples with deployment considerations can be found here: bumblebee/examples/phoenix at main · elixir-nx/bumblebee · GitHub

6 Likes

Thanks, was thinking more about GPU friendly vendor support, i.e. fly.io or …

2 Likes

I’ve come across https://www.vultr.com/products/cloud-gpu/ as an option but haven’t tried it out (also not sure how the pricing compares to other options…)

1 Like

Vultr is good value:performance for sure

3 Likes

There is a related thread here that may be of interest:

(Might be worth posting some of those cloud providers there as well :D)

1 Like

Also came across this: A Google Colab that runs Elixir livebook w/ BumbleeBee and CUDA acceleration (probably only useful for development / personal use though)

Edit: it is indeed limited, some attempts at using Bumblebee fail with an error like Allocator (GPU_0_bfc) ran out of memory trying to allocate 4.41GiB requested by op

1 Like

Consider carefully assessing whether you need a GPU in production. Inference is not as compute intensive as training models. Can you meet business domain needs with just the CPU version of XLA or Torchx? Many initial product MVPs can live with the sub-second latency of some models running on the CPU.

4 Likes