Run LLM inference on Cloud Run GPUs with Hugging Face Transformers.js (services)

The following codelab shows how to run a backend service that runs the Transformers.js package. The Transformers.js package is functionally equivalent to the Hugging Face transformers python library together with Google's Gemma 2 model.

See the entire codelab at How to Run Transformers.js on Cloud Run GPUs.