arrow Blog

Announcing the new Score Examples Hub

date November 3, 2025

date 2 min read

Announcement of the new place for all the examples about Score: The Score Examples Hub!
Tobias Babin Tobias Babin, Technical Documentation Specialist
Announcing the new Score Examples Hub

SHARE

linkedin share twitter share reddit share

Hi, there! Tobi here, Tech Writer at Humanitec and Score contributor.

🎉 Today, Score is officially announcing the new Score Examples Hub! 🎉

One of the feedbacks we have been capturing is to show more concrete examples and use cases with Score.

When we celebrated the first anniversary of Score as CNCF Sandbox a couple of months ago, we provided a sneak peek about this Score Examples Hub. Since then, we have been adding more examples of Score files, resource provisioners and patch templates.

Screenshot of the Score Examples Hub Homepage

Score specification and resources examples

Find Score files examples illustrating how to use the Score specification as well as the resources provisioners with either score-compose or score-k8s:

Screenshot of the Score Examples Hub for Spec & Resources

As an example, that’s how in there you will be able to find how you can use a local LLM model with Score and score-compose:

score.yaml:

apiVersion: score.dev/v1b1
metadata:
  name: my-workload
containers:
  my-container:
    image: busybox
    command: ["/bin/sh"]
    args: ["-c", "while true; do echo $LLM_BASE_URL; sleep 5; done"]
    variables:
      LLM_MODEL_NAME: "${resources.model.model}"
      LLM_BASE_URL: "${resources.model.url}"
resources:
  model:
    type: llm-model
    params:
      model: ai/smollm2:135M-Q4_0

Run:

score-compose init \
    --provisioners https://raw.githubusercontent.com/score-spec/community-provisioners/refs/heads/main/llm-model/score-compose/10-dmr-llm-model-via-service-provider.provisioners.yaml

score-compose generate score.yaml -o compose.yaml

docker compose up -d --wait

In the score.yaml file, the Developer can request a llm-model resource to use it in their app by injecting the corresponding url generated when this Score file will be deployed. By using score-compose init --provisioners, the actual implementation of the llm-model resource is downloaded locally (you can look at its definition here, Docker Model Runner is used). Then, with score-compose generate, both the my-container and the llm-model services are generated into a compose.yaml file. Finally, docker compose up will deploy them.

Score resources provisioners examples

Find examples of resources provisioners with either score-compose or score-k8s:

Screenshot of the Score Examples Hub for Resources Provisioners

Score patch templates examples

Find examples of patch templates with either score-compose or score-k8s:

Screenshot of the Score Examples Hub for Patch Templates

More advanced examples

Find more advanced examples illustrating how to deploy a NodeJS application talking to a PostgreSQL database, how to deploy Backstage and Nginx, how to use Dapr with your workloads, etc. with either score-compose or score-k8s:

Screenshot of the Score Examples Hub for more advanced use cases

Kudos!

Huge kudos to Santiago Beroch for collaborating with me to design and implement this very powerful Score Examples Hub! And thanks, Humanitec for your support on this!

What’s next?

Please share your feedback with us and please do let us know if there is anything in particular you would like to see in this growing list of examples!

See you in Atlanta for KubeCon NA 2025!?

arrow Blog overview