It is now possible to run AI models directly through the docker model runner. GPU support etc comes natively. We should extend the GMT to allow energy benchmarking of these runs.
https://docs.docker.com/ai/model-runner/api-reference/
We should also check if it is possible to somehow get single prompts measured. This should also become a blog post.