To run the TheBloke/Mistral-7B-OpenOrca-AWQ
model on Hugging Face infrastructure using a Clojure clj-http
client,
you’ll need to leverage Hugging Face's Inference API, which allows you to make HTTP requests to run models hosted on
their platform. Since you have a Hugging Face Pro account, you can use this API with your authentication token for
enhanced access (e.g., higher rate limits or access to private models, though this specific model is public).
Here’s a step-by-step guide to achieve this: