We will set up an LMQL environment with Hugging Face models using Docker.
LMQL is a query language for large language models (LLMs). It facilitates LLM interaction by combining the benefits of natural language prompting with the expressiveness of Python. With only a few lines of LMQL code, users can express advanced, multi-part and tool-augmented LM queries, which then are optimized by the LMQL runtime to run efficiently as part of the LM decoding loop.
"The no-GPU version was tested to work on Ubuntu 22.04 and macOS 13.2 Ventura or Windows 10 via WSL2."
Otherwise, you may have to run an Ubuntu image with Docker: