This repository hosts code related to Large Language Models (LLM) inference, including references and examples. It aims to provide resources for leveraging LLMs in various applications.
Inside the mk1
folder, you will find code for setting up a FastAPI endpoint using MK1 for inference with Flywheel. The endpoint allows you to deploy and interact with LLM models efficiently and securely.
To set up the development environment using Poetry:
-
Install Poetry if you haven't already:
curl -sSL https://raw.githubusercontent.com/python-poetry/poetry/master/get-poetry.py | python -
-
Clone this repository and navigate to the project directory:
git clone https://github.com/albertoecf/llm_inferece.git cd llm_inference
-
Install dependencies using Poetry:
poetry install
-
Activate the virtual environment created by Poetry:
poetry shell
Now you're ready to start working with the LLM inference code in this repository! Refer to individual folders and files for specific instructions and examples.