In today's event, we'll create an E2E Application through Hugging Face Inference Endpoints!
There are 2 main sections to this event:
Select "Inference Endpoint" from the "Solutions" button in Hugging Face:
Create a "+ New Endpoint" from the Inference Endpoints dashboard.
Select the ai-maker-space/gen-z-translate-llama-3-instruct-v1
model repository and name your endpoint. Select N. Virginia as your region (us-east-1
). Give your endpoint an appropriate name.
Select the following settings for your Advanced Configuration
.
Create a Protected
endpoint.
If you were successful, you should see the following screen:
You'll repeat the same process for your embedding model!
First, we fine-tune Llama 3 8B Instruct for a specific task, in this case: A translation task!
Then, we create a Docker Hugging Face space powering a Chainlit UI - code available here
Please go to each endpoint's settings and select Delete Endpoint
. To delete the resources, you will need to type the endpoint's name.