Skip to content

AmpereComputingAI/llama.cpp

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

9 Commits
 
 
 
 
 
 

Repository files navigation

llama.cpp

Ampere® optimized llama.cpp

llama.cpp pull count

Ampere® optimized build of llama.cpp with full support for rich collection of GGUF models available at HuggingFace: GGUF models

This Docker image can be run on bare metal Ampere® CPUs and Ampere® based VMs available in the cloud.

Release notes and binary executables are available on our GitHub

Starting container

Default entrypoint runs the server binary of llama.cpp, mimicking behavior of original llama.cpp server image: docker image

To launch shell instead, do this:

sudo docker run --privileged=true --name llama --entrypoint /bin/bash -it amperecomputingai/llama.cpp:latest

Quick start example will be presented at docker container launch:

quick start

Make sure to visit us at Ampere Solutions Portal!

Support

Please contact us at [email protected]

LEGAL NOTICE

By accessing, downloading or using this software and any required dependent software (the “Ampere AI Software”), you agree to the terms and conditions of the software license agreements for the Ampere AI Software, which may also include notices, disclaimers, or license terms for third party software included with the Ampere AI Software. Please refer to the Ampere AI Software EULA v1.6 or other similarly-named text file for additional details.