Skip to content

Shunmo17/llama_ros

 
 

Repository files navigation

llama_ros

This is a ROS 1 wrapper for llama.cpp. Original is llama_ros. This package support only llama2. If you want to use other LLMs, please add launch files for them.

Requirement

  • ROS1 Noetic

Usage

Launch llama2 server.

Run on GPU

roslaunch llama_bringup llama2.launch

run on CPU

roslaunch llama_bringup llama2.launch n_gpu_layers:=0

Send prompt

roslaunch llama_ros llama_client_node.launch prompt:=<YOUR PROMPT>

Installation

Firstly, you have to an environment for llama2. Then, please follow the instalation procudure.

Setup workspace

Please modify the workspace path to adapt ypur environment.

mkdir -p workspace/src
cd workspace/src
git clone --recursive [email protected]:Shunmo17/llama_ros.git
catkin b

Download models

Llama2 models finetuned for chat are available here: 7B, 13B, 70B

After downloding a model, please set the correct model path in llama2.launch.

Parameters

  • use_default_sampling_config

    If it is true, we ignore the samping config in the requested goal. This parameter is added not to use uninitialized parameters because we cannot set default parameters of action messages in ROS1.

Maintainer

Shunmo17

About

llama.cpp for ROS 1

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • C++ 90.9%
  • Python 6.1%
  • CMake 3.0%