In this chapter, we will introduce the basic usage workflow of D-Robotics-LLM to help you get started quickly. Here, we take the DeepSeek-R1-Distill-Qwen-1.5B model as an example to demonstrate its usage.
Download the provided D-Robotics_LLM_{version}.tar.gz deployment package and extract it.
Download the provided DeepSeek_R1_Distill_Qwen_1.5B_1024.hbm model.
For the download link of the hbm model, please refer to the resolve_model.txt file in the model folder of oellm_runtime.
After preparing both the deployment package and the model, integrate the model (*.hbm) with the oellm_runtime SDK from the deployment package. The reference directory structure is as follows:
Create a working directory on the device using the following commands as a reference:
Copy the integrated folder from your development machine to this device directory. Reference command:
Finally, configure LD_LIBRARY_PATH under the path /home/root/llm/D-Robotics_LLM_{version}/oellm_runtime. Reference commands:
Reference command for running the model on the device:
The runtime parameters are as follows:
| Parameter | Description | Optional/Required |
|---|---|---|
-h, --help | Display help information. | / |
--hbm_path | Specifies the path to the quantized model file (*.hbm). | Required |
--tokenizer_dir | Specifies the tokenizer configuration path. | Required |
--template_path | Specifies the conversation template path. | Required |
--model_type | Specifies the model type to run; currently, the DeepSeek model type is 1. | Required |
A simple dialogue test example is shown below: