Skip to content

Commit fc816c3

Browse files
authored
Changed word in README about loading topology data for clarification
1 parent 1264d3f commit fc816c3

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

finetuning/StarCoder2/inference.ipynb

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,7 @@
1111
"cell_type": "markdown",
1212
"metadata": {},
1313
"source": [
14-
"In the previous [notebook](https://github.com/NVIDIA/GenerativeAIExamples/blob/main/finetuning/StarCoder2/lora.ipynb), we show how to parameter efficiently finetune StarCoder2 model with a custom code (instruction, completion) pair dataset. We choose LoRA as our PEFT algorithnm and finetune for 50 interations. In this notebook, the goal is to demonstrate how to compile fintuned .nemo model into optimized TensorRT-LLM engines. The converted model engine can perform accelerated inference locally or be deployed to Triton Inference Server."
14+
"In the previous [notebook](https://github.com/NVIDIA/GenerativeAIExamples/blob/main/finetuning/StarCoder2/lora.ipynb), we show how to parameter efficiently finetune StarCoder2 model with a custom code (instruction, completion) pair dataset. We choose LoRA as our PEFT algorithnm and finetune for 50 iterations. In this notebook, the goal is to demonstrate how to compile fintuned .nemo model into optimized TensorRT-LLM engines. The converted model engine can perform accelerated inference locally or be deployed to Triton Inference Server."
1515
]
1616
},
1717
{

0 commit comments

Comments
 (0)