-
Notifications
You must be signed in to change notification settings - Fork 13
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
- Loading branch information
1 parent
ba59df4
commit 76ac4dc
Showing
10 changed files
with
145 additions
and
0 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,20 @@ | ||
+++ | ||
title = 'Minimalist Team Leader, First Edition (2017)' | ||
date = 2024-10-20 | ||
|
||
+++ | ||
|
||
This booklet was quickly completed by the author in August 2017, intended as a quick-start guide for "senior" interns and doctoral students preparing to lead others. It also serves as a summary of the author's five years of experience in team management. | ||
|
||
[Read the full article](https://mp.weixin.qq.com/s/8sq0rTfkcGQjJU2ZfnjzQw) | ||
|
||
|
||
|
||
|
||
|
||
|
||
|
||
|
||
|
||
|
||
|
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,11 @@ | ||
+++ | ||
title = 'Progress in the RISC-V + AI Ecosystem: The llama.cpp Optimization for RVV 1.0 is Complete, Achieving up to a 350% Performance Boost! The Code is Open-Source—Feel Free to Replicate and Explore' | ||
date = 2024-10-26 | ||
|
||
+++ | ||
|
||
llama.cpp is a large language model (LLM) inference framework implemented entirely in C/C++, relying heavily on its derivative project, ggml, for tensor operations, which requires high computational performance. | ||
|
||
Recently, **xctan**, an intern at the PLCT Lab, enhanced ggml’s Q4_0_8_8 quantized matrix multiplication operator by adding optimized support for RISC-V Vector 1.0, achieving significant performance improvements. | ||
|
||
[Read the full article](https://mp.weixin.qq.com/s/e36oemQkHoSTUpCrcEqFOg) |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,20 @@ | ||
+++ | ||
title = 'Minimalist Team Leader, First Edition (2017)' | ||
date = 2024-10-20 | ||
|
||
+++ | ||
|
||
This booklet was quickly completed by the author in August 2017, intended as a quick-start guide for "senior" interns and doctoral students preparing to lead others. It also serves as a summary of the author's five years of experience in team management. | ||
|
||
[Read the full article](https://mp.weixin.qq.com/s/8sq0rTfkcGQjJU2ZfnjzQw) | ||
|
||
|
||
|
||
|
||
|
||
|
||
|
||
|
||
|
||
|
||
|
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,11 @@ | ||
+++ | ||
title = 'Progress in the RISC-V + AI Ecosystem: The llama.cpp Optimization for RVV 1.0 is Complete, Achieving up to a 350% Performance Boost! The Code is Open-Source—Feel Free to Replicate and Explore' | ||
date = 2024-10-26 | ||
|
||
+++ | ||
|
||
llama.cpp is a large language model (LLM) inference framework implemented entirely in C/C++, relying heavily on its derivative project, ggml, for tensor operations, which requires high computational performance. | ||
|
||
Recently, **xctan**, an intern at the PLCT Lab, enhanced ggml’s Q4_0_8_8 quantized matrix multiplication operator by adding optimized support for RISC-V Vector 1.0, achieving significant performance improvements. | ||
|
||
[Read the full article](https://mp.weixin.qq.com/s/e36oemQkHoSTUpCrcEqFOg) |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,20 @@ | ||
+++ | ||
title = 'Minimalist Team Leader, First Edition (2017)' | ||
date = 2024-10-20 | ||
|
||
+++ | ||
|
||
This booklet was quickly completed by the author in August 2017, intended as a quick-start guide for "senior" interns and doctoral students preparing to lead others. It also serves as a summary of the author's five years of experience in team management. | ||
|
||
[Read the full article](https://mp.weixin.qq.com/s/8sq0rTfkcGQjJU2ZfnjzQw) | ||
|
||
|
||
|
||
|
||
|
||
|
||
|
||
|
||
|
||
|
||
|
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,11 @@ | ||
+++ | ||
title = 'Progress in the RISC-V + AI Ecosystem: The llama.cpp Optimization for RVV 1.0 is Complete, Achieving up to a 350% Performance Boost! The Code is Open-Source—Feel Free to Replicate and Explore' | ||
date = 2024-10-26 | ||
|
||
+++ | ||
|
||
llama.cpp is a large language model (LLM) inference framework implemented entirely in C/C++, relying heavily on its derivative project, ggml, for tensor operations, which requires high computational performance. | ||
|
||
Recently, **xctan**, an intern at the PLCT Lab, enhanced ggml’s Q4_0_8_8 quantized matrix multiplication operator by adding optimized support for RISC-V Vector 1.0, achieving significant performance improvements. | ||
|
||
[Read the full article](https://mp.weixin.qq.com/s/e36oemQkHoSTUpCrcEqFOg) |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,20 @@ | ||
+++ | ||
title = 'Minimalist Team Leader, First Edition (2017)' | ||
date = 2024-10-20 | ||
|
||
+++ | ||
|
||
This booklet was quickly completed by the author in August 2017, intended as a quick-start guide for "senior" interns and doctoral students preparing to lead others. It also serves as a summary of the author's five years of experience in team management. | ||
|
||
[Read the full article](https://mp.weixin.qq.com/s/8sq0rTfkcGQjJU2ZfnjzQw) | ||
|
||
|
||
|
||
|
||
|
||
|
||
|
||
|
||
|
||
|
||
|
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,11 @@ | ||
+++ | ||
title = 'Progress in the RISC-V + AI Ecosystem: The llama.cpp Optimization for RVV 1.0 is Complete, Achieving up to a 350% Performance Boost! The Code is Open-Source—Feel Free to Replicate and Explore' | ||
date = 2024-10-26 | ||
|
||
+++ | ||
|
||
llama.cpp is a large language model (LLM) inference framework implemented entirely in C/C++, relying heavily on its derivative project, ggml, for tensor operations, which requires high computational performance. | ||
|
||
Recently, **xctan**, an intern at the PLCT Lab, enhanced ggml’s Q4_0_8_8 quantized matrix multiplication operator by adding optimized support for RISC-V Vector 1.0, achieving significant performance improvements. | ||
|
||
[Read the full article](https://mp.weixin.qq.com/s/e36oemQkHoSTUpCrcEqFOg) |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,10 @@ | ||
+++ | ||
title = '《极简项目管理》第一版(2017年)全文' | ||
date = 2024-10-20 | ||
|
||
+++ | ||
|
||
这是笔者在 2017 年 8 月一气呵成的很小的册子,写作初衷是要为即将带人的“老”实习生和“老”博士生提供一个速成指导,同时将自己过去 5 年的团队管理经验进行一次汇总。 | ||
|
||
[阅读全文](https://mp.weixin.qq.com/s/8sq0rTfkcGQjJU2ZfnjzQw) | ||
|
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,11 @@ | ||
+++ | ||
title = 'RISC-V+AI 生态进展:llama.cpp RVV 1.0 优化搞定了,最高 350% 性能提升!代码开源,欢迎复现' | ||
date = 2024-10-26 | ||
|
||
+++ | ||
|
||
llama.cpp 是使用纯 C/C++ 实现的大语言模型(LLM)推理框架,内部大量使用其衍生项目 ggml 进行张量运算,对计算性能要求较高。 | ||
|
||
近日,PLCT 实验室的实习生 **xctan** 为 ggml 的 Q4_0_8_8 量化方式的矩阵乘法算子增加了优化后的 RISC-V Vector 1.0 支持,取得了显著的性能提升。 | ||
|
||
[阅读全文](https://mp.weixin.qq.com/s/e36oemQkHoSTUpCrcEqFOg) |