Skip to content

Commit ad44687

Browse files
committed
Fixed line breaks and spellcheck errors
1 parent 425aaa1 commit ad44687

File tree

2 files changed

+15
-1
lines changed

2 files changed

+15
-1
lines changed

.github/actions/spelling/allow/terms.txt

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -11,6 +11,7 @@ GSoC
1111
HSF
1212
JIT'd
1313
Jacobians
14+
LLMs
1415
LLVM
1516
NVIDIA
1617
NVMe
@@ -28,6 +29,7 @@ cppyy
2829
cytokine
2930
cytokines
3031
gitlab
32+
gridlay
3133
gsoc
3234
llm
3335
linkedin

_data/contributors.yml

Lines changed: 13 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -322,7 +322,19 @@
322322
- title: "Enhancing LLM Training Efficiency with Clad for Automatic Differentiation"
323323
status: Ongoing
324324
description: |
325-
Training Large Language Models is computationally expensive, often limited by the performance limitations of Python-based frameworks. This project addresses this challenge by enhancing LLM training efficiency within a C++ environment through the integration of Clad, a Clang/LLVM compiler plugin for automatic differentiation (AD). We will develop a custom C++ tensor library specifically designed for optimal interaction with Clad. The core objective is to replace traditional runtime or manual gradient computations with Clad's efficient compile-time differentiation for key LLM operations within a GPT-2 training pipeline. This involves investigating effective strategies to bridge Clad's static analysis with dynamic neural network computations, benchmarking the resulting performance gains in speed and memory usage against a non-Clad baseline, and leveraging OpenMP for further parallelization.
325+
Training Large Language Models is computationally expensive, often
326+
limited by the performance limitations of Python-based frameworks. This
327+
project addresses this challenge by enhancing LLM training efficiency
328+
within a C++ environment through the integration of Clad, a Clang/LLVM
329+
compiler plugin for automatic differentiation (AD). We will develop a
330+
custom C++ tensor library specifically designed for optimal interaction
331+
with Clad. The core objective is to replace traditional runtime or
332+
manual gradient computations with Clad's efficient compile-time
333+
differentiation for key LLM operations within a GPT-2 training pipeline.
334+
This involves investigating effective strategies to bridge Clad's static
335+
analysis with dynamic neural network computations, benchmarking the
336+
resulting performance gains in speed and memory usage against a non-Clad
337+
baseline, and leveraging OpenMP for further parallelization.
326338
proposal: /assets/docs/Rohan_Timmaraju_Proposal_2025.pdf
327339
mentors: Vassil Vassilev, David Lange, Jonas Rembser, Christina Koutsou
328340

0 commit comments

Comments
 (0)