Skip to content

Commit 3f542b7

Browse files
committed
remove comments
1 parent 7b1d172 commit 3f542b7

2 files changed

Lines changed: 0 additions & 32 deletions

File tree

src/main/java/org/beehive/gpullama3/tornadovm/layers/type/fp16/LlamaFP16FFNLayers.java

Lines changed: 0 additions & 16 deletions
Original file line numberDiff line numberDiff line change
@@ -50,7 +50,6 @@ public GridScheduler updateGridScheduler(GridScheduler tornadoForwardScheduler)
5050
for (int i = 0; i < config.numberOfLayers(); i++) {
5151
// === Attention Block ===
5252
tornadoForwardScheduler.addWorkerGrid("layer_" + i + ".attn_rms_reduce", rmsNormWorker);
53-
//tornadoForwardScheduler.addWorkerGrid("layer_" + i + ".attn_rms_apply_fp16", rmsNormWorker);
5453
tornadoForwardScheduler.addWorkerGrid("layer_" + i + ".qkv_projection", fusedQKVWorker);
5554
tornadoForwardScheduler.addWorkerGrid("layer_" + i + ".rope_and_kv_cache", ropeWithCacheWorker);
5655
tornadoForwardScheduler.addWorkerGrid("layer_" + i + ".attention", parallelAttentionWorker);
@@ -202,21 +201,6 @@ TaskGraph setupSingleFFNLayer(LlamaTornadoWeights weights, Configuration config,
202201
TransformerComputeKernelsLayered::reductionOneBlockWithLayerFuseFP16,
203202
context, state.wrapXbFP16, state.wrapX, weights.rms_att_weightLayered[layerIndex].asFloatArray(), state.temp,
204203
config.dim(), config.rmsNormEps(), state.localSize);
205-
/*unifiedLayer.task("attn_rms_reduce",
206-
TransformerComputeKernelsLayered::reductionOneBlockWithLayer,
207-
context, state.temp, state.wrapX,
208-
config.dim(), config.rmsNormEps(), state.localSize);
209-
210-
if (shouldUseFinalNormalization()) {
211-
unifiedLayer.task("attn_rms_finalize",
212-
TransformerComputeKernelsLayered::reductionFinalNormalization,
213-
context, state.temp, config.dim(), config.rmsNormEps());
214-
}
215-
216-
unifiedLayer.task("attn_rms_apply_fp16",
217-
TransformerComputeKernels::mapContextWithQuantize,
218-
context, state.wrapXbFP16, state.wrapX,
219-
weights.rms_att_weightLayered[layerIndex].asFloatArray(), state.temp);*/
220204

221205
// QKV Projection (fused)
222206
unifiedLayer.task("qkv_projection",

src/main/java/org/beehive/gpullama3/tornadovm/layers/type/q8_0/LlamaQ8_0FFNLayers.java

Lines changed: 0 additions & 16 deletions
Original file line numberDiff line numberDiff line change
@@ -164,21 +164,6 @@ TaskGraph setupSingleFFNLayer(LlamaTornadoWeights weights, Configuration config,
164164
TransformerComputeKernelsLayered::reductionOneBlockWithLayerFuse,
165165
context, state.wrapXb, state.wrapX, weights.rms_att_weightLayered[layerIndex].asFloatArray(), state.temp,
166166
config.dim(), config.rmsNormEps(), state.localSize);
167-
/*unifiedLayer.task("attn_rms_reduce",
168-
TransformerComputeKernelsLayered::reductionOneBlockWithLayer,
169-
context, state.temp, state.wrapX,
170-
config.dim(), config.rmsNormEps(), state.localSize);
171-
172-
if (shouldUseFinalNormalization()) {
173-
unifiedLayer.task("attn_rms_finalize",
174-
TransformerComputeKernelsLayered::reductionFinalNormalization,
175-
context, state.temp, config.dim(), config.rmsNormEps());
176-
}
177-
178-
unifiedLayer.task("attn_rms_apply",
179-
TransformerComputeKernelsLayered::reductionOneBlock2WithLayer,
180-
context, state.wrapXb, state.wrapX,
181-
weights.rms_att_weightLayered[layerIndex].asFloatArray(), state.temp);*/
182167

183168
// QKV Projection (fused with Q8 dequantization)
184169
unifiedLayer.task("qkv_projection",
@@ -310,7 +295,6 @@ public GridScheduler updateGridScheduler(GridScheduler tornadoForwardScheduler)
310295
// --- Attention Block ---
311296
// RMS Normalization
312297
tornadoForwardScheduler.addWorkerGrid("layer_" + i + ".attn_rms_reduce", rmsNormWorker);
313-
//tornadoForwardScheduler.addWorkerGrid("layer_" + i + ".attn_rms_apply", rmsNormWorker);
314298
tornadoForwardScheduler.addWorkerGrid("layer_" + i + ".qkv_projection", fusedQkvWorker);
315299
tornadoForwardScheduler.addWorkerGrid("layer_" + i + ".rope_and_kv_cache", ropeWithCacheWorker);
316300
tornadoForwardScheduler.addWorkerGrid("layer_" + i + ".attention", parallelAttentionWorker);

0 commit comments

Comments
 (0)