@@ -544,6 +544,10 @@ static std::set<llm_tensor> llm_get_tensor_names(llm_arch arch) {
544544 case LLM_ARCH_CLIP:
545545 return {};
546546 case LLM_ARCH_LLAMA:
547+ case LLM_ARCH_REFACT:
548+ case LLM_ARCH_MINICPM:
549+ case LLM_ARCH_GRANITE:
550+ case LLM_ARCH_GRANITE_MOE:
547551 case LLM_ARCH_DECI:
548552 case LLM_ARCH_MISTRAL3:
549553 case LLM_ARCH_LLAMA_EMBED:
@@ -744,11 +748,9 @@ static std::set<llm_tensor> llm_get_tensor_names(llm_arch arch) {
744748 LLM_TENSOR_ATTN_Q_NORM,
745749 LLM_TENSOR_ATTN_K_NORM,
746750 };
747- case LLM_ARCH_REFACT:
748751 case LLM_ARCH_QWEN2:
749752 case LLM_ARCH_QWEN2VL:
750753 case LLM_ARCH_INTERNLM2:
751- case LLM_ARCH_GRANITE:
752754 case LLM_ARCH_ERNIE4_5:
753755 case LLM_ARCH_PADDLEOCR:
754756 case LLM_ARCH_SMOLLM3:
@@ -759,6 +761,7 @@ static std::set<llm_tensor> llm_get_tensor_names(llm_arch arch) {
759761 LLM_TENSOR_TOKEN_EMBD,
760762 LLM_TENSOR_OUTPUT_NORM,
761763 LLM_TENSOR_OUTPUT,
764+ LLM_TENSOR_ROPE_FREQS,
762765 LLM_TENSOR_ATTN_NORM,
763766 LLM_TENSOR_ATTN_Q,
764767 LLM_TENSOR_ATTN_K,
@@ -1232,29 +1235,6 @@ static std::set<llm_tensor> llm_get_tensor_names(llm_arch arch) {
12321235 LLM_TENSOR_FFN_DOWN,
12331236 LLM_TENSOR_FFN_UP,
12341237 };
1235- case LLM_ARCH_MINICPM:
1236- return {
1237- LLM_TENSOR_TOKEN_EMBD,
1238- LLM_TENSOR_OUTPUT_NORM,
1239- LLM_TENSOR_OUTPUT,
1240- LLM_TENSOR_ROPE_FREQS,
1241- LLM_TENSOR_ROPE_FACTORS_LONG,
1242- LLM_TENSOR_ROPE_FACTORS_SHORT,
1243- LLM_TENSOR_ATTN_NORM,
1244- LLM_TENSOR_ATTN_Q,
1245- LLM_TENSOR_ATTN_K,
1246- LLM_TENSOR_ATTN_V,
1247- LLM_TENSOR_ATTN_OUT,
1248- LLM_TENSOR_ATTN_ROT_EMBD,
1249- LLM_TENSOR_FFN_GATE_INP,
1250- LLM_TENSOR_FFN_NORM,
1251- LLM_TENSOR_FFN_GATE,
1252- LLM_TENSOR_FFN_DOWN,
1253- LLM_TENSOR_FFN_UP,
1254- LLM_TENSOR_FFN_GATE_EXP,
1255- LLM_TENSOR_FFN_DOWN_EXP,
1256- LLM_TENSOR_FFN_UP_EXP,
1257- };
12581238 case LLM_ARCH_MINICPM3:
12591239 return {
12601240 LLM_TENSOR_TOKEN_EMBD,
@@ -1442,6 +1422,7 @@ static std::set<llm_tensor> llm_get_tensor_names(llm_arch arch) {
14421422 LLM_TENSOR_TOKEN_EMBD,
14431423 LLM_TENSOR_OUTPUT,
14441424 LLM_TENSOR_OUTPUT_NORM,
1425+ LLM_TENSOR_ROPE_FREQS,
14451426 LLM_TENSOR_ATTN_NORM,
14461427 LLM_TENSOR_ATTN_Q,
14471428 LLM_TENSOR_ATTN_K,
@@ -1657,7 +1638,9 @@ static std::set<llm_tensor> llm_get_tensor_names(llm_arch arch) {
16571638 LLM_TENSOR_ROPE_FREQS,
16581639 LLM_TENSOR_OUTPUT_NORM,
16591640 LLM_TENSOR_OUTPUT,
1641+ LLM_TENSOR_TOKEN_EMBD,
16601642 LLM_TENSOR_ATTN_NORM,
1643+ LLM_TENSOR_ATTN_QKV,
16611644 LLM_TENSOR_ATTN_Q,
16621645 LLM_TENSOR_ATTN_K,
16631646 LLM_TENSOR_ATTN_V,
@@ -2061,30 +2044,12 @@ static std::set<llm_tensor> llm_get_tensor_names(llm_arch arch) {
20612044 LLM_TENSOR_FFN_DOWN,
20622045 LLM_TENSOR_FFN_UP,
20632046 };
2064- case LLM_ARCH_GRANITE_MOE:
2065- return {
2066- LLM_TENSOR_TOKEN_EMBD,
2067- LLM_TENSOR_OUTPUT_NORM,
2068- LLM_TENSOR_OUTPUT,
2069- LLM_TENSOR_ATTN_NORM,
2070- LLM_TENSOR_ATTN_Q,
2071- LLM_TENSOR_ATTN_K,
2072- LLM_TENSOR_ATTN_V,
2073- LLM_TENSOR_ATTN_OUT,
2074- LLM_TENSOR_FFN_NORM,
2075- LLM_TENSOR_FFN_GATE_INP,
2076- LLM_TENSOR_FFN_GATE_EXPS,
2077- LLM_TENSOR_FFN_DOWN_EXPS,
2078- LLM_TENSOR_FFN_UP_EXPS,
2079- LLM_TENSOR_FFN_GATE_SHEXP,
2080- LLM_TENSOR_FFN_DOWN_SHEXP,
2081- LLM_TENSOR_FFN_UP_SHEXP,
2082- };
20832047 case LLM_ARCH_GRANITE_HYBRID:
20842048 return {
20852049 LLM_TENSOR_TOKEN_EMBD,
20862050 LLM_TENSOR_OUTPUT_NORM,
20872051 LLM_TENSOR_OUTPUT,
2052+ LLM_TENSOR_ROPE_FREQS,
20882053 LLM_TENSOR_ATTN_NORM,
20892054 LLM_TENSOR_SSM_IN,
20902055 LLM_TENSOR_SSM_CONV1D,
@@ -2412,6 +2377,7 @@ static std::set<llm_tensor> llm_get_tensor_names(llm_arch arch) {
24122377 LLM_TENSOR_TOKEN_EMBD,
24132378 LLM_TENSOR_OUTPUT_NORM,
24142379 LLM_TENSOR_OUTPUT,
2380+ LLM_TENSOR_ROPE_FREQS,
24152381 LLM_TENSOR_ATTN_NORM,
24162382 LLM_TENSOR_ATTN_QKV,
24172383 LLM_TENSOR_ATTN_OUT,
@@ -2789,7 +2755,12 @@ std::string LLM_TN_IMPL::str() const {
27892755 }
27902756
27912757 if (model_tensors.find (tensor) == model_tensors.end ()) {
2792- return LLM_TENSOR_NAMES.at (tensor);
2758+ const char * name = LLM_TENSOR_NAMES.at (tensor);
2759+ if (suffix != nullptr || bid != -1 || xid != -1 ) {
2760+ LLAMA_LOG_WARN (" %s: cannot properly format tensor name %s with suffix=%s bid=%d xid=%d\n " ,
2761+ __func__, name, suffix, bid, xid);
2762+ }
2763+ return name;
27932764 }
27942765
27952766 std::string name = ::format (LLM_TENSOR_NAMES.at (tensor), bid, xid);
0 commit comments