Skip to content

Commit

Permalink
added extra newline in convert_hf_to_gguf.py. Replaced a tab with spa…
Browse files Browse the repository at this point in the history
…ces in src/llama.cpp
  • Loading branch information
ymcki committed Dec 18, 2024
1 parent f77260c commit d7652ca
Show file tree
Hide file tree
Showing 2 changed files with 3 additions and 1 deletion.
2 changes: 2 additions & 0 deletions convert_hf_to_gguf.py
Original file line number Diff line number Diff line change
Expand Up @@ -1671,6 +1671,7 @@ def prepare_tensors(self):
if len(experts) > 0:
raise ValueError(f"Unprocessed experts: {experts}")


@Model.register("DeciLMForCausalLM")
class DeciModel(Model):
model_arch = gguf.MODEL_ARCH.DECI
Expand Down Expand Up @@ -1848,6 +1849,7 @@ def generate_extra_tensors(self) -> Iterable[tuple[str, Tensor]]:
def prepare_tensors(self):
super().prepare_tensors()


@Model.register("BitnetForCausalLM")
class BitnetModel(Model):
model_arch = gguf.MODEL_ARCH.BITNET
Expand Down
2 changes: 1 addition & 1 deletion src/llama.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -11195,7 +11195,7 @@ struct llm_build_context {

if (n_head == 0) { // attention-free layer of Llama-3_1-Nemotron-51B
cur = inpL;
} else {
} else {
// norm
cur = llm_build_norm(ctx0, inpL, hparams,
model.layers[il].attn_norm, NULL,
Expand Down

0 comments on commit d7652ca

Please sign in to comment.