Skip to content

Commit 2b1e5ea

Browse files
committed
convert-hf: add missing ftype
1 parent d7e199e commit 2b1e5ea

File tree

2 files changed

+10
-1
lines changed

2 files changed

+10
-1
lines changed

convert-hf-to-gguf.py

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1202,6 +1202,7 @@ def set_gguf_parameters(self):
12021202
self.gguf_writer.add_head_count_kv(hparams["num_key_value_heads"])
12031203
self.gguf_writer.add_parallel_residual(hparams["use_parallel_residual"] if "use_parallel_residual" in hparams else True)
12041204
self.gguf_writer.add_layer_norm_eps(self.find_hparam(["layer_norm_eps", "norm_eps"]))
1205+
self.gguf_writer.add_file_type(self.ftype)
12051206

12061207
_q_norms: list[dict[str, Tensor]] | None = None
12071208
_k_norms: list[dict[str, Tensor]] | None = None
@@ -1578,6 +1579,7 @@ def set_gguf_parameters(self):
15781579
self.gguf_writer.add_rope_dimension_count(self.hparams["hidden_size"] // self.hparams["num_attention_heads"])
15791580
self.gguf_writer.add_head_count(self.hparams["num_attention_heads"])
15801581
self.gguf_writer.add_layer_norm_rms_eps(self.hparams["layer_norm_epsilon"])
1582+
self.gguf_writer.add_file_type(self.ftype)
15811583

15821584

15831585
@Model.register("Qwen2ForCausalLM")
@@ -1815,6 +1817,7 @@ def set_gguf_parameters(self):
18151817
self.gguf_writer.add_head_count(hparams["num_attention_heads"])
18161818
self.gguf_writer.add_head_count_kv(5) # hparams["num_key_value_heads"]) is wrong
18171819
self.gguf_writer.add_layer_norm_rms_eps(hparams["rms_norm_eps"])
1820+
self.gguf_writer.add_file_type(self.ftype)
18181821

18191822
def shuffle_attn_q_weight(self, data_torch):
18201823
assert data_torch.size() == (5120, 5120)
@@ -1994,6 +1997,7 @@ def set_gguf_parameters(self):
19941997
self.gguf_writer.add_head_count(self.hparams["num_attention_heads"])
19951998
self.gguf_writer.add_layer_norm_rms_eps(self.hparams["rms_norm_eps"])
19961999
self.gguf_writer.add_head_count_kv(self.hparams["num_key_value_heads"])
2000+
self.gguf_writer.add_file_type(self.ftype)
19972001

19982002
def modify_tensors(self, data_torch: Tensor, name: str, bid: int | None) -> Iterable[tuple[str, Tensor]]:
19992003
num_heads = self.hparams["num_attention_heads"]

gguf-py/gguf/lazy.py

Lines changed: 6 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -174,7 +174,12 @@ def already_eager_to_eager(_t: LazyBase) -> Any:
174174
while _t._data is None:
175175
lt = _t._lazy.popleft()
176176
if lt._data is not None:
177-
raise ValueError(f"{lt} did not belong in the lazy queue")
177+
# Lazy tensor did not belong in the lazy queue.
178+
# Weirdly only happens with Bloom models...
179+
# likely because tensors aren't unique in the queue.
180+
# The final output is still the same as in eager mode,
181+
# so it's safe to ignore this.
182+
continue
178183
assert lt._func is not None
179184
lt._args = cls._recurse_apply(lt._args, already_eager_to_eager)
180185
lt._data = lt._func(lt._args)

0 commit comments

Comments
 (0)