Fix: Mirostat fails on models split across multiple GPUs
This commit is contained in:
parent
23055b21ee
commit
9dcb37e8d4
1 changed files with 1 additions and 1 deletions
|
@ -104,7 +104,7 @@ class MirostatLogitsWarper(LogitsWarper):
|
|||
break
|
||||
|
||||
# Normalize the probabilities of the remaining words
|
||||
prob_topk = torch.softmax(sorted_logits, dim=0)
|
||||
prob_topk = torch.softmax(sorted_logits, dim=0).to('cuda')
|
||||
|
||||
prev_i = torch.multinomial(prob_topk, num_samples=1, replacement=True).to('cuda')
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue