Skip to content

Commit

Permalink
Greatly improved LLaMA sampling defaults. (#2)
Browse files Browse the repository at this point in the history
By default `drama_llama` was using greedy sampling with no repetition
penalty. This was a mistake in the implementation of `Default` for
various settings structs. The default has now been changed to locally
typical sampling with a minor repetition penalty. Quality of generation
should be greatly improved.

Additionally, `llama.cpp` has been updated. This means any models will
need to be updated since the tokenizer code has changed. The user will
be warned in the terminal if that is the case.
  • Loading branch information
mdegans authored May 24, 2024
1 parent 0246432 commit 6b36ea2
Show file tree
Hide file tree
Showing 2 changed files with 42 additions and 43 deletions.
83 changes: 41 additions & 42 deletions Cargo.lock

Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.

2 changes: 1 addition & 1 deletion Cargo.toml
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
[package]
name = "weave-writer"
version = "0.0.0"
version = "0.0.1"
edition = "2021"
description = "A tool for collaborative generative writing."
license-file = "LICENSE.md"
Expand Down

0 comments on commit 6b36ea2

Please sign in to comment.