Skip to content

Commit

Permalink
Merge branch 'main' into dev
Browse files Browse the repository at this point in the history
  • Loading branch information
sarthaklangde committed Apr 3, 2023
2 parents e4e6e64 + 7910a7e commit 3c2a8a2
Show file tree
Hide file tree
Showing 4 changed files with 6 additions and 6 deletions.
Binary file modified .github/ui-playground.gif
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
2 changes: 1 addition & 1 deletion pyproject.toml
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
[project]
name = "xturing"
version = "0.0.4"
version = "0.0.5"
description = "Fine-tuning, evaluation and data generation for LLMs"

authors = [
Expand Down
2 changes: 1 addition & 1 deletion src/xturing/__about__.py
Original file line number Diff line number Diff line change
@@ -1 +1 @@
__version__ = "0.0.4"
__version__ = "0.0.5"
8 changes: 4 additions & 4 deletions src/xturing/ui/playground.py
Original file line number Diff line number Diff line change
Expand Up @@ -218,7 +218,7 @@ def model(history):
value=0.6,
label="Penalty alpha",
interactive=True,
info="Choose betwen 0 and 1",
info="Choose between 0 and 1",
)
penalty_alpha.release(
lambda penalty_alpha: self.set_penalty_alpha(penalty_alpha),
Expand All @@ -230,7 +230,7 @@ def model(history):
value=4,
label="Top-k",
interactive=True,
info="Choose betwen 1 and 40",
info="Choose between 1 and 40",
)
top_k.release(lambda top_k: self.set_top_k(top_k), top_k)
max_new_tokens = gr.Slider(
Expand All @@ -239,7 +239,7 @@ def model(history):
value=self.max_new_tokens,
label="Max new tokens",
interactive=True,
info="Choose betwen 1 and 512",
info="Choose between 1 and 512",
)
max_new_tokens.release(
lambda max_new_tokens: self.set_max_new_tokens(
Expand Down Expand Up @@ -283,7 +283,7 @@ def model(history):
value=self.max_new_tokens,
label="Max new tokens",
interactive=True,
info="Choose betwen 1 and 512",
info="Choose between 1 and 512",
)
max_new_tokens2.release(
lambda max_new_tokens: self.set_max_new_tokens(
Expand Down

0 comments on commit 3c2a8a2

Please sign in to comment.