Skip to content
This repository has been archived by the owner on Jun 24, 2024. It is now read-only.

chore: update rustix 0.38 #441

Merged
merged 1 commit into from
Nov 12, 2023
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
70 changes: 25 additions & 45 deletions Cargo.lock

Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.

11 changes: 8 additions & 3 deletions Cargo.toml
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@ members = [
"crates/llm",
"crates/llm-base",
"crates/models/*",
"binaries/*"
"binaries/*",
]
resolver = "2"
default-members = ["binaries/llm-cli", "crates/llm"]
Expand All @@ -27,7 +27,7 @@ anyhow = "1.0"
rustyline = { version = "11.0.0", features = ["derive"] }
serde = { version = "1.0", features = ["derive"] }
serde_json = { version = "1.0" }
spinoff = { version = "0.7.0", default-features = false, features = ["dots2"] }
spinoff = { version = "0.8.0", default-features = false, features = ["dots2"] }
clap = { version = "4.1.8", features = ["derive"] }
memmap2 = "0.5.10"
tracing-subscriber = { version = "0.3", features = ["env-filter"] }
Expand All @@ -45,7 +45,12 @@ ci = ["github"]
# The installers to generate for each app
installers = ["shell", "powershell"]
# Target platforms to build apps for (Rust target-triple syntax)
targets = ["x86_64-unknown-linux-gnu", "x86_64-apple-darwin", "x86_64-pc-windows-msvc", "aarch64-apple-darwin"]
targets = [
"x86_64-unknown-linux-gnu",
"x86_64-apple-darwin",
"x86_64-pc-windows-msvc",
"aarch64-apple-darwin",
]

# The profile that 'cargo dist' will build with
[profile.dist]
Expand Down
6 changes: 3 additions & 3 deletions binaries/llm-cli/src/cli_args.rs
Original file line number Diff line number Diff line change
Expand Up @@ -542,7 +542,7 @@ impl ModelLoad {
let tokenizer_source = match self.model_and_tokenizer.to_source() {
Ok(vs) => vs,
Err(err) => {
if let Some(sp) = sp.take() {
if let Some(mut sp) = sp.take() {
sp.fail(&format!("Failed to load tokenizer: {}", err));
}
return Err(err);
Expand Down Expand Up @@ -595,7 +595,7 @@ impl ModelLoad {
file_size,
tensor_count,
} => {
if let Some(sp) = sp.take() {
if let Some(mut sp) = sp.take() {
sp.success(&format!(
"Loaded {tensor_count} tensors ({}) after {}ms",
bytesize::to_string(file_size, false),
Expand All @@ -610,7 +610,7 @@ impl ModelLoad {
if model.is_err() {
// If we've failed at loading the model, we probably haven't stopped the spinner yet.
// Cancel it now if needed.
if let Some(sp) = sp {
if let Some(mut sp) = sp {
sp.fail("Failed to load model")
}
}
Expand Down
2 changes: 1 addition & 1 deletion binaries/llm-cli/src/interactive.rs
Original file line number Diff line number Diff line change
Expand Up @@ -141,7 +141,7 @@ fn feed_prompt_with_spinner(
prompt.insert(0, '\n');
}

let sp = spinoff::Spinner::new(spinoff::spinners::Dots2, "".to_string(), None);
let mut sp = spinoff::Spinner::new(spinoff::spinners::Dots2, "".to_string(), None);
let result = session.feed_prompt(
model,
&prompt,
Expand Down
Loading