Skip to content

Commit 04d450b

Browse files
committed
style: cleanup
Signed-off-by: YdrMaster <ydrml@hotmail.com>
1 parent c555ba4 commit 04d450b

File tree

5 files changed

+10
-56
lines changed

5 files changed

+10
-56
lines changed

models/llama/common-cpu/src/lib.rs

Lines changed: 8 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,10 @@ use common_cpu::{
44
tensor::{reslice, slice, udim, Tensor},
55
CpuKernels, Kernels, KernelsA, KernelsB, ThisThread,
66
};
7-
use llama::{ComputeConst, ComputeStream, Handle, LayerStorage, QueueOf, SliceOn, Storage, Weight};
7+
use llama::{
8+
ComputeConst, ComputeStream, Handle, InferenceConfig, LayerStorage, QueueOf, SliceOn, Storage,
9+
Weight,
10+
};
811
use std::{iter::repeat, ops::Deref, path::Path, slice::from_raw_parts};
912

1013
pub struct Transformer {
@@ -133,12 +136,10 @@ impl CausalLM for Transformer {
133136
}
134137
#[inline]
135138
fn duplicate_cache(&self, cache: &Tensor<Self::Storage>, pos: upos) -> Tensor<Self::Storage> {
136-
self.s
137-
.config
138-
.duplicate_cache(cache, pos, Blob::new, |dst, src| {
139-
src.map_physical(|u| &**u)
140-
.reform_to(&mut dst.map_physical(|u| &mut **u))
141-
})
139+
InferenceConfig::duplicate_cache(cache, pos, Blob::new, |dst, src| {
140+
src.map_physical(|u| &**u)
141+
.reform_to(&mut dst.map_physical(|u| &mut **u))
142+
})
142143
}
143144

144145
fn token_embed(&self, queries: impl IntoIterator<Item = utok>) -> Tensor<Self::Storage> {

models/llama/common/src/lib.rs

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -84,7 +84,6 @@ impl InferenceConfig {
8484
}
8585

8686
pub fn duplicate_cache<S>(
87-
&self,
8887
cache: &Tensor<S>,
8988
pos: upos,
9089
malloc: impl FnOnce(usize) -> S,

models/llama/nvidia-gpu-distributed/src/lib.rs

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -143,7 +143,7 @@ impl CausalLM for Transformer {
143143

144144
fn duplicate_cache(&self, cache: &Tensor<Self::Storage>, pos: upos) -> Tensor<Self::Storage> {
145145
let contexts = Arc::new(self.comms.contexts().collect::<Vec<_>>());
146-
self.config.duplicate_cache(
146+
InferenceConfig::duplicate_cache(
147147
cache,
148148
pos,
149149
|len| Cache {

models/llama/nvidia-gpu/src/lib.rs

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -175,7 +175,7 @@ impl CausalLM for Transformer {
175175
}
176176

177177
fn duplicate_cache(&self, cache: &Tensor<Self::Storage>, pos: upos) -> Tensor<Self::Storage> {
178-
self.0.config.duplicate_cache(
178+
InferenceConfig::duplicate_cache(
179179
cache,
180180
pos,
181181
|len| self.cache(len),

tensor/src/compatibility.rs

Lines changed: 0 additions & 46 deletions
This file was deleted.

0 commit comments

Comments
 (0)