mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-09-11 04:14:52 +00:00
Remove special, it's not correct enough (and not necessarily useful).
This commit is contained in:
parent
c12ff38974
commit
048bc5b4b7
@ -696,13 +696,13 @@ async fn tokenize(
|
||||
.get_ids()
|
||||
.iter()
|
||||
.zip(encoding.get_offsets())
|
||||
.map(|(&id, (start, stop))| {
|
||||
let text: String = input.chars().skip(*start).take(stop - start).collect();
|
||||
.map(|(&id, &(start, stop))| {
|
||||
let text: String = input.chars().skip(start).take(stop - start).collect();
|
||||
SimpleToken {
|
||||
id,
|
||||
text,
|
||||
start: *start,
|
||||
stop: *stop,
|
||||
start,
|
||||
stop,
|
||||
}
|
||||
})
|
||||
.collect();
|
||||
|
@ -365,7 +365,7 @@ fn prepare_input(
|
||||
) -> Result<(tokenizers::Encoding, String), ValidationError> {
|
||||
// Get the number of tokens in the input
|
||||
let mut encoding = tokenizer
|
||||
.encode(inputs.clone(), true)
|
||||
.encode(inputs.clone(), false)
|
||||
.map_err(|err| ValidationError::Tokenizer(err.to_string()))?;
|
||||
|
||||
// Optionally truncate
|
||||
|
Loading…
Reference in New Issue
Block a user