mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-09-11 20:34:54 +00:00
Revert "Removing params test (seems flaky in CI ?)"
This reverts commit b93d4ec604
.
This commit is contained in:
parent
b93d4ec604
commit
43796ce4f9
@ -38,27 +38,27 @@ async def test_flash_idefics2_next_simple(flash_idefics2_next, response_snapshot
|
|||||||
assert response == response_snapshot
|
assert response == response_snapshot
|
||||||
|
|
||||||
|
|
||||||
# @pytest.mark.asyncio
|
@pytest.mark.asyncio
|
||||||
# @pytest.mark.private
|
@pytest.mark.private
|
||||||
# async def test_flash_idefics2_next_all_params(flash_idefics2_next, response_snapshot):
|
async def test_flash_idefics2_next_all_params(flash_idefics2_next, response_snapshot):
|
||||||
# response = await flash_idefics2_next.generate(
|
response = await flash_idefics2_next.generate(
|
||||||
# "Test request",
|
"Test request",
|
||||||
# max_new_tokens=10,
|
max_new_tokens=10,
|
||||||
# repetition_penalty=1.2,
|
repetition_penalty=1.2,
|
||||||
# return_full_text=True,
|
return_full_text=True,
|
||||||
# stop_sequences=["test"],
|
stop_sequences=["test"],
|
||||||
# temperature=0.5,
|
temperature=0.5,
|
||||||
# top_p=0.9,
|
top_p=0.9,
|
||||||
# top_k=10,
|
top_k=10,
|
||||||
# truncate=5,
|
truncate=5,
|
||||||
# typical_p=0.9,
|
typical_p=0.9,
|
||||||
# watermark=True,
|
watermark=True,
|
||||||
# decoder_input_details=True,
|
decoder_input_details=True,
|
||||||
# seed=0,
|
seed=0,
|
||||||
# )
|
)
|
||||||
#
|
|
||||||
# assert response.details.generated_tokens == 10
|
assert response.details.generated_tokens == 10
|
||||||
# assert response == response_snapshot
|
assert response == response_snapshot
|
||||||
|
|
||||||
|
|
||||||
@pytest.mark.asyncio
|
@pytest.mark.asyncio
|
||||||
|
@ -510,21 +510,17 @@ fn fetch_image(input: &str) -> Result<(String, usize, usize), ValidationError> {
|
|||||||
/// Get input length and optionally truncate it
|
/// Get input length and optionally truncate it
|
||||||
fn prepare_input(
|
fn prepare_input(
|
||||||
mut inputs: String,
|
mut inputs: String,
|
||||||
truncate: Option<usize>,
|
_truncate: Option<usize>,
|
||||||
tokenizer: &Tokenizer,
|
tokenizer: &Tokenizer,
|
||||||
config: &Option<Config>,
|
config: &Option<Config>,
|
||||||
) -> Result<(tokenizers::Encoding, String), ValidationError> {
|
) -> Result<(tokenizers::Encoding, String), ValidationError> {
|
||||||
static RE: Lazy<Regex> = Lazy::new(|| Regex::new(r"!\[\]\([^\)]*\)").unwrap());
|
static RE: Lazy<Regex> = Lazy::new(|| Regex::new(r"!\[\]\([^\)]*\)").unwrap());
|
||||||
tracing::info!("Truncate {truncate:?}");
|
|
||||||
let tokenizer_query = match config {
|
let tokenizer_query = match config {
|
||||||
Some(Config::LlavaNext(config)) => {
|
Some(Config::LlavaNext(config)) => {
|
||||||
let mut modified_inputs = String::with_capacity(inputs.len());
|
let mut modified_inputs = String::with_capacity(inputs.len());
|
||||||
let mut tokenizer_query = String::with_capacity(inputs.len());
|
let mut tokenizer_query = String::with_capacity(inputs.len());
|
||||||
let mut start = 0;
|
let mut start = 0;
|
||||||
for chunk in RE.find_iter(&inputs) {
|
for chunk in RE.find_iter(&inputs) {
|
||||||
if let Some(truncate) = truncate {
|
|
||||||
return Err(ValidationError::TruncateImage(truncate));
|
|
||||||
}
|
|
||||||
let chunk_start = chunk.start();
|
let chunk_start = chunk.start();
|
||||||
let chunk_end = chunk.end();
|
let chunk_end = chunk.end();
|
||||||
if chunk_start != start {
|
if chunk_start != start {
|
||||||
@ -549,9 +545,6 @@ fn prepare_input(
|
|||||||
let mut tokenizer_query = String::with_capacity(inputs.len());
|
let mut tokenizer_query = String::with_capacity(inputs.len());
|
||||||
let mut start = 0;
|
let mut start = 0;
|
||||||
for chunk in RE.find_iter(&inputs) {
|
for chunk in RE.find_iter(&inputs) {
|
||||||
if let Some(truncate) = truncate {
|
|
||||||
return Err(ValidationError::TruncateImage(truncate));
|
|
||||||
}
|
|
||||||
let chunk_start = chunk.start();
|
let chunk_start = chunk.start();
|
||||||
let chunk_end = chunk.end();
|
let chunk_end = chunk.end();
|
||||||
if chunk_start != start {
|
if chunk_start != start {
|
||||||
@ -688,10 +681,6 @@ pub enum ValidationError {
|
|||||||
InvalidImageContent(String),
|
InvalidImageContent(String),
|
||||||
#[error("Could not fetch image: {0}")]
|
#[error("Could not fetch image: {0}")]
|
||||||
FailedFetchImage(#[from] reqwest::Error),
|
FailedFetchImage(#[from] reqwest::Error),
|
||||||
#[error(
|
|
||||||
"`truncate` cannot be used with VLM and images as it is truncating the image in the middle"
|
|
||||||
)]
|
|
||||||
TruncateImage(usize),
|
|
||||||
}
|
}
|
||||||
|
|
||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
|
Loading…
Reference in New Issue
Block a user