mirror of
https://github.com/huggingface/text-generation-inference.git
synced 2025-04-21 23:12:07 +00:00
27 lines
907 B
Rust
27 lines
907 B
Rust
|
use std::sync::LazyLock;
|
||
|
|
||
|
pub static COMPUTE_CAPABILITY: LazyLock<Option<(usize, usize)>> =
|
||
|
LazyLock::new(get_cuda_capability);
|
||
|
|
||
|
fn get_cuda_capability() -> Option<(usize, usize)> {
|
||
|
use pyo3::prelude::*;
|
||
|
|
||
|
let py_get_capability = |py: Python| -> PyResult<(isize, isize)> {
|
||
|
let torch = py.import_bound("torch.cuda")?;
|
||
|
let get_device_capability = torch.getattr("get_device_capability")?;
|
||
|
get_device_capability.call0()?.extract()
|
||
|
};
|
||
|
|
||
|
match pyo3::Python::with_gil(py_get_capability) {
|
||
|
Ok((major, minor)) if major < 0 || minor < 0 => {
|
||
|
tracing::warn!("Ignoring negative GPU compute capabilities: {major}.{minor}");
|
||
|
None
|
||
|
}
|
||
|
Ok((major, minor)) => Some((major as usize, minor as usize)),
|
||
|
Err(err) => {
|
||
|
tracing::warn!("Cannot determine GPU compute capability: {}", err);
|
||
|
None
|
||
|
}
|
||
|
}
|
||
|
}
|