Default Branch

1ff9d185d5 · Remove useless packages (#3253) · Updated 2025-06-03 11:42:29 +00:00

Branches

45a383bda4 · fix(neuron): adjust test expectations for llama on nxd · Updated 2025-06-04 14:41:27 +00:00

2
10

30bdf922bd · feat: improve llava next pooling for granite vision · Updated 2025-06-04 13:50:39 +00:00

0
1

8e41da951d · Release 3.3.2 · Updated 2025-05-30 14:19:18 +00:00

0
1

55cdfbfce3 · enable multi-card test · Updated 2025-05-22 14:13:59 +00:00

2
16

767a65202d · Release 3.3.1 · Updated 2025-05-22 07:47:12 +00:00

3
1

8275bdcfe9 · Fix? · Updated 2025-05-18 22:04:41 +00:00

10
2

3245b8972a · Merge branch 'main' into add_logs_gaudi_warmup · Updated 2025-05-11 09:59:20 +00:00

13
4

03a8b8d751 · Release 3.3.0 · Updated 2025-05-09 13:53:38 +00:00

14
1

6c25a98b49 · Prepare for 3.2.4 · Updated 2025-05-09 09:53:25 +00:00

15
1

551ee3a365 · fix: linter · Updated 2025-05-06 00:03:17 +00:00

19
14

338cdc2eb8 · Tiny fix. · Updated 2025-04-30 15:50:14 +00:00

150
9

70c616ca27 · feat: lock updated kernel versions · Updated 2025-04-29 15:05:41 +00:00

29
1

7237e8e6bf · update pixel_values · Updated 2025-04-19 17:12:23 +00:00

31
6

2a10a28d08 · force attn to flashdecoding · Updated 2025-04-11 15:24:12 +00:00

40
5

3d71c06aff · flashinfer: head_dim -> head_dim_qk · Updated 2025-04-11 12:38:17 +00:00

40
2

73d0876f12 · Fixing the updating logic of backends. · Updated 2025-04-10 09:04:03 +00:00

152
11

d93ad244a3 · add attn · Updated 2025-04-09 16:37:34 +00:00

41
1

e5618d6e40 · add chunked attn support · Updated 2025-04-09 16:36:06 +00:00

41
1

a1f3ebe17c · Release 3.2.3 · Updated 2025-04-08 08:17:51 +00:00

42
1

c67546fd40 · Release 3.2.2 · Updated 2025-04-06 09:40:52 +00:00

47
1