What's Changed
- Fix images building by @IlyasMoutawwakil in #242
- Faster quality check by @IlyasMoutawwakil in #243
- Decode output of
nvmlDeviceGetName
to avoid JSON serialize issue by @KeitaW in #240 - Fix makefile typo by @IlyasMoutawwakil in #244
- fix neural compressor backend by @baptistecolle in #245
- Update cuda images by @IlyasMoutawwakil in #246
- Add t4 for llm perf leaderboard by @baptistecolle in #238
- add optimum-intel ipex backend into benchmark by @yao-matrix in #250
- WIP fix rocm runners by @baptistecolle in #249
- Code Style by @IlyasMoutawwakil in #254
- Update ROCm by @IlyasMoutawwakil in #253
- Build from source quantization packages by @baptistecolle in #239
- Fix py-txi ci by @IlyasMoutawwakil in #255
- Fix API tests on ROCm by @IlyasMoutawwakil in #256
-
- refine cpu Dockerfile for better performance 2. add ipex_bert example by @yao-matrix in #257
- Add support for intel in leaderboard by @baptistecolle in #248
- fix broken canonical list by @baptistecolle in #262
- Fix broken canonical list by @baptistecolle in #264
- Fix issue with CodeCarbon lock by @regisss in #265
- Set is_distributed false by default in vllm by @asesorov in #266
- fix broken cuda and rocm images by @baptistecolle in #263
- Styling by @IlyasMoutawwakil in #267
- Labeling system in CI by @IlyasMoutawwakil in #268
- Multi-gpu vllm by @IlyasMoutawwakil in #269
- fix multi gpu ipc by @IlyasMoutawwakil in #270
- Allow multiple runs and handle connection communication errors by @IlyasMoutawwakil in #271
- Removing barriers by @IlyasMoutawwakil in #273
- Update readme with IPEX by @IlyasMoutawwakil in #274
- Distributed trt-llm by @IlyasMoutawwakil in #275
- ipex backend enhancements by @yao-matrix in #272
- Bump version by @IlyasMoutawwakil in #278
- Pass backend name to EnergyTracker in Training scenario by @asesorov in #279
- move to new runners by @glegendre01 in #281
- Markdown Report by @IlyasMoutawwakil in #280
- dev version by @IlyasMoutawwakil in #284
- Using intermediate env vars in CI by @IlyasMoutawwakil in #290
- remove old code linked to llm-perf leaderboard by @baptistecolle in #291
- Image Text To Text Support by @IlyasMoutawwakil in #296
- Feat: reimplement vllm backend beam search using logprobs by @vicoooo26 in #293
- Add the logic for Energy Star by @regisss in #261
- Add torchao to optimum as a pytorch backend configuration by @jerryzh168 in #297
- fix llamacpp and windows libuv by @IlyasMoutawwakil in #298
- Remove DP vs TP distinction and simplify aggregation across processes by @IlyasMoutawwakil in #299
- Fix misc test by @IlyasMoutawwakil in #300
- Adding latency and memory to energy star by @IlyasMoutawwakil in #302
- Add per_step diffusion measurments by @IlyasMoutawwakil in #303
- Fixes by @IlyasMoutawwakil in #304
- Secure Instinct CI by @IlyasMoutawwakil in #301
- Remove non maintained backends (llm-swarm, inc) by @IlyasMoutawwakil in #305
- Test examples by @IlyasMoutawwakil in #306
- Optional backend kwargs by @IlyasMoutawwakil in #307
- Fix trt llm by @IlyasMoutawwakil in #308
- Protect hf token by @IlyasMoutawwakil in #309
- Preparing for version 0.5 and checking CI by @IlyasMoutawwakil in #310
New Contributors
- @KeitaW made their first contribution in #240
- @yao-matrix made their first contribution in #250
- @asesorov made their first contribution in #266
- @glegendre01 made their first contribution in #281
- @vicoooo26 made their first contribution in #293
- @jerryzh168 made their first contribution in #297
Full Changelog: v0.4.0...v0.5.0