mirror of
https://git.datalinker.icu/vllm-project/vllm.git
synced 2026-04-30 21:50:08 +08:00
[Chore] Bump lm-eval version (#31264)
Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk>
This commit is contained in:
parent
aa3868ecfe
commit
d201807339
@ -2,7 +2,7 @@
|
|||||||
# We can use this script to compute baseline accuracy on chartqa for vllm.
|
# We can use this script to compute baseline accuracy on chartqa for vllm.
|
||||||
#
|
#
|
||||||
# Make sure you have lm-eval-harness installed:
|
# Make sure you have lm-eval-harness installed:
|
||||||
# pip install lm-eval==0.4.9
|
# pip install "lm-eval[api]>=0.4.9.2"
|
||||||
|
|
||||||
usage() {
|
usage() {
|
||||||
echo``
|
echo``
|
||||||
|
|||||||
@ -2,7 +2,7 @@
|
|||||||
# We can use this script to compute baseline accuracy on GSM for transformers.
|
# We can use this script to compute baseline accuracy on GSM for transformers.
|
||||||
#
|
#
|
||||||
# Make sure you have lm-eval-harness installed:
|
# Make sure you have lm-eval-harness installed:
|
||||||
# pip install git+https://github.com/EleutherAI/lm-evaluation-harness.git@206b7722158f58c35b7ffcd53b035fdbdda5126d#egg=lm-eval[api]
|
# pip install "lm-eval[api]>=0.4.9.2"
|
||||||
|
|
||||||
usage() {
|
usage() {
|
||||||
echo``
|
echo``
|
||||||
|
|||||||
@ -3,7 +3,7 @@
|
|||||||
# We use this for fp8, which HF does not support.
|
# We use this for fp8, which HF does not support.
|
||||||
#
|
#
|
||||||
# Make sure you have lm-eval-harness installed:
|
# Make sure you have lm-eval-harness installed:
|
||||||
# pip install git+https://github.com/EleutherAI/lm-evaluation-harness.git@206b7722158f58c35b7ffcd53b035fdbdda5126d#egg=lm-eval[api]
|
# pip install "lm-eval[api]>=0.4.9.2"
|
||||||
|
|
||||||
usage() {
|
usage() {
|
||||||
echo``
|
echo``
|
||||||
|
|||||||
@ -3,7 +3,7 @@
|
|||||||
# We use this for fp8, which HF does not support.
|
# We use this for fp8, which HF does not support.
|
||||||
#
|
#
|
||||||
# Make sure you have lm-eval-harness installed:
|
# Make sure you have lm-eval-harness installed:
|
||||||
# pip install git+https://github.com/EleutherAI/lm-evaluation-harness.git@206b7722158f58c35b7ffcd53b035fdbdda5126d#egg=lm-eval[api]
|
# pip install "lm-eval[api]>=0.4.9.2"
|
||||||
|
|
||||||
usage() {
|
usage() {
|
||||||
echo``
|
echo``
|
||||||
|
|||||||
@ -61,7 +61,7 @@ echo "Results will be stored in: $RESULTS_DIR"
|
|||||||
echo "--- Installing Python dependencies ---"
|
echo "--- Installing Python dependencies ---"
|
||||||
python3 -m pip install --progress-bar off git+https://github.com/thuml/depyf.git \
|
python3 -m pip install --progress-bar off git+https://github.com/thuml/depyf.git \
|
||||||
&& python3 -m pip install --progress-bar off pytest pytest-asyncio tpu-info \
|
&& python3 -m pip install --progress-bar off pytest pytest-asyncio tpu-info \
|
||||||
&& python3 -m pip install --progress-bar off "lm-eval @ git+https://github.com/EleutherAI/lm-evaluation-harness.git@206b7722158f58c35b7ffcd53b035fdbdda5126d" \
|
&& python3 -m pip install --progress-bar off "lm-eval[api]>=0.4.9.2" \
|
||||||
&& python3 -m pip install --progress-bar off hf-transfer tblib==3.1.0
|
&& python3 -m pip install --progress-bar off hf-transfer tblib==3.1.0
|
||||||
echo "--- Python dependencies installed ---"
|
echo "--- Python dependencies installed ---"
|
||||||
|
|
||||||
|
|||||||
@ -61,7 +61,7 @@ echo "Results will be stored in: $RESULTS_DIR"
|
|||||||
echo "--- Installing Python dependencies ---"
|
echo "--- Installing Python dependencies ---"
|
||||||
python3 -m pip install --progress-bar off git+https://github.com/thuml/depyf.git \
|
python3 -m pip install --progress-bar off git+https://github.com/thuml/depyf.git \
|
||||||
&& python3 -m pip install --progress-bar off pytest pytest-asyncio tpu-info \
|
&& python3 -m pip install --progress-bar off pytest pytest-asyncio tpu-info \
|
||||||
&& python3 -m pip install --progress-bar off "lm-eval @ git+https://github.com/EleutherAI/lm-evaluation-harness.git@206b7722158f58c35b7ffcd53b035fdbdda5126d" \
|
&& python3 -m pip install --progress-bar off "lm-eval[api]>=0.4.9.2" \
|
||||||
&& python3 -m pip install --progress-bar off hf-transfer tblib==3.1.0
|
&& python3 -m pip install --progress-bar off hf-transfer tblib==3.1.0
|
||||||
echo "--- Python dependencies installed ---"
|
echo "--- Python dependencies installed ---"
|
||||||
|
|
||||||
|
|||||||
@ -84,7 +84,7 @@ Since simple RTN does not require data for weight quantization and the activatio
|
|||||||
Install `vllm` and `lm-evaluation-harness` for evaluation:
|
Install `vllm` and `lm-evaluation-harness` for evaluation:
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
pip install vllm git+https://github.com/EleutherAI/lm-evaluation-harness.git@206b7722158f58c35b7ffcd53b035fdbdda5126d#egg=lm-eval[api]
|
pip install vllm "lm-eval[api]>=0.4.9.2"
|
||||||
```
|
```
|
||||||
|
|
||||||
Load and run the model in `vllm`:
|
Load and run the model in `vllm`:
|
||||||
|
|||||||
@ -18,7 +18,7 @@ pip install llmcompressor
|
|||||||
Additionally, install `vllm` and `lm-evaluation-harness` for evaluation:
|
Additionally, install `vllm` and `lm-evaluation-harness` for evaluation:
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
pip install vllm git+https://github.com/EleutherAI/lm-evaluation-harness.git@206b7722158f58c35b7ffcd53b035fdbdda5126d#egg=lm-eval[api]
|
pip install vllm "lm-eval[api]>=0.4.9.2"
|
||||||
```
|
```
|
||||||
|
|
||||||
## Quantization Process
|
## Quantization Process
|
||||||
|
|||||||
@ -23,7 +23,7 @@ pip install llmcompressor
|
|||||||
Additionally, install `vllm` and `lm-evaluation-harness` for evaluation:
|
Additionally, install `vllm` and `lm-evaluation-harness` for evaluation:
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
pip install vllm git+https://github.com/EleutherAI/lm-evaluation-harness.git@206b7722158f58c35b7ffcd53b035fdbdda5126d#egg=lm-eval[api]
|
pip install vllm "lm-eval[api]>=0.4.9.2"
|
||||||
```
|
```
|
||||||
|
|
||||||
## Quantization Process
|
## Quantization Process
|
||||||
|
|||||||
@ -20,7 +20,7 @@ for more installation details.
|
|||||||
Additionally, install `vllm` and `lm-evaluation-harness` for evaluation:
|
Additionally, install `vllm` and `lm-evaluation-harness` for evaluation:
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
pip install vllm git+https://github.com/EleutherAI/lm-evaluation-harness.git@206b7722158f58c35b7ffcd53b035fdbdda5126d#egg=lm-eval[api]
|
pip install vllm "lm-eval[api]>=0.4.9.2"
|
||||||
```
|
```
|
||||||
|
|
||||||
## Quantization Process
|
## Quantization Process
|
||||||
|
|||||||
@ -27,7 +27,7 @@ mistral_common[image,audio] >= 1.8.5 # required for voxtral test
|
|||||||
num2words # required for smolvlm test
|
num2words # required for smolvlm test
|
||||||
opencv-python-headless >= 4.11.0 # required for video test
|
opencv-python-headless >= 4.11.0 # required for video test
|
||||||
datamodel_code_generator # required for minicpm3 test
|
datamodel_code_generator # required for minicpm3 test
|
||||||
lm-eval[api] @ git+https://github.com/EleutherAI/lm-evaluation-harness.git@206b7722158f58c35b7ffcd53b035fdbdda5126d # required for model evaluation test
|
lm-eval[api]>=0.4.9.2 # required for model evaluation test
|
||||||
mteb>=1.38.11, <2 # required for mteb test
|
mteb>=1.38.11, <2 # required for mteb test
|
||||||
transformers==4.57.3
|
transformers==4.57.3
|
||||||
tokenizers==0.22.0
|
tokenizers==0.22.0
|
||||||
|
|||||||
@ -58,7 +58,7 @@ schemathesis==3.39.15
|
|||||||
# OpenAI schema test
|
# OpenAI schema test
|
||||||
|
|
||||||
# Evaluation and benchmarking
|
# Evaluation and benchmarking
|
||||||
lm-eval[api] @ git+https://github.com/EleutherAI/lm-evaluation-harness.git@206b7722158f58c35b7ffcd53b035fdbdda5126d
|
lm-eval[api]>=0.4.9.2
|
||||||
jiwer==4.0.0
|
jiwer==4.0.0
|
||||||
|
|
||||||
# Required for multiprocessed tests that use spawn method, Datasets and Evaluate Test
|
# Required for multiprocessed tests that use spawn method, Datasets and Evaluate Test
|
||||||
|
|||||||
@ -34,8 +34,7 @@ num2words # required for smolvlm test
|
|||||||
open_clip_torch==2.32.0 # Required for nemotron_vl test
|
open_clip_torch==2.32.0 # Required for nemotron_vl test
|
||||||
opencv-python-headless >= 4.11.0 # required for video test
|
opencv-python-headless >= 4.11.0 # required for video test
|
||||||
datamodel_code_generator # required for minicpm3 test
|
datamodel_code_generator # required for minicpm3 test
|
||||||
# TODO: Use lm-eval[api]==0.4.10 once released
|
lm-eval[api]>=0.4.9.2 # required for model evaluation test
|
||||||
lm-eval[api] @ git+https://github.com/EleutherAI/lm-evaluation-harness.git@206b7722158f58c35b7ffcd53b035fdbdda5126d # required for model evaluation test
|
|
||||||
mteb[bm25s]>=2, <3 # required for mteb test
|
mteb[bm25s]>=2, <3 # required for mteb test
|
||||||
transformers==4.57.3
|
transformers==4.57.3
|
||||||
tokenizers==0.22.0
|
tokenizers==0.22.0
|
||||||
|
|||||||
@ -441,7 +441,7 @@ lightning-utilities==0.14.3
|
|||||||
# torchmetrics
|
# torchmetrics
|
||||||
llvmlite==0.44.0
|
llvmlite==0.44.0
|
||||||
# via numba
|
# via numba
|
||||||
lm-eval @ git+https://github.com/EleutherAI/lm-evaluation-harness.git@206b7722158f58c35b7ffcd53b035fdbdda5126d
|
lm-eval==0.4.9.2
|
||||||
# via -r requirements/test.in
|
# via -r requirements/test.in
|
||||||
lxml==5.3.0
|
lxml==5.3.0
|
||||||
# via
|
# via
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user