openteamsinc
pypi/lm-eval
Risk Profile
Package
Source
Risk Profile
Maturity:
Mature
Health:
Healthy
Legal:
Healthy
Security:
Healthy
Package
pypi
lm-eval
Version
0.4.11
Last Release Date
19 days ago
License
MIT
Dependencies
datasets
>=2.16.0
numpy
evaluate
>=0.4.0
jinja2
jsonlines
pytablewriter
rouge-score
>=0.0.4
sacrebleu
>=1.5.0
scikit-learn
>=0.24.1
sqlitedict
zstandard
dill
word2number
more_itertools
typing_extensions
api:
requests
aiohttp
tenacity
tqdm
tiktoken
hf:
transformers
>=4.1
torch
>=1.8
accelerate
>=0.26.0
peft
>=0.2.0
vllm:
vllm
>=0.4.2
gptq:
auto-gptq
[triton]
>=0.6.0
gptqmodel:
gptqmodel
>=1.0.9
ipex:
optimum-intel
ibm-watsonx-ai:
ibm_watsonx_ai
>=1.1.22
python-dotenv
optimum:
optimum
[openvino]
sparsify:
sparsify
acpbench:
lark
>=1.1.9
tarski
[clingo]
==0.8.2
pddl
==0.4.2
kstar-planner
==1.4.2
audiolm-qwen:
librosa
soundfile
dev:
pytest
pytest-cov
pytest-xdist
pre-commit
requests
aiohttp
tenacity
tqdm
tiktoken
sentencepiece
ruff
ifeval:
langdetect
immutabledict
nltk
>=3.9.1
japanese-leaderboard:
emoji
==2.14.0
neologdn
==0.5.3
fugashi
[unidic-lite]
rouge_score
>=0.1.2
longbench:
jieba
fuzzywuzzy
rouge
libra:
pymorphy2
math:
sympy
>=1.12
antlr4-python3-runtime
==4.11
math_verify
[antlr4_11_0]
multilingual:
nagisa
>=0.2.7
jieba
>=0.42.1
pycountry
ruler:
nltk
wonderwords
scipy
sentencepiece:
sentencepiece
>=0.1.98
discrim-eval:
statsmodels
==0.14.4
unitxt:
unitxt
==1.22.0
wandb:
wandb
>=0.16.3
pandas
numpy
zeno:
pandas
zeno-client
tasks:
lm_eval
[discrim_eval]
lm_eval
[ifeval]
lm_eval
[japanese_leaderboard]
lm_eval
[longbench]
lm_eval
[libra]
lm_eval
[math]
lm_eval
[multilingual]
lm_eval
[ruler]
Source
Location
EleutherAI/lm-evaluation-harness
Last Source Update
about 9 hours ago
Licenses
MIT License
Distribution Destinations
pypi/lm-eval