|
| 1 | +[metadata] |
| 2 | +name = crfm-helm |
| 3 | +version = 0.2.2 |
| 4 | +author = Stanford CRFM |
| 5 | + |
| 6 | +description = Benchmark for language models |
| 7 | +long_description = Benchmark for language models |
| 8 | +keywords = language models benchmarking |
| 9 | +license = Apache License 2.0 |
| 10 | +classifiers = |
| 11 | + Programming Language :: Python :: 3 :: Only |
| 12 | + Programming Language :: Python :: 3.8 |
| 13 | + License :: OSI Approved :: Apache Software License |
| 14 | +url = "https://github.com/stanford-crfm/helm" |
| 15 | + |
| 16 | +[options] |
| 17 | +python_requires = ~=3.8 |
| 18 | +package_dir = |
| 19 | + =src |
| 20 | +packages = find: |
| 21 | +zip_safe = False |
| 22 | +include_package_data = True |
| 23 | + |
| 24 | +install_requires= |
| 25 | + # Common |
| 26 | + zstandard~=0.18.0 |
| 27 | + tqdm~=4.64.1 |
| 28 | + pyhocon~=0.3.59 |
| 29 | + dacite~=1.6.0 |
| 30 | + |
| 31 | + # Proxy |
| 32 | + aleph-alpha-client~=2.14.0 |
| 33 | + anthropic~=0.2.5 |
| 34 | + bottle~=0.12.23 |
| 35 | + gunicorn~=20.1.0 |
| 36 | + Mako~=1.2.3 |
| 37 | + # sqlitedict==2.0.0 is slow! https://github.com/RaRe-Technologies/sqlitedict/issues/152 |
| 38 | + # Keep the version at 1.7.0. |
| 39 | + sqlitedict~=1.7.0 |
| 40 | + pymongo~=4.2.0 |
| 41 | + retrying~=1.3.3 |
| 42 | + websocket-client~=1.3.2 # For Anthropic (Legacy stanford-online-all-v4-s3) |
| 43 | + openai~=0.27.0 |
| 44 | + tiktoken~=0.3.3 |
| 45 | + transformers~=4.28.1 |
| 46 | + tokenizers~=0.13.2 |
| 47 | + icetk~=0.0.4 |
| 48 | + protobuf~=3.20.2 # Can't use 4.21.0 due to backward incompatibility |
| 49 | + google-api-python-client~=2.64.0 |
| 50 | + revChatGPT~=0.1.1 |
| 51 | + |
| 52 | + # Scenarios |
| 53 | + datasets~=2.5.2 # For math scenarios |
| 54 | + gdown~=4.4.0 |
| 55 | + jsonlines~=3.1.0 # Not really needed |
| 56 | + sympy~=1.11.1 # For math scenarios |
| 57 | + sentencepiece~=0.1.97 |
| 58 | + numba~=0.56.4 |
| 59 | + cattrs~=22.2.0 |
| 60 | + xlrd~=2.0.1 # Used by pandas.read_excel in ice_scenario |
| 61 | + |
| 62 | + # Metrics |
| 63 | + importlib-resources~=5.10.0 |
| 64 | + nltk~=3.7 |
| 65 | + scipy~=1.9.1 |
| 66 | + uncertainty-calibration~=0.1.3 |
| 67 | + rouge-score~=0.1.2 |
| 68 | + pyext~=0.7 |
| 69 | + pytrec_eval==0.5 |
| 70 | + sacrebleu~=2.2.1 |
| 71 | + # Work around https://github.com/p-lambda/verified_calibration/issues/11 |
| 72 | + # TODO: Remove after this issue is resolved |
| 73 | + scikit-learn~=1.1.2 |
| 74 | + spacy~=3.2.4 |
| 75 | + summ-eval~=0.892 |
| 76 | + surge-api~=1.1.0 |
| 77 | + # End users should install a CUDA version of PyTorch manually if needed |
| 78 | + torch~=1.12.1 # Summarization metrics |
| 79 | + torchvision~=0.13.1 |
| 80 | + |
| 81 | + # plotting |
| 82 | + colorcet~=3.0.1 |
| 83 | + matplotlib~=3.6.0 |
| 84 | + numpy~=1.23.3 |
| 85 | + seaborn~=0.11.0 |
| 86 | + |
| 87 | +[options.entry_points] |
| 88 | +console_scripts = |
| 89 | + helm-run = helm.benchmark.run:main |
| 90 | + helm-summarize = helm.benchmark.presentation.summarize:main |
| 91 | + helm-server = helm.benchmark.server:main |
| 92 | + helm-create-plots = helm.benchmark.presentation.create_plots:main |
| 93 | + crfm-proxy-server = helm.proxy.server:main |
| 94 | + crfm-proxy-cli = helm.proxy.cli:main |
| 95 | + |
| 96 | +[options.packages.find] |
| 97 | +where = src |
| 98 | +exclude = |
| 99 | + tests* |
| 100 | + |
1 | 101 | # Settings for Flake8: Tool For Style Guide Enforcement
|
2 | 102 | [flake8]
|
3 | 103 | max-line-length = 120
|
|
0 commit comments