8000 GitHub - aakarsh/rl-llm-calibration-test: Attempt at replication of the parts of the paper "Language models (mostly) know what they know", on open datasets, and models.
[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Skip to content

Attempt at replication of the parts of the paper "Language models (mostly) know what they know", on open datasets, and models.

Notifications You must be signed in to change notification settings

aakarsh/rl-llm-calibration-test

Repository files navigation

LLM Calibration Benchmark.

This repository attempts to run benchmarks on some popular openly available language models.

Installation

pip install -r requirements. tx

When running in the colab environment it is recommended ot use.

pip install -r requirements-colab.txt

Unit Tests

Running unit tests requires pytest module invoked as follows:

    python -m pytest test

Running Individual Experiments

Any individual experiment can be rerun using the following command

python  ../llm_calibration/run_experiment.py --model_name='meta-llama/Llama-2-13b-hf' --dataset='STEM' 

The experimental result will produce a json result files which can be parsed offline to generate the requisite plots.

About

Attempt at replication of the parts of the paper "Language models (mostly) know what they know", on open datasets, and models.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 3

  •  
  •  
  •  
0