Skip to content

MantisAI/prompt_engineering

Repository files navigation

LLM Evaluation

This project aims to evaluate Large Language Models performance on different NLP tasks in combination with various prompts

Environment

Create a virtualenv and install requirements

make virtualenv

Then pull the data

dvc pull

Note that for DVC to work you need access to Mantis AWS

Data

To be filled

About

Code that accompanies the PyData New York (2022) talk: Addressing the sensitivity of Large language models

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Sponsor this project

 

Packages

No packages published