Iprompt
Finding semantically meaningful and accurate prompts.
Install / Use
/learn @csinva/IpromptREADME
<b>Official code for using / reproducing iPrompt from the paper "Explaining Patterns in Data with Language Models via Interpretable Autoprompting" (<a href="https://arxiv.org/abs/2210.01848">Singh*, Morris*, Aneja, Rush, & Gao, 2022</a>) </b> iPrompt generates a human-interpretable prompt that explains patterns in data while still inducing strong generalization performance.
https://user-images.githubusercontent.com/4960970/197355573-e5a1af4c-0784-4344-a314-79793f284b97.mov
Quickstart
Installation: pip install imodelsx (or, for more control, clone and install from source)
Usage example (see imodelsX for more details):
from imodelsx import explain_dataset_iprompt, get_add_two_numbers_dataset
# get a simple dataset of adding two numbers
input_strings, output_strings = get_add_two_numbers_dataset(num_examples=100)
for i in range(5):
print(repr(input_strings[i]), repr(output_strings[i]))
# explain the relationship between the inputs and outputs
# with a natural-language prompt string
prompts, metadata = explain_dataset_iprompt(
input_strings=input_strings,
output_strings=output_strings,
checkpoint='EleutherAI/gpt-j-6B', # which language model to use
num_learned_tokens=3, # how long of a prompt to learn
n_shots=3, # shots per example
n_epochs=15, # how many epochs to search
verbose=0, # how much to print
llm_float16=True, # whether to load the model in float_16
)
--------
prompts is a list of found natural-language prompt strings
Docs
<blockquote> <b>Abstract</b>: Large language models (LLMs) have displayed an impressive ability to harness natural language to perform complex tasks. In this work, we explore whether we can leverage this learned ability to find and explain patterns in data. Specifically, given a pre-trained LLM and data examples, we introduce interpretable autoprompting (iPrompt), an algorithm that generates a natural-language string explaining the data. iPrompt iteratively alternates between generating explanations with an LLM and reranking them based on their performance when used as a prompt. Experiments on a wide range of datasets, from synthetic mathematics to natural-language understanding, show that iPrompt can yield meaningful insights by accurately finding groundtruth dataset descriptions. Moreover, the prompts produced by iPrompt are simultaneously human-interpretable and highly effective for generalization: on real-world sentiment classification datasets, iPrompt produces prompts that match or even improve upon human-written prompts for GPT-3. Finally, experiments with an fMRI dataset show the potential for iPrompt to aid in scientific discovery. </blockquote>- the main api requires simply importing
imodelsx - the
experimentsandexperiments/scriptsfolders contain hyperparameters for running sweeps contained in the paper- note: args that start with
use_are boolean
- note: args that start with
- the
notebooksfolder contains notebooks for analyzing the outputs + making figures
Related work
- fMRI data experiment: Uses scientific data/code from https://github.com/HuthLab/speechmodeltutorial linked to the paper "Natural speech reveals the semantic maps that tile human cerebral cortex" Huth, A. G. et al., (2016) Nature.
- AutoPrompt: find an (uninterpretable) prompt using input-gradients (paper; github)
- Aug-imodels: Explain a dataset by fitting an interpretable linear model/decision tree leveraging a pre-trained language model (paper; github)
Testing
- to check if the pipeline seems to work, install pytest then run
pytestfrom the repo's root directory
If this package is useful for you, please cite the following!
@article{singh2022iprompt,
title = {Explaining Patterns in Data with Language Models via Interpretable Autoprompting},
author = {Singh, Chandan and Morris, John X. and Aneja, Jyoti and Rush, Alexander M. and Gao, Jianfeng},
year = {2022},
url = {https://arxiv.org/abs/2210.01848},
publisher = {arXiv},
doi = {10.48550/ARXIV.2210.01848}
}
Related Skills
YC-Killer
2.7kA library of enterprise-grade AI agents designed to democratize artificial intelligence and provide free, open-source alternatives to overvalued Y Combinator startups. If you are excited about democratizing AI access & AI agents, please star ⭐️ this repository and use the link in the readme to join our open source AI research team.
best-practices-researcher
The most comprehensive Claude Code skills registry | Web Search: https://skills-registry-web.vercel.app
groundhog
399Groundhog's primary purpose is to teach people how Cursor and all these other coding agents work under the hood. If you understand how these coding assistants work from first principles, then you can drive these tools harder (or perhaps make your own!).
last30days-skill
10.3kAI agent skill that researches any topic across Reddit, X, YouTube, HN, Polymarket, and the web - then synthesizes a grounded summary
