You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+1-2Lines changed: 1 addition & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -82,6 +82,7 @@ People are writing great tools and papers for improving outputs from GPT. Here a
82
82
-[Scale Spellbook](https://scale.com/spellbook): A paid product for building, comparing, and shipping language model apps.
83
83
-[PromptPerfect](https://promptperfect.jina.ai/prompts): A paid product for testing and improving prompts.
84
84
-[Weights & Biases](https://wandb.ai/site/solutions/llmops): A paid product for tracking model training and prompt engineering experiments.
85
+
-[OpenAI Evals](https://github.com/openai/evals): An open-source library for evaluating task performance of language models and prompts.
85
86
86
87
### Prompting guides
87
88
@@ -97,7 +98,6 @@ People are writing great tools and papers for improving outputs from GPT. Here a
97
98
-[Andrej Karpathy's Let's build GPT](https://www.youtube.com/watch?v=kCc8FmEb1nY): A detailed dive into the machine learning underlying GPT.
98
99
-[Prompt Engineering by DAIR.AI](https://www.youtube.com/watch?v=dOxUroR57xs): A one-hour video on various prompt engineering techniques.
99
100
100
-
101
101
### Papers on advanced prompting to improve reasoning
102
102
103
103
-[Chain-of-Thought Prompting Elicits Reasoning in Large Language Models (2022)](https://arxiv.org/abs/2201.11903): Using few-shot prompts to ask models to think step by step improves their reasoning. PaLM's score on math word problems (GSM8K) go from 18% to 57%.
@@ -111,7 +111,6 @@ People are writing great tools and papers for improving outputs from GPT. Here a
111
111
-[Reflexion: an autonomous agent with dynamic memory and self-reflection (2023)](https://arxiv.org/abs/2303.11366): Retrying tasks with memory of prior failures improves subsequent performance.
112
112
-[Demonstrate-Search-Predict: Composing retrieval and language models for knowledge-intensive NLP (2023)](https://arxiv.org/abs/2212.14024): Models augmented with knowledge via a "retrieve-then-read" can be improved with multi-hop chains of searches.
113
113
114
-
115
114
## Contributing
116
115
117
116
If there are examples or guides you'd like to see, feel free to suggest them on the [issues page]. We are also happy to accept high quality pull requests, as long as they fit the scope of the repo.
0 commit comments