2024 chatgpt update

This commit is contained in:
Your a
2024-05-06 10:24:52 +03:00
parent 6e6445d44a
commit c8302bf3ae
8709 changed files with 190023 additions and 18711 deletions
+120 -2
View File
@@ -1,5 +1,5 @@
[![Multiple choice quiz generator](https://flow-prompt-covers.s3.us-west-1.amazonaws.com/icon/realistic/real_1.png)](https://gptcall.net/chat.html?data=%7B%22contact%22%3A%7B%22id%22%3A%22PVAQ7e8kglwUrLe0u07lm%22%2C%22flow%22%3Atrue%7D%7D)
# Multiple choice quiz generator | [Start Chat](https://gptcall.net/chat.html?data=%7B%22contact%22%3A%7B%22id%22%3A%22PVAQ7e8kglwUrLe0u07lm%22%2C%22flow%22%3Atrue%7D%7D)
This prompt is used to generate {n} multiple choice questions based on long text information {text}. The text can be a summary of a transcript, or excerpt form a book. The output is a JSON to allow integration with code or apps.
@@ -34,6 +34,125 @@ The text delimited in triple backticks:
```{{text}}```
```
## Welcome Message
{
"questions": [
{
"question": "What is the main purpose of the Think GPT library?",
"options": [
"A) Enhancing large language models",
"B) Storing external knowledge",
"C) Performing code generation",
"D) Enabling natural language conditions"
],
"correct_answer": "A) Enhancing large language models",
"explanation": "The Think GPT library aims to overcome the limitations of large language models (LLMs) and enhance their capabilities."
},
{
"question": "Which feature of the Think GPT library allows users to fit large amounts of knowledge into the limited context size of the LLM?",
"options": [
"A) Long memory",
"B) Self-refinement",
"C) Compression",
"D) Natural language conditions"
],
"correct_answer": "C) Compression",
"explanation": "The Think GPT library offers two methods for compression, samurai and summarize, to fit large amounts of knowledge into the limited context size of the LLM."
},
{
"question": "What is the purpose of the LLM library?",
"options": [
"A) Summarization",
"B) Self-refinement",
"C) Code generation",
"D) All of the above"
],
"correct_answer": "D) All of the above",
"explanation": "The LLM library provides functionalities such as summarization, self-refinement, and generative agents."
},
{
"question": "What parameter can users specify when using LLM for summarization?",
"options": [
"A) Max tokens",
"B) Target limit",
"C) Chunk size",
"D) Memory store"
],
"correct_answer": "A) Max tokens",
"explanation": "Users can specify parameters such as max tokens and target limit for summarization using LLM."
},
{
"question": "What infrastructure is needed to get started with the library for implementing prompts?",
"options": [
"A) Vector database",
"B) Cluster setup",
"C) OpenAI API key",
"D) Memory store"
],
"correct_answer": "C) OpenAI API key",
"explanation": "Users only need an OpenAI API key and the library installed to get started with the library for implementing prompts."
}
]
}
## Conversation
**ASSISTANT**: {
@@ -154,4 +273,3 @@ The text delimited in triple backticks:
}