lm-sys / FastChat

An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
Apache License 2.0
37.1k stars 4.57k forks source link

Too many requests in 1 hour. Try again later. #385

Closed chinoll closed 1 year ago

chinoll commented 1 year ago

When entering structured text (such as JSON), there is a possibility of receiving the following error message: "Too many requests in 1 hour. Try again later." Why does this situation occur? Was the corpus contaminated during the training process?

merrymercy commented 1 year ago

Could you upload some screenshots or full conversations?

chinoll commented 1 year ago

Could you upload some screenshots or full conversations?

#1 Task Planning Stage: The AI assistant can parse user input to several tasks: [{"task": task, "id": task_id, "dep": dependency_task_id, "args": {"text": text or <GENERATED>-dep_id, "image": image_url or <GENERATED>-dep_id, "audio": audio_url or <GENERATED>-dep_id}}]. The special tag "<GENERATED>-dep_id" refer to the one generated text/image/audio in the dependency task (Please consider whether the dependency task generates resources of this type.) and "dep_id" must be in "dep" list. The "dep" field denotes the ids of the previous prerequisite tasks which generate a new resource that the current task relies on. The "args" field must in ["text", "image", "audio"], nothing else. The task MUST be selected from the following options: "token-classification", "text2text-generation", "summarization", "translation", "question-answering", "conversational", "text-generation", "sentence-similarity", "tabular-classification", "object-detection", "image-classification", "image-to-image", "image-to-text", "text-to-image", "text-to-video", "visual-question-answering", "document-question-answering", "image-segmentation", "depth-estimation", "text-to-speech", "automatic-speech-recognition", "audio-to-audio", "audio-classification", "canny-control", "hed-control", "mlsd-control", "normal-control", "openpose-control", "canny-text-to-image", "depth-text-to-image", "hed-text-to-image", "mlsd-text-to-image", "normal-text-to-image", "openpose-text-to-image", "seg-text-to-image". There may be multiple tasks of the same type. Think step by step about all the tasks needed to resolve the user\'s request. Parse out as few tasks as possible while ensuring that the user request can be resolved. Pay attention to the dependencies and order among tasks. If the user input can\'t be parsed, you need to reply empty JSON []. <im_start>user\nGive you some pictures e1.jpg, e2.png, e3.jpg, help me count the number of sheep?<im_end>\n<im_start>assistant\n[{"task": "image-to-text", "id": 0, "dep": [-1], "args": {"image": "e1.jpg" }}, {"task": "object-detection", "id": 1, "dep": [-1], "args": {"image": "e1.jpg" }}, {"task": "visual-question-answering", "id": 2, "dep": [1], "args": {"image": "<GENERATED>-1", "text": "How many sheep in the picture"}} }}, {"task": "image-to-text", "id": 3, "dep": [-1], "args": {"image": "e2.png" }}, {"task": "object-detection", "id": 4, "dep": [-1], "args": {"image": "e2.png" }}, {"task": "visual-question-answering", "id": 5, "dep": [4], "args": {"image": "<GENERATED>-4", "text": "How many sheep in the picture"}} }}, {"task": "image-to-text", "id": 6, "dep": [-1], "args": {"image": "e3.jpg" }},  {"task": "object-detection", "id": 7, "dep": [-1], "args": {"image": "e3.jpg" }}, {"task": "visual-question-answering", "id": 8, "dep": [7], "args": {"image": "<GENERATED>-7", "text": "How many sheep in the picture"}}]<im_end>\n<im_start>user\nLook at /e.jpg, can you tell me how many objects in the picture? Give me a picture and video similar to this one.<im_end>\n<im_start>assistant\n[{"task": "image-to-text", "id": 0, "dep": [-1], "args": {"image": "/e.jpg" }}, {"task": "object-detection", "id": 1, "dep": [-1], "args": {"image": "/e.jpg" }}, {"task": "visual-question-answering", "id": 2, "dep": [1], "args": {"image": "<GENERATED>-1", "text": "how many objects in the picture?" }}, {"task": "text-to-image", "id": 3, "dep": [0], "args": {"text": "<GENERATED-0>" }}, {"task": "image-to-image", "id": 4, "dep": [-1], "args": {"image": "/e.jpg" }}, {"task": "text-to-video", "id": 5, "dep": [0], "args": {"text": "<GENERATED-0>" }}]<im_end>\n<im_start>user\ngiven a document /images/e.jpeg, answer me what is the student amount? And describe the image with your voice<im_end>\n<im_start>assistant\n{"task": "document-question-answering", "id": 0, "dep": [-1], "args": {"image": "/images/e.jpeg", "text": "what is the student amount?" }}, {"task": "visual-question-answering", "id": 1, "dep": [-1], "args": {"image": "/images/e.jpeg", "text": "what is the student amount?" }}, {"task": "image-to-text", "id": 2, "dep": [-1], "args": {"image": "/images/e.jpg" }}, {"task": "text-to-speech", "id": 3, "dep": [2], "args": {"text": "<GENERATED>-2" }}]<im_end>\n<im_start>user\nGiven an image /example.jpg, first generate a hed image, then based on the hed image generate a new image where a girl is reading a book<im_end>\n<im_start>assistant\n[{"task": "openpose-control", "id": 0, "dep": [-1], "args": {"image": "/example.jpg" }},  {"task": "openpose-text-to-image", "id": 1, "dep": [0], "args": {"text": "a girl is reading a book", "image": "<GENERATED>-0" }}]<im_end>\n<im_start>user\nplease show me a video and an image of (based on the text) \'a boy is running\' and dub it<im_end>\n<im_start>assistant\n[{"task": "text-to-video", "id": 0, "dep": [-1], "args": {"text": "a boy is running" }}, {"task": "text-to-speech", "id": 1, "dep": [-1], "args": {"text": "a boy is running" }}, {"task": "text-to-image", "id": 2, "dep": [-1], "args": {"text": "a boy is running" }}]<im_end>\n<im_start>user\nplease show me a joke and an image of cat<im_end>\n<im_start>assistant\n[{"task": "conversational", "id": 0, "dep": [-1], "args": {"text": "please show me a joke of cat" }}, {"task": "text-to-image", "id": 1, "dep": [-1], "args": {"text": "a photo of cat" }}]<im_end>\n<im_start>user\nThe chat log [ [] ] may contain the resources I mentioned. Now I input { what is in the examples/a.jpg }. Pay attention to the input and output types of tasks and the dependencies between tasks.<im_end>\n<im_start>assistant

prompt comes from huggingGPT

zhisbug commented 1 year ago

it seems your error is because of using some other service (like openAI's API). This issue is not related to Vicuna or fastchat.

chinoll commented 1 year ago

it seems your error is because of using some other service (like openAI's API). This issue is not related to Vicuna or fastchat.

However, this is the result outputted by serve/cli.py

Junjie-Chu commented 8 months ago

I also meet this, even when I do inference on offline local servers. I guess probably the training data is polluted.