Upload data.jsonl
Browse files- data.jsonl +3 -3
data.jsonl
CHANGED
|
@@ -887,7 +887,7 @@
|
|
| 887 |
{"organization": "binary-husky", "repo_name": "gpt_academic", "base_commit": "bfa6661367b7592e82225515e5e4845c4aad95bb", "is_iss": 0, "iss_html_url": "https://github.com/binary-husky/gpt_academic/issues/252", "iss_label": "", "title": "\u80fd\u4e0d\u80fd\u4f7f\u7528azure openai key?", "body": "\u4ee3\u7406\u670d\u52a1\u5668\u4e0d\u591f\u7a33\u5b9a\uff0c\u66f4\u9ebb\u70e6\u7684\u662f\u7ed9openai\u7eed\u8d39\uff0c\u8fd8\u8981\u4e2a\u7f8e\u56fd\u4fe1\u7528\u5361\r\n\r\n\u975e\u5e38\u597d\u7684\u5e94\u7528\uff0c\u5e0c\u671b\u51fa\u66f4\u591a\u7684\u63d2\u4ef6\u529f\u80fd\uff0c\u8c22\u8c22", "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {"base_commit": "bfa6661367b7592e82225515e5e4845c4aad95bb", "files": [{"path": "config.py", "Loc": {}}]}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [], "analysis": {"iss_type": "3", "iss_reason": "5", "loc_way": "comment", "loc_scope": "0", "info_type": "Code"}, "max_topk": 0, "file_topk": 0, "loctype": {"code": ["config.py"], "doc": [], "test": [], "config": [], "asset": []}}
|
| 888 |
{"organization": "binary-husky", "repo_name": "gpt_academic", "base_commit": "2d2e02040d7d91d2f2a4c34f4d0bf677873b5f4d", "is_iss": 0, "iss_html_url": "https://github.com/binary-husky/gpt_academic/issues/1328", "iss_label": "", "title": "[Bug]: \u7cbe\u51c6\u7ffb\u8bd1PDF\u6587\u6863(NOUGAT)\u529f\u80fd\u51fa\u9519\uff0c", "body": "### Installation Method | \u5b89\u88c5\u65b9\u6cd5\u4e0e\u5e73\u53f0\n\nOthers (Please Describe)\n\n### Version | \u7248\u672c\n\nPlease choose | \u8bf7\u9009\u62e9\n\n### OS | \u64cd\u4f5c\u7cfb\u7edf\n\nPlease choose | \u8bf7\u9009\u62e9\n\n### Describe the bug | \u7b80\u8ff0\n\n\u6d4b\u8bd5\u670d\u52a1\u5668\uff0c\u7cbe\u51c6\u7ffb\u8bd1PDF\u6587\u6863(NOUGAT)\u529f\u80fd\u51fa\u9519\uff0c\u4f46\u662f\u53ef\u4ee5\u4f7f\u7528\u7cbe\u51c6\u7ffb\u8bd1PDF\u7684\u529f\u80fd\r\n\r\n\r\n\r\n\u62a5\u9519\u4fe1\u606f\u5982\u4e0b\r\n\r\n\n\n### Screen Shot | \u6709\u5e2e\u52a9\u7684\u622a\u56fe\n\n\r\nTraceback (most recent call last):\r\n File \"./toolbox.py\", line 159, in decorated\r\n yield from f(main_input, llm_kwargs, plugin_kwargs, chatbot_with_cookie, history, *args, **kwargs)\r\n File \"./crazy_functions/\u6279\u91cf\u7ffb\u8bd1PDF\u6587\u6863_NOUGAT.py\", line 93, in \u6279\u91cf\u7ffb\u8bd1PDF\u6587\u6863\r\n yield from \u89e3\u6790PDF_\u57fa\u4e8eNOUGAT(file_manifest, project_folder, llm_kwargs, plugin_kwargs, chatbot, history, system_prompt)\r\n File \"./crazy_functions/\u6279\u91cf\u7ffb\u8bd1PDF\u6587\u6863_NOUGAT.py\", line 111, in \u89e3\u6790PDF_\u57fa\u4e8eNOUGAT\r\n fpp = yield from nougat_handle.NOUGAT_parse_pdf(fp, chatbot, history)\r\n File \"./crazy_functions/crazy_utils.py\", line 761, in NOUGAT_parse_pdf\r\n raise RuntimeError(\"Nougat\u89e3\u6790\u8bba\u6587\u5931\u8d25\u3002\")\r\nRuntimeError: Nougat\u89e3\u6790\u8bba\u6587\u5931\u8d25\u3002\n\n### Terminal Traceback & Material to Help Reproduce Bugs | \u7ec8\u7aeftraceback\uff08\u5982\u6709\uff09 + \u5e2e\u52a9\u6211\u4eec\u590d\u73b0\u7684\u6d4b\u8bd5\u6750\u6599\u6837\u672c\uff08\u5982\u6709\uff09\n\n_No response_", "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {"base_commit": "2d2e02040d7d91d2f2a4c34f4d0bf677873b5f4d", "files": [{"path": "crazy_functions/crazy_utils.py", "Loc": {"('nougat_interface', 'NOUGAT_parse_pdf', 739)": {"mod": [752]}, "('nougat_interface', None, 719)": {"mod": [723]}}, "status": "modified"}]}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [], "analysis": {"iss_type": "1", "iss_reason": "5", "loc_way": "comment", "loc_scope": "0", "info_type": "Code"}, "max_topk": 2, "file_topk": 1, "loctype": {"code": ["crazy_functions/crazy_utils.py"], "doc": [], "test": [], "config": [], "asset": []}}
|
| 889 |
{"organization": "binary-husky", "repo_name": "gpt_academic", "base_commit": "17abd29d5035b5b227deaad69d32cf437b23e542", "is_iss": 0, "iss_html_url": "https://github.com/binary-husky/gpt_academic/issues/94", "iss_label": "", "title": "[\u4e00\u4e9b\u5efa\u8bae]input\u6846\u8fd8\u662f\u592a\u5c0f\u4e86", "body": "RT \u591a\u884c\u8f93\u5165\u8fd8\u662f\u4e0d\u65b9\u4fbf\uff0c\u5982\u679c\u9002\u5f53\u8c03\u6574\u4f1a\u66f4\u597d\u7528\u3002\r\n\r\n\u5e0c\u671b\u91c7\u7eb3\uff0c\u611f\u8c22\u5206\u4eab\u3002", "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {"base_commit": "17abd29d5035b5b227deaad69d32cf437b23e542", "files": [{"path": "main.py", "Loc": {"(None, None, None)": {"mod": [1]}}, "status": "modified"}]}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [], "analysis": {"iss_type": "4", "iss_reason": "5", "loc_way": "comment", "loc_scope": "0", "info_type": "Code"}, "max_topk": 1, "file_topk": 1, "loctype": {"code": ["main.py"], "doc": [], "test": [], "config": [], "asset": []}}
|
| 890 |
-
{"organization": "binary-husky", "repo_name": "gpt_academic", "base_commit": "37744a9cb173477398a2609f02d5e7cef47eb677", "is_iss": 1, "iss_html_url": "https://github.com/binary-husky/gpt_academic/issues/1438", "iss_label": "", "title": "[Bug]: \u6d6e\u52a8\u8f93\u5165\u6846\u5728\u62d6\u81f3\u9876\u90e8\u540e\uff0c\u65e0\u6cd5\u91cd\u65b0\u79fb\u4f4d", "body": "### Installation Method | \u5b89\u88c5\u65b9\u6cd5\u4e0e\u5e73\u53f0\n\nOthers (Please Describe)\n\n### Version | \u7248\u672c\n\nPlease choose | \u8bf7\u9009\u62e9\n\n### OS | \u64cd\u4f5c\u7cfb\u7edf\n\nMac\n\n### Describe the bug | \u7b80\u8ff0\n\n\u6d6e\u52a8\u8f93\u5165\u6846\u5728\u62d6\u81f3\u9876\u90e8\u540e\uff0c\u65e0\u6cd5\u91cd\u65b0\u79fb\u4f4d\r\n\r\n\u671f\u671b\uff1a\u91cd\u65b0\u52fe\u9009\u540e\uff0c\u5e94\u8be5\u56de\u5230\u521d\u59cb\u4f4d\u7f6e\n\n### Screen Shot | \u6709\u5e2e\u52a9\u7684\u622a\u56fe\n\n\n\n### Terminal Traceback & Material to Help Reproduce Bugs | \u7ec8\u7aeftraceback\uff08\u5982\u6709\uff09 + \u5e2e\u52a9\u6211\u4eec\u590d\u73b0\u7684\u6d4b\u8bd5\u6750\u6599\u6837\u672c\uff08\u5982\u6709\uff09\n\n_No response_", "code": null, "pr_html_url": null, "commit_html_url": "https://github.com/binary-husky/gradio-fix/commit/fb67dd12f58aa53c75a90378cddbc811ac3c01d2", "file_loc": {}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [{"org": "binary-husky", "pro": "gradio-fix", "path": {"base_commit": "fb67dd12f58aa53c75a90378cddbc811ac3c01d2", "files": [{"path": "js/app/src/components/Floating/StaticFloating.svelte", "status": "modified", "Loc": {"(None, None, 48)": {"add": [48]}}}]}}], "analysis": {"iss_type": "2", "iss_reason": "1", "loc_way": "commit", "loc_scope": "0", "info_type": "Code"}, "loctype": {"code": [], "doc": [], "test": [], "config": [], "asset": ["gradio-fix", "js/app/src/components/Floating/StaticFloating.svelte"]}}
|
| 891 |
{"organization": "binary-husky", "repo_name": "gpt_academic", "base_commit": "6538c58b8e5a4a7ae08dfa1ae9970bc422158096", "is_iss": 0, "iss_html_url": "https://github.com/binary-husky/gpt_academic/issues/620", "iss_label": "", "title": "\u60f3\u95ee\u95eenewbing\u7684cookies\u600e\u4e48\u586b\u5199\uff0c\u6211\u4ecejavascript:alert(document.cookie)\u627e\u5230\u4e86cookies\u4f46\u662f\u4e00\u76f4\u663e\u793acookies\u6709\u9519", "body": "\r\n\r\n", "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {"base_commit": "6538c58b8e5a4a7ae08dfa1ae9970bc422158096", "files": [{"path": "config.py", "Loc": {"(None, None, None)": {"mod": [69]}}, "status": "modified"}]}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [], "analysis": {"iss_type": "1", "iss_reason": "5", "loc_way": "comment", "loc_scope": "", "info_type": "Other"}, "max_topk": 1, "file_topk": 1, "loctype": {"code": ["config.py"], "doc": [], "test": [], "config": [], "asset": []}}
|
| 892 |
{"organization": "binary-husky", "repo_name": "gpt_academic", "base_commit": "6d8c8cd3f0b9d2b6fe8d412b83f902cbd43fa0bd", "is_iss": 0, "iss_html_url": "https://github.com/binary-husky/gpt_academic/issues/150", "iss_label": "documentation\nhigh value issue", "title": "\u6709\u6ca1\u6709\u5b8c\u5168\u90e8\u7f72\u6210\u529f\u7684\u5927\u795e\u51fa\u4e2a\u8be6\u7ec6\u7684\u90e8\u7f72\u6b65\u9aa4\u5440\uff1fWindows \u6709\u622a\u56fe\uff0c\u8dea\u6c42", "body": "Windows\u5b89\u88c5\u90e8\u7f72\r\n\u57fa\u672c\u73af\uff1a\u5b89\u88c5anaconda\r\n1.\u4e0b\u8f7d\u9879\u76ee CMD\r\n\u9009\u62e9\u8def\u5f84\r\ngit clone https://github.com/binary-husky/chatgpt_academic.git\r\ncd chatgpt_academic\r\n\u6211\u4eec\u5efa\u8bae\u5c06config.py\u590d\u5236\u4e3aconfig_private.py\u5e76\u5c06\u540e\u8005\u7528\u4f5c\u4e2a\u6027\u5316\u914d\u7f6e\u6587\u4ef6\u4ee5\u907f\u514dconfig.py\u4e2d\u7684\u53d8\u66f4\u5f71\u54cd\u4f60\u7684\u4f7f\u7528\u6216\u4e0d\u5c0f\u5fc3\u5c06\u5305\u542b\u4f60\u7684OpenAI API KEY\u7684config.py\u63d0\u4ea4\u81f3\u672c\u9879\u76ee\u3002\r\ncp config.py config_private.py\r\n2.\u521b\u5efa\u865a\u62df\u73af\u5883 python 3.11\r\nconda create -n chatgpt python=3.11.0 #\u65b0\u5efa\u73af\u5883\u3001\r\n3.\u8fdb\u5165\u9879\u76ee\u4e0b\u8f7d\u8def\u5f84\r\n\u4f8b\u5982 cd G:\\python\\Program\\chatgpt_academic\r\n4.\u542f\u52a8\u865a\u62df\u73af\u5883\r\nconda activate chatgpt\r\n5. \u5b89\u88c5 gradio>=3.23\r\n\uff081\uff09\u5230https://pypi.org/project/gradio/ \u4e0b\u8f7dwhl\u7248\u672c\r\n\uff082\uff09pip install G:\\python\\Program\\chatgpt_academic\\gradio-3.23.0-py3-none-any.whl\r\n6.\u914d\u7f6e\u5176\u4ed6\u73af\u5883\r\n\uff081\uff09\u6253\u5f00requirements.txt\uff0c\u6ce8\u91ca\u6389gradio\uff0c\u7136\u540e\u4fdd\u5b58\r\n\uff082\uff09\u8fd0\u884c python -m pip install -r requirements.txt\r\n7.\u542f\u52a8\u4ee3\u7406\r\n8. \u914d\u7f6econfig_private.py\r\n\uff081\uff09\u6dfb\u52a0API_KEY\r\n\uff082\uff09\u4fee\u6539USE_PROXY = Ture\r\n\uff083\uff09\u4fee\u6539proxies\r\n\u5728\u6d4f\u89c8\u5668\u8f93\u5165: https://ipapi.co/json/\r\n\u6d4f\u89c8\u5668\u4e0a\u53f3\u952e->\u68c0\u67e5->\u7f51\u7edc->ctrl+r\r\n\u6253\u5f00json\uff0c\u5c06\u8fdc\u7a0b\u5730\u5740\u4fee\u6539\u5230proxies = { \"http\": \"104.26.9.44:443\", \"https\": \"104.26.9.44:443\", }\r\n9.\u542f\u52a8\u7a0b\u5e8f\r\npython main.py", "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {"base_commit": "6d8c8cd3f0b9d2b6fe8d412b83f902cbd43fa0bd", "files": [{"path": "requirements.txt", "Loc": {}}]}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [], "analysis": {"iss_type": "3", "iss_reason": "5", "loc_way": "comment", "loc_scope": "0", "info_type": "Code\n+ \nDoc"}, "max_topk": 0, "file_topk": 0, "loctype": {"code": [], "doc": [], "test": [], "config": ["requirements.txt"], "asset": []}}
|
| 893 |
{"organization": "binary-husky", "repo_name": "gpt_academic", "base_commit": "e20070939c6c7eeca33a8438041c9e038836957b", "is_iss": 1, "iss_html_url": "https://github.com/binary-husky/gpt_academic/issues/568", "iss_label": "enhancement", "title": "\u80fd\u5426\u589e\u52a0\u804a\u5929\u5185\u5bb9\u5bfc\u51fa\u529f\u80fd\uff1f", "body": null, "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {}, "own_code_loc": [], "ass_file_loc": ["gpt_log/chat_secrets.log"], "other_rep_loc": [], "analysis": {"iss_type": "4", "iss_reason": "5", "loc_way": "comment", "loc_scope": "1", "info_type": "Config"}, "loctype": {"code": [], "doc": [], "test": [], "config": [], "asset": ["gpt_log/chat_secrets.log"]}}
|
|
@@ -1004,7 +1004,7 @@
|
|
| 1004 |
{"organization": "huggingface", "repo_name": "transformers", "base_commit": "82c7e879876822864b5ceaf2c99eb01159266bcd", "is_iss": 0, "iss_html_url": "https://github.com/huggingface/transformers/issues/27200", "iss_label": "", "title": "dataset download error in speech recognition examples", "body": "### System Info\n\n- `transformers` version: 4.35.0.dev0\r\n- Platform: Linux-5.15.0-43-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.18\r\n- Huggingface_hub version: 0.17.3\r\n- Safetensors version: 0.4.0\r\n- Accelerate version: 0.24.1\r\n- Accelerate config: not found\r\n- PyTorch version (GPU?): 1.10.0+cu111 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Flax version (CPU?/GPU?/TPU?): not installed (NA)\r\n- Jax version: not installed\r\n- JaxLib version: not installed\r\n- Using GPU in script?: <fill in>\r\n- Using distributed or parallel set-up in script?: <fill in>\n\n### Who can help?\n\n@stevhliu and @MKhalusova\n\n### Information\n\n- [x] The official example scripts\n- [ ] My own modified scripts\n\n### Tasks\n\n- [X] An officially supported task in the `examples` folder (such as GLUE/SQuAD, ...)\n- [ ] My own task or dataset (give details below)\n\n### Reproduction\n\nCUDA_VISIBLE_DEVICES=0 python run_speech_recognition_ctc.py \\\r\n\t--dataset_name=\"common_voice\" \\\r\n\t--model_name_or_path=\"facebook/wav2vec2-large-xlsr-53\" \\\r\n\t--dataset_config_name=\"tr\" \\\r\n\t--output_dir=\"./wav2vec2-common_voice-tr-demo\" \\\r\n\t--overwrite_output_dir \\\r\n\t--num_train_epochs=\"15\" \\\r\n\t--per_device_train_batch_size=\"16\" \\\r\n\t--gradient_accumulation_steps=\"2\" \\\r\n\t--learning_rate=\"3e-4\" \\\r\n\t--warmup_steps=\"500\" \\\r\n\t--evaluation_strategy=\"steps\" \\\r\n\t--text_column_name=\"sentence\" \\\r\n\t--length_column_name=\"input_length\" \\\r\n\t--save_steps=\"400\" \\\r\n\t--eval_steps=\"100\" \\\r\n\t--layerdrop=\"0.0\" \\\r\n\t--save_total_limit=\"3\" \\\r\n\t--freeze_feature_encoder \\\r\n\t--gradient_checkpointing \\\r\n\t--chars_to_ignore , ? . ! - \\; \\: \\\" \u201c % \u2018 \u201d \ufffd \\\r\n\t--fp16 \\\r\n\t--group_by_length \\\r\n\t--push_to_hub \\\r\n\t--do_train --do_eval \n\n### Expected behavior\n\nWhen I run the default command, which set `dataset_name` as \"common_voice\", and I got a warning:\r\n```\r\n/home/xintong/.cache/huggingface/modules/datasets_modules/datasets/common_voice/220833898d6a60c50f621126e51fb22eb2dfe5244392c70dccd8e6e2f055f4bf/common_voice.py:634: FutureWarning: \r\n This version of the Common Voice dataset is deprecated.\r\n You can download the latest one with\r\n >>> load_dataset(\"mozilla-foundation/common_voice_11_0\", \"en\")\r\n \r\n warnings.warn(\r\nGenerating train split: 0%| | 0/1831 [00:00<?, ? examples/s]\r\nTraceback (most recent call last):\r\n File \"/home/xintong/miniconda3/envs/test/lib/python3.8/tarfile.py\", line 2578, in next\r\n tarinfo = self.tarinfo.fromtarfile(self)\r\n File \"/home/xintong/miniconda3/envs/test/lib/python3.8/tarfile.py\", line 1283, in fromtarfile\r\n obj = cls.frombuf(buf, tarfile.encoding, tarfile.errors)\r\n File \"/home/xintong/miniconda3/envs/test/lib/python3.8/tarfile.py\", line 1221, in frombuf\r\n raise TruncatedHeaderError(\"truncated header\")\r\ntarfile.TruncatedHeaderError: truncated header\r\n```\r\nI modified this into `mozilla-foundation/common_voice_11_0`, it passed. \r\n```\r\nDownloading builder script: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 8.13k/8.13k [00:00<00:00, 30.3MB/s]\r\nDownloading readme: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 14.4k/14.4k [00:00<00:00, 19.2MB/s]\r\nDownloading extra modules: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 3.44k/3.44k [00:00<00:00, 19.9MB/s]\r\nDownloading extra modules: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 60.9k/60.9k [00:00<00:00, 304kB/s]\r\nDownloading data: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 12.2k/12.2k [00:00<00:00, 25.6MB/s]\r\nDownloading data: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 568M/568M [00:07<00:00, 71.7MB/s]\r\nDownloading data: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 233M/233M [00:02<00:00, 78.6MB/s]\r\nDownloading data: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 285M/285M [00:04<00:00, 67.7MB/s]\r\nDownloading data: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 4.86M/4.86M [00:00<00:00, 73.3MB/s]\r\nDownloading data: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 109M/109M [00:01<00:00, 80.4MB/s]\r\nDownloading data files: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 5/5 [00:21<00:00, 4.24s/it]\r\nExtracting data files: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 5/5 [00:07<00:00, 1.54s/it]\r\nDownloading data: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 5.76M/5.76M [00:00<00:00, 56.0MB/s]\r\nDownloading data: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2.17M/2.17M [00:00<00:00, 54.1MB/s]\r\nDownloading data: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2.18M/2.18M [00:00<00:00, 64.3MB/s]\r\nDownloading data: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 32.8k/32.8k [00:00<00:00, 53.1MB/s]\r\nDownloading data: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 800k/800k [00:00<00:00, 59.8MB/s]\r\nDownloading data files: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 5/5 [00:05<00:00, 1.01s/it]\r\nExtracting data files: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 5/5 [00:00<00:00, 2954.98it/s]\r\n```", "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {"base_commit": "82c7e879876822864b5ceaf2c99eb01159266bcd", "files": [{"path": "examples/pytorch/speech-recognition/README.md", "Loc": {"(None, None, 69)": {"mod": [69]}}, "status": "modified"}]}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [], "analysis": {"iss_type": "1", "iss_reason": "2", "loc_way": "comment", "loc_scope": "0", "info_type": "Doc"}, "max_topk": 1, "file_topk": 1, "loctype": {"code": [], "doc": ["examples/pytorch/speech-recognition/README.md"], "test": [], "config": [], "asset": []}}
|
| 1005 |
{"organization": "huggingface", "repo_name": "transformers", "base_commit": "0e82f0cbc28b41b3d87a5e4069dc0e20bacc2494", "is_iss": 0, "iss_html_url": "https://github.com/huggingface/transformers/issues/12081", "iss_label": "", "title": "GPT2 Flax \"TypeError: JAX only supports number and bool dtypes, got dtype object in array\"", "body": "On GPU\r\n\r\n```\r\n>>> from transformers import AutoTokenizer, FlaxAutoModelForCausalLM\r\n\r\n>>> tokenizer = AutoTokenizer.from_pretrained(\"gpt2-medium\")\r\n>>> model = FlaxAutoModelForCausalLM.from_pretrained(\"gpt2-medium\")\r\n>>> input_context = \"The dog\"\r\n>>> # encode input context\r\n>>> input_ids = tokenizer(input_context, return_tensors=\"jax\").input_ids\r\n>>> # generate candidates using sampling\r\n>>> outputs = model.generate(input_ids=input_ids, max_length=20, top_k=30, do_sample=True)\r\n\r\nTypeError: JAX only supports number and bool dtypes, got dtype object in array\r\n```\r\n\r\n@patrickvonplaten @patil-suraj ", "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {"base_commit": "0e82f0cbc28b41b3d87a5e4069dc0e20bacc2494", "files": [{"path": "src/transformers/models/gpt2/modeling_flax_gpt2.py", "Loc": {"('FlaxGPT2LMHeadModule', None, 553)": {"mod": []}}, "status": "modified"}, {"path": "src/transformers/models/gpt2/tokenization_gpt2_fast.py", "Loc": {"('GPT2TokenizerFast', None, 70)": {"mod": []}}, "status": "modified"}, {"Loc": [6, 7], "path": null}]}, "own_code_loc": [{"Loc": [6, 7], "path": null}], "ass_file_loc": [], "other_rep_loc": [], "analysis": {"iss_type": "1", "iss_reason": "3", "loc_way": "comment", "loc_scope": "3", "info_type": "Code"}, "max_topk": 2, "file_topk": 2, "loctype": {"code": [null, "src/transformers/models/gpt2/tokenization_gpt2_fast.py", "src/transformers/models/gpt2/modeling_flax_gpt2.py"], "doc": [], "test": [], "config": [], "asset": []}}
|
| 1006 |
{"organization": "huggingface", "repo_name": "transformers", "base_commit": "322037e842e5e89080918c824998c17722df6f19", "is_iss": 0, "iss_html_url": "https://github.com/huggingface/transformers/issues/10079", "iss_label": "", "title": "Unclear error \"NotImplementedError: \"while saving tokenizer. How fix it?", "body": "Here is my tokenizer code and how I save it to a json file\" /content/bert-datas7.json\"\r\n\r\n````\r\nfrom tokenizers import normalizers\r\nfrom tokenizers.normalizers import Lowercase, NFD, StripAccents\r\n\r\nbert_tokenizer.pre_tokenizer = Whitespace()\r\n\r\nfrom tokenizers.processors import TemplateProcessing\r\n\r\nbert_tokenizer.post_processor = TemplateProcessing(\r\n single=\"[CLS] $A [SEP]\",\r\n pair=\"[CLS] $A [SEP] $B:1 [SEP]:1\",\r\n special_tokens=[\r\n (\"[CLS]\", 1),\r\n (\"[SEP]\", 2),\r\n (\"[PAD]\", 3),\r\n ],\r\n \r\n)\r\nfrom tokenizers.trainers import WordPieceTrainer\r\n\r\ntrainer = WordPieceTrainer(\r\n vocab_size=30522, special_tokens=[\"[UNK]\", \"[CLS]\", \"[SEP]\", \"[PAD]\", \"[MASK]\"], pad_to_max_length=True\r\n)\r\nfiles = [f\"/content/For_ITMO.txt\" for split in [\"test\", \"train\", \"valid\"]]\r\nbert_tokenizer.train(trainer, files)\r\n\r\nmodel_files = bert_tokenizer.model.save(\"data\", \"/content/For_ITMO.txt\")\r\n\r\nbert_tokenizer.model = WordPiece.from_file(*model_files, unk_token=\"[UNK]\", pad_to_max_length=True)\r\n\r\nbert_tokenizer.save(\"/content/bert-datas7.json\") \r\n````\r\n\r\nWhen I output tokenizer name_or_path = nothing is displayed. This is normal?\r\n\r\n\r\n````\r\ntokenizer = PreTrainedTokenizerFast(tokenizer_file='/content/bert-datas7.json')\r\ntokenizer.add_special_tokens({'pad_token': '[PAD]'})\r\n\r\nprint(tokenizer)\r\n>>> PreTrainedTokenizerFast(name_or_path='', vocab_size=1435, model_max_len=1000000000000000019884624838656, is_fast=True, padding_side='right', special_tokens={'pad_token': '[PAD]'})\r\n````\r\nAlso, when I try to save my tokenizer, I get an error without explanation. How can I rewrite the code so that all this???\r\n#9658 \r\n#10039 \r\n[For_ITMO.txt-vocab (1) (1).txt](https://github.com/huggingface/transformers/files/5945659/For_ITMO.txt-vocab.1.1.txt)\r\n \r\n````\r\ntokenizer.save_pretrained(\"/content/tokennizerrrr\")\r\n\r\nNotImplementedError Traceback (most recent call last)\r\n<ipython-input-11-efc48254a528> in <module>()\r\n----> 1 tokenizer.save_pretrained(\"/content/tokennizerrrr\")\r\n\r\n2 frames\r\n/usr/local/lib/python3.6/dist-packages/transformers/tokenization_utils_base.py in save_vocabulary(self, save_directory, filename_prefix)\r\n 2042 :obj:`Tuple(str)`: Paths to the files saved.\r\n 2043 \"\"\"\r\n-> 2044 raise NotImplementedError\r\n 2045 \r\n 2046 def tokenize(self, text: str, pair: Optional[str] = None, add_special_tokens: bool = False, **kwargs) -> List[str]:\r\n\r\nNotImplementedError: \r\n````\r\n", "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {"base_commit": "322037e842e5e89080918c824998c17722df6f19", "files": [{"path": "src/transformers/tokenization_utils_fast.py", "Loc": {"('PreTrainedTokenizerFast', '_save_pretrained', 505)": {"mod": [509]}}, "status": "modified"}]}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [], "analysis": {"iss_type": "1", "iss_reason": "5", "loc_way": "comment", "loc_scope": "0", "info_type": "Code"}, "max_topk": 1, "file_topk": 1, "loctype": {"code": ["src/transformers/tokenization_utils_fast.py"], "doc": [], "test": [], "config": [], "asset": []}}
|
| 1007 |
-
{"organization": "huggingface", "repo_name": "transformers", "base_commit": "77a257fc210a56f1fd0d75166ecd654cf58111f3", "is_iss": 1, "iss_html_url": "https://github.com/huggingface/transformers/issues/8403", "iss_label": "", "title": "[s2s finetune] huge increase in memory demands with --fp16 native amp", "body": "While working on https://github.com/huggingface/transformers/issues/8353 I discovered that `--fp16` causes a 10x+ increase in gpu memory demands.\r\n\r\ne.g. I can run bs=12 w/o `--fp16` \r\n\r\n```\r\ncd examples/seq2seq\r\nexport BS=12; rm -rf distilbart-cnn-12-6; python finetune.py --learning_rate=3e-5 --gpus 1 \\\r\n--do_train --do_predict --val_check_interval 0.25 --n_val 500 --num_train_epochs 2 --freeze_encoder \\\r\n--freeze_embeds --data_dir cnn_dm --max_target_length 142 --val_max_target_length=142 \\\r\n--train_batch_size=$BS --eval_batch_size=$BS --gradient_accumulation_steps 1 \\\r\n--model_name_or_path sshleifer/student_cnn_12_6 --tokenizer_name facebook/bart-large \\\r\n--warmup_steps 500 --output_dir distilbart-cnn-12-6\r\n\r\n```\r\nBut if I add:\r\n```\r\n--fp16\r\n```\r\n\r\n(w/ or w/o `--fp16_opt_level O1`)\r\n\r\nI get OOM even with bs=1 on a 8GB card and it barely manages on a 24GB card - I think the increase in memory demand is more than 10x.\r\n\r\nThe OOM either right away when it does the sanity check step, or after just 10-20 batches - so within a few secs\r\n\r\nThis is with pytorch-1.6. Same goes for pytorch-1.7 and 1.8-nightly.\r\n\r\nI wasn't able to test `--fp16` with pytorch-1.5, since I can't build apex on ubuntu-20.04. Without `--fp16` pytorch-1.5 works the same as pytorch-1.6 gpu memory-wise.\r\n\r\nI tested with pytorch-1.5 + apex and there is no problem there. Memory consumption is about half.\r\n\r\nHere is the table of the batch sizes that fit into a 8gb rtx-1070 (bigger BS leads to an instant OOM):\r\n\r\nbs | version\r\n---|--------\r\n12 | pt15\r\n20 | pt15+fp16\r\n12 | pt16\r\n1 | pt16+fp16\r\n\r\n\r\n\r\nIf you'd like to reproduce the problem here are the full steps:\r\n\r\n```\r\n# prep library\r\ngit clone https://github.com/huggingface/transformers\r\ncd transformers\r\npip install -e .[dev]\r\npip install -r examples/requirements.txt\r\ncd examples/seq2seq\r\n\r\n# prep data\r\nwget https://cdn-datasets.huggingface.co/summarization/cnn_dm_v2.tgz\r\ntar -xzvf cnn_dm_v2.tgz # empty lines removed\r\nmv cnn_cln cnn_dm\r\n\r\n# run\r\nexport BS=12; \r\nrm -rf distilbart-cnn-12-6\r\npython finetune.py --learning_rate=3e-5 --gpus 1 \\\r\n--do_train --do_predict --val_check_interval 0.25 --n_val 500 --num_train_epochs 2 --freeze_encoder \\\r\n--freeze_embeds --data_dir cnn_dm --max_target_length 142 --val_max_target_length=142 \\\r\n--train_batch_size=$BS --eval_batch_size=$BS --gradient_accumulation_steps 1 \\\r\n--model_name_or_path sshleifer/student_cnn_12_6 --tokenizer_name facebook/bart-large \\\r\n--warmup_steps 500 --output_dir distilbart-cnn-12-6 \r\n```\r\n\r\nThis issue is to track the problem and hopefully finding a solution.\r\n\r\n@sshleifer ", "code": null, "pr_html_url": null, "commit_html_url": "https://github.com/pytorch/pytorch/commit/57bffc3a8e4fee0cce31e1ff1f662ccf7b16db57", "file_loc": {}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [{"org": "pytorch", "pro": "pytorch", "path": {"base_commit": "57bffc3a8e4fee0cce31e1ff1f662ccf7b16db57", "files": [{"path": "aten/src/ATen/autocast_mode.cpp", "status": "modified", "Loc": {"(None, 'cached_cast', 67)": {"mod": [71]}}}, {"path": "test/test_cuda.py", "status": "modified", "Loc": {"('TestCuda', None, 92)": {"add": [2708]}}}]}}], "analysis": {"iss_type": "2", "iss_reason": "1", "loc_way": "commit", "loc_scope": "2", "info_type": "Code"}, "loctype": {"code": ["aten/src/ATen/autocast_mode.cpp"], "doc": [], "test": ["test/test_cuda.py"], "config": [], "asset": ["pytorch"]}}
|
| 1008 |
{"organization": "huggingface", "repo_name": "transformers", "base_commit": "1a688709b34b10bd372e3e0860c8d39d170ebf53", "is_iss": 0, "iss_html_url": "https://github.com/huggingface/transformers/issues/17201", "iss_label": "", "title": "a memory leak in qqp prediction using bart", "body": "### System Info\n\n```shell\n- `transformers` version: 4.19.0.dev0\r\n- Platform: Linux-5.11.0-43-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- Huggingface_hub version: 0.4.0\r\n- PyTorch version (GPU?): 1.10.1 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Flax version (CPU?/GPU?/TPU?): not installed (NA)\r\n- Jax version: not installed\r\n- JaxLib version: not installed\r\n- Using GPU in script?: Yes\r\n- Using distributed or parallel set-up in script?: No\n```\n\n\n### Who can help?\n\n@sgugger\n\n### Information\n\n- [X] The official example scripts\n- [ ] My own modified scripts\n\n### Tasks\n\n- [X] An officially supported task in the `examples` folder (such as GLUE/SQuAD, ...)\n- [ ] My own task or dataset (give details below)\n\n### Reproduction\n\nI met the same issue #11011. If not using `--eval_accumulation_steps`, it caused CUDA out of memory. If using it, it caused out of RAM and killed by system.\r\n\r\nI only did prediction on GLUE QQP dataset using bart without fine-tuning. Considering QQP having a large test set (300k), the prediction got slower and slower, and finally got out of memory.\r\n\r\nThis is the script to reproduce:\r\n```\r\nCUDA_VISIBLE_DEVICES=0 python run_glue.py --model_name_or_path facebook/bart-large --task_name qqp --output_dir bart-large_qqp --eval_accumulation_steps 100 --do_predict --per_device_eval_batch_size 24\r\n```\n\n### Expected behavior\n\n```shell\nPrediction without out memory.\n```\n", "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {"base_commit": "1a688709b34b10bd372e3e0860c8d39d170ebf53", "files": [{"path": "src/transformers/trainer.py", "Loc": {"('Trainer', 'evaluation_loop', 2549)": {"mod": [2635]}}, "status": "modified"}]}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [], "analysis": {"iss_type": "1", "iss_reason": "2\nOr\n5", "loc_way": "comment", "loc_scope": "0", "info_type": "Code"}, "max_topk": 1, "file_topk": 1, "loctype": {"code": ["src/transformers/trainer.py"], "doc": [], "test": [], "config": [], "asset": []}}
|
| 1009 |
{"organization": "huggingface", "repo_name": "transformers", "base_commit": "cef2e40e0f8eaad13b8d32817a48fdddc32eb2a5", "is_iss": 0, "iss_html_url": "https://github.com/huggingface/transformers/issues/28435", "iss_label": "", "title": "Skip some weights for load_in_8bit and keep them as fp16/32?", "body": "### Feature request\r\n\r\nHello,\r\n\r\nI am looking for a way to load a checkpoint where I only load some of the weights in 8 bit and keep others in 16/32 bit.\r\n\r\n### Motivation\r\n\r\nMy motivation is for vision-language models like Llava or BLIP2 where I want to load the LLM part in 8 bit but the image encoder should stay in 16 bit because I notice performance degradations with CLIP in 8 bit and also want to be able to train this part without LoRA.\r\n\r\nAs far as I can see in the documentation, issues and with Google (both here and for bitsandbytes), there is currently no way to do this.\r\n\r\n### Your contribution\r\n\r\nI can in theory help implement something like this but I don't know where and how in the code this should be done.", "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {"base_commit": "cef2e40e0f8eaad13b8d32817a48fdddc32eb2a5", "files": [{"path": "src/transformers/modeling_utils.py", "Loc": {"('PreTrainedModel', 'from_pretrained', 2528)": {"mod": [3524]}}, "status": "modified"}, {"path": "src/transformers/utils/quantization_config.py", "Loc": {"('BitsAndBytesConfig', None, 151)": {"mod": [176]}}, "status": "modified"}]}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [], "analysis": {"iss_type": "3", "iss_reason": "5", "loc_way": "comment", "loc_scope": "0", "info_type": "Code"}, "max_topk": 2, "file_topk": 2, "loctype": {"code": ["src/transformers/modeling_utils.py", "src/transformers/utils/quantization_config.py"], "doc": [], "test": [], "config": [], "asset": []}}
|
| 1010 |
{"organization": "huggingface", "repo_name": "transformers", "base_commit": "705ca7f21b2b557e0cfd5d0853b297fa53489d20", "is_iss": 1, "iss_html_url": "https://github.com/huggingface/transformers/issues/14938", "iss_label": "", "title": "Question: Object of type EncoderDecoderConfig is not JSON serializable", "body": "Hi.\r\nAn error occurred when I used Trainer to train and save EncoderDecoderModel.\r\n\r\n```python\r\n File \"/home/jwli/ljw/study/hotpotqa/roberta_seq2seq/roberta_for_seq2seq.py\", line 482, in <module>\r\n run(model_args, data_args, training_args)\r\n File \"/home/jwli/ljw/study/hotpotqa/roberta_seq2seq/roberta_for_seq2seq.py\", line 465, in run\r\n train_result = trainer.train(resume_from_checkpoint=checkpoint)\r\n File \"/home/jwli/anaconda3/envs/study/lib/python3.7/site-packages/transformers/trainer.py\", line 1391, in train\r\n self._maybe_log_save_evaluate(tr_loss, model, trial, epoch, ignore_keys_for_eval)\r\n File \"/home/jwli/anaconda3/envs/study/lib/python3.7/site-packages/transformers/trainer.py\", line 1495, in _maybe_log_save_evaluate\r\n self._save_checkpoint(model, trial, metrics=metrics)\r\n File \"/home/jwli/anaconda3/envs/study/lib/python3.7/site-packages/transformers/trainer.py\", line 1557, in _save_checkpoint\r\n self.save_model(output_dir)\r\n File \"/home/jwli/anaconda3/envs/study/lib/python3.7/site-packages/transformers/trainer.py\", line 1961, in save_model\r\n self._save(output_dir)\r\n File \"/home/jwli/anaconda3/envs/study/lib/python3.7/site-packages/transformers/trainer.py\", line 2009, in _save\r\n self.model.save_pretrained(output_dir, state_dict=state_dict)\r\n File \"/home/jwli/anaconda3/envs/study/lib/python3.7/site-packages/transformers/modeling_utils.py\", line 1053, in save_pretrained\r\n model_to_save.config.save_pretrained(save_directory)\r\n File \"/home/jwli/anaconda3/envs/study/lib/python3.7/site-packages/transformers/configuration_utils.py\", line 416, in save_pretrained\r\n self.to_json_file(output_config_file, use_diff=True)\r\n File \"/home/jwli/anaconda3/envs/study/lib/python3.7/site-packages/transformers/configuration_utils.py\", line 739, in to_json_file\r\n writer.write(self.to_json_string(use_diff=use_diff))\r\n File \"/home/jwli/anaconda3/envs/study/lib/python3.7/site-packages/transformers/configuration_utils.py\", line 725, in to_json_string\r\n return json.dumps(config_dict, indent=2, sort_keys=True) + \"\\n\"\r\n File \"/home/jwli/anaconda3/envs/study/lib/python3.7/json/__init__.py\", line 238, in dumps\r\n **kw).encode(obj)\r\n File \"/home/jwli/anaconda3/envs/study/lib/python3.7/json/encoder.py\", line 201, in encode\r\n chunks = list(chunks)\r\n File \"/home/jwli/anaconda3/envs/study/lib/python3.7/json/encoder.py\", line 431, in _iterencode\r\n yield from _iterencode_dict(o, _current_indent_level)\r\n File \"/home/jwli/anaconda3/envs/study/lib/python3.7/json/encoder.py\", line 405, in _iterencode_dict\r\n yield from chunks\r\n File \"/home/jwli/anaconda3/envs/study/lib/python3.7/json/encoder.py\", line 438, in _iterencode\r\n o = _default(o)\r\n File \"/home/jwli/anaconda3/envs/study/lib/python3.7/json/encoder.py\", line 179, in default\r\n raise TypeError(f'Object of type {o.__class__.__name__} '\r\nTypeError: Object of type EncoderDecoderConfig is not JSON serializable\r\n```\r\nMy model and Config define the following code. \r\n```python\r\n tokenizer = RobertaTokenizerFast.from_pretrained(model_args.tokenizer_name)\r\n encoder_config = RobertaConfig.from_pretrained(model_args.encoder_model_name_or_path)\r\n decoder_config = RobertaConfig.from_pretrained(model_args.decoder_model_name_or_path)\r\n encoder_decoder_config = EncoderDecoderConfig.from_encoder_decoder_configs(encoder_config, decoder_config)\r\n model = RobertaForSeq2Seq.from_encoder_decoder_pretrained(model_args.encoder_model_name_or_path,\r\n model_args.decoder_model_name_or_path,\r\n config=encoder_decoder_config, tie_encoder_decoder=True)\r\n model.config.decoder_start_token_id = tokenizer.bos_token_id\r\n model.config.eos_token_id = tokenizer.eos_token_id\r\n model.config.max_length = 64\r\n model.config.early_stopping = True\r\n model.config.no_repeat_ngram_size = 3\r\n model.config.length_penalty = 2.0\r\n model.config.num_beams = 4\r\n model.config.pad_token_id = tokenizer.pad_token_id\r\n```\r\nThis error occurred because EncoderDecoderConfig cannot be converted to json format. But I don't know how to modify it.\r\n```python\r\nERROR OCCURRED:\r\n\r\n if use_diff is True:\r\n config_dict = self.to_diff_dict()\r\n else:\r\n config_dict = self.to_dict()\r\n return json.dumps(config_dict, indent=2, sort_keys=True) + \"\\n\"\r\n```\r\n\r\nI look forward to your help! Thanks!\r\n @jplu @patrickvonplaten ", "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {}, "own_code_loc": [{"Loc": [46, 47], "path": null}], "ass_file_loc": [], "other_rep_loc": [], "analysis": {"iss_type": "1", "iss_reason": "3", "loc_way": "comment", "loc_scope": "3", "info_type": "Code"}, "loctype": {"code": [null], "doc": [], "test": [], "config": [], "asset": []}}
|
|
@@ -1085,7 +1085,7 @@
|
|
| 1085 |
{"organization": "abi", "repo_name": "screenshot-to-code", "base_commit": "fcd305d0d26e7ef7b93dd605cbd5ed0e1a5a5e9c", "is_iss": 0, "iss_html_url": "https://github.com/abi/screenshot-to-code/issues/150", "iss_label": "", "title": "Error generating code. Check the Developer Console AND the backend logs for details", "body": "My ChatGPT has access to GPT-VISION. and the web app loads well but when I upload an image. it returns this error 'Error generating code. Check the Developer Console AND the backend logs for details'\r\n<img width=\"466\" alt=\"error\" src=\"https://github.com/abi/screenshot-to-code/assets/100529823/97c337b7-de54-45f9-8def-f984ade50a6d\">\r\n", "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {"base_commit": "fcd305d0d26e7ef7b93dd605cbd5ed0e1a5a5e9c", "files": [{"path": "docker-compose.yml", "Loc": {"(None, None, 20)": {"mod": [20]}}, "status": "modified"}]}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [], "analysis": {"iss_type": "1", "iss_reason": "3", "loc_way": "comment", "loc_scope": "0", "info_type": "Code"}, "max_topk": 1, "file_topk": 1, "loctype": {"code": [], "doc": ["docker-compose.yml"], "test": [], "config": [], "asset": []}}
|
| 1086 |
{"organization": "pytorch", "repo_name": "pytorch", "base_commit": "4622b3395276b37e10141fab43ffea33941ca0c2", "is_iss": 0, "iss_html_url": "https://github.com/pytorch/pytorch/issues/2384", "iss_label": "", "title": "How the grad is transferred between layer", "body": "consider a simple example here:\r\n```python\r\nimport torch\r\nfrom torch.autograd import Variable\r\n\r\ninput = Variable(torch.randn(20, 3, 28, 28), requires_grad=True)\r\nm = torch.nn.Conv2d(3, 16, 5)\r\noutput = m(input)\r\n\r\nloss = torch.sum(output)# define loss to perform backprop\r\nm.zero_grad()\r\nloss.backward()\r\n\r\nprint(type(input))\r\nprint(input.grad.size())\r\nprint(type(output))\r\nprint(output.grad)\r\n```\r\nthe output is:\r\n```\r\n<class 'torch.autograd.variable.Variable'>\r\ntorch.Size([20, 3, 28, 28])\r\n<class 'torch.autograd.variable.Variable'>\r\nNone\r\n```\r\nI find the `output.grad` is `None`. I don't know how the `input.grad` is calculated without `output.grad`.\r\nand want to know how to get the values of `output.grad`.\r\n\r\nthanks!", "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {"base_commit": "4622b3395276b37e10141fab43ffea33941ca0c2", "files": [{"path": "torch/autograd/variable.py", "Loc": {"('Variable', 'retain_grad', 236)": {"mod": []}}, "status": "modified"}]}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [], "analysis": {"iss_type": "3", "iss_reason": "3", "loc_way": "comment", "loc_scope": "0", "info_type": "Code"}, "max_topk": 1, "file_topk": 1, "loctype": {"code": ["torch/autograd/variable.py"], "doc": [], "test": [], "config": [], "asset": []}}
|
| 1087 |
{"organization": "pytorch", "repo_name": "pytorch", "base_commit": "2abcafcfd8beb4f6a22e08532d58f9f09c490f0f", "is_iss": 0, "iss_html_url": "https://github.com/pytorch/pytorch/issues/96983", "iss_label": "module: binaries\ntriaged\nmodule: arm", "title": "PyTorch 2.0 aarch64 wheels are missing the mkldnn+acl backend support", "body": "### \ud83d\udc1b Describe the bug\r\n\r\nPyTorch 2.0 aarch64 wheels are missing the mkldnn+acl backend support, where as PyTorch 1.13.0 had support.\r\n\r\nSolution:\r\nthe wheels need to be built with the `--enable-mkldnn` option while building them from the pytorch/builder repo.\r\n\r\nexample command for pytorch wheel builder script:\r\n`./build_aarch64_wheel.py --python-version 3.8 --use-docker --keep-running --os ubuntu20_04 --enable-mkldnn --branch release/2.0`\r\n\r\nTo reproduce the issue, create c6g or c7g instance from AWS EC2, and in the below output, look for `USE_MKLDNN=`, this was ON for PyTorch 1.13.0 but OFF for PyTorch2.0.0.\r\n\r\nnon-working scenario\r\n```\r\npip install torch==2.0.0\r\n\r\ntime python3 -c \"import torch; torch.set_num_threads(8); print(torch.__version__, torch.__config__.show(), torch.get_num_threads());a=torch.rand(100, 100, 100); b=torch.rand(100,100, 100); [torch.bmm(a,b).sum() for i in range(1000)]\"\r\n2.0.0 PyTorch built with:\r\n - GCC 10.2\r\n - C++ Version: 201703\r\n - OpenMP 201511 (a.k.a. OpenMP 4.5)\r\n - LAPACK is enabled (usually provided by MKL)\r\n - NNPACK is enabled\r\n - CPU capability usage: NO AVX\r\n - Build settings: BLAS_INFO=open, BUILD_TYPE=Release, CXX_COMPILER=/opt/rh/devtoolset-10/root/usr/bin/c++, CXX_FLAGS= -D_GLIBCXX_USE_CXX11_ABI=0 -fabi-version=11 -Wno-deprecated -fvisibility-inlines-hidden -DUSE_PTHREADPOOL -DNDEBUG -DUSE_KINETO -DLIBKINETO_NOCUPTI -DLIBKINETO_NOROCTRACER -DUSE_QNNPACK -DUSE_PYTORCH_QNNPACK -DUSE_XNNPACK -DSYMBOLICATE_MOBILE_DEBUG_HANDLE -O2 -fPIC -Wall -Wextra -Werror=return-type -Werror=non-virtual-dtor -Werror=bool-operation -Wnarrowing -Wno-missing-field-initializers -Wno-type-limits -Wno-array-bounds -Wno-unknown-pragmas -Wunused-local-typedefs -Wno-unused-parameter -Wno-unused-function -Wno-unused-result -Wno-strict-overflow -Wno-strict-aliasing -Wno-error=deprecated-declarations -Wno-stringop-overflow -Wno-psabi -Wno-error=pedantic -Wno-error=redundant-decls -Wno-error=old-style-cast -fdiagnostics-color=always -faligned-new -Wno-unused-but-set-variable -Wno-maybe-uninitialized -fno-math-errno -fno-trapping-math -Werror=format -Werror=cast-function-type -Wno-stringop-overflow, LAPACK_INFO=open, TORCH_DISABLE_GPU_ASSERTS=ON, TORCH_VERSION=2.0.0, USE_CUDA=OFF, USE_CUDNN=OFF, USE_EIGEN_FOR_BLAS=ON, USE_EXCEPTION_PTR=1, USE_GFLAGS=OFF, USE_GLOG=OFF, USE_MKL=OFF, USE_MKLDNN=OFF, USE_MPI=OFF, USE_NCCL=OFF, USE_NNPACK=ON, USE_OPENMP=ON, USE_ROCM=OFF, \r\n\r\n```\r\n\r\nworking scenario:\r\n\r\n```\r\npip3 install torch==1.13.0\r\n\r\ntime python3 -c \"import torch; torch.set_num_threads(8); print(torch.__version__, torch.__config__.show(), torch.get_num_threads());a=torch.rand(100, 100, 100); b=torch.rand(100,100, 100); [torch.bmm(a,b).sum() for i in range(1000)]\"\r\n\r\n1.13.0 PyTorch built with:\r\n - GCC 10.2\r\n - C++ Version: 201402\r\n - Intel(R) MKL-DNN v2.6.0 (Git Hash 52b5f107dd9cf10910aaa19cb47f3abf9b349815)\r\n - OpenMP 201511 (a.k.a. OpenMP 4.5)\r\n - LAPACK is enabled (usually provided by MKL)\r\n - NNPACK is enabled\r\n - CPU capability usage: NO AVX\r\n - Build settings: BLAS_INFO=open, BUILD_TYPE=Release, CXX_COMPILER=/opt/rh/devtoolset-10/root/usr/bin/c++, CXX_FLAGS= -Wno-deprecated -fvisibility-inlines-hidden -DUSE_PTHREADPOOL -fopenmp -DNDEBUG -DUSE_KINETO -DLIBKINETO_NOCUPTI -DUSE_QNNPACK -DUSE_PYTORCH_QNNPACK -DUSE_XNNPACK -DSYMBOLICATE_MOBILE_DEBUG_HANDLE -DEDGE_PROFILER_USE_KINETO -O2 -fPIC -Wno-narrowing -Wall -Wextra -Werror=return-type -Werror=non-virtual-dtor -Wno-missing-field-initializers -Wno-type-limits -Wno-array-bounds -Wno-unknown-pragmas -Wunused-local-typedefs -Wno-unused-parameter -Wno-unused-function -Wno-unused-result -Wno-strict-overflow -Wno-strict-aliasing -Wno-error=deprecated-declarations -Wno-stringop-overflow -Wno-psabi -Wno-error=pedantic -Wno-error=redundant-decls -Wno-error=old-style-cast -fdiagnostics-color=always -faligned-new -Wno-unused-but-set-variable -Wno-maybe-uninitialized -fno-math-errno -fno-trapping-math -Werror=format -Werror=cast-function-type -Wno-stringop-overflow, LAPACK_INFO=open, TORCH_VERSION=1.13.0, USE_CUDA=OFF, USE_CUDNN=OFF, USE_EIGEN_FOR_BLAS=ON, USE_EXCEPTION_PTR=1, USE_GFLAGS=OFF, USE_GLOG=OFF, USE_MKL=OFF, USE_MKLDNN=ON, USE_MPI=OFF, USE_NCCL=OFF, USE_NNPACK=ON, USE_OPENMP=ON, USE_ROCM=OFF, \r\n \r\n\r\n\r\n```\r\n\r\n### Versions\r\n```\r\nCollecting environment information...\r\nPyTorch version: 2.0.0\r\nIs debug build: False\r\nCUDA used to build PyTorch: None\r\nROCM used to build PyTorch: N/A\r\n\r\nOS: Ubuntu 20.04.5 LTS (aarch64)\r\nGCC version: (Ubuntu 10.3.0-1ubuntu1~20.04) 10.3.0\r\nClang version: Could not collect\r\nCMake version: version 3.25.2\r\nLibc version: glibc-2.31\r\n\r\nPython version: 3.8.10 (default, Nov 14 2022, 12:59:47) [GCC 9.4.0] (64-bit runtime)\r\nPython platform: Linux-5.15.0-1028-aws-aarch64-with-glibc2.29\r\nIs CUDA available: False\r\nCUDA runtime version: No CUDA\r\nCUDA_MODULE_LOADING set to: N/A\r\nGPU models and configuration: No CUDA\r\nNvidia driver version: No CUDA\r\ncuDNN version: No CUDA\r\nHIP runtime version: N/A\r\nMIOpen runtime version: N/A\r\nIs XNNPACK available: True\r\n\r\nCPU:\r\nArchitecture: aarch64\r\nCPU op-mode(s): 32-bit, 64-bit\r\nByte Order: Little Endian\r\nCPU(s): 16\r\nOn-line CPU(s) list: 0-15\r\nThread(s) per core: 1\r\nCore(s) per socket: 16\r\nSocket(s): 1\r\nNUMA node(s): 1\r\nVendor ID: ARM\r\nModel: 1\r\nStepping: r1p1\r\nBogoMIPS: 2100.00\r\nL1d cache: 1 MiB\r\nL1i cache: 1 MiB\r\nL2 cache: 16 MiB\r\nL3 cache: 32 MiB\r\nNUMA node0 CPU(s): 0-15\r\nVulnerability Itlb multihit: Not affected\r\nVulnerability L1tf: Not affected\r\nVulnerability Mds: Not affected\r\nVulnerability Meltdown: Not affected\r\nVulnerability Mmio stale data: Not affected\r\nVulnerability Retbleed: Not affected\r\nVulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl\r\nVulnerability Spectre v1: Mitigation; __user pointer sanitization\r\nVulnerability Spectre v2: Mitigation; CSV2, BHB\r\nVulnerability Srbds: Not affected\r\nVulnerability Tsx async abort: Not affected\r\nFlags: fp asimd evtstrm aes pmull sha1 sha2 crc32 atomics fphp asimdhp cpuid asimdrdm jscvt fcma lrcpc dcpop sha3 sm3 sm4 asimddp sha512 sve asimdfhm dit uscat ilrcpc flagm ssbs paca pacg dcpodp svei8mm svebf16 i8mm bf16 dgh rng\r\n\r\nVersions of relevant libraries:\r\n[pip3] numpy==1.24.2\r\n[pip3] torch==2.0.0\r\n[pip3] torchvision==0.14.1\r\n[conda] Could not collect\r\n```\r\n\r\ncc @ezyang @seemethere @malfet", "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {"base_commit": "2abcafcfd8beb4f6a22e08532d58f9f09c490f0f", "files": [{"path": ".ci/aarch64_linux/build_aarch64_wheel.py", "Loc": {"(None, None, None)": {"mod": [8]}}, "status": "modified"}]}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [], "analysis": {"iss_type": "1", "iss_reason": "2", "loc_way": "comment", "loc_scope": "0", "info_type": "Code"}, "max_topk": 1, "file_topk": 1, "loctype": {"code": [".ci/aarch64_linux/build_aarch64_wheel.py"], "doc": [], "test": [], "config": [], "asset": []}}
|
| 1088 |
-
{"organization": "pytorch", "repo_name": "pytorch", "base_commit": "2dff0b3e918530719f7667cb31541f036a25e3f2", "is_iss": 1, "iss_html_url": "https://github.com/pytorch/pytorch/issues/48435", "iss_label": "", "title": "AttributeError: module 'torch.cuda' has no attribute 'comm'", "body": "## \u2753 Questions and Help\r\n\r\nI'm using torch 1.7.0, and get this kind of error\r\n\r\nmy torch is installed via \r\n\r\npip install torch==1.7.0+cu101 torchvision==0.8.1+cu101 torchaudio===0.7.0 -f https://download.pytorch.org/whl/torch_stable.html\r\n\r\nmy os is win10", "code": null, "pr_html_url": null, "commit_html_url": "https://github.com/facebookresearch/InterHand2.6M/commit/874eb9f740ef54c275433d1bd27f8fb8f6a8f17d", "file_loc": {}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [{"org": "facebookresearch", "pro": "InterHand2.6M", "path": {"base_commit": "874eb9f740ef54c275433d1bd27f8fb8f6a8f17d", "files": [{"path": "common/nets/module.py", "status": "modified", "Loc": {"('PoseNet', 'soft_argmax_1d', 41)": {"mod": [43]}}}]}}], "analysis": {"iss_type": "1", "iss_reason": "1", "loc_way": "commit", "loc_scope": "2", "info_type": "Code"}, "loctype": {"code": ["common/nets/module.py"], "doc": [], "test": [], "config": [], "asset": ["InterHand2.6M"]}}
|
| 1089 |
{"organization": "xtekky", "repo_name": "gpt4free", "base_commit": "e8f6013d0349229fd8f7d298952cfe56fc4b8761", "is_iss": 0, "iss_html_url": "https://github.com/xtekky/gpt4free/issues/2070", "iss_label": "bug\nstale", "title": "Liaobots and You don't work", "body": "Liaobots and You do not work, they give the following errors:\r\n\r\n```\r\nLiaobots: ResponseStatusError: Response 500: Error\r\n``` \r\n\r\n```\r\nYou: ResponseStatusError: Response 401: {\"status_code\":401,\"request_id\":\"request-id-live-183191e7-adc1-4838-8e29-6e0c5c3ca048\",\"error_type\":\"endpoint_not_authorized_for_sdk\",\"error_message\":\"The project owner has not authorized the SDK to call this endpoint. Please enable it in the dashboard to continue: https://stytch.com/dashboard/sdk-configuration.\",\"error_url\":\"https://stytch.com/docs/api/errors/401#endpoint_not_authorized_for_sdk\"}\r\n``` \r\n@xtekky @hlohaus ", "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {"base_commit": "e8f6013d0349229fd8f7d298952cfe56fc4b8761", "files": [{"path": "g4f/Provider/Liaobots.py", "Loc": {"('Liaobots', 'create_async_generator', 111)": {"mod": [149]}}, "status": "modified"}]}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [], "analysis": {"iss_type": "1", "iss_reason": "5", "loc_way": "comment", "loc_scope": "0", "info_type": "Code"}, "max_topk": 1, "file_topk": 1, "loctype": {"code": ["g4f/Provider/Liaobots.py"], "doc": [], "test": [], "config": [], "asset": []}}
|
| 1090 |
{"organization": "xtekky", "repo_name": "gpt4free", "base_commit": "fa2d608822540c9b73350bfa036e8822ade4e23f", "is_iss": 0, "iss_html_url": "https://github.com/xtekky/gpt4free/issues/2305", "iss_label": "stale", "title": "ValueError: Unknown model: dall-e-3", "body": "```\r\nC:\\Users\\MAX\\Desktop>pip install -U g4f[all]\r\nDefaulting to user installation because normal site-packages is not writeable\r\nRequirement already satisfied: g4f[all] in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (0.3.3.2)\r\nRequirement already satisfied: requests in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (2.32.3)\r\nRequirement already satisfied: aiohttp in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (3.9.3)\r\nRequirement already satisfied: brotli in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (1.1.0)\r\nRequirement already satisfied: pycryptodome in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (3.20.0)\r\nRequirement already satisfied: curl-cffi>=0.6.2 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (0.7.3)\r\nRequirement already satisfied: cloudscraper in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (1.2.71)\r\nRequirement already satisfied: certifi in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (2024.8.30)\r\nRequirement already satisfied: browser-cookie3 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (0.19.1)\r\nRequirement already satisfied: PyExecJS in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (1.5.1)\r\nRequirement already satisfied: duckduckgo-search>=5.0 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (6.3.2)\r\nRequirement already satisfied: beautifulsoup4 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (4.12.3)\r\nRequirement already satisfied: pywebview in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (5.2)\r\nRequirement already satisfied: platformdirs in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (4.2.2)\r\nRequirement already satisfied: plyer in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (2.1.0)\r\nRequirement already satisfied: cryptography in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (43.0.0)\r\nRequirement already satisfied: aiohttp-socks in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (0.8.4)\r\nRequirement already satisfied: pillow in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (10.2.0)\r\nRequirement already satisfied: cairosvg in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (2.7.1)\r\nRequirement already satisfied: werkzeug in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (3.0.1)\r\nRequirement already satisfied: flask in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (3.0.2)\r\nRequirement already satisfied: loguru in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (0.7.2)\r\nRequirement already satisfied: fastapi in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (0.109.2)\r\nRequirement already satisfied: uvicorn in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (0.27.0.post1)\r\nRequirement already satisfied: nest-asyncio in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (1.6.0)\r\nRequirement already satisfied: cffi>=1.12.0 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from curl-cffi>=0.6.2->g4f[all]) (1.17.0)\r\nRequirement already satisfied: typing-extensions in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from curl-cffi>=0.6.2->g4f[all]) (4.12.2)\r\nRequirement already satisfied: click>=8.1.7 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from duckduckgo-search>=5.0->g4f[all]) (8.1.7)\r\nRequirement already satisfied: primp>=0.6.4 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from duckduckgo-search>=5.0->g4f[all]) (0.6.4)\r\nRequirement already satisfied: aiosignal>=1.1.2 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from aiohttp->g4f[all]) (1.3.1)\r\nRequirement already satisfied: attrs>=17.3.0 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from aiohttp->g4f[all]) (23.2.0)\r\nRequirement already satisfied: frozenlist>=1.1.1 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from aiohttp->g4f[all]) (1.4.1)\r\nRequirement already satisfied: multidict<7.0,>=4.5 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from aiohttp->g4f[all]) (6.0.5)\r\nRequirement already satisfied: yarl<2.0,>=1.0 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from aiohttp->g4f[all]) (1.9.4)\r\nRequirement already satisfied: python-socks<3.0.0,>=2.4.3 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from python-socks[asyncio]<3.0.0,>=2.4.3->aiohttp-socks->g4f[all]) (2.4.4)\r\nRequirement already satisfied: soupsieve>1.2 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from beautifulsoup4->g4f[all]) (2.5)\r\nRequirement already satisfied: lz4 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from browser-cookie3->g4f[all]) (4.3.3)\r\nRequirement already satisfied: pycryptodomex in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from browser-cookie3->g4f[all]) (3.20.0)\r\nRequirement already satisfied: cairocffi in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from cairosvg->g4f[all]) (1.6.1)\r\nRequirement already satisfied: cssselect2 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from cairosvg->g4f[all]) (0.7.0)\r\nRequirement already satisfied: defusedxml in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from cairosvg->g4f[all]) (0.7.1)\r\nRequirement already satisfied: tinycss2 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from cairosvg->g4f[all]) (1.2.1)\r\nRequirement already satisfied: pyparsing>=2.4.7 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from cloudscraper->g4f[all]) (3.1.2)\r\nRequirement already satisfied: requests-toolbelt>=0.9.1 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from cloudscraper->g4f[all]) (1.0.0)\r\nRequirement already satisfied: charset-normalizer<4,>=2 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from requests->g4f[all]) (3.3.2)\r\nRequirement already satisfied: idna<4,>=2.5 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from requests->g4f[all]) (3.6)\r\nRequirement already satisfied: urllib3<3,>=1.21.1 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from requests->g4f[all]) (2.1.0)\r\nRequirement already satisfied: pydantic!=1.8,!=1.8.1,!=2.0.0,!=2.0.1,!=2.1.0,<3.0.0,>=1.7.4 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from fastapi->g4f[all]) (2.6.1)\r\nRequirement already satisfied: starlette<0.37.0,>=0.36.3 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from fastapi->g4f[all]) (0.36.3)\r\nRequirement already satisfied: Jinja2>=3.1.2 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from flask->g4f[all]) (3.1.3)\r\nRequirement already satisfied: itsdangerous>=2.1.2 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from flask->g4f[all]) (2.1.2)\r\nRequirement already satisfied: blinker>=1.6.2 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from flask->g4f[all]) (1.7.0)\r\nRequirement already satisfied: MarkupSafe>=2.1.1 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from werkzeug->g4f[all]) (2.1.5)\r\nRequirement already satisfied: colorama>=0.3.4 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from loguru->g4f[all]) (0.4.6)\r\nRequirement already satisfied: win32-setctime>=1.0.0 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from loguru->g4f[all]) (1.1.0)\r\nRequirement already satisfied: six>=1.10.0 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from PyExecJS->g4f[all]) (1.16.0)\r\nRequirement already satisfied: proxy-tools in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from pywebview->g4f[all]) (0.1.0)\r\nRequirement already satisfied: bottle in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from pywebview->g4f[all]) (0.13.1)\r\nRequirement already satisfied: pythonnet in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from pywebview->g4f[all]) (3.0.3)\r\nRequirement already satisfied: h11>=0.8 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from uvicorn->g4f[all]) (0.14.0)\r\nRequirement already satisfied: pycparser in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from cffi>=1.12.0->curl-cffi>=0.6.2->g4f[all]) (2.22)\r\nRequirement already satisfied: annotated-types>=0.4.0 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from pydantic!=1.8,!=1.8.1,!=2.0.0,!=2.0.1,!=2.1.0,<3.0.0,>=1.7.4->fastapi->g4f[all]) (0.6.0)\r\nRequirement already satisfied: pydantic-core==2.16.2 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from pydantic!=1.8,!=1.8.1,!=2.0.0,!=2.0.1,!=2.1.0,<3.0.0,>=1.7.4->fastapi->g4f[all]) (2.16.2)\r\nRequirement already satisfied: async-timeout>=3.0.1 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from python-socks[asyncio]<3.0.0,>=2.4.3->aiohttp-socks->g4f[all]) (4.0.3)\r\nRequirement already satisfied: anyio<5,>=3.4.0 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from starlette<0.37.0,>=0.36.3->fastapi->g4f[all]) (4.2.0)\r\nRequirement already satisfied: webencodings in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from cssselect2->cairosvg->g4f[all]) (0.5.1)\r\nRequirement already satisfied: clr-loader<0.3.0,>=0.2.6 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from pythonnet->pywebview->g4f[all]) (0.2.6)\r\nRequirement already satisfied: sniffio>=1.1 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from anyio<5,>=3.4.0->starlette<0.37.0,>=0.36.3->fastapi->g4f[all]) (1.3.0)\r\n\r\nC:\\Users\\MAX\\Desktop>\r\nTraceback (most recent call last):.py\r\n File \"C:\\Users\\MAX\\Desktop\\gptimg.py\", line 4, in <module>\r\n response = client.images.generate(\r\n ^^^^^^^^^^^^^^^^^^^^^^^\r\n File \"C:\\Users\\MAX\\AppData\\Local\\Packages\\PythonSoftwareFoundation.Python.3.12_qbz5n2kfra8p0\\LocalCache\\local-packages\\Python312\\site-packages\\g4f\\client\\client.py\", line 421, in generate\r\n return asyncio.run(self.async_generate(prompt, model, response_format=response_format, **kwargs))\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File \"C:\\Program Files\\WindowsApps\\PythonSoftwareFoundation.Python.3.12_3.12.2032.0_x64__qbz5n2kfra8p0\\Lib\\asyncio\\runners.py\", line 194, in run\r\n return runner.run(main)\r\n ^^^^^^^^^^^^^^^^\r\n File \"C:\\Program Files\\WindowsApps\\PythonSoftwareFoundation.Python.3.12_3.12.2032.0_x64__qbz5n2kfra8p0\\Lib\\asyncio\\runners.py\", line 118, in run\r\n return self._loop.run_until_complete(task)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File \"C:\\Program Files\\WindowsApps\\PythonSoftwareFoundation.Python.3.12_3.12.2032.0_x64__qbz5n2kfra8p0\\Lib\\asyncio\\base_events.py\", line 687, in run_until_complete\r\n return future.result()\r\n ^^^^^^^^^^^^^^^\r\n File \"C:\\Users\\MAX\\AppData\\Local\\Packages\\PythonSoftwareFoundation.Python.3.12_qbz5n2kfra8p0\\LocalCache\\local-packages\\Python312\\site-packages\\g4f\\client\\client.py\", line 426, in async_generate\r\n raise ValueError(f\"Unknown model: {model}\")\r\nValueError: Unknown model: dall-e-3\r\n```", "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {"base_commit": "fa2d608822540c9b73350bfa036e8822ade4e23f", "files": [{"path": "g4f/models.py", "Loc": {}}]}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [], "analysis": {"iss_type": "1", "iss_reason": "5", "loc_way": "comment", "loc_scope": "0", "info_type": "Code"}, "max_topk": 0, "file_topk": 0, "loctype": {"code": ["g4f/models.py"], "doc": [], "test": [], "config": [], "asset": []}}
|
| 1091 |
{"organization": "xtekky", "repo_name": "gpt4free", "base_commit": "1ade1d959cbc9aea7cf653bbe5b6c414ba486c97", "is_iss": 0, "iss_html_url": "https://github.com/xtekky/gpt4free/issues/1292", "iss_label": "bug\nstale", "title": "RecursionError: maximum recursion depth exceeded while calling a Python object", "body": "Ubuntu 22, g4f-0.1.9.0, pip installation method, python3.10\r\n\r\n**Bug description**\r\nG4F API has these errors after 5-10 requests. I have to restart constantly. It is very uncomfortable. This problem did not exist in the previous version.\r\n\r\n**Errors**\r\n```\r\nRecursionError: maximum recursion depth exceeded in comparison\r\nRecursionError: maximum recursion depth exceeded while calling a Python object\r\nRuntimeError: RetryProvider failed:\r\nYou: RecursionError: maximum recursion depth exceeded\r\nChatgpt4Online: RecursionError: maximum recursion depth exceeded in comparison\r\nChatAnywhere: RecursionError: maximum recursion depth exceeded while encoding a JSON object\r\nChatgptX: RecursionError: maximum recursion depth exceeded in comparison\r\nGptForLove: RuntimeUnavailableError: Could not find an available JavaScript runtime.\r\nChatBase: RecursionError: maximum recursion depth exceeded while encoding a JSON object\r\nGptGo: RecursionError: maximum recursion depth exceeded while calling a Python object\r\n```\r\n\r\n**Traceback**\r\n```\r\nERROR: Exception in ASGI application\r\nTraceback (most recent call last):\r\n File \"/usr/local/lib/python3.10/dist-packages/g4f/api/__init__.py\", line 85, in chat_completions\r\n response = g4f.ChatCompletion.create(\r\n File \"/usr/local/lib/python3.10/dist-packages/g4f/__init__.py\", line 76, in create\r\n return result if stream else ''.join(result)\r\n File \"/usr/local/lib/python3.10/dist-packages/g4f/Provider/retry_provider.py\", line 59, in create_completion\r\n self.raise_exceptions()\r\n File \"/usr/local/lib/python3.10/dist-packages/g4f/Provider/retry_provider.py\", line 87, in raise_exceptions\r\n raise RuntimeError(\"\\n\".join([\"RetryProvider failed:\"] + [\r\nRuntimeError: RetryProvider failed:\r\nChatAnywhere: RecursionError: maximum recursion depth exceeded\r\nChatBase: RecursionError: maximum recursion depth exceeded\r\nChatgptX: RecursionError: maximum recursion depth exceeded\r\nYou: RecursionError: maximum recursion depth exceeded while calling a Python object\r\nGptGo: RecursionError: maximum recursion depth exceeded\r\nChatgpt4Online: RecursionError: maximum recursion depth exceeded\r\nGptForLove: RecursionError: maximum recursion depth exceeded\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"/usr/local/lib/python3.10/dist-packages/uvicorn/protocols/http/h11_impl.py\", line 408, in run_asgi\r\n result = await app( # type: ignore[func-returns-value]\r\n File \"/usr/local/lib/python3.10/dist-packages/uvicorn/middleware/proxy_headers.py\", line 84, in __call__\r\n return await self.app(scope, receive, send)\r\n File \"/usr/local/lib/python3.10/dist-packages/fastapi/applications.py\", line 1106, in __call__\r\n await super().__call__(scope, receive, send)\r\n File \"/usr/local/lib/python3.10/dist-packages/starlette/applications.py\", line 122, in __call__\r\n await self.middleware_stack(scope, receive, send)\r\n File \"/usr/local/lib/python3.10/dist-packages/starlette/middleware/errors.py\", line 184, in __call__\r\n raise exc\r\n File \"/usr/local/lib/python3.10/dist-packages/starlette/middleware/errors.py\", line 162, in __call__\r\n await self.app(scope, receive, _send)\r\n File \"/usr/local/lib/python3.10/dist-packages/starlette/middleware/exceptions.py\", line 79, in __call__\r\n raise exc\r\n File \"/usr/local/lib/python3.10/dist-packages/starlette/middleware/exceptions.py\", line 68, in __call__\r\n await self.app(scope, receive, sender)\r\n File \"/usr/local/lib/python3.10/dist-packages/fastapi/middleware/asyncexitstack.py\", line 20, in __call__\r\n raise e\r\n File \"/usr/local/lib/python3.10/dist-packages/fastapi/middleware/asyncexitstack.py\", line 17, in __call__\r\n await self.app(scope, receive, send)\r\n File \"/usr/local/lib/python3.10/dist-packages/starlette/routing.py\", line 718, in __call__\r\n await route.handle(scope, receive, send)\r\n File \"/usr/local/lib/python3.10/dist-packages/starlette/routing.py\", line 276, in handle\r\n await self.app(scope, receive, send)\r\n File \"/usr/local/lib/python3.10/dist-packages/starlette/routing.py\", line 66, in app\r\n response = await func(request)\r\n File \"/usr/local/lib/python3.10/dist-packages/fastapi/routing.py\", line 274, in app\r\n raw_response = await run_endpoint_function(\r\n File \"/usr/local/lib/python3.10/dist-packages/fastapi/routing.py\", line 191, in run_endpoint_function\r\n return await dependant.call(**values)\r\n File \"/usr/local/lib/python3.10/dist-packages/g4f/api/__init__.py\", line 91, in chat_completions\r\n logging.exception(e)\r\n File \"/usr/lib/python3.10/logging/__init__.py\", line 2113, in exception\r\n error(msg, *args, exc_info=exc_info, **kwargs)\r\n File \"/usr/lib/python3.10/logging/__init__.py\", line 2105, in error\r\n root.error(msg, *args, **kwargs)\r\n File \"/usr/lib/python3.10/logging/__init__.py\", line 1506, in error\r\n self._log(ERROR, msg, args, **kwargs)\r\n File \"/usr/lib/python3.10/logging/__init__.py\", line 1624, in _log\r\n self.handle(record)\r\n File \"/usr/lib/python3.10/logging/__init__.py\", line 1634, in handle\r\n self.callHandlers(record)\r\n File \"/usr/lib/python3.10/logging/__init__.py\", line 1696, in callHandlers\r\n hdlr.handle(record)\r\n File \"/usr/lib/python3.10/logging/__init__.py\", line 968, in handle\r\n self.emit(record)\r\n File \"/usr/lib/python3.10/logging/__init__.py\", line 1100, in emit\r\n msg = self.format(record)\r\n File \"/usr/lib/python3.10/logging/__init__.py\", line 943, in format\r\n return fmt.format(record)\r\n File \"/usr/lib/python3.10/logging/__init__.py\", line 686, in format\r\n record.exc_text = self.formatException(record.exc_info)\r\n File \"/usr/lib/python3.10/logging/__init__.py\", line 636, in formatException\r\n traceback.print_exception(ei[0], ei[1], tb, None, sio)\r\n File \"/usr/lib/python3.10/traceback.py\", line 120, in print_exception\r\n for line in te.format(chain=chain):\r\n File \"/usr/local/lib/python3.10/dist-packages/exceptiongroup/_formatting.py\", line 248, in format\r\n yield from _ctx.emit(exc.format_exception_only())\r\n File \"/usr/local/lib/python3.10/dist-packages/exceptiongroup/_formatting.py\", line 64, in emit\r\n for text in text_gen:\r\n File \"/usr/local/lib/python3.10/dist-packages/exceptiongroup/_formatting.py\", line 335, in format_exception_only\r\n if isinstance(self.__notes__, collections.abc.Sequence):\r\n File \"/usr/lib/python3.10/abc.py\", line 119, in __instancecheck__\r\n return _abc_instancecheck(cls, instance)\r\nRecursionError: maximum recursion depth exceeded in comparison\r\n```\r\n\r\n\r\n", "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {"base_commit": "1ade1d959cbc9aea7cf653bbe5b6c414ba486c97", "files": [{"path": "g4f/cli.py", "Loc": {}}]}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [], "analysis": {"iss_type": "1", "iss_reason": "5", "loc_way": "comment", "loc_scope": "0", "info_type": "Code"}, "max_topk": 0, "file_topk": 0, "loctype": {"code": ["g4f/cli.py"], "doc": [], "test": [], "config": [], "asset": []}}
|
|
|
|
| 887 |
{"organization": "binary-husky", "repo_name": "gpt_academic", "base_commit": "bfa6661367b7592e82225515e5e4845c4aad95bb", "is_iss": 0, "iss_html_url": "https://github.com/binary-husky/gpt_academic/issues/252", "iss_label": "", "title": "\u80fd\u4e0d\u80fd\u4f7f\u7528azure openai key?", "body": "\u4ee3\u7406\u670d\u52a1\u5668\u4e0d\u591f\u7a33\u5b9a\uff0c\u66f4\u9ebb\u70e6\u7684\u662f\u7ed9openai\u7eed\u8d39\uff0c\u8fd8\u8981\u4e2a\u7f8e\u56fd\u4fe1\u7528\u5361\r\n\r\n\u975e\u5e38\u597d\u7684\u5e94\u7528\uff0c\u5e0c\u671b\u51fa\u66f4\u591a\u7684\u63d2\u4ef6\u529f\u80fd\uff0c\u8c22\u8c22", "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {"base_commit": "bfa6661367b7592e82225515e5e4845c4aad95bb", "files": [{"path": "config.py", "Loc": {}}]}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [], "analysis": {"iss_type": "3", "iss_reason": "5", "loc_way": "comment", "loc_scope": "0", "info_type": "Code"}, "max_topk": 0, "file_topk": 0, "loctype": {"code": ["config.py"], "doc": [], "test": [], "config": [], "asset": []}}
|
| 888 |
{"organization": "binary-husky", "repo_name": "gpt_academic", "base_commit": "2d2e02040d7d91d2f2a4c34f4d0bf677873b5f4d", "is_iss": 0, "iss_html_url": "https://github.com/binary-husky/gpt_academic/issues/1328", "iss_label": "", "title": "[Bug]: \u7cbe\u51c6\u7ffb\u8bd1PDF\u6587\u6863(NOUGAT)\u529f\u80fd\u51fa\u9519\uff0c", "body": "### Installation Method | \u5b89\u88c5\u65b9\u6cd5\u4e0e\u5e73\u53f0\n\nOthers (Please Describe)\n\n### Version | \u7248\u672c\n\nPlease choose | \u8bf7\u9009\u62e9\n\n### OS | \u64cd\u4f5c\u7cfb\u7edf\n\nPlease choose | \u8bf7\u9009\u62e9\n\n### Describe the bug | \u7b80\u8ff0\n\n\u6d4b\u8bd5\u670d\u52a1\u5668\uff0c\u7cbe\u51c6\u7ffb\u8bd1PDF\u6587\u6863(NOUGAT)\u529f\u80fd\u51fa\u9519\uff0c\u4f46\u662f\u53ef\u4ee5\u4f7f\u7528\u7cbe\u51c6\u7ffb\u8bd1PDF\u7684\u529f\u80fd\r\n\r\n\r\n\r\n\u62a5\u9519\u4fe1\u606f\u5982\u4e0b\r\n\r\n\n\n### Screen Shot | \u6709\u5e2e\u52a9\u7684\u622a\u56fe\n\n\r\nTraceback (most recent call last):\r\n File \"./toolbox.py\", line 159, in decorated\r\n yield from f(main_input, llm_kwargs, plugin_kwargs, chatbot_with_cookie, history, *args, **kwargs)\r\n File \"./crazy_functions/\u6279\u91cf\u7ffb\u8bd1PDF\u6587\u6863_NOUGAT.py\", line 93, in \u6279\u91cf\u7ffb\u8bd1PDF\u6587\u6863\r\n yield from \u89e3\u6790PDF_\u57fa\u4e8eNOUGAT(file_manifest, project_folder, llm_kwargs, plugin_kwargs, chatbot, history, system_prompt)\r\n File \"./crazy_functions/\u6279\u91cf\u7ffb\u8bd1PDF\u6587\u6863_NOUGAT.py\", line 111, in \u89e3\u6790PDF_\u57fa\u4e8eNOUGAT\r\n fpp = yield from nougat_handle.NOUGAT_parse_pdf(fp, chatbot, history)\r\n File \"./crazy_functions/crazy_utils.py\", line 761, in NOUGAT_parse_pdf\r\n raise RuntimeError(\"Nougat\u89e3\u6790\u8bba\u6587\u5931\u8d25\u3002\")\r\nRuntimeError: Nougat\u89e3\u6790\u8bba\u6587\u5931\u8d25\u3002\n\n### Terminal Traceback & Material to Help Reproduce Bugs | \u7ec8\u7aeftraceback\uff08\u5982\u6709\uff09 + \u5e2e\u52a9\u6211\u4eec\u590d\u73b0\u7684\u6d4b\u8bd5\u6750\u6599\u6837\u672c\uff08\u5982\u6709\uff09\n\n_No response_", "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {"base_commit": "2d2e02040d7d91d2f2a4c34f4d0bf677873b5f4d", "files": [{"path": "crazy_functions/crazy_utils.py", "Loc": {"('nougat_interface', 'NOUGAT_parse_pdf', 739)": {"mod": [752]}, "('nougat_interface', None, 719)": {"mod": [723]}}, "status": "modified"}]}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [], "analysis": {"iss_type": "1", "iss_reason": "5", "loc_way": "comment", "loc_scope": "0", "info_type": "Code"}, "max_topk": 2, "file_topk": 1, "loctype": {"code": ["crazy_functions/crazy_utils.py"], "doc": [], "test": [], "config": [], "asset": []}}
|
| 889 |
{"organization": "binary-husky", "repo_name": "gpt_academic", "base_commit": "17abd29d5035b5b227deaad69d32cf437b23e542", "is_iss": 0, "iss_html_url": "https://github.com/binary-husky/gpt_academic/issues/94", "iss_label": "", "title": "[\u4e00\u4e9b\u5efa\u8bae]input\u6846\u8fd8\u662f\u592a\u5c0f\u4e86", "body": "RT \u591a\u884c\u8f93\u5165\u8fd8\u662f\u4e0d\u65b9\u4fbf\uff0c\u5982\u679c\u9002\u5f53\u8c03\u6574\u4f1a\u66f4\u597d\u7528\u3002\r\n\r\n\u5e0c\u671b\u91c7\u7eb3\uff0c\u611f\u8c22\u5206\u4eab\u3002", "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {"base_commit": "17abd29d5035b5b227deaad69d32cf437b23e542", "files": [{"path": "main.py", "Loc": {"(None, None, None)": {"mod": [1]}}, "status": "modified"}]}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [], "analysis": {"iss_type": "4", "iss_reason": "5", "loc_way": "comment", "loc_scope": "0", "info_type": "Code"}, "max_topk": 1, "file_topk": 1, "loctype": {"code": ["main.py"], "doc": [], "test": [], "config": [], "asset": []}}
|
| 890 |
+
{"organization": "binary-husky", "repo_name": "gpt_academic", "base_commit": "37744a9cb173477398a2609f02d5e7cef47eb677", "is_iss": 1, "iss_html_url": "https://github.com/binary-husky/gpt_academic/issues/1438", "iss_label": "", "title": "[Bug]: \u6d6e\u52a8\u8f93\u5165\u6846\u5728\u62d6\u81f3\u9876\u90e8\u540e\uff0c\u65e0\u6cd5\u91cd\u65b0\u79fb\u4f4d", "body": "### Installation Method | \u5b89\u88c5\u65b9\u6cd5\u4e0e\u5e73\u53f0\n\nOthers (Please Describe)\n\n### Version | \u7248\u672c\n\nPlease choose | \u8bf7\u9009\u62e9\n\n### OS | \u64cd\u4f5c\u7cfb\u7edf\n\nMac\n\n### Describe the bug | \u7b80\u8ff0\n\n\u6d6e\u52a8\u8f93\u5165\u6846\u5728\u62d6\u81f3\u9876\u90e8\u540e\uff0c\u65e0\u6cd5\u91cd\u65b0\u79fb\u4f4d\r\n\r\n\u671f\u671b\uff1a\u91cd\u65b0\u52fe\u9009\u540e\uff0c\u5e94\u8be5\u56de\u5230\u521d\u59cb\u4f4d\u7f6e\n\n### Screen Shot | \u6709\u5e2e\u52a9\u7684\u622a\u56fe\n\n\n\n### Terminal Traceback & Material to Help Reproduce Bugs | \u7ec8\u7aeftraceback\uff08\u5982\u6709\uff09 + \u5e2e\u52a9\u6211\u4eec\u590d\u73b0\u7684\u6d4b\u8bd5\u6750\u6599\u6837\u672c\uff08\u5982\u6709\uff09\n\n_No response_", "code": null, "pr_html_url": null, "commit_html_url": "https://github.com/binary-husky/gradio-fix/commit/fb67dd12f58aa53c75a90378cddbc811ac3c01d2", "file_loc": {}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [{"org": "binary-husky", "pro": "gradio-fix", "path": [{"base_commit": "fb67dd12f58aa53c75a90378cddbc811ac3c01d2", "files": [{"path": "js/app/src/components/Floating/StaticFloating.svelte", "status": "modified", "Loc": {"(None, None, 48)": {"add": [48]}}}]}]}], "analysis": {"iss_type": "2", "iss_reason": "1", "loc_way": "commit", "loc_scope": "0", "info_type": "Code"}, "loctype": {"code": [], "doc": [], "test": [], "config": [], "asset": ["gradio-fix", "js/app/src/components/Floating/StaticFloating.svelte"]}}
|
| 891 |
{"organization": "binary-husky", "repo_name": "gpt_academic", "base_commit": "6538c58b8e5a4a7ae08dfa1ae9970bc422158096", "is_iss": 0, "iss_html_url": "https://github.com/binary-husky/gpt_academic/issues/620", "iss_label": "", "title": "\u60f3\u95ee\u95eenewbing\u7684cookies\u600e\u4e48\u586b\u5199\uff0c\u6211\u4ecejavascript:alert(document.cookie)\u627e\u5230\u4e86cookies\u4f46\u662f\u4e00\u76f4\u663e\u793acookies\u6709\u9519", "body": "\r\n\r\n", "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {"base_commit": "6538c58b8e5a4a7ae08dfa1ae9970bc422158096", "files": [{"path": "config.py", "Loc": {"(None, None, None)": {"mod": [69]}}, "status": "modified"}]}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [], "analysis": {"iss_type": "1", "iss_reason": "5", "loc_way": "comment", "loc_scope": "", "info_type": "Other"}, "max_topk": 1, "file_topk": 1, "loctype": {"code": ["config.py"], "doc": [], "test": [], "config": [], "asset": []}}
|
| 892 |
{"organization": "binary-husky", "repo_name": "gpt_academic", "base_commit": "6d8c8cd3f0b9d2b6fe8d412b83f902cbd43fa0bd", "is_iss": 0, "iss_html_url": "https://github.com/binary-husky/gpt_academic/issues/150", "iss_label": "documentation\nhigh value issue", "title": "\u6709\u6ca1\u6709\u5b8c\u5168\u90e8\u7f72\u6210\u529f\u7684\u5927\u795e\u51fa\u4e2a\u8be6\u7ec6\u7684\u90e8\u7f72\u6b65\u9aa4\u5440\uff1fWindows \u6709\u622a\u56fe\uff0c\u8dea\u6c42", "body": "Windows\u5b89\u88c5\u90e8\u7f72\r\n\u57fa\u672c\u73af\uff1a\u5b89\u88c5anaconda\r\n1.\u4e0b\u8f7d\u9879\u76ee CMD\r\n\u9009\u62e9\u8def\u5f84\r\ngit clone https://github.com/binary-husky/chatgpt_academic.git\r\ncd chatgpt_academic\r\n\u6211\u4eec\u5efa\u8bae\u5c06config.py\u590d\u5236\u4e3aconfig_private.py\u5e76\u5c06\u540e\u8005\u7528\u4f5c\u4e2a\u6027\u5316\u914d\u7f6e\u6587\u4ef6\u4ee5\u907f\u514dconfig.py\u4e2d\u7684\u53d8\u66f4\u5f71\u54cd\u4f60\u7684\u4f7f\u7528\u6216\u4e0d\u5c0f\u5fc3\u5c06\u5305\u542b\u4f60\u7684OpenAI API KEY\u7684config.py\u63d0\u4ea4\u81f3\u672c\u9879\u76ee\u3002\r\ncp config.py config_private.py\r\n2.\u521b\u5efa\u865a\u62df\u73af\u5883 python 3.11\r\nconda create -n chatgpt python=3.11.0 #\u65b0\u5efa\u73af\u5883\u3001\r\n3.\u8fdb\u5165\u9879\u76ee\u4e0b\u8f7d\u8def\u5f84\r\n\u4f8b\u5982 cd G:\\python\\Program\\chatgpt_academic\r\n4.\u542f\u52a8\u865a\u62df\u73af\u5883\r\nconda activate chatgpt\r\n5. \u5b89\u88c5 gradio>=3.23\r\n\uff081\uff09\u5230https://pypi.org/project/gradio/ \u4e0b\u8f7dwhl\u7248\u672c\r\n\uff082\uff09pip install G:\\python\\Program\\chatgpt_academic\\gradio-3.23.0-py3-none-any.whl\r\n6.\u914d\u7f6e\u5176\u4ed6\u73af\u5883\r\n\uff081\uff09\u6253\u5f00requirements.txt\uff0c\u6ce8\u91ca\u6389gradio\uff0c\u7136\u540e\u4fdd\u5b58\r\n\uff082\uff09\u8fd0\u884c python -m pip install -r requirements.txt\r\n7.\u542f\u52a8\u4ee3\u7406\r\n8. \u914d\u7f6econfig_private.py\r\n\uff081\uff09\u6dfb\u52a0API_KEY\r\n\uff082\uff09\u4fee\u6539USE_PROXY = Ture\r\n\uff083\uff09\u4fee\u6539proxies\r\n\u5728\u6d4f\u89c8\u5668\u8f93\u5165: https://ipapi.co/json/\r\n\u6d4f\u89c8\u5668\u4e0a\u53f3\u952e->\u68c0\u67e5->\u7f51\u7edc->ctrl+r\r\n\u6253\u5f00json\uff0c\u5c06\u8fdc\u7a0b\u5730\u5740\u4fee\u6539\u5230proxies = { \"http\": \"104.26.9.44:443\", \"https\": \"104.26.9.44:443\", }\r\n9.\u542f\u52a8\u7a0b\u5e8f\r\npython main.py", "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {"base_commit": "6d8c8cd3f0b9d2b6fe8d412b83f902cbd43fa0bd", "files": [{"path": "requirements.txt", "Loc": {}}]}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [], "analysis": {"iss_type": "3", "iss_reason": "5", "loc_way": "comment", "loc_scope": "0", "info_type": "Code\n+ \nDoc"}, "max_topk": 0, "file_topk": 0, "loctype": {"code": [], "doc": [], "test": [], "config": ["requirements.txt"], "asset": []}}
|
| 893 |
{"organization": "binary-husky", "repo_name": "gpt_academic", "base_commit": "e20070939c6c7eeca33a8438041c9e038836957b", "is_iss": 1, "iss_html_url": "https://github.com/binary-husky/gpt_academic/issues/568", "iss_label": "enhancement", "title": "\u80fd\u5426\u589e\u52a0\u804a\u5929\u5185\u5bb9\u5bfc\u51fa\u529f\u80fd\uff1f", "body": null, "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {}, "own_code_loc": [], "ass_file_loc": ["gpt_log/chat_secrets.log"], "other_rep_loc": [], "analysis": {"iss_type": "4", "iss_reason": "5", "loc_way": "comment", "loc_scope": "1", "info_type": "Config"}, "loctype": {"code": [], "doc": [], "test": [], "config": [], "asset": ["gpt_log/chat_secrets.log"]}}
|
|
|
|
| 1004 |
{"organization": "huggingface", "repo_name": "transformers", "base_commit": "82c7e879876822864b5ceaf2c99eb01159266bcd", "is_iss": 0, "iss_html_url": "https://github.com/huggingface/transformers/issues/27200", "iss_label": "", "title": "dataset download error in speech recognition examples", "body": "### System Info\n\n- `transformers` version: 4.35.0.dev0\r\n- Platform: Linux-5.15.0-43-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.18\r\n- Huggingface_hub version: 0.17.3\r\n- Safetensors version: 0.4.0\r\n- Accelerate version: 0.24.1\r\n- Accelerate config: not found\r\n- PyTorch version (GPU?): 1.10.0+cu111 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Flax version (CPU?/GPU?/TPU?): not installed (NA)\r\n- Jax version: not installed\r\n- JaxLib version: not installed\r\n- Using GPU in script?: <fill in>\r\n- Using distributed or parallel set-up in script?: <fill in>\n\n### Who can help?\n\n@stevhliu and @MKhalusova\n\n### Information\n\n- [x] The official example scripts\n- [ ] My own modified scripts\n\n### Tasks\n\n- [X] An officially supported task in the `examples` folder (such as GLUE/SQuAD, ...)\n- [ ] My own task or dataset (give details below)\n\n### Reproduction\n\nCUDA_VISIBLE_DEVICES=0 python run_speech_recognition_ctc.py \\\r\n\t--dataset_name=\"common_voice\" \\\r\n\t--model_name_or_path=\"facebook/wav2vec2-large-xlsr-53\" \\\r\n\t--dataset_config_name=\"tr\" \\\r\n\t--output_dir=\"./wav2vec2-common_voice-tr-demo\" \\\r\n\t--overwrite_output_dir \\\r\n\t--num_train_epochs=\"15\" \\\r\n\t--per_device_train_batch_size=\"16\" \\\r\n\t--gradient_accumulation_steps=\"2\" \\\r\n\t--learning_rate=\"3e-4\" \\\r\n\t--warmup_steps=\"500\" \\\r\n\t--evaluation_strategy=\"steps\" \\\r\n\t--text_column_name=\"sentence\" \\\r\n\t--length_column_name=\"input_length\" \\\r\n\t--save_steps=\"400\" \\\r\n\t--eval_steps=\"100\" \\\r\n\t--layerdrop=\"0.0\" \\\r\n\t--save_total_limit=\"3\" \\\r\n\t--freeze_feature_encoder \\\r\n\t--gradient_checkpointing \\\r\n\t--chars_to_ignore , ? . ! - \\; \\: \\\" \u201c % \u2018 \u201d \ufffd \\\r\n\t--fp16 \\\r\n\t--group_by_length \\\r\n\t--push_to_hub \\\r\n\t--do_train --do_eval \n\n### Expected behavior\n\nWhen I run the default command, which set `dataset_name` as \"common_voice\", and I got a warning:\r\n```\r\n/home/xintong/.cache/huggingface/modules/datasets_modules/datasets/common_voice/220833898d6a60c50f621126e51fb22eb2dfe5244392c70dccd8e6e2f055f4bf/common_voice.py:634: FutureWarning: \r\n This version of the Common Voice dataset is deprecated.\r\n You can download the latest one with\r\n >>> load_dataset(\"mozilla-foundation/common_voice_11_0\", \"en\")\r\n \r\n warnings.warn(\r\nGenerating train split: 0%| | 0/1831 [00:00<?, ? examples/s]\r\nTraceback (most recent call last):\r\n File \"/home/xintong/miniconda3/envs/test/lib/python3.8/tarfile.py\", line 2578, in next\r\n tarinfo = self.tarinfo.fromtarfile(self)\r\n File \"/home/xintong/miniconda3/envs/test/lib/python3.8/tarfile.py\", line 1283, in fromtarfile\r\n obj = cls.frombuf(buf, tarfile.encoding, tarfile.errors)\r\n File \"/home/xintong/miniconda3/envs/test/lib/python3.8/tarfile.py\", line 1221, in frombuf\r\n raise TruncatedHeaderError(\"truncated header\")\r\ntarfile.TruncatedHeaderError: truncated header\r\n```\r\nI modified this into `mozilla-foundation/common_voice_11_0`, it passed. \r\n```\r\nDownloading builder script: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 8.13k/8.13k [00:00<00:00, 30.3MB/s]\r\nDownloading readme: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 14.4k/14.4k [00:00<00:00, 19.2MB/s]\r\nDownloading extra modules: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 3.44k/3.44k [00:00<00:00, 19.9MB/s]\r\nDownloading extra modules: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 60.9k/60.9k [00:00<00:00, 304kB/s]\r\nDownloading data: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 12.2k/12.2k [00:00<00:00, 25.6MB/s]\r\nDownloading data: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 568M/568M [00:07<00:00, 71.7MB/s]\r\nDownloading data: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 233M/233M [00:02<00:00, 78.6MB/s]\r\nDownloading data: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 285M/285M [00:04<00:00, 67.7MB/s]\r\nDownloading data: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 4.86M/4.86M [00:00<00:00, 73.3MB/s]\r\nDownloading data: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 109M/109M [00:01<00:00, 80.4MB/s]\r\nDownloading data files: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 5/5 [00:21<00:00, 4.24s/it]\r\nExtracting data files: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 5/5 [00:07<00:00, 1.54s/it]\r\nDownloading data: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 5.76M/5.76M [00:00<00:00, 56.0MB/s]\r\nDownloading data: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2.17M/2.17M [00:00<00:00, 54.1MB/s]\r\nDownloading data: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 2.18M/2.18M [00:00<00:00, 64.3MB/s]\r\nDownloading data: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 32.8k/32.8k [00:00<00:00, 53.1MB/s]\r\nDownloading data: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 800k/800k [00:00<00:00, 59.8MB/s]\r\nDownloading data files: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 5/5 [00:05<00:00, 1.01s/it]\r\nExtracting data files: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 5/5 [00:00<00:00, 2954.98it/s]\r\n```", "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {"base_commit": "82c7e879876822864b5ceaf2c99eb01159266bcd", "files": [{"path": "examples/pytorch/speech-recognition/README.md", "Loc": {"(None, None, 69)": {"mod": [69]}}, "status": "modified"}]}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [], "analysis": {"iss_type": "1", "iss_reason": "2", "loc_way": "comment", "loc_scope": "0", "info_type": "Doc"}, "max_topk": 1, "file_topk": 1, "loctype": {"code": [], "doc": ["examples/pytorch/speech-recognition/README.md"], "test": [], "config": [], "asset": []}}
|
| 1005 |
{"organization": "huggingface", "repo_name": "transformers", "base_commit": "0e82f0cbc28b41b3d87a5e4069dc0e20bacc2494", "is_iss": 0, "iss_html_url": "https://github.com/huggingface/transformers/issues/12081", "iss_label": "", "title": "GPT2 Flax \"TypeError: JAX only supports number and bool dtypes, got dtype object in array\"", "body": "On GPU\r\n\r\n```\r\n>>> from transformers import AutoTokenizer, FlaxAutoModelForCausalLM\r\n\r\n>>> tokenizer = AutoTokenizer.from_pretrained(\"gpt2-medium\")\r\n>>> model = FlaxAutoModelForCausalLM.from_pretrained(\"gpt2-medium\")\r\n>>> input_context = \"The dog\"\r\n>>> # encode input context\r\n>>> input_ids = tokenizer(input_context, return_tensors=\"jax\").input_ids\r\n>>> # generate candidates using sampling\r\n>>> outputs = model.generate(input_ids=input_ids, max_length=20, top_k=30, do_sample=True)\r\n\r\nTypeError: JAX only supports number and bool dtypes, got dtype object in array\r\n```\r\n\r\n@patrickvonplaten @patil-suraj ", "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {"base_commit": "0e82f0cbc28b41b3d87a5e4069dc0e20bacc2494", "files": [{"path": "src/transformers/models/gpt2/modeling_flax_gpt2.py", "Loc": {"('FlaxGPT2LMHeadModule', None, 553)": {"mod": []}}, "status": "modified"}, {"path": "src/transformers/models/gpt2/tokenization_gpt2_fast.py", "Loc": {"('GPT2TokenizerFast', None, 70)": {"mod": []}}, "status": "modified"}, {"Loc": [6, 7], "path": null}]}, "own_code_loc": [{"Loc": [6, 7], "path": null}], "ass_file_loc": [], "other_rep_loc": [], "analysis": {"iss_type": "1", "iss_reason": "3", "loc_way": "comment", "loc_scope": "3", "info_type": "Code"}, "max_topk": 2, "file_topk": 2, "loctype": {"code": [null, "src/transformers/models/gpt2/tokenization_gpt2_fast.py", "src/transformers/models/gpt2/modeling_flax_gpt2.py"], "doc": [], "test": [], "config": [], "asset": []}}
|
| 1006 |
{"organization": "huggingface", "repo_name": "transformers", "base_commit": "322037e842e5e89080918c824998c17722df6f19", "is_iss": 0, "iss_html_url": "https://github.com/huggingface/transformers/issues/10079", "iss_label": "", "title": "Unclear error \"NotImplementedError: \"while saving tokenizer. How fix it?", "body": "Here is my tokenizer code and how I save it to a json file\" /content/bert-datas7.json\"\r\n\r\n````\r\nfrom tokenizers import normalizers\r\nfrom tokenizers.normalizers import Lowercase, NFD, StripAccents\r\n\r\nbert_tokenizer.pre_tokenizer = Whitespace()\r\n\r\nfrom tokenizers.processors import TemplateProcessing\r\n\r\nbert_tokenizer.post_processor = TemplateProcessing(\r\n single=\"[CLS] $A [SEP]\",\r\n pair=\"[CLS] $A [SEP] $B:1 [SEP]:1\",\r\n special_tokens=[\r\n (\"[CLS]\", 1),\r\n (\"[SEP]\", 2),\r\n (\"[PAD]\", 3),\r\n ],\r\n \r\n)\r\nfrom tokenizers.trainers import WordPieceTrainer\r\n\r\ntrainer = WordPieceTrainer(\r\n vocab_size=30522, special_tokens=[\"[UNK]\", \"[CLS]\", \"[SEP]\", \"[PAD]\", \"[MASK]\"], pad_to_max_length=True\r\n)\r\nfiles = [f\"/content/For_ITMO.txt\" for split in [\"test\", \"train\", \"valid\"]]\r\nbert_tokenizer.train(trainer, files)\r\n\r\nmodel_files = bert_tokenizer.model.save(\"data\", \"/content/For_ITMO.txt\")\r\n\r\nbert_tokenizer.model = WordPiece.from_file(*model_files, unk_token=\"[UNK]\", pad_to_max_length=True)\r\n\r\nbert_tokenizer.save(\"/content/bert-datas7.json\") \r\n````\r\n\r\nWhen I output tokenizer name_or_path = nothing is displayed. This is normal?\r\n\r\n\r\n````\r\ntokenizer = PreTrainedTokenizerFast(tokenizer_file='/content/bert-datas7.json')\r\ntokenizer.add_special_tokens({'pad_token': '[PAD]'})\r\n\r\nprint(tokenizer)\r\n>>> PreTrainedTokenizerFast(name_or_path='', vocab_size=1435, model_max_len=1000000000000000019884624838656, is_fast=True, padding_side='right', special_tokens={'pad_token': '[PAD]'})\r\n````\r\nAlso, when I try to save my tokenizer, I get an error without explanation. How can I rewrite the code so that all this???\r\n#9658 \r\n#10039 \r\n[For_ITMO.txt-vocab (1) (1).txt](https://github.com/huggingface/transformers/files/5945659/For_ITMO.txt-vocab.1.1.txt)\r\n \r\n````\r\ntokenizer.save_pretrained(\"/content/tokennizerrrr\")\r\n\r\nNotImplementedError Traceback (most recent call last)\r\n<ipython-input-11-efc48254a528> in <module>()\r\n----> 1 tokenizer.save_pretrained(\"/content/tokennizerrrr\")\r\n\r\n2 frames\r\n/usr/local/lib/python3.6/dist-packages/transformers/tokenization_utils_base.py in save_vocabulary(self, save_directory, filename_prefix)\r\n 2042 :obj:`Tuple(str)`: Paths to the files saved.\r\n 2043 \"\"\"\r\n-> 2044 raise NotImplementedError\r\n 2045 \r\n 2046 def tokenize(self, text: str, pair: Optional[str] = None, add_special_tokens: bool = False, **kwargs) -> List[str]:\r\n\r\nNotImplementedError: \r\n````\r\n", "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {"base_commit": "322037e842e5e89080918c824998c17722df6f19", "files": [{"path": "src/transformers/tokenization_utils_fast.py", "Loc": {"('PreTrainedTokenizerFast', '_save_pretrained', 505)": {"mod": [509]}}, "status": "modified"}]}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [], "analysis": {"iss_type": "1", "iss_reason": "5", "loc_way": "comment", "loc_scope": "0", "info_type": "Code"}, "max_topk": 1, "file_topk": 1, "loctype": {"code": ["src/transformers/tokenization_utils_fast.py"], "doc": [], "test": [], "config": [], "asset": []}}
|
| 1007 |
+
{"organization": "huggingface", "repo_name": "transformers", "base_commit": "77a257fc210a56f1fd0d75166ecd654cf58111f3", "is_iss": 1, "iss_html_url": "https://github.com/huggingface/transformers/issues/8403", "iss_label": "", "title": "[s2s finetune] huge increase in memory demands with --fp16 native amp", "body": "While working on https://github.com/huggingface/transformers/issues/8353 I discovered that `--fp16` causes a 10x+ increase in gpu memory demands.\r\n\r\ne.g. I can run bs=12 w/o `--fp16` \r\n\r\n```\r\ncd examples/seq2seq\r\nexport BS=12; rm -rf distilbart-cnn-12-6; python finetune.py --learning_rate=3e-5 --gpus 1 \\\r\n--do_train --do_predict --val_check_interval 0.25 --n_val 500 --num_train_epochs 2 --freeze_encoder \\\r\n--freeze_embeds --data_dir cnn_dm --max_target_length 142 --val_max_target_length=142 \\\r\n--train_batch_size=$BS --eval_batch_size=$BS --gradient_accumulation_steps 1 \\\r\n--model_name_or_path sshleifer/student_cnn_12_6 --tokenizer_name facebook/bart-large \\\r\n--warmup_steps 500 --output_dir distilbart-cnn-12-6\r\n\r\n```\r\nBut if I add:\r\n```\r\n--fp16\r\n```\r\n\r\n(w/ or w/o `--fp16_opt_level O1`)\r\n\r\nI get OOM even with bs=1 on a 8GB card and it barely manages on a 24GB card - I think the increase in memory demand is more than 10x.\r\n\r\nThe OOM either right away when it does the sanity check step, or after just 10-20 batches - so within a few secs\r\n\r\nThis is with pytorch-1.6. Same goes for pytorch-1.7 and 1.8-nightly.\r\n\r\nI wasn't able to test `--fp16` with pytorch-1.5, since I can't build apex on ubuntu-20.04. Without `--fp16` pytorch-1.5 works the same as pytorch-1.6 gpu memory-wise.\r\n\r\nI tested with pytorch-1.5 + apex and there is no problem there. Memory consumption is about half.\r\n\r\nHere is the table of the batch sizes that fit into a 8gb rtx-1070 (bigger BS leads to an instant OOM):\r\n\r\nbs | version\r\n---|--------\r\n12 | pt15\r\n20 | pt15+fp16\r\n12 | pt16\r\n1 | pt16+fp16\r\n\r\n\r\n\r\nIf you'd like to reproduce the problem here are the full steps:\r\n\r\n```\r\n# prep library\r\ngit clone https://github.com/huggingface/transformers\r\ncd transformers\r\npip install -e .[dev]\r\npip install -r examples/requirements.txt\r\ncd examples/seq2seq\r\n\r\n# prep data\r\nwget https://cdn-datasets.huggingface.co/summarization/cnn_dm_v2.tgz\r\ntar -xzvf cnn_dm_v2.tgz # empty lines removed\r\nmv cnn_cln cnn_dm\r\n\r\n# run\r\nexport BS=12; \r\nrm -rf distilbart-cnn-12-6\r\npython finetune.py --learning_rate=3e-5 --gpus 1 \\\r\n--do_train --do_predict --val_check_interval 0.25 --n_val 500 --num_train_epochs 2 --freeze_encoder \\\r\n--freeze_embeds --data_dir cnn_dm --max_target_length 142 --val_max_target_length=142 \\\r\n--train_batch_size=$BS --eval_batch_size=$BS --gradient_accumulation_steps 1 \\\r\n--model_name_or_path sshleifer/student_cnn_12_6 --tokenizer_name facebook/bart-large \\\r\n--warmup_steps 500 --output_dir distilbart-cnn-12-6 \r\n```\r\n\r\nThis issue is to track the problem and hopefully finding a solution.\r\n\r\n@sshleifer ", "code": null, "pr_html_url": null, "commit_html_url": "https://github.com/pytorch/pytorch/commit/57bffc3a8e4fee0cce31e1ff1f662ccf7b16db57", "file_loc": {}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [{"org": "pytorch", "pro": "pytorch", "path": [{"base_commit": "57bffc3a8e4fee0cce31e1ff1f662ccf7b16db57", "files": [{"path": "aten/src/ATen/autocast_mode.cpp", "status": "modified", "Loc": {"(None, 'cached_cast', 67)": {"mod": [71]}}}, {"path": "test/test_cuda.py", "status": "modified", "Loc": {"('TestCuda', None, 92)": {"add": [2708]}}}]}]}], "analysis": {"iss_type": "2", "iss_reason": "1", "loc_way": "commit", "loc_scope": "2", "info_type": "Code"}, "loctype": {"code": ["aten/src/ATen/autocast_mode.cpp"], "doc": [], "test": ["test/test_cuda.py"], "config": [], "asset": ["pytorch"]}}
|
| 1008 |
{"organization": "huggingface", "repo_name": "transformers", "base_commit": "1a688709b34b10bd372e3e0860c8d39d170ebf53", "is_iss": 0, "iss_html_url": "https://github.com/huggingface/transformers/issues/17201", "iss_label": "", "title": "a memory leak in qqp prediction using bart", "body": "### System Info\n\n```shell\n- `transformers` version: 4.19.0.dev0\r\n- Platform: Linux-5.11.0-43-generic-x86_64-with-glibc2.17\r\n- Python version: 3.8.10\r\n- Huggingface_hub version: 0.4.0\r\n- PyTorch version (GPU?): 1.10.1 (True)\r\n- Tensorflow version (GPU?): not installed (NA)\r\n- Flax version (CPU?/GPU?/TPU?): not installed (NA)\r\n- Jax version: not installed\r\n- JaxLib version: not installed\r\n- Using GPU in script?: Yes\r\n- Using distributed or parallel set-up in script?: No\n```\n\n\n### Who can help?\n\n@sgugger\n\n### Information\n\n- [X] The official example scripts\n- [ ] My own modified scripts\n\n### Tasks\n\n- [X] An officially supported task in the `examples` folder (such as GLUE/SQuAD, ...)\n- [ ] My own task or dataset (give details below)\n\n### Reproduction\n\nI met the same issue #11011. If not using `--eval_accumulation_steps`, it caused CUDA out of memory. If using it, it caused out of RAM and killed by system.\r\n\r\nI only did prediction on GLUE QQP dataset using bart without fine-tuning. Considering QQP having a large test set (300k), the prediction got slower and slower, and finally got out of memory.\r\n\r\nThis is the script to reproduce:\r\n```\r\nCUDA_VISIBLE_DEVICES=0 python run_glue.py --model_name_or_path facebook/bart-large --task_name qqp --output_dir bart-large_qqp --eval_accumulation_steps 100 --do_predict --per_device_eval_batch_size 24\r\n```\n\n### Expected behavior\n\n```shell\nPrediction without out memory.\n```\n", "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {"base_commit": "1a688709b34b10bd372e3e0860c8d39d170ebf53", "files": [{"path": "src/transformers/trainer.py", "Loc": {"('Trainer', 'evaluation_loop', 2549)": {"mod": [2635]}}, "status": "modified"}]}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [], "analysis": {"iss_type": "1", "iss_reason": "2\nOr\n5", "loc_way": "comment", "loc_scope": "0", "info_type": "Code"}, "max_topk": 1, "file_topk": 1, "loctype": {"code": ["src/transformers/trainer.py"], "doc": [], "test": [], "config": [], "asset": []}}
|
| 1009 |
{"organization": "huggingface", "repo_name": "transformers", "base_commit": "cef2e40e0f8eaad13b8d32817a48fdddc32eb2a5", "is_iss": 0, "iss_html_url": "https://github.com/huggingface/transformers/issues/28435", "iss_label": "", "title": "Skip some weights for load_in_8bit and keep them as fp16/32?", "body": "### Feature request\r\n\r\nHello,\r\n\r\nI am looking for a way to load a checkpoint where I only load some of the weights in 8 bit and keep others in 16/32 bit.\r\n\r\n### Motivation\r\n\r\nMy motivation is for vision-language models like Llava or BLIP2 where I want to load the LLM part in 8 bit but the image encoder should stay in 16 bit because I notice performance degradations with CLIP in 8 bit and also want to be able to train this part without LoRA.\r\n\r\nAs far as I can see in the documentation, issues and with Google (both here and for bitsandbytes), there is currently no way to do this.\r\n\r\n### Your contribution\r\n\r\nI can in theory help implement something like this but I don't know where and how in the code this should be done.", "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {"base_commit": "cef2e40e0f8eaad13b8d32817a48fdddc32eb2a5", "files": [{"path": "src/transformers/modeling_utils.py", "Loc": {"('PreTrainedModel', 'from_pretrained', 2528)": {"mod": [3524]}}, "status": "modified"}, {"path": "src/transformers/utils/quantization_config.py", "Loc": {"('BitsAndBytesConfig', None, 151)": {"mod": [176]}}, "status": "modified"}]}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [], "analysis": {"iss_type": "3", "iss_reason": "5", "loc_way": "comment", "loc_scope": "0", "info_type": "Code"}, "max_topk": 2, "file_topk": 2, "loctype": {"code": ["src/transformers/modeling_utils.py", "src/transformers/utils/quantization_config.py"], "doc": [], "test": [], "config": [], "asset": []}}
|
| 1010 |
{"organization": "huggingface", "repo_name": "transformers", "base_commit": "705ca7f21b2b557e0cfd5d0853b297fa53489d20", "is_iss": 1, "iss_html_url": "https://github.com/huggingface/transformers/issues/14938", "iss_label": "", "title": "Question: Object of type EncoderDecoderConfig is not JSON serializable", "body": "Hi.\r\nAn error occurred when I used Trainer to train and save EncoderDecoderModel.\r\n\r\n```python\r\n File \"/home/jwli/ljw/study/hotpotqa/roberta_seq2seq/roberta_for_seq2seq.py\", line 482, in <module>\r\n run(model_args, data_args, training_args)\r\n File \"/home/jwli/ljw/study/hotpotqa/roberta_seq2seq/roberta_for_seq2seq.py\", line 465, in run\r\n train_result = trainer.train(resume_from_checkpoint=checkpoint)\r\n File \"/home/jwli/anaconda3/envs/study/lib/python3.7/site-packages/transformers/trainer.py\", line 1391, in train\r\n self._maybe_log_save_evaluate(tr_loss, model, trial, epoch, ignore_keys_for_eval)\r\n File \"/home/jwli/anaconda3/envs/study/lib/python3.7/site-packages/transformers/trainer.py\", line 1495, in _maybe_log_save_evaluate\r\n self._save_checkpoint(model, trial, metrics=metrics)\r\n File \"/home/jwli/anaconda3/envs/study/lib/python3.7/site-packages/transformers/trainer.py\", line 1557, in _save_checkpoint\r\n self.save_model(output_dir)\r\n File \"/home/jwli/anaconda3/envs/study/lib/python3.7/site-packages/transformers/trainer.py\", line 1961, in save_model\r\n self._save(output_dir)\r\n File \"/home/jwli/anaconda3/envs/study/lib/python3.7/site-packages/transformers/trainer.py\", line 2009, in _save\r\n self.model.save_pretrained(output_dir, state_dict=state_dict)\r\n File \"/home/jwli/anaconda3/envs/study/lib/python3.7/site-packages/transformers/modeling_utils.py\", line 1053, in save_pretrained\r\n model_to_save.config.save_pretrained(save_directory)\r\n File \"/home/jwli/anaconda3/envs/study/lib/python3.7/site-packages/transformers/configuration_utils.py\", line 416, in save_pretrained\r\n self.to_json_file(output_config_file, use_diff=True)\r\n File \"/home/jwli/anaconda3/envs/study/lib/python3.7/site-packages/transformers/configuration_utils.py\", line 739, in to_json_file\r\n writer.write(self.to_json_string(use_diff=use_diff))\r\n File \"/home/jwli/anaconda3/envs/study/lib/python3.7/site-packages/transformers/configuration_utils.py\", line 725, in to_json_string\r\n return json.dumps(config_dict, indent=2, sort_keys=True) + \"\\n\"\r\n File \"/home/jwli/anaconda3/envs/study/lib/python3.7/json/__init__.py\", line 238, in dumps\r\n **kw).encode(obj)\r\n File \"/home/jwli/anaconda3/envs/study/lib/python3.7/json/encoder.py\", line 201, in encode\r\n chunks = list(chunks)\r\n File \"/home/jwli/anaconda3/envs/study/lib/python3.7/json/encoder.py\", line 431, in _iterencode\r\n yield from _iterencode_dict(o, _current_indent_level)\r\n File \"/home/jwli/anaconda3/envs/study/lib/python3.7/json/encoder.py\", line 405, in _iterencode_dict\r\n yield from chunks\r\n File \"/home/jwli/anaconda3/envs/study/lib/python3.7/json/encoder.py\", line 438, in _iterencode\r\n o = _default(o)\r\n File \"/home/jwli/anaconda3/envs/study/lib/python3.7/json/encoder.py\", line 179, in default\r\n raise TypeError(f'Object of type {o.__class__.__name__} '\r\nTypeError: Object of type EncoderDecoderConfig is not JSON serializable\r\n```\r\nMy model and Config define the following code. \r\n```python\r\n tokenizer = RobertaTokenizerFast.from_pretrained(model_args.tokenizer_name)\r\n encoder_config = RobertaConfig.from_pretrained(model_args.encoder_model_name_or_path)\r\n decoder_config = RobertaConfig.from_pretrained(model_args.decoder_model_name_or_path)\r\n encoder_decoder_config = EncoderDecoderConfig.from_encoder_decoder_configs(encoder_config, decoder_config)\r\n model = RobertaForSeq2Seq.from_encoder_decoder_pretrained(model_args.encoder_model_name_or_path,\r\n model_args.decoder_model_name_or_path,\r\n config=encoder_decoder_config, tie_encoder_decoder=True)\r\n model.config.decoder_start_token_id = tokenizer.bos_token_id\r\n model.config.eos_token_id = tokenizer.eos_token_id\r\n model.config.max_length = 64\r\n model.config.early_stopping = True\r\n model.config.no_repeat_ngram_size = 3\r\n model.config.length_penalty = 2.0\r\n model.config.num_beams = 4\r\n model.config.pad_token_id = tokenizer.pad_token_id\r\n```\r\nThis error occurred because EncoderDecoderConfig cannot be converted to json format. But I don't know how to modify it.\r\n```python\r\nERROR OCCURRED:\r\n\r\n if use_diff is True:\r\n config_dict = self.to_diff_dict()\r\n else:\r\n config_dict = self.to_dict()\r\n return json.dumps(config_dict, indent=2, sort_keys=True) + \"\\n\"\r\n```\r\n\r\nI look forward to your help! Thanks!\r\n @jplu @patrickvonplaten ", "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {}, "own_code_loc": [{"Loc": [46, 47], "path": null}], "ass_file_loc": [], "other_rep_loc": [], "analysis": {"iss_type": "1", "iss_reason": "3", "loc_way": "comment", "loc_scope": "3", "info_type": "Code"}, "loctype": {"code": [null], "doc": [], "test": [], "config": [], "asset": []}}
|
|
|
|
| 1085 |
{"organization": "abi", "repo_name": "screenshot-to-code", "base_commit": "fcd305d0d26e7ef7b93dd605cbd5ed0e1a5a5e9c", "is_iss": 0, "iss_html_url": "https://github.com/abi/screenshot-to-code/issues/150", "iss_label": "", "title": "Error generating code. Check the Developer Console AND the backend logs for details", "body": "My ChatGPT has access to GPT-VISION. and the web app loads well but when I upload an image. it returns this error 'Error generating code. Check the Developer Console AND the backend logs for details'\r\n<img width=\"466\" alt=\"error\" src=\"https://github.com/abi/screenshot-to-code/assets/100529823/97c337b7-de54-45f9-8def-f984ade50a6d\">\r\n", "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {"base_commit": "fcd305d0d26e7ef7b93dd605cbd5ed0e1a5a5e9c", "files": [{"path": "docker-compose.yml", "Loc": {"(None, None, 20)": {"mod": [20]}}, "status": "modified"}]}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [], "analysis": {"iss_type": "1", "iss_reason": "3", "loc_way": "comment", "loc_scope": "0", "info_type": "Code"}, "max_topk": 1, "file_topk": 1, "loctype": {"code": [], "doc": ["docker-compose.yml"], "test": [], "config": [], "asset": []}}
|
| 1086 |
{"organization": "pytorch", "repo_name": "pytorch", "base_commit": "4622b3395276b37e10141fab43ffea33941ca0c2", "is_iss": 0, "iss_html_url": "https://github.com/pytorch/pytorch/issues/2384", "iss_label": "", "title": "How the grad is transferred between layer", "body": "consider a simple example here:\r\n```python\r\nimport torch\r\nfrom torch.autograd import Variable\r\n\r\ninput = Variable(torch.randn(20, 3, 28, 28), requires_grad=True)\r\nm = torch.nn.Conv2d(3, 16, 5)\r\noutput = m(input)\r\n\r\nloss = torch.sum(output)# define loss to perform backprop\r\nm.zero_grad()\r\nloss.backward()\r\n\r\nprint(type(input))\r\nprint(input.grad.size())\r\nprint(type(output))\r\nprint(output.grad)\r\n```\r\nthe output is:\r\n```\r\n<class 'torch.autograd.variable.Variable'>\r\ntorch.Size([20, 3, 28, 28])\r\n<class 'torch.autograd.variable.Variable'>\r\nNone\r\n```\r\nI find the `output.grad` is `None`. I don't know how the `input.grad` is calculated without `output.grad`.\r\nand want to know how to get the values of `output.grad`.\r\n\r\nthanks!", "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {"base_commit": "4622b3395276b37e10141fab43ffea33941ca0c2", "files": [{"path": "torch/autograd/variable.py", "Loc": {"('Variable', 'retain_grad', 236)": {"mod": []}}, "status": "modified"}]}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [], "analysis": {"iss_type": "3", "iss_reason": "3", "loc_way": "comment", "loc_scope": "0", "info_type": "Code"}, "max_topk": 1, "file_topk": 1, "loctype": {"code": ["torch/autograd/variable.py"], "doc": [], "test": [], "config": [], "asset": []}}
|
| 1087 |
{"organization": "pytorch", "repo_name": "pytorch", "base_commit": "2abcafcfd8beb4f6a22e08532d58f9f09c490f0f", "is_iss": 0, "iss_html_url": "https://github.com/pytorch/pytorch/issues/96983", "iss_label": "module: binaries\ntriaged\nmodule: arm", "title": "PyTorch 2.0 aarch64 wheels are missing the mkldnn+acl backend support", "body": "### \ud83d\udc1b Describe the bug\r\n\r\nPyTorch 2.0 aarch64 wheels are missing the mkldnn+acl backend support, where as PyTorch 1.13.0 had support.\r\n\r\nSolution:\r\nthe wheels need to be built with the `--enable-mkldnn` option while building them from the pytorch/builder repo.\r\n\r\nexample command for pytorch wheel builder script:\r\n`./build_aarch64_wheel.py --python-version 3.8 --use-docker --keep-running --os ubuntu20_04 --enable-mkldnn --branch release/2.0`\r\n\r\nTo reproduce the issue, create c6g or c7g instance from AWS EC2, and in the below output, look for `USE_MKLDNN=`, this was ON for PyTorch 1.13.0 but OFF for PyTorch2.0.0.\r\n\r\nnon-working scenario\r\n```\r\npip install torch==2.0.0\r\n\r\ntime python3 -c \"import torch; torch.set_num_threads(8); print(torch.__version__, torch.__config__.show(), torch.get_num_threads());a=torch.rand(100, 100, 100); b=torch.rand(100,100, 100); [torch.bmm(a,b).sum() for i in range(1000)]\"\r\n2.0.0 PyTorch built with:\r\n - GCC 10.2\r\n - C++ Version: 201703\r\n - OpenMP 201511 (a.k.a. OpenMP 4.5)\r\n - LAPACK is enabled (usually provided by MKL)\r\n - NNPACK is enabled\r\n - CPU capability usage: NO AVX\r\n - Build settings: BLAS_INFO=open, BUILD_TYPE=Release, CXX_COMPILER=/opt/rh/devtoolset-10/root/usr/bin/c++, CXX_FLAGS= -D_GLIBCXX_USE_CXX11_ABI=0 -fabi-version=11 -Wno-deprecated -fvisibility-inlines-hidden -DUSE_PTHREADPOOL -DNDEBUG -DUSE_KINETO -DLIBKINETO_NOCUPTI -DLIBKINETO_NOROCTRACER -DUSE_QNNPACK -DUSE_PYTORCH_QNNPACK -DUSE_XNNPACK -DSYMBOLICATE_MOBILE_DEBUG_HANDLE -O2 -fPIC -Wall -Wextra -Werror=return-type -Werror=non-virtual-dtor -Werror=bool-operation -Wnarrowing -Wno-missing-field-initializers -Wno-type-limits -Wno-array-bounds -Wno-unknown-pragmas -Wunused-local-typedefs -Wno-unused-parameter -Wno-unused-function -Wno-unused-result -Wno-strict-overflow -Wno-strict-aliasing -Wno-error=deprecated-declarations -Wno-stringop-overflow -Wno-psabi -Wno-error=pedantic -Wno-error=redundant-decls -Wno-error=old-style-cast -fdiagnostics-color=always -faligned-new -Wno-unused-but-set-variable -Wno-maybe-uninitialized -fno-math-errno -fno-trapping-math -Werror=format -Werror=cast-function-type -Wno-stringop-overflow, LAPACK_INFO=open, TORCH_DISABLE_GPU_ASSERTS=ON, TORCH_VERSION=2.0.0, USE_CUDA=OFF, USE_CUDNN=OFF, USE_EIGEN_FOR_BLAS=ON, USE_EXCEPTION_PTR=1, USE_GFLAGS=OFF, USE_GLOG=OFF, USE_MKL=OFF, USE_MKLDNN=OFF, USE_MPI=OFF, USE_NCCL=OFF, USE_NNPACK=ON, USE_OPENMP=ON, USE_ROCM=OFF, \r\n\r\n```\r\n\r\nworking scenario:\r\n\r\n```\r\npip3 install torch==1.13.0\r\n\r\ntime python3 -c \"import torch; torch.set_num_threads(8); print(torch.__version__, torch.__config__.show(), torch.get_num_threads());a=torch.rand(100, 100, 100); b=torch.rand(100,100, 100); [torch.bmm(a,b).sum() for i in range(1000)]\"\r\n\r\n1.13.0 PyTorch built with:\r\n - GCC 10.2\r\n - C++ Version: 201402\r\n - Intel(R) MKL-DNN v2.6.0 (Git Hash 52b5f107dd9cf10910aaa19cb47f3abf9b349815)\r\n - OpenMP 201511 (a.k.a. OpenMP 4.5)\r\n - LAPACK is enabled (usually provided by MKL)\r\n - NNPACK is enabled\r\n - CPU capability usage: NO AVX\r\n - Build settings: BLAS_INFO=open, BUILD_TYPE=Release, CXX_COMPILER=/opt/rh/devtoolset-10/root/usr/bin/c++, CXX_FLAGS= -Wno-deprecated -fvisibility-inlines-hidden -DUSE_PTHREADPOOL -fopenmp -DNDEBUG -DUSE_KINETO -DLIBKINETO_NOCUPTI -DUSE_QNNPACK -DUSE_PYTORCH_QNNPACK -DUSE_XNNPACK -DSYMBOLICATE_MOBILE_DEBUG_HANDLE -DEDGE_PROFILER_USE_KINETO -O2 -fPIC -Wno-narrowing -Wall -Wextra -Werror=return-type -Werror=non-virtual-dtor -Wno-missing-field-initializers -Wno-type-limits -Wno-array-bounds -Wno-unknown-pragmas -Wunused-local-typedefs -Wno-unused-parameter -Wno-unused-function -Wno-unused-result -Wno-strict-overflow -Wno-strict-aliasing -Wno-error=deprecated-declarations -Wno-stringop-overflow -Wno-psabi -Wno-error=pedantic -Wno-error=redundant-decls -Wno-error=old-style-cast -fdiagnostics-color=always -faligned-new -Wno-unused-but-set-variable -Wno-maybe-uninitialized -fno-math-errno -fno-trapping-math -Werror=format -Werror=cast-function-type -Wno-stringop-overflow, LAPACK_INFO=open, TORCH_VERSION=1.13.0, USE_CUDA=OFF, USE_CUDNN=OFF, USE_EIGEN_FOR_BLAS=ON, USE_EXCEPTION_PTR=1, USE_GFLAGS=OFF, USE_GLOG=OFF, USE_MKL=OFF, USE_MKLDNN=ON, USE_MPI=OFF, USE_NCCL=OFF, USE_NNPACK=ON, USE_OPENMP=ON, USE_ROCM=OFF, \r\n \r\n\r\n\r\n```\r\n\r\n### Versions\r\n```\r\nCollecting environment information...\r\nPyTorch version: 2.0.0\r\nIs debug build: False\r\nCUDA used to build PyTorch: None\r\nROCM used to build PyTorch: N/A\r\n\r\nOS: Ubuntu 20.04.5 LTS (aarch64)\r\nGCC version: (Ubuntu 10.3.0-1ubuntu1~20.04) 10.3.0\r\nClang version: Could not collect\r\nCMake version: version 3.25.2\r\nLibc version: glibc-2.31\r\n\r\nPython version: 3.8.10 (default, Nov 14 2022, 12:59:47) [GCC 9.4.0] (64-bit runtime)\r\nPython platform: Linux-5.15.0-1028-aws-aarch64-with-glibc2.29\r\nIs CUDA available: False\r\nCUDA runtime version: No CUDA\r\nCUDA_MODULE_LOADING set to: N/A\r\nGPU models and configuration: No CUDA\r\nNvidia driver version: No CUDA\r\ncuDNN version: No CUDA\r\nHIP runtime version: N/A\r\nMIOpen runtime version: N/A\r\nIs XNNPACK available: True\r\n\r\nCPU:\r\nArchitecture: aarch64\r\nCPU op-mode(s): 32-bit, 64-bit\r\nByte Order: Little Endian\r\nCPU(s): 16\r\nOn-line CPU(s) list: 0-15\r\nThread(s) per core: 1\r\nCore(s) per socket: 16\r\nSocket(s): 1\r\nNUMA node(s): 1\r\nVendor ID: ARM\r\nModel: 1\r\nStepping: r1p1\r\nBogoMIPS: 2100.00\r\nL1d cache: 1 MiB\r\nL1i cache: 1 MiB\r\nL2 cache: 16 MiB\r\nL3 cache: 32 MiB\r\nNUMA node0 CPU(s): 0-15\r\nVulnerability Itlb multihit: Not affected\r\nVulnerability L1tf: Not affected\r\nVulnerability Mds: Not affected\r\nVulnerability Meltdown: Not affected\r\nVulnerability Mmio stale data: Not affected\r\nVulnerability Retbleed: Not affected\r\nVulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl\r\nVulnerability Spectre v1: Mitigation; __user pointer sanitization\r\nVulnerability Spectre v2: Mitigation; CSV2, BHB\r\nVulnerability Srbds: Not affected\r\nVulnerability Tsx async abort: Not affected\r\nFlags: fp asimd evtstrm aes pmull sha1 sha2 crc32 atomics fphp asimdhp cpuid asimdrdm jscvt fcma lrcpc dcpop sha3 sm3 sm4 asimddp sha512 sve asimdfhm dit uscat ilrcpc flagm ssbs paca pacg dcpodp svei8mm svebf16 i8mm bf16 dgh rng\r\n\r\nVersions of relevant libraries:\r\n[pip3] numpy==1.24.2\r\n[pip3] torch==2.0.0\r\n[pip3] torchvision==0.14.1\r\n[conda] Could not collect\r\n```\r\n\r\ncc @ezyang @seemethere @malfet", "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {"base_commit": "2abcafcfd8beb4f6a22e08532d58f9f09c490f0f", "files": [{"path": ".ci/aarch64_linux/build_aarch64_wheel.py", "Loc": {"(None, None, None)": {"mod": [8]}}, "status": "modified"}]}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [], "analysis": {"iss_type": "1", "iss_reason": "2", "loc_way": "comment", "loc_scope": "0", "info_type": "Code"}, "max_topk": 1, "file_topk": 1, "loctype": {"code": [".ci/aarch64_linux/build_aarch64_wheel.py"], "doc": [], "test": [], "config": [], "asset": []}}
|
| 1088 |
+
{"organization": "pytorch", "repo_name": "pytorch", "base_commit": "2dff0b3e918530719f7667cb31541f036a25e3f2", "is_iss": 1, "iss_html_url": "https://github.com/pytorch/pytorch/issues/48435", "iss_label": "", "title": "AttributeError: module 'torch.cuda' has no attribute 'comm'", "body": "## \u2753 Questions and Help\r\n\r\nI'm using torch 1.7.0, and get this kind of error\r\n\r\nmy torch is installed via \r\n\r\npip install torch==1.7.0+cu101 torchvision==0.8.1+cu101 torchaudio===0.7.0 -f https://download.pytorch.org/whl/torch_stable.html\r\n\r\nmy os is win10", "code": null, "pr_html_url": null, "commit_html_url": "https://github.com/facebookresearch/InterHand2.6M/commit/874eb9f740ef54c275433d1bd27f8fb8f6a8f17d", "file_loc": {}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [{"org": "facebookresearch", "pro": "InterHand2.6M", "path": [{"base_commit": "874eb9f740ef54c275433d1bd27f8fb8f6a8f17d", "files": [{"path": "common/nets/module.py", "status": "modified", "Loc": {"('PoseNet', 'soft_argmax_1d', 41)": {"mod": [43]}}}]}]}], "analysis": {"iss_type": "1", "iss_reason": "1", "loc_way": "commit", "loc_scope": "2", "info_type": "Code"}, "loctype": {"code": ["common/nets/module.py"], "doc": [], "test": [], "config": [], "asset": ["InterHand2.6M"]}}
|
| 1089 |
{"organization": "xtekky", "repo_name": "gpt4free", "base_commit": "e8f6013d0349229fd8f7d298952cfe56fc4b8761", "is_iss": 0, "iss_html_url": "https://github.com/xtekky/gpt4free/issues/2070", "iss_label": "bug\nstale", "title": "Liaobots and You don't work", "body": "Liaobots and You do not work, they give the following errors:\r\n\r\n```\r\nLiaobots: ResponseStatusError: Response 500: Error\r\n``` \r\n\r\n```\r\nYou: ResponseStatusError: Response 401: {\"status_code\":401,\"request_id\":\"request-id-live-183191e7-adc1-4838-8e29-6e0c5c3ca048\",\"error_type\":\"endpoint_not_authorized_for_sdk\",\"error_message\":\"The project owner has not authorized the SDK to call this endpoint. Please enable it in the dashboard to continue: https://stytch.com/dashboard/sdk-configuration.\",\"error_url\":\"https://stytch.com/docs/api/errors/401#endpoint_not_authorized_for_sdk\"}\r\n``` \r\n@xtekky @hlohaus ", "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {"base_commit": "e8f6013d0349229fd8f7d298952cfe56fc4b8761", "files": [{"path": "g4f/Provider/Liaobots.py", "Loc": {"('Liaobots', 'create_async_generator', 111)": {"mod": [149]}}, "status": "modified"}]}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [], "analysis": {"iss_type": "1", "iss_reason": "5", "loc_way": "comment", "loc_scope": "0", "info_type": "Code"}, "max_topk": 1, "file_topk": 1, "loctype": {"code": ["g4f/Provider/Liaobots.py"], "doc": [], "test": [], "config": [], "asset": []}}
|
| 1090 |
{"organization": "xtekky", "repo_name": "gpt4free", "base_commit": "fa2d608822540c9b73350bfa036e8822ade4e23f", "is_iss": 0, "iss_html_url": "https://github.com/xtekky/gpt4free/issues/2305", "iss_label": "stale", "title": "ValueError: Unknown model: dall-e-3", "body": "```\r\nC:\\Users\\MAX\\Desktop>pip install -U g4f[all]\r\nDefaulting to user installation because normal site-packages is not writeable\r\nRequirement already satisfied: g4f[all] in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (0.3.3.2)\r\nRequirement already satisfied: requests in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (2.32.3)\r\nRequirement already satisfied: aiohttp in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (3.9.3)\r\nRequirement already satisfied: brotli in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (1.1.0)\r\nRequirement already satisfied: pycryptodome in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (3.20.0)\r\nRequirement already satisfied: curl-cffi>=0.6.2 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (0.7.3)\r\nRequirement already satisfied: cloudscraper in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (1.2.71)\r\nRequirement already satisfied: certifi in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (2024.8.30)\r\nRequirement already satisfied: browser-cookie3 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (0.19.1)\r\nRequirement already satisfied: PyExecJS in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (1.5.1)\r\nRequirement already satisfied: duckduckgo-search>=5.0 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (6.3.2)\r\nRequirement already satisfied: beautifulsoup4 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (4.12.3)\r\nRequirement already satisfied: pywebview in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (5.2)\r\nRequirement already satisfied: platformdirs in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (4.2.2)\r\nRequirement already satisfied: plyer in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (2.1.0)\r\nRequirement already satisfied: cryptography in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (43.0.0)\r\nRequirement already satisfied: aiohttp-socks in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (0.8.4)\r\nRequirement already satisfied: pillow in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (10.2.0)\r\nRequirement already satisfied: cairosvg in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (2.7.1)\r\nRequirement already satisfied: werkzeug in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (3.0.1)\r\nRequirement already satisfied: flask in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (3.0.2)\r\nRequirement already satisfied: loguru in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (0.7.2)\r\nRequirement already satisfied: fastapi in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (0.109.2)\r\nRequirement already satisfied: uvicorn in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (0.27.0.post1)\r\nRequirement already satisfied: nest-asyncio in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from g4f[all]) (1.6.0)\r\nRequirement already satisfied: cffi>=1.12.0 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from curl-cffi>=0.6.2->g4f[all]) (1.17.0)\r\nRequirement already satisfied: typing-extensions in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from curl-cffi>=0.6.2->g4f[all]) (4.12.2)\r\nRequirement already satisfied: click>=8.1.7 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from duckduckgo-search>=5.0->g4f[all]) (8.1.7)\r\nRequirement already satisfied: primp>=0.6.4 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from duckduckgo-search>=5.0->g4f[all]) (0.6.4)\r\nRequirement already satisfied: aiosignal>=1.1.2 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from aiohttp->g4f[all]) (1.3.1)\r\nRequirement already satisfied: attrs>=17.3.0 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from aiohttp->g4f[all]) (23.2.0)\r\nRequirement already satisfied: frozenlist>=1.1.1 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from aiohttp->g4f[all]) (1.4.1)\r\nRequirement already satisfied: multidict<7.0,>=4.5 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from aiohttp->g4f[all]) (6.0.5)\r\nRequirement already satisfied: yarl<2.0,>=1.0 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from aiohttp->g4f[all]) (1.9.4)\r\nRequirement already satisfied: python-socks<3.0.0,>=2.4.3 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from python-socks[asyncio]<3.0.0,>=2.4.3->aiohttp-socks->g4f[all]) (2.4.4)\r\nRequirement already satisfied: soupsieve>1.2 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from beautifulsoup4->g4f[all]) (2.5)\r\nRequirement already satisfied: lz4 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from browser-cookie3->g4f[all]) (4.3.3)\r\nRequirement already satisfied: pycryptodomex in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from browser-cookie3->g4f[all]) (3.20.0)\r\nRequirement already satisfied: cairocffi in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from cairosvg->g4f[all]) (1.6.1)\r\nRequirement already satisfied: cssselect2 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from cairosvg->g4f[all]) (0.7.0)\r\nRequirement already satisfied: defusedxml in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from cairosvg->g4f[all]) (0.7.1)\r\nRequirement already satisfied: tinycss2 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from cairosvg->g4f[all]) (1.2.1)\r\nRequirement already satisfied: pyparsing>=2.4.7 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from cloudscraper->g4f[all]) (3.1.2)\r\nRequirement already satisfied: requests-toolbelt>=0.9.1 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from cloudscraper->g4f[all]) (1.0.0)\r\nRequirement already satisfied: charset-normalizer<4,>=2 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from requests->g4f[all]) (3.3.2)\r\nRequirement already satisfied: idna<4,>=2.5 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from requests->g4f[all]) (3.6)\r\nRequirement already satisfied: urllib3<3,>=1.21.1 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from requests->g4f[all]) (2.1.0)\r\nRequirement already satisfied: pydantic!=1.8,!=1.8.1,!=2.0.0,!=2.0.1,!=2.1.0,<3.0.0,>=1.7.4 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from fastapi->g4f[all]) (2.6.1)\r\nRequirement already satisfied: starlette<0.37.0,>=0.36.3 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from fastapi->g4f[all]) (0.36.3)\r\nRequirement already satisfied: Jinja2>=3.1.2 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from flask->g4f[all]) (3.1.3)\r\nRequirement already satisfied: itsdangerous>=2.1.2 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from flask->g4f[all]) (2.1.2)\r\nRequirement already satisfied: blinker>=1.6.2 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from flask->g4f[all]) (1.7.0)\r\nRequirement already satisfied: MarkupSafe>=2.1.1 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from werkzeug->g4f[all]) (2.1.5)\r\nRequirement already satisfied: colorama>=0.3.4 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from loguru->g4f[all]) (0.4.6)\r\nRequirement already satisfied: win32-setctime>=1.0.0 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from loguru->g4f[all]) (1.1.0)\r\nRequirement already satisfied: six>=1.10.0 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from PyExecJS->g4f[all]) (1.16.0)\r\nRequirement already satisfied: proxy-tools in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from pywebview->g4f[all]) (0.1.0)\r\nRequirement already satisfied: bottle in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from pywebview->g4f[all]) (0.13.1)\r\nRequirement already satisfied: pythonnet in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from pywebview->g4f[all]) (3.0.3)\r\nRequirement already satisfied: h11>=0.8 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from uvicorn->g4f[all]) (0.14.0)\r\nRequirement already satisfied: pycparser in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from cffi>=1.12.0->curl-cffi>=0.6.2->g4f[all]) (2.22)\r\nRequirement already satisfied: annotated-types>=0.4.0 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from pydantic!=1.8,!=1.8.1,!=2.0.0,!=2.0.1,!=2.1.0,<3.0.0,>=1.7.4->fastapi->g4f[all]) (0.6.0)\r\nRequirement already satisfied: pydantic-core==2.16.2 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from pydantic!=1.8,!=1.8.1,!=2.0.0,!=2.0.1,!=2.1.0,<3.0.0,>=1.7.4->fastapi->g4f[all]) (2.16.2)\r\nRequirement already satisfied: async-timeout>=3.0.1 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from python-socks[asyncio]<3.0.0,>=2.4.3->aiohttp-socks->g4f[all]) (4.0.3)\r\nRequirement already satisfied: anyio<5,>=3.4.0 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from starlette<0.37.0,>=0.36.3->fastapi->g4f[all]) (4.2.0)\r\nRequirement already satisfied: webencodings in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from cssselect2->cairosvg->g4f[all]) (0.5.1)\r\nRequirement already satisfied: clr-loader<0.3.0,>=0.2.6 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from pythonnet->pywebview->g4f[all]) (0.2.6)\r\nRequirement already satisfied: sniffio>=1.1 in c:\\users\\max\\appdata\\local\\packages\\pythonsoftwarefoundation.python.3.12_qbz5n2kfra8p0\\localcache\\local-packages\\python312\\site-packages (from anyio<5,>=3.4.0->starlette<0.37.0,>=0.36.3->fastapi->g4f[all]) (1.3.0)\r\n\r\nC:\\Users\\MAX\\Desktop>\r\nTraceback (most recent call last):.py\r\n File \"C:\\Users\\MAX\\Desktop\\gptimg.py\", line 4, in <module>\r\n response = client.images.generate(\r\n ^^^^^^^^^^^^^^^^^^^^^^^\r\n File \"C:\\Users\\MAX\\AppData\\Local\\Packages\\PythonSoftwareFoundation.Python.3.12_qbz5n2kfra8p0\\LocalCache\\local-packages\\Python312\\site-packages\\g4f\\client\\client.py\", line 421, in generate\r\n return asyncio.run(self.async_generate(prompt, model, response_format=response_format, **kwargs))\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File \"C:\\Program Files\\WindowsApps\\PythonSoftwareFoundation.Python.3.12_3.12.2032.0_x64__qbz5n2kfra8p0\\Lib\\asyncio\\runners.py\", line 194, in run\r\n return runner.run(main)\r\n ^^^^^^^^^^^^^^^^\r\n File \"C:\\Program Files\\WindowsApps\\PythonSoftwareFoundation.Python.3.12_3.12.2032.0_x64__qbz5n2kfra8p0\\Lib\\asyncio\\runners.py\", line 118, in run\r\n return self._loop.run_until_complete(task)\r\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\r\n File \"C:\\Program Files\\WindowsApps\\PythonSoftwareFoundation.Python.3.12_3.12.2032.0_x64__qbz5n2kfra8p0\\Lib\\asyncio\\base_events.py\", line 687, in run_until_complete\r\n return future.result()\r\n ^^^^^^^^^^^^^^^\r\n File \"C:\\Users\\MAX\\AppData\\Local\\Packages\\PythonSoftwareFoundation.Python.3.12_qbz5n2kfra8p0\\LocalCache\\local-packages\\Python312\\site-packages\\g4f\\client\\client.py\", line 426, in async_generate\r\n raise ValueError(f\"Unknown model: {model}\")\r\nValueError: Unknown model: dall-e-3\r\n```", "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {"base_commit": "fa2d608822540c9b73350bfa036e8822ade4e23f", "files": [{"path": "g4f/models.py", "Loc": {}}]}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [], "analysis": {"iss_type": "1", "iss_reason": "5", "loc_way": "comment", "loc_scope": "0", "info_type": "Code"}, "max_topk": 0, "file_topk": 0, "loctype": {"code": ["g4f/models.py"], "doc": [], "test": [], "config": [], "asset": []}}
|
| 1091 |
{"organization": "xtekky", "repo_name": "gpt4free", "base_commit": "1ade1d959cbc9aea7cf653bbe5b6c414ba486c97", "is_iss": 0, "iss_html_url": "https://github.com/xtekky/gpt4free/issues/1292", "iss_label": "bug\nstale", "title": "RecursionError: maximum recursion depth exceeded while calling a Python object", "body": "Ubuntu 22, g4f-0.1.9.0, pip installation method, python3.10\r\n\r\n**Bug description**\r\nG4F API has these errors after 5-10 requests. I have to restart constantly. It is very uncomfortable. This problem did not exist in the previous version.\r\n\r\n**Errors**\r\n```\r\nRecursionError: maximum recursion depth exceeded in comparison\r\nRecursionError: maximum recursion depth exceeded while calling a Python object\r\nRuntimeError: RetryProvider failed:\r\nYou: RecursionError: maximum recursion depth exceeded\r\nChatgpt4Online: RecursionError: maximum recursion depth exceeded in comparison\r\nChatAnywhere: RecursionError: maximum recursion depth exceeded while encoding a JSON object\r\nChatgptX: RecursionError: maximum recursion depth exceeded in comparison\r\nGptForLove: RuntimeUnavailableError: Could not find an available JavaScript runtime.\r\nChatBase: RecursionError: maximum recursion depth exceeded while encoding a JSON object\r\nGptGo: RecursionError: maximum recursion depth exceeded while calling a Python object\r\n```\r\n\r\n**Traceback**\r\n```\r\nERROR: Exception in ASGI application\r\nTraceback (most recent call last):\r\n File \"/usr/local/lib/python3.10/dist-packages/g4f/api/__init__.py\", line 85, in chat_completions\r\n response = g4f.ChatCompletion.create(\r\n File \"/usr/local/lib/python3.10/dist-packages/g4f/__init__.py\", line 76, in create\r\n return result if stream else ''.join(result)\r\n File \"/usr/local/lib/python3.10/dist-packages/g4f/Provider/retry_provider.py\", line 59, in create_completion\r\n self.raise_exceptions()\r\n File \"/usr/local/lib/python3.10/dist-packages/g4f/Provider/retry_provider.py\", line 87, in raise_exceptions\r\n raise RuntimeError(\"\\n\".join([\"RetryProvider failed:\"] + [\r\nRuntimeError: RetryProvider failed:\r\nChatAnywhere: RecursionError: maximum recursion depth exceeded\r\nChatBase: RecursionError: maximum recursion depth exceeded\r\nChatgptX: RecursionError: maximum recursion depth exceeded\r\nYou: RecursionError: maximum recursion depth exceeded while calling a Python object\r\nGptGo: RecursionError: maximum recursion depth exceeded\r\nChatgpt4Online: RecursionError: maximum recursion depth exceeded\r\nGptForLove: RecursionError: maximum recursion depth exceeded\r\n\r\nDuring handling of the above exception, another exception occurred:\r\n\r\nTraceback (most recent call last):\r\n File \"/usr/local/lib/python3.10/dist-packages/uvicorn/protocols/http/h11_impl.py\", line 408, in run_asgi\r\n result = await app( # type: ignore[func-returns-value]\r\n File \"/usr/local/lib/python3.10/dist-packages/uvicorn/middleware/proxy_headers.py\", line 84, in __call__\r\n return await self.app(scope, receive, send)\r\n File \"/usr/local/lib/python3.10/dist-packages/fastapi/applications.py\", line 1106, in __call__\r\n await super().__call__(scope, receive, send)\r\n File \"/usr/local/lib/python3.10/dist-packages/starlette/applications.py\", line 122, in __call__\r\n await self.middleware_stack(scope, receive, send)\r\n File \"/usr/local/lib/python3.10/dist-packages/starlette/middleware/errors.py\", line 184, in __call__\r\n raise exc\r\n File \"/usr/local/lib/python3.10/dist-packages/starlette/middleware/errors.py\", line 162, in __call__\r\n await self.app(scope, receive, _send)\r\n File \"/usr/local/lib/python3.10/dist-packages/starlette/middleware/exceptions.py\", line 79, in __call__\r\n raise exc\r\n File \"/usr/local/lib/python3.10/dist-packages/starlette/middleware/exceptions.py\", line 68, in __call__\r\n await self.app(scope, receive, sender)\r\n File \"/usr/local/lib/python3.10/dist-packages/fastapi/middleware/asyncexitstack.py\", line 20, in __call__\r\n raise e\r\n File \"/usr/local/lib/python3.10/dist-packages/fastapi/middleware/asyncexitstack.py\", line 17, in __call__\r\n await self.app(scope, receive, send)\r\n File \"/usr/local/lib/python3.10/dist-packages/starlette/routing.py\", line 718, in __call__\r\n await route.handle(scope, receive, send)\r\n File \"/usr/local/lib/python3.10/dist-packages/starlette/routing.py\", line 276, in handle\r\n await self.app(scope, receive, send)\r\n File \"/usr/local/lib/python3.10/dist-packages/starlette/routing.py\", line 66, in app\r\n response = await func(request)\r\n File \"/usr/local/lib/python3.10/dist-packages/fastapi/routing.py\", line 274, in app\r\n raw_response = await run_endpoint_function(\r\n File \"/usr/local/lib/python3.10/dist-packages/fastapi/routing.py\", line 191, in run_endpoint_function\r\n return await dependant.call(**values)\r\n File \"/usr/local/lib/python3.10/dist-packages/g4f/api/__init__.py\", line 91, in chat_completions\r\n logging.exception(e)\r\n File \"/usr/lib/python3.10/logging/__init__.py\", line 2113, in exception\r\n error(msg, *args, exc_info=exc_info, **kwargs)\r\n File \"/usr/lib/python3.10/logging/__init__.py\", line 2105, in error\r\n root.error(msg, *args, **kwargs)\r\n File \"/usr/lib/python3.10/logging/__init__.py\", line 1506, in error\r\n self._log(ERROR, msg, args, **kwargs)\r\n File \"/usr/lib/python3.10/logging/__init__.py\", line 1624, in _log\r\n self.handle(record)\r\n File \"/usr/lib/python3.10/logging/__init__.py\", line 1634, in handle\r\n self.callHandlers(record)\r\n File \"/usr/lib/python3.10/logging/__init__.py\", line 1696, in callHandlers\r\n hdlr.handle(record)\r\n File \"/usr/lib/python3.10/logging/__init__.py\", line 968, in handle\r\n self.emit(record)\r\n File \"/usr/lib/python3.10/logging/__init__.py\", line 1100, in emit\r\n msg = self.format(record)\r\n File \"/usr/lib/python3.10/logging/__init__.py\", line 943, in format\r\n return fmt.format(record)\r\n File \"/usr/lib/python3.10/logging/__init__.py\", line 686, in format\r\n record.exc_text = self.formatException(record.exc_info)\r\n File \"/usr/lib/python3.10/logging/__init__.py\", line 636, in formatException\r\n traceback.print_exception(ei[0], ei[1], tb, None, sio)\r\n File \"/usr/lib/python3.10/traceback.py\", line 120, in print_exception\r\n for line in te.format(chain=chain):\r\n File \"/usr/local/lib/python3.10/dist-packages/exceptiongroup/_formatting.py\", line 248, in format\r\n yield from _ctx.emit(exc.format_exception_only())\r\n File \"/usr/local/lib/python3.10/dist-packages/exceptiongroup/_formatting.py\", line 64, in emit\r\n for text in text_gen:\r\n File \"/usr/local/lib/python3.10/dist-packages/exceptiongroup/_formatting.py\", line 335, in format_exception_only\r\n if isinstance(self.__notes__, collections.abc.Sequence):\r\n File \"/usr/lib/python3.10/abc.py\", line 119, in __instancecheck__\r\n return _abc_instancecheck(cls, instance)\r\nRecursionError: maximum recursion depth exceeded in comparison\r\n```\r\n\r\n\r\n", "code": null, "pr_html_url": null, "commit_html_url": null, "file_loc": {"base_commit": "1ade1d959cbc9aea7cf653bbe5b6c414ba486c97", "files": [{"path": "g4f/cli.py", "Loc": {}}]}, "own_code_loc": [], "ass_file_loc": [], "other_rep_loc": [], "analysis": {"iss_type": "1", "iss_reason": "5", "loc_way": "comment", "loc_scope": "0", "info_type": "Code"}, "max_topk": 0, "file_topk": 0, "loctype": {"code": ["g4f/cli.py"], "doc": [], "test": [], "config": [], "asset": []}}
|