organization
string
repo_name
string
base_commit
string
iss_html_url
string
iss_label
string
title
string
body
string
code
null
pr_html_url
string
commit_html_url
string
file_loc
string
own_code_loc
list
ass_file_loc
list
other_rep_loc
list
analysis
dict
loctype
dict
iss_has_pr
int64
AUTOMATIC1111
stable-diffusion-webui
d9499f4301018ebd2977685d098381aa4111d2ae
https://github.com/AUTOMATIC1111/stable-diffusion-webui/issues/13724
enhancement
[Feature Request]: Sort items by date by default
### Is there an existing issue for this? - [X] I have searched the existing issues and checked the recent builds/commits ### What would your feature do ? I hope to use time sorting by default when opening additional interfaces, so that I can immediately try the new model I just downloaded. ### Proposed workflow 1. Go to .... 2. Press .... 3. ... ### Additional information _No response_
null
null
https://github.com/AUTOMATIC1111/stable-diffusion-webui/commit/d9499f4301018ebd2977685d098381aa4111d2ae
{'base_commit': 'd9499f4301018ebd2977685d098381aa4111d2ae', 'files': [{'path': 'javascript/extraNetworks.js', 'status': 'modified', 'Loc': {"(None, 'setupExtraNetworksForTab', 18)": {'add': [51, 54, 98, 99], 'mod': [30, 56, 57, 58, 59, 65, 91, 92, 93, 94]}, '(None, None, None)': {'add': [115]}, "(None, 'applyExtraNetworkSort', 116)": {'add': [116]}}}, {'path': 'modules/shared_options.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [236]}}}, {'path': 'modules/ui_extra_networks.py', 'status': 'modified', 'Loc': {"(None, 'create_ui', 357)": {'add': [397], 'mod': [384, 385]}}}]}
[]
[]
[]
{ "iss_type": "4", "iss_reason": "2", "loc_way": "commit", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "modules/shared_options.py", "modules/ui_extra_networks.py", "javascript/extraNetworks.js" ], "doc": [], "test": [], "config": [], "asset": [] }
null
AUTOMATIC1111
stable-diffusion-webui
22bcc7be428c94e9408f589966c2040187245d81
https://github.com/AUTOMATIC1111/stable-diffusion-webui/issues/9102
bug-report
[Bug]: Model Dropdown Select on Firefox is obscured by svelte pre-loader
### Is there an existing issue for this? - [X] I have searched the existing issues and checked the recent builds/commits ### What happened? Seems the latest pull request has added pre-loaders and I noticed that the model dropdown is constantly loading and therefore obscuring the dropdown from the user. This is only happening in Firefox, Chrome for example is fine. ``` .wrap.default.svelte-gjihhp { inset: 0; } ``` I just set it to `display: none` to access it ### Steps to reproduce the problem Load up in Firefox and try to change the model See attached screenshot ![Screenshot_1](https://user-images.githubusercontent.com/3169931/228311409-22be3832-0348-424c-9298-08e76cb166a7.jpg) ### What should have happened? Have access to the model dropdown select ### Commit where the problem happens f1db987 ### What platforms do you use to access the UI ? Windows ### What browsers do you use to access the UI ? Mozilla Firefox ### Command Line Arguments ```Shell No ``` ### List of extensions No ### Console logs ```Shell Uncaught (in promise) TypeError: q[R[H]] is undefined ct http://127.0.0.1:7860/assets/Blocks.5efe22d4.js:76 ct http://127.0.0.1:7860/assets/Blocks.5efe22d4.js:76 ze http://127.0.0.1:7860/assets/Blocks.5efe22d4.js:76 ``` ### Additional information _No response_
null
null
https://github.com/joysun545/stable-diffusion-webui/commit/22bcc7be428c94e9408f589966c2040187245d81
{'base_commit': '22bcc7be428c94e9408f589966c2040187245d81', 'files': [{'path': 'modules/ui.py', 'status': 'modified', 'Loc': {"(None, 'create_ui', 437)": {'add': [1630]}}}]}
[]
[]
[]
{ "iss_type": "2", "iss_reason": "1", "loc_way": "commit", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "modules/ui.py" ], "doc": [], "test": [], "config": [], "asset": [] }
null
huggingface
transformers
a8e3336a850e856188350a93e67d77c07c85b8af
https://github.com/huggingface/transformers/issues/2008
wontfix
Expand run_lm_finetuning.py to all models
## 🚀 Feature [run_lm_finetuning.py](https://github.com/huggingface/transformers/blob/b0ee7c7df3d49a819c4d6cef977214bd91f5c075/examples/run_lm_finetuning.py) is a very useful tool for finetuning many models the library provided. But it doesn't cover all the models. Currently available models are: - gpt2 - openai-gpt - bert - roberta - distilbert - camembert And not available ones: - ctrl - xlm - xlnet - transfo-xl - albert ## Motivation Most important part of such a library is that it can be easily finetuned. `run_lm_finetuning.py` gives us that opportunity but why say no more :)
null
null
https://github.com/huggingface/transformers/commit/3dcb748e31be8c7c9e4f62926c5c144c62d07218 https://github.com/huggingface/transformers/commit/a8e3336a850e856188350a93e67d77c07c85b8af
{'base_commit': 'a8e3336a850e856188350a93e67d77c07c85b8af', 'files': [{'path': 'examples/ner/run_ner.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [33], 'mod': [41]}}}, {'path': 'examples/ner/run_tf_ner.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [16, 17, 18, 19, 21, 22, 23, 24, 25, 37, 38, 39, 41, 42, 43, 44, 45, 52]}, "(None, 'main', 457)": {'mod': [512, 513, 523, 530, 565, 587, 614, 615]}}}, {'path': 'examples/run_glue.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [32], 'mod': [35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 676, 677, 683, 695]}, "(None, 'train', 69)": {'mod': [75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101]}, "(None, 'main', 386)": {'mod': [445, 625, 626, 632, 637]}}}, {'path': 'examples/run_language_modeling.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [40], 'mod': [43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 60, 61, 62, 789]}, "('TextDataset', '__init__', 68)": {'mod': [76, 77, 78, 79, 80, 81, 82, 83]}, "(None, 'main', 464)": {'mod': [696, 699, 701, 703, 706, 708, 712, 722, 730, 771, 772]}}}, {'path': 'examples/run_squad.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [32], 'mod': [35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 75, 76, 77, 78, 79, 80, 81, 845]}, "(None, 'train', 76)": {'mod': [83, 84, 85, 86, 87, 88, 89, 90, 91]}, "(None, 'main', 477)": {'mod': [516, 760, 761, 765, 770, 820, 821]}}}, {'path': 'src/transformers/__init__.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [160, 319]}}}, {'path': 'templates/adding_a_new_example_script/run_xxx.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [30], 'mod': [33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 75, 76, 77, 78, 79, 80, 709]}, "(None, 'set_seed', 69)": {'mod': [71, 72, 73]}, "(None, 'main', 388)": {'mod': [421, 629, 630, 634, 639, 690, 691]}}}]}
[]
[]
[]
{ "iss_type": "4", "iss_reason": "2", "loc_way": "commit", "loc_scope": "", "info_type": "Code" }
{ "code": [ "examples/run_squad.py", "templates/adding_a_new_example_script/run_xxx.py", "examples/run_glue.py", "src/transformers/__init__.py", "examples/ner/run_tf_ner.py", "examples/run_language_modeling.py", "examples/ner/run_ner.py" ], "doc": [], "test": [], "config": [], "asset": [] }
null
huggingface
transformers
88d7f96e33c3f3e541bcdd913f2ff1e50aa18c1b
https://github.com/huggingface/transformers/issues/5212
BartConfig wrong decoder_start_token_id?
# 🐛 Bug ## Information Model I am using (Bert, XLNet ...): Bart Language I am using the model on (English, Chinese ...): English ## To reproduce Steps to reproduce the behavior: ``` from transformers import BartConfig, BartTokenizer config = BartConfig.from_pretrained('facebook/bart-large') tokenizer = BartTokenizer.from_pretrained('facebook/bart-large') config.decoder_start_token_id >>> 2 tokenizer.bos_token_id >>> 0 # != config.decoder_start_token_id tokenizer.eos_token_id >>> 2 ``` It is misleading in the documentation of the function ```generate```` *decoder_start_token_id=None – (optional) int If an encoder-decoder model starts decoding with a different token than BOS. Defaults to None and is changed to BOS later.* ## Expected behavior I expect that decoder_start_token_id = tokenizer.bos_token_id, but maybe the model is designed to start decoding with EOS token.
null
https://github.com/huggingface/transformers/pull/5306
null
{'base_commit': '88d7f96e33c3f3e541bcdd913f2ff1e50aa18c1b', 'files': [{'path': 'src/transformers/modeling_tf_utils.py', 'status': 'modified', 'Loc': {"('TFPreTrainedModel', 'generate', 551)": {'mod': [645, 646]}}}, {'path': 'src/transformers/modeling_utils.py', 'status': 'modified', 'Loc': {"('PreTrainedModel', 'generate', 871)": {'mod': [965, 966]}}}]}
[]
[]
[]
{ "iss_type": "2", "iss_reason": "1", "loc_way": "pr", "loc_scope": "0", "info_type": "Code\n里的docstring" }
{ "code": [ "src/transformers/modeling_utils.py", "src/transformers/modeling_tf_utils.py" ], "doc": [], "test": [], "config": [], "asset": [] }
null
geekan
MetaGPT
ba9e3fe6267ea81a3d546b7aa5bcf0122f365e51
https://github.com/geekan/MetaGPT/issues/1153
bug
mermaid: Generating ..seq_flow/20240402032019.svg.. Error: Failed to launch the browser process!
**Bug description** Generating /app/metagpt/workspace/jqdlhwap/resources/seq_flow/20240402032019.svg.. Error: Failed to launch the browser process! **Bug solved method** **Environment information** "docker compose up -d" after clone already run "npm install -g @mermaid-js/mermaid-cli": root@84a2e77496b0:/app/metagpt# mmdc -h Usage: mmdc [options] Options: -V, --version output the version number -t, --theme [theme] Theme of the chart (choices: "default", "forest", "dark", "neutral", default: "default") -w, --width [width] Width of the page (default: 800) -H, --height [height] Height of the page (default: 600) -i, --input <input> Input mermaid file. Files ending in .md will be treated as Markdown and all charts (e.g. ```mermaid (...)```) will be extracted and generated. Use `-` to read from stdin. -o, --output [output] Output file. It should be either md, svg, png or pdf. Optional. Default: input + ".svg" -e, --outputFormat [format] Output format for the generated image. (choices: "svg", "png", "pdf", default: Loaded from the output file extension) -b, --backgroundColor [backgroundColor] Background color for pngs/svgs (not pdfs). Example: transparent, red, '#F0F0F0'. (default: "white") -c, --configFile [configFile] JSON configuration file for mermaid. -C, --cssFile [cssFile] CSS file for the page. -s, --scale [scale] Puppeteer scale factor (default: 1) -f, --pdfFit Scale PDF to fit chart -q, --quiet Suppress log output -p --puppeteerConfigFile [puppeteerConfigFile] JSON configuration file for puppeteer. -h, --help display help for command - LLM type and model name: zhipu-api / GLM-4 - System version: - Python version: - MetaGPT version or branch: main `run in docker` - packages version: - installation method: **Screenshots or logs** 2024-04-02 03:20:46.126 | INFO | metagpt.utils.mermaid:mermaid_to_file:48 - Generating /app/metagpt/workspace/jqdlhwap/resources/seq_flow/20240402032019.svg.. 2024-04-02 03:20:46.460 | WARNING | metagpt.utils.mermaid:mermaid_to_file:74 - Error: Failed to launch the browser process! [0402/032046.449080:ERROR:zygote_host_impl_linux.cc(100)] Running as root without --no-sandbox is not supported. See https://crbug.com/638180. TROUBLESHOOTING: https://pptr.dev/troubleshooting at Interface.onClose (file:///usr/lib/node_modules/@mermaid-js/mermaid-cli/node_modules/@puppeteer/browsers/lib/esm/launch.js:253:24) at Interface.emit (node:events:524:35) at Interface.close (node:internal/readline/interface:526:10) at Socket.onend (node:internal/readline/interface:252:10) at Socket.emit (node:events:524:35) at endReadableNT (node:internal/streams/readable:1378:12) at process.processTicksAndRejections (node:internal/process/task_queues:82:21)
null
https://github.com/FoundationAgents/MetaGPT/pull/1155
null
{'base_commit': 'ba9e3fe6267ea81a3d546b7aa5bcf0122f365e51', 'files': [{'path': 'metagpt/configs/mermaid_config.py', 'status': 'modified', 'Loc': {"('MermaidConfig', None, 13)": {'mod': [16]}}}, {'path': 'metagpt/utils/mermaid.py', 'status': 'modified', 'Loc': {"(None, 'mermaid_to_file', 17)": {'add': [83]}}}, {'path': 'config/config2.yaml', 'Loc': {}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "pr", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "metagpt/configs/mermaid_config.py", "metagpt/utils/mermaid.py" ], "doc": [], "test": [], "config": [ "config/config2.yaml" ], "asset": [] }
null
geekan
MetaGPT
c779f6977ecbdba075d7c81519edd5eaf6de2d0e
https://github.com/geekan/MetaGPT/issues/1197
Support for Cohere API
Please add support for Cohere API with all the built in RAG and tool use functionalities. Essentially, RAG and tool use in Cohere are just chat parameters definable by users. More information can be found at https://docs.cohere.com/reference/chat .
null
https://github.com/FoundationAgents/MetaGPT/pull/1193
null
{'base_commit': 'c779f6977ecbdba075d7c81519edd5eaf6de2d0e', 'files': [{'path': 'metagpt/const.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [53]}}}, {'path': 'metagpt/rag/factories/ranker.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [13]}, "('RankerFactory', '__init__', 20)": {'add': [24]}, "('RankerFactory', None, 17)": {'add': [47]}}}, {'path': 'metagpt/rag/schema.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [121]}}}, {'path': 'setup.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [42]}}}]}
[]
[]
[]
{ "iss_type": "4", "iss_reason": "2", "loc_way": "pr", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "metagpt/rag/schema.py", "metagpt/rag/factories/ranker.py", "setup.py", "metagpt/const.py" ], "doc": [], "test": [], "config": [], "asset": [] }
null
geekan
MetaGPT
b5bb4d7e63e72c3d118e449a3763c1ff4411f159
https://github.com/geekan/MetaGPT/issues/1547
ERROR | metagpt.utils.mmdc_playwright:mermaid_to_file:118 - Page.goto: net::ERR_FILE_NOT_FOUND at file:///root/anaconda3/envs/metagpt/lib/python3.11/site-packages/metagpt/utils/index.html
**Bug description** I installed playwright and its chrominum with the guidance and made configuration of mermaid. But it seems that the mermaid didn't work normally. ``` ERROR | metagpt.utils.mmdc_playwright:mermaid_to_file:118 - Page.goto: net::ERR_FILE_NOT_FOUND at file:///root/anaconda3/envs/metagpt/lib/python3.11/site-packages/metagpt/utils/index.html Call log: navigating to "file:///root/anaconda3/envs/metagpt/lib/python3.11/site-packages/metagpt/utils/index.html", waiting until "load" ``` **Bug solved method** I think it may do with the configuration but the document of this part is not so clear. I don't know how to fill in the "path" if I'm using playwright and whether it is right to keep that stuff in default. This is my configuration: ```yaml llm: api_type: 'openai' # or azure / ollama / groq etc. Check LLMType for more options api_key: '[MY_API_KEY]' # MY_API_KEY model: 'yi-lightning' # or gpt-3.5-turbo base_url: 'https://api.lingyiwanwu.com/v1' # or any forward url. # proxy: 'YOUR_LLM_PROXY_IF_NEEDED' # Optional. If you want to use a proxy, set it here. # pricing_plan: 'YOUR_PRICING_PLAN' # Optional. If your pricing plan uses a different name than the `model`. mermaid: engine: 'playwright' # nodejs/ink/playwright/pyppeteer # path: 'mmdc' # such as './node_modules/.bin/mmdc' # puppeteer_config: './config/puppeteer-config' # only for nodejs # pyppeteer_path: '/usr/bin/google-chrome-stable' # only for pyppeteer ``` This is your example: ```yaml mermaid: engine: 'nodejs' # nodejs/ink/playwright/pyppeteer path: 'mmdc' # such as './node_modules/.bin/mmdc' puppeteer_config: './config/puppeteer-config' # only for nodejs pyppeteer_path: '/usr/bin/google-chrome-stable' # only for pyppeteer ``` **Environment information** - LLM type and model name: - System version: ubuntu 22.04 - Python version: 3.11 - MetaGPT version or branch: 0.8 <!-- Dependent packagess:the packages version cause the bug(like `pydantic 1.10.8`), installation method(like `pip install metagpt` or `pip install from source` or `run in docker`) --> - packages version: - installation method: pip **Screenshots or logs** 2024-10-29 16:22:16.900 | WARNING | metagpt.utils.cost_manager:update_cost:49 - Model yi-lightning not found in TOKEN_COSTS. 2024-10-29 16:22:16.903 | INFO | metagpt.utils.git_repository:rename_root:219 - Rename directory /root/workspace/cited_papaer_eval/workspace/20241029162137 to /root/workspace/cited_papaer_eval/workspace/scholar_cited_evaluation 2024-10-29 16:22:16.904 | INFO | metagpt.utils.file_repository:save:57 - save to: /root/workspace/cited_papaer_eval/workspace/scholar_cited_evaluation/docs/prd/20241029162216.json 2024-10-29 16:22:17.435 | ERROR | metagpt.utils.mmdc_playwright:mermaid_to_file:118 - Page.goto: net::ERR_FILE_NOT_FOUND at file:///root/anaconda3/envs/metagpt/lib/python3.11/site-packages/metagpt/utils/index.html Call log: navigating to "file:///root/anaconda3/envs/metagpt/lib/python3.11/site-packages/metagpt/utils/index.html", waiting until "load"
null
https://github.com/FoundationAgents/MetaGPT/pull/1564
null
{'base_commit': 'b5bb4d7e63e72c3d118e449a3763c1ff4411f159', 'files': [{'path': 'metagpt/utils/mmdc_playwright.py', 'status': 'modified', 'Loc': {"(None, 'mermaid_to_file', 17)": {'mod': [84, 85, 86, 87]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "pr", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "metagpt/utils/mmdc_playwright.py" ], "doc": [], "test": [], "config": [], "asset": [] }
null
langflow-ai
langflow
7276f699fc85c611f1c3f83a19a368da9841e3a4
https://github.com/langflow-ai/langflow/issues/2892
question
Flow as tool Usage
### Discussed in https://github.com/langflow-ai/langflow/discussions/2891 <div type='discussions-op-text'> <sup>Originally posted by **pavansandeep2910** July 23, 2024</sup> I cannot understand how to load files to see them in flow as tool component. can anyone help me direct to flow as tool usage?</div>
null
https://github.com/langflow-ai/langflow/pull/3093
null
{'base_commit': '7276f699fc85c611f1c3f83a19a368da9841e3a4', 'files': [{'path': 'src/backend/base/langflow/components/prototypes/SubFlow.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [2], 'mod': [4, 6, 9, 10, 11, 12]}, "('SubFlowComponent', 'build', 98)": {'add': [103], 'mod': [102, 105, 108, 112, 114, 115]}, "('SubFlowComponent', None, 15)": {'mod': [15, 17, 18, 19, 22, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 98, 99, 100]}, "('SubFlowComponent', 'get_flow_names', 24)": {'mod': [25, 26]}, "('SubFlowComponent', 'update_build_config', 35)": {'mod': [36, 39, 41]}, "('SubFlowComponent', 'add_inputs_to_build_config', 58)": {'mod': [71]}}}, {'path': 'src/backend/base/langflow/initial_setup/setup.py', 'status': 'modified', 'Loc': {"(None, 'load_starter_projects', 342)": {'mod': [346]}}}, {'path': 'src/backend/base/langflow/inputs/inputs.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [231]}, "('SecretStrInput', None, 215)": {'mod': [229]}}}, {'path': 'src/frontend/src/CustomNodes/GenericNode/components/handleRenderComponent/index.tsx', 'status': 'modified', 'Loc': {'(None, None, 1)': {'mod': [1]}, '(None, None, 4)': {'mod': [4]}, '(None, None, 40)': {'mod': [40, 41, 42, 43]}, '(None, None, 103)': {'mod': [103]}}}, {'path': 'src/frontend/src/CustomNodes/GenericNode/components/parameterComponent/index.tsx', 'status': 'modified', 'Loc': {'(None, None, 352)': {'mod': [352, 353]}, '(None, None, 365)': {'mod': [365, 366]}}}, {'path': 'src/frontend/src/CustomNodes/hooks/use-handle-new-value.tsx', 'status': 'modified', 'Loc': {'(None, None, 66)': {'mod': [66, 67]}}}, {'path': 'src/frontend/src/components/parameterRenderComponent/component/refreshParameterComponent/index.tsx', 'status': 'modified', 'Loc': {'(None, None, 32)': {'mod': [32]}}}]}
[]
[]
[]
{ "iss_type": "3", "iss_reason": "1", "loc_way": "pr", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "src/backend/base/langflow/components/prototypes/SubFlow.py", "src/backend/base/langflow/initial_setup/setup.py", "src/backend/base/langflow/inputs/inputs.py", "src/frontend/src/components/parameterRenderComponent/component/refreshParameterComponent/index.tsx", "src/frontend/src/CustomNodes/GenericNode/components/parameterComponent/index.tsx", "src/frontend/src/CustomNodes/GenericNode/components/handleRenderComponent/index.tsx", "src/frontend/src/CustomNodes/hooks/use-handle-new-value.tsx" ], "doc": [], "test": [], "config": [], "asset": [] }
null
langflow-ai
langflow
b3b5290598f5970fd6a1a092fe4d11211008a04d
https://github.com/langflow-ai/langflow/issues/5378
bug
URL component lost imported urls in tool mode when refresh UI
### Bug Description I have this issue when i build multi agent flow like import json below - URL Component Before refresh UI: ![image](https://github.com/user-attachments/assets/f361501d-3d11-4e4a-b560-203faf8a4935) ![image](https://github.com/user-attachments/assets/4ee6c93f-914b-4efa-b0eb-5a13f5c404fa) After refresh UI: ![image](https://github.com/user-attachments/assets/2224a76e-1dd7-42d4-8531-a5e86719653c) ![image](https://github.com/user-attachments/assets/3753754f-3a16-4bfc-a82e-1286c7504fe3) I have check in normal mode of URL Component, and this bug does not appear. ### Reproduction 1. Create flow 2. Add URL component, change to Tool mode 3. Input Urls 4. Save flow 5. Reload UI (Press F5) ### Expected behavior URL component has keep Urls has input. In Tool mode ### Who can help? _No response_ ### Operating System Windows 11/Docker ### Langflow Version v1.1.1 ### Python Version None ### Screenshot _No response_ ### Flow File [Simple Agent (bug).json](https://github.com/user-attachments/files/18206463/Simple.Agent.bug.json)
null
https://github.com/langflow-ai/langflow/pull/5316
null
{'base_commit': 'b3b5290598f5970fd6a1a092fe4d11211008a04d', 'files': [{'path': 'src/frontend/src/components/core/parameterRenderComponent/components/toggleShadComponent/index.tsx', 'status': 'modified', 'Loc': {'(None, None, 39)': {'mod': [39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54]}}}, {'path': 'src/frontend/src/pages/FlowPage/components/nodeToolbarComponent/index.tsx', 'status': 'modified', 'Loc': {'(None, None, 5)': {'add': [5]}, '(None, None, 44)': {'add': [44]}, '(None, None, 79)': {'add': [79]}, '(None, None, 98)': {'add': [98]}, '(None, None, 149)': {'add': [149]}, '(None, None, 11)': {'mod': [11]}, '(None, None, 23)': {'mod': [23, 24, 25]}, '(None, None, 145)': {'mod': [145, 146]}, '(None, None, 158)': {'mod': [158, 159]}, '(None, None, 174)': {'mod': [174]}, '(None, None, 291)': {'mod': [291]}, '(None, None, 403)': {'mod': [403, 404, 405, 407, 408, 409]}, '(None, None, 421)': {'mod': [421, 422, 423, 424, 425, 426]}, '(None, None, 471)': {'mod': [471, 472, 473, 474, 475, 476, 477, 478, 479, 480, 481, 482, 483, 484, 485]}}}]}
[]
[]
[]
{ "iss_type": "2", "iss_reason": "5", "loc_way": "pr", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "src/frontend/src/pages/FlowPage/components/nodeToolbarComponent/index.tsx", "src/frontend/src/components/core/parameterRenderComponent/components/toggleShadComponent/index.tsx" ], "doc": [], "test": [], "config": [], "asset": [] }
null
langflow-ai
langflow
a4dc5381b2cf31c507cc32f9027f76bf00d61ccc
https://github.com/langflow-ai/langflow/issues/3536
bug
Prompt component does not pass variables correctly
### Bug Description I have prompt with two variables. {image_url} Value of image is https://oaidalleapiprodscus.blob.core.windows.net/private/org-N0bzhj17kGCdvkPCuGgpUuhO/user-RwGFQKsTTGw8hOO1ResFbpwQ/img-GH9Zkmf4RabhW48gVeZslvyb.png?st=2024-08-23T19%3A00%3A31Z&se=2024-08-23T21%3A00%3A31Z&sp=r&sv=2024-08-04&sr=b&rscd=inline&rsct=image/png&skoid=d505667d-d6c1-4a0a-bac7-5c84a87759f8&sktid=a48cca56-e6da-484e-a814-9c849652bcb3&skt=2024-08-23T18%3A20%3A28Z&ske=2024-08-24T18%3A20%3A28Z&sks=b&skv=2024-08-04&sig=HXgLmH4hDFvbuV/rPEX//ifP0PLXUUlCYP9ZDpFEAHM%3D {post_id} Value of Post ID is 11620 Prompt is Task: Upload the image from the provided URL to WordPress and set it as the featured image for the specified post. Image URL: {image_url} Post ID: {post_id} This worked on 1.0.15, but after I upgraded to 1.0.16 the second variable is not passed on, it repeats the image one. Component output Task: Upload the image from the provided URL to WordPress and set it as the featured image for the specified post. Image URL: https://oaidalleapiprodscus.blob.core.windows.net/private/org-N0bzhj17kGCdvkPCuGgpUuhO/user-RwGFQKsTTGw8hOO1ResFbpwQ/img-GH9Zkmf4RabhW48gVeZslvyb.png?st=2024-08-23T19%3A00%3A31Z&se=2024-08-23T21%3A00%3A31Z&sp=r&sv=2024-08-04&sr=b&rscd=inline&rsct=image/png&skoid=d505667d-d6c1-4a0a-bac7-5c84a87759f8&sktid=a48cca56-e6da-484e-a814-9c849652bcb3&skt=2024-08-23T18%3A20%3A28Z&ske=2024-08-24T18%3A20%3A28Z&sks=b&skv=2024-08-04&sig=HXgLmH4hDFvbuV/rPEX//ifP0PLXUUlCYP9ZDpFEAHM%3D Post ID: https://oaidalleapiprodscus.blob.core.windows.net/private/org-N0bzhj17kGCdvkPCuGgpUuhO/user-RwGFQKsTTGw8hOO1ResFbpwQ/img-GH9Zkmf4RabhW48gVeZslvyb.png?st=2024-08-23T19%3A00%3A31Z&se=2024-08-23T21%3A00%3A31Z&sp=r&sv=2024-08-04&sr=b&rscd=inline&rsct=image/png&skoid=d505667d-d6c1-4a0a-bac7-5c84a87759f8&sktid=a48cca56-e6da-484e-a814-9c849652bcb3&skt=2024-08-23T18%3A20%3A28Z&ske=2024-08-24T18%3A20%3A28Z&sks=b&skv=2024-08-04&sig=HXgLmH4hDFvbuV/rPEX//ifP0PLXUUlCYP9ZDpFEAHM%3D ### Reproduction Create prompt with multiple variables. Add value in each. Try to build prompt and you will find that only first variable is being pulled. ![image](https://github.com/user-attachments/assets/cc478377-6c3d-4a71-ba13-ab6e1773413e) ![image](https://github.com/user-attachments/assets/fad07891-e9d1-474f-b7a9-efb3084c4caf) ### Expected behavior multiple variables should work ### Who can help? _No response_ ### Operating System Render ### Langflow Version 1.0.16 ### Python Version 3.12 ### Screenshot _No response_ ### Flow File _No response_
null
https://github.com/langflow-ai/langflow/pull/3698
null
{'base_commit': 'a4dc5381b2cf31c507cc32f9027f76bf00d61ccc', 'files': [{'path': 'src/backend/base/langflow/custom/custom_component/component.py', 'status': 'modified', 'Loc': {"('Component', '__init__', 42)": {'add': [71], 'mod': [45, 56]}, "('Component', '_reset_all_output_values', 88)": {'mod': [89, 90]}, "('Component', '_build_state_model', 92)": {'mod': [98]}, "('Component', '__deepcopy__', 112)": {'mod': [119]}, "('Component', 'list_outputs', 166)": {'mod': [170]}, "('Component', 'get_output', 210)": {'mod': [223, 224]}, "('Component', 'set_output_value', 234)": {'mod': [235, 236]}, "('Component', 'map_outputs', 240)": {'mod': [253, 257]}, "('Component', 'map_inputs', 259)": {'mod': [270]}, "('Component', '_set_output_types', 290)": {'mod': [291]}, "('Component', 'get_output_by_method', 296)": {'mod': [299]}, "('Component', '_find_matching_output_method', 327)": {'mod': [329]}, "('Component', '__getattr__', 440)": {'mod': [445, 446]}, "('Component', '_set_outputs', 577)": {'mod': [581]}, "('Component', '_build_results', 619)": {'mod': [623]}}}, {'path': 'src/backend/base/langflow/graph/graph/base.py', 'status': 'modified', 'Loc': {"('Graph', '__apply_config', 318)": {'mod': [322]}}}, {'path': 'src/backend/base/langflow/template/field/base.py', 'status': 'modified', 'Loc': {"('Output', None, 161)": {'add': [180]}}}, {'path': 'src/backend/tests/unit/test_custom_component.py', 'status': 'modified', 'Loc': {"(None, 'test_custom_component_get_function_entrypoint_args_no_args', 397)": {'add': [402]}}}, {'path': 'src/backend/tests/unit/test_database.py', 'status': 'modified', 'Loc': {"(None, 'test_read_flow', 76)": {'mod': [76, 79]}}}]}
[]
[]
[]
{ "iss_type": "2", "iss_reason": "1", "loc_way": "pr", "loc_scope": "", "info_type": "Code" }
{ "code": [ "src/backend/base/langflow/template/field/base.py", "src/backend/base/langflow/graph/graph/base.py", "src/backend/base/langflow/custom/custom_component/component.py" ], "doc": [], "test": [ "src/backend/tests/unit/test_custom_component.py", "src/backend/tests/unit/test_database.py" ], "config": [], "asset": [] }
null
langflow-ai
langflow
20ceb42504087c712aaee41bfc17a870ae0109d4
https://github.com/langflow-ai/langflow/issues/2039
enhancement
[Feat] Firecrawl🔥 Integration
Hi all, Opening this issue after chatting with Rodrigo. It would be awesome to add a [Firecrawl](https://firecrawl.dev) web loader / tool for people to use it to scrape, crawl and extract LLM ready data from the web. Would love to hear your thoughts on how we can best integrate it.
null
https://github.com/langflow-ai/langflow/pull/2359
null
{'base_commit': '20ceb42504087c712aaee41bfc17a870ae0109d4', 'files': [{'path': 'poetry.lock', 'status': 'modified', 'Loc': {'(None, None, 2114)': {'add': [2114]}, '(None, None, 2436)': {'add': [2436]}, '(None, None, 2596)': {'add': [2596]}, '(None, None, 2600)': {'add': [2600]}, '(None, None, 4618)': {'add': [4618]}, '(None, None, 6082)': {'add': [6082]}, '(None, None, 2435)': {'mod': [2435]}, '(None, None, 2595)': {'mod': [2595]}, '(None, None, 2599)': {'mod': [2599]}, '(None, None, 4617)': {'mod': [4617]}, '(None, None, 6085)': {'mod': [6085]}, '(None, None, 10555)': {'mod': [10555]}}}, {'path': 'pyproject.toml', 'status': 'modified', 'Loc': {'(None, None, 94)': {'add': [94]}}}, {'path': 'src/backend/base/poetry.lock', 'status': 'modified', 'Loc': {'(None, None, 741)': {'add': [741]}, '(None, None, 3238)': {'mod': [3238]}}}, {'path': 'src/backend/base/pyproject.toml', 'status': 'modified', 'Loc': {'(None, None, 66)': {'add': [66]}}}, {'path': 'src/frontend/src/utils/styleUtils.ts', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [173, 365]}}}]}
[]
[]
[]
{ "iss_type": "4", "iss_reason": "2", "loc_way": "pr", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "src/frontend/src/utils/styleUtils.ts" ], "doc": [], "test": [], "config": [ "pyproject.toml", "poetry.lock", "src/backend/base/poetry.lock", "src/backend/base/pyproject.toml" ], "asset": [] }
null
langflow-ai
langflow
9d8009f2f5c5e3fd3bf47760debc787deb454b1a
https://github.com/langflow-ai/langflow/issues/3004
bug
Problem with Global Variables Setting Page
### Bug Description When entering http://127.0.0.1:7860/settings/global-variables I am getting error in JS console. ``` `DialogContent` requires a `DialogTitle` for the component to be accessible for screen reader users. If you want to hide the `DialogTitle`, you can wrap it with our VisuallyHidden component. For more information, see https://radix-ui.com/primitives/docs/components/dialog [index-BMduUo-e.js:3231:165](http://127.0.0.1:7860/assets/index-BMduUo-e.js) TitleWarning http://127.0.0.1:7860/assets/index-BMduUo-e.js:3231 Qj http://127.0.0.1:7860/assets/index-BMduUo-e.js:984 Hk http://127.0.0.1:7860/assets/index-BMduUo-e.js:984 Wk http://127.0.0.1:7860/assets/index-BMduUo-e.js:984 Pk http://127.0.0.1:7860/assets/index-BMduUo-e.js:984 Ek http://127.0.0.1:7860/assets/index-BMduUo-e.js:984 jg http://127.0.0.1:7860/assets/index-BMduUo-e.js:982 Wk http://127.0.0.1:7860/assets/index-BMduUo-e.js:984 Pk http://127.0.0.1:7860/assets/index-BMduUo-e.js:984 Gk http://127.0.0.1:7860/assets/index-BMduUo-e.js:984 Ct http://127.0.0.1:7860/assets/index-BMduUo-e.js:969 xt http://127.0.0.1:7860/assets/index-BMduUo-e.js:969 ``` I am also getting error when adding again earlier deleted variable: "Sorry, we found an unexpected error! Please report errors with detailed tracebacks on the [GitHub Issues](https://github.com/langflow-ai/langflow/issues) page. Thank you!" So as asked, I am kindly reporting it. Also, There is no feature to edit fields. ### Reproduction JS error: 1. Just enter the page Saving Error: 1. saving new variable 2. deleting this new variable 3. adding it again with same name ### Expected behavior It should work without error ### Who can help? _No response_ ### Operating System Windows 11 pro ### Langflow Version 1.13 ### Python Version None ### Screenshot _No response_ ### Flow File _No response_
null
https://github.com/langflow-ai/langflow/pull/3284
null
{'base_commit': '9d8009f2f5c5e3fd3bf47760debc787deb454b1a', 'files': [{'path': 'src/backend/base/langflow/api/v1/variable.py', 'status': 'modified', 'Loc': {"(None, 'create_variable', 17)": {'add': [22], 'mod': [26, 27, 28, 29, 30, 31, 32, 33, 34, 36, 37, 39, 40, 42, 44, 46, 47, 48, 49, 50, 51, 52]}, "(None, 'read_variables', 60)": {'add': [63], 'mod': [67, 68]}, "(None, 'update_variable', 74)": {'add': [79], 'mod': [83, 84, 85, 86, 87, 89, 90, 91, 92, 93, 94, 95]}, "(None, 'delete_variable', 101)": {'add': [105], 'mod': [109, 110, 111, 112, 113, 114, 115]}, '(None, None, None)': {'mod': [1, 5, 7, 10, 11]}}}, {'path': 'src/backend/base/langflow/services/variable/base.py', 'status': 'modified', 'Loc': {"('VariableService', None, 11)": {'add': [84], 'mod': [72]}}}, {'path': 'src/backend/base/langflow/services/variable/kubernetes.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [4], 'mod': [12, 13]}, "('KubernetesSecretService', None, 16)": {'add': [123], 'mod': [113, 114, 115, 116, 117, 118]}}}, {'path': 'src/backend/base/langflow/services/variable/service.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [1], 'mod': [11]}, "('DatabaseVariableService', 'get_variable', 68)": {'add': [78], 'mod': [86, 87]}, "('DatabaseVariableService', None, 22)": {'add': [90, 111]}, "('DatabaseVariableService', 'list_variables', 91)": {'mod': [92, 93]}, "('DatabaseVariableService', 'delete_variable', 112)": {'mod': [118, 123]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "2", "loc_way": "pr", "loc_scope": "", "info_type": "Code" }
{ "code": [ "src/backend/base/langflow/api/v1/variable.py", "src/backend/base/langflow/services/variable/base.py", "src/backend/base/langflow/services/variable/service.py", "src/backend/base/langflow/services/variable/kubernetes.py" ], "doc": [], "test": [], "config": [], "asset": [] }
null
Significant-Gravitas
AutoGPT
1e98d349877305a8ee9c84901282b5731675578f
https://github.com/Significant-Gravitas/AutoGPT/issues/803
debug mode not debug
### Duplicates - [X] I have searched the existing issues ### Steps to reproduce 🕹 attribute in chat.py is not named correctly ### Current behavior 😯 wrong name can't call the attrb on the object ### Expected behavior 🤔 to use the attrb in a call without error ### Your prompt 📝 ```yaml # Paste your prompt here
null
https://github.com/Significant-Gravitas/AutoGPT/pull/888
null
{'base_commit': '1e98d349877305a8ee9c84901282b5731675578f', 'files': [{'path': 'scripts/chat.py', 'status': 'modified', 'Loc': {"(None, 'chat_with_ai', 42)": {'mod': [67, 74, 113]}}}, {'path': 'scripts/json_parser.py', 'status': 'modified', 'Loc': {"(None, 'fix_json', 76)": {'mod': [94]}}}, {'path': 'scripts/main.py', 'status': 'modified', 'Loc': {"(None, 'parse_arguments', 266)": {'add': [268]}, '(None, None, None)': {'add': [294]}}}]}
[]
[]
[]
{ "iss_type": "2", "iss_reason": "1", "loc_way": "pr", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "scripts/json_parser.py", "scripts/main.py", "scripts/chat.py" ], "doc": [], "test": [], "config": [], "asset": [] }
null
fastapi
fastapi
0f152b4e97a102c0105f26d76d6e1bba3b12fc2a
https://github.com/fastapi/fastapi/issues/894
bug answered reviewed
RecursionError from response model in 0.47.1
### Describe the bug FastAPI 0.47.1 will not be able to start due to a `RecursionError` when there is a circular reference among models. The issue seems to originate from https://github.com/tiangolo/fastapi/pull/889. This works fine in 0.46.0. ### Environment - OS: Windows - FastAPI Version: 0.47.1 - Python version: 3.7.0 ### To Reproduce ```Python from typing import Optional from fastapi import FastAPI from pydantic import BaseModel, Field class Group(BaseModel): representative: Optional['Person'] = Field(None) class Person(BaseModel): group: Optional[Group] = Field(None) Group.update_forward_refs() app = FastAPI() @app.get('/group/{group_id}', response_model=Group) def get_group(group_id): return [] ``` ### Expected behavior No exception ### Actual output ``` Traceback (most recent call last): File "test.py", line 21, in <module> @app.get('/group/{group_id}', response_model=Group) File "D:\virtualenvs\test\lib\site-packages\fastapi\routing.py", line 494, in decorator callbacks=callbacks, File "D:\virtualenvs\test\lib\site-packages\fastapi\routing.py", line 438, in add_api_route callbacks=callbacks, File "D:\virtualenvs\test\lib\site-packages\fastapi\routing.py", line 275, in __init__ ] = create_cloned_field(self.response_field) File "D:\virtualenvs\test\lib\site-packages\fastapi\utils.py", line 100, in create_cloned_field use_type.__fields__[f.name] = create_cloned_field(f) File "D:\virtualenvs\test\lib\site-packages\fastapi\utils.py", line 100, in create_cloned_field use_type.__fields__[f.name] = create_cloned_field(f) File "D:\virtualenvs\test\lib\site-packages\fastapi\utils.py", line 100, in create_cloned_field use_type.__fields__[f.name] = create_cloned_field(f) [Previous line repeated 981 more times] File "D:\virtualenvs\test\lib\site-packages\fastapi\utils.py", line 97, in create_cloned_field original_type.__name__, __config__=original_type.__config__ File "D:\virtualenvs\test\lib\site-packages\pydantic\main.py", line 773, in create_model return type(model_name, (__base__,), namespace) File "D:\virtualenvs\test\lib\site-packages\pydantic\main.py", line 152, in __new__ if issubclass(base, BaseModel) and base != BaseModel: File "D:\virtualenvs\test\lib\abc.py", line 143, in __subclasscheck__ return _abc_subclasscheck(cls, subclass) RecursionError: maximum recursion depth exceeded in comparison ```
null
null
https://github.com/fastapi/fastapi/commit/0f152b4e97a102c0105f26d76d6e1bba3b12fc2a
{'base_commit': '0f152b4e97a102c0105f26d76d6e1bba3b12fc2a', 'files': [{'path': 'fastapi/utils.py', 'status': 'modified', 'Loc': {"(None, 'create_cloned_field', 134)": {'mod': [134, 141, 142, 143, 160, 163]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "commit", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "fastapi/utils.py" ], "doc": [], "test": [], "config": [], "asset": [] }
null
oobabooga
text-generation-webui
05676caf70db7f3715cf6a3b4680f15efd45977a
https://github.com/oobabooga/text-generation-webui/issues/6202
bug stale
Llama-cpp-python 0.2.81 'already loaded' fails to load models
### Describe the bug Attempting to load a model after running the update-wizard-macos today (the version from a day or two ago worked fine) fails with the stack trace log included below. Notably, the error message references [this new issue in llama-cpp-python](https://github.com/abetlen/llama-cpp-python/issues/1575). ### Is there an existing issue for this? - [X] I have searched the existing issues ### Reproduction - Run the update wizard to update software. - Attempt to load a gguf model using the GPU and llama.cpp - Observe that loading fails. ### Screenshot ![Screenshot 2024-07-04 at 11 10 47 PM](https://github.com/oobabooga/text-generation-webui/assets/9359101/72148b05-8a43-4d2e-9fd5-7ba6fa57b317) ### Logs ```shell Traceback (most recent call last): File "/Users/patrickleiser/Documents/Programming/AI/text-generation-webui/modules/ui_model_menu.py", line 246, in load_model_wrapper shared.model, shared.tokenizer = load_model(selected_model, loader) File "/Users/patrickleiser/Documents/Programming/AI/text-generation-webui/modules/models.py", line 94, in load_model output = load_func_map[loader](model_name) File "/Users/patrickleiser/Documents/Programming/AI/text-generation-webui/modules/models.py", line 275, in llamacpp_loader model, tokenizer = LlamaCppModel.from_pretrained(model_file) File "/Users/patrickleiser/Documents/Programming/AI/text-generation-webui/modules/llamacpp_model.py", line 39, in from_pretrained LlamaCache = llama_cpp_lib().LlamaCache File "/Users/patrickleiser/Documents/Programming/AI/text-generation-webui/modules/llama_cpp_python_hijack.py", line 38, in llama_cpp_lib raise Exception(f"Cannot import 'llama_cpp_cuda' because '{imported_module}' is already imported. See issue #1575 in llama-cpp-python. Please restart the server before attempting to use a different version of llama-cpp-python.") Exception: Cannot import 'llama_cpp_cuda' because 'llama_cpp' is already imported. See issue #1575 in llama-cpp-python. Please restart the server before attempting to use a different version of llama-cpp-python. ``` ### System Info ```shell M1 Max Macbook Pro, MacOS 14.5 ``` Edit: Just realized that Ooobabooga was the one that created that issue on the llama-cpp-python project, so I guess this error was already known. Sorry if this issue is therefore somewhat redundant
null
https://github.com/oobabooga/text-generation-webui/pull/6227
null
{'base_commit': '05676caf70db7f3715cf6a3b4680f15efd45977a', 'files': [{'path': 'modules/llama_cpp_python_hijack.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [1]}, "(None, 'llama_cpp_lib', 13)": {'mod': [16, 17, 18, 19, 20, 21, 22, 24, 26, 28, 29, 30, 31, 32, 33, 34, 35, 37, 38, 39, 40, 41, 42, 43, 44, 46, 47, 48, 49, 50, 51, 52, 53, 55, 56, 57, 58, 59, 60, 61, 62, 64, 65, 67]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "pr", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "modules/llama_cpp_python_hijack.py" ], "doc": [], "test": [], "config": [], "asset": [] }
null
oobabooga
text-generation-webui
3c076c3c8096fa83440d701ba4d7d49606aaf61f
https://github.com/oobabooga/text-generation-webui/issues/2958
bug
Latest version of Pillow breaks current implementation in html_generator.py.
### Describe the bug Pillow 10.0.0 removed `ANTIALIAS` from `PIL.Image`. Current implementation requires 9.5.0, however the requirements.txt currently allows for 10.0.0 to be installed. ### Is there an existing issue for this? - [X] I have searched the existing issues ### Reproduction Add new characters with png images and load the webui in chat mode. ### Screenshot _No response_ ### Logs ```shell Traceback (most recent call last): File "G:\F\Projects\AI\text-generation-webui\one-click-installers-test\installer_files\env\lib\site-packages\gradio\routes.py", line 427, in run_predict output = await app.get_blocks().process_api( File "G:\F\Projects\AI\text-generation-webui\one-click-installers-test\installer_files\env\lib\site-packages\gradio\blocks.py", line 1323, in process_api result = await self.call_function( File "G:\F\Projects\AI\text-generation-webui\one-click-installers-test\installer_files\env\lib\site-packages\gradio\blocks.py", line 1051, in call_function prediction = await anyio.to_thread.run_sync( File "G:\F\Projects\AI\text-generation-webui\one-click-installers-test\installer_files\env\lib\site-packages\anyio\to_thread.py", line 33, in run_sync return await get_asynclib().run_sync_in_worker_thread( File "G:\F\Projects\AI\text-generation-webui\one-click-installers-test\installer_files\env\lib\site-packages\anyio\_backends\_asyncio.py", line 877, in run_sync_in_worker_thread return await future File "G:\F\Projects\AI\text-generation-webui\one-click-installers-test\installer_files\env\lib\site-packages\anyio\_backends\_asyncio.py", line 807, in run result = context.run(func, *args) File "G:\F\Projects\AI\text-generation-webui\one-click-installers-test\text-generation-webui\extensions\gallery\script.py", line 71, in generate_html image_html = f'<img src="file/{get_image_cache(path)}">' File "G:\F\Projects\AI\text-generation-webui\one-click-installers-test\text-generation-webui\modules\html_generator.py", line 144, in get_image_cache img = make_thumbnail(Image.open(path)) File "G:\F\Projects\AI\text-generation-webui\one-click-installers-test\text-generation-webui\modules\html_generator.py", line 132, in make_thumbnail image = ImageOps.fit(image, (350, 470), Image.ANTIALIAS) AttributeError: module 'PIL.Image' has no attribute 'ANTIALIAS' ``` ### System Info ```shell Windows 10 GPU: GTX 1080ti ```
null
https://github.com/oobabooga/text-generation-webui/pull/2954
null
{'base_commit': '3c076c3c8096fa83440d701ba4d7d49606aaf61f', 'files': [{'path': 'modules/html_generator.py', 'status': 'modified', 'Loc': {"(None, 'make_thumbnail', 129)": {'mod': [132]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "pr", "loc_scope": "0", "info_type": "Other\n依赖声明 " }
{ "code": [ "modules/html_generator.py" ], "doc": [], "test": [], "config": [], "asset": [] }
null
oobabooga
text-generation-webui
cf2c4e740b1d06e145c1992515d9b34e18affc95
https://github.com/oobabooga/text-generation-webui/issues/801
enhancement
How can we disable Gradio analytics?
**Description** How where / can this be implemented? https://github.com/brkirch/stable-diffusion-webui/commit/a534959cbcabc95af50fbbe4654f8c0ee1cdd41c `os.environ['GRADIO_ANALYTICS_ENABLED'] = 'False'` **Additional Context** For [stable-diffusion-webui](https://github.com/AUTOMATIC1111/stable-diffusion-webui) [preserve privacy by disabling gradio analytics globally #8658 ](https://github.com/AUTOMATIC1111/stable-diffusion-webui/pull/8658)
null
null
https://github.com/oobabooga/text-generation-webui/commit/cf2c4e740b1d06e145c1992515d9b34e18affc95
{'base_commit': 'cf2c4e740b1d06e145c1992515d9b34e18affc95', 'files': [{'path': 'server.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [0]}}}]}
[]
[]
[]
{ "iss_type": "4", "iss_reason": "2", "loc_way": "commit", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "server.py" ], "doc": [], "test": [], "config": [], "asset": [] }
null
oobabooga
text-generation-webui
a3085dba073fe8bdcfb5120729a84560f5d024c3
https://github.com/oobabooga/text-generation-webui/issues/1000
bug
Bot spams random numbers or does not load
### Describe the bug Hello, I installed oobabooga with the one click installer and I can not load the facebook_opt-2.7b (I copied the console into the log). I also installed the gpt4x alpaca model with the automatic installer(download-model.bat). If I chat with it, it just spams random 2 and 4 (I took a screenshot and pasted it down below). If I manually install the gpt4x model (with the help of this tutorial: https://youtu.be/nVC9D9fRyNU?t=162 ), I get the same output as the Facebook model in the log. ### Is there an existing issue for this? - [X] I have searched the existing issues ### Reproduction 1. Automatic installer 2. let download-model.bat download anon8231489123/gpt4-x-alpaca-13b-native-4bit-128g or follow this tutorial: 3. let download-model.bat download one model from the list 4. start-webui.bat has following arguments: python server.py --auto-devices --chat --wbits 4 --groupsize 128 ### Screenshot ![image](https://user-images.githubusercontent.com/125409728/230895729-d2f12173-81a5-4de6-9296-71845906ab01.png) ### Logs ```shell Starting the web UI... ===================================BUG REPORT=================================== Welcome to bitsandbytes. For bug reports, please submit your error trace to: https://github.com/TimDettmers/bitsandbytes/issues ================================================================================ CUDA SETUP: CUDA runtime path found: C:\Users\Caner\Documents\oobabooga-windows\installer_files\env\bin\cudart64_110.dll CUDA SETUP: Highest compute capability among GPUs detected: 8.6 CUDA SETUP: Detected CUDA version 117 CUDA SETUP: Loading binary C:\Users\Caner\Documents\oobabooga-windows\installer_files\env\lib\site-packages\bitsandbytes\libbitsandbytes_cuda117.dll... The following models are available: 1. anon8231489123_gpt4-x-alpaca-13b-native-4bit-128g 2. facebook_opt-2.7b Which one do you want to load? 1-2 2 Loading facebook_opt-2.7b... Could not find the quantized model in .pt or .safetensors format, exiting... Drücken Sie eine beliebige Taste . . . ``` ### System Info ```shell Windows 10 Version 22H2, Amd Ryzen 5800x, Palit Gamingpro Rtx 3080. ```
null
null
https://github.com/oobabooga/text-generation-webui/commit/a3085dba073fe8bdcfb5120729a84560f5d024c3
{'base_commit': 'a3085dba073fe8bdcfb5120729a84560f5d024c3', 'files': [{'path': 'modules/models.py', 'status': 'modified', 'Loc': {"(None, 'load_model', 40)": {'add': [176]}}}]}
[]
[]
[]
{ "iss_type": "2", "iss_reason": "1", "loc_way": "commit", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "modules/models.py" ], "doc": [], "test": [], "config": [], "asset": [] }
null
oobabooga
text-generation-webui
c4aa1a42b156b9c5ddcfb060cc497b2fba55430f
https://github.com/oobabooga/text-generation-webui/issues/1203
bug
When I click "Click Me" on (Character screen) it no longer generates a log of the current chat/instruction.
### Describe the bug When I click "Click Me" on (Character screen) it no longer generates a log of the current chat/instruction. This either got caused when I did an update to the whole program. Reinstall does not fix. I also ran some Antivirus and registry changes. I can generate new note pads and I reinstalled notepad on my PC. I've of course restarted my pc and I've tried firefox and opera as the host browser. This is a new problem just from today. But I did a few things on my PC. Note pad is also missing from my right click "Create new" list. However it is is in my folders create new list. The one you can click on in the top menu settings inside explorer. The first thing is I guess I need to rule out if others having the issue or not. Next it would be nice to get some utility to auto save all logs or something as an option. Thanks for any advice anyone. If it's a bug I will simply wait for a fix. It's possible I may have messed something but but it may also be a design flaw of the program if it's dependant on just one specific thing to save the chat log. I have no logs or other info to share at this time. ### Is there an existing issue for this? - [X] I have searched the existing issues ### Reproduction For me it's easy. Issue persists even if I restart PC, update or reinstall oobabooga. It worked last night. Today it does not work. ### Screenshot _No response_ ### Logs ```shell Not applicable. ``` ### System Info ```shell 12700K, Nvidia 4080, Windows 10. Running locally on my PC not a colab etc. Like I said I tried firefox and opera. Issue seems persistent. ```
null
null
https://github.com/oobabooga/text-generation-webui/commit/c4aa1a42b156b9c5ddcfb060cc497b2fba55430f
{'base_commit': 'c4aa1a42b156b9c5ddcfb060cc497b2fba55430f', 'files': [{'path': 'server.py', 'status': 'modified', 'Loc': {"(None, 'create_model_menus', 251)": {'mod': [324, 349]}}}]}
[]
[]
[]
{ "iss_type": "2", "iss_reason": "1", "loc_way": "commit", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "server.py" ], "doc": [], "test": [], "config": [], "asset": [] }
null
oobabooga
text-generation-webui
2b7ba9586fb80cfbc47c77ad7bbbb03f7d6bc0df
https://github.com/oobabooga/text-generation-webui/issues/2326
bug
extensions/openai KeyError: 'assistant'
### Describe the bug Starting after [https://github.com/oobabooga/text-generation-webui/pull/2291] Which I think it's a great improvement. ### Is there an existing issue for this? - [X] I have searched the existing issues ### Reproduction Start server with extension --openai --model openaccess-ai-collective_manticore-13b. Starting [DGdev91 Auto-GPT](https://github.com/DGdev91/Auto-GPT), runs 1 cycle, give 'y' for the second, the error appears. ### Screenshot _No response_ ### Logs ```shell openaccess-ai-collectiveException occurred during processing of request from ('127.0.0.1', 42032) Traceback (most recent call last): File "/home/mihai/miniconda3/envs/textgen/lib/python3.10/socketserver.py", line 683, in process_request_thread self.finish_request(request, client_address) File "/home/mihai/miniconda3/envs/textgen/lib/python3.10/socketserver.py", line 360, in finish_request self.RequestHandlerClass(request, client_address, self) File "/home/mihai/miniconda3/envs/textgen/lib/python3.10/socketserver.py", line 747, in __init__ self.handle() File "/home/mihai/miniconda3/envs/textgen/lib/python3.10/http/server.py", line 433, in handle self.handle_one_request() File "/home/mihai/miniconda3/envs/textgen/lib/python3.10/http/server.py", line 421, in handle_one_request method() File "/home/mihai/text-generation-webui/extensions/openai/script.py", line 310, in do_POST msg = role_formats[role].format(message=content) KeyError: 'assistant' ---------------------------------------- ``` ### System Info ```shell Win11 WSL2 Ubuntu 20.04 Python 3.10 ```
null
null
https://github.com/oobabooga/text-generation-webui/commit/2b7ba9586fb80cfbc47c77ad7bbbb03f7d6bc0df
{'base_commit': '2b7ba9586fb80cfbc47c77ad7bbbb03f7d6bc0df', 'files': [{'path': 'extensions/openai/script.py', 'status': 'modified', 'Loc': {"('Handler', 'do_POST', 159)": {'mod': [262]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "commit", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "extensions/openai/script.py" ], "doc": [], "test": [], "config": [], "asset": [] }
null
hacksider
Deep-Live-Cam
92a0994f01ec6ae7756951312a70e101fb33c7e5
https://github.com/hacksider/Deep-Live-Cam/issues/597
The app starts transparent!
Seems like everything was ok with the install. When I run I get the error: / warning: ``` > python run.py --execution-provider cuda Exception in Tkinter callback Traceback (most recent call last): File "C:\Users\dulci\.pyenv\pyenv-win\versions\3.10.6\lib\tkinter\__init__.py", line 1921, in __call__ return self.func(*args) File "C:\Users\dulci\.pyenv\pyenv-win\versions\3.10.6\lib\tkinter\__init__.py", line 839, in callit func(*args) File "C:\Users\dulci\Documents\Development\Deep-Live-Cam\.venv\lib\site-packages\customtkinter\windows\widgets\scaling\scaling_tracker.py", line 186, in check_dpi_scaling window.block_update_dimensions_event() File "C:\Users\dulci\.pyenv\pyenv-win\versions\3.10.6\lib\tkinter\__init__.py", line 2383, in __getattr__ return getattr(self.tk, attr) AttributeError: '_tkinter.tkapp' object has no attribute 'block_update_dimensions_event' (.venv) PS C:\Users\dulci\Documents\Development\Deep-Live-Cam> ``` And the application starts so transparent (opacity super low) that I can barely see it and I can't use it, because it is almost invisible against the desktop background. Can anybody suggest how to fix it? I am working with venv python 3.10.6.
null
https://github.com/hacksider/Deep-Live-Cam/pull/632
null
{'base_commit': '92a0994f01ec6ae7756951312a70e101fb33c7e5', 'files': [{'path': 'README.md', 'status': 'modified', 'Loc': {'(None, None, 2)': {'add': [2]}, '(None, None, 3)': {'add': [3]}, '(None, None, 37)': {'add': [37]}, '(None, None, 47)': {'add': [47]}, '(None, None, 68)': {'add': [68]}, '(None, None, 168)': {'add': [168]}, '(None, None, 365)': {'add': [365]}, '(None, None, 6)': {'mod': [6]}, '(None, None, 8)': {'mod': [8]}, '(None, None, 10)': {'mod': [10]}, '(None, None, 12)': {'mod': [12]}, '(None, None, 15)': {'mod': [15]}, '(None, None, 20)': {'mod': [20]}, '(None, None, 24)': {'mod': [24]}, '(None, None, 28)': {'mod': [28]}, '(None, None, 32)': {'mod': [32]}, '(None, None, 36)': {'mod': [36]}, '(None, None, 39)': {'mod': [39, 40, 41]}, '(None, None, 43)': {'mod': [43, 44, 46]}, '(None, None, 49)': {'mod': [49, 50, 51, 52, 53, 54, 55, 56, 57]}, '(None, None, 59)': {'mod': [59]}, '(None, None, 61)': {'mod': [61, 62]}, '(None, None, 64)': {'mod': [64]}, '(None, None, 66)': {'mod': [66, 67]}, '(None, None, 71)': {'mod': [71, 72]}, '(None, None, 75)': {'mod': [75]}, '(None, None, 77)': {'mod': [77]}, '(None, None, 82)': {'mod': [82]}, '(None, None, 84)': {'mod': [84, 85, 86]}, '(None, None, 91)': {'mod': [91, 92]}, '(None, None, 96)': {'mod': [96]}, '(None, None, 98)': {'mod': [98, 100]}, '(None, None, 105)': {'mod': [105, 106]}, '(None, None, 110)': {'mod': [110]}, '(None, None, 112)': {'mod': [112, 113]}, '(None, None, 118)': {'mod': [118, 119]}, '(None, None, 123)': {'mod': [123]}, '(None, None, 125)': {'mod': [125, 126]}, '(None, None, 131)': {'mod': [131, 132]}, '(None, None, 136)': {'mod': [136]}, '(None, None, 138)': {'mod': [138, 139]}, '(None, None, 144)': {'mod': [144, 145]}, '(None, None, 150)': {'mod': [150, 151]}, '(None, None, 153)': {'mod': [153, 154]}, '(None, None, 156)': {'mod': [156]}, '(None, None, 158)': {'mod': [158, 159, 160, 161, 162]}, '(None, None, 164)': {'mod': [164]}, '(None, None, 166)': {'mod': [166, 167]}, '(None, None, 170)': {'mod': [170]}, '(None, None, 197)': {'mod': [197]}, '(None, None, 206)': {'mod': [206]}, '(None, None, 210)': {'mod': [210]}, '(None, None, 224)': {'mod': [224]}, '(None, None, 237)': {'mod': [237]}, '(None, None, 247)': {'mod': [247]}, '(None, None, 274)': {'mod': [274]}, '(None, None, 306)': {'mod': [306]}, '(None, None, 314)': {'mod': [314]}, '(None, None, 343)': {'mod': [343, 344]}, '(None, None, 346)': {'mod': [346, 347]}, '(None, None, 353)': {'mod': [353]}, '(None, None, 360)': {'mod': [360]}}}, {'path': 'modules/ui.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [6, 548], 'mod': [10, 13, 18, 19, 22, 23, 24, 25, 27, 28, 29, 30, 32, 33, 34, 35, 37, 38]}, "(None, 'analyze_target', 154)": {'add': [167], 'mod': [155, 163, 166]}, "(None, 'create_source_target_popup', 176)": {'add': [182], 'mod': [191, 192, 200, 201, 203, 204, 206, 207, 210, 211, 214, 215, 217, 218, 221, 222, 224]}, "(None, 'update_popup_source', 221)": {'add': [233], 'mod': [228, 229, 238, 240, 241, 242, 243, 245, 246, 249, 250]}, "(None, 'select_source_path', 290)": {'add': [299, 302], 'mod': [294]}, "(None, 'swap_faces_paths', 305)": {'add': [323, 326]}, "(None, 'select_target_path', 329)": {'add': [338, 343, 346], 'mod': [333]}, "(None, 'update_preview', 435)": {'add': [445], 'mod': [437, 438, 441, 443, 444, 450]}, "(None, 'init', 61)": {'mod': [61]}, "(None, 'create_root', 70)": {'mod': [70, 73, 74, 75, 77, 78, 79, 80, 81, 83, 84, 86, 87, 89, 90, 92, 93, 95, 96, 99, 100, 103, 104, 106, 107, 108, 109, 112, 113, 116, 117, 119, 121, 122, 124, 125, 126, 129, 130, 132, 133, 135, 136, 138, 139, 141, 142, 144, 145, 147, 148, 149, 150]}, "(None, 'on_submit_click', 184)": {'mod': [189]}, "(None, 'on_button_click', 194)": {'mod': [195, 197, 198]}, "(None, 'create_preview', 258)": {'mod': [263, 265, 268, 269, 271]}, "(None, 'select_output_path', 349)": {'mod': [353, 355]}, "(None, 'check_and_ignore_nsfw', 364)": {'mod': [365, 367, 370, 372, 375, 376, 378]}, "(None, 'fit_image_to_size', 381)": {'mod': [390]}, "(None, 'toggle_preview', 417)": {'mod': [418]}, "(None, 'init_preview', 425)": {'mod': [431]}, "(None, 'create_webcam_preview', 463)": {'mod': [466, 467, 468, 469, 471, 484, 487, 512]}, "(None, 'on_submit_click', 527)": {'mod': [533]}, "(None, 'create_source_target_popup_for_webcam', 519)": {'mod': [540, 543, 546]}, "(None, 'refresh_data', 550)": {'mod': [553, 554, 563, 565, 566, 568, 569, 571, 572, 575, 576, 579, 580, 581, 584, 585, 588, 589, 590]}, "(None, 'update_webcam_source', 593)": {'mod': [596, 600, 601, 610, 612, 613, 614, 615, 617, 618, 621, 622, 623, 624]}, "(None, 'update_webcam_target', 629)": {'mod': [629, 632, 636, 637, 646, 648, 649, 650, 651, 653, 654, 657, 658, 659, 660]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "pr", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "modules/ui.py" ], "doc": [ "README.md" ], "test": [], "config": [], "asset": [] }
null
Textualize
rich
4be4f6bd24d2a35da0e50df943209ad24c068159
https://github.com/Textualize/rich/issues/388
enhancement done
[REQUEST] minimal table width
hey @willmcgugan, great package! in `Table.add_column` method it would be nice to have an _actual_ minimal with option. [The documentation](https://github.com/willmcgugan/rich/blob/c98bf070e4f3785dbb050b72c09663021c5b1d73/rich/table.py#L303) says that `width` argument sets the minimal with for a column, but in fact in my tests it sets a constant width for one. I'd like my column to be at least `width` wide but expand if there is a longer string to display.
null
https://github.com/Textualize/rich/pull/391
null
{'base_commit': '4be4f6bd24d2a35da0e50df943209ad24c068159', 'files': [{'path': 'CHANGELOG.md', 'status': 'modified', 'Loc': {'(None, None, 26)': {'add': [26]}}}, {'path': 'rich/console.py', 'status': 'modified', 'Loc': {"('Console', 'rule', 991)": {'mod': [993]}}}, {'path': 'rich/measure.py', 'status': 'modified', 'Loc': {"('Measurement', None, 11)": {'add': [45]}, "('Measurement', 'with_maximum', 34)": {'mod': [41]}}}, {'path': 'rich/table.py', 'status': 'modified', 'Loc': {"('Column', None, 29)": {'add': [50, 54]}, "('Table', '__init__', 118)": {'add': [123, 157]}, "('Table', 'add_column', 278)": {'add': [288, 303, 321]}, "('Table', '_calculate_column_widths', 410)": {'add': [417], 'mod': [458, 459]}, "('Table', '_measure_column', 558)": {'add': [587]}, "('Table', '__rich_measure__', 241)": {'mod': [249, 254, 255, 265]}, '(None, None, None)': {'mod': [740, 741, 743, 745, 746, 747, 748, 749, 750, 751, 753, 754, 755, 756, 757, 758, 759, 761]}}}, {'path': 'tests/test_measure.py', 'status': 'modified', 'Loc': {"(None, 'test_measure_renderables', 29)": {'add': [33]}}}, {'path': 'tests/test_table.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [3, 125]}}}]}
[]
[]
[]
{ "iss_type": "4", "iss_reason": "2", "loc_way": "pr", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "rich/measure.py", "rich/console.py", "rich/table.py" ], "doc": [ "CHANGELOG.md" ], "test": [ "tests/test_table.py", "tests/test_measure.py" ], "config": [], "asset": [] }
null
ytdl-org
youtube-dl
9fa57892790ce205634f6a7c83de2b9e52ab5284
https://github.com/ytdl-org/youtube-dl/issues/8799
site-support-request account-needed
Request support site: Viceland
Viceland is a new channel from Vice. Website at https://www.viceland.com. Appears to use Uplynk, and may be encrypted, so it may not be possible.
null
null
https://github.com/ytdl-org/youtube-dl/commit/9fa57892790ce205634f6a7c83de2b9e52ab5284
{'base_commit': '9fa57892790ce205634f6a7c83de2b9e52ab5284', 'files': [{'path': 'youtube_dl/extractor/uplynk.py', 'status': 'modified', 'Loc': {"('UplynkIE', None, 13)": {'add': [51], 'mod': [29, 30]}, "('UplynkIE', '_real_extract', 52)": {'mod': [53]}, "('UplynkPreplayIE', '_real_extract', 59)": {'mod': [64]}}}, {'path': 'youtube_dl/extractor/viceland.py', 'status': 'modified', 'Loc': {"('VicelandIE', None, 20)": {'add': [27]}}}]}
[]
[]
[]
{ "iss_type": "4", "iss_reason": "2", "loc_way": "commit", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "youtube_dl/extractor/uplynk.py", "youtube_dl/extractor/viceland.py" ], "doc": [], "test": [], "config": [], "asset": [] }
null
ytdl-org
youtube-dl
5dbe81a1d35ae704b5ea208698a6bb785923d71a
https://github.com/ytdl-org/youtube-dl/issues/8171
Vimeo ondemand download preiview only
I try to download a video from vimeo on demand but i only get the preview of. could someone help me plaese youtube-dl -u .com -p https://vimeo.com/ondemand/thelastcolony/150274832 --verbo [debug] System config: [] [debug] User config: [] [debug] Command-line args: [u'-u', u'PRIVATE', u'-p', u'PRIVATE', u'https://vimeo.com/ondemand/thelastcolony/150274832', u'--verbo'] [debug] Encodings: locale cp1252, fs mbcs, out cp437, pref cp1252 [debug] youtube-dl version 2016.01.01 [debug] Python version 2.7.10 - Windows-8-6.2.9200 [debug] exe versions: none [debug] Proxy map: {} [vimeo] Logging in [vimeo] 150274832: Downloading webpage [vimeo] 150274832: Extracting information [vimeo] 150274832: Downloading webpage [vimeo] 150274832: Downloading JSON metadata [vimeo] 150274832: Downloading m3u8 information [debug] Invoking downloader on u'https://01-lvl3-gcs-pdl.vimeocdn.com/vimeo-prod-skyfire-std-us/01/54/6/150274832/459356950.mp4?expires=1452223995&token=00c3c5830ebe84f9310d4' [download] The Last Colony-150274832.mp4 has already been downloaded [download] 100% of 119.97MiB
null
null
https://github.com/ytdl-org/youtube-dl/commit/5dbe81a1d35ae704b5ea208698a6bb785923d71a
{'base_commit': '5dbe81a1d35ae704b5ea208698a6bb785923d71a', 'files': [{'path': 'youtube_dl/extractor/vimeo.py', 'status': 'modified', 'Loc': {"('VimeoIE', '_real_extract', 264)": {'add': [356], 'mod': [265, 267, 269, 345]}, "('VimeoIE', '_extract_vimeo_url', 214)": {'mod': [220]}}}]}
[]
[]
[]
{ "iss_type": "2", "iss_reason": "5", "loc_way": "commit", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "youtube_dl/extractor/vimeo.py" ], "doc": [], "test": [], "config": [], "asset": [] }
null
ytdl-org
youtube-dl
365343131d752bece96d2279a3e0bcd7e9f0000f
https://github.com/ytdl-org/youtube-dl/issues/17728
[PluralSight] Unable to download captions JSON: HTTP Error 404: Not Found
last version I tested youtube-dl is up-to-date (2018.09.26) I'm try to download video from PluralSight. Video is ok but subtitle is cannot download. The error is WARNING: Unable to download captions JSON: HTTP Error 404: Not Found My command: youtube-dl --username xxx --password xxxx --sleep-interval 35 --max-sleep-interval 120 --sub-lang en --sub-format srt --write-sub https://app.pluralsight.com/library/courses/design-database-structure-sql-server-2014-70-465/table-of-contents Thanks.
null
null
https://github.com/ytdl-org/youtube-dl/commit/365343131d752bece96d2279a3e0bcd7e9f0000f
{'base_commit': '365343131d752bece96d2279a3e0bcd7e9f0000f', 'files': [{'path': 'youtube_dl/extractor/pluralsight.py', 'status': 'modified', 'Loc': {"('PluralsightIE', None, 112)": {'mod': [213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224]}, "('PluralsightIE', '_real_extract', 271)": {'mod': [416]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "commit", "loc_scope": "", "info_type": "Code" }
{ "code": [ "youtube_dl/extractor/pluralsight.py" ], "doc": [], "test": [], "config": [], "asset": [] }
null
ytdl-org
youtube-dl
50f84a9ae171233c08ada41e03f6555c5ed95236
https://github.com/ytdl-org/youtube-dl/issues/7427
"ERROR: Signature extraction failed" for youtube video
Hi, I have encountered an error with a video: https://www.youtube.com/watch?v=LDvVYqUMuJ0 ``` $ /tmp/ydl/youtube-dl/youtube-dl https://www.youtube.com/watch?v=LDvVYqUMuJ0 --verbose [debug] System config: [] [debug] User config: [] [debug] Command-line args: [u'https://www.youtube.com/watch?v=LDvVYqUMuJ0', u'--verbose'] [debug] Encodings: locale UTF-8, fs UTF-8, out UTF-8, pref UTF-8 [debug] youtube-dl version 2015.11.02 [debug] Python version 2.7.10+ - Linux-4.2.0-1-amd64-x86_64-with-debian-stretch-sid [debug] exe versions: ffmpeg 1.0.6, ffprobe 1.0.6, rtmpdump 2.4 [debug] Proxy map: {} [youtube] LDvVYqUMuJ0: Downloading webpage [youtube] LDvVYqUMuJ0: Downloading video info webpage [youtube] LDvVYqUMuJ0: Extracting video information WARNING: unable to extract html5 player; please report this issue on https://yt-dl.org/bug . Make sure you are using the latest version; type youtube-dl -U to update. Be sure to call youtube-dl with the --verbose flag and include its complete output. [youtube] {22} signature length 40.44, html5 player None ERROR: Signature extraction failed: Traceback (most recent call last): File "/tmp/ydl/youtube-dl/youtube-dl/youtube_dl/extractor/youtube.py", line 817, in _decrypt_signature video_id, player_url, s File "/tmp/ydl/youtube-dl/youtube-dl/youtube_dl/extractor/youtube.py", line 709, in _extract_signature_function raise ExtractorError('Cannot identify player %r' % player_url) ExtractorError: Cannot identify player u'https://s.ytimg.com/yts/jsbin/player-en_US-vfljDEtYP/base.js'; please report this issue on https://yt-dl.org/bug . Make sure you are using the latest version; type youtube-dl -U to update. Be sure to call youtube-dl with the --verbose flag and include its complete output. (caused by ExtractorError(u"Cannot identify player u'https://s.ytimg.com/yts/jsbin/player-en_US-vfljDEtYP/base.js'; please report this issue on https://yt-dl.org/bug . Make sure you are using the latest version; type youtube-dl -U to update. Be sure to call youtube-dl with the --verbose flag and include its complete output.",)); please report this issue on https://yt-dl.org/bug . Make sure you are using the latest version; type youtube-dl -U to update. Be sure to call youtube-dl with the --verbose flag and include its complete output. Traceback (most recent call last): File "/tmp/ydl/youtube-dl/youtube-dl/youtube_dl/extractor/youtube.py", line 817, in _decrypt_signature video_id, player_url, s File "/tmp/ydl/youtube-dl/youtube-dl/youtube_dl/extractor/youtube.py", line 709, in _extract_signature_function raise ExtractorError('Cannot identify player %r' % player_url) ExtractorError: Cannot identify player u'https://s.ytimg.com/yts/jsbin/player-en_US-vfljDEtYP/base.js'; please report this issue on https://yt-dl.org/bug . Make sure you are using the latest version; type youtube-dl -U to update. Be sure to call youtube-dl with the --verbose flag and include its complete output. Traceback (most recent call last): File "/tmp/ydl/youtube-dl/youtube-dl/youtube_dl/YoutubeDL.py", line 661, in extract_info ie_result = ie.extract(url) File "/tmp/ydl/youtube-dl/youtube-dl/youtube_dl/extractor/common.py", line 290, in extract return self._real_extract(url) File "/tmp/ydl/youtube-dl/youtube-dl/youtube_dl/extractor/youtube.py", line 1345, in _real_extract encrypted_sig, video_id, player_url, age_gate) File "/tmp/ydl/youtube-dl/youtube-dl/youtube_dl/extractor/youtube.py", line 827, in _decrypt_signature 'Signature extraction failed: ' + tb, cause=e) ExtractorError: Signature extraction failed: Traceback (most recent call last): File "/tmp/ydl/youtube-dl/youtube-dl/youtube_dl/extractor/youtube.py", line 817, in _decrypt_signature video_id, player_url, s File "/tmp/ydl/youtube-dl/youtube-dl/youtube_dl/extractor/youtube.py", line 709, in _extract_signature_function raise ExtractorError('Cannot identify player %r' % player_url) ExtractorError: Cannot identify player u'https://s.ytimg.com/yts/jsbin/player-en_US-vfljDEtYP/base.js'; please report this issue on https://yt-dl.org/bug . Make sure you are using the latest version; type youtube-dl -U to update. Be sure to call youtube-dl with the --verbose flag and include its complete output. (caused by ExtractorError(u"Cannot identify player u'https://s.ytimg.com/yts/jsbin/player-en_US-vfljDEtYP/base.js'; please report this issue on https://yt-dl.org/bug . Make sure you are using the latest version; type youtube-dl -U to update. Be sure to call youtube-dl with the --verbose flag and include its complete output.",)); please report this issue on https://yt-dl.org/bug . Make sure you are using the latest version; type youtube-dl -U to update. Be sure to call youtube-dl with the --verbose flag and include its complete output. ``` Thanks, Corey
null
null
https://github.com/ytdl-org/youtube-dl/commit/50f84a9ae171233c08ada41e03f6555c5ed95236
{'base_commit': '50f84a9ae171233c08ada41e03f6555c5ed95236', 'files': [{'path': 'youtube_dl/extractor/youtube.py', 'status': 'modified', 'Loc': {"('YoutubeIE', '_extract_signature_function', 704)": {'mod': [706]}, "('YoutubeIE', '_real_extract', 1008)": {'mod': [1346]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "commit", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "youtube_dl/extractor/youtube.py" ], "doc": [], "test": [], "config": [], "asset": [] }
null
ytdl-org
youtube-dl
ea1f5e5dbd6c58d4f0872a65b97611732f4b29bd
https://github.com/ytdl-org/youtube-dl/issues/16139
fixed
ITV BTCC videos support?
## Please follow the guide below - You will be asked some questions and requested to provide some information, please read them **carefully** and answer honestly - Put an `x` into all the boxes [ ] relevant to your *issue* (like this: `[x]`) - Use the *Preview* tab to see what your issue will actually look like --- ### Make sure you are using the *latest* version: run `youtube-dl --version` and ensure your version is *2018.04.09*. If it's not, read [this FAQ entry](https://github.com/rg3/youtube-dl/blob/master/README.md#how-do-i-update-youtube-dl) and update. Issues with outdated version will be rejected. - [*] I've **verified** and **I assure** that I'm running youtube-dl **2018.04.09** ### Before submitting an *issue* make sure you have: - [*] At least skimmed through the [README](https://github.com/rg3/youtube-dl/blob/master/README.md), **most notably** the [FAQ](https://github.com/rg3/youtube-dl#faq) and [BUGS](https://github.com/rg3/youtube-dl#bugs) sections - [*] [Searched](https://github.com/rg3/youtube-dl/search?type=Issues) the bugtracker for similar issues including closed ones - [*] Checked that provided video/audio/playlist URLs (if any) are alive and playable in a browser ### What is the purpose of your *issue*? - [ ] Bug report (encountered problems with youtube-dl) - [*] Site support request (request for adding support for a new site) - [ ] Feature request (request for a new functionality) - [ ] Question - [ ] Other --- ### The following sections concretize particular purposed issues, you can erase any section (the contents between triple ---) not applicable to your *issue* --- ### If the purpose of this *issue* is a *bug report*, *site support request* or you are not completely sure provide the full verbose output as follows: Add the `-v` flag to **your command line** you run youtube-dl with (`youtube-dl -v <your command line>`), copy the **whole** output and insert it here. It should look similar to one below (replace it with **your** log inserted between triple ```): ``` $ youtube-dl -v https://pastebin.com/raw/KxD6rhpF --geo-bypass-country UK [debug] System config: [] [debug] User config: [] [debug] Custom config: [] [debug] Command-line args: [u'-v', u'https://pastebin.com/raw/KxD6rhpF', u'--geo-bypass-country', u'UK'] [debug] Encodings: locale UTF-8, fs UTF-8, out UTF-8, pref UTF-8 [debug] youtube-dl version 2018.04.09 [debug] Python version 2.7.13 (CPython) - Linux-4.9.62-v7+-armv7l-with-debian-9.4 [debug] exe versions: ffmpeg 3.2.10-1, ffprobe 3.2.10-1 [debug] Proxy map: {} [debug] Using fake IP None (UK) as X-Forwarded-For. [generic] KxD6rhpF: Requesting header WARNING: Falling back on generic information extractor. [generic] KxD6rhpF: Downloading webpage [generic] KxD6rhpF: Extracting information [download] Downloading playlist: Brightcove video tester [generic] playlist Brightcove video tester: Collected 1 video ids (downloading 1 of them) [download] Downloading video 1 of 1 [debug] Using fake IP None (UK) as X-Forwarded-For. [debug] Using fake IP None (UK) as X-Forwarded-For. [brightcove:new] 5766870719001: Downloading webpage [brightcove:new] 5766870719001: Downloading JSON metadata ERROR: Access to this resource is forbidden by access policy. You might want to use a VPN or a proxy server (with --proxy) to workaround. Traceback (most recent call last): File "/usr/local/lib/python2.7/dist-packages/youtube_dl/extractor/brightcove.py", line 706, in _real_extract json_data = self._download_json(api_url, video_id, headers=headers) File "/usr/local/lib/python2.7/dist-packages/youtube_dl/extractor/common.py", line 692, in _download_json encoding=encoding, data=data, headers=headers, query=query) File "/usr/local/lib/python2.7/dist-packages/youtube_dl/extractor/common.py", line 634, in _download_webpage res = self._download_webpage_handle(url_or_request, video_id, note, errnote, fatal, encoding=encoding, data=data, headers=headers, query=query) File "/usr/local/lib/python2.7/dist-packages/youtube_dl/extractor/adobepass.py", line 1332, in _download_webpage_handle *args, **compat_kwargs(kwargs)) File "/usr/local/lib/python2.7/dist-packages/youtube_dl/extractor/common.py", line 539, in _download_webpage_handle urlh = self._request_webpage(url_or_request, video_id, note, errnote, fatal, data=data, headers=headers, query=query) File "/usr/local/lib/python2.7/dist-packages/youtube_dl/extractor/common.py", line 528, in _request_webpage raise ExtractorError(errmsg, sys.exc_info()[2], cause=err) ExtractorError: Unable to download JSON metadata: HTTP Error 403: Forbidden (caused by HTTPError()); please report this issue on https://yt-dl.org/bug . Make sure you are using the latest version; see https://yt-dl.org/update on how to update. Be sure to call youtube-dl with the --verbose flag and include its complete output. Traceback (most recent call last): File "/usr/local/lib/python2.7/dist-packages/youtube_dl/YoutubeDL.py", line 789, in extract_info ie_result = ie.extract(url) File "/usr/local/lib/python2.7/dist-packages/youtube_dl/extractor/common.py", line 440, in extract ie_result = self._real_extract(url) File "/usr/local/lib/python2.7/dist-packages/youtube_dl/extractor/brightcove.py", line 712, in _real_extract self.raise_geo_restricted(msg=message) File "/usr/local/lib/python2.7/dist-packages/youtube_dl/extractor/common.py", line 743, in raise_geo_restricted raise GeoRestrictedError(msg, countries=countries) GeoRestrictedError: Access to this resource is forbidden by access policy. ``` --- ### If the purpose of this *issue* is a *site support request* please provide all kinds of example URLs support for which should be included (replace following example URLs by **yours**): ITV separated the BTCC race videos from the hub (which also seems to be having issues as per https://github.com/rg3/youtube-dl/issues/15925) Lately the video are hosted at http://www.itv.com/btcc/races (ie for a particular weekend all videos are posted at individual pages like: http://www.itv.com/btcc/races/btcc-2018-all-the-action-from-brands-hatch) Skimming the source code of this sample weekend page, I extracted the vid params and built a test page: - Single video: https://pastebin.com/raw/KxD6rhpF **Question 1:** is the log error above pointing just to a geo restriction error or is there anything else involved that I missed? (ie: like writing some header to force the ITV scrapper to act instead of a generic one) ``` <!doctype html> <html> <head> <meta charset="UTF-8"> <title>Brightcove video tester</title> <style> .video-js { height: 360px; width: 640px; } .video-js.vjs-mouse .vjs-big-play-button { background-color: red; opacity: .3; } </style> <script src="https://ajax.googleapis.com/ajax/libs/jquery/3.2.1/jquery.min.js"></script> </head> <body> <video data-id="id3809566268703" data-video-id="5766870719001" data-account="1582188683001" data-player="HkiHLnNRx" data-embed="default" data-application-id="" data-fc-id="87459" class="video-js" controls=""> </video> <script src="https://players.brightcove.net/1582188683001/HkiHLnNRx_default/index.min.js"> </script> </body> </html> ``` **Question 2:** is there any way to generate a playlist of downloadable items based on pages like http://www.itv.com/btcc/races/btcc-2018-all-the-action-from-brands-hatch with youtube-dl?
null
null
https://github.com/ytdl-org/youtube-dl/commit/ea1f5e5dbd6c58d4f0872a65b97611732f4b29bd
{'base_commit': 'ea1f5e5dbd6c58d4f0872a65b97611732f4b29bd', 'files': [{'path': 'youtube_dl/extractor/extractors.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [480]}}}, {'path': 'youtube_dl/extractor/itv.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [9, 20]}, "('ITVIE', '_real_extract', 56)": {'add': [262]}}}]}
[]
[]
[]
{ "iss_type": "4", "iss_reason": "2", "loc_way": "commit", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "youtube_dl/extractor/itv.py", "youtube_dl/extractor/extractors.py" ], "doc": [], "test": [], "config": [], "asset": [] }
null
localstack
localstack
78f635ad3a8f819645f3991dfd244ff09f06a7f0
https://github.com/localstack/localstack/issues/8833
type: bug aws:cloudformation status: backlog
bug: CDK Table build with replicationRegions failing on latest
### Is there an existing issue for this? - [X] I have searched the existing issues ### Current Behavior Trying to deploy a [CDK Table](https://docs.aws.amazon.com/cdk/api/v2/docs/aws-cdk-lib.aws_dynamodb.Table.html) to a localstack environment results in the following: ``` stack | 0/3 | 8:56:58 PM | CREATE_FAILED | AWS::CloudFormation::Stack | stack Waiter StackCreateComplete failed: Waiter encountered a terminal failure state: For expression "Stacks[].StackStatus" we matched expected path: "CREATE_FAILED" at least once ``` ``` localstack | 2023-08-06T03:56:28.709 DEBUG --- [ asgi_gw_4] localstack.packages.api : Installation of dynamodb-local skipped (already installed). localstack | 2023-08-06T03:56:28.709 DEBUG --- [ asgi_gw_4] l.services.dynamodb.server : Starting DynamoDB Local: ['java', '-Xmx256m', '-javaagent:/usr/lib/localstack/dynamodb-local/latest/ddb-local-loader-0.1.jar', '-Djava.library.path=/usr/lib/localstack/dynamodb-local/latest/DynamoDBLocal_lib', '-jar', '/usr/lib/localstack/dynamodb-local/latest/DynamoDBLocal.jar', '-port', '35799', '-dbPath', '/var/lib/localstack/tmp/state/dynamodb'] localstack | 2023-08-06T03:56:28.710 DEBUG --- [uncthread160] localstack.utils.run : Executing command: ['java', '-Xmx256m', '-javaagent:/usr/lib/localstack/dynamodb-local/latest/ddb-local-loader-0.1.jar', '-Djava.library.path=/usr/lib/localstack/dynamodb-local/latest/DynamoDBLocal_lib', '-jar', '/usr/lib/localstack/dynamodb-local/latest/DynamoDBLocal.jar', '-port', '35799', '-dbPath', '/var/lib/localstack/tmp/state/dynamodb'] localstack | 2023-08-06T03:56:28.939 DEBUG --- [uncthread160] l.services.dynamodb.server : Initializing DynamoDB Local with the following configuration: localstack | 2023-08-06T03:56:28.939 DEBUG --- [uncthread160] l.services.dynamodb.server : Port: 35799 localstack | 2023-08-06T03:56:28.939 DEBUG --- [uncthread160] l.services.dynamodb.server : InMemory: false localstack | 2023-08-06T03:56:28.939 DEBUG --- [uncthread160] l.services.dynamodb.server : DbPath: /var/lib/localstack/tmp/state/dynamodb localstack | 2023-08-06T03:56:28.940 DEBUG --- [uncthread160] l.services.dynamodb.server : SharedDb: false localstack | 2023-08-06T03:56:28.940 DEBUG --- [uncthread160] l.services.dynamodb.server : shouldDelayTransientStatuses: false localstack | 2023-08-06T03:56:28.940 DEBUG --- [uncthread160] l.services.dynamodb.server : CorsParams: null localstack | 2023-08-06T03:56:28.950 DEBUG --- [uncthread160] l.services.dynamodb.server : localstack | 2023-08-06T03:56:29.715 INFO --- [ asgi_gw_4] botocore.credentials : Found credentials in environment variables. localstack | 2023-08-06T03:56:30.532 DEBUG --- [ asgi_gw_4] l.services.plugins : checking service health dynamodb:4566 localstack | 2023-08-06T03:56:30.534 INFO --- [ asgi_gw_4] localstack.utils.bootstrap : Execution of "require" took 1887.18ms localstack | 2023-08-06T03:56:30.879 DEBUG --- [ asgi_gw_1] l.services.plugins : checking service health kinesis:4566 localstack | 2023-08-06T03:56:30.886 INFO --- [ asgi_gw_1] l.s.k.kinesis_mock_server : Creating kinesis backend for account 000000000000 localstack | 2023-08-06T03:56:30.887 DEBUG --- [ asgi_gw_1] localstack.packages.api : Starting installation of kinesis-local... localstack | 2023-08-06T03:56:30.887 DEBUG --- [ asgi_gw_1] localstack.utils.run : Executing command: ['npm', 'install', '--prefix', '/var/lib/localstack/lib/kinesis-local/0.4.2', 'kinesis-local@0.4.2'] localstack | 2023-08-06T03:56:32.575 DEBUG --- [ asgi_gw_1] localstack.packages.core : Setting ownership root:root on /var/lib/localstack/lib/kinesis-local/0.4.2 localstack | 2023-08-06T03:56:32.575 DEBUG --- [ asgi_gw_1] localstack.packages.api : Installation of kinesis-local finished. localstack | 2023-08-06T03:56:32.576 DEBUG --- [ asgi_gw_1] l.s.k.kinesis_mock_server : starting kinesis process ['node', PosixPath('/var/lib/localstack/lib/kinesis-local/0.4.2/node_modules/kinesis-local/main.js')] with env vars {'KINESIS_MOCK_CERT_PATH': '/var/lib/localstack/lib/kinesis-local/0.4.2/node_modules/kinesis-local/server.json', 'KINESIS_MOCK_PLAIN_PORT': 42209, 'KINESIS_MOCK_TLS_PORT': 34279, 'SHARD_LIMIT': '100', 'ON_DEMAND_STREAM_COUNT_LIMIT': '10', 'AWS_ACCOUNT_ID': '000000000000', 'CREATE_STREAM_DURATION': '500ms', 'DELETE_STREAM_DURATION': '500ms', 'REGISTER_STREAM_CONSUMER_DURATION': '500ms', 'START_STREAM_ENCRYPTION_DURATION': '500ms', 'STOP_STREAM_ENCRYPTION_DURATION': '500ms', 'DEREGISTER_STREAM_CONSUMER_DURATION': '500ms', 'MERGE_SHARDS_DURATION': '500ms', 'SPLIT_SHARD_DURATION': '500ms', 'UPDATE_SHARD_COUNT_DURATION': '500ms', 'UPDATE_STREAM_MODE_DURATION': '500ms', 'SHOULD_PERSIST_DATA': 'true', 'PERSIST_PATH': '../../../var/lib/localstack/tmp/state/kinesis', 'PERSIST_FILE_NAME': '000000000000.json', 'PERSIST_INTERVAL': '5s', 'LOG_LEVEL': 'INFO'} localstack | 2023-08-06T03:56:32.576 DEBUG --- [uncthread166] localstack.utils.run : Executing command: ['node', PosixPath('/var/lib/localstack/lib/kinesis-local/0.4.2/node_modules/kinesis-local/main.js')] localstack | 2023-08-06T03:56:32.834 INFO --- [uncthread166] l.s.k.kinesis_mock_server : [info] kinesis.mock.KinesisMockService$ 2023-08-06T03:56:32.823005Z contextId=6956dd23-c61e-4aa1-80ce-b8bfc8d0894b, cacheConfig={"awsAccountId":"000000000000","awsRegion":"us-east-1","createStreamDuration":{"length":500,"unit":"MILLISECONDS"},"deleteStreamDuration":{"length":500,"unit":"MILLISECONDS"},"deregisterStreamConsumerDuration":{"length":500,"unit":"MILLISECONDS"},"initializeStreams":null,"logLevel":"INFO","mergeShardsDuration":{"length":500,"unit":"MILLISECONDS"},"onDemandStreamCountLimit":10,"persistConfig":{"fileName":"000000000000.json","interval":{"length":5,"unit":"SECONDS"},"loadIfExists":true,"path":"../../../var/lib/localstack/tmp/state/kinesis","shouldPersist":true},"registerStreamConsumerDuration":{"length":500,"unit":"MILLISECONDS"},"shardLimit":100,"splitShardDuration":{"length":500,"unit":"MILLISECONDS"},"startStreamEncryptionDuration":{"length":500,"unit":"MILLISECONDS"},"stopStreamEncryptionDuration":{"length":500,"unit":"MILLISECONDS"},"updateShardCountDuration":{"length":500,"unit":"MILLISECONDS"}} Logging Cache Config localstack | 2023-08-06T03:56:32.986 INFO --- [uncthread166] l.s.k.kinesis_mock_server : [info] kinesis.mock.KinesisMockService$ 2023-08-06T03:56:32.986197Z Starting Kinesis TLS Mock Service on port 34279 localstack | 2023-08-06T03:56:32.987 INFO --- [uncthread166] l.s.k.kinesis_mock_server : [info] kinesis.mock.KinesisMockService$ 2023-08-06T03:56:32.986862Z Starting Kinesis Plain Mock Service on port 42209 localstack | 2023-08-06T03:56:32.994 INFO --- [uncthread166] l.s.k.kinesis_mock_server : [info] kinesis.mock.KinesisMockService$ 2023-08-06T03:56:32.994001Z contextId=1d81ef53-1648-4fbc-8b16-f09375d77ece Starting persist data loop localstack | 2023-08-06T03:56:33.215 DEBUG --- [uncthread158] l.s.c.resource_provider : Executing callback method for AWS::DynamoDB::Table:ddbFooTabletable735E488F localstack | 2023-08-06T03:56:33.330 DEBUG --- [uncthread158] l.s.c.e.template_deployer : Error applying changes for CloudFormation stack "stack-ddbFooTableNestedStackdd-f1d922ad": 'NoneType' object has no attribute 'get' Traceback (most recent call last): localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/engine/template_deployer.py", line 965, in _run localstack | self.do_apply_changes_in_loop(changes, stack) localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/engine/template_deployer.py", line 1011, in do_apply_changes_in_loop localstack | should_deploy = self.prepare_should_deploy_change( localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/engine/template_deployer.py", line 1105, in prepare_should_deploy_change localstack | resolve_refs_recursively( localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack_ext/services/cloudformation/cloudformation_extended.py.enc", line 35, in resolve_refs_recursively localstack | A=resolve_refs_recursively_orig(*E,**F) localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/utils/functions.py", line 80, in func localstack | return wrapped(*args, **kwargs) localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/engine/template_deployer.py", line 178, in resolve_refs_recursively localstack | result = _resolve_refs_recursively( localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/utils/functions.py", line 80, in func localstack | return wrapped(*args, **kwargs) localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/engine/template_deployer.py", line 478, in _resolve_refs_recursively localstack | value[key] = resolve_refs_recursively( localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack_ext/services/cloudformation/cloudformation_extended.py.enc", line 35, in resolve_refs_recursively localstack | A=resolve_refs_recursively_orig(*E,**F) localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/utils/functions.py", line 80, in func localstack | return wrapped(*args, **kwargs) localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/engine/template_deployer.py", line 178, in resolve_refs_recursively localstack | result = _resolve_refs_recursively( localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/utils/functions.py", line 80, in func localstack | return wrapped(*args, **kwargs) localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/engine/template_deployer.py", line 478, in _resolve_refs_recursively localstack | value[key] = resolve_refs_recursively( localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack_ext/services/cloudformation/cloudformation_extended.py.enc", line 35, in resolve_refs_recursively localstack | A=resolve_refs_recursively_orig(*E,**F) localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/utils/functions.py", line 80, in func localstack | return wrapped(*args, **kwargs) localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/engine/template_deployer.py", line 178, in resolve_refs_recursively localstack | result = _resolve_refs_recursively( localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/utils/functions.py", line 80, in func localstack | return wrapped(*args, **kwargs) localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/engine/template_deployer.py", line 497, in _resolve_refs_recursively localstack | value[i] = resolve_refs_recursively( localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack_ext/services/cloudformation/cloudformation_extended.py.enc", line 35, in resolve_refs_recursively localstack | A=resolve_refs_recursively_orig(*E,**F) localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/utils/functions.py", line 80, in func localstack | return wrapped(*args, **kwargs) localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/engine/template_deployer.py", line 178, in resolve_refs_recursively localstack | result = _resolve_refs_recursively( localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/utils/functions.py", line 80, in func localstack | return wrapped(*args, **kwargs) localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/engine/template_deployer.py", line 290, in _resolve_refs_recursively localstack | resolved_getatt = get_attr_from_model_instance( localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/engine/template_deployer.py", line 102, in get_attr_from_model_instance localstack | attribute = attribute.get(part) localstack | AttributeError: 'NoneType' object has no attribute 'get' localstack | localstack | localstack | 2023-08-06T03:56:33.426 INFO --- [ asgi_gw_3] localstack.request.aws : AWS cloudformation.DescribeStackEvents => 200 localstack | 2023-08-06T03:56:33.445 INFO --- [ asgi_gw_2] localstack.request.aws : AWS cloudformation.DescribeStacks => 200 localstack | 2023-08-06T03:56:38.447 INFO --- [ asgi_gw_4] localstack.request.aws : AWS cloudformation.DescribeStackEvents => 200 localstack | 2023-08-06T03:56:38.458 INFO --- [ asgi_gw_3] localstack.request.aws : AWS cloudformation.DescribeStacks => 200 localstack | 2023-08-06T03:56:43.464 INFO --- [ asgi_gw_2] localstack.request.aws : AWS cloudformation.DescribeStackEvents => 200 localstack | 2023-08-06T03:56:43.473 INFO --- [ asgi_gw_4] localstack.request.aws : AWS cloudformation.DescribeStacks => 200 localstack | 2023-08-06T03:56:48.479 INFO --- [ asgi_gw_2] localstack.request.aws : AWS cloudformation.DescribeStackEvents => 200 localstack | 2023-08-06T03:56:48.489 INFO --- [ asgi_gw_4] localstack.request.aws : AWS cloudformation.DescribeStacks => 200 localstack | 2023-08-06T03:56:53.495 INFO --- [ asgi_gw_3] localstack.request.aws : AWS cloudformation.DescribeStackEvents => 200 localstack | 2023-08-06T03:56:53.502 INFO --- [ asgi_gw_2] localstack.request.aws : AWS cloudformation.DescribeStacks => 200 localstack | 2023-08-06T03:56:58.487 INFO --- [ asgi_gw_3] localstack.request.aws : AWS cloudformation.DescribeStackEvents => 200 localstack | 2023-08-06T03:56:58.495 INFO --- [ asgi_gw_2] localstack.request.aws : AWS cloudformation.DescribeStacks => 200 localstack | 2023-08-06T03:56:58.743 DEBUG --- [uncthread154] l.s.c.e.template_deployer : Error applying changes for CloudFormation stack "stack": Waiter StackCreateComplete failed: Waiter encountered a terminal failure state: For expression "Stacks[].StackStatus" we matched expected path: "CREATE_FAILED" at least once Traceback (most recent call last): localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/engine/template_deployer.py", line 965, in _run localstack | self.do_apply_changes_in_loop(changes, stack) localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/engine/template_deployer.py", line 1039, in do_apply_changes_in_loop localstack | self.apply_change(change, stack=stack) localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/engine/template_deployer.py", line 1152, in apply_change localstack | progress_event = executor.deploy_loop(resource_provider_payload) # noqa localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/resource_provider.py", line 572, in deploy_loop localstack | event = self.execute_action(resource_provider, payload) localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/resource_provider.py", line 638, in execute_action localstack | return resource_provider.create(request) localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/resource_provider.py", line 350, in create localstack | return self.create_or_delete(request) localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/resource_provider.py", line 499, in create_or_delete localstack | result_handler( localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/models/cloudformation.py", line 55, in _handle_result localstack | connect_to().cloudformation.get_waiter("stack_create_complete").wait( localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/botocore/waiter.py", line 55, in wait localstack | Waiter.wait(self, **kwargs) localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/botocore/waiter.py", line 375, in wait localstack | raise WaiterError( localstack | botocore.exceptions.WaiterError: Waiter StackCreateComplete failed: Waiter encountered a terminal failure state: For expression "Stacks[].StackStatus" we matched expected path: "CREATE_FAILED" at least once localstack | localstack | localstack | 2023-08-06T03:57:03.504 INFO --- [ asgi_gw_3] localstack.request.aws : AWS cloudformation.DescribeStackEvents => 200 localstack | 2023-08-06T03:57:03.511 INFO --- [ asgi_gw_2] localstack.request.aws : AWS cloudformation.DescribeStacks => 200 localstack | 2023-08-06T03:57:03.520 INFO --- [ asgi_gw_4] localstack.request.aws : AWS cloudformation.DescribeStackEvents => 200 ``` I am using the CDK Table class with a stream enabled, which requires replicas to enable the stream. There's no creative way around this issue that I'm aware of, due to the stack failing before the fully deployment of the Table class. ### Expected Behavior I'm expecting the build to succeed locally, so I can continue with our stack deployment chain for our local environments. The behavior is not present on `localstack/localstack-pro:2.2.0` or any release prior to the update this week. I am successfully able to deploy to live environments with the same CDK stack without error. ### How are you starting LocalStack? With a docker-compose file ### Steps To Reproduce #### How are you starting localstack (e.g., `bin/localstack` command, arguments, or `docker-compose.yml`) docker-compose up ``` version: "3.8" networks: ls: name: ls services: localstack: container_name: "${LOCALSTACK_DOCKER_NAME-localstack}" environment: - DEBUG=${DEBUG-1} - DISABLE_CORS_CHECKS=1 - DISABLE_CUSTOM_CORS_APIGATEWAY=1 - DOCKER_HOST=unix:///var/run/docker.sock - EXTRA_CORS_ALLOWED_ORIGINS=* - MAIN_DOCKER_NETWORK=ls - PERSISTENCE=${PERSISTENCE-} env_file: - ./localstack.local.env image: "localstack/localstack-pro:${LOCALSTACK_VERSION-latest}" networks: - ls ports: - "127.0.0.1:4566:4566" # LocalStack Gateway - "127.0.0.1:4510-4559:4510-4559" # external services port range - "127.0.0.1:53:53" # DNS config (required for Pro) - "127.0.0.1:53:53/udp" # DNS config (required for Pro) - "127.0.0.1:443:443" # LocalStack HTTPS Gateway (required for Pro) volumes: - "/var/run/docker.sock:/var/run/docker.sock" ``` .env file contains: ``` LOCALSTACK_API_KEY=xxxxxxxxxx ``` #### Client commands (e.g., AWS SDK code snippet, or sequence of "awslocal" commands) npx cdklocal deploy api ``` const account = process.env.CDK_ACCOUNT || "000000000000"; const region = "us-west-2"; new APIStack(app, "api", { crossRegionReferences: true, env: { account, region }, stackName: "api", }); ``` ``` new DynamoStack(this, "ddbFooTable", { billingMode: BillingMode.PROVISIONED, deletionProtection: false, encryption: TableEncryption.AWS_MANAGED, partitionKey: { name: "id", type: AttributeType.STRING }, pointInTimeRecovery: false, removalPolicy: false ? RemovalPolicy.RETAIN : RemovalPolicy.DESTROY, replicationRegions: ["us-west-1"], stream: StreamViewType.NEW_AND_OLD_IMAGES, tableName: "foo", }); ``` ### Environment ```markdown - OS: OSX 13.4 - LocalStack: latest (70e077bf43491cc0954698c1240159caa9cecc0ac6652b890b52aaf0801d5fcb) - aws-cdk-lib: 2.90.0 - aws-cdk-local: 2.18.0 ``` ### Anything else? Unfortunately, I'm in a position where I need the latest update due to Cognito User Pool domains not being functional in prior releases. I'm trying to get our devs to authenticate locally with an Oauth2 IDP flow instead of forcing them to authenticate with a live-stable environment. More information [here](https://github.com/localstack/localstack/issues/8700).
null
https://github.com/localstack/localstack/pull/8882
null
{'base_commit': '78f635ad3a8f819645f3991dfd244ff09f06a7f0', 'files': [{'path': 'localstack/services/cloudformation/engine/template_deployer.py', 'status': 'modified', 'Loc': {"(None, 'get_attr_from_model_instance', 75)": {'add': [100, 101]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "pr", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "localstack/services/cloudformation/engine/template_deployer.py" ], "doc": [], "test": [], "config": [], "asset": [] }
null
localstack
localstack
f4a188b6d51155a0831a3246f1d8e4f4be835861
https://github.com/localstack/localstack/issues/4652
type: bug status: triage needed
bug: LAMBDA_DOCKER_FLAGS doesn't work with -e
### Is there an existing issue for this? - [X] I have searched the existing issues ### Current Behavior Attempting to add environment variables to containers created to service lambda requests no longer works in the latest version of localstack. This works in localStack version 0.12.16 ### Expected Behavior Setting the environment variable `LAMBDA_DOCKER_FLAGS=-e TEST_VAL=True` on localstack's docker container will result in spawned containers created for serving lambda functions having the environment variable TEST_VAL set to True. ### How are you starting LocalStack? With a docker-compose file ### Steps To Reproduce 1. Run `docker-compose up -d` with the following docker-compose.yml ```yaml version: '2.1' services: localstack_ltest: container_name: "ltest" image: localstack/localstack:0.12.18 ports: - "4566:4566" environment: - DOCKER_HOST=unix:///var/run/docker.sock - LOCALSTACK_API_KEY=<removed> - LAMBDA_EXECUTOR=docker-reuse - LAMBDA_DOCKER_FLAGS=-e TEST_VAL=True volumes: - "/var/run/docker.sock:/var/run/docker.sock" restart: always ``` 2. Create the file logs-template.yml ```yaml --- AWSTemplateFormatVersion: '2010-09-09' Resources: LambdaFunctionLogGroup: Type: AWS::Logs::LogGroup Properties: RetentionInDays: 60 LogGroupName: !Join ["", ["/aws/lambda/", !Ref LambdaFunction]] LambdaFunctionRole: Type: AWS::IAM::Role Properties: AssumeRolePolicyDocument: Version: '2012-10-17' Statement: - Effect: Allow Principal: Service: - lambda.amazonaws.com Action: - sts:AssumeRole Path: / Policies: - PolicyName: LambdaRolePolicy PolicyDocument: Statement: - Effect: Allow Action: - 'logs:*' Resource: 'arn:aws:logs:*:*:*' LambdaFunction: Type: AWS::Lambda::Function Properties: FunctionName: "test-function" Role: !GetAtt LambdaFunctionRole.Arn Handler: index.lambda_handler Runtime: python3.8 Code: ZipFile: | import os def lambda_handler(event, context): print("environ: " + str(os.environ)) ``` 3. Run ` aws cloudformation deploy --stack-name test --template-file .\logs-template.yml --endpoint-url http://127.0.0.1:4566 --region us-east-1` 4. Run `aws --endpoint-url http://127.0.0.1:4566 --region us-east-1 lambda invoke --function-name test-function out.txt` 5. Run `aws --endpoint-url=http://localhost:4566 --region us-east-1 logs tail /aws/lambda/test-function` 6. Check for `"TEST_VAL": "True",` being in the output of the above command. ### Environment ```markdown Current configuration (broken) - OS: Ubuntu 20.04 - LocalStack version: 0.12.18 - LocalStack build date: 2021-09-27 - LocalStack build git hash: 00797f9e Working configuration - OS: Ubuntu 20.04 - LocalStack version: 0.12.16 - LocalStack Docker container id: b0137bad2045 - LocalStack build date: 2021-07-31 - LocalStack build git hash: f1262f74 ``` ### Anything else? _No response_
null
null
https://github.com/localstack/localstack/commit/f4a188b6d51155a0831a3246f1d8e4f4be835861
{'base_commit': 'f4a188b6d51155a0831a3246f1d8e4f4be835861', 'files': [{'path': 'localstack/services/awslambda/lambda_executors.py', 'status': 'modified', 'Loc': {"('LambdaExecutorContainers', 'run_lambda_executor', 495)": {'mod': [543, 544]}}}, {'path': 'tests/integration/test_lambda.py', 'status': 'modified', 'Loc': {"('TestLambdaBaseFeatures', 'test_large_payloads', 477)": {'add': [492]}}}]}
[]
[]
[]
{ "iss_type": "2", "iss_reason": "1", "loc_way": "commit", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "localstack/services/awslambda/lambda_executors.py" ], "doc": [], "test": [ "tests/integration/test_lambda.py" ], "config": [], "asset": [] }
null
localstack
localstack
dec1ba1b94153a4380cc94a0c8bd805f8922b6e3
https://github.com/localstack/localstack/issues/3202
status: triage needed area: configuration
Illegal path is passed into the HEAD request during the object download
# Type of request: This is a ... [ ] bug report [ ] feature request # Detailed description I have tried to upgrade localstack to 0.11.5(and after) and use the service port 4566. On local, I got passes to all tests we have been doing. But on CircleCI, I got errors when download object from s3. ``` An error occurred (404) when calling the HeadObject operation: Not Found ``` It goes through to 0.11.4, so I'm sure it's a bug, but what do you think? Would you have any advice? ## Expected behavior My test does as below 1. upload object to s3 bucket named "test-bucket" 1. list v2 for the bucket 1. download it So, I expected to succeed to this. Of course, this test has been passed until localstack upgrade. ## Actual behavior At CircleCI, I got below. Somehow "/test-bucket/test-bucket" is being passed as a HEAD request parameter at download time. This path is double of bucket name "/test-bucket". ``` : 2020-10-30 05:47:54,523:API: 172.20.0.5 - - [30/Oct/2020 05:47:54] "PUT /test-bucket HTTP/1.1" 200 - 2020-10-30 05:47:54,536:API: 172.20.0.5 - - [30/Oct/2020 05:47:54] "PUT /test-bucket/loadable/2020/04/06/000000_2_e462a109-916a-4b8f-b393-f5b01a6a8c10 HTTP/1.1" 200 - 2020-10-30 05:47:54,554:API: 172.20.0.5 - - [30/Oct/2020 05:47:54] "GET /test-bucket?list-type=2&max-keys=200&prefix=loadable%2F HTTP/1.1" 200 - 2020-10-30 05:47:54,566:API: 172.20.0.5 - - [30/Oct/2020 05:47:54] "GET /test-bucket/loadable/2020/04/06/000000_2_e462a109-916a-4b8f-b393-f5b01a6a8c10 HTTP/1.1" 206 - 2020-10-30 05:47:54,578:API: 172.20.0.5 - - [30/Oct/2020 05:47:54] "HEAD /test-bucket/test-bucket HTTP/1.1" 404 - 2020-10-30T05:47:54:WARNING:bootstrap.py: Thread run method <function AdaptiveThreadPool.submit.<locals>._run at 0x7f333a234f70>(None) failed: An error occurred (404) when calling the HeadObject operation: Not Found Traceback (most recent call last): File "/opt/code/localstack/localstack/utils/bootstrap.py", line 534, in run result = self.func(self.params) File "/opt/code/localstack/localstack/utils/async_utils.py", line 28, in _run return fn(*args, **kwargs) File "/opt/code/localstack/localstack/services/generic_proxy.py", line 560, in handler response = modify_and_forward(method=method, path=path_with_params, data_bytes=data, headers=headers, File "/opt/code/localstack/localstack/services/generic_proxy.py", line 333, in modify_and_forward listener_result = listener.forward_request(method=method, File "/opt/code/localstack/localstack/services/edge.py", line 81, in forward_request return do_forward_request(api, port, method, path, data, headers) File "/opt/code/localstack/localstack/services/edge.py", line 86, in do_forward_request result = do_forward_request_inmem(api, port, method, path, data, headers) File "/opt/code/localstack/localstack/services/edge.py", line 106, in do_forward_request_inmem response = modify_and_forward(method=method, path=path, data_bytes=data, headers=headers, File "/opt/code/localstack/localstack/services/generic_proxy.py", line 401, in modify_and_forward updated_response = update_listener.return_response(**kwargs) File "/opt/code/localstack/localstack/services/s3/s3_listener.py", line 1254, in return_response fix_range_content_type(bucket_name, path, headers, response) File "/opt/code/localstack/localstack/services/s3/s3_listener.py", line 465, in fix_range_content_type result = s3_client.head_object(Bucket=bucket_name, Key=key_name) File "/opt/code/localstack/.venv/lib/python3.8/site-packages/botocore/client.py", line 357, in _api_call return self._make_api_call(operation_name, kwargs) File "/opt/code/localstack/.venv/lib/python3.8/site-packages/botocore/client.py", line 676, in _make_api_call raise error_class(parsed_response, operation_name) botocore.exceptions.ClientError: An error occurred (404) when calling the HeadObject operation: Not Found ``` # Steps to reproduce ## Command used to start LocalStack sorry... ## Client code (AWS SDK code snippet, or sequence of "awslocal" commands) sorry... ┆Issue is synchronized with this [Jira Task](https://localstack.atlassian.net/browse/LOC-71) by [Unito](https://www.unito.io/learn-more)
null
https://github.com/localstack/localstack/pull/3370
null
{'base_commit': 'dec1ba1b94153a4380cc94a0c8bd805f8922b6e3', 'files': [{'path': 'localstack/services/s3/s3_listener.py', 'status': 'modified', 'Loc': {"(None, 'uses_path_addressing', 891)": {'mod': [892]}}}, {'path': 'tests/integration/test_s3.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [36]}, "('S3ListenerTest', None, 59)": {'add': [1454]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "pr", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "localstack/services/s3/s3_listener.py" ], "doc": [], "test": [ "tests/integration/test_s3.py" ], "config": [], "asset": [] }
null
OpenInterpreter
open-interpreter
7a53ba3dad9f7b2e31dac3fbb3162838eb9441c6
https://github.com/OpenInterpreter/open-interpreter/issues/58
Please support Azure OpenAI
null
null
https://github.com/openinterpreter/open-interpreter/pull/62
null
{'base_commit': '7a53ba3dad9f7b2e31dac3fbb3162838eb9441c6', 'files': [{'path': 'interpreter/cli.py', 'status': 'modified', 'Loc': {"(None, 'cli', 4)": {'add': [28, 39]}}}, {'path': 'interpreter/interpreter.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [52]}, "('Interpreter', '__init__', 62)": {'add': [69]}, "('Interpreter', 'verify_api_key', 255)": {'add': [263], 'mod': [260, 262, 271, 272, 275, 276, 278, 279, 280, 281, 284, 285, 286, 287, 289]}, "('Interpreter', 'respond', 296)": {'add': [340], 'mod': [312, 313, 314, 315, 316, 317, 318]}}}]}
[]
[]
[]
{ "iss_type": "4", "iss_reason": "2", "loc_way": "pr", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "interpreter/cli.py", "interpreter/interpreter.py" ], "doc": [], "test": [], "config": [], "asset": [] }
null
abi
screenshot-to-code
e69269d844b7089dec636516d6edb4f70911ebf6
https://github.com/abi/screenshot-to-code/issues/54
Support OPENAI_ API_ BASE for proxy URLs
How to add OPENAI_ API_ BASE code to use other proxy keys?
null
null
https://github.com/abi/screenshot-to-code/commit/e69269d844b7089dec636516d6edb4f70911ebf6
{'base_commit': 'e69269d844b7089dec636516d6edb4f70911ebf6', 'files': [{'path': 'backend/image_generation.py', 'status': 'modified', 'Loc': {"(None, 'process_tasks', 8)": {'mod': [8, 9]}, "(None, 'generate_image', 23)": {'mod': [23, 24]}, "(None, 'generate_images', 63)": {'mod': [63, 90]}}}, {'path': 'backend/llm.py', 'status': 'modified', 'Loc': {"(None, 'stream_openai_response', 8)": {'mod': [9, 11]}}}, {'path': 'backend/main.py', 'status': 'modified', 'Loc': {"(None, 'stream_code_test', 62)": {'add': [75, 85, 119], 'mod': [132]}}}, {'path': 'frontend/src/App.tsx', 'status': 'modified', 'Loc': {'(None, None, 39)': {'add': [39]}}}, {'path': 'frontend/src/components/SettingsDialog.tsx', 'status': 'modified', 'Loc': {'(None, None, 78)': {'add': [78]}}}, {'path': 'frontend/src/types.ts', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [2]}}}]}
[]
[]
[]
{ "iss_type": "3", "iss_reason": "5", "loc_way": "commit", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "frontend/src/App.tsx", "frontend/src/types.ts", "backend/main.py", "frontend/src/components/SettingsDialog.tsx", "backend/image_generation.py", "backend/llm.py" ], "doc": [], "test": [], "config": [], "asset": [] }
null
pytorch
pytorch
d363cf4639aacdaefbb8f69919f3c787a4519b7b
https://github.com/pytorch/pytorch/issues/38479
triaged module: numpy
torch.einsum does not pass equation argument to __torch_function__ API
## 🐛 Bug when delegating torch.einsum call to an object which implements `__torch_function__` API the equation argument is not passed resulting in the error. ```TypeError: einsum(): argument 'equation' (position 1) must be str, not Tensor``` this was tested on pytorch 1.5.0 I've actually found the cause of this bug and have written a fix. the following script illustrates the problem and the proposed solution ## To Reproduce ```python import torch class Wrapper(): def __init__(self,data): self.data = data def __torch_function__(self, func, types, args=(), kwargs=None): if kwargs is None: kwargs = {} #unwrap inputs if necessary def unwrap(v): return v.data if isinstance(v,Wrapper) else v args = map(unwrap,args) kwargs = {k:unwrap(v) for k,v in kwargs.items()} return func(*args, **kwargs) # fixed einsum implementation from torch import Tensor,_VF from torch._overrides import has_torch_function,handle_torch_function def fixed_einsum(equation,*operands): if not torch.jit.is_scripting(): if any(type(t) is not Tensor for t in operands) and has_torch_function(operands): # equation is not passed # return handle_torch_function(einsum, operands, *operands) return handle_torch_function(fixed_einsum, operands, equation,*operands) if len(operands) == 1 and isinstance(operands[0], (list, tuple)): # the old interface of passing the operands as one list argument operands = operands[0] # recurse incase operands contains value that has torch function #in the original implementation this line is omitted return fixed_einsum(equation,*operands) return _VF.einsum(equation, operands) if __name__ == "__main__": print(torch.__version__) # uncomment to use fixed einsum # torch.einsum = fixed_einsum #operands are wrapped x = Wrapper(torch.randn(5)) y = Wrapper(torch.randn(4)) assert torch.allclose(torch.einsum('i,j->ij',x, y),torch.ger(x,y)) # outer product print("works with wrapped inputs") #old interface operands is a list a = Wrapper(torch.randn(2,3)) b = Wrapper(torch.randn(5,3,7)) c = Wrapper(torch.randn(2,7)) assert torch.allclose(torch.einsum('ik,jkl,il->ij', [a, b, c]),torch.nn.functional.bilinear(a,c,b)) # bilinear interpolation print("works with old API operands is list") #equation is wrapped As = Wrapper(torch.randn(3,2,5)) Bs = Wrapper(torch.randn(3,5,4)) equation = Wrapper('bij,bjk->bik') assert torch.allclose(torch.einsum(equation, As, Bs),torch.matmul(As,Bs)) # batch matrix multiplication print("works with equation wrapped") #see that it also works with plain tensors x = torch.randn(5) y = torch.randn(4) assert torch.allclose(torch.einsum('i,j->ij',x, y),torch.ger(x,y)) print("works with no wrapped values") ``` cc @albanD @mruberry
null
null
https://github.com/pytorch/pytorch/commit/d363cf4639aacdaefbb8f69919f3c787a4519b7b
{'base_commit': 'd363cf4639aacdaefbb8f69919f3c787a4519b7b', 'files': [{'path': 'test/test_overrides.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [490]}}}, {'path': 'torch/_overrides.py', 'status': 'modified', 'Loc': {"(None, 'get_testing_overrides', 144)": {'mod': [264]}}}, {'path': 'torch/functional.py', 'status': 'modified', 'Loc': {"(None, 'einsum', 222)": {'add': [300], 'mod': [297]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "5", "loc_way": "commit", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "torch/functional.py", "torch/_overrides.py" ], "doc": [], "test": [ "test/test_overrides.py" ], "config": [], "asset": [] }
null
pytorch
pytorch
81a4aeabdf9d550ceda52a5060f19568de61b265
https://github.com/pytorch/pytorch/issues/93667
triaged tracker oncall: pt2 module: dynamo
14k github models on PyTorch 2.0 pass rates dashboard
We are weekly running dynamo-eager, dynamo-eager-fullgraph, export and inductor on 14k ```nn.Modules``` crawled from 1.4k github projects to get coverage level, find and fix bugs. This dashboard page is used to track the pass rates of different backends. How to repro: * Checkout https://github.com/jansel/pytorch-jit-paritybench * Batch evaluation with different backends: * dynamo-eager: ```python main.py --backend eager``` * dynamo-eager-fullgraph: ```python main.py --backend eager --fullgraph``` * export: ```python main.py --compile_mode export``` * inductor: ```python main.py``` * Adhoc evaluation: * ```pytest ./generated/{filename}.py -k test_{n}``` (e.g, ```pytest ./generated/test_KunpengLi1994_VSRN.py -k test_002```) * ```python -e ./generated/{filename}.py --backend eager``` Bugs umbrella task(#92670) cc @ezyang @msaroufim @bdhirsh @anijain2305 @zou3519 @voznesenskym @penguinwu @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @wenzhe-nrv @jiayisunx @chenyang78 @aakhundov @kadeng @soumith @wconstab @ngimel @Xia-Weiwen @desertfire @davidberard98
null
null
https://github.com/pytorch/pytorch/commit/81a4aeabdf9d550ceda52a5060f19568de61b265
{'base_commit': '81a4aeabdf9d550ceda52a5060f19568de61b265', 'files': [{'path': 'test/dynamo/test_misc.py', 'status': 'modified', 'Loc': {"('MiscTests', None, 40)": {'add': [2965]}, "('MiscTests', 'fn', 421)": {'mod': [422]}, "('MiscTests', 'test_numel', 420)": {'mod': [425]}}}, {'path': 'torch/_dynamo/variables/tensor.py', 'status': 'modified', 'Loc': {"('TensorVariable', 'call_method', 178)": {'mod': [206]}}}, {'path': 'torch/_dynamo/variables/torch.py', 'status': 'modified', 'Loc': {"('TorchVariable', 'can_constant_fold_through', 159)": {'add': [165]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "commit", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "torch/_dynamo/variables/torch.py", "torch/_dynamo/variables/tensor.py" ], "doc": [], "test": [ "test/dynamo/test_misc.py" ], "config": [], "asset": [] }
null
pytorch
pytorch
aac9e5288f7a9666884705e2b716c260cb5f9afc
https://github.com/pytorch/pytorch/issues/67002
module: windows module: multiprocessing triaged skipped
DISABLED test_fs_sharing (__main__.TestMultiprocessing)
Flaky failures in the last week: https://fburl.com/scuba/opensource_ci_jobs/inmj698k. They only appear to be on windows Platforms: rocm cc @peterjc123 @mszhanyi @skyline75489 @nbcsm @VitalyFedyunin
null
null
https://github.com/pytorch/pytorch/commit/aac9e5288f7a9666884705e2b716c260cb5f9afc
{'base_commit': 'aac9e5288f7a9666884705e2b716c260cb5f9afc', 'files': [{'path': 'test/test_multiprocessing.py', 'status': 'modified', 'Loc': {"('TestMultiprocessing', 'test_receive', 289)": {'add': [293]}, '(None, None, None)': {'mod': [19, 27]}, "('TestMultiprocessing', 'test_fill', 259)": {'mod': [269, 270]}, "('TestMultiprocessing', None, 251)": {'mod': [361]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "2", "loc_way": "commit", "loc_scope": "0", "info_type": "Code" }
{ "code": [], "doc": [], "test": [ "test/test_multiprocessing.py" ], "config": [], "asset": [] }
null
pytorch
pytorch
e37a22128eca7ccac6e289659587a9e1bfe6d242
https://github.com/pytorch/pytorch/issues/15052
oncall: jit
Tracer doesn't work with join/wait
Reported error: `RuntimeError: output 0 of traced region did not have observable data dependence with trace inputs; this probably indicates your program cannot be understood by the tracer.` To reproduce: ```python def test_async_script_trace(self): class Module(torch.jit.ScriptModule): def __init__(self): super(Module, self).__init__(False) @torch.jit.script_method def forward(self, x): future = torch.jit._fork(torch.neg, x) outputs = [] outputs.append(torch.jit._wait(future)) return outputs class Tuple(nn.Module): def __init__(self): super(Tuple, self).__init__() self.module = Module() def forward(self, x): return tuple(self.module(x)) x = torch.rand(3, 4) module = torch.jit.trace(Tuple(), (x), _force_outplace=True) self.assertEqual(module(x), torch.neg(x))
null
null
https://github.com/pytorch/pytorch/commit/e37a22128eca7ccac6e289659587a9e1bfe6d242
{'base_commit': 'e37a22128eca7ccac6e289659587a9e1bfe6d242', 'files': [{'path': 'aten/src/ATen/core/jit_type.h', 'status': 'modified', 'Loc': {"(None, 'FutureType', 516)": {'add': [534]}}}, {'path': 'test/test_jit.py', 'status': 'modified', 'Loc': {"('TestAsync', None, 11055)": {'add': [11224]}}}, {'path': 'torch/csrc/jit/graph_executor.cpp', 'status': 'modified', 'Loc': {'(None, None, 505)': {'mod': [516, 530, 531, 532, 533]}}}, {'path': 'torch/csrc/jit/tracer.cpp', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [39]}}}, {'path': 'torch/csrc/jit/tracer.h', 'status': 'modified', 'Loc': {"(None, 'function', 42)": {'add': [44]}, "(None, 'tracer', 24)": {'mod': [35, 36, 37, 38]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "commit", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "torch/csrc/jit/tracer.h", "torch/csrc/jit/tracer.cpp", "aten/src/ATen/core/jit_type.h", "torch/csrc/jit/graph_executor.cpp" ], "doc": [], "test": [ "test/test_jit.py" ], "config": [], "asset": [] }
null
pytorch
pytorch
0c091380cc03b23e68dde7368f3b910c21deb010
https://github.com/pytorch/pytorch/issues/21680
high priority module: cudnn module: nn triaged small
Disable nondeterministic CTCLoss from cuDNN
## 🐛 Bug <!-- A clear and concise description of what the bug is. --> ## To Reproduce Steps to reproduce the behavior: 1.I i updated pytorch version and ctc,use pytorch_nightly, but in my train ,nn.CTCloss() is still zero,so,i would like to ask if the version pytorch(nightly) has been solved this problem 1. 1. <!-- If you have a code sample, error messages, stack traces, please provide it here as well --> ## Expected behavior <!-- A clear and concise description of what you expected to happen. --> ## Environment Please copy and paste the output from our [environment collection script](https://raw.githubusercontent.com/pytorch/pytorch/master/torch/utils/collect_env.py) (or fill out the checklist below manually). You can get the script and run it with: ``` wget https://raw.githubusercontent.com/pytorch/pytorch/master/torch/utils/collect_env.py # For security purposes, please check the contents of collect_env.py before running it. python collect_env.py ``` - PyTorch Version (e.g., 1.0): - OS (e.g., Linux): - How you installed PyTorch (`conda`, `pip`, source): - Build command you used (if compiling from source): - Python version: - CUDA/cuDNN version: - GPU models and configuration: - Any other relevant information: ## Additional context <!-- Add any other context about the problem here. -->
null
null
https://github.com/pytorch/pytorch/commit/0c091380cc03b23e68dde7368f3b910c21deb010
{'base_commit': '0c091380cc03b23e68dde7368f3b910c21deb010', 'files': [{'path': 'aten/src/ATen/native/LossCTC.cpp', 'status': 'modified', 'Loc': {"(None, 'ctc_loss', 341)": {'mod': [367]}}}]}
[]
[]
[]
{ "iss_type": "2", "iss_reason": "2", "loc_way": "commit", "loc_scope": "0", "info_type": "Code\nC" }
{ "code": [ "aten/src/ATen/native/LossCTC.cpp" ], "doc": [], "test": [], "config": [], "asset": [] }
null
xtekky
gpt4free
5d08c7201fa5b4641f4277bf248c944b2c297b94
https://github.com/xtekky/gpt4free/issues/843
bug
permission denied
**Bug description** I'm running basic G4F code from README: ``` import g4f response = g4f.ChatCompletion.create( model=g4f.models.gpt_4, messages=[{"role": "user", "content": "hi"}], ) # alterative model setting print(response) ``` Error: ``` Traceback (most recent call last): File "C:\Users\IonE\Desktop\main.py", line 3, in <module> import g4f File "D:\Program Files\Python399\lib\site-packages\g4f\__init__.py", line 1, in <module> from . import models File "D:\Program Files\Python399\lib\site-packages\g4f\models.py", line 3, in <module> from .Provider import Bard, BaseProvider, GetGpt, H2o, Liaobots, Vercel, Equing File "D:\Program Files\Python399\lib\site-packages\g4f\Provider\__init__.py", line 6, in <module> from .Bard import Bard File "D:\Program Files\Python399\lib\site-packages\g4f\Provider\Bard.py", line 11, in <module> class Bard(AsyncProvider): File "D:\Program Files\Python399\lib\site-packages\g4f\Provider\Bard.py", line 22, in Bard cookies: dict = get_cookies(".google.com"), File "D:\Program Files\Python399\lib\site-packages\g4f\Provider\base_provider.py", line 45, in get_cookies for cookie in browser_cookie3.load(cookie_domain): File "D:\Program Files\Python399\lib\site-packages\browser_cookie3\__init__.py", line 1233, in load for cookie in cookie_fn(domain_name=domain_name): File "D:\Program Files\Python399\lib\site-packages\browser_cookie3\__init__.py", line 1160, in chrome return Chrome(cookie_file, domain_name, key_file).load() File "D:\Program Files\Python399\lib\site-packages\browser_cookie3\__init__.py", line 489, in load with _DatabaseConnetion(self.cookie_file) as con: File "D:\Program Files\Python399\lib\site-packages\browser_cookie3\__init__.py", line 349, in __enter__ return self.get_connection() File "D:\Program Files\Python399\lib\site-packages\browser_cookie3\__init__.py", line 383, in get_connection con = method() File "D:\Program Files\Python399\lib\site-packages\browser_cookie3\__init__.py", line 374, in __get_connection_legacy shutil.copyfile(self.__database_file, self.__temp_cookie_file) File "D:\Program Files\Python399\lib\shutil.py", line 264, in copyfile with open(src, 'rb') as fsrc: PermissionError: [Errno 13] Permission denied: 'C:\\Users\\IonE\\AppData\\Roaming\\..\\Local\\Google\\Chrome\\User Data\\Default\\Network\\Cookies' ``` **Environement** - python 3.9.9 - ukraine
null
https://github.com/xtekky/gpt4free/pull/847
null
{'base_commit': '5d08c7201fa5b4641f4277bf248c944b2c297b94', 'files': [{'path': 'g4f/Provider/Bard.py', 'status': 'modified', 'Loc': {"('Bard', 'create_async', 17)": {'add': [33], 'mod': [22]}}}, {'path': 'g4f/Provider/Bing.py', 'status': 'modified', 'Loc': {"('Bing', 'create_async_generator', 21)": {'add': [34], 'mod': [24]}}}, {'path': 'g4f/Provider/Hugchat.py', 'status': 'modified', 'Loc': {"('Hugchat', 'create_completion', 17)": {'add': [25], 'mod': [23]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "2", "loc_way": "pr", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "g4f/Provider/Bing.py", "g4f/Provider/Hugchat.py", "g4f/Provider/Bard.py" ], "doc": [], "test": [], "config": [], "asset": [] }
null
xtekky
gpt4free
3430b04f870d982d7fba34e3b9d6e5cf3bd3b847
https://github.com/xtekky/gpt4free/issues/1003
bug
please delete site chat.aivvm.com
**Known Issues** // delete this please delete site `chat.aivvm.com` **Delete site description** Gpt4free maintainers, I am the administrator of chat.aivvm.com and request to delete this site. My website is already under high load
null
null
https://github.com/xtekky/gpt4free/commit/3430b04f870d982d7fba34e3b9d6e5cf3bd3b847
{'base_commit': '3430b04f870d982d7fba34e3b9d6e5cf3bd3b847', 'files': [{'path': 'README.md', 'status': 'modified', 'Loc': {'(None, None, 218)': {'mod': [218]}, '(None, None, 281)': {'mod': [281]}, '(None, None, 374)': {'mod': [374]}}}, {'path': 'etc/testing/test_chat_completion.py', 'status': 'modified', 'Loc': {"(None, 'run_async', 19)": {'mod': [22]}}}, {'path': 'g4f/Provider/__init__.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [9]}}}, {'path': 'g4f/Provider/Aivvm.py', 'status': 'renamed', 'Loc': {'(None, None, None)': {'mod': [3, 4, 5, 21, 22]}}}, {'path': 'g4f/Provider/deprecated/__init__.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [14]}}}, {'path': 'g4f/models.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [23, 47, 56, 66, 170, 171, 172, 173, 178, 179, 183, 184, 188, 189]}}}]}
[]
[]
[]
{ "iss_type": "3\n用户请求项目删除自己网址", "iss_reason": "2", "loc_way": "commit", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "g4f/Provider/Aivvm.py", "g4f/Provider/deprecated/__init__.py", "g4f/Provider/__init__.py", "g4f/models.py" ], "doc": [ "README.md" ], "test": [ "etc/testing/test_chat_completion.py" ], "config": [], "asset": [] }
null
Z4nzu
hackingtool
0a4faeac9c4f93a61c937b0e57023b693beeca6f
https://github.com/Z4nzu/hackingtool/issues/188
fixed
can not run hackingtool !!!!! SyntaxError: invalid syntax ???
# python3 ./hackingtool.py Traceback (most recent call last): File "/root/hackingtool/./hackingtool.py", line 11, in <module> from tools.ddos import DDOSTools File "/root/hackingtool/tools/ddos.py", line 29 "sudo", "python3 ddos", method, url, socks_type5.4.1, threads, proxylist, multiple, timer]) ^ SyntaxError: invalid syntax __________________________________________________________________________________
null
https://github.com/Z4nzu/hackingtool/pull/176
null
{'base_commit': '0a4faeac9c4f93a61c937b0e57023b693beeca6f', 'files': [{'path': 'tools/ddos.py', 'status': 'modified', 'Loc': {"('ddos', 'run', 20)": {'mod': [29]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "pr", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "tools/ddos.py" ], "doc": [], "test": [], "config": [], "asset": [] }
null
Z4nzu
hackingtool
0a4faeac9c4f93a61c937b0e57023b693beeca6f
https://github.com/Z4nzu/hackingtool/issues/177
Help me please
Traceback (most recent call last): File "/usr/share/doc/hackingtool/hackingtool.py", line 11, in <module> from tools.ddos import DDOSTools File "/usr/share/doc/hackingtool/tools/ddos.py", line 29 "sudo", "python3 ddos", method, url, socks_type5.4.1, threads, proxylist, multiple, timer]) ^ SyntaxError: invalid syntax
null
https://github.com/Z4nzu/hackingtool/pull/176
null
{'base_commit': '0a4faeac9c4f93a61c937b0e57023b693beeca6f', 'files': [{'path': 'tools/ddos.py', 'status': 'modified', 'Loc': {"('ddos', 'run', 20)": {'mod': [29]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "pr", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "tools/ddos.py" ], "doc": [], "test": [], "config": [], "asset": [] }
null
Z4nzu
hackingtool
0a4faeac9c4f93a61c937b0e57023b693beeca6f
https://github.com/Z4nzu/hackingtool/issues/189
look
![image](https://user-images.githubusercontent.com/93758292/155468650-b9d57e21-6c82-4005-a3ee-1783699e7f11.png)
null
https://github.com/Z4nzu/hackingtool/pull/176
null
{'base_commit': '0a4faeac9c4f93a61c937b0e57023b693beeca6f', 'files': [{'path': 'tools/ddos.py', 'status': 'modified', 'Loc': {"('ddos', 'run', 20)": {'mod': [29]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "pr", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "tools/ddos.py" ], "doc": [], "test": [], "config": [], "asset": [] }
null
Z4nzu
hackingtool
0a4faeac9c4f93a61c937b0e57023b693beeca6f
https://github.com/Z4nzu/hackingtool/issues/209
Running Issue
**Describe the bug** A clear and concise description of what the bug is. I have installed this tool successfully though when i go to run it with <sudo hackingtool> or <hackingtool> it comes up with the error I have attached as a screenshot. Why would this be happening? **To Reproduce** Steps to reproduce the behavior: 1. Go to '...' 2. Click on '....' 3. Scroll down to '....' 4. See error **Expected behavior** A clear and concise description of what you expected to happen. **Screenshots** If applicable, add screenshots to help explain your problem. ![image](https://user-images.githubusercontent.com/13176339/161368818-26bf3219-9ba7-4bab-a451-65d379c6d405.jpeg) **Desktop (please complete the following information):** - OS: Kali - Browser [e.g. chrome, safari] - Version [e.g. 22] **Smartphone (please complete the following information):** - Device: rpi4 - OS: [e.g. iOS8.1] - Browser [e.g. stock browser, safari] - Version [e.g. 22] **Additional context** Add any other context about the problem here.
null
https://github.com/Z4nzu/hackingtool/pull/176
null
{'base_commit': '0a4faeac9c4f93a61c937b0e57023b693beeca6f', 'files': [{'path': 'tools/ddos.py', 'status': 'modified', 'Loc': {"('ddos', 'run', 20)": {'mod': [29]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "pr", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "tools/ddos.py" ], "doc": [], "test": [], "config": [], "asset": [] }
null
Z4nzu
hackingtool
0a4faeac9c4f93a61c937b0e57023b693beeca6f
https://github.com/Z4nzu/hackingtool/issues/185
fixed
SyntaxError: invalid syntax traceback most recent call last
┌──(root💀localhost)-[~] └─# hackingtool Traceback (most recent call last): File "/usr/share/doc/hackingtool/hackingtool.py", line 11, in <module> from tools.ddos import DDOSTools File "/usr/share/doc/hackingtool/tools/ddos.py", line 29 "sudo", "python3 ddos", method, url, socks_type5.4.1, threads, proxylist, multiple, timer]) ^ SyntaxError: invalid syntax this happens when i type in "hackingtool" in terminal. any fixes?
null
https://github.com/Z4nzu/hackingtool/pull/176
null
{'base_commit': '0a4faeac9c4f93a61c937b0e57023b693beeca6f', 'files': [{'path': 'tools/ddos.py', 'status': 'modified', 'Loc': {"('ddos', 'run', 20)": {'mod': [29]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "pr", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "tools/ddos.py" ], "doc": [], "test": [], "config": [], "asset": [] }
null
Z4nzu
hackingtool
0a4faeac9c4f93a61c937b0e57023b693beeca6f
https://github.com/Z4nzu/hackingtool/issues/187
fixed
from tools.ddos import DDOSTools
# sudo hackingtool Traceback (most recent call last): File "/usr/share/doc/hackingtool/hackingtool.py", line 11, in <module> from tools.ddos import DDOSTools File "/usr/share/doc/hackingtool/tools/ddos.py", line 29 "sudo", "python3 ddos", method, url, socks_type5.4.1, threads, proxylist, multiple, timer]) ^ SyntaxError: invalid syntax
null
https://github.com/Z4nzu/hackingtool/pull/176
null
{'base_commit': '0a4faeac9c4f93a61c937b0e57023b693beeca6f', 'files': [{'path': 'tools/ddos.py', 'status': 'modified', 'Loc': {"('ddos', 'run', 20)": {'mod': [29]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "pr", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "tools/ddos.py" ], "doc": [], "test": [], "config": [], "asset": [] }
null
scikit-learn
scikit-learn
bf0886bae0ccbc8c5d285b6e2affe7e40474f970
https://github.com/scikit-learn/scikit-learn/issues/16924
Bug Easy module:metrics
Matthews correlation coefficient metric throws misleading division by zero RuntimeWarning
#### Description With tested values all equal, `sklearn.metrics.matthews_corrcoef` throws a `RuntimeWarning` reporting a division by zero. This behavior was already reported in #1937 and reported fixed, but reappears in recent versions. #### Steps/Code to Reproduce The snippet below reproduces the warning. ```python import sklearn.metrics trues = [1,0,1,1,0] preds = [0,0,0,0,0] sklearn.metrics.matthews_corrcoef(trues, preds) ``` #### Expected Results No warning is thrown. #### Actual Results The following warning is thrown: ``` C:\anaconda\envs\sklearn-test\lib\site-packages\sklearn\metrics\_classification.py:900: RuntimeWarning: invalid value encountered in double_scalars mcc = cov_ytyp / np.sqrt(cov_ytyt * cov_ypyp) ``` #### Versions ``` System: python: 3.8.2 (default, Mar 25 2020, 08:56:29) [MSC v.1916 64 bit (AMD64)] executable: C:\anaconda\envs\sklearn-test\python.exe machine: Windows-10-10.0.18362-SP0 Python dependencies: pip: 20.0.2 setuptools: 46.1.3.post20200330 sklearn: 0.22.1 numpy: 1.18.1 scipy: 1.4.1 Cython: None pandas: None matplotlib: None joblib: 0.14.1 ```
null
https://github.com/scikit-learn/scikit-learn/pull/19977
null
{'base_commit': 'bf0886bae0ccbc8c5d285b6e2affe7e40474f970', 'files': [{'path': 'sklearn/metrics/_classification.py', 'status': 'modified', 'Loc': {"(None, 'matthews_corrcoef', 800)": {'mod': [881, 883, 886]}}}, {'path': 'sklearn/metrics/tests/test_classification.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [23, 625]}, "(None, 'test_matthews_corrcoef', 671)": {'mod': [687, 688, 690, 691, 694, 696, 697]}, "(None, 'test_matthews_corrcoef_multiclass', 713)": {'mod': [734, 737, 738, 739, 757, 761, 762, 763, 765, 766]}}}, {'path': 'sklearn/utils/_testing.py', 'status': 'modified', 'Loc': {"(None, 'assert_warns_div0', 190)": {'mod': [190, 191, 193, 194, 196, 197, 198, 199, 200, 201, 203, 204, 205, 206, 207, 208, 209, 210, 211]}}}]}
[]
[]
[]
{ "iss_type": "2", "iss_reason": "2", "loc_way": "pr", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "sklearn/metrics/_classification.py", "sklearn/utils/_testing.py" ], "doc": [], "test": [ "sklearn/metrics/tests/test_classification.py" ], "config": [], "asset": [] }
null
scikit-learn
scikit-learn
4ac6a90a82e4a8d7b5338c18ae8a16559c98ba10
https://github.com/scikit-learn/scikit-learn/issues/5101
LatentDirichletAllocation has superfluous attributes
It has `dirichlet_component_` (undocumented) and `exp_dirichlet_component_` (exponential of same). I propose to get rid of at least the latter.
null
https://github.com/scikit-learn/scikit-learn/pull/5111
null
{'base_commit': '4ac6a90a82e4a8d7b5338c18ae8a16559c98ba10', 'files': [{'path': 'sklearn/decomposition/online_lda.py', 'status': 'modified', 'Loc': {"('LatentDirichletAllocation', '_approx_bound', 542)": {'add': [579], 'mod': [597, 612]}, "('LatentDirichletAllocation', '_init_latent_vars', 283)": {'mod': [305, 306, 308]}, "('LatentDirichletAllocation', '_em_step', 366)": {'mod': [407, 408]}}}]}
[]
[]
[]
{ "iss_type": "1\n多余的选项", "iss_reason": "2", "loc_way": "pr", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "sklearn/decomposition/online_lda.py" ], "doc": [], "test": [], "config": [], "asset": [] }
null
pandas-dev
pandas
05123af1b2f8db1bc4f05c22515ef378cbeefbd3
https://github.com/pandas-dev/pandas/issues/76
Bug
Sparse cumsum functions do not work
e.g. SparseSeries.cumsum
null
null
https://github.com/wesm/pandas/commit/05123af1b2f8db1bc4f05c22515ef378cbeefbd3
{'base_commit': '05123af1b2f8db1bc4f05c22515ef378cbeefbd3', 'files': [{'path': 'pandas/core/frame.py', 'status': 'modified', 'Loc': {"('DataFrame', None, 97)": {'mod': [1962, 1963, 1964, 1966, 1967, 1968, 1969, 1971, 1972, 1973, 1974, 1975, 1976, 1977, 1978, 1979, 1980, 1981, 1982, 1983, 1984, 1985, 2021, 2022, 2023, 2025, 2026, 2027, 2028, 2030, 2031, 2032, 2033, 2034, 2035, 2036, 2037, 2038, 2039, 2041, 2043]}}}, {'path': 'pandas/core/generic.py', 'status': 'modified', 'Loc': {"('PandasGeneric', '_reindex_axis', 162)": {'add': [168]}}}, {'path': 'pandas/core/series.py', 'status': 'modified', 'Loc': {"('Series', 'cumsum', 570)": {'mod': [580, 581, 582, 583, 584, 585, 591, 592, 593, 594]}}}, {'path': 'pandas/core/sparse.py', 'status': 'modified', 'Loc': {"('SparseSeries', None, 152)": {'add': [512]}, "('SparseDataFrame', 'count', 1058)": {'add': [1059]}, '(None, None, None)': {'mod': [13]}}}, {'path': 'pandas/tests/test_frame.py', 'status': 'modified', 'Loc': {"('TestDataFrame', None, 539)": {'add': [2271]}, "('TestDataFrame', 'test_cumsum', 2271)": {'add': [2276, 2283, 2284], 'mod': [2273, 2274, 2286, 2287]}}}, {'path': 'pandas/tests/test_sparse.py', 'status': 'modified', 'Loc': {"('TestSparseSeries', None, 111)": {'add': [577]}, "('TestSparseDataFrame', 'test_count', 1065)": {'add': [1068]}}}]}
[]
[]
[]
{ "iss_type": "2", "iss_reason": "1", "loc_way": "commit", "loc_scope": "0", "info_type": "Code" }
{ "code": [ "pandas/core/frame.py", "pandas/core/generic.py", "pandas/core/series.py", "pandas/core/sparse.py" ], "doc": [], "test": [ "pandas/tests/test_frame.py", "pandas/tests/test_sparse.py" ], "config": [], "asset": [] }
null
pandas-dev
pandas
65c0441a41b2dcaeebb648274d30978419a8661a
https://github.com/pandas-dev/pandas/issues/16607
Datetime Compat
to_datetime should support ISO week year
`to_datetime` does not currently seem to support `ISO week year` like `strptime` does: ``` In [38]: datetime.date(2016, 1, 1).strftime('%G-%V') Out[38]: '2015-53' In [39]: datetime.datetime.strptime(datetime.date(2016, 1, 1).strftime('%G-%V')+'-1', '%G-%V-%u') Out[39]: datetime.datetime(2015, 12, 28, 0, 0) In [41]: pd.to_datetime(datetime.date(2016, 1, 1).strftime('%G-%V')+'-1', format='%G-%V-%u') --------------------------------------------------------------------------- TypeError Traceback (most recent call last) /Users/Robin/.pyenv/versions/3.6.1/lib/python3.6/site-packages/pandas/core/tools/datetimes.py in _convert_listlike(arg, box, format, name, tz) 443 try: --> 444 values, tz = tslib.datetime_to_datetime64(arg) 445 return DatetimeIndex._simple_new(values, name=name, tz=tz) pandas/_libs/tslib.pyx in pandas._libs.tslib.datetime_to_datetime64 (pandas/_libs/tslib.c:33275)() TypeError: Unrecognized value type: <class 'str'> During handling of the above exception, another exception occurred: ValueError Traceback (most recent call last) <ipython-input-41-7ce30c959690> in <module>() ----> 1 pd.to_datetime(datetime.date(2016, 1, 1).strftime('%G-%V')+'-1', format='%G-%V-%u') /Users/Robin/.pyenv/versions/3.6.1/lib/python3.6/site-packages/pandas/core/tools/datetimes.py in to_datetime(arg, errors, dayfirst, yearfirst, utc, box, format, exact, unit, infer_datetime_format, origin) 516 result = _convert_listlike(arg, box, format) 517 else: --> 518 result = _convert_listlike(np.array([arg]), box, format)[0] 519 520 return result /Users/Robin/.pyenv/versions/3.6.1/lib/python3.6/site-packages/pandas/core/tools/datetimes.py in _convert_listlike(arg, box, format, name, tz) 445 return DatetimeIndex._simple_new(values, name=name, tz=tz) 446 except (ValueError, TypeError): --> 447 raise e 448 449 if arg is None: /Users/Robin/.pyenv/versions/3.6.1/lib/python3.6/site-packages/pandas/core/tools/datetimes.py in _convert_listlike(arg, box, format, name, tz) 412 try: 413 result = tslib.array_strptime(arg, format, exact=exact, --> 414 errors=errors) 415 except tslib.OutOfBoundsDatetime: 416 if errors == 'raise': pandas/_libs/tslib.pyx in pandas._libs.tslib.array_strptime (pandas/_libs/tslib.c:63124)() pandas/_libs/tslib.pyx in pandas._libs.tslib.array_strptime (pandas/_libs/tslib.c:63003)() ValueError: 'G' is a bad directive in format '%G-%V-%u' ``` <details> INSTALLED VERSIONS ------------------ commit: None pandas: 0.20.1 pytest: 3.1.0 pip: 9.0.1 setuptools: 28.8.0 Cython: 0.25.2 numpy: 1.12.1 scipy: 0.19.0 xarray: None IPython: 6.0.0 sphinx: None patsy: 0.4.1 dateutil: 2.6.0 pytz: 2017.2 blosc: None bottleneck: None tables: 3.4.2 numexpr: 2.6.2 feather: None matplotlib: 2.0.2 openpyxl: None xlrd: None xlwt: None xlsxwriter: None lxml: None bs4: None html5lib: 0.999999999 sqlalchemy: 1.1.10 pymysql: None psycopg2: 2.7.1 (dt dec pq3 ext lo64) jinja2: 2.9.6 s3fs: None pandas_gbq: None pandas_datareader: None </details>
null
https://github.com/pandas-dev/pandas/pull/25541
null
{'base_commit': '65c0441a41b2dcaeebb648274d30978419a8661a', 'files': [{'path': 'doc/source/whatsnew/v0.25.0.rst', 'status': 'modified', 'Loc': {'(None, None, 21)': {'add': [21]}}}, {'path': 'pandas/_libs/tslibs/strptime.pyx', 'status': 'modified', 'Loc': {'(None, None, 79)': {'add': [79]}, '(None, None, 171)': {'add': [171]}, '(None, None, 267)': {'add': [267]}, '(None, None, 513)': {'add': [513]}, '(None, None, 520)': {'add': [520]}, '(None, None, 521)': {'add': [521]}, '(None, None, 622)': {'add': [622]}, '(None, None, 57)': {'mod': [57]}, '(None, None, 178)': {'mod': [178]}, '(None, None, 271)': {'mod': [271, 272, 273, 274]}, '(None, None, 596)': {'mod': [596, 597]}, '(None, None, 600)': {'mod': [600]}}}, {'path': 'pandas/core/tools/datetimes.py', 'status': 'modified', 'Loc': {"(None, 'to_datetime', 403)": {'add': [457]}}}, {'path': 'pandas/tests/indexes/datetimes/test_tools.py', 'status': 'modified', 'Loc': {"('TestToDatetime', None, 246)": {'add': [246]}}}]}
[]
[]
[]
{ "iss_type": "4", "iss_reason": "2", "loc_way": "pr", "loc_scope": "0", "info_type": "Code/Doc" }
{ "code": [ "pandas/core/tools/datetimes.py", "pandas/_libs/tslibs/strptime.pyx" ], "doc": [ "doc/source/whatsnew/v0.25.0.rst" ], "test": [ "pandas/tests/indexes/datetimes/test_tools.py" ], "config": [], "asset": [] }
null
pallets
flask
6ed68f015a50ab35b84a8ea71b0f846ca6a75281
https://github.com/pallets/flask/issues/3074
send_file doesn't urlencode ':/' in unicode attachment_filename
### Expected Behavior When sending files with unicode filename (with `:` or `/`) they should be downloaded with name from `filename*` field. ```python # -*- coding: utf-8 -*- import os from flask import Flask, send_from_directory app = Flask(__name__) @app.route('/test/', methods=['GET']) def test_route(): tmp_dir = os.getcwd() tmp_filename = __file__ attachment_filename = u'тест:тест_тест.py' return send_from_directory( tmp_dir, tmp_filename, as_attachment=True, attachment_filename=attachment_filename ) if __name__ == '__main__': app.run(host='::', port=5000) ``` ### Actual Behavior Some browsers (Chrome-based/Safari) ignore `filename*` field when it contains colon or slash. For example file `тест:тест_тест.py` gets downloaded in Chrome/Safari as `__.py` but in Firefox as `тест_тест_тест.py` which is acceptable in my opinion. Flask response: `Content-Disposition: attachment; filename*="UTF-8''%D1%82%D0%B5%D1%81%D1%82:%D1%82%D0%B5%D1%81%D1%82_%D1%82%D0%B5%D1%81%D1%82.py"; filename=":_.py"` ### Environment * Python version: 2.7.15 * Flask version: 1.0.2 * Werkzeug version: 0.14.1
null
https://github.com/pallets/flask/pull/3273
null
{'base_commit': '6ed68f015a50ab35b84a8ea71b0f846ca6a75281', 'files': [{'path': 'CHANGES.rst', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [10]}}}, {'path': 'flask/helpers.py', 'status': 'modified', 'Loc': {"(None, 'send_file', 454)": {'mod': [579]}}}, {'path': 'tests/test_helpers.py', 'status': 'modified', 'Loc': {"('TestSendfile', None, 436)": {'add': [648]}}}]}
[]
[]
[]
{ "iss_type": "2", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "flask/helpers.py" ], "doc": [ "CHANGES.rst" ], "test": [ "tests/test_helpers.py" ], "config": [], "asset": [] }
1
pallets
flask
50b7dcbab343c93bb6738bbf116a177e72b1d9ec
https://github.com/pallets/flask/issues/4099
docs
Harmless race condition in tutorial
I was browsing the flaskr tutorial when I noticed an (admittedly quite unlikely) race condition in the `register` view, specifically: ```py if not username: error = 'Username is required.' elif not password: error = 'Password is required.' elif db.execute( 'SELECT id FROM user WHERE username = ?', (username,) ).fetchone() is not None: error = f"User {username} is already registered." if error is None: db.execute( 'INSERT INTO user (username, password) VALUES (?, ?)', (username, generate_password_hash(password)) ) db.commit() return redirect(url_for('auth.login')) ``` If two requests arrive with the right timing, the following can happen: ``` Request 1: Request 2: SELECT id FROM user WHERE username = abc | v empty, no such user SELECT id FROM user WHERE username = abc | v empty, no such user INSERT INTO user (username, password) VALUES (abc, 123) | v ok INSERT INTO user (username, password) VALUES (abc, 456) | v failed UNIQUE constraint -> -> sqlite3.IntegrityError -> -> user gets HTTP 500 ``` While the likelihood of this happening is pretty small and the harm practically zero (user gets HTTP 500 and has to manually login/choose a different username), I feel like this is not really the sort of good practice the tutorial should teach. I also believe it's important the developer understands that it's the UNIQUE constraint that ensures their app works correctly and not the if condition in the application code (the tutorial mentions SQL injection attacks and explains what protects the developer against them, so I don't really feel this is out of scope). In my own app I've modified the code to the following: ```py if not username: error = 'Username is required.' elif not password: error = 'Password is required.' else: try: db.execute( 'INSERT INTO users (username, password) VALUES (?, ?)', (username, generate_password_hash(password)) ) db.commit() except IntegrityError: error = f"User {username} is already registered." else: return redirect(url_for('auth.login')) ``` I suggest something similar be incorporated into the tutorial, with a short explanation (maybe a comment) of how the UNIQUE constraint does the work for the developer and maybe a note about the principle that one should "ask forgiveness, not permission." I'm not sure on how it's better worded, so I'm making this an issue instead of a pull request. Cheers, and thank you for your great work!
null
https://github.com/pallets/flask/pull/4139
null
{'base_commit': '50b7dcbab343c93bb6738bbf116a177e72b1d9ec', 'files': [{'path': 'docs/tutorial/views.rst', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [202], 'mod': [94, 95, 96, 97, 100, 101, 102, 103, 104, 105, 128, 129, 130, 131, 132, 133, 134, 136, 137, 138, 139, 141, 142, 143, 144, 145, 146, 147]}}}, {'path': 'examples/tutorial/flaskr/auth.py', 'status': 'modified', 'Loc': {"(None, 'register', 47)": {'mod': [63, 64, 65, 66, 67, 70, 71, 72, 73, 74, 75, 76, 77]}}}]}
[]
[]
[]
{ "iss_type": "4", "iss_reason": "2", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "examples/tutorial/flaskr/auth.py" ], "doc": [ "docs/tutorial/views.rst" ], "test": [], "config": [], "asset": [] }
1
pallets
flask
f17e6061fcffdc290f615d3fdc9d949e9e719574
https://github.com/pallets/flask/issues/1443
json_encoder not invoked from flask.jsonify
I created a custom JSON encoder class extended from flask.json.JSONEncoder but it is not called when calling flask.jsonify. Additionally, I removed my custom JSON encoder and confirmed that flask.json.JSONEncoder isn't called either via a break statement in Pycharm. ``` from flask import Flask from flask import jsonify from flask.json import JSONEncoder class MyEncoder(JSONEncoder): def default(self, obj): if hasattr(obj, '__json__'): return obj.__json__() else: try: iterable = iter(obj) except TypeError: pass else: return list(iterable) return JSONEncoder.default(self, obj) class MyClass(object): key = 'a' value = 'b' def __json__(self): return {'key': self.key, 'value': self.value} app = Flask(__name__) app.json_encoder = MyEncoder @app.route('/') def hello_world(): return jsonify(MyClass()) if __name__ == '__main__': app.run(debug=True) ```
null
https://github.com/pallets/flask/pull/1671
null
{'base_commit': 'f17e6061fcffdc290f615d3fdc9d949e9e719574', 'files': [{'path': 'AUTHORS', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [17, 20], 'mod': [35]}}}, {'path': 'CHANGES', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [10]}}}, {'path': 'docs/security.rst', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [98, 100, 101, 102, 103, 105, 106, 107, 108, 109, 111, 112, 113, 114, 115, 116, 117, 119, 120, 121, 122, 123, 125, 127, 128, 129, 130, 132, 133, 134, 135, 137, 138, 140, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 161, 162, 163, 164, 165, 166, 167, 168, 170, 171, 172, 173, 174, 175]}}}, {'path': 'flask/json.py', 'status': 'modified', 'Loc': {"(None, 'jsonify', 201)": {'add': [244], 'mod': [202, 203, 204, 205, 225, 226, 248, 249]}}}, {'path': 'tests/test_helpers.py', 'status': 'modified', 'Loc': {"('TestJSON', 'test_json_as_unicode', 121)": {'add': [122], 'mod': [124, 125, 126, 127, 129, 130, 131, 132]}, "('TestJSON', None, 32)": {'mod': [34, 35, 37, 38, 39, 40, 42, 43, 45, 46, 47, 48, 49, 50, 106, 107, 121]}}}]}
[]
[]
[]
{ "iss_type": "2", "iss_reason": "2", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "flask/json.py" ], "doc": [ "docs/security.rst", "CHANGES" ], "test": [ "tests/test_helpers.py" ], "config": [], "asset": [ "AUTHORS" ] }
1
pallets
flask
f808c20139649b747f604492bc33b61a7dd3e13a
https://github.com/pallets/flask/issues/2731
bug blueprints
Flask 1.0 backwards-incompat with double-slash/no-slash re. #2629
This is a major backwards-compat breaking change, but I suspect not the intended design and hopefully easy to fix. The issue is related to PR #2629, and this example follows from that: Given blueprint `bp` and app `app`: ```python @bp.route('b/') def tmp(): return "URI should be '/a/b/" app.register_blueprint(bp, url_prefix='/a/') ``` In Flask 0.12 the URL is correctly `/a/b`, but in Flask 1.0 it's `/ab`. Since issue #2629 relates to resolve double-slashes, I imagine this is a bug (and not a design decision) - and the correct solution would be to remove a slash only when there are two.
null
https://github.com/pallets/flask/pull/2629
null
{'base_commit': 'f808c20139649b747f604492bc33b61a7dd3e13a', 'files': [{'path': 'CHANGES.rst', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [145, 188]}}}, {'path': 'flask/blueprints.py', 'status': 'modified', 'Loc': {"('BlueprintSetupState', '__init__', 25)": {'add': [55]}}}, {'path': 'tests/test_blueprints.py', 'status': 'modified', 'Loc': {"(None, 'test_blueprint_url_definitions', 117)": {'mod': [117]}}}]}
[]
[]
[]
{ "iss_type": "2", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "flask/blueprints.py" ], "doc": [ "CHANGES.rst" ], "test": [ "tests/test_blueprints.py" ], "config": [], "asset": [] }
1
pallets
flask
22708b048d224a5590fa28d86ca02bac52294f90
https://github.com/pallets/flask/issues/2594
cli
add ssl_context option to `flask run`
### Expected Behaviour I expect to be able to pass the `flask run` command any of the options which are valid for the `Flask.run()` method: ```sh $ FLASK_APP=myapp/run.py FLASK_DEBUG=1 flask run --host=0.0.0.0 --ssl_context=adhoc * Running on https://0.0.0.0:5000/ (Press CTRL+C to quit) ``` Specifically, I want to pass `ssl_context=adhoc`, but it seems sensible to extend the command to accept all valid keyword arguments for `Flask.run()` / `werkzeug.serving.run_simple()`. ### Actual Behaviour ``` Error: no such option: --ssl_context flask run --host=0.0.0.0 --ssl_context=adhoc exited with code 2 ``` ### Environment * Python version: 3.5.2 * Flask version: 0.12.2 * Werkzeug version: 0.12.2
null
https://github.com/pallets/flask/pull/2606
null
{'base_commit': '22708b048d224a5590fa28d86ca02bac52294f90', 'files': [{'path': 'CHANGES', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [120, 156]}}}, {'path': 'flask/cli.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [16, 23, 601, 606], 'mod': [26, 608, 611, 614]}, "(None, 'run_command', 619)": {'mod': [620, 645]}}}, {'path': 'tests/test_cli.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [16, 17], 'mod': [27, 28]}, "(None, 'test_dotenv_optional', 462)": {'add': [466]}}}]}
[]
[]
[]
{ "iss_type": "4", "iss_reason": "2", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "flask/cli.py" ], "doc": [ "CHANGES" ], "test": [ "tests/test_cli.py" ], "config": [], "asset": [] }
1
pallets
flask
e4c712ffd2682f963906e1d0d27e67b7f83d95ce
https://github.com/pallets/flask/issues/266
docs good first issue blueprints
Blueprint template lookup not documented enough
The new blueprint template lookup scheme where the templates folder is just added to the searchpath instead of doing some weird stuff with the names as before. The documentation has to be clearer about that.
null
https://github.com/pallets/flask/pull/1843
null
{'base_commit': 'e4c712ffd2682f963906e1d0d27e67b7f83d95ce', 'files': [{'path': 'docs/blueprints.rst', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [179, 180, 181, 182, 183, 188]}}}]}
[]
[]
[]
{ "iss_type": "4", "iss_reason": "2", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [], "doc": [ "docs/blueprints.rst" ], "test": [], "config": [], "asset": [] }
1
pallets
flask
8cd0b03beeac4a41c398ea365475c651c484a9ee
https://github.com/pallets/flask/issues/2118
bug
config.from_pyfile crashes on Python 3 when source isn't encoded in default encoding
when I read my instance config file, I get an error. > exec(compile(config_file.read(), filename, 'exec'), d.__dict__) > UnicodeDecodeError: 'gbk' codec can't decode byte 0x80 in position 437: illegal multibyte sequence Then I modify the code of config.from_pyfile to this > with open(filename, 'rb') as config_file: The problem is resolved.
null
https://github.com/pallets/flask/pull/2123
null
{'base_commit': '8cd0b03beeac4a41c398ea365475c651c484a9ee', 'files': [{'path': 'CHANGES', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [10]}}}, {'path': 'flask/config.py', 'status': 'modified', 'Loc': {"('Config', 'from_pyfile', 111)": {'mod': [129]}}}, {'path': 'tests/test_config.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [13, 14], 'mod': [10, 12]}, "(None, 'test_get_namespace', 168)": {'add': [189]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "flask/config.py" ], "doc": [ "CHANGES" ], "test": [ "tests/test_config.py" ], "config": [], "asset": [] }
1
pallets
flask
85fa8aabf5a7bd0adf204f0c2dacbba1fa6683de
https://github.com/pallets/flask/issues/2023
discussion logging
How should logging in Flask look like?
Flask started to ship with a default, hardcoded logging handler. Unfortunately this setup makes it harder to install custom logging setups, because then you'll have to undo all the things Flask did to the app logger, or replace the `app.logger` entirely. A symptom of this is #1993, where Flask's own logger had to be tweaked yet again such that messages didn't get logged twice (once via Flask's setup, once via the custom one). My question is: **Do we even want Flask to do any logging setup?** It appears that this sort of default logging is only useful during development, so maybe it makes sense to set up a default logging handler in the new Flask CLI instead of from within the application.
null
https://github.com/pallets/flask/pull/2436
null
{'base_commit': '85fa8aabf5a7bd0adf204f0c2dacbba1fa6683de', 'files': [{'path': 'CHANGES', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [108]}}}, {'path': 'docs/config.rst', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [331], 'mod': [202, 204, 205, 207, 209, 211, 212, 213, 215]}}}, {'path': 'docs/contents.rst.inc', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [18]}}}, {'path': 'docs/errorhandling.rst', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [146, 147, 149, 150, 151, 152, 153, 155, 156, 157, 158, 159, 161, 162, 163, 165, 166, 167, 168, 169, 170, 171, 172, 173, 175, 176, 177, 178, 179, 180, 181, 183, 184, 185, 187, 188, 189, 192, 193, 195, 196, 197, 198, 199, 200, 202, 203, 204, 206, 207, 208, 209, 210, 211, 212, 213, 214, 216, 217, 218, 220, 221, 222, 223, 224, 225, 227, 229, 230, 232, 233, 234, 235, 236, 238, 239, 240, 242, 244, 245, 247, 249, 250, 251, 252, 253, 254, 255, 257, 259, 260, 262, 263, 265, 267, 268, 269, 270, 271, 274, 275, 277, 278, 279, 281, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 314, 315, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 329, 332, 333, 335, 336, 337, 338, 339, 340, 341, 343, 344, 345, 347, 348, 349, 350, 351, 352]}}}, {'path': 'flask/app.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [32], 'mod': [19, 20, 21, 22, 40, 41]}, "('Flask', None, 70)": {'mod': [267, 268, 269, 270, 271, 297, 298, 616]}, "('Flask', '__init__', 352)": {'mod': [395, 396, 397]}, "('Flask', 'logger', 617)": {'mod': [618, 619, 620, 621, 623, 624, 625, 629, 630, 631, 632, 633, 634, 635, 636]}}}, {'path': 'flask/logging.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [13, 41, 49], 'mod': [1, 2, 3, 4, 6, 8, 9, 10, 17, 18, 19, 22, 23, 24, 25, 26, 27, 28]}, "(None, '_proxy_stream', 32)": {'mod': [32, 33, 34, 35, 37, 38, 39, 40]}, "(None, '_should_log_for', 43)": {'mod': [43, 44, 45, 46]}, "(None, 'create_logger', 50)": {'mod': [51, 52, 53, 54, 55, 57, 59, 60, 61, 62, 63, 65, 66, 67, 68, 70, 71, 72, 73, 75, 76, 77, 79, 80, 81, 83, 84, 85, 86, 87, 88, 89, 91, 92]}}}, {'path': 'tests/test_basic.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [1023]}, "(None, 'test_teardown_request_handler_error', 739)": {'mod': [741]}, "(None, 'test_error_handling', 816)": {'mod': [817]}, "(None, 'test_error_handling_processing', 862)": {'mod': [863]}, "(None, 'test_baseexception_error_handling', 884)": {'mod': [885]}, "(None, 'apprunner', 1427)": {'mod': [1428]}}}, {'path': 'tests/test_helpers.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [12, 16, 18, 19, 22, 23, 25]}, "('TestLogging', None, 663)": {'mod': [663, 664, 665, 666, 667, 668, 669, 670, 672, 673, 675, 676, 677, 678, 679, 681, 682, 683, 685, 686, 687, 688, 689, 690, 691, 693, 694, 696, 697, 698, 699, 700, 702, 703, 704, 705, 706, 707, 709, 710, 711, 713, 714, 715, 717, 718, 719, 720, 721, 723, 724, 725, 727, 728, 729, 730, 732, 733, 734, 735, 736, 738, 739, 740, 742, 743, 744, 746, 747, 748, 749]}}}, {'path': 'tests/test_subclassing.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [13]}, "(None, 'test_suppressed_exception_logging', 18)": {'mod': [25, 26, 32, 36, 37]}, "(None, 'index', 29)": {'mod': [30]}}}, {'path': 'tests/test_templating.py', 'status': 'modified', 'Loc': {"(None, 'test_template_loader_debugging', 402)": {'mod': [402, 422, 423, 424, 425, 426, 428, 429, 431, 432, 433, 434]}}}, {'path': 'tests/test_testing.py', 'status': 'modified', 'Loc': {"(None, 'test_test_client_context_binding', 209)": {'mod': [210]}}}]}
[]
[]
[]
{ "iss_type": "4", "iss_reason": "2", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "flask/logging.py", "flask/app.py" ], "doc": [ "docs/contents.rst.inc", "docs/config.rst", "docs/errorhandling.rst", "CHANGES" ], "test": [ "tests/test_templating.py", "tests/test_testing.py", "tests/test_basic.py", "tests/test_helpers.py", "tests/test_subclassing.py" ], "config": [], "asset": [] }
1
pallets
flask
465da9f610a04d379bb39a0ff03fb6c0b0ea1c45
https://github.com/pallets/flask/issues/2866
logging
DispatcherMiddleware with different loggers per app in flask 1.0
After upgrading to flask 1.0 logging from different apps using DispatcherMiddleware, each log emitted is written to all handlers in the different apps. I assume this caused by `app.logger` always having the name `flask.app`, maybe? Here is a example: ``` from werkzeug.wsgi import DispatcherMiddleware from flask import Flask from logging.handlers import RotatingFileHandler handler1 = RotatingFileHandler('app1.log') app1 = Flask('app1') app1.logger.addHandler(handler1) handler2 = RotatingFileHandler('app2.log') app2 = Flask('app2') app2.logger.addHandler(handler2) @app1.route("/") def hello(): app1.logger.error("from app1") return '' @app2.route("/") def hello2(): app2.logger.error("from app2") return '' app = DispatcherMiddleware(app1, { '/app2': app2 }) ``` Run with ``` uwsgi --socket 0.0.0.0:8000 --protocol=http -w app --callable app ``` And then make a request to / and /app2/. Each error log will be written in both logfiles. ### Environment * Python version: 3.6.5 * Flask version: 1.0.2 * Werkzeug version: 0.14.1 My actual app is using `current_app.logger` with blueprints with the same behaviour, but I assume it the same issue.
null
https://github.com/pallets/flask/pull/3282
null
{'base_commit': '465da9f610a04d379bb39a0ff03fb6c0b0ea1c45', 'files': [{'path': 'CHANGES.rst', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [20]}}}, {'path': 'docs/config.rst', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [385]}}}, {'path': 'docs/errorhandling.rst', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [234]}}}, {'path': 'docs/logging.rst', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [1, 6, 7, 8, 9]}}}, {'path': 'src/flask/app.py', 'status': 'modified', 'Loc': {"('Flask', 'logger', 655)": {'mod': [656, 657, 659, 660, 662, 663, 665, 667, 668, 669, 670, 671]}}}, {'path': 'src/flask/logging.py', 'status': 'modified', 'Loc': {"(None, 'create_logger', 59)": {'mod': [60, 69]}}}, {'path': 'tests/test_logging.py', 'status': 'modified', 'Loc': {"(None, 'reset_logging', 21)": {'mod': [26]}, "(None, 'test_logger', 44)": {'mod': [45]}}}, {'path': 'tests/test_templating.py', 'status': 'modified', 'Loc': {"(None, 'test_template_loader_debugging', 409)": {'mod': [433]}}}]}
[]
[]
[]
{ "iss_type": "2", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "src/flask/logging.py", "src/flask/app.py" ], "doc": [ "docs/config.rst", "docs/logging.rst", "docs/errorhandling.rst", "CHANGES.rst" ], "test": [ "tests/test_templating.py", "tests/test_logging.py" ], "config": [], "asset": [] }
1
pallets
flask
c8cf4694c60f0d81809468a1b45ec730496cc546
https://github.com/pallets/flask/issues/5160
Switch to importlib breaks scripts with `app.run()`
With a trivial script [using `app.run()`](https://flask.palletsprojects.com/en/2.3.x/server/#in-code) such as: ```python3 from flask import Flask app = Flask(__name__) if __name__ == "__main__": app.run(debug=True) ``` The current git `main` breaks with: ```pytb Traceback (most recent call last): File "/home/florian/tmp/flask/app.py", line 3, in <module> app = Flask(__name__) ^^^^^^^^^^^^^^^ File "/home/florian/tmp/flask/src/flask/app.py", line 376, in __init__ instance_path = self.auto_find_instance_path() ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/florian/tmp/flask/src/flask/app.py", line 630, in auto_find_instance_path prefix, package_path = find_package(self.import_name) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/florian/tmp/flask/src/flask/scaffold.py", line 898, in find_package package_path = _find_package_path(import_name) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/florian/tmp/flask/src/flask/scaffold.py", line 858, in _find_package_path spec = importlib.util.find_spec(root_mod_name) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "<frozen importlib.util>", line 114, in find_spec ValueError: __main__.__spec__ is None ``` This seems to be a regression due to 84e11a1e827c0f55f9b9ee15952eddcf8a6492e0 from #5157. Environment: - Python version: 3.11.4 - Flask version: git main
null
https://github.com/pallets/flask/pull/5161
null
{'base_commit': 'c8cf4694c60f0d81809468a1b45ec730496cc546', 'files': [{'path': 'CHANGES.rst', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [7]}}}, {'path': 'src/flask/helpers.py', 'status': 'modified', 'Loc': {"(None, 'get_root_path', 562)": {'mod': [578, 579, 584]}}}, {'path': 'src/flask/scaffold.py', 'status': 'modified', 'Loc': {"(None, '_matching_loader_thinks_module_is_package', 782)": {'mod': [782, 783, 785, 786, 787, 788, 789, 790, 792, 794, 795, 796, 797, 799, 800, 801, 802, 803, 804]}, "(None, '_find_package_path', 816)": {'mod': [825, 826, 827, 828, 829, 831, 832, 833, 834, 835, 836, 837, 838, 839, 840, 841, 842, 843, 844, 845, 846, 847, 848, 849, 850, 851, 852, 853, 854, 855, 857, 858, 859, 861, 862, 865, 866, 867, 868, 869, 870, 871, 872, 873, 875, 877, 878, 879, 880, 882]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "src/flask/helpers.py", "src/flask/scaffold.py" ], "doc": [ "CHANGES.rst" ], "test": [], "config": [], "asset": [] }
1
sherlock-project
sherlock
a2655907389f1625540c0c643aed33bc26e63da8
https://github.com/sherlock-project/sherlock/issues/762
false positive
linkedin false positive
null
https://github.com/sherlock-project/sherlock/pull/773
null
{'base_commit': 'a2655907389f1625540c0c643aed33bc26e63da8', 'files': [{'path': 'removed_sites.json', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [484], 'mod': [149, 150]}}}, {'path': 'removed_sites.md', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [956]}}}, {'path': 'sherlock/resources/data.json', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [891, 892, 893, 894, 895, 896, 897, 898, 899, 1521, 1522, 1523, 1524, 1525, 1526, 1527, 1528, 2259, 2260, 2261, 2262, 2263, 2264, 2265, 2266, 2267]}}}, {'path': 'sites.md', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [1, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300]}}}]}
[]
[]
[]
{ "iss_type": "2", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "removed_sites.json", "sherlock/resources/data.json" ], "doc": [ "removed_sites.md", "sites.md" ], "test": [], "config": [], "asset": [] }
1
sherlock-project
sherlock
b9ff06593288bcf94211878e41e5d010eb1e00f2
https://github.com/sherlock-project/sherlock/issues/30
enhancement
Optional csv Result Output
While the text output file is good to have, it only shows the sites that do have an account with the username. It would be useful to also have a list of the sites that still had an opening. I think that it would be useful to have a csv output as well. It could be optional, so the user would have to add a `--csv` switch before the file would be output. Then the output would be available in LibreOffice or Excel for users to sort and organize as they wish. I am thinking of the following columns: - Username The user name that is being queried. - Social Network Name For example, "Twitter", "GitHub", ... - Social Network Main URL For example, "https://twitter.com/" for Twitter - Social Network User URL This is the URL that Sherlock would be trying to check the existence of. For example, "https://twitter.com/user123" for Twitter. - Username Exists Will be either "True" or "False". - Request Response Data I think that this would be useful for debug of the program. Especially for those sites where Sherlock depends on the response text (as opposed to an explicit HTTP response code).
null
https://github.com/sherlock-project/sherlock/pull/32
null
{'base_commit': 'b9ff06593288bcf94211878e41e5d010eb1e00f2', 'files': [{'path': 'sherlock.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [0, 6, 20], 'mod': [9]}, "(None, 'print_error', 26)": {'mod': [26]}, "(None, 'make_request', 33)": {'mod': [33, 37, 46]}, "(None, 'sherlock', 49)": {'mod': [57, 58, 60, 61, 62, 63, 64, 65, 67, 68, 69, 70, 72, 73, 74, 75, 77, 79, 80, 81, 82, 83, 84, 86, 87, 89, 90, 91, 92, 93, 95, 96, 98, 99, 100, 101, 102, 103, 104, 105, 107, 108]}, "(None, 'main', 115)": {'mod': [134, 155, 157, 158]}}}]}
[]
[]
[]
{ "iss_type": "4", "iss_reason": "2", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "sherlock.py" ], "doc": [], "test": [], "config": [], "asset": [] }
1
sherlock-project
sherlock
e781ed5e260e1fdb58410a3103adb2b95ce547c8
https://github.com/sherlock-project/sherlock/issues/503
site support request
Add search "kali community"
``` "kali community": { "errorType": "status_code", "rank": 10313, "url": "https://forums.kali.org/member.php?username={}", "urlMain": "https://forums.kali.org/", "username_claimed": "blue", "username_unclaimed": "noonewouldeverusethis7" }, ```
null
https://github.com/sherlock-project/sherlock/pull/567
null
{'base_commit': 'e781ed5e260e1fdb58410a3103adb2b95ce547c8', 'files': [{'path': 'data.json', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [288, 545, 665, 939, 974, 1374, 1390, 1628]}}}]}
[]
[]
[]
{ "iss_type": "4", "iss_reason": "2", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "data.json" ], "doc": [], "test": [], "config": [], "asset": [] }
1
sherlock-project
sherlock
7b3cf0aaf28e0e4051dffedb6d1416b9fa9ef456
https://github.com/sherlock-project/sherlock/issues/1305
bug
[ISSUE] Sherlock stops after Countable search
- [X] I'm reporting a bug in Sherlock's functionality - [X] The bug I'm reporting is not a false positive or a false negative - [X] I've verified that I'm running the latest version of Sherlock - [X] I've checked for similar bug reports including closed ones - [X] I've checked for pull requests that attempt to fix this bug ## Description Upon running `python3 sherlock username`, the checks stop right after a return on Countable. (See screenshot below) As I am unsure of if this is a social media downtime issue, I am consulting here. <img width="401" alt="Screen_Shot_2022-04-16_at_11 00 42_PM" src="https://user-images.githubusercontent.com/60671860/163698401-dd7c35e0-b6ec-468c-9b9d-6ac29806065c.png"> Suggested fix: Improve 404 messages and insert timeouts if sites are not reachable within a few seconds.
null
https://github.com/sherlock-project/sherlock/pull/1307
null
{'base_commit': '7b3cf0aaf28e0e4051dffedb6d1416b9fa9ef456', 'files': [{'path': 'removed_sites.json', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [655]}}}, {'path': 'removed_sites.md', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [1267]}}}, {'path': 'sherlock/resources/data.json', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [545, 546, 547, 548, 549, 550, 551]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "removed_sites.json", "sherlock/resources/data.json" ], "doc": [ "removed_sites.md" ], "test": [], "config": [], "asset": [] }
1
sherlock-project
sherlock
588616d6155edb4184e5ff8f4cf0104e0a0e2468
https://github.com/sherlock-project/sherlock/issues/326
bug
Codepen False Negative/Trip False Positive
![image](https://user-images.githubusercontent.com/12257112/65722187-a5f37e00-e0a3-11e9-9a69-676ab5c5da78.png) ![image](https://user-images.githubusercontent.com/12257112/65722337-fc60bc80-e0a3-11e9-95da-fc8ba98cbbeb.png)
null
https://github.com/sherlock-project/sherlock/pull/328
null
{'base_commit': '588616d6155edb4184e5ff8f4cf0104e0a0e2468', 'files': [{'path': 'README.md', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [71]}}}, {'path': 'data.json', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [1257, 1258]}}}, {'path': 'sherlock.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [29]}}}]}
[]
[]
[]
{ "iss_type": "2", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "sherlock.py", "data.json" ], "doc": [ "README.md" ], "test": [], "config": [], "asset": [] }
1
sherlock-project
sherlock
508eb88724dbe20dedf07dc00527ad4f32c93a77
https://github.com/sherlock-project/sherlock/issues/5
bug
Formatting on windows Powershell very off
[+] Unsplash: https://unsplash.com/@User [*] Saved: User.txt That's how all the lines look like in windows powershell. I guess the print is designed for a linux platform?
null
https://github.com/sherlock-project/sherlock/pull/71
null
{'base_commit': '508eb88724dbe20dedf07dc00527ad4f32c93a77', 'files': [{'path': 'requirements.txt', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [3]}}}, {'path': 'sherlock.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [16]}, "(None, 'main', 247)": {'add': [247]}}}]}
[]
[]
[]
{ "iss_type": "2", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "sherlock.py" ], "doc": [], "test": [], "config": [ "requirements.txt" ], "asset": [] }
1
sherlock-project
sherlock
e6642737462aaadf2d1f36ea0d417b8cf8a40541
https://github.com/sherlock-project/sherlock/issues/301
bug
Error Connecting for most websites (Failed to establish a new connection: [Errno -2] Name or service not known' in urllib3
I believe this might be a partial duplicate of #230, however, because of lack of information and error output I made my own issue just to be safe. With the latest python3.7.3 and the correct requirements I constantly run into the issue of websites not connecting. Specifically this happens after trying to connect to Bandcamp, subsequently, all other urls following bandcamp fail as well. I printed the `requests.exceptions.ConnectionError` at line 128 as well as the future object class. For some reason, urllib3 seems to fail finding the name or service (so this might be a DNS issue?). Further details below (snippet from e.g. YouTube): ``` [-] Error Connecting: YouTube [-] YouTube: Error! {'errorType': 'response_url', 'errorUrl': 'https://www.zhihu.com/people/{}', 'rank': 0, 'url': 'https://www.zhihu.com/people/{}', 'urlMain': 'https://www.zhihu.com/', 'username_claimed': 'blue', 'username_unclaimed': 'noonewouldeverusethis7', 'request_future': <Future at 0x7f318072ce80 state=finished raised ConnectionError>} <Future at 0x7f318072ce80 state=finished raised ConnectionError> HTTPSConnectionPool(host='www.zhihu.com', port=443): Max retries exceeded with url: /people/berghopper (Caused by NewConnectionError('<urllib3.connection.VerifiedHTTPSConnection object at 0x7f318070c2e8>: Failed to establish a new connection: [Errno -2] Name or service not known')) <Future at 0x7f318072ce80 state=finished raised ConnectionError> ```
null
https://github.com/sherlock-project/sherlock/pull/471
null
{'base_commit': 'e6642737462aaadf2d1f36ea0d417b8cf8a40541', 'files': [{'path': 'README.md', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [72]}}}, {'path': 'data.json', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [50, 315, 714, 722, 1032, 1118, 1302, 1617, 1872, 1880, 1896, 1939, 1971, 2005, 2013], 'mod': [4, 13, 21, 29, 37, 45, 53, 61, 69, 78, 87, 95, 104, 112, 120, 128, 137, 146, 154, 162, 171, 179, 189, 197, 205, 215, 224, 232, 241, 250, 259, 267, 276, 284, 293, 302, 310, 318, 326, 334, 343, 351, 359, 368, 377, 386, 395, 403, 411, 419, 428, 444, 453, 461, 469, 477, 487, 497, 506, 514, 523, 531, 539, 549, 557, 565, 574, 582, 590, 598, 608, 616, 624, 633, 641, 649, 657, 665, 675, 684, 692, 700, 709, 717, 726, 735, 744, 753, 762, 770, 779, 788, 797, 805, 814, 832, 840, 848, 856, 865, 873, 881, 889, 898, 907, 916, 925, 933, 941, 950, 958, 968, 976, 985, 993, 1002, 1010, 1019, 1027, 1035, 1043, 1052, 1061, 1070, 1078, 1097, 1105, 1113, 1121, 1130, 1139, 1148, 1156, 1164, 1172, 1180, 1188, 1196, 1204, 1221, 1229, 1237, 1245, 1254, 1262, 1271, 1280, 1289, 1297, 1323, 1332, 1342, 1351, 1359, 1367, 1384, 1393, 1401, 1410, 1418, 1426, 1434, 1442, 1451, 1459, 1467, 1475, 1484, 1493, 1501, 1510, 1519, 1527, 1535, 1544, 1552, 1561, 1569, 1578, 1587, 1596, 1612, 1621, 1630, 1639, 1657, 1665, 1673, 1682, 1690, 1698, 1706, 1714, 1723, 1732, 1741, 1750, 1758, 1767, 1776, 1784, 1793, 1803, 1811, 1820, 1829, 1838, 1856, 1867, 1875, 1883, 1891, 1900, 1908, 1916, 1925, 1934, 1942, 1950, 1958, 1966, 1975, 1984, 1992, 2000, 2008, 2016, 2025, 2033, 2042, 2051, 2059, 2068, 2076, 2082, 2083, 2084, 2085, 2086, 2087, 2088, 2089, 2090, 2091, 2092, 2093, 2094, 2095, 2096, 2097, 2098, 2099, 2100, 2101, 2102, 2103, 2104, 2105, 2106, 2107, 2108, 2109, 2110, 2111, 2112, 2113, 2114, 2115, 2116, 2117, 2118, 2119, 2120, 2121, 2122, 2123, 2124, 2125, 2126, 2127, 2128, 2129, 2130, 2131, 2132, 2133, 2134, 2135, 2136, 2137, 2138, 2139, 2140, 2141, 2142, 2143, 2144, 2145, 2146, 2147, 2148, 2149, 2150, 2151, 2152, 2153, 2154, 2155, 2156, 2157, 2158, 2159, 2160, 2161, 2162, 2163, 2164, 2165, 2166, 2167, 2168, 2169, 2170, 2171, 2172, 2173, 2174, 2175, 2176, 2177, 2178, 2179, 2180, 2181, 2182, 2183, 2184, 2185, 2186, 2187, 2188, 2189, 2190, 2191, 2192, 2193, 2194, 2195, 2196, 2197, 2198, 2199, 2200, 2201, 2202, 2203, 2204, 2205, 2206, 2207, 2208, 2209, 2210, 2211, 2212, 2213, 2216, 2222, 2224, 2225, 2226, 2227, 2228, 2229, 2230, 2231, 2232, 2233, 2234, 2235, 2236, 2242, 2248, 2249, 2250, 2251, 2252, 2253, 2254, 2255, 2256, 2258, 2259, 2260, 2264, 2266, 2267, 2268, 2272, 2273, 2274, 2275, 2276, 2277, 2278, 2281, 2283, 2284, 2285, 2287]}}}, {'path': 'sherlock.py', 'status': 'modified', 'Loc': {"(None, 'sherlock', 151)": {'add': [189, 191], 'mod': [183, 184, 187, 188, 193, 194, 195]}, '(None, None, None)': {'mod': [29]}}}, {'path': 'sites.md', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [1, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 248]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "3", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "sherlock.py", "data.json" ], "doc": [ "README.md", "sites.md" ], "test": [], "config": [], "asset": [] }
1
keras-team
keras
96130040540e1405ffe746ddf2b2cceb9b8b8f65
https://github.com/keras-team/keras/issues/16453
type:docs
The vocabulary_size method of preprocessing layers does not work in graph mode
**System information**. - Have I written custom code (as opposed to using a stock example script provided in Keras): **yes** - OS Platform and Distribution (e.g., Linux Ubuntu 16.04): **Colab** - TensorFlow installed from (source or binary): **binary** - TensorFlow version (use command below): **TF 2.8, nightly 2.10.0-dev20220422** **Describe the problem**. Using the `vocabulary_size()` method of preprocessing layers (like `tf.keras.layers.StringLookup`) fails in graph mode, because the implementation looks like this: https://github.com/keras-team/keras/blob/aea9728313bcaa8262774699c21976288171b209/keras/layers/preprocessing/index_lookup.py#L346-L352 **Describe the current behavior**. Calling `vocabulary_size()` in graph mode fails. **Describe the expected behavior**. Calling `vocabulary_size()` in graph mode succeeds and returns a `tf.Tensor` with the size. **Standalone code to reproduce the issue**. Colab notebook showing the issue is at https://colab.research.google.com/drive/1Mq9G8eUvNLw6ykk4ARKf6jurLREAvRhu?usp=sharing both for TF 2.8 and TF nightly. **Source code / logs**. An obvious fix is to avoid using the `numpy()` method when in graph mode, i.e., instead of ```python return int(self.lookup_table.size().numpy()) + self._token_start_index() ``` use just ```python return self.lookup_table.size() + self._token_start_index() ``` Note that the above notebook also shows that this implementation works in the graph mode.
null
https://github.com/keras-team/keras/pull/16460
null
{'base_commit': '96130040540e1405ffe746ddf2b2cceb9b8b8f65', 'files': [{'path': 'keras/layers/preprocessing/index_lookup.py', 'status': 'modified', 'Loc': {"('IndexLookup', 'vocabulary_size', 346)": {'mod': [352]}}}]}
[]
[]
[]
{ "iss_type": "2", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "keras/layers/preprocessing/index_lookup.py" ], "doc": [], "test": [], "config": [], "asset": [] }
1
keras-team
keras
5b299743442b64afaeeec01e925ddbeb112aad3c
https://github.com/keras-team/keras/issues/20712
type:support
Tensorflow BackupAndRestore method does not work
I copied the [example code here](https://keras.io/api/callbacks/backup_and_restore/) and it raises a ValueError with Python 3.11 and Tensorflow 2.17: ``` import keras import numpy as np class InterruptingCallback(keras.callbacks.Callback): def on_epoch_begin(self, epoch, logs=None): if epoch == 4: raise RuntimeError('Interrupting!') callback = keras.callbacks.BackupAndRestore(backup_dir="/tmp/backup") model = keras.models.Sequential([keras.layers.Dense(10)]) model.compile(keras.optimizers.SGD(), loss='mse') try: model.fit(np.arange(100).reshape(5, 20), np.zeros(5), epochs=10, batch_size=1, callbacks=[callback, InterruptingCallback()], verbose=0) except Exception as e: print(e) history = model.fit(np.arange(100).reshape(5, 20), np.zeros(5), epochs=10, batch_size=1, callbacks=[callback], verbose=0) len(history.history['loss']) ValueError: To use the BackupAndRestore method, your model must be built before you call `fit()`. Model is unbuilt. You can build it beforehand by calling it on a batch of data. ``` Doesn't that defeat the purpose of backupAndRestore?
null
https://github.com/keras-team/keras/pull/20714
null
{'base_commit': '5b299743442b64afaeeec01e925ddbeb112aad3c', 'files': [{'path': 'keras/src/callbacks/backup_and_restore.py', 'status': 'modified', 'Loc': {"('BackupAndRestore', None, 9)": {'add': [39]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "keras/src/callbacks/backup_and_restore.py" ], "doc": [], "test": [], "config": [], "asset": [] }
1
keras-team
keras
a56b16fffec6e4a431bf14e13e7dabeeb5904cd8
https://github.com/keras-team/keras/issues/8426
vgg16 predictions: class probabilities change significantly in 2.0.9
Please make sure that the boxes below are checked before you submit your issue. If your issue is an implementation question, please ask your question on [StackOverflow](http://stackoverflow.com/questions/tagged/keras) or [join the Keras Slack channel](https://keras-slack-autojoin.herokuapp.com/) and ask there instead of filing a GitHub issue. Thank you! - [x] Check that you are up-to-date with the master branch of Keras. You can update with: pip install git+git://github.com/fchollet/keras.git --upgrade --no-deps - [x] If running on TensorFlow, check that you are up-to-date with the latest version. The installation instructions can be found [here](https://www.tensorflow.org/get_started/os_setup). - [x] If running on Theano, check that you are up-to-date with the master branch of Theano. You can update with: pip install git+git://github.com/Theano/Theano.git --upgrade --no-deps - [x] Provide a link to a GitHub Gist of a Python script that can reproduce your issue (or just copy the script here if it is short). I've noticed when updating to Keras 2.0.9 that the class probabilities for predictions on the "creative_commons_elephant.jpg" image used in tutorials change significantly (i.e. African_elephant goes from `0.909` to `0.789`). Here is my script: ```python from keras.applications.vgg16 import VGG16 model = VGG16(weights='imagenet') from keras.preprocessing import image from keras.applications.vgg16 import preprocess_input, decode_predictions import numpy as np img_path = 'creative_commons_elephant.jpg' img = image.load_img(img_path, target_size=(224, 224)) x = image.img_to_array(img) x = np.expand_dims(x, axis=0) x = preprocess_input(x) preds = model.predict(x) print('Predicted:', decode_predictions(preds, top=3)[0]) ``` Under Keras 2.0.8 this gives: ``` ('Predicted:', [(u'n02504458', u'African_elephant', 0.90942073), (u'n01871265', u'tusker', 0.086183183), (u'n02504013', u'Indian_elephant', 0.0043545808)]) ``` However under Keras 2.0.9 it gives: ``` ('Predicted:', [(u'n02504458', u'African_elephant', 0.78988522), (u'n01871265', u'tusker', 0.1987267), (u'n02504013', u'Indian_elephant', 0.011142471)]) ```
null
https://github.com/keras-team/keras/pull/8435
null
{'base_commit': 'a56b16fffec6e4a431bf14e13e7dabeeb5904cd8', 'files': [{'path': 'keras/preprocessing/image.py', 'status': 'modified', 'Loc': {"(None, 'load_img', 321)": {'mod': [322, 335]}}}]}
[]
[]
[]
{ "iss_type": "2", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "keras/preprocessing/image.py" ], "doc": [], "test": [], "config": [], "asset": [] }
1
keras-team
keras
ffb8b813818221a9b43d51e251d489a40b116607
https://github.com/keras-team/keras/issues/9434
type:bug/performance
Error with multiprocessing on Sequence in fit_generator()
I'm trying to use a `Sequence` as the generator for `model.fit_generator()`. With `use_multiprocessing=False` it works fine, but with `use_multiprocessing=True` an error is produced. **Minimal working example:** ```python from keras.utils import Sequence from keras.models import Sequential from keras.layers import Dense from keras.utils import to_categorical import numpy as np class DummySequence(Sequence): def __init__(self, x_set, y_set, batch_size): self.x, self.y = x_set, y_set self.batch_size = batch_size def __len__(self): return int(np.ceil(len(self.x) / float(self.batch_size))) def __getitem__(self, idx): batch_x = self.x[idx * self.batch_size:(idx + 1) * self.batch_size] batch_y = self.y[idx * self.batch_size:(idx + 1) * self.batch_size] return np.array(batch_x), np.array(batch_y) if __name__ == '__main__': x = np.random.random((100, 3)) y = to_categorical(np.random.random(100) > .5).astype(int) seq = DummySequence(x, y, 10) model = Sequential() model.add(Dense(32, input_dim=3)) model.add(Dense(2, activation='softmax')) model.compile(optimizer='rmsprop', loss='categorical_crossentropy', metrics=['accuracy']) model.fit_generator(generator=seq, workers=2, use_multiprocessing=True) ``` **Error:** ``` Traceback (most recent call last): File "C:\Users\elcombato\AppData\Local\Continuum\Anaconda3\envs\ml\lib\multiprocessing\pool.py", line 119, in worker result = (True, func(*args, **kwds)) File "C:\Users\elcombato\AppData\Local\Continuum\Anaconda3\envs\ml\lib\site-packages\keras\utils\data_utils.py", line 392, in get_index return _SHARED_SEQUENCES[uid][i] KeyError: 0 ``` **Setup** Windows 10 Python 3.6.4 Keras 2.1.3 Tensorflow 1.2.1
null
https://github.com/keras-team/keras/pull/9436
null
{'base_commit': 'ffb8b813818221a9b43d51e251d489a40b116607', 'files': [{'path': 'keras/utils/data_utils.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [375]}, "('OrderedEnqueuer', 'start', 500)": {'mod': [509, 511]}, "('OrderedEnqueuer', '_run', 526)": {'mod': [534]}}}, {'path': 'tests/keras/utils/data_utils_test.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [9, 11, 26, 221, 240, 251, 295]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "keras/utils/data_utils.py", "tests/keras/utils/data_utils_test.py" ], "doc": [], "test": [], "config": [], "asset": [] }
1
keras-team
keras
fed28a7357e13aeb955f891747a1f9b26d5bc581
https://github.com/keras-team/keras/issues/19520
type:docs stat:awaiting response from contributor
kl divergence outputs confusing result when the label contains negative
Here are some APIs implemented for kl divergence: ``` tf.losses.kullback_leibler_divergence tf.keras.losses.KLDivergence tf.keras.metrics.kl_divergence ``` However, when the `y_true` or `y_pred` is negative, these APIs output incorrect result which is inconsistent with the documentation (https://www.tensorflow.org/api_docs/python/tf/keras/losses/KLDivergence). Here is the code to reproduce: ``` import os os.environ['TF_CPP_MIN_LOG_LEVEL'] = '2' os.environ['OPENBLAS_NUM_THREADS'] = '1' os.environ['CUDA_VISIBLE_DEVICES'] = '' import numpy as np np.random.seed(35) import tensorflow as tf y_true = tf.constant(np.random.randn(1), dtype='float32') y_pred = tf.constant(np.random.randn(1), dtype='float32') print(y_true, y_pred) out1 = tf.losses.kullback_leibler_divergence(y_true,y_pred) out2 = tf.keras.losses.KLDivergence( reduction='sum_over_batch_size', name='kl_divergence' )(y_true, y_pred) out3 = tf.keras.metrics.kl_divergence(y_true, y_pred) print(out1, out2, out3) # 0.0, 0.0, 0.0 print(f"Expected result: {y_true*(np.log(y_true/y_pred))}") # Expected result following the equation: [-2.8711002] ``` I notice that the current code will silently clip the negative value close to zero: https://github.com/keras-team/keras/blob/61bbff593a0914f5a2c426c14caadb7372f56da0/keras/losses/losses.py#L1466-L1467 If negative value is not allowed for these concerned APIs, maybe some descriptions on the documentation or validation check inside the source code can provide more information.
null
https://github.com/keras-team/keras/pull/19526
null
{'base_commit': 'fed28a7357e13aeb955f891747a1f9b26d5bc581', 'files': [{'path': 'keras/losses/losses.py', 'status': 'modified', 'Loc': {"(None, 'kl_divergence', 1441)": {'mod': [1464]}}}]}
[]
[]
[]
{ "iss_type": "2", "iss_reason": "2", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "keras/losses/losses.py" ], "doc": [], "test": [], "config": [], "asset": [] }
1
keras-team
keras
4bb90a5bfd762269adeee772e233a733a6d318a9
https://github.com/keras-team/keras/issues/12013
To investigate
Nasnet weight errors (no_top version)
I"m running into this error: ValueError: You are trying to load a weight file containing 532 layers into a model with 526 layers. My keras version is: `>>> keras.__version__ '2.2.4'` `Tensorflow version: '1.12.0'` Bascially, when the model tries to load I get this value error ` ValueError: You are trying to load a weight file containing 532 layers into a model with 526 layers. ` I"ve looked at this thread: https://github.com/keras-team/keras/issues/10109 However, I"m trying to run the `no_top version`, so It shouldn't matter what my input vector is. base_model(weights='imagenet', include_top=False) Thank you. Here is the script Im using: ################IMPORTS######################## #--Keras imports--# from keras.applications import resnet50, xception,inception_v3,inception_resnet_v2, densenet, nasnet, imagenet_utils from keras.preprocessing import image from keras.preprocessing.image import ImageDataGenerator from keras.layers import Input, Dense from keras import backend as k from keras.models import Model, clone_model from keras.layers import Dense, GlobalAveragePooling2D,Dropout, BatchNormalization from keras import optimizers from keras.optimizers import Adam from keras.callbacks import ModelCheckpoint,ProgbarLogger from keras.utils import print_summary from keras import __version__ from keras.datasets import cifar10 #--python imports--# import os import numpy as np import datetime import h5py import json import time ################# -- PARAMETERS -- ################################## img_width, img_height = 480, 480 (x_train, y_train), _ = cifar10.load_data() classes = len(y_train[0]) ##------initial training parameters -----## i_epochs = 10 i_batch_size = 20 i_steps_per_epoch = 100 i_optimizer = optimizers.SGD(lr=0.0001, momentum=0.9) #################### MODELS ###################################### def basemodel(): base_model = nasnet.NASNetLarge(weights='imagenet', include_top=False) preprocess = nasnet.preprocess_input return base_model, preprocess def full_model(): base_model, preprocess = basemodel() x = base_model.output x = GlobalAveragePooling2D()(x) x = Dense(2048, activation='relu')(x) x = Dropout(.60)(x) x= Dense(512, activation='relu')(x) predictions = Dense(classes, activation='softmax')(x) model = Model(inputs= base_model.input, outputs=predictions) return model,preprocess def initial_training_full(): model, preprocess = full_model() for layer in model.layers[:-5]: layer.trainable = False model.compile(optimizer= i_optimizer, loss='sparse_categorical_crossentropy', metrics = ['accuracy']) print('Starting model training') history = model.fit(x_train, y_train, steps_per_epoch = i_steps_per_epoch, epochs = i_epochs, shuffle= True, verbose = 1) return history if __name__ == "__main__": initial_training_full()
null
https://github.com/keras-team/keras/pull/62
null
{'base_commit': '4bb90a5bfd762269adeee772e233a733a6d318a9', 'files': [{'path': 'keras/datasets/data_utils.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [3, 4]}, "(None, 'get_file', 7)": {'mod': [32]}}}, {'path': 'setup.py', 'status': 'modified', 'Loc': {}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "keras/datasets/data_utils.py", "setup.py" ], "doc": [], "test": [], "config": [], "asset": [] }
1
keras-team
keras
3d9428d3445a429b535a247168d93b8a5910219d
https://github.com/keras-team/keras/issues/6298
Performance Using HDF5 File vs In-Memory Dataset
I am trying out using the HDF5 file as the format of training/testing data, so that I can avoid the problem of big dataset which cannot fit into the memory. As a test, I used a dataset with about 20 million rows and 10 columns as my input, and my response variable is a real-valued vector. The model is a simple sequential model with 2 hidden layers and 50 neurons per hidden layer, and the training batch size is 10000. I am still using Keras version 1.2.0. When the dataset is fully loaded into the memory, one epoch of training takes about 30 seconds. However, if I used HDF5 matrix with HDF5Matrix, one epoch of training takes about 360 seconds. Is this huge slow down common due to the I/O bottleneck? Shall I implement some perhaps better data loading process with fit_generator by myself by considering the specifications of my dataset?
null
https://github.com/keras-team/keras/pull/6891
null
{'base_commit': '3d9428d3445a429b535a247168d93b8a5910219d', 'files': [{'path': 'keras/engine/training.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [12], 'mod': [7, 9, 10]}, "('Model', 'fit_generator', 1715)": {'add': [1732, 1856], 'mod': [1723, 1725, 1734, 1759, 1762, 1807, 1860, 1861, 1869, 1870, 1871, 1872, 1873, 1874, 1875, 1926, 1928]}, "('Model', 'evaluate_generator', 1956)": {'add': [1965, 1994], 'mod': [1957, 1968, 1971, 1998, 1999, 2002, 2003, 2004, 2005, 2006, 2007, 2008]}, "('Model', 'predict_generator', 2056)": {'add': [2091], 'mod': [2057, 2058, 2065, 2068, 2071, 2095, 2096, 2102, 2103, 2104, 2105, 2106, 2107, 2108]}, "('GeneratorEnqueuer', None, 582)": {'mod': [582, 583, 585, 587, 588, 589, 590, 592, 593, 594, 595, 596, 597, 599, 600, 602, 603, 604, 605, 606, 608, 609, 610, 611, 612, 613, 614, 615, 616, 617, 618, 620, 621, 622, 623, 624, 625, 626, 627, 628, 629, 630, 632, 633, 634, 635, 636, 637, 638, 639, 640, 641, 642, 643, 644, 645, 647, 648, 650, 651, 653, 655, 656, 657, 658, 659, 661, 662, 663, 664, 665, 666, 668, 669, 670, 672, 673, 674]}}}, {'path': 'keras/legacy/interfaces.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [604]}}}, {'path': 'keras/models.py', 'status': 'modified', 'Loc': {"('Sequential', 'fit_generator', 1028)": {'mod': [1036, 1038, 1075, 1077, 1121, 1123]}, "('Sequential', 'evaluate_generator', 1127)": {'mod': [1128, 1129, 1140, 1142, 1162, 1164]}, "('Sequential', 'predict_generator', 1167)": {'mod': [1168, 1169, 1179, 1181, 1194, 1196]}}}, {'path': 'keras/utils/__init__.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [10]}}}, {'path': 'keras/utils/data_utils.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [14, 17], 'mod': [5, 6, 8, 10, 12, 13, 16]}, "(None, 'validate_file', 261)": {'add': [284]}}}, {'path': 'tests/keras/engine/test_training.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [3, 10, 14, 299]}, "(None, 'test_model_methods', 16)": {'mod': [83, 207]}}}, {'path': 'tests/keras/legacy/interface_test.py', 'status': 'modified', 'Loc': {"(None, 'test_generator_methods_interface', 781)": {'mod': [810, 811, 812, 813, 814, 815, 818, 821]}}}, {'path': 'tests/keras/test_sequential_model.py', 'status': 'modified', 'Loc': {"(None, 'test_sequential_fit_generator', 69)": {'mod': [101]}, "(None, 'test_sequential', 106)": {'mod': [136, 137]}}}, {'path': 'tests/keras/utils/data_utils_test.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [5, 8, 66], 'mod': [4, 7, 11]}}}, {'path': 'tests/test_multiprocessing.py', 'status': 'modified', 'Loc': {"(None, 'test_multiprocessing_training', 22)": {'mod': [52, 54, 60, 61]}, "(None, 'test_multiprocessing_training_fromfile', 105)": {'mod': [134, 136, 142, 143]}, "(None, 'test_multiprocessing_predicting', 149)": {'mod': [169, 171, 174, 175]}, "(None, 'test_multiprocessing_evaluating', 179)": {'mod': [202, 204, 207, 208]}, "(None, 'test_multiprocessing_fit_error', 212)": {'mod': [229, 232, 235, 238]}, "(None, 'test_multiprocessing_evaluate_error', 243)": {'mod': [258, 261, 264, 267]}, "(None, 'test_multiprocessing_predict_error', 272)": {'mod': [287, 290, 293, 296]}}}]}
[]
[]
[]
{ "iss_type": "3", "iss_reason": "3", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "tests/keras/utils/data_utils_test.py", "keras/utils/data_utils.py", "keras/utils/__init__.py", "tests/keras/legacy/interface_test.py", "keras/legacy/interfaces.py", "keras/engine/training.py", "keras/models.py" ], "doc": [], "test": [ "tests/keras/test_sequential_model.py", "tests/test_multiprocessing.py", "tests/keras/engine/test_training.py" ], "config": [], "asset": [] }
1
keras-team
keras
2bfd1f2c950df5fc3f40b903c1966f1b0a48bee4
https://github.com/keras-team/keras/issues/10855
fit_generator crashes though keras.utils.data_utils.Sequence was used
When `model.fit_generator` is used with `workers=0` and subclasses of [`keras.utils.data_utils.Sequence`](https://keras.io/utils/#sequence) for both training and validation data, API of `Sequence` is not recognized inside `evaluate_generator`, it raises: ```python File ".../keras/legacy/interfaces.py", line 91, in wrapper return func(*args, **kwargs) File ".../keras/engine/training.py", line 1415, in fit_generator initial_epoch=initial_epoch) File ".../keras/engine/training_generator.py", line 230, in fit_generator validation_steps, File ".../keras/legacy/interfaces.py", line 91, in wrapper return func(*args, **kwargs) File ".../keras/engine/training.py", line 1469, in evaluate_generator verbose=verbose) File ".../keras/engine/training_generator.py", line 298, in evaluate_generator else: ValueError: `steps=None` is only valid for a generator based on the `keras.utils.Sequence` class. Please specify `steps` or use the `keras.utils.Sequence` class. ``` Example code: ```python from keras import Sequential from keras.layers import Dense from keras.utils.data_utils import Sequence import numpy as np class Dataset(Sequence): def __getitem__(self, index): return np.random.uniform(size=(16, 8)), np.random.uniform(size=(16, 1)) def __len__(self): return 128 model = Sequential([Dense(4, activation='relu', input_shape=(8,)), Dense(1, activation='sigmoid')]) model.compile(loss='mse', optimizer='adam') model.fit_generator(generator=Dataset(), validation_data=Dataset(), workers=0) ``` Issue can be fixed [here](https://github.com/keras-team/keras/blob/7205d903fbc079bb99fbae0e3c02e6d2b4d227f0/keras/engine/training_generator.py#L124) by replacing: ```python if isinstance(val_data, Sequence): val_enqueuer_gen = iter(val_data) ``` with ```python if isinstance(val_data, Sequence): val_enqueuer_gen = iter(val_data) validation_steps = len(val_data) ```
null
https://github.com/keras-team/keras/pull/11285
null
{'base_commit': '2bfd1f2c950df5fc3f40b903c1966f1b0a48bee4', 'files': [{'path': 'keras/engine/training_generator.py', 'status': 'modified', 'Loc': {"(None, 'fit_generator', 21)": {'mod': [127]}}}, {'path': 'tests/keras/engine/test_training.py', 'status': 'modified', 'Loc': {"(None, 'test_model_methods', 126)": {'add': [470]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "3", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "keras/engine/training_generator.py" ], "doc": [], "test": [ "tests/keras/engine/test_training.py" ], "config": [], "asset": [] }
1
keras-team
keras
c0d95fd6c2cd8ffc0738819825c3065e3c89977c
https://github.com/keras-team/keras/issues/3859
clipnorm doesn't work with Embedding
I'm getting a Traceback every time "clipnorm" is used in NN with Embedding layer. Here is a simple script where the problem is obvious: ``` python import numpy as np from keras.layers import Input, Embedding from keras.optimizers import Adam from keras.models import Model input_layer = Input(shape = (1,) ) embedding = Embedding(input_dim = 1, output_dim = 1)(input_layer) model = Model(input = input_layer, output = embedding) model.compile(optimizer = Adam(clipnorm = 1.0), loss = 'mse') X = np.array([[1]]) Y = np.array([[[0.5]]]) model.fit(X, Y, nb_epoch = 1) ``` Failure: ``` shell I tensorflow/core/common_runtime/gpu/gpu_device.cc:867] Creating TensorFlow device (/gpu:0) -> (device: 0, name: GeForce GTX TITAN X, pci bus id: 0000:01:00.0) I tensorflow/core/common_runtime/gpu/gpu_device.cc:867] Creating TensorFlow device (/gpu:1) -> (device: 1, name: GeForce GTX TITAN X, pci bus id: 0000:02:00.0) Traceback (most recent call last): File "./clipnorm-bug.py", line 20, in <module> model.fit(X, Y, nb_epoch = 1) File "/usr/local/lib/python3.5/dist-packages/keras/engine/training.py", line 1079, in fit self._make_train_function() File "/usr/local/lib/python3.5/dist-packages/keras/engine/training.py", line 696, in _make_train_function self.total_loss) File "/usr/local/lib/python3.5/dist-packages/keras/optimizers.py", line 379, in get_updates grads = self.get_gradients(loss, params) File "/usr/local/lib/python3.5/dist-packages/keras/optimizers.py", line 71, in get_gradients grads = [clip_norm(g, self.clipnorm, norm) for g in grads] File "/usr/local/lib/python3.5/dist-packages/keras/optimizers.py", line 71, in <listcomp> grads = [clip_norm(g, self.clipnorm, norm) for g in grads] File "/usr/local/lib/python3.5/dist-packages/keras/optimizers.py", line 9, in clip_norm g = K.switch(n >= c, g * c / n, g) TypeError: unsupported operand type(s) for *: 'IndexedSlices' and 'float' ``` Keras version is 1.1.0, TensorFlow is 0.10rc clipvalue on the other hand works fine.
null
https://github.com/keras-team/keras/pull/4815
null
{'base_commit': 'c0d95fd6c2cd8ffc0738819825c3065e3c89977c', 'files': [{'path': '.travis.yml', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [52, 54]}}}, {'path': 'keras/backend/tensorflow_backend.py', 'status': 'modified', 'Loc': {"(None, 'rnn', 1640)": {'add': [1699], 'mod': [1720, 1738, 1745, 1746, 1747]}, "(None, 'batch_dot', 811)": {'mod': [870, 871, 872, 873]}, "(None, '_step', 1798)": {'mod': [1808, 1809]}, "(None, 'elu', 1923)": {'mod': [1934]}, "(None, 'random_binomial', 2402)": {'mod': [2407, 2408, 2409]}}}, {'path': 'keras/optimizers.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [1], 'mod': [4]}}}, {'path': 'tests/keras/test_optimizers.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [76]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "keras/optimizers.py", "keras/backend/tensorflow_backend.py" ], "doc": [], "test": [ "tests/keras/test_optimizers.py" ], "config": [ ".travis.yml" ], "asset": [] }
1
keras-team
keras
80fbbc3a6a2a30f391bad2aa85e7558c50ca0709
https://github.com/keras-team/keras/issues/4452
PReLU should be channelwise
``` from keras.models import Sequential from keras.layers.advanced_activations import PReLU from keras.layers import Convolution2D, MaxPooling2D model = Sequential() model.add(Convolution2D(32, 5, 5, input_shape=(28,28,1))) model.add(PReLU()) model.summary() ``` This script produces the PReLU layer with 18432 parameters. If the PReLU was implemented according to the paper (https://arxiv.org/abs/1502.01852), number of parameters would be 32. Shouldn't this not be implemented that way? It adds an unnecessarily large number of parameters to the model.
null
https://github.com/keras-team/keras/pull/4141
null
{'base_commit': '80fbbc3a6a2a30f391bad2aa85e7558c50ca0709', 'files': [{'path': 'keras/backend/theano_backend.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [726]}}}, {'path': 'keras/layers/advanced_activations.py', 'status': 'modified', 'Loc': {"('PReLU', None, 38)": {'add': [54], 'mod': [59]}, "('PReLU', '__init__', 59)": {'add': [62]}, "('ParametricSoftplus', None, 118)": {'add': [133]}, "('ParametricSoftplus', '__init__', 138)": {'add': [143], 'mod': [139]}, "('SReLU', None, 201)": {'add': [216]}, "('SReLU', '__init__', 221)": {'add': [227], 'mod': [222]}, "('PReLU', 'build', 65)": {'mod': [66]}, "('PReLU', 'call', 74)": {'mod': [76]}, "('ParametricSoftplus', 'build', 146)": {'mod': [147, 148, 150]}, "('ParametricSoftplus', 'call', 158)": {'mod': [159]}, "('SReLU', 'build', 230)": {'mod': [231, 238, 240, 242, 244]}, "('SReLU', 'call', 251)": {'mod': [252, 253, 254, 255]}}}, {'path': 'tests/keras/layers/test_advanced_activations.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [19, 51]}, "(None, 'test_parametric_softplus', 29)": {'mod': [31, 32, 33, 34, 35]}}}]}
[]
[]
[]
{ "iss_type": "2", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "keras/backend/theano_backend.py", "keras/layers/advanced_activations.py" ], "doc": [], "test": [ "tests/keras/layers/test_advanced_activations.py" ], "config": [], "asset": [] }
1
nvbn
thefuck
070bb2ff28b108e5fc627efa1d95ded00ca067c8
https://github.com/nvbn/thefuck/issues/537
Args are lost with git_push
From the README: > `git_push` – adds `--set-upstream origin $branch` to previous failed `git push`; This isn't actually the case, `--set-upstream` isn't added to the previous push, the entire command is replaced with just `git push --set-upstream origin $branch` regardless of whether there were any args. Example: ``` $ git push --quiet fatal: The current branch new-branch has no upstream branch. To push the current branch and set the remote as upstream, use git push --set-upstream origin new-branch $ fuck git push --set-upstream origin new-branch [enter/↑/↓/ctrl+c] ``` The `--quiet` arg is lost, ideally `git push --set-upstream origin new-branch --quiet` would be suggested.
null
https://github.com/nvbn/thefuck/pull/538
null
{'base_commit': '070bb2ff28b108e5fc627efa1d95ded00ca067c8', 'files': [{'path': 'tests/rules/test_git_push.py', 'status': 'modified', 'Loc': {"(None, 'test_match', 16)": {'add': [16]}, "(None, 'test_get_new_command', 22)": {'add': [24]}}}, {'path': 'thefuck/rules/git_push.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [0]}, "(None, 'get_new_command', 11)": {'mod': [12]}}}]}
[]
[]
[]
{ "iss_type": "2", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "thefuck/rules/git_push.py" ], "doc": [], "test": [ "tests/rules/test_git_push.py" ], "config": [], "asset": [] }
1
nvbn
thefuck
86efc6a252c39526c643ea3335db02c4621798e9
https://github.com/nvbn/thefuck/issues/718
next release
Cannot run fuck in zsh without export errors
Installed via pip. Ran eval "$(thefuck --alias)" in both bash and zsh. $ fuck then produces 'No fucks given' in bash while in zsh it complains about invalid arguments. ``` $ eval "$(thefuck --alias)" $ fuck fuck:export:3: not valid in this context: -' fuck:3: not an identifier: [skip ``` Also FWIW: ``` $ zsh --version zsh 4.3.11 (x86_64-redhat-linux-gnu) $ cat /etc/redhat-release Red Hat Enterprise Linux Server release 6.8 (Santiago) ```
null
https://github.com/nvbn/thefuck/pull/820
null
{'base_commit': '86efc6a252c39526c643ea3335db02c4621798e9', 'files': [{'path': 'thefuck/shells/zsh.py', 'status': 'modified', 'Loc': {"('Zsh', 'app_alias', 12)": {'mod': [19, 20]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "thefuck/shells/zsh.py" ], "doc": [], "test": [], "config": [], "asset": [] }
1
nvbn
thefuck
f700b23f5725e14f8ee6ffb0a5c44ab9eaf42b53
https://github.com/nvbn/thefuck/issues/717
alias produces invalid code for fish as of 3.24
As of 3.24 the output from `thefuck --alias` is incompatible with fish **Output from 3.24** ``` $ thefuck --alias function fuck () { TF_PYTHONIOENCODING=$PYTHONIOENCODING; export TF_ALIAS=fuck; export TF_SHELL_ALIASES=$(alias); export TF_HISTORY=$(fc -ln -10); export PYTHONIOENCODING=utf-8; TF_CMD=$( thefuck THEFUCK_ARGUMENT_PLACEHOLDER $@ ) && eval $TF_CMD; unset TF_HISTORY; export PYTHONIOENCODING=$TF_PYTHONIOENCODING; history -s $TF_CMD; } $ thefuck --version The Fuck 3.24 using Python 3.5.2 ``` **Output from 3.23** ``` $ thefuck --alias function fuck -d "Correct your previous console command" set -l fucked_up_command $history[1] env TF_ALIAS=fuck PYTHONIOENCODING=utf-8 thefuck $fucked_up_command | read -l unfucked_command if [ "$unfucked_command" != "" ] eval $unfucked_command builtin history delete --exact --case-sensitive -- $fucked_up_command builtin history merge ^ /dev/null end end $ thefuck --version The Fuck 3.23 using Python 3.5.2 ``` **Fish version output** ``` $ fish --version fish, version 2.6.0 ```
null
https://github.com/nvbn/thefuck/pull/744
null
{'base_commit': 'f700b23f5725e14f8ee6ffb0a5c44ab9eaf42b53', 'files': [{'path': 'tests/shells/test_bash.py', 'status': 'modified', 'Loc': {"('TestBash', 'test_app_alias_variables_correctly_set', 51)": {'add': [53]}}}, {'path': 'tests/shells/test_fish.py', 'status': 'modified', 'Loc': {"('TestFish', 'test_app_alias', 70)": {'add': [73]}}}, {'path': 'tests/shells/test_tcsh.py', 'status': 'modified', 'Loc': {"('TestTcsh', None, 8)": {'add': [46]}}}, {'path': 'tests/shells/test_zsh.py', 'status': 'modified', 'Loc': {"('TestZsh', 'test_app_alias_variables_correctly_set', 51)": {'add': [53]}}}, {'path': 'thefuck/shells/fish.py', 'status': 'modified', 'Loc': {"('Fish', 'app_alias', 28)": {'mod': [38]}}}, {'path': 'thefuck/shells/tcsh.py', 'status': 'modified', 'Loc': {"('Tcsh', 'app_alias', 9)": {'mod': [10]}}}]}
[]
[]
[]
{ "iss_type": "2", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "thefuck/shells/tcsh.py", "thefuck/shells/fish.py" ], "doc": [], "test": [ "tests/shells/test_zsh.py", "tests/shells/test_fish.py", "tests/shells/test_tcsh.py", "tests/shells/test_bash.py" ], "config": [], "asset": [] }
1
nvbn
thefuck
18992f246a84331832c399283ab930408db21d86
https://github.com/nvbn/thefuck/issues/652
git push to branch not of same name: use git's suggestion
I checked out a branch with a different name than the remote counterpart: ``` $ git checkout -b local-branch myupstream/remote-branch ``` When I attempted to push my changes upstream, I got the following error: ``` $ git push -f fatal: The upstream branch of your current branch does not match the name of your current branch. To push to the upstream branch on the remote, use git push myupstream HEAD:remote-branch To push to the branch of the same name on the remote, use git push myupstream local-branch ``` However, thefuck didn't use this information to provide a suggestion: ``` $ fuck git push -d [enter/↑/↓/ctrl+c] ```
null
https://github.com/nvbn/thefuck/pull/703
null
{'base_commit': '18992f246a84331832c399283ab930408db21d86', 'files': [{'path': 'README.md', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [198]}}}]}
[]
[]
[]
{ "iss_type": "4", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [], "doc": [ "README.md" ], "test": [], "config": [], "asset": [] }
1
nvbn
thefuck
f604756cb7639ab3efb1cd50c85e633a18cd9f9d
https://github.com/nvbn/thefuck/issues/486
windows
werid behavior
python 2.7.10 windows powershell when I type: ``` bash > git comit -am 'test' > fuck ``` it gives me something that is completely weird: ``` C:\Python27\lib\site-packages\win_unicode_console\__init__.py:27: RuntimeWarning: sys.stdin.encoding == 'utf-8', whereas sys.stdout.encoding == None, readline hook consumer may assume they are the same readline_hook.enable(use_pyreadline=use_pyreadline) git commit -am 'test' [enter/↑/↓/ctrl+c] ``` and then give an exception: ``` Traceback (most recent call last): File "C:\Python27\lib\runpy.py", line 162, in _run_module_as_main "__main__", fname, loader, pkg_name) File "C:\Python27\lib\runpy.py", line 72, in _run_code exec code in run_globals File "C:\Python27\Scripts\thefuck.exe\__main__.py", line 9, in <module> File "C:\Python27\lib\site-packages\thefuck\main.py", line 80, in main fix_command() File "C:\Python27\lib\site-packages\thefuck\main.py", line 32, in fix_command selected_command = select_command(corrected_commands) File "C:\Python27\lib\site-packages\thefuck\ui.py", line 80, in select_command for action in read_actions(): File "C:\Python27\lib\site-packages\thefuck\ui.py", line 13, in read_actions key = get_key() File "C:\Python27\lib\site-packages\thefuck\system\win32.py", line 25, in get_key return ch.decode(sys.stdout.encoding) TypeError: decode() argument 1 must be string, not None ``` I think this is related to UTF-8 encoding in python2... I cannot add python3 to path, because my vim plugins will require python2 in the path... Is there any way I can fix this?
null
https://github.com/nvbn/thefuck/pull/487
null
{'base_commit': 'f604756cb7639ab3efb1cd50c85e633a18cd9f9d', 'files': [{'path': 'thefuck/system/win32.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [0]}, "(None, 'get_key', 13)": {'mod': [25]}}}]}
[]
[]
[]
{ "iss_type": "3", "iss_reason": "2", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "thefuck/system/win32.py" ], "doc": [], "test": [], "config": [], "asset": [] }
1
nvbn
thefuck
926e9ef963464075184f5c2e04ffdce9cb971997
https://github.com/nvbn/thefuck/issues/809
windows
Powershell, The term 'thefuck' is not recognized as the name of a cmdlet
**The output of `thefuck --version` (something like `The Fuck 3.1 using Python 3.5.0`):** The Fuck 3.26 using Python 3.6.5 **Your shell and its version (`bash`, `zsh`, *Windows PowerShell*, etc.):** Windows Powershell, 5.1.16299.251 **Your system (Debian 7, ArchLinux, Windows, etc.):** Windows 10 **How to reproduce the bug:** - install thefuck `py -m pip install thefuck` - put the alias into your powershell profile ``` $env:PYTHONIOENCODING="utf-8" iex "$(thefuck --alias)" ``` - start a new powershell Result: ``` thefuck : The term 'thefuck' is not recognized as the name of a cmdlet, function, script file, or operable program. Check the spelling of the name, or if a path was included, verify that the path is correct and try again. ``` **The output of The Fuck with `THEFUCK_DEBUG=true` exported (typically execute `export THEFUCK_DEBUG=true` in your shell before The Fuck):** Hmm I can't manage to output this. **Anything else you think is relevant:** I also tried it with the following in my profile: ``` ... iex "$(C:\Users\MyUserName\AppData\Local\Programs\Python\Python36-32\Scripts\thefuck.exe --alias)" ``` Then I get a similar error after executing `fuck`: ``` thefuck : The term 'thefuck' is not recognized as the name of a cmdlet, function, script file, or operable program. Check the spelling of the name, or if a path was included, verify that the path is correct and try again. At line:1 char:141 + ... g]::IsNullOrWhiteSpace($history)) { $fuck = $(thefuck $args $ ... + ~~~~~~~ + CategoryInfo : ObjectNotFound: (thefuck:String) [], CommandNotFoundException + FullyQualifiedErrorId : CommandNotFoundException ```
null
https://github.com/nvbn/thefuck/pull/844
null
{'base_commit': '926e9ef963464075184f5c2e04ffdce9cb971997', 'files': [{'path': 'thefuck/shells/__init__.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [19]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "2", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "thefuck/shells/__init__.py" ], "doc": [], "test": [], "config": [], "asset": [] }
1
nvbn
thefuck
18992f246a84331832c399283ab930408db21d86
https://github.com/nvbn/thefuck/issues/658
hacktoberfest
exception while "run fuck second time for configuring it automatically."
``` luca@g550jk ~> fuck Seems like fuck alias isn't configured! Please put eval (thefuck --alias | tr ' ' ';') in your ~/.config/fish/config.fish and apply changes with fish or restart your shell. Or run fuck second time for configuring it automatically. More details - https://github.com/nvbn/thefuck#manual-installation luca@g550jk ~> fuck Traceback (most recent call last): File "/usr/bin/fuck", line 11, in <module> load_entry_point('thefuck==3.18', 'console_scripts', 'fuck')() File "/usr/lib/python3.6/site-packages/thefuck/not_configured.py", line 80, in main elif _is_second_run(): File "/usr/lib/python3.6/site-packages/thefuck/not_configured.py", line 40, in _is_second_run if not tracker_path.exists() or not shell.get_history()[-1] == 'fuck': IndexError: list index out of range luca@g550jk ~> thefuck --version The Fuck 3.18 using Python 3.6.1 luca@g550jk ~> lsb_release -d Description: Arch Linux luca@g550jk ~> echo $SHELL /usr/bin/fish luca@g550jk ~> fish --version fish, version 2.5.0 ```
null
https://github.com/nvbn/thefuck/pull/704
null
{'base_commit': '18992f246a84331832c399283ab930408db21d86', 'files': [{'path': 'thefuck/shells/fish.py', 'status': 'modified', 'Loc': {"('Fish', 'how_to_configure', 72)": {'mod': [74]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "thefuck/shells/fish.py" ], "doc": [], "test": [], "config": [], "asset": [] }
1
nvbn
thefuck
328e65179e8a886b44611850b908109868884c14
https://github.com/nvbn/thefuck/issues/229
Should suggest 'git tag' when typing 'git tags' instead of 'git stage'
``` $ git tags git: 'tags' is not a git command. See 'git --help'. Did you mean one of these? stage tag $ fuck git stage Nothing specified, nothing added. Maybe you wanted to say 'git add .'? ``` Any way to improve the suggestion in this case?
null
https://github.com/nvbn/thefuck/pull/285
null
{'base_commit': '328e65179e8a886b44611850b908109868884c14', 'files': [{'path': 'tests/rules/test_tmux.py', 'status': 'modified', 'Loc': {"(None, 'test_get_new_command', 17)": {'mod': [19]}}}, {'path': 'thefuck/rules/tmux.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [0]}, "(None, 'get_new_command', 10)": {'mod': [11, 14]}}}]}
[]
[]
[]
{ "iss_type": "4", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "thefuck/rules/tmux.py" ], "doc": [], "test": [ "tests/rules/test_tmux.py" ], "config": [], "asset": [] }
1
nvbn
thefuck
470c0ef699f31bba76445f18d7cbb289af388565
https://github.com/nvbn/thefuck/issues/439
Windows - ImportError: No module named dbm
I have installed thefuck with pip using Python 2.7.11 in Windows. When running `thefuck` from Powershell, I get the error message `ImportError: No module named dbm`. Any ideas?
null
https://github.com/nvbn/thefuck/pull/458
null
{'base_commit': '470c0ef699f31bba76445f18d7cbb289af388565', 'files': [{'path': 'thefuck/utils.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [1, 19, 21, 22]}, "(None, '_cache', 211)": {'mod': [230]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "thefuck/utils.py" ], "doc": [], "test": [], "config": [], "asset": [] }
1
nvbn
thefuck
d92765d5df6607cb2f2fb67cee7b63f64ac7aa6b
https://github.com/nvbn/thefuck/issues/301
bug
UnicodeEncodeError on non-english systems
When a command outputs a sequence with non-ascii characters, no fucks are given with a dirty UnicodeEncodeError followed by a grep --compact not recognized option. Example : ``` ~$ aptget blah La commande « aptget » est introuvable, vouliez-vous dire : La commande « apt-get » du paquet « apt » (main) aptget : commande introuvable ~$ fuck apt-get blah [enter/ctrl+c] E: L'opération blah n'est pas valable ~$ fuck Traceback (most recent call last): File "/usr/local/lib/python2.7/dist-packages/thefuck-2.1-py2.7.egg/thefuck/main.py", line 100, in get_matched_rule if rule.match(command, settings): File "/usr/local/lib/python2.7/dist-packages/thefuck-2.1-py2.7.egg/thefuck/rules/history.py", line 28, in match _history_of_exists_without_current(command))) File "/usr/local/lib/python2.7/dist-packages/thefuck-2.1-py2.7.egg/thefuck/utils.py", line 109, in wrapper memo[key] = fn(*args, **kwargs) File "/usr/local/lib/python2.7/dist-packages/thefuck-2.1-py2.7.egg/thefuck/rules/history.py", line 19, in _history_of_exists_without_current history = get_history() File "/usr/local/lib/python2.7/dist-packages/thefuck-2.1-py2.7.egg/thefuck/utils.py", line 109, in wrapper memo[key] = fn(*args, **kwargs) File "/usr/local/lib/python2.7/dist-packages/thefuck-2.1-py2.7.egg/thefuck/shells.py", line 267, in get_history return list(_get_shell().get_history()) File "/usr/local/lib/python2.7/dist-packages/thefuck-2.1-py2.7.egg/thefuck/shells.py", line 67, in get_history prepared = self._script_from_history(line)\ File "/usr/local/lib/python2.7/dist-packages/thefuck-2.1-py2.7.egg/thefuck/shells.py", line 103, in _script_from_history print(line) UnicodeEncodeError: 'ascii' codec can't encode character u'\xe9' in position 4: ordinal not in range(128) ---------------------------- No fuck given grep : option non reconnue « --compact » Utilisation : grep [OPTION]... MOTIF [FICHIER]... Exécutez « grep --help » pour obtenir des renseignements complémentaires. sync: ignorer tous les arguments ```
null
https://github.com/nvbn/thefuck/pull/458
null
{'base_commit': '470c0ef699f31bba76445f18d7cbb289af388565', 'files': [{'path': 'thefuck/utils.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [1, 19, 21, 22]}, "(None, '_cache', 211)": {'mod': [230]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "thefuck/utils.py" ], "doc": [], "test": [], "config": [], "asset": [] }
1
home-assistant
core
7d71a2c979bc9ae300e177cbedb059a6133174b1
https://github.com/home-assistant/core/issues/34897
integration: homekit
homekit integration: work but error in log
<!-- READ THIS FIRST: - If you need additional help with this template, please refer to https://www.home-assistant.io/help/reporting_issues/ - Make sure you are running the latest version of Home Assistant before reporting an issue: https://github.com/home-assistant/core/releases - Do not report issues for integrations if you are using custom components or integrations. - Provide as many details as possible. Paste logs, configuration samples and code into the backticks. DO NOT DELETE ANY TEXT from this template! Otherwise, your issue may be closed without comment. --> ## The problem <!-- Describe the issue you are experiencing here to communicate to the maintainers. Tell us what you were trying to do and what happened. -->No problem for functionality but a lot of same error in log ## Environment <!-- Provide details about the versions you are using, which helps us to reproduce and find the issue quicker. Version information is found in the Home Assistant frontend: Developer tools -> Info. -->0.109 - Home Assistant Core release with the issue: 0.109 - Last working Home Assistant Core release (if known): 0.108.9 - Operating environment (Home Assistant/Supervised/Docker/venv): Home Assistant in venv - Integration causing this issue: homekit - Link to integration documentation on our website: integration:homekit ## Problem-relevant `configuration.yaml` <!-- An example configuration that caused the problem for you. Fill this out even if it seems unimportant to you. Please be sure to remove personal information like passwords, private URLs and other credentials. --> ``` homekit: filter: exclude_domains: - automation - device_tracker - person - script - group exclude_entities: - climate.netatmo_ingresso - climate.netatmo_cucina - binary_sensor.telecomando_a - binary_sensor.telecomando_b - binary_sensor.telecomando_c - binary_sensor.telecomando_d - light.led_ingresso2 - light.led_ingresso3 - light.led_ingresso4 - light.led_ingresso5 - switch.ovunque_do_not_disturb_switch - switch.ovunque_repeat_switch - switch.altoparlante_do_not_disturb_switch - switch.davide_s_reverb_ai_do_not_disturb_switch - switch.echo_studio_repeat_switch - switch.dappertutto_do_not_disturb_switch - switch.echo_dot_ingresso_repeat_switch - switch.davide_s_reverb_ai_shuffle_switch - switch.dappertutto_repeat_switch - switch.this_device_do_not_disturb_switch - switch.altoparlante_repeat_switch - switch.davide_s_alexa_apps_do_not_disturb_switch - switch.echo_dot_camera_letto_do_not_disturb_switch - switch.echo_dot_camera_letto_shuffle_switch - switch.echo_dot_ingresso_do_not_disturb_switch - switch.echo_show_shuffle_switch - switch.davide_s_reverb_ai_repeat_switch - switch.echo_dot_camera_letto_repeat_switch - switch.echo_dot_ingresso_shuffle_switch - switch.altoparlante_shuffle_switch - switch.dappertutto_shuffle_switch - switch.ovunque_shuffle_switch - switch.echo_show_repeat_switch - switch.echo_studio_shuffle_switch - switch.echo_show_do_not_disturb_switch - switch.echo_studio_do_not_disturb_switch - binary_sensor.updater entity_config: media_player.lg_tv: name: TV ``` ## Traceback/Error logs <!-- If you come across any trace or error logs, please provide them. --> ``` Apr 29 23:03:37 raspberrypi hass[6324]: 2020-04-29 23:03:37 ERROR (MainThread) [homeassistant.core] Error doing job: Future exception was never retrieved Apr 29 23:03:37 raspberrypi hass[6324]: Traceback (most recent call last): Apr 29 23:03:37 raspberrypi hass[6324]: File "/usr/lib/python3.7/concurrent/futures/thread.py", line 57, in run Apr 29 23:03:37 raspberrypi hass[6324]: result = self.fn(*self.args, **self.kwargs) Apr 29 23:03:37 raspberrypi hass[6324]: File "/srv/homeassistant/lib/python3.7/site-packages/homeassistant/components/homekit/accessories.py", line 252, in update_battery Apr 29 23:03:37 raspberrypi hass[6324]: if self._char_battery.value != battery_level: Apr 29 23:03:37 raspberrypi hass[6324]: AttributeError: 'TemperatureSensor' object has no attribute '_char_battery' ``` ## Additional information
null
https://github.com/home-assistant/core/pull/34906
null
{'base_commit': '7d71a2c979bc9ae300e177cbedb059a6133174b1', 'files': [{'path': 'homeassistant/components/homekit/accessories.py', 'status': 'modified', 'Loc': {"('HomeAccessory', '__init__', 85)": {'add': [101]}, "('HomeAccessory', 'update_battery', 245)": {'add': [249], 'mod': [261]}}}, {'path': 'tests/components/homekit/test_accessories.py', 'status': 'modified', 'Loc': {"(None, 'test_missing_linked_battery_sensor', 344)": {'mod': [366, 367, 368]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "homeassistant/components/homekit/accessories.py" ], "doc": [], "test": [ "tests/components/homekit/test_accessories.py" ], "config": [], "asset": [] }
1
home-assistant
core
acf41d03db6293b9897b8194f5530793de89952f
https://github.com/home-assistant/core/issues/26146
integration: mopar
mopar component still does not work
Home Assistant release with the issue: 0.97.2 Last working Home Assistant release (if known): Operating environment (Hass.io/Docker/Windows/etc.): arch | x86_64 dev | false docker | true hassio | true os_name | Linux python_version | 3.7.4 timezone | America/Denver version | 0.97.2 virtualenv | false **Description of problem:** Componets show up but nothing works. ``` 'utf-8' codec can't decode byte 0x80 in position 0: invalid start byte ``` **Problem-relevant `configuration.yaml` entries and (fill out even if it seems unimportant):** ```yaml mopar: username: !secret mopar_username password: !secret mopar_password pin: !secret mopar_pin ``` **Traceback (if applicable):** ``` File "/usr/local/lib/python3.7/site-packages/requests/models.py", line 897, in json return complexjson.loads(self.text, **kwargs) File "/usr/local/lib/python3.7/site-packages/simplejson/__init__.py", line 518, in loads return _default_decoder.decode(s) File "/usr/local/lib/python3.7/site-packages/simplejson/decoder.py", line 370, in decode obj, end = self.raw_decode(s) File "/usr/local/lib/python3.7/site-packages/simplejson/decoder.py", line 400, in raw_decode return self.scan_once(s, idx=_w(s, idx).end()) simplejson.errors.JSONDecodeError: Expecting value: line 1 column 1 (char 0) ``` **Additional information:**
null
https://github.com/home-assistant/core/pull/33066
null
{'base_commit': 'acf41d03db6293b9897b8194f5530793de89952f', 'files': [{'path': '.coveragerc', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [435]}}}, {'path': 'homeassistant/components/mopar/__init__.py', 'status': 'removed', 'Loc': {}}, {'path': 'homeassistant/components/mopar/lock.py', 'status': 'removed', 'Loc': {}}, {'path': 'homeassistant/components/mopar/manifest.json', 'status': 'removed', 'Loc': {}}, {'path': 'homeassistant/components/mopar/sensor.py', 'status': 'removed', 'Loc': {}}, {'path': 'homeassistant/components/mopar/services.yaml', 'status': 'removed', 'Loc': {}}, {'path': 'homeassistant/components/mopar/switch.py', 'status': 'removed', 'Loc': {}}, {'path': 'requirements_all.txt', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [885, 886]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "homeassistant/components/mopar/sensor.py", ".coveragerc", "homeassistant/components/mopar/manifest.json", "homeassistant/components/mopar/__init__.py", "homeassistant/components/mopar/lock.py", "homeassistant/components/mopar/switch.py" ], "doc": [], "test": [], "config": [ "homeassistant/components/mopar/services.yaml", "requirements_all.txt" ], "asset": [] }
1
home-assistant
core
f020d65416177e4647a846ec017f441fe08c0696
https://github.com/home-assistant/core/issues/54258
integration: fronius energy
Fronius energy sensors missing device_class and last_reset
### The problem The Fronius integration with `sensor_type: inverter` has energy sensors, but these are not marked with any device_class or state_class and there's no last_reset. eg the sensor `sensor.energy_day_fronius_inverter_1_http_solar` value `12455` and only the following attributes: ``` unit_of_measurement: Wh friendly_name: Energy day Fronius Inverter 1 http://solar ``` The energy sensor from `sensor_type: power_flow` has the state_class and device_class but no last_reset. eg `sensor.energy_day_fronius_power_flow_0_http_solar`. Also the device_class is power but should be energy. ``` state_class: measurement unit_of_measurement: Wh friendly_name: Energy day Fronius Power flow 0 http://solar device_class: power ``` The total energy from the inverter sensor `sensor.energy_total_fronius_inverter_1_http_solar` is correct though. ### What is version of Home Assistant Core has the issue? core-2021.8.3 ### What was the last working version of Home Assistant Core? _No response_ ### What type of installation are you running? Home Assistant Container ### Integration causing the issue fronius ### Link to integration documentation on our website _No response_ ### Example YAML snippet ```yaml sensor: - platform: fronius resource: http://solar monitored_conditions: - sensor_type: power_flow - sensor_type: meter - sensor_type: inverter ``` ``` ### Anything in the logs that might be useful for us? _No response_ ### Additional information _No response_
null
https://github.com/home-assistant/core/pull/54758
null
{'base_commit': '013b998974c889c8a80d04637a9ea8e43c7e2fc3', 'files': [{'path': 'homeassistant/components/fronius/manifest.json', 'status': 'modified', 'Loc': {'(None, None, 5)': {'mod': [5]}}}, {'path': 'requirements_all.txt', 'status': 'modified', 'Loc': {'(None, None, 1469)': {'mod': [1469]}}}]}
[]
[]
[]
{ "iss_type": "4", "iss_reason": "2", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "homeassistant/components/fronius/manifest.json" ], "doc": [], "test": [], "config": [ "requirements_all.txt" ], "asset": [] }
1
home-assistant
core
832bc15daacfe1294538bb4a17600e4ada36a47e
https://github.com/home-assistant/core/issues/83076
integration: twinkly
Bug: Twinkly unavailable
### The problem Twinkly device unavailable after update to latest beta. ### What version of Home Assistant Core has the issue? core-2022.12.0b0 ### What was the last working version of Home Assistant Core? core-2022.11.5 ### What type of installation are you running? Home Assistant Container ### Integration causing the issue twinkly ### Link to integration documentation on our website https://www.home-assistant.io/integrations/twinkly/ ### Diagnostics information n/a ### Example YAML snippet ```yaml n/a ``` ### Anything in the logs that might be useful for us? ```txt 2022-12-01 14:11:15.238 DEBUG (MainThread) [ttls.client] Authentication token refreshed 2022-12-01 14:11:15.239 DEBUG (MainThread) [ttls.client] GET endpoint gestalt 2022-12-01 14:11:15.264 DEBUG (MainThread) [ttls.client] GET response 200 2022-12-01 14:11:15.265 DEBUG (MainThread) [homeassistant.components.twinkly.light] Updating 'Twinkly_6A5D49.collective.lan' 2022-12-01 14:11:15.265 DEBUG (MainThread) [ttls.client] Authentication token still valid 2022-12-01 14:11:15.265 DEBUG (MainThread) [ttls.client] GET endpoint led/mode 2022-12-01 14:11:15.290 DEBUG (MainThread) [ttls.client] GET response 200 2022-12-01 14:11:15.290 DEBUG (MainThread) [ttls.client] Authentication token still valid 2022-12-01 14:11:15.290 DEBUG (MainThread) [ttls.client] GET endpoint led/out/brightness 2022-12-01 14:11:15.358 DEBUG (MainThread) [ttls.client] GET response 200 2022-12-01 14:11:15.359 DEBUG (MainThread) [ttls.client] Authentication token still valid 2022-12-01 14:11:15.359 DEBUG (MainThread) [ttls.client] GET endpoint gestalt 2022-12-01 14:11:15.374 DEBUG (MainThread) [ttls.client] GET response 200 2022-12-01 14:11:15.374 DEBUG (MainThread) [ttls.client] Authentication token still valid 2022-12-01 14:11:15.374 DEBUG (MainThread) [ttls.client] GET endpoint movies ``` ### Additional information The device reports `unavailable`. I am able to control it via the Twinkly app.
null
https://github.com/home-assistant/core/pull/83145
null
{'base_commit': '832bc15daacfe1294538bb4a17600e4ada36a47e', 'files': [{'path': 'homeassistant/components/twinkly/const.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [11, 29]}}}, {'path': 'homeassistant/components/twinkly/light.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [9, 27, 39]}, "('TwinklyLight', '__init__', 63)": {'add': [98]}, "('TwinklyLight', 'device_info', 126)": {'add': [132]}, "('TwinklyLight', None, 60)": {'add': [167], 'mod': [135, 136, 137, 138]}, "('TwinklyLight', 'async_turn_on', 168)": {'mod': [181, 182, 183, 185, 187, 188, 189, 190, 191, 192, 193, 194, 196, 197, 199, 200, 201, 203, 205, 206, 207, 208, 210]}, "('TwinklyLight', 'async_update', 226)": {'mod': [271, 272]}}}, {'path': 'tests/components/twinkly/__init__.py', 'status': 'modified', 'Loc': {"('ClientMock', '__init__', 19)": {'add': [27]}, "('ClientMock', 'turn_on', 53)": {'add': [57]}, "('ClientMock', 'set_static_colour', 81)": {'add': [83]}, "('ClientMock', 'set_mode', 100)": {'mod': [103, 105]}}}, {'path': 'tests/components/twinkly/test_light.py', 'status': 'modified', 'Loc': {"(None, 'test_turn_on_with_brightness', 68)": {'add': [80, 92], 'mod': [82, 94]}, "(None, 'test_turn_on_with_color_rgbw', 101)": {'add': [109, 114, 122], 'mod': [102, 116]}, "(None, 'test_turn_on_with_color_rgb', 125)": {'add': [133, 138, 146], 'mod': [126, 140]}, "(None, 'test_turn_on_with_effect', 149)": {'add': [163, 171], 'mod': [150, 158, 165]}, '(None, None, None)': {'mod': [6]}, "(None, 'test_turn_on_off', 48)": {'mod': [58, 60]}, "(None, 'test_turn_off', 174)": {'mod': [181, 183]}, "(None, 'test_update_name', 190)": {'mod': [202, 204]}}}]}
[]
[]
[]
{ "iss_type": "2", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "homeassistant/components/twinkly/light.py", "tests/components/twinkly/__init__.py", "homeassistant/components/twinkly/const.py" ], "doc": [], "test": [ "tests/components/twinkly/test_light.py" ], "config": [], "asset": [] }
1
home-assistant
core
20f94d7ad47082ec2f1c204130abe318c4b014dd
https://github.com/home-assistant/core/issues/54678
integration: myq
MyQ Login Issue
### The problem Myq fails to initialize with "Authentication failed: Form containing fields for email, password and submit not found.Unable to continue login process.". Issue started happening today for me. There are more people reporting it here: https://community.home-assistant.io/t/myq-login-unavailable/330322 ### What is version of Home Assistant Core has the issue? 2021.7.4 ### What was the last working version of Home Assistant Core? _No response_ ### What type of installation are you running? Home Assistant Container ### Integration causing the issue myq ### Link to integration documentation on our website https://www.home-assistant.io/integrations/myq/ ### Example YAML snippet _No response_ ### Anything in the logs that might be useful for us? ```txt Logger: pymyq.api Source: /usr/local/lib/python3.9/site-packages/pymyq/api.py:701 First occurred: 9:24:30 PM (13 occurrences) Last logged: 9:37:04 PM Authentication failed: Form containing fields for email, password and submit not found.Unable to continue login process. ``` ### Additional information _No response_
null
https://github.com/home-assistant/core/pull/55099
null
{'base_commit': '20f94d7ad47082ec2f1c204130abe318c4b014dd', 'files': [{'path': 'homeassistant/components/myq/manifest.json', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [5]}}}, {'path': 'requirements_all.txt', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [1626]}}}, {'path': 'requirements_test_all.txt', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [939]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "homeassistant/components/myq/manifest.json" ], "doc": [], "test": [], "config": [ "requirements_test_all.txt", "requirements_all.txt" ], "asset": [] }
1
home-assistant
core
b960ebeb8bbf791a8072e8503b417474fec134ac
https://github.com/home-assistant/core/issues/2290
Smoke detector are not changing from "Panic" state
Make sure you are running the latest version of Home Assistant before reporting an issue. You should only file an issue if you found a bug. Feature and enhancement requests should go in [the Feature Requests section](https://community.home-assistant.io/c/feature-requests) of our community forum: **Home Assistant release (`hass --version`): 0.21.1 **Component/platform: Rfxtrx **Description of problem: First time I press the test button on my Nexa KD101 smoke detector it appears in HASS UI, but after that the UI is never changed, It keep showing panic and nothing happen in the UI after first test. The log are showing the smoke detector signal, so it is received. ![image](https://cloud.githubusercontent.com/assets/7728206/16024413/5e301276-31ca-11e6-824a-b1020fc66239.png) ![image](https://cloud.githubusercontent.com/assets/7728206/16024419/61d9ef78-31ca-11e6-9f31-20d61aeb7ffb.png)
null
https://github.com/home-assistant/core/pull/2498
null
{'base_commit': '09a4336bc5d9a2f01f475b786492bb517ca36b5f', 'files': [{'path': 'homeassistant/components/sensor/rfxtrx.py', 'status': 'modified', 'Loc': {"(None, 'setup_platform', 29)": {'mod': [50]}, "('RfxtrxSensor', 'state', 111)": {'mod': [113]}, "('RfxtrxSensor', 'device_state_attributes', 123)": {'mod': [125]}}}, {'path': 'tests/components/sensor/test_rfxtrx.py', 'status': 'modified', 'Loc': {"('TestSensorRfxtrx', 'test_several_sensors', 93)": {'add': [122, 135], 'mod': [115, 116, 117, 118, 119, 120, 125, 126, 127, 128, 129, 130, 137, 138, 139, 140, 141, 142, 143]}, "('TestSensorRfxtrx', 'test_update_of_sensors', 233)": {'add': [262, 275], 'mod': [255, 256, 257, 258, 259, 260, 265, 266, 267, 268, 269, 270, 277, 278, 279, 280, 281, 282, 283]}, "('TestSensorRfxtrx', 'test_old_config_sensor', 33)": {'mod': [47, 48, 49, 50, 51, 52]}, "('TestSensorRfxtrx', 'test_one_sensor', 54)": {'mod': [67, 68, 69, 70, 71, 72]}, "('TestSensorRfxtrx', 'test_one_sensor_no_datatype', 74)": {'mod': [86, 87, 88, 89, 90, 91]}}}]}
[]
[]
[]
{ "iss_type": "2", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "homeassistant/components/sensor/rfxtrx.py" ], "doc": [], "test": [ "tests/components/sensor/test_rfxtrx.py" ], "config": [], "asset": [] }
1
home-assistant
core
d9e3c02df3a2690e74d1b606e8db0a4dd686e872
https://github.com/home-assistant/core/issues/79275
integration: unifi
Ubiquiti POE toggle (off) : aiounifi.models.event:Unsupported event key EVT_SW_PoeDisconnect
### The problem as title, toggling a POE off throws the Warning they do toggle in the end, although it can take up to 45 secs or so. which is the same behavior we see on the camera's issue, extremely slow response times on the toggle ### What version of Home Assistant Core has the issue? 2022.10.0b0 ### What was the last working version of Home Assistant Core? _No response_ ### What type of installation are you running? Home Assistant OS ### Integration causing the issue UniFi Network ### Link to integration documentation on our website https://www.home-assistant.io/integrations/unifi/ ### Diagnostics information _No response_ ### Example YAML snippet _No response_ ### Anything in the logs that might be useful for us? ```txt 2022-09-29 09:31:08.184 WARNING (MainThread) [aiounifi.models.event] Unsupported event key EVT_SW_PoeDisconnect 2022-09-29 09:31:08.190 WARNING (MainThread) [aiounifi.models.event] Unsupported event key EVT_SW_PoeDisconnect 2022-09-29 09:31:08.192 WARNING (MainThread) [aiounifi.models.event] Unsupported event key EVT_SW_PoeDisconnect 2022-09-29 09:31:08.196 WARNING (MainThread) [aiounifi.models.event] Unsupported event key EVT_SW_PoeDisconnect ``` or: ``` Logger: aiounifi.models.event Source: components/unifi/controller.py:232 First occurred: 09:31:08 (1 occurrences) Last logged: 09:31:08 Unsupported event key EVT_SW_PoeDisconnect ``` ### Additional information these are the bt proxies I have in the integration: <img width="811" alt="Schermafbeelding 2022-09-29 om 09 44 02" src="https://user-images.githubusercontent.com/33354141/192971228-1269e885-6e6e-499b-8261-734ce122fcec.png"> have [another issue on POE switching](https://github.com/home-assistant/core/issues/79088), Ubiquiti Cams not responding correctly), not sure if its related
null
https://github.com/home-assistant/core/pull/79487
null
{'base_commit': '069818940e5a16c412544e1e14e69b5d0964f157', 'files': [{'path': 'homeassistant/components/unifi/manifest.json', 'status': 'modified', 'Loc': {'(None, None, 6)': {'mod': [6]}}}, {'path': 'requirements_all.txt', 'status': 'modified', 'Loc': {'(None, None, 279)': {'mod': [279]}}}, {'path': 'requirements_test_all.txt', 'status': 'modified', 'Loc': {'(None, None, 254)': {'mod': [254]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "homeassistant/components/unifi/manifest.json" ], "doc": [], "test": [], "config": [ "requirements_test_all.txt", "requirements_all.txt" ], "asset": [] }
1
home-assistant
core
b71a0c5d4bb22f1d7ec84e892cff851ad1d6f283
https://github.com/home-assistant/core/issues/54573
integration: ambiclimate
Ambi Climate Integration not listing any Entities
### The problem The ambiclimate integration is appearing in the Integration Screen but not showing any entities even though 1 is available and was working correctly in the previous version ### What is version of Home Assistant Core has the issue? Home Assistant 2021.8.6 ### What was the last working version of Home Assistant Core? Home Assistant 2021.8.1 ### What type of installation are you running? Home Assistant Core ### Integration causing the issue Ambiclimate ### Link to integration documentation on our website https://www.home-assistant.io/integrations/ambiclimate/ ### Example YAML snippet _No response_ ### Anything in the logs that might be useful for us? ```txt Logger: homeassistant.components.climate Source: components/ambiclimate/climate.py:157 Integration: Climate (documentation, issues) First occurred: 2:32:41 PM (1 occurrences) Last logged: 2:32:41 PM Error while setting up ambiclimate platform for climate Traceback (most recent call last): File "/usr/src/homeassistant/homeassistant/helpers/entity_platform.py", line 249, in _async_setup_platform await asyncio.shield(task) File "/usr/src/homeassistant/homeassistant/components/ambiclimate/climate.py", line 94, in async_setup_entry devs.append(AmbiclimateEntity(heater, store)) File "/usr/src/homeassistant/homeassistant/components/ambiclimate/climate.py", line 157, in __init__ self._attr_min_temp = heater.get_min_temp() File "/usr/local/lib/python3.9/site-packages/ambiclimate/__init__.py", line 251, in get_min_temp data = self.ir_features['data'][self.ac_data[0].get('mode').lower()]['temperature']['value'] TypeError: 'NoneType' object is not subscriptable ``` ### Additional information _No response_
null
https://github.com/home-assistant/core/pull/54579
null
{'base_commit': 'b71a0c5d4bb22f1d7ec84e892cff851ad1d6f283', 'files': [{'path': 'homeassistant/components/ambiclimate/climate.py', 'status': 'modified', 'Loc': {"('AmbiclimateEntity', 'async_update', 175)": {'add': [186]}, "('AmbiclimateEntity', '__init__', 146)": {'mod': [157, 158]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "homeassistant/components/ambiclimate/climate.py" ], "doc": [], "test": [], "config": [], "asset": [] }
1
home-assistant
core
f18c6ae72ccf9075c3f5442b1aff2381379ca9d6
https://github.com/home-assistant/core/issues/39472
integration: arlo
arlo integration is broken
<!-- READ THIS FIRST: - If you need additional help with this template, please refer to https://www.home-assistant.io/help/reporting_issues/ - Make sure you are running the latest version of Home Assistant before reporting an issue: https://github.com/home-assistant/core/releases - Do not report issues for integrations if you are using custom components or integrations. - Provide as many details as possible. Paste logs, configuration samples and code into the backticks. DO NOT DELETE ANY TEXT from this template! Otherwise, your issue may be closed without comment. --> ## The problem <!-- Describe the issue you are experiencing here to communicate to the maintainers. Tell us what you were trying to do and what happened. --> `arlo` integration fails to initialize with message ```txt ERROR (MainThread) [homeassistant.setup] Setup failed for arlo: Integration failed to initialize. ``` ## Environment <!-- Provide details about the versions you are using, which helps us to reproduce and find the issue quicker. Version information is found in the Home Assistant frontend: Configuration -> Info. --> - Home Assistant Core release with the issue: 0.114.1 - Last working Home Assistant Core release (if known): N/A - Operating environment (OS/Container/Supervised/Core): Container - Integration causing this issue: arlo - Link to integration documentation on our website: https://www.home-assistant.io/integrations/arlo/ ## Problem-relevant `configuration.yaml` <!-- An example configuration that caused the problem for you. Fill this out even if it seems unimportant to you. Please be sure to remove personal information like passwords, private URLs and other credentials. --> ```yaml arlo: !include arlo.yaml ``` `arlo.yaml` ```yaml username: !secret arlo_username password: !secret arlo_password ``` `secrets.yaml` ```yaml arlo_username: REDACTED arlo_password: REDACTED ``` ## Traceback/Error logs <!-- If you come across any trace or error logs, please provide them. --> ```txt > grep arlo home-assistant.log 2020-08-31 18:29:28 INFO (MainThread) [homeassistant.bootstrap] Setting up stage 2: {'noonlight', 'met', 'sun', 'mobile_app', 'input_datetime', 'hue', 'ffmpeg', 'default_config', 'map', 'input_text', 'alarm_control_panel', 'notify', 'input_number', 'arlo', 'scene', 'system_health', 'input_boolean', 'zeroconf', 'weather', 'light', 'automation', 'zha', 'updater', 'tts', 'ssdp', 'input_select', 'history', 'zone', 'logbook', 'zwave', 'group', 'script'} 2020-08-31 18:29:29 DEBUG (MainThread) [homeassistant.setup] Dependency arlo will wait for ['ffmpeg'] 2020-08-31 18:29:34 INFO (MainThread) [homeassistant.setup] Setting up arlo 2020-08-31 18:29:34 DEBUG (SyncWorker_6) [pyarlo] Creating Arlo session 2020-08-31 18:29:34 DEBUG (SyncWorker_6) [pyarlo] Params: {'email': 'REDACTED', 'password': 'REDACTED'} 2020-08-31 18:29:34 DEBUG (SyncWorker_6) [pyarlo] Headers: {'Content-Type': 'application/json', 'Authorization': None} 2020-08-31 18:29:34 DEBUG (SyncWorker_6) [pyarlo] Querying https://arlo.netgear.com/hmsweb/login/v2 on attempt: 0/3 2020-08-31 18:29:34 DEBUG (SyncWorker_6) [pyarlo] Params: {'email': 'REDACTED', 'password': 'REDACTED'} 2020-08-31 18:29:34 DEBUG (SyncWorker_6) [pyarlo] Headers: {'Content-Type': 'application/json', 'Authorization': None} 2020-08-31 18:29:34 DEBUG (SyncWorker_6) [pyarlo] Querying https://arlo.netgear.com/hmsweb/login/v2 on attempt: 1/3 2020-08-31 18:29:34 DEBUG (SyncWorker_6) [pyarlo] Params: {'email': 'REDACTED', 'password': 'REDACTED'} 2020-08-31 18:29:34 DEBUG (SyncWorker_6) [pyarlo] Headers: {'Content-Type': 'application/json', 'Authorization': None} 2020-08-31 18:29:34 DEBUG (SyncWorker_6) [pyarlo] Querying https://arlo.netgear.com/hmsweb/login/v2 on attempt: 2/3 2020-08-31 18:29:35 DEBUG (SyncWorker_6) [pyarlo] Params: {'email': 'REDACTED', 'password': 'REDACTED'} 2020-08-31 18:29:35 DEBUG (SyncWorker_6) [pyarlo] Headers: {'Content-Type': 'application/json', 'Authorization': None} 2020-08-31 18:29:35 DEBUG (SyncWorker_6) [pyarlo] Querying https://arlo.netgear.com/hmsweb/login/v2 on attempt: 3/3 2020-08-31 18:29:35 INFO (MainThread) [homeassistant.setup] Setup of domain arlo took 0.6 seconds 2020-08-31 18:29:35 ERROR (MainThread) [homeassistant.setup] Setup failed for arlo: Integration failed to initialize. ``` ## Additional information I have tested the exact same configuration using https://github.com/twrecked/hass-aarlo instead (by renaming config references from `arlo` to `aarlo`) and the integration initializes without issue. My Arlo account does not have 2FA enabled.
null
https://github.com/home-assistant/core/pull/44034
null
{'base_commit': 'f18c6ae72ccf9075c3f5442b1aff2381379ca9d6', 'files': [{'path': 'homeassistant/components/arlo/manifest.json', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [5]}}}, {'path': 'requirements_all.txt', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [1274]}}}, {'path': 'requirements_test_all.txt', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [643]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "homeassistant/components/arlo/manifest.json" ], "doc": [], "test": [], "config": [ "requirements_test_all.txt", "requirements_all.txt" ], "asset": [] }
1
home-assistant
core
91962e2681dde1b23612df06633b16aa0867c950
https://github.com/home-assistant/core/issues/14989
Mysensor HVAC frontend shows multiple values
I think since 0.6x mysensors hvac node shows several temp settings ![image](https://user-images.githubusercontent.com/15857592/41496413-4743c5d4-7148-11e8-96c4-ae99923dc6e8.png)
null
https://github.com/home-assistant/core/pull/15110
null
{'base_commit': '91962e2681dde1b23612df06633b16aa0867c950', 'files': [{'path': 'homeassistant/components/climate/mysensors.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [29, 30, 31]}, "('MySensorsHVAC', 'supported_features', 46)": {'mod': [48]}, "('MySensorsHVAC', 'operation_list', 104)": {'mod': [106]}, "('MySensorsHVAC', 'fan_list', 114)": {'mod': [116]}}}]}
[]
[]
[]
{ "iss_type": "2", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "homeassistant/components/climate/mysensors.py" ], "doc": [], "test": [], "config": [], "asset": [] }
1
home-assistant
core
62054b84336e4d4a8a6a1ea649724e9eb86876b4
https://github.com/home-assistant/core/issues/35779
integration: denon
Error while setting up denon platform for media_player with Marantz
<!-- READ THIS FIRST: - If you need additional help with this template, please refer to https://www.home-assistant.io/help/reporting_issues/ - Make sure you are running the latest version of Home Assistant before reporting an issue: https://github.com/home-assistant/core/releases - Do not report issues for integrations if you are using custom components or integrations. - Provide as many details as possible. Paste logs, configuration samples and code into the backticks. DO NOT DELETE ANY TEXT from this template! Otherwise, your issue may be closed without comment. --> ## The problem I've used the denon integration (not DenonAVR) to connect to my Marantz NR1609. It worked for a while, but doesn't now ## Environment - Home Assistant Core release with the issue: 0.109.6 - Last working Home Assistant Core release (if known): 0.99 worked. Don't know when it was switched over. - Operating environment (Home Assistant/Supervised/Docker/venv): Supervised - Integration causing this issue: Denon - Link to integration documentation on our website: https://www.home-assistant.io/integrations/denon/ ## Problem-relevant `configuration.yaml` Didn't change this configuration recently, since it only has one variable (and a name) ``` media_player: ... - platform: denon host: 192.168.xx.xx name: Marantz ``` ## Traceback/Error logs ``` Logger: homeassistant.components.media_player Source: components/denon/media_player.py:117 Traceback (most recent call last): File "/usr/src/homeassistant/homeassistant/helpers/entity_platform.py", line 178, in _async_setup_platform await asyncio.wait_for(asyncio.shield(task), SLOW_SETUP_MAX_WAIT) File "/usr/local/lib/python3.7/asyncio/tasks.py", line 442, in wait_for return fut.result() File "/usr/local/lib/python3.7/concurrent/futures/thread.py", line 57, in run result = self.fn(*self.args, **self.kwargs) File "/usr/src/homeassistant/homeassistant/components/denon/media_player.py", line 85, in setup_platform if denon.update(): File "/usr/src/homeassistant/homeassistant/components/denon/media_player.py", line 162, in update self._setup_sources(telnet) File "/usr/src/homeassistant/homeassistant/components/denon/media_player.py", line 117, in _setup_sources source, configured_name = line[len("SSFUN") :].split(" ", 1) ValueError: not enough values to unpack (expected 2, got 1) ``` ## Additional information
null
https://github.com/home-assistant/core/pull/40514
null
{'base_commit': '62054b84336e4d4a8a6a1ea649724e9eb86876b4', 'files': [{'path': 'homeassistant/components/denon/media_player.py', 'status': 'modified', 'Loc': {"('DenonDevice', '_setup_sources', 108)": {'mod': [114, 117]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "homeassistant/components/denon/media_player.py" ], "doc": [], "test": [], "config": [], "asset": [] }
1
home-assistant
core
e11ead410bbd9179ded05ba5e07b6de9919ec0ea
https://github.com/home-assistant/core/issues/128046
integration: gree
Gree climate does not update tha data (status)
### The problem The HA does not update the climate status (switch) if the change is made with a remote control. It does not follow the climate continuously. ### What version of Home Assistant Core has the issue? 2024.10.1 ### What was the last working version of Home Assistant Core? 2024.09. ### What type of installation are you running? Home Assistant OS ### Integration causing the issue Gree ### Link to integration documentation on our website https://www.home-assistant.io/integrations/gree ### Diagnostics information _No response_ ### Example YAML snippet _No response_ ### Anything in the logs that might be useful for us? _No response_ ### Additional information _No response_
null
https://github.com/home-assistant/core/pull/139469
null
{'base_commit': 'e11ead410bbd9179ded05ba5e07b6de9919ec0ea', 'files': [{'path': 'homeassistant/components/gree/const.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [22]}}}, {'path': 'homeassistant/components/gree/coordinator.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [4, 26]}, "('DeviceDataUpdateCoordinator', '_async_update_data', 65)": {'add': [102], 'mod': [91, 99, 108]}, "('DeviceDataUpdateCoordinator', '__init__', 38)": {'mod': [51]}}}, {'path': 'tests/components/gree/test_climate.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [54]}, "(None, 'run_update', 348)": {'mod': [349]}}}]}
[]
[]
[]
{ "iss_type": "2", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "homeassistant/components/gree/coordinator.py", "homeassistant/components/gree/const.py" ], "doc": [], "test": [ "tests/components/gree/test_climate.py" ], "config": [], "asset": [] }
1
home-assistant
core
80af2f4279828419a9112a5478029f2c506896e7
https://github.com/home-assistant/core/issues/53678
integration: sonos
Sonos Alarm switch changes immediatly after click
### The problem I have sonos alarm switches. When I click on it, the status changes, but immediately back again. After a while (approx. 60s) the status changes to the correct state. In the sonos app, the alarm is switched correctly with the first click. See the GIF below. The behavior when switching on and then switching off. ![Animation](https://user-images.githubusercontent.com/15085873/127458431-ea5a0ed0-25e6-4459-837a-41f9fad044d8.gif) ### What is version of Home Assistant Core has the issue? core-2021.7.4 ### What was the last working version of Home Assistant Core? _No response_ ### What type of installation are you running? Home Assistant Core ### Integration causing the issue Sonos ### Link to integration documentation on our website https://github.com/home-assistant/core/tree/dev/homeassistant/components/sonos ### Example YAML snippet _No response_ ### Anything in the logs that might be useful for us? _No response_ ### Additional information _No response_
null
https://github.com/home-assistant/core/pull/55529
null
{'base_commit': '80af2f4279828419a9112a5478029f2c506896e7', 'files': [{'path': 'homeassistant/components/sonos/alarms.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [11], 'mod': [9]}, "('SonosAlarms', 'async_update_entities', 37)": {'add': [38], 'mod': [40, 42, 45, 46, 47, 48, 49, 50, 51, 53, 54, 56, 57, 58, 59, 61, 62, 63, 64]}, "('SonosAlarms', '__init__', 23)": {'mod': [26]}, "('SonosAlarms', '__iter__', 28)": {'mod': [30, 31]}, "('SonosAlarms', 'get', 33)": {'mod': [35]}, "('SonosAlarms', None, 20)": {'mod': [37]}, "('SonosAlarms', 'update_cache', 53)": {'mod': [66, 67, 68, 70]}}}, {'path': 'homeassistant/components/sonos/favorites.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [5, 11]}, "('SonosFavorites', '__init__', 23)": {'add': [26]}, "('SonosFavorites', 'update_cache', 46)": {'add': [50, 63, 64], 'mod': [49, 61]}, "('SonosFavorites', None, 20)": {'mod': [33, 46, 47]}, "('SonosFavorites', 'async_update_entities', 33)": {'mod': [35, 36, 37, 38, 39, 44]}}}, {'path': 'homeassistant/components/sonos/household_coordinator.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [9], 'mod': [4, 7, 11]}, "('SonosHouseholdCoordinator', '__init__', 22)": {'add': [27], 'mod': [26]}, "('SonosHouseholdCoordinator', None, 19)": {'add': [46], 'mod': [34, 35, 37, 38, 60, 61, 62, 63, 64, 65, 66, 68, 72, 73]}, "('SonosHouseholdCoordinator', 'setup', 29)": {'mod': [32]}, "('SonosHouseholdCoordinator', '_async_poll', 47)": {'mod': [52, 53, 55]}}}, {'path': 'homeassistant/components/sonos/manifest.json', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [6]}}}, {'path': 'homeassistant/components/sonos/speaker.py', 'status': 'modified', 'Loc': {"('SonosSpeaker', 'async_dispatch_favorites', 449)": {'mod': [453]}}}, {'path': 'homeassistant/components/sonos/switch.py', 'status': 'modified', 'Loc': {"(None, '_async_create_entity', 38)": {'add': [39, 40, 41]}}}, {'path': 'requirements_all.txt', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [2175]}}}, {'path': 'requirements_test_all.txt', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [1215]}}}, {'path': 'tests/components/sonos/conftest.py', 'status': 'modified', 'Loc': {"('SonosMockEvent', 'increment_variable', 34)": {'add': [41]}, "(None, 'alarm_clock_fixture', 111)": {'add': [115], 'mod': [121]}, "(None, 'alarm_clock_fixture_extended', 127)": {'add': [131], 'mod': [141]}, "(None, 'music_library_fixture', 103)": {'mod': [105, 106]}}}, {'path': 'tests/components/sonos/test_switch.py', 'status': 'modified', 'Loc': {"(None, 'test_alarm_create_delete', 53)": {'add': [71], 'mod': [78]}}}]}
[]
[]
[]
{ "iss_type": "2", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "homeassistant/components/sonos/alarms.py", "homeassistant/components/sonos/household_coordinator.py", "homeassistant/components/sonos/manifest.json", "tests/components/sonos/conftest.py", "homeassistant/components/sonos/switch.py", "homeassistant/components/sonos/speaker.py", "homeassistant/components/sonos/favorites.py" ], "doc": [], "test": [ "tests/components/sonos/test_switch.py" ], "config": [ "requirements_test_all.txt", "requirements_all.txt" ], "asset": [] }
1
home-assistant
core
f29e0bf53eec22dca65e3a62336cf84797e36bf2
https://github.com/home-assistant/core/issues/7306
Problems sending push message with HTML5 notify
**Home Assistant release (`hass --version`):** 0.43.0 **Python release (`python3 --version`):** Python 3.4.2 **Component/platform:** HTML5 notify on Raspberry PI 3 **Description of problem:** I've followed the installation instructions here: https://home-assistant.io/components/notify.html5/ When I try to send a message, manually triggering it through the services pane, I get the error below (see traceback). At the bottom of the trace it says: `AttributeError: /usr/lib/arm-linux-gnueabihf/libcrypto.so.1.0.0: undefined symbol: EVP_CIPHER_CTX_reset` **Expected:** **Problem-relevant `configuration.yaml` entries and steps to reproduce:** I followed the installation instructions and installed `libffi-dev`, `libpython-dev`, and `libssl-dev` before installing `pywebpush`. I've tried `pywebpush` 0.5, 0.6, 0.6.1 and 0.8. All resulted in the same error, which leads me to believe the error is elsewhere. I'm using a SSL certificate from let's encrypt and duckdns.org. ```yaml notify: - name: html5 platform: html5 gcm_api_key: !secret gcm_api_key gcm_sender_id: !secret gcm_sender_id ``` 1. Go to services pane 2. Use the `notify.html5push` service with data `{"title": "Title", "message": "Message"}` 3. Call service, and the error pops up in the HASS logs. **Traceback (if applicable):** ```bash Apr 25 20:13:36 raspberrypi hass[2828]: INFO:homeassistant.core:Bus:Handling <Event call_service[L]: service_call_id=1978163280-1, service_data=title=hei, message=hei, service=html5push, domain=notify> Apr 25 20:13:37 raspberrypi hass[2828]: ERROR:homeassistant.core:Error doing job: Task exception was never retrieved Apr 25 20:13:37 raspberrypi hass[2828]: Traceback (most recent call last): Apr 25 20:13:37 raspberrypi hass[2828]: File "/usr/lib/python3.4/asyncio/tasks.py", line 233, in _step Apr 25 20:13:37 raspberrypi hass[2828]: result = coro.throw(exc) Apr 25 20:13:37 raspberrypi hass[2828]: File "/srv/homeassistant/lib/python3.4/site-packages/homeassistant/core.py", line 1015, in _event_to_service_call Apr 25 20:13:37 raspberrypi hass[2828]: yield from service_handler.func(service_call) Apr 25 20:13:37 raspberrypi hass[2828]: File "/srv/homeassistant/lib/python3.4/site-packages/homeassistant/components/notify/__init__.py", line 136, in async_notify_message Apr 25 20:13:37 raspberrypi hass[2828]: yield from notify_service.async_send_message(**kwargs) Apr 25 20:13:37 raspberrypi hass[2828]: File "/usr/lib/python3.4/asyncio/futures.py", line 388, in __iter__ Apr 25 20:13:37 raspberrypi hass[2828]: yield self # This tells Task to wait for completion. Apr 25 20:13:37 raspberrypi hass[2828]: File "/usr/lib/python3.4/asyncio/tasks.py", line 286, in _wakeup Apr 25 20:13:37 raspberrypi hass[2828]: value = future.result() Apr 25 20:13:37 raspberrypi hass[2828]: File "/usr/lib/python3.4/asyncio/futures.py", line 277, in result Apr 25 20:13:37 raspberrypi hass[2828]: raise self._exception Apr 25 20:13:37 raspberrypi hass[2828]: File "/usr/lib/python3.4/concurrent/futures/thread.py", line 54, in run Apr 25 20:13:37 raspberrypi hass[2828]: result = self.fn(*self.args, **self.kwargs) Apr 25 20:13:37 raspberrypi hass[2828]: File "/srv/homeassistant/lib/python3.4/site-packages/homeassistant/components/notify/html5.py", line 349, in send_message Apr 25 20:13:37 raspberrypi hass[2828]: from pywebpush import WebPusher Apr 25 20:13:37 raspberrypi hass[2828]: File "/home/homeassistant/.homeassistant/deps/pywebpush/__init__.py", line 10, in <module> Apr 25 20:13:37 raspberrypi hass[2828]: import http_ece Apr 25 20:13:37 raspberrypi hass[2828]: File "/home/homeassistant/.homeassistant/deps/http_ece/__init__.py", line 8, in <module> Apr 25 20:13:37 raspberrypi hass[2828]: import pyelliptic Apr 25 20:13:37 raspberrypi hass[2828]: File "/home/homeassistant/.homeassistant/deps/pyelliptic/__init__.py", line 43, in <module> Apr 25 20:13:37 raspberrypi hass[2828]: from .openssl import OpenSSL Apr 25 20:13:37 raspberrypi hass[2828]: File "/home/homeassistant/.homeassistant/deps/pyelliptic/openssl.py", line 310, in <module> Apr 25 20:13:37 raspberrypi hass[2828]: OpenSSL = _OpenSSL(libname) Apr 25 20:13:37 raspberrypi hass[2828]: File "/home/homeassistant/.homeassistant/deps/pyelliptic/openssl.py", line 144, in __init__ Apr 25 20:13:37 raspberrypi hass[2828]: self.EVP_CIPHER_CTX_reset = self._lib.EVP_CIPHER_CTX_reset Apr 25 20:13:37 raspberrypi hass[2828]: File "/usr/lib/python3.4/ctypes/__init__.py", line 364, in __getattr__ Apr 25 20:13:37 raspberrypi hass[2828]: func = self.__getitem__(name) Apr 25 20:13:37 raspberrypi hass[2828]: File "/usr/lib/python3.4/ctypes/__init__.py", line 369, in __getitem__ Apr 25 20:13:37 raspberrypi hass[2828]: func = self._FuncPtr((name_or_ordinal, self)) Apr 25 20:13:37 raspberrypi hass[2828]: AttributeError: /usr/lib/arm-linux-gnueabihf/libcrypto.so.1.0.0: undefined symbol: EVP_CIPHER_CTX_reset ``` **Additional info:**
null
https://github.com/home-assistant/core/pull/7310
null
{'base_commit': 'f29e0bf53eec22dca65e3a62336cf84797e36bf2', 'files': [{'path': 'homeassistant/components/notify/html5.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [28]}}}, {'path': 'requirements_all.txt', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [526]}}}]}
[]
[]
[]
{ "iss_type": "1", "iss_reason": "1", "loc_way": "pr", "loc_scope": null, "info_type": null }
{ "code": [ "homeassistant/components/notify/html5.py" ], "doc": [], "test": [], "config": [ "requirements_all.txt" ], "asset": [] }
1