organization string | repo_name string | base_commit string | iss_html_url string | iss_label string | title string | body string | code null | pr_html_url string | commit_html_url string | file_loc string | own_code_loc list | ass_file_loc list | other_rep_loc list | analysis dict | loctype dict | iss_has_pr int64 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
AUTOMATIC1111 | stable-diffusion-webui | d9499f4301018ebd2977685d098381aa4111d2ae | https://github.com/AUTOMATIC1111/stable-diffusion-webui/issues/13724 | enhancement | [Feature Request]: Sort items by date by default | ### Is there an existing issue for this?
- [X] I have searched the existing issues and checked the recent builds/commits
### What would your feature do ?
I hope to use time sorting by default when opening additional interfaces, so that I can immediately try the new model I just downloaded.
### Proposed workflow
1. Go to ....
2. Press ....
3. ...
### Additional information
_No response_ | null | null | https://github.com/AUTOMATIC1111/stable-diffusion-webui/commit/d9499f4301018ebd2977685d098381aa4111d2ae | {'base_commit': 'd9499f4301018ebd2977685d098381aa4111d2ae', 'files': [{'path': 'javascript/extraNetworks.js', 'status': 'modified', 'Loc': {"(None, 'setupExtraNetworksForTab', 18)": {'add': [51, 54, 98, 99], 'mod': [30, 56, 57, 58, 59, 65, 91, 92, 93, 94]}, '(None, None, None)': {'add': [115]}, "(None, 'applyExtraNetworkSort', 116)": {'add': [116]}}}, {'path': 'modules/shared_options.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [236]}}}, {'path': 'modules/ui_extra_networks.py', 'status': 'modified', 'Loc': {"(None, 'create_ui', 357)": {'add': [397], 'mod': [384, 385]}}}]} | [] | [] | [] | {
"iss_type": "4",
"iss_reason": "2",
"loc_way": "commit",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"modules/shared_options.py",
"modules/ui_extra_networks.py",
"javascript/extraNetworks.js"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | null |
AUTOMATIC1111 | stable-diffusion-webui | 22bcc7be428c94e9408f589966c2040187245d81 | https://github.com/AUTOMATIC1111/stable-diffusion-webui/issues/9102 | bug-report | [Bug]: Model Dropdown Select on Firefox is obscured by svelte pre-loader | ### Is there an existing issue for this?
- [X] I have searched the existing issues and checked the recent builds/commits
### What happened?
Seems the latest pull request has added pre-loaders and I noticed that the model dropdown is constantly loading and therefore obscuring the dropdown from the user. This is only happening in Firefox, Chrome for example is fine.
```
.wrap.default.svelte-gjihhp {
inset: 0;
}
```
I just set it to `display: none` to access it
### Steps to reproduce the problem
Load up in Firefox and try to change the model
See attached screenshot

### What should have happened?
Have access to the model dropdown select
### Commit where the problem happens
f1db987
### What platforms do you use to access the UI ?
Windows
### What browsers do you use to access the UI ?
Mozilla Firefox
### Command Line Arguments
```Shell
No
```
### List of extensions
No
### Console logs
```Shell
Uncaught (in promise) TypeError: q[R[H]] is undefined
ct http://127.0.0.1:7860/assets/Blocks.5efe22d4.js:76
ct http://127.0.0.1:7860/assets/Blocks.5efe22d4.js:76
ze http://127.0.0.1:7860/assets/Blocks.5efe22d4.js:76
```
### Additional information
_No response_ | null | null | https://github.com/joysun545/stable-diffusion-webui/commit/22bcc7be428c94e9408f589966c2040187245d81 | {'base_commit': '22bcc7be428c94e9408f589966c2040187245d81', 'files': [{'path': 'modules/ui.py', 'status': 'modified', 'Loc': {"(None, 'create_ui', 437)": {'add': [1630]}}}]} | [] | [] | [] | {
"iss_type": "2",
"iss_reason": "1",
"loc_way": "commit",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"modules/ui.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | null |
huggingface | transformers | a8e3336a850e856188350a93e67d77c07c85b8af | https://github.com/huggingface/transformers/issues/2008 | wontfix | Expand run_lm_finetuning.py to all models | ## 🚀 Feature
[run_lm_finetuning.py](https://github.com/huggingface/transformers/blob/b0ee7c7df3d49a819c4d6cef977214bd91f5c075/examples/run_lm_finetuning.py) is a very useful tool for finetuning many models the library provided. But it doesn't cover all the models. Currently available models are:
- gpt2
- openai-gpt
- bert
- roberta
- distilbert
- camembert
And not available ones:
- ctrl
- xlm
- xlnet
- transfo-xl
- albert
## Motivation
Most important part of such a library is that it can be easily finetuned. `run_lm_finetuning.py` gives us that opportunity but why say no more :)
| null | null | https://github.com/huggingface/transformers/commit/3dcb748e31be8c7c9e4f62926c5c144c62d07218
https://github.com/huggingface/transformers/commit/a8e3336a850e856188350a93e67d77c07c85b8af | {'base_commit': 'a8e3336a850e856188350a93e67d77c07c85b8af', 'files': [{'path': 'examples/ner/run_ner.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [33], 'mod': [41]}}}, {'path': 'examples/ner/run_tf_ner.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [16, 17, 18, 19, 21, 22, 23, 24, 25, 37, 38, 39, 41, 42, 43, 44, 45, 52]}, "(None, 'main', 457)": {'mod': [512, 513, 523, 530, 565, 587, 614, 615]}}}, {'path': 'examples/run_glue.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [32], 'mod': [35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 676, 677, 683, 695]}, "(None, 'train', 69)": {'mod': [75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101]}, "(None, 'main', 386)": {'mod': [445, 625, 626, 632, 637]}}}, {'path': 'examples/run_language_modeling.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [40], 'mod': [43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 60, 61, 62, 789]}, "('TextDataset', '__init__', 68)": {'mod': [76, 77, 78, 79, 80, 81, 82, 83]}, "(None, 'main', 464)": {'mod': [696, 699, 701, 703, 706, 708, 712, 722, 730, 771, 772]}}}, {'path': 'examples/run_squad.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [32], 'mod': [35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 75, 76, 77, 78, 79, 80, 81, 845]}, "(None, 'train', 76)": {'mod': [83, 84, 85, 86, 87, 88, 89, 90, 91]}, "(None, 'main', 477)": {'mod': [516, 760, 761, 765, 770, 820, 821]}}}, {'path': 'src/transformers/__init__.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [160, 319]}}}, {'path': 'templates/adding_a_new_example_script/run_xxx.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [30], 'mod': [33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 75, 76, 77, 78, 79, 80, 709]}, "(None, 'set_seed', 69)": {'mod': [71, 72, 73]}, "(None, 'main', 388)": {'mod': [421, 629, 630, 634, 639, 690, 691]}}}]} | [] | [] | [] | {
"iss_type": "4",
"iss_reason": "2",
"loc_way": "commit",
"loc_scope": "",
"info_type": "Code"
} | {
"code": [
"examples/run_squad.py",
"templates/adding_a_new_example_script/run_xxx.py",
"examples/run_glue.py",
"src/transformers/__init__.py",
"examples/ner/run_tf_ner.py",
"examples/run_language_modeling.py",
"examples/ner/run_ner.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | null |
huggingface | transformers | 88d7f96e33c3f3e541bcdd913f2ff1e50aa18c1b | https://github.com/huggingface/transformers/issues/5212 | BartConfig wrong decoder_start_token_id? | # 🐛 Bug
## Information
Model I am using (Bert, XLNet ...): Bart
Language I am using the model on (English, Chinese ...): English
## To reproduce
Steps to reproduce the behavior:
```
from transformers import BartConfig, BartTokenizer
config = BartConfig.from_pretrained('facebook/bart-large')
tokenizer = BartTokenizer.from_pretrained('facebook/bart-large')
config.decoder_start_token_id
>>> 2
tokenizer.bos_token_id
>>> 0 # != config.decoder_start_token_id
tokenizer.eos_token_id
>>> 2
```
It is misleading in the documentation of the function ```generate````
*decoder_start_token_id=None – (optional) int If an encoder-decoder model starts decoding with a different token than BOS. Defaults to None and is changed to BOS later.*
## Expected behavior
I expect that decoder_start_token_id = tokenizer.bos_token_id, but maybe the model is designed to start decoding with EOS token.
| null | https://github.com/huggingface/transformers/pull/5306 | null | {'base_commit': '88d7f96e33c3f3e541bcdd913f2ff1e50aa18c1b', 'files': [{'path': 'src/transformers/modeling_tf_utils.py', 'status': 'modified', 'Loc': {"('TFPreTrainedModel', 'generate', 551)": {'mod': [645, 646]}}}, {'path': 'src/transformers/modeling_utils.py', 'status': 'modified', 'Loc': {"('PreTrainedModel', 'generate', 871)": {'mod': [965, 966]}}}]} | [] | [] | [] | {
"iss_type": "2",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": "0",
"info_type": "Code\n里的docstring"
} | {
"code": [
"src/transformers/modeling_utils.py",
"src/transformers/modeling_tf_utils.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | null | |
geekan | MetaGPT | ba9e3fe6267ea81a3d546b7aa5bcf0122f365e51 | https://github.com/geekan/MetaGPT/issues/1153 | bug | mermaid: Generating ..seq_flow/20240402032019.svg.. Error: Failed to launch the browser process! | **Bug description**
Generating /app/metagpt/workspace/jqdlhwap/resources/seq_flow/20240402032019.svg.. Error: Failed to launch the browser process!
**Bug solved method**
**Environment information**
"docker compose up -d" after clone
already run "npm install -g @mermaid-js/mermaid-cli":
root@84a2e77496b0:/app/metagpt# mmdc -h
Usage: mmdc [options]
Options:
-V, --version output the version number
-t, --theme [theme] Theme of the chart (choices: "default", "forest", "dark", "neutral", default: "default")
-w, --width [width] Width of the page (default: 800)
-H, --height [height] Height of the page (default: 600)
-i, --input <input> Input mermaid file. Files ending in .md will be treated as Markdown and all charts (e.g. ```mermaid (...)```) will be extracted and generated.
Use `-` to read from stdin.
-o, --output [output] Output file. It should be either md, svg, png or pdf. Optional. Default: input + ".svg"
-e, --outputFormat [format] Output format for the generated image. (choices: "svg", "png", "pdf", default: Loaded from the output file extension)
-b, --backgroundColor [backgroundColor] Background color for pngs/svgs (not pdfs). Example: transparent, red, '#F0F0F0'. (default: "white")
-c, --configFile [configFile] JSON configuration file for mermaid.
-C, --cssFile [cssFile] CSS file for the page.
-s, --scale [scale] Puppeteer scale factor (default: 1)
-f, --pdfFit Scale PDF to fit chart
-q, --quiet Suppress log output
-p --puppeteerConfigFile [puppeteerConfigFile] JSON configuration file for puppeteer.
-h, --help display help for command
- LLM type and model name: zhipu-api / GLM-4
- System version:
- Python version:
- MetaGPT version or branch: main
`run in docker`
- packages version:
- installation method:
**Screenshots or logs**
2024-04-02 03:20:46.126 | INFO | metagpt.utils.mermaid:mermaid_to_file:48 - Generating /app/metagpt/workspace/jqdlhwap/resources/seq_flow/20240402032019.svg..
2024-04-02 03:20:46.460 | WARNING | metagpt.utils.mermaid:mermaid_to_file:74 -
Error: Failed to launch the browser process!
[0402/032046.449080:ERROR:zygote_host_impl_linux.cc(100)] Running as root without --no-sandbox is not supported. See https://crbug.com/638180.
TROUBLESHOOTING: https://pptr.dev/troubleshooting
at Interface.onClose (file:///usr/lib/node_modules/@mermaid-js/mermaid-cli/node_modules/@puppeteer/browsers/lib/esm/launch.js:253:24)
at Interface.emit (node:events:524:35)
at Interface.close (node:internal/readline/interface:526:10)
at Socket.onend (node:internal/readline/interface:252:10)
at Socket.emit (node:events:524:35)
at endReadableNT (node:internal/streams/readable:1378:12)
at process.processTicksAndRejections (node:internal/process/task_queues:82:21)
| null | https://github.com/FoundationAgents/MetaGPT/pull/1155 | null | {'base_commit': 'ba9e3fe6267ea81a3d546b7aa5bcf0122f365e51', 'files': [{'path': 'metagpt/configs/mermaid_config.py', 'status': 'modified', 'Loc': {"('MermaidConfig', None, 13)": {'mod': [16]}}}, {'path': 'metagpt/utils/mermaid.py', 'status': 'modified', 'Loc': {"(None, 'mermaid_to_file', 17)": {'add': [83]}}}, {'path': 'config/config2.yaml', 'Loc': {}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"metagpt/configs/mermaid_config.py",
"metagpt/utils/mermaid.py"
],
"doc": [],
"test": [],
"config": [
"config/config2.yaml"
],
"asset": []
} | null |
geekan | MetaGPT | c779f6977ecbdba075d7c81519edd5eaf6de2d0e | https://github.com/geekan/MetaGPT/issues/1197 | Support for Cohere API | Please add support for Cohere API with all the built in RAG and tool use functionalities. Essentially, RAG and tool use in Cohere are just chat parameters definable by users. More information can be found at https://docs.cohere.com/reference/chat . | null | https://github.com/FoundationAgents/MetaGPT/pull/1193 | null | {'base_commit': 'c779f6977ecbdba075d7c81519edd5eaf6de2d0e', 'files': [{'path': 'metagpt/const.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [53]}}}, {'path': 'metagpt/rag/factories/ranker.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [13]}, "('RankerFactory', '__init__', 20)": {'add': [24]}, "('RankerFactory', None, 17)": {'add': [47]}}}, {'path': 'metagpt/rag/schema.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [121]}}}, {'path': 'setup.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [42]}}}]} | [] | [] | [] | {
"iss_type": "4",
"iss_reason": "2",
"loc_way": "pr",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"metagpt/rag/schema.py",
"metagpt/rag/factories/ranker.py",
"setup.py",
"metagpt/const.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | null | |
geekan | MetaGPT | b5bb4d7e63e72c3d118e449a3763c1ff4411f159 | https://github.com/geekan/MetaGPT/issues/1547 | ERROR | metagpt.utils.mmdc_playwright:mermaid_to_file:118 - Page.goto: net::ERR_FILE_NOT_FOUND at file:///root/anaconda3/envs/metagpt/lib/python3.11/site-packages/metagpt/utils/index.html | **Bug description**
I installed playwright and its chrominum with the guidance and made configuration of mermaid. But it seems that the mermaid didn't work normally.
```
ERROR | metagpt.utils.mmdc_playwright:mermaid_to_file:118 - Page.goto: net::ERR_FILE_NOT_FOUND at file:///root/anaconda3/envs/metagpt/lib/python3.11/site-packages/metagpt/utils/index.html
Call log:
navigating to "file:///root/anaconda3/envs/metagpt/lib/python3.11/site-packages/metagpt/utils/index.html", waiting until "load"
```
**Bug solved method**
I think it may do with the configuration but the document of this part is not so clear. I don't know how to fill in the "path" if I'm using playwright and whether it is right to keep that stuff in default.
This is my configuration:
```yaml
llm:
api_type: 'openai' # or azure / ollama / groq etc. Check LLMType for more options
api_key: '[MY_API_KEY]' # MY_API_KEY
model: 'yi-lightning' # or gpt-3.5-turbo
base_url: 'https://api.lingyiwanwu.com/v1' # or any forward url.
# proxy: 'YOUR_LLM_PROXY_IF_NEEDED' # Optional. If you want to use a proxy, set it here.
# pricing_plan: 'YOUR_PRICING_PLAN' # Optional. If your pricing plan uses a different name than the `model`.
mermaid:
engine: 'playwright' # nodejs/ink/playwright/pyppeteer
# path: 'mmdc' # such as './node_modules/.bin/mmdc'
# puppeteer_config: './config/puppeteer-config' # only for nodejs
# pyppeteer_path: '/usr/bin/google-chrome-stable' # only for pyppeteer
```
This is your example:
```yaml
mermaid:
engine: 'nodejs' # nodejs/ink/playwright/pyppeteer
path: 'mmdc' # such as './node_modules/.bin/mmdc'
puppeteer_config: './config/puppeteer-config' # only for nodejs
pyppeteer_path: '/usr/bin/google-chrome-stable' # only for pyppeteer
```
**Environment information**
- LLM type and model name:
- System version: ubuntu 22.04
- Python version: 3.11
- MetaGPT version or branch: 0.8
<!-- Dependent packagess:the packages version cause the bug(like `pydantic 1.10.8`), installation method(like `pip install metagpt` or `pip install from source` or `run in docker`) -->
- packages version:
- installation method: pip
**Screenshots or logs**
2024-10-29 16:22:16.900 | WARNING | metagpt.utils.cost_manager:update_cost:49 - Model yi-lightning not found in TOKEN_COSTS.
2024-10-29 16:22:16.903 | INFO | metagpt.utils.git_repository:rename_root:219 - Rename directory /root/workspace/cited_papaer_eval/workspace/20241029162137 to /root/workspace/cited_papaer_eval/workspace/scholar_cited_evaluation
2024-10-29 16:22:16.904 | INFO | metagpt.utils.file_repository:save:57 - save to: /root/workspace/cited_papaer_eval/workspace/scholar_cited_evaluation/docs/prd/20241029162216.json
2024-10-29 16:22:17.435 | ERROR | metagpt.utils.mmdc_playwright:mermaid_to_file:118 - Page.goto: net::ERR_FILE_NOT_FOUND at file:///root/anaconda3/envs/metagpt/lib/python3.11/site-packages/metagpt/utils/index.html
Call log:
navigating to "file:///root/anaconda3/envs/metagpt/lib/python3.11/site-packages/metagpt/utils/index.html", waiting until "load"
| null | https://github.com/FoundationAgents/MetaGPT/pull/1564 | null | {'base_commit': 'b5bb4d7e63e72c3d118e449a3763c1ff4411f159', 'files': [{'path': 'metagpt/utils/mmdc_playwright.py', 'status': 'modified', 'Loc': {"(None, 'mermaid_to_file', 17)": {'mod': [84, 85, 86, 87]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"metagpt/utils/mmdc_playwright.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | null | |
langflow-ai | langflow | 7276f699fc85c611f1c3f83a19a368da9841e3a4 | https://github.com/langflow-ai/langflow/issues/2892 | question | Flow as tool Usage | ### Discussed in https://github.com/langflow-ai/langflow/discussions/2891
<div type='discussions-op-text'>
<sup>Originally posted by **pavansandeep2910** July 23, 2024</sup>
I cannot understand how to load files to see them in flow as tool component. can anyone help me direct to flow as tool usage?</div> | null | https://github.com/langflow-ai/langflow/pull/3093 | null | {'base_commit': '7276f699fc85c611f1c3f83a19a368da9841e3a4', 'files': [{'path': 'src/backend/base/langflow/components/prototypes/SubFlow.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [2], 'mod': [4, 6, 9, 10, 11, 12]}, "('SubFlowComponent', 'build', 98)": {'add': [103], 'mod': [102, 105, 108, 112, 114, 115]}, "('SubFlowComponent', None, 15)": {'mod': [15, 17, 18, 19, 22, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 98, 99, 100]}, "('SubFlowComponent', 'get_flow_names', 24)": {'mod': [25, 26]}, "('SubFlowComponent', 'update_build_config', 35)": {'mod': [36, 39, 41]}, "('SubFlowComponent', 'add_inputs_to_build_config', 58)": {'mod': [71]}}}, {'path': 'src/backend/base/langflow/initial_setup/setup.py', 'status': 'modified', 'Loc': {"(None, 'load_starter_projects', 342)": {'mod': [346]}}}, {'path': 'src/backend/base/langflow/inputs/inputs.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [231]}, "('SecretStrInput', None, 215)": {'mod': [229]}}}, {'path': 'src/frontend/src/CustomNodes/GenericNode/components/handleRenderComponent/index.tsx', 'status': 'modified', 'Loc': {'(None, None, 1)': {'mod': [1]}, '(None, None, 4)': {'mod': [4]}, '(None, None, 40)': {'mod': [40, 41, 42, 43]}, '(None, None, 103)': {'mod': [103]}}}, {'path': 'src/frontend/src/CustomNodes/GenericNode/components/parameterComponent/index.tsx', 'status': 'modified', 'Loc': {'(None, None, 352)': {'mod': [352, 353]}, '(None, None, 365)': {'mod': [365, 366]}}}, {'path': 'src/frontend/src/CustomNodes/hooks/use-handle-new-value.tsx', 'status': 'modified', 'Loc': {'(None, None, 66)': {'mod': [66, 67]}}}, {'path': 'src/frontend/src/components/parameterRenderComponent/component/refreshParameterComponent/index.tsx', 'status': 'modified', 'Loc': {'(None, None, 32)': {'mod': [32]}}}]} | [] | [] | [] | {
"iss_type": "3",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"src/backend/base/langflow/components/prototypes/SubFlow.py",
"src/backend/base/langflow/initial_setup/setup.py",
"src/backend/base/langflow/inputs/inputs.py",
"src/frontend/src/components/parameterRenderComponent/component/refreshParameterComponent/index.tsx",
"src/frontend/src/CustomNodes/GenericNode/components/parameterComponent/index.tsx",
"src/frontend/src/CustomNodes/GenericNode/components/handleRenderComponent/index.tsx",
"src/frontend/src/CustomNodes/hooks/use-handle-new-value.tsx"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | null |
langflow-ai | langflow | b3b5290598f5970fd6a1a092fe4d11211008a04d | https://github.com/langflow-ai/langflow/issues/5378 | bug | URL component lost imported urls in tool mode when refresh UI | ### Bug Description
I have this issue when i build multi agent flow like import json below
- URL Component Before refresh UI:


After refresh UI:


I have check in normal mode of URL Component, and this bug does not appear.
### Reproduction
1. Create flow
2. Add URL component, change to Tool mode
3. Input Urls
4. Save flow
5. Reload UI (Press F5)
### Expected behavior
URL component has keep Urls has input. In Tool mode
### Who can help?
_No response_
### Operating System
Windows 11/Docker
### Langflow Version
v1.1.1
### Python Version
None
### Screenshot
_No response_
### Flow File
[Simple Agent (bug).json](https://github.com/user-attachments/files/18206463/Simple.Agent.bug.json)
| null | https://github.com/langflow-ai/langflow/pull/5316 | null | {'base_commit': 'b3b5290598f5970fd6a1a092fe4d11211008a04d', 'files': [{'path': 'src/frontend/src/components/core/parameterRenderComponent/components/toggleShadComponent/index.tsx', 'status': 'modified', 'Loc': {'(None, None, 39)': {'mod': [39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54]}}}, {'path': 'src/frontend/src/pages/FlowPage/components/nodeToolbarComponent/index.tsx', 'status': 'modified', 'Loc': {'(None, None, 5)': {'add': [5]}, '(None, None, 44)': {'add': [44]}, '(None, None, 79)': {'add': [79]}, '(None, None, 98)': {'add': [98]}, '(None, None, 149)': {'add': [149]}, '(None, None, 11)': {'mod': [11]}, '(None, None, 23)': {'mod': [23, 24, 25]}, '(None, None, 145)': {'mod': [145, 146]}, '(None, None, 158)': {'mod': [158, 159]}, '(None, None, 174)': {'mod': [174]}, '(None, None, 291)': {'mod': [291]}, '(None, None, 403)': {'mod': [403, 404, 405, 407, 408, 409]}, '(None, None, 421)': {'mod': [421, 422, 423, 424, 425, 426]}, '(None, None, 471)': {'mod': [471, 472, 473, 474, 475, 476, 477, 478, 479, 480, 481, 482, 483, 484, 485]}}}]} | [] | [] | [] | {
"iss_type": "2",
"iss_reason": "5",
"loc_way": "pr",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"src/frontend/src/pages/FlowPage/components/nodeToolbarComponent/index.tsx",
"src/frontend/src/components/core/parameterRenderComponent/components/toggleShadComponent/index.tsx"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | null |
langflow-ai | langflow | a4dc5381b2cf31c507cc32f9027f76bf00d61ccc | https://github.com/langflow-ai/langflow/issues/3536 | bug | Prompt component does not pass variables correctly | ### Bug Description
I have prompt with two variables.
{image_url} Value of image is https://oaidalleapiprodscus.blob.core.windows.net/private/org-N0bzhj17kGCdvkPCuGgpUuhO/user-RwGFQKsTTGw8hOO1ResFbpwQ/img-GH9Zkmf4RabhW48gVeZslvyb.png?st=2024-08-23T19%3A00%3A31Z&se=2024-08-23T21%3A00%3A31Z&sp=r&sv=2024-08-04&sr=b&rscd=inline&rsct=image/png&skoid=d505667d-d6c1-4a0a-bac7-5c84a87759f8&sktid=a48cca56-e6da-484e-a814-9c849652bcb3&skt=2024-08-23T18%3A20%3A28Z&ske=2024-08-24T18%3A20%3A28Z&sks=b&skv=2024-08-04&sig=HXgLmH4hDFvbuV/rPEX//ifP0PLXUUlCYP9ZDpFEAHM%3D
{post_id} Value of Post ID is 11620
Prompt is
Task: Upload the image from the provided URL to WordPress and set it as the featured image for the specified post.
Image URL: {image_url}
Post ID: {post_id}
This worked on 1.0.15, but after I upgraded to 1.0.16 the second variable is not passed on, it repeats the image one.
Component output
Task: Upload the image from the provided URL to WordPress and set it as the featured image for the specified post.
Image URL: https://oaidalleapiprodscus.blob.core.windows.net/private/org-N0bzhj17kGCdvkPCuGgpUuhO/user-RwGFQKsTTGw8hOO1ResFbpwQ/img-GH9Zkmf4RabhW48gVeZslvyb.png?st=2024-08-23T19%3A00%3A31Z&se=2024-08-23T21%3A00%3A31Z&sp=r&sv=2024-08-04&sr=b&rscd=inline&rsct=image/png&skoid=d505667d-d6c1-4a0a-bac7-5c84a87759f8&sktid=a48cca56-e6da-484e-a814-9c849652bcb3&skt=2024-08-23T18%3A20%3A28Z&ske=2024-08-24T18%3A20%3A28Z&sks=b&skv=2024-08-04&sig=HXgLmH4hDFvbuV/rPEX//ifP0PLXUUlCYP9ZDpFEAHM%3D
Post ID: https://oaidalleapiprodscus.blob.core.windows.net/private/org-N0bzhj17kGCdvkPCuGgpUuhO/user-RwGFQKsTTGw8hOO1ResFbpwQ/img-GH9Zkmf4RabhW48gVeZslvyb.png?st=2024-08-23T19%3A00%3A31Z&se=2024-08-23T21%3A00%3A31Z&sp=r&sv=2024-08-04&sr=b&rscd=inline&rsct=image/png&skoid=d505667d-d6c1-4a0a-bac7-5c84a87759f8&sktid=a48cca56-e6da-484e-a814-9c849652bcb3&skt=2024-08-23T18%3A20%3A28Z&ske=2024-08-24T18%3A20%3A28Z&sks=b&skv=2024-08-04&sig=HXgLmH4hDFvbuV/rPEX//ifP0PLXUUlCYP9ZDpFEAHM%3D
### Reproduction
Create prompt with multiple variables.
Add value in each.
Try to build prompt and you will find that only first variable is being pulled.


### Expected behavior
multiple variables should work
### Who can help?
_No response_
### Operating System
Render
### Langflow Version
1.0.16
### Python Version
3.12
### Screenshot
_No response_
### Flow File
_No response_ | null | https://github.com/langflow-ai/langflow/pull/3698 | null | {'base_commit': 'a4dc5381b2cf31c507cc32f9027f76bf00d61ccc', 'files': [{'path': 'src/backend/base/langflow/custom/custom_component/component.py', 'status': 'modified', 'Loc': {"('Component', '__init__', 42)": {'add': [71], 'mod': [45, 56]}, "('Component', '_reset_all_output_values', 88)": {'mod': [89, 90]}, "('Component', '_build_state_model', 92)": {'mod': [98]}, "('Component', '__deepcopy__', 112)": {'mod': [119]}, "('Component', 'list_outputs', 166)": {'mod': [170]}, "('Component', 'get_output', 210)": {'mod': [223, 224]}, "('Component', 'set_output_value', 234)": {'mod': [235, 236]}, "('Component', 'map_outputs', 240)": {'mod': [253, 257]}, "('Component', 'map_inputs', 259)": {'mod': [270]}, "('Component', '_set_output_types', 290)": {'mod': [291]}, "('Component', 'get_output_by_method', 296)": {'mod': [299]}, "('Component', '_find_matching_output_method', 327)": {'mod': [329]}, "('Component', '__getattr__', 440)": {'mod': [445, 446]}, "('Component', '_set_outputs', 577)": {'mod': [581]}, "('Component', '_build_results', 619)": {'mod': [623]}}}, {'path': 'src/backend/base/langflow/graph/graph/base.py', 'status': 'modified', 'Loc': {"('Graph', '__apply_config', 318)": {'mod': [322]}}}, {'path': 'src/backend/base/langflow/template/field/base.py', 'status': 'modified', 'Loc': {"('Output', None, 161)": {'add': [180]}}}, {'path': 'src/backend/tests/unit/test_custom_component.py', 'status': 'modified', 'Loc': {"(None, 'test_custom_component_get_function_entrypoint_args_no_args', 397)": {'add': [402]}}}, {'path': 'src/backend/tests/unit/test_database.py', 'status': 'modified', 'Loc': {"(None, 'test_read_flow', 76)": {'mod': [76, 79]}}}]} | [] | [] | [] | {
"iss_type": "2",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": "",
"info_type": "Code"
} | {
"code": [
"src/backend/base/langflow/template/field/base.py",
"src/backend/base/langflow/graph/graph/base.py",
"src/backend/base/langflow/custom/custom_component/component.py"
],
"doc": [],
"test": [
"src/backend/tests/unit/test_custom_component.py",
"src/backend/tests/unit/test_database.py"
],
"config": [],
"asset": []
} | null |
langflow-ai | langflow | 20ceb42504087c712aaee41bfc17a870ae0109d4 | https://github.com/langflow-ai/langflow/issues/2039 | enhancement | [Feat] Firecrawl🔥 Integration | Hi all,
Opening this issue after chatting with Rodrigo. It would be awesome to add a [Firecrawl](https://firecrawl.dev) web loader / tool for people to use it to scrape, crawl and extract LLM ready data from the web.
Would love to hear your thoughts on how we can best integrate it.
| null | https://github.com/langflow-ai/langflow/pull/2359 | null | {'base_commit': '20ceb42504087c712aaee41bfc17a870ae0109d4', 'files': [{'path': 'poetry.lock', 'status': 'modified', 'Loc': {'(None, None, 2114)': {'add': [2114]}, '(None, None, 2436)': {'add': [2436]}, '(None, None, 2596)': {'add': [2596]}, '(None, None, 2600)': {'add': [2600]}, '(None, None, 4618)': {'add': [4618]}, '(None, None, 6082)': {'add': [6082]}, '(None, None, 2435)': {'mod': [2435]}, '(None, None, 2595)': {'mod': [2595]}, '(None, None, 2599)': {'mod': [2599]}, '(None, None, 4617)': {'mod': [4617]}, '(None, None, 6085)': {'mod': [6085]}, '(None, None, 10555)': {'mod': [10555]}}}, {'path': 'pyproject.toml', 'status': 'modified', 'Loc': {'(None, None, 94)': {'add': [94]}}}, {'path': 'src/backend/base/poetry.lock', 'status': 'modified', 'Loc': {'(None, None, 741)': {'add': [741]}, '(None, None, 3238)': {'mod': [3238]}}}, {'path': 'src/backend/base/pyproject.toml', 'status': 'modified', 'Loc': {'(None, None, 66)': {'add': [66]}}}, {'path': 'src/frontend/src/utils/styleUtils.ts', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [173, 365]}}}]} | [] | [] | [] | {
"iss_type": "4",
"iss_reason": "2",
"loc_way": "pr",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"src/frontend/src/utils/styleUtils.ts"
],
"doc": [],
"test": [],
"config": [
"pyproject.toml",
"poetry.lock",
"src/backend/base/poetry.lock",
"src/backend/base/pyproject.toml"
],
"asset": []
} | null |
langflow-ai | langflow | 9d8009f2f5c5e3fd3bf47760debc787deb454b1a | https://github.com/langflow-ai/langflow/issues/3004 | bug | Problem with Global Variables Setting Page | ### Bug Description
When entering http://127.0.0.1:7860/settings/global-variables
I am getting error in JS console.
```
`DialogContent` requires a `DialogTitle` for the component to be accessible for screen reader users.
If you want to hide the `DialogTitle`, you can wrap it with our VisuallyHidden component.
For more information, see https://radix-ui.com/primitives/docs/components/dialog [index-BMduUo-e.js:3231:165](http://127.0.0.1:7860/assets/index-BMduUo-e.js)
TitleWarning http://127.0.0.1:7860/assets/index-BMduUo-e.js:3231
Qj http://127.0.0.1:7860/assets/index-BMduUo-e.js:984
Hk http://127.0.0.1:7860/assets/index-BMduUo-e.js:984
Wk http://127.0.0.1:7860/assets/index-BMduUo-e.js:984
Pk http://127.0.0.1:7860/assets/index-BMduUo-e.js:984
Ek http://127.0.0.1:7860/assets/index-BMduUo-e.js:984
jg http://127.0.0.1:7860/assets/index-BMduUo-e.js:982
Wk http://127.0.0.1:7860/assets/index-BMduUo-e.js:984
Pk http://127.0.0.1:7860/assets/index-BMduUo-e.js:984
Gk http://127.0.0.1:7860/assets/index-BMduUo-e.js:984
Ct http://127.0.0.1:7860/assets/index-BMduUo-e.js:969
xt http://127.0.0.1:7860/assets/index-BMduUo-e.js:969
```
I am also getting error when adding again earlier deleted variable:
"Sorry, we found an unexpected error!
Please report errors with detailed tracebacks on the [GitHub Issues](https://github.com/langflow-ai/langflow/issues) page.
Thank you!"
So as asked, I am kindly reporting it.
Also, There is no feature to edit fields.
### Reproduction
JS error:
1. Just enter the page
Saving Error:
1. saving new variable
2. deleting this new variable
3. adding it again with same name
### Expected behavior
It should work without error
### Who can help?
_No response_
### Operating System
Windows 11 pro
### Langflow Version
1.13
### Python Version
None
### Screenshot
_No response_
### Flow File
_No response_ | null | https://github.com/langflow-ai/langflow/pull/3284 | null | {'base_commit': '9d8009f2f5c5e3fd3bf47760debc787deb454b1a', 'files': [{'path': 'src/backend/base/langflow/api/v1/variable.py', 'status': 'modified', 'Loc': {"(None, 'create_variable', 17)": {'add': [22], 'mod': [26, 27, 28, 29, 30, 31, 32, 33, 34, 36, 37, 39, 40, 42, 44, 46, 47, 48, 49, 50, 51, 52]}, "(None, 'read_variables', 60)": {'add': [63], 'mod': [67, 68]}, "(None, 'update_variable', 74)": {'add': [79], 'mod': [83, 84, 85, 86, 87, 89, 90, 91, 92, 93, 94, 95]}, "(None, 'delete_variable', 101)": {'add': [105], 'mod': [109, 110, 111, 112, 113, 114, 115]}, '(None, None, None)': {'mod': [1, 5, 7, 10, 11]}}}, {'path': 'src/backend/base/langflow/services/variable/base.py', 'status': 'modified', 'Loc': {"('VariableService', None, 11)": {'add': [84], 'mod': [72]}}}, {'path': 'src/backend/base/langflow/services/variable/kubernetes.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [4], 'mod': [12, 13]}, "('KubernetesSecretService', None, 16)": {'add': [123], 'mod': [113, 114, 115, 116, 117, 118]}}}, {'path': 'src/backend/base/langflow/services/variable/service.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [1], 'mod': [11]}, "('DatabaseVariableService', 'get_variable', 68)": {'add': [78], 'mod': [86, 87]}, "('DatabaseVariableService', None, 22)": {'add': [90, 111]}, "('DatabaseVariableService', 'list_variables', 91)": {'mod': [92, 93]}, "('DatabaseVariableService', 'delete_variable', 112)": {'mod': [118, 123]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "2",
"loc_way": "pr",
"loc_scope": "",
"info_type": "Code"
} | {
"code": [
"src/backend/base/langflow/api/v1/variable.py",
"src/backend/base/langflow/services/variable/base.py",
"src/backend/base/langflow/services/variable/service.py",
"src/backend/base/langflow/services/variable/kubernetes.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | null |
Significant-Gravitas | AutoGPT | 1e98d349877305a8ee9c84901282b5731675578f | https://github.com/Significant-Gravitas/AutoGPT/issues/803 | debug mode not debug | ### Duplicates
- [X] I have searched the existing issues
### Steps to reproduce 🕹
attribute in chat.py is not named correctly
### Current behavior 😯
wrong name can't call the attrb on the object
### Expected behavior 🤔
to use the attrb in a call without error
### Your prompt 📝
```yaml
# Paste your prompt here
| null | https://github.com/Significant-Gravitas/AutoGPT/pull/888 | null | {'base_commit': '1e98d349877305a8ee9c84901282b5731675578f', 'files': [{'path': 'scripts/chat.py', 'status': 'modified', 'Loc': {"(None, 'chat_with_ai', 42)": {'mod': [67, 74, 113]}}}, {'path': 'scripts/json_parser.py', 'status': 'modified', 'Loc': {"(None, 'fix_json', 76)": {'mod': [94]}}}, {'path': 'scripts/main.py', 'status': 'modified', 'Loc': {"(None, 'parse_arguments', 266)": {'add': [268]}, '(None, None, None)': {'add': [294]}}}]} | [] | [] | [] | {
"iss_type": "2",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"scripts/json_parser.py",
"scripts/main.py",
"scripts/chat.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | null | |
fastapi | fastapi | 0f152b4e97a102c0105f26d76d6e1bba3b12fc2a | https://github.com/fastapi/fastapi/issues/894 | bug
answered
reviewed | RecursionError from response model in 0.47.1 | ### Describe the bug
FastAPI 0.47.1 will not be able to start due to a `RecursionError` when there is a circular reference among models. The issue seems to originate from https://github.com/tiangolo/fastapi/pull/889. This works fine in 0.46.0.
### Environment
- OS: Windows
- FastAPI Version: 0.47.1
- Python version: 3.7.0
### To Reproduce
```Python
from typing import Optional
from fastapi import FastAPI
from pydantic import BaseModel, Field
class Group(BaseModel):
representative: Optional['Person'] = Field(None)
class Person(BaseModel):
group: Optional[Group] = Field(None)
Group.update_forward_refs()
app = FastAPI()
@app.get('/group/{group_id}', response_model=Group)
def get_group(group_id):
return []
```
### Expected behavior
No exception
### Actual output
```
Traceback (most recent call last):
File "test.py", line 21, in <module>
@app.get('/group/{group_id}', response_model=Group)
File "D:\virtualenvs\test\lib\site-packages\fastapi\routing.py", line 494, in decorator
callbacks=callbacks,
File "D:\virtualenvs\test\lib\site-packages\fastapi\routing.py", line 438, in add_api_route
callbacks=callbacks,
File "D:\virtualenvs\test\lib\site-packages\fastapi\routing.py", line 275, in __init__
] = create_cloned_field(self.response_field)
File "D:\virtualenvs\test\lib\site-packages\fastapi\utils.py", line 100, in create_cloned_field
use_type.__fields__[f.name] = create_cloned_field(f)
File "D:\virtualenvs\test\lib\site-packages\fastapi\utils.py", line 100, in create_cloned_field
use_type.__fields__[f.name] = create_cloned_field(f)
File "D:\virtualenvs\test\lib\site-packages\fastapi\utils.py", line 100, in create_cloned_field
use_type.__fields__[f.name] = create_cloned_field(f)
[Previous line repeated 981 more times]
File "D:\virtualenvs\test\lib\site-packages\fastapi\utils.py", line 97, in create_cloned_field
original_type.__name__, __config__=original_type.__config__
File "D:\virtualenvs\test\lib\site-packages\pydantic\main.py", line 773, in create_model
return type(model_name, (__base__,), namespace)
File "D:\virtualenvs\test\lib\site-packages\pydantic\main.py", line 152, in __new__
if issubclass(base, BaseModel) and base != BaseModel:
File "D:\virtualenvs\test\lib\abc.py", line 143, in __subclasscheck__
return _abc_subclasscheck(cls, subclass)
RecursionError: maximum recursion depth exceeded in comparison
``` | null | null | https://github.com/fastapi/fastapi/commit/0f152b4e97a102c0105f26d76d6e1bba3b12fc2a | {'base_commit': '0f152b4e97a102c0105f26d76d6e1bba3b12fc2a', 'files': [{'path': 'fastapi/utils.py', 'status': 'modified', 'Loc': {"(None, 'create_cloned_field', 134)": {'mod': [134, 141, 142, 143, 160, 163]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "commit",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"fastapi/utils.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | null |
oobabooga | text-generation-webui | 05676caf70db7f3715cf6a3b4680f15efd45977a | https://github.com/oobabooga/text-generation-webui/issues/6202 | bug
stale | Llama-cpp-python 0.2.81 'already loaded' fails to load models | ### Describe the bug
Attempting to load a model after running the update-wizard-macos today (the version from a day or two ago worked fine) fails with the stack trace log included below.
Notably, the error message references [this new issue in llama-cpp-python](https://github.com/abetlen/llama-cpp-python/issues/1575).
### Is there an existing issue for this?
- [X] I have searched the existing issues
### Reproduction
- Run the update wizard to update software.
- Attempt to load a gguf model using the GPU and llama.cpp
- Observe that loading fails.
### Screenshot

### Logs
```shell
Traceback (most recent call last):
File "/Users/patrickleiser/Documents/Programming/AI/text-generation-webui/modules/ui_model_menu.py", line 246, in load_model_wrapper
shared.model, shared.tokenizer = load_model(selected_model, loader)
File "/Users/patrickleiser/Documents/Programming/AI/text-generation-webui/modules/models.py", line 94, in load_model
output = load_func_map[loader](model_name)
File "/Users/patrickleiser/Documents/Programming/AI/text-generation-webui/modules/models.py", line 275, in llamacpp_loader
model, tokenizer = LlamaCppModel.from_pretrained(model_file)
File "/Users/patrickleiser/Documents/Programming/AI/text-generation-webui/modules/llamacpp_model.py", line 39, in from_pretrained
LlamaCache = llama_cpp_lib().LlamaCache
File "/Users/patrickleiser/Documents/Programming/AI/text-generation-webui/modules/llama_cpp_python_hijack.py", line 38, in llama_cpp_lib
raise Exception(f"Cannot import 'llama_cpp_cuda' because '{imported_module}' is already imported. See issue #1575 in llama-cpp-python. Please restart the server before attempting to use a different version of llama-cpp-python.")
Exception: Cannot import 'llama_cpp_cuda' because 'llama_cpp' is already imported. See issue #1575 in llama-cpp-python. Please restart the server before attempting to use a different version of llama-cpp-python.
```
### System Info
```shell
M1 Max Macbook Pro, MacOS 14.5
```
Edit: Just realized that Ooobabooga was the one that created that issue on the llama-cpp-python project, so I guess this error was already known. Sorry if this issue is therefore somewhat redundant
| null | https://github.com/oobabooga/text-generation-webui/pull/6227 | null | {'base_commit': '05676caf70db7f3715cf6a3b4680f15efd45977a', 'files': [{'path': 'modules/llama_cpp_python_hijack.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [1]}, "(None, 'llama_cpp_lib', 13)": {'mod': [16, 17, 18, 19, 20, 21, 22, 24, 26, 28, 29, 30, 31, 32, 33, 34, 35, 37, 38, 39, 40, 41, 42, 43, 44, 46, 47, 48, 49, 50, 51, 52, 53, 55, 56, 57, 58, 59, 60, 61, 62, 64, 65, 67]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"modules/llama_cpp_python_hijack.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | null |
oobabooga | text-generation-webui | 3c076c3c8096fa83440d701ba4d7d49606aaf61f | https://github.com/oobabooga/text-generation-webui/issues/2958 | bug | Latest version of Pillow breaks current implementation in html_generator.py. | ### Describe the bug
Pillow 10.0.0 removed `ANTIALIAS` from `PIL.Image`. Current implementation requires 9.5.0, however the requirements.txt currently allows for 10.0.0 to be installed.
### Is there an existing issue for this?
- [X] I have searched the existing issues
### Reproduction
Add new characters with png images and load the webui in chat mode.
### Screenshot
_No response_
### Logs
```shell
Traceback (most recent call last):
File "G:\F\Projects\AI\text-generation-webui\one-click-installers-test\installer_files\env\lib\site-packages\gradio\routes.py", line 427, in run_predict
output = await app.get_blocks().process_api(
File "G:\F\Projects\AI\text-generation-webui\one-click-installers-test\installer_files\env\lib\site-packages\gradio\blocks.py", line 1323, in process_api
result = await self.call_function(
File "G:\F\Projects\AI\text-generation-webui\one-click-installers-test\installer_files\env\lib\site-packages\gradio\blocks.py", line 1051, in call_function
prediction = await anyio.to_thread.run_sync(
File "G:\F\Projects\AI\text-generation-webui\one-click-installers-test\installer_files\env\lib\site-packages\anyio\to_thread.py", line 33, in run_sync
return await get_asynclib().run_sync_in_worker_thread(
File "G:\F\Projects\AI\text-generation-webui\one-click-installers-test\installer_files\env\lib\site-packages\anyio\_backends\_asyncio.py", line 877, in run_sync_in_worker_thread
return await future
File "G:\F\Projects\AI\text-generation-webui\one-click-installers-test\installer_files\env\lib\site-packages\anyio\_backends\_asyncio.py", line 807, in run
result = context.run(func, *args)
File "G:\F\Projects\AI\text-generation-webui\one-click-installers-test\text-generation-webui\extensions\gallery\script.py", line 71, in generate_html
image_html = f'<img src="file/{get_image_cache(path)}">'
File "G:\F\Projects\AI\text-generation-webui\one-click-installers-test\text-generation-webui\modules\html_generator.py", line 144, in get_image_cache
img = make_thumbnail(Image.open(path))
File "G:\F\Projects\AI\text-generation-webui\one-click-installers-test\text-generation-webui\modules\html_generator.py", line 132, in make_thumbnail
image = ImageOps.fit(image, (350, 470), Image.ANTIALIAS)
AttributeError: module 'PIL.Image' has no attribute 'ANTIALIAS'
```
### System Info
```shell
Windows 10
GPU: GTX 1080ti
```
| null | https://github.com/oobabooga/text-generation-webui/pull/2954 | null | {'base_commit': '3c076c3c8096fa83440d701ba4d7d49606aaf61f', 'files': [{'path': 'modules/html_generator.py', 'status': 'modified', 'Loc': {"(None, 'make_thumbnail', 129)": {'mod': [132]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": "0",
"info_type": "Other\n依赖声明 "
} | {
"code": [
"modules/html_generator.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | null |
oobabooga | text-generation-webui | cf2c4e740b1d06e145c1992515d9b34e18affc95 | https://github.com/oobabooga/text-generation-webui/issues/801 | enhancement | How can we disable Gradio analytics? | **Description**
How where / can this be implemented?
https://github.com/brkirch/stable-diffusion-webui/commit/a534959cbcabc95af50fbbe4654f8c0ee1cdd41c
`os.environ['GRADIO_ANALYTICS_ENABLED'] = 'False'`
**Additional Context**
For [stable-diffusion-webui](https://github.com/AUTOMATIC1111/stable-diffusion-webui)
[preserve privacy by disabling gradio analytics globally
#8658 ](https://github.com/AUTOMATIC1111/stable-diffusion-webui/pull/8658)
| null | null | https://github.com/oobabooga/text-generation-webui/commit/cf2c4e740b1d06e145c1992515d9b34e18affc95 | {'base_commit': 'cf2c4e740b1d06e145c1992515d9b34e18affc95', 'files': [{'path': 'server.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [0]}}}]} | [] | [] | [] | {
"iss_type": "4",
"iss_reason": "2",
"loc_way": "commit",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"server.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | null |
oobabooga | text-generation-webui | a3085dba073fe8bdcfb5120729a84560f5d024c3 | https://github.com/oobabooga/text-generation-webui/issues/1000 | bug | Bot spams random numbers or does not load | ### Describe the bug
Hello,
I installed oobabooga with the one click installer and I can not load the facebook_opt-2.7b (I copied the console into the log).
I also installed the gpt4x alpaca model with the automatic installer(download-model.bat). If I chat with it, it just spams random 2 and 4 (I took a screenshot and pasted it down below). If I manually install the gpt4x model (with the help of this tutorial: https://youtu.be/nVC9D9fRyNU?t=162 ), I get the same output as the Facebook model in the log.
### Is there an existing issue for this?
- [X] I have searched the existing issues
### Reproduction
1. Automatic installer
2. let download-model.bat download anon8231489123/gpt4-x-alpaca-13b-native-4bit-128g or follow this tutorial:
3. let download-model.bat download one model from the list
4. start-webui.bat has following arguments: python server.py --auto-devices --chat --wbits 4 --groupsize 128
### Screenshot

### Logs
```shell
Starting the web UI...
===================================BUG REPORT===================================
Welcome to bitsandbytes. For bug reports, please submit your error trace to: https://github.com/TimDettmers/bitsandbytes/issues
================================================================================
CUDA SETUP: CUDA runtime path found: C:\Users\Caner\Documents\oobabooga-windows\installer_files\env\bin\cudart64_110.dll
CUDA SETUP: Highest compute capability among GPUs detected: 8.6
CUDA SETUP: Detected CUDA version 117
CUDA SETUP: Loading binary C:\Users\Caner\Documents\oobabooga-windows\installer_files\env\lib\site-packages\bitsandbytes\libbitsandbytes_cuda117.dll...
The following models are available:
1. anon8231489123_gpt4-x-alpaca-13b-native-4bit-128g
2. facebook_opt-2.7b
Which one do you want to load? 1-2
2
Loading facebook_opt-2.7b...
Could not find the quantized model in .pt or .safetensors format, exiting...
Drücken Sie eine beliebige Taste . . .
```
### System Info
```shell
Windows 10 Version 22H2, Amd Ryzen 5800x, Palit Gamingpro Rtx 3080.
```
| null | null | https://github.com/oobabooga/text-generation-webui/commit/a3085dba073fe8bdcfb5120729a84560f5d024c3 | {'base_commit': 'a3085dba073fe8bdcfb5120729a84560f5d024c3', 'files': [{'path': 'modules/models.py', 'status': 'modified', 'Loc': {"(None, 'load_model', 40)": {'add': [176]}}}]} | [] | [] | [] | {
"iss_type": "2",
"iss_reason": "1",
"loc_way": "commit",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"modules/models.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | null |
oobabooga | text-generation-webui | c4aa1a42b156b9c5ddcfb060cc497b2fba55430f | https://github.com/oobabooga/text-generation-webui/issues/1203 | bug | When I click "Click Me" on (Character screen) it no longer generates a log of the current chat/instruction. | ### Describe the bug
When I click "Click Me" on (Character screen) it no longer generates a log of the current chat/instruction.
This either got caused when I did an update to the whole program. Reinstall does not fix. I also ran some Antivirus and registry changes. I can generate new note pads and I reinstalled notepad on my PC.
I've of course restarted my pc and I've tried firefox and opera as the host browser. This is a new problem just from today. But I did a few things on my PC.
Note pad is also missing from my right click "Create new" list. However it is is in my folders create new list. The one you can click on in the top menu settings inside explorer. The first thing is I guess I need to rule out if others having the issue or not. Next it would be nice to get some utility to auto save all logs or something as an option.
Thanks for any advice anyone. If it's a bug I will simply wait for a fix. It's possible I may have messed something but but it may also be a design flaw of the program if it's dependant on just one specific thing to save the chat log.
I have no logs or other info to share at this time.
### Is there an existing issue for this?
- [X] I have searched the existing issues
### Reproduction
For me it's easy. Issue persists even if I restart PC, update or reinstall oobabooga. It worked last night. Today it does not work.
### Screenshot
_No response_
### Logs
```shell
Not applicable.
```
### System Info
```shell
12700K, Nvidia 4080, Windows 10. Running locally on my PC not a colab etc. Like I said I tried firefox and opera. Issue seems persistent.
```
| null | null | https://github.com/oobabooga/text-generation-webui/commit/c4aa1a42b156b9c5ddcfb060cc497b2fba55430f | {'base_commit': 'c4aa1a42b156b9c5ddcfb060cc497b2fba55430f', 'files': [{'path': 'server.py', 'status': 'modified', 'Loc': {"(None, 'create_model_menus', 251)": {'mod': [324, 349]}}}]} | [] | [] | [] | {
"iss_type": "2",
"iss_reason": "1",
"loc_way": "commit",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"server.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | null |
oobabooga | text-generation-webui | 2b7ba9586fb80cfbc47c77ad7bbbb03f7d6bc0df | https://github.com/oobabooga/text-generation-webui/issues/2326 | bug | extensions/openai KeyError: 'assistant' | ### Describe the bug
Starting after [https://github.com/oobabooga/text-generation-webui/pull/2291]
Which I think it's a great improvement.
### Is there an existing issue for this?
- [X] I have searched the existing issues
### Reproduction
Start server with extension --openai --model openaccess-ai-collective_manticore-13b.
Starting [DGdev91 Auto-GPT](https://github.com/DGdev91/Auto-GPT), runs 1 cycle, give 'y' for the second, the error appears.
### Screenshot
_No response_
### Logs
```shell
openaccess-ai-collectiveException occurred during processing of request from ('127.0.0.1', 42032)
Traceback (most recent call last):
File "/home/mihai/miniconda3/envs/textgen/lib/python3.10/socketserver.py", line 683, in process_request_thread
self.finish_request(request, client_address)
File "/home/mihai/miniconda3/envs/textgen/lib/python3.10/socketserver.py", line 360, in finish_request
self.RequestHandlerClass(request, client_address, self)
File "/home/mihai/miniconda3/envs/textgen/lib/python3.10/socketserver.py", line 747, in __init__
self.handle()
File "/home/mihai/miniconda3/envs/textgen/lib/python3.10/http/server.py", line 433, in handle
self.handle_one_request()
File "/home/mihai/miniconda3/envs/textgen/lib/python3.10/http/server.py", line 421, in handle_one_request
method()
File "/home/mihai/text-generation-webui/extensions/openai/script.py", line 310, in do_POST
msg = role_formats[role].format(message=content)
KeyError: 'assistant'
----------------------------------------
```
### System Info
```shell
Win11 WSL2 Ubuntu 20.04
Python 3.10
```
| null | null | https://github.com/oobabooga/text-generation-webui/commit/2b7ba9586fb80cfbc47c77ad7bbbb03f7d6bc0df | {'base_commit': '2b7ba9586fb80cfbc47c77ad7bbbb03f7d6bc0df', 'files': [{'path': 'extensions/openai/script.py', 'status': 'modified', 'Loc': {"('Handler', 'do_POST', 159)": {'mod': [262]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "commit",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"extensions/openai/script.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | null |
hacksider | Deep-Live-Cam | 92a0994f01ec6ae7756951312a70e101fb33c7e5 | https://github.com/hacksider/Deep-Live-Cam/issues/597 | The app starts transparent! | Seems like everything was ok with the install.
When I run I get the error: / warning:
```
> python run.py --execution-provider cuda
Exception in Tkinter callback
Traceback (most recent call last):
File "C:\Users\dulci\.pyenv\pyenv-win\versions\3.10.6\lib\tkinter\__init__.py", line 1921, in __call__
return self.func(*args)
File "C:\Users\dulci\.pyenv\pyenv-win\versions\3.10.6\lib\tkinter\__init__.py", line 839, in callit
func(*args)
File "C:\Users\dulci\Documents\Development\Deep-Live-Cam\.venv\lib\site-packages\customtkinter\windows\widgets\scaling\scaling_tracker.py", line
186, in check_dpi_scaling
window.block_update_dimensions_event()
File "C:\Users\dulci\.pyenv\pyenv-win\versions\3.10.6\lib\tkinter\__init__.py", line 2383, in __getattr__
return getattr(self.tk, attr)
AttributeError: '_tkinter.tkapp' object has no attribute 'block_update_dimensions_event'
(.venv) PS C:\Users\dulci\Documents\Development\Deep-Live-Cam>
```
And the application starts so transparent (opacity super low) that I can barely see it and I can't use it, because it is almost invisible against the desktop background.
Can anybody suggest how to fix it?
I am working with venv python 3.10.6.
| null | https://github.com/hacksider/Deep-Live-Cam/pull/632 | null | {'base_commit': '92a0994f01ec6ae7756951312a70e101fb33c7e5', 'files': [{'path': 'README.md', 'status': 'modified', 'Loc': {'(None, None, 2)': {'add': [2]}, '(None, None, 3)': {'add': [3]}, '(None, None, 37)': {'add': [37]}, '(None, None, 47)': {'add': [47]}, '(None, None, 68)': {'add': [68]}, '(None, None, 168)': {'add': [168]}, '(None, None, 365)': {'add': [365]}, '(None, None, 6)': {'mod': [6]}, '(None, None, 8)': {'mod': [8]}, '(None, None, 10)': {'mod': [10]}, '(None, None, 12)': {'mod': [12]}, '(None, None, 15)': {'mod': [15]}, '(None, None, 20)': {'mod': [20]}, '(None, None, 24)': {'mod': [24]}, '(None, None, 28)': {'mod': [28]}, '(None, None, 32)': {'mod': [32]}, '(None, None, 36)': {'mod': [36]}, '(None, None, 39)': {'mod': [39, 40, 41]}, '(None, None, 43)': {'mod': [43, 44, 46]}, '(None, None, 49)': {'mod': [49, 50, 51, 52, 53, 54, 55, 56, 57]}, '(None, None, 59)': {'mod': [59]}, '(None, None, 61)': {'mod': [61, 62]}, '(None, None, 64)': {'mod': [64]}, '(None, None, 66)': {'mod': [66, 67]}, '(None, None, 71)': {'mod': [71, 72]}, '(None, None, 75)': {'mod': [75]}, '(None, None, 77)': {'mod': [77]}, '(None, None, 82)': {'mod': [82]}, '(None, None, 84)': {'mod': [84, 85, 86]}, '(None, None, 91)': {'mod': [91, 92]}, '(None, None, 96)': {'mod': [96]}, '(None, None, 98)': {'mod': [98, 100]}, '(None, None, 105)': {'mod': [105, 106]}, '(None, None, 110)': {'mod': [110]}, '(None, None, 112)': {'mod': [112, 113]}, '(None, None, 118)': {'mod': [118, 119]}, '(None, None, 123)': {'mod': [123]}, '(None, None, 125)': {'mod': [125, 126]}, '(None, None, 131)': {'mod': [131, 132]}, '(None, None, 136)': {'mod': [136]}, '(None, None, 138)': {'mod': [138, 139]}, '(None, None, 144)': {'mod': [144, 145]}, '(None, None, 150)': {'mod': [150, 151]}, '(None, None, 153)': {'mod': [153, 154]}, '(None, None, 156)': {'mod': [156]}, '(None, None, 158)': {'mod': [158, 159, 160, 161, 162]}, '(None, None, 164)': {'mod': [164]}, '(None, None, 166)': {'mod': [166, 167]}, '(None, None, 170)': {'mod': [170]}, '(None, None, 197)': {'mod': [197]}, '(None, None, 206)': {'mod': [206]}, '(None, None, 210)': {'mod': [210]}, '(None, None, 224)': {'mod': [224]}, '(None, None, 237)': {'mod': [237]}, '(None, None, 247)': {'mod': [247]}, '(None, None, 274)': {'mod': [274]}, '(None, None, 306)': {'mod': [306]}, '(None, None, 314)': {'mod': [314]}, '(None, None, 343)': {'mod': [343, 344]}, '(None, None, 346)': {'mod': [346, 347]}, '(None, None, 353)': {'mod': [353]}, '(None, None, 360)': {'mod': [360]}}}, {'path': 'modules/ui.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [6, 548], 'mod': [10, 13, 18, 19, 22, 23, 24, 25, 27, 28, 29, 30, 32, 33, 34, 35, 37, 38]}, "(None, 'analyze_target', 154)": {'add': [167], 'mod': [155, 163, 166]}, "(None, 'create_source_target_popup', 176)": {'add': [182], 'mod': [191, 192, 200, 201, 203, 204, 206, 207, 210, 211, 214, 215, 217, 218, 221, 222, 224]}, "(None, 'update_popup_source', 221)": {'add': [233], 'mod': [228, 229, 238, 240, 241, 242, 243, 245, 246, 249, 250]}, "(None, 'select_source_path', 290)": {'add': [299, 302], 'mod': [294]}, "(None, 'swap_faces_paths', 305)": {'add': [323, 326]}, "(None, 'select_target_path', 329)": {'add': [338, 343, 346], 'mod': [333]}, "(None, 'update_preview', 435)": {'add': [445], 'mod': [437, 438, 441, 443, 444, 450]}, "(None, 'init', 61)": {'mod': [61]}, "(None, 'create_root', 70)": {'mod': [70, 73, 74, 75, 77, 78, 79, 80, 81, 83, 84, 86, 87, 89, 90, 92, 93, 95, 96, 99, 100, 103, 104, 106, 107, 108, 109, 112, 113, 116, 117, 119, 121, 122, 124, 125, 126, 129, 130, 132, 133, 135, 136, 138, 139, 141, 142, 144, 145, 147, 148, 149, 150]}, "(None, 'on_submit_click', 184)": {'mod': [189]}, "(None, 'on_button_click', 194)": {'mod': [195, 197, 198]}, "(None, 'create_preview', 258)": {'mod': [263, 265, 268, 269, 271]}, "(None, 'select_output_path', 349)": {'mod': [353, 355]}, "(None, 'check_and_ignore_nsfw', 364)": {'mod': [365, 367, 370, 372, 375, 376, 378]}, "(None, 'fit_image_to_size', 381)": {'mod': [390]}, "(None, 'toggle_preview', 417)": {'mod': [418]}, "(None, 'init_preview', 425)": {'mod': [431]}, "(None, 'create_webcam_preview', 463)": {'mod': [466, 467, 468, 469, 471, 484, 487, 512]}, "(None, 'on_submit_click', 527)": {'mod': [533]}, "(None, 'create_source_target_popup_for_webcam', 519)": {'mod': [540, 543, 546]}, "(None, 'refresh_data', 550)": {'mod': [553, 554, 563, 565, 566, 568, 569, 571, 572, 575, 576, 579, 580, 581, 584, 585, 588, 589, 590]}, "(None, 'update_webcam_source', 593)": {'mod': [596, 600, 601, 610, 612, 613, 614, 615, 617, 618, 621, 622, 623, 624]}, "(None, 'update_webcam_target', 629)": {'mod': [629, 632, 636, 637, 646, 648, 649, 650, 651, 653, 654, 657, 658, 659, 660]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"modules/ui.py"
],
"doc": [
"README.md"
],
"test": [],
"config": [],
"asset": []
} | null | |
Textualize | rich | 4be4f6bd24d2a35da0e50df943209ad24c068159 | https://github.com/Textualize/rich/issues/388 | enhancement
done | [REQUEST] minimal table width | hey @willmcgugan, great package!
in `Table.add_column` method it would be nice to have an _actual_ minimal with option. [The documentation](https://github.com/willmcgugan/rich/blob/c98bf070e4f3785dbb050b72c09663021c5b1d73/rich/table.py#L303) says that `width` argument sets the minimal with for a column, but in fact in my tests it sets a constant width for one. I'd like my column to be at least `width` wide but expand if there is a longer string to display. | null | https://github.com/Textualize/rich/pull/391 | null | {'base_commit': '4be4f6bd24d2a35da0e50df943209ad24c068159', 'files': [{'path': 'CHANGELOG.md', 'status': 'modified', 'Loc': {'(None, None, 26)': {'add': [26]}}}, {'path': 'rich/console.py', 'status': 'modified', 'Loc': {"('Console', 'rule', 991)": {'mod': [993]}}}, {'path': 'rich/measure.py', 'status': 'modified', 'Loc': {"('Measurement', None, 11)": {'add': [45]}, "('Measurement', 'with_maximum', 34)": {'mod': [41]}}}, {'path': 'rich/table.py', 'status': 'modified', 'Loc': {"('Column', None, 29)": {'add': [50, 54]}, "('Table', '__init__', 118)": {'add': [123, 157]}, "('Table', 'add_column', 278)": {'add': [288, 303, 321]}, "('Table', '_calculate_column_widths', 410)": {'add': [417], 'mod': [458, 459]}, "('Table', '_measure_column', 558)": {'add': [587]}, "('Table', '__rich_measure__', 241)": {'mod': [249, 254, 255, 265]}, '(None, None, None)': {'mod': [740, 741, 743, 745, 746, 747, 748, 749, 750, 751, 753, 754, 755, 756, 757, 758, 759, 761]}}}, {'path': 'tests/test_measure.py', 'status': 'modified', 'Loc': {"(None, 'test_measure_renderables', 29)": {'add': [33]}}}, {'path': 'tests/test_table.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [3, 125]}}}]} | [] | [] | [] | {
"iss_type": "4",
"iss_reason": "2",
"loc_way": "pr",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"rich/measure.py",
"rich/console.py",
"rich/table.py"
],
"doc": [
"CHANGELOG.md"
],
"test": [
"tests/test_table.py",
"tests/test_measure.py"
],
"config": [],
"asset": []
} | null |
ytdl-org | youtube-dl | 9fa57892790ce205634f6a7c83de2b9e52ab5284 | https://github.com/ytdl-org/youtube-dl/issues/8799 | site-support-request
account-needed | Request support site: Viceland | Viceland is a new channel from Vice. Website at https://www.viceland.com. Appears to use Uplynk, and may be encrypted, so it may not be possible.
| null | null | https://github.com/ytdl-org/youtube-dl/commit/9fa57892790ce205634f6a7c83de2b9e52ab5284 | {'base_commit': '9fa57892790ce205634f6a7c83de2b9e52ab5284', 'files': [{'path': 'youtube_dl/extractor/uplynk.py', 'status': 'modified', 'Loc': {"('UplynkIE', None, 13)": {'add': [51], 'mod': [29, 30]}, "('UplynkIE', '_real_extract', 52)": {'mod': [53]}, "('UplynkPreplayIE', '_real_extract', 59)": {'mod': [64]}}}, {'path': 'youtube_dl/extractor/viceland.py', 'status': 'modified', 'Loc': {"('VicelandIE', None, 20)": {'add': [27]}}}]} | [] | [] | [] | {
"iss_type": "4",
"iss_reason": "2",
"loc_way": "commit",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"youtube_dl/extractor/uplynk.py",
"youtube_dl/extractor/viceland.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | null |
ytdl-org | youtube-dl | 5dbe81a1d35ae704b5ea208698a6bb785923d71a | https://github.com/ytdl-org/youtube-dl/issues/8171 | Vimeo ondemand download preiview only | I try to download a video from vimeo on demand but i only get the preview of.
could someone help me plaese
youtube-dl -u .com -p https://vimeo.com/ondemand/thelastcolony/150274832 --verbo
[debug] System config: []
[debug] User config: []
[debug] Command-line args: [u'-u', u'PRIVATE', u'-p', u'PRIVATE', u'https://vimeo.com/ondemand/thelastcolony/150274832', u'--verbo']
[debug] Encodings: locale cp1252, fs mbcs, out cp437, pref cp1252
[debug] youtube-dl version 2016.01.01
[debug] Python version 2.7.10 - Windows-8-6.2.9200
[debug] exe versions: none
[debug] Proxy map: {}
[vimeo] Logging in
[vimeo] 150274832: Downloading webpage
[vimeo] 150274832: Extracting information
[vimeo] 150274832: Downloading webpage
[vimeo] 150274832: Downloading JSON metadata
[vimeo] 150274832: Downloading m3u8 information
[debug] Invoking downloader on u'https://01-lvl3-gcs-pdl.vimeocdn.com/vimeo-prod-skyfire-std-us/01/54/6/150274832/459356950.mp4?expires=1452223995&token=00c3c5830ebe84f9310d4'
[download] The Last Colony-150274832.mp4 has already been downloaded
[download] 100% of 119.97MiB
| null | null | https://github.com/ytdl-org/youtube-dl/commit/5dbe81a1d35ae704b5ea208698a6bb785923d71a | {'base_commit': '5dbe81a1d35ae704b5ea208698a6bb785923d71a', 'files': [{'path': 'youtube_dl/extractor/vimeo.py', 'status': 'modified', 'Loc': {"('VimeoIE', '_real_extract', 264)": {'add': [356], 'mod': [265, 267, 269, 345]}, "('VimeoIE', '_extract_vimeo_url', 214)": {'mod': [220]}}}]} | [] | [] | [] | {
"iss_type": "2",
"iss_reason": "5",
"loc_way": "commit",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"youtube_dl/extractor/vimeo.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | null | |
ytdl-org | youtube-dl | 365343131d752bece96d2279a3e0bcd7e9f0000f | https://github.com/ytdl-org/youtube-dl/issues/17728 | [PluralSight] Unable to download captions JSON: HTTP Error 404: Not Found | last version I tested
youtube-dl is up-to-date (2018.09.26)
I'm try to download video from PluralSight. Video is ok but subtitle is cannot download. The error is
WARNING: Unable to download captions JSON: HTTP Error 404: Not Found
My command:
youtube-dl --username xxx --password xxxx --sleep-interval 35 --max-sleep-interval 120 --sub-lang en --sub-format srt --write-sub https://app.pluralsight.com/library/courses/design-database-structure-sql-server-2014-70-465/table-of-contents
Thanks. | null | null | https://github.com/ytdl-org/youtube-dl/commit/365343131d752bece96d2279a3e0bcd7e9f0000f | {'base_commit': '365343131d752bece96d2279a3e0bcd7e9f0000f', 'files': [{'path': 'youtube_dl/extractor/pluralsight.py', 'status': 'modified', 'Loc': {"('PluralsightIE', None, 112)": {'mod': [213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224]}, "('PluralsightIE', '_real_extract', 271)": {'mod': [416]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "commit",
"loc_scope": "",
"info_type": "Code"
} | {
"code": [
"youtube_dl/extractor/pluralsight.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | null | |
ytdl-org | youtube-dl | 50f84a9ae171233c08ada41e03f6555c5ed95236 | https://github.com/ytdl-org/youtube-dl/issues/7427 | "ERROR: Signature extraction failed" for youtube video | Hi,
I have encountered an error with a video:
https://www.youtube.com/watch?v=LDvVYqUMuJ0
```
$ /tmp/ydl/youtube-dl/youtube-dl https://www.youtube.com/watch?v=LDvVYqUMuJ0 --verbose
[debug] System config: []
[debug] User config: []
[debug] Command-line args: [u'https://www.youtube.com/watch?v=LDvVYqUMuJ0', u'--verbose']
[debug] Encodings: locale UTF-8, fs UTF-8, out UTF-8, pref UTF-8
[debug] youtube-dl version 2015.11.02
[debug] Python version 2.7.10+ - Linux-4.2.0-1-amd64-x86_64-with-debian-stretch-sid
[debug] exe versions: ffmpeg 1.0.6, ffprobe 1.0.6, rtmpdump 2.4
[debug] Proxy map: {}
[youtube] LDvVYqUMuJ0: Downloading webpage
[youtube] LDvVYqUMuJ0: Downloading video info webpage
[youtube] LDvVYqUMuJ0: Extracting video information
WARNING: unable to extract html5 player; please report this issue on https://yt-dl.org/bug . Make sure you are using the latest version; type youtube-dl -U to update. Be sure to call youtube-dl with the --verbose flag and include its complete output.
[youtube] {22} signature length 40.44, html5 player None
ERROR: Signature extraction failed: Traceback (most recent call last):
File "/tmp/ydl/youtube-dl/youtube-dl/youtube_dl/extractor/youtube.py", line 817, in _decrypt_signature
video_id, player_url, s
File "/tmp/ydl/youtube-dl/youtube-dl/youtube_dl/extractor/youtube.py", line 709, in _extract_signature_function
raise ExtractorError('Cannot identify player %r' % player_url)
ExtractorError: Cannot identify player u'https://s.ytimg.com/yts/jsbin/player-en_US-vfljDEtYP/base.js'; please report this issue on https://yt-dl.org/bug . Make sure you are using the latest version; type youtube-dl -U to update. Be sure to call youtube-dl with the --verbose flag and include its complete output.
(caused by ExtractorError(u"Cannot identify player u'https://s.ytimg.com/yts/jsbin/player-en_US-vfljDEtYP/base.js'; please report this issue on https://yt-dl.org/bug . Make sure you are using the latest version; type youtube-dl -U to update. Be sure to call youtube-dl with the --verbose flag and include its complete output.",)); please report this issue on https://yt-dl.org/bug . Make sure you are using the latest version; type youtube-dl -U to update. Be sure to call youtube-dl with the --verbose flag and include its complete output.
Traceback (most recent call last):
File "/tmp/ydl/youtube-dl/youtube-dl/youtube_dl/extractor/youtube.py", line 817, in _decrypt_signature
video_id, player_url, s
File "/tmp/ydl/youtube-dl/youtube-dl/youtube_dl/extractor/youtube.py", line 709, in _extract_signature_function
raise ExtractorError('Cannot identify player %r' % player_url)
ExtractorError: Cannot identify player u'https://s.ytimg.com/yts/jsbin/player-en_US-vfljDEtYP/base.js'; please report this issue on https://yt-dl.org/bug . Make sure you are using the latest version; type youtube-dl -U to update. Be sure to call youtube-dl with the --verbose flag and include its complete output.
Traceback (most recent call last):
File "/tmp/ydl/youtube-dl/youtube-dl/youtube_dl/YoutubeDL.py", line 661, in extract_info
ie_result = ie.extract(url)
File "/tmp/ydl/youtube-dl/youtube-dl/youtube_dl/extractor/common.py", line 290, in extract
return self._real_extract(url)
File "/tmp/ydl/youtube-dl/youtube-dl/youtube_dl/extractor/youtube.py", line 1345, in _real_extract
encrypted_sig, video_id, player_url, age_gate)
File "/tmp/ydl/youtube-dl/youtube-dl/youtube_dl/extractor/youtube.py", line 827, in _decrypt_signature
'Signature extraction failed: ' + tb, cause=e)
ExtractorError: Signature extraction failed: Traceback (most recent call last):
File "/tmp/ydl/youtube-dl/youtube-dl/youtube_dl/extractor/youtube.py", line 817, in _decrypt_signature
video_id, player_url, s
File "/tmp/ydl/youtube-dl/youtube-dl/youtube_dl/extractor/youtube.py", line 709, in _extract_signature_function
raise ExtractorError('Cannot identify player %r' % player_url)
ExtractorError: Cannot identify player u'https://s.ytimg.com/yts/jsbin/player-en_US-vfljDEtYP/base.js'; please report this issue on https://yt-dl.org/bug . Make sure you are using the latest version; type youtube-dl -U to update. Be sure to call youtube-dl with the --verbose flag and include its complete output.
(caused by ExtractorError(u"Cannot identify player u'https://s.ytimg.com/yts/jsbin/player-en_US-vfljDEtYP/base.js'; please report this issue on https://yt-dl.org/bug . Make sure you are using the latest version; type youtube-dl -U to update. Be sure to call youtube-dl with the --verbose flag and include its complete output.",)); please report this issue on https://yt-dl.org/bug . Make sure you are using the latest version; type youtube-dl -U to update. Be sure to call youtube-dl with the --verbose flag and include its complete output.
```
Thanks,
Corey
| null | null | https://github.com/ytdl-org/youtube-dl/commit/50f84a9ae171233c08ada41e03f6555c5ed95236 | {'base_commit': '50f84a9ae171233c08ada41e03f6555c5ed95236', 'files': [{'path': 'youtube_dl/extractor/youtube.py', 'status': 'modified', 'Loc': {"('YoutubeIE', '_extract_signature_function', 704)": {'mod': [706]}, "('YoutubeIE', '_real_extract', 1008)": {'mod': [1346]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "commit",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"youtube_dl/extractor/youtube.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | null | |
ytdl-org | youtube-dl | ea1f5e5dbd6c58d4f0872a65b97611732f4b29bd | https://github.com/ytdl-org/youtube-dl/issues/16139 | fixed | ITV BTCC videos support? | ## Please follow the guide below
- You will be asked some questions and requested to provide some information, please read them **carefully** and answer honestly
- Put an `x` into all the boxes [ ] relevant to your *issue* (like this: `[x]`)
- Use the *Preview* tab to see what your issue will actually look like
---
### Make sure you are using the *latest* version: run `youtube-dl --version` and ensure your version is *2018.04.09*. If it's not, read [this FAQ entry](https://github.com/rg3/youtube-dl/blob/master/README.md#how-do-i-update-youtube-dl) and update. Issues with outdated version will be rejected.
- [*] I've **verified** and **I assure** that I'm running youtube-dl **2018.04.09**
### Before submitting an *issue* make sure you have:
- [*] At least skimmed through the [README](https://github.com/rg3/youtube-dl/blob/master/README.md), **most notably** the [FAQ](https://github.com/rg3/youtube-dl#faq) and [BUGS](https://github.com/rg3/youtube-dl#bugs) sections
- [*] [Searched](https://github.com/rg3/youtube-dl/search?type=Issues) the bugtracker for similar issues including closed ones
- [*] Checked that provided video/audio/playlist URLs (if any) are alive and playable in a browser
### What is the purpose of your *issue*?
- [ ] Bug report (encountered problems with youtube-dl)
- [*] Site support request (request for adding support for a new site)
- [ ] Feature request (request for a new functionality)
- [ ] Question
- [ ] Other
---
### The following sections concretize particular purposed issues, you can erase any section (the contents between triple ---) not applicable to your *issue*
---
### If the purpose of this *issue* is a *bug report*, *site support request* or you are not completely sure provide the full verbose output as follows:
Add the `-v` flag to **your command line** you run youtube-dl with (`youtube-dl -v <your command line>`), copy the **whole** output and insert it here. It should look similar to one below (replace it with **your** log inserted between triple ```):
```
$ youtube-dl -v https://pastebin.com/raw/KxD6rhpF --geo-bypass-country UK
[debug] System config: []
[debug] User config: []
[debug] Custom config: []
[debug] Command-line args: [u'-v', u'https://pastebin.com/raw/KxD6rhpF', u'--geo-bypass-country', u'UK']
[debug] Encodings: locale UTF-8, fs UTF-8, out UTF-8, pref UTF-8
[debug] youtube-dl version 2018.04.09
[debug] Python version 2.7.13 (CPython) - Linux-4.9.62-v7+-armv7l-with-debian-9.4
[debug] exe versions: ffmpeg 3.2.10-1, ffprobe 3.2.10-1
[debug] Proxy map: {}
[debug] Using fake IP None (UK) as X-Forwarded-For.
[generic] KxD6rhpF: Requesting header
WARNING: Falling back on generic information extractor.
[generic] KxD6rhpF: Downloading webpage
[generic] KxD6rhpF: Extracting information
[download] Downloading playlist: Brightcove video tester
[generic] playlist Brightcove video tester: Collected 1 video ids (downloading 1 of them)
[download] Downloading video 1 of 1
[debug] Using fake IP None (UK) as X-Forwarded-For.
[debug] Using fake IP None (UK) as X-Forwarded-For.
[brightcove:new] 5766870719001: Downloading webpage
[brightcove:new] 5766870719001: Downloading JSON metadata
ERROR: Access to this resource is forbidden by access policy.
You might want to use a VPN or a proxy server (with --proxy) to workaround.
Traceback (most recent call last):
File "/usr/local/lib/python2.7/dist-packages/youtube_dl/extractor/brightcove.py", line 706, in _real_extract
json_data = self._download_json(api_url, video_id, headers=headers)
File "/usr/local/lib/python2.7/dist-packages/youtube_dl/extractor/common.py", line 692, in _download_json
encoding=encoding, data=data, headers=headers, query=query)
File "/usr/local/lib/python2.7/dist-packages/youtube_dl/extractor/common.py", line 634, in _download_webpage
res = self._download_webpage_handle(url_or_request, video_id, note, errnote, fatal, encoding=encoding, data=data, headers=headers, query=query)
File "/usr/local/lib/python2.7/dist-packages/youtube_dl/extractor/adobepass.py", line 1332, in _download_webpage_handle
*args, **compat_kwargs(kwargs))
File "/usr/local/lib/python2.7/dist-packages/youtube_dl/extractor/common.py", line 539, in _download_webpage_handle
urlh = self._request_webpage(url_or_request, video_id, note, errnote, fatal, data=data, headers=headers, query=query)
File "/usr/local/lib/python2.7/dist-packages/youtube_dl/extractor/common.py", line 528, in _request_webpage
raise ExtractorError(errmsg, sys.exc_info()[2], cause=err)
ExtractorError: Unable to download JSON metadata: HTTP Error 403: Forbidden (caused by HTTPError()); please report this issue on https://yt-dl.org/bug . Make sure you are using the latest version; see https://yt-dl.org/update on how to update. Be sure to call youtube-dl with the --verbose flag and include its complete output.
Traceback (most recent call last):
File "/usr/local/lib/python2.7/dist-packages/youtube_dl/YoutubeDL.py", line 789, in extract_info
ie_result = ie.extract(url)
File "/usr/local/lib/python2.7/dist-packages/youtube_dl/extractor/common.py", line 440, in extract
ie_result = self._real_extract(url)
File "/usr/local/lib/python2.7/dist-packages/youtube_dl/extractor/brightcove.py", line 712, in _real_extract
self.raise_geo_restricted(msg=message)
File "/usr/local/lib/python2.7/dist-packages/youtube_dl/extractor/common.py", line 743, in raise_geo_restricted
raise GeoRestrictedError(msg, countries=countries)
GeoRestrictedError: Access to this resource is forbidden by access policy.
```
---
### If the purpose of this *issue* is a *site support request* please provide all kinds of example URLs support for which should be included (replace following example URLs by **yours**):
ITV separated the BTCC race videos from the hub (which also seems to be having issues as per https://github.com/rg3/youtube-dl/issues/15925)
Lately the video are hosted at http://www.itv.com/btcc/races (ie for a particular weekend all videos are posted at individual pages like: http://www.itv.com/btcc/races/btcc-2018-all-the-action-from-brands-hatch)
Skimming the source code of this sample weekend page, I extracted the vid params and built a test page:
- Single video: https://pastebin.com/raw/KxD6rhpF
**Question 1:** is the log error above pointing just to a geo restriction error or is there anything else involved that I missed? (ie: like writing some header to force the ITV scrapper to act instead of a generic one)
```
<!doctype html>
<html>
<head>
<meta charset="UTF-8">
<title>Brightcove video tester</title>
<style>
.video-js {
height: 360px;
width: 640px;
}
.video-js.vjs-mouse .vjs-big-play-button {
background-color: red;
opacity: .3;
}
</style>
<script src="https://ajax.googleapis.com/ajax/libs/jquery/3.2.1/jquery.min.js"></script>
</head>
<body>
<video data-id="id3809566268703" data-video-id="5766870719001" data-account="1582188683001" data-player="HkiHLnNRx" data-embed="default" data-application-id="" data-fc-id="87459" class="video-js" controls="">
</video>
<script src="https://players.brightcove.net/1582188683001/HkiHLnNRx_default/index.min.js">
</script>
</body>
</html>
```
**Question 2:** is there any way to generate a playlist of downloadable items based on pages like http://www.itv.com/btcc/races/btcc-2018-all-the-action-from-brands-hatch with youtube-dl? | null | null | https://github.com/ytdl-org/youtube-dl/commit/ea1f5e5dbd6c58d4f0872a65b97611732f4b29bd | {'base_commit': 'ea1f5e5dbd6c58d4f0872a65b97611732f4b29bd', 'files': [{'path': 'youtube_dl/extractor/extractors.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [480]}}}, {'path': 'youtube_dl/extractor/itv.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [9, 20]}, "('ITVIE', '_real_extract', 56)": {'add': [262]}}}]} | [] | [] | [] | {
"iss_type": "4",
"iss_reason": "2",
"loc_way": "commit",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"youtube_dl/extractor/itv.py",
"youtube_dl/extractor/extractors.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | null |
localstack | localstack | 78f635ad3a8f819645f3991dfd244ff09f06a7f0 | https://github.com/localstack/localstack/issues/8833 | type: bug
aws:cloudformation
status: backlog | bug: CDK Table build with replicationRegions failing on latest | ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Current Behavior
Trying to deploy a [CDK Table](https://docs.aws.amazon.com/cdk/api/v2/docs/aws-cdk-lib.aws_dynamodb.Table.html) to a localstack environment results in the following:
```
stack | 0/3 | 8:56:58 PM | CREATE_FAILED | AWS::CloudFormation::Stack | stack Waiter StackCreateComplete failed: Waiter encountered a terminal failure state: For expression "Stacks[].StackStatus" we matched expected path: "CREATE_FAILED" at least once
```
```
localstack | 2023-08-06T03:56:28.709 DEBUG --- [ asgi_gw_4] localstack.packages.api : Installation of dynamodb-local skipped (already installed).
localstack | 2023-08-06T03:56:28.709 DEBUG --- [ asgi_gw_4] l.services.dynamodb.server : Starting DynamoDB Local: ['java', '-Xmx256m', '-javaagent:/usr/lib/localstack/dynamodb-local/latest/ddb-local-loader-0.1.jar', '-Djava.library.path=/usr/lib/localstack/dynamodb-local/latest/DynamoDBLocal_lib', '-jar', '/usr/lib/localstack/dynamodb-local/latest/DynamoDBLocal.jar', '-port', '35799', '-dbPath', '/var/lib/localstack/tmp/state/dynamodb']
localstack | 2023-08-06T03:56:28.710 DEBUG --- [uncthread160] localstack.utils.run : Executing command: ['java', '-Xmx256m', '-javaagent:/usr/lib/localstack/dynamodb-local/latest/ddb-local-loader-0.1.jar', '-Djava.library.path=/usr/lib/localstack/dynamodb-local/latest/DynamoDBLocal_lib', '-jar', '/usr/lib/localstack/dynamodb-local/latest/DynamoDBLocal.jar', '-port', '35799', '-dbPath', '/var/lib/localstack/tmp/state/dynamodb']
localstack | 2023-08-06T03:56:28.939 DEBUG --- [uncthread160] l.services.dynamodb.server : Initializing DynamoDB Local with the following configuration:
localstack | 2023-08-06T03:56:28.939 DEBUG --- [uncthread160] l.services.dynamodb.server : Port: 35799
localstack | 2023-08-06T03:56:28.939 DEBUG --- [uncthread160] l.services.dynamodb.server : InMemory: false
localstack | 2023-08-06T03:56:28.939 DEBUG --- [uncthread160] l.services.dynamodb.server : DbPath: /var/lib/localstack/tmp/state/dynamodb
localstack | 2023-08-06T03:56:28.940 DEBUG --- [uncthread160] l.services.dynamodb.server : SharedDb: false
localstack | 2023-08-06T03:56:28.940 DEBUG --- [uncthread160] l.services.dynamodb.server : shouldDelayTransientStatuses: false
localstack | 2023-08-06T03:56:28.940 DEBUG --- [uncthread160] l.services.dynamodb.server : CorsParams: null
localstack | 2023-08-06T03:56:28.950 DEBUG --- [uncthread160] l.services.dynamodb.server :
localstack | 2023-08-06T03:56:29.715 INFO --- [ asgi_gw_4] botocore.credentials : Found credentials in environment variables.
localstack | 2023-08-06T03:56:30.532 DEBUG --- [ asgi_gw_4] l.services.plugins : checking service health dynamodb:4566
localstack | 2023-08-06T03:56:30.534 INFO --- [ asgi_gw_4] localstack.utils.bootstrap : Execution of "require" took 1887.18ms
localstack | 2023-08-06T03:56:30.879 DEBUG --- [ asgi_gw_1] l.services.plugins : checking service health kinesis:4566
localstack | 2023-08-06T03:56:30.886 INFO --- [ asgi_gw_1] l.s.k.kinesis_mock_server : Creating kinesis backend for account 000000000000
localstack | 2023-08-06T03:56:30.887 DEBUG --- [ asgi_gw_1] localstack.packages.api : Starting installation of kinesis-local...
localstack | 2023-08-06T03:56:30.887 DEBUG --- [ asgi_gw_1] localstack.utils.run : Executing command: ['npm', 'install', '--prefix', '/var/lib/localstack/lib/kinesis-local/0.4.2', 'kinesis-local@0.4.2']
localstack | 2023-08-06T03:56:32.575 DEBUG --- [ asgi_gw_1] localstack.packages.core : Setting ownership root:root on /var/lib/localstack/lib/kinesis-local/0.4.2
localstack | 2023-08-06T03:56:32.575 DEBUG --- [ asgi_gw_1] localstack.packages.api : Installation of kinesis-local finished.
localstack | 2023-08-06T03:56:32.576 DEBUG --- [ asgi_gw_1] l.s.k.kinesis_mock_server : starting kinesis process ['node', PosixPath('/var/lib/localstack/lib/kinesis-local/0.4.2/node_modules/kinesis-local/main.js')] with env vars {'KINESIS_MOCK_CERT_PATH': '/var/lib/localstack/lib/kinesis-local/0.4.2/node_modules/kinesis-local/server.json', 'KINESIS_MOCK_PLAIN_PORT': 42209, 'KINESIS_MOCK_TLS_PORT': 34279, 'SHARD_LIMIT': '100', 'ON_DEMAND_STREAM_COUNT_LIMIT': '10', 'AWS_ACCOUNT_ID': '000000000000', 'CREATE_STREAM_DURATION': '500ms', 'DELETE_STREAM_DURATION': '500ms', 'REGISTER_STREAM_CONSUMER_DURATION': '500ms', 'START_STREAM_ENCRYPTION_DURATION': '500ms', 'STOP_STREAM_ENCRYPTION_DURATION': '500ms', 'DEREGISTER_STREAM_CONSUMER_DURATION': '500ms', 'MERGE_SHARDS_DURATION': '500ms', 'SPLIT_SHARD_DURATION': '500ms', 'UPDATE_SHARD_COUNT_DURATION': '500ms', 'UPDATE_STREAM_MODE_DURATION': '500ms', 'SHOULD_PERSIST_DATA': 'true', 'PERSIST_PATH': '../../../var/lib/localstack/tmp/state/kinesis', 'PERSIST_FILE_NAME': '000000000000.json', 'PERSIST_INTERVAL': '5s', 'LOG_LEVEL': 'INFO'}
localstack | 2023-08-06T03:56:32.576 DEBUG --- [uncthread166] localstack.utils.run : Executing command: ['node', PosixPath('/var/lib/localstack/lib/kinesis-local/0.4.2/node_modules/kinesis-local/main.js')]
localstack | 2023-08-06T03:56:32.834 INFO --- [uncthread166] l.s.k.kinesis_mock_server : [info] kinesis.mock.KinesisMockService$ 2023-08-06T03:56:32.823005Z contextId=6956dd23-c61e-4aa1-80ce-b8bfc8d0894b, cacheConfig={"awsAccountId":"000000000000","awsRegion":"us-east-1","createStreamDuration":{"length":500,"unit":"MILLISECONDS"},"deleteStreamDuration":{"length":500,"unit":"MILLISECONDS"},"deregisterStreamConsumerDuration":{"length":500,"unit":"MILLISECONDS"},"initializeStreams":null,"logLevel":"INFO","mergeShardsDuration":{"length":500,"unit":"MILLISECONDS"},"onDemandStreamCountLimit":10,"persistConfig":{"fileName":"000000000000.json","interval":{"length":5,"unit":"SECONDS"},"loadIfExists":true,"path":"../../../var/lib/localstack/tmp/state/kinesis","shouldPersist":true},"registerStreamConsumerDuration":{"length":500,"unit":"MILLISECONDS"},"shardLimit":100,"splitShardDuration":{"length":500,"unit":"MILLISECONDS"},"startStreamEncryptionDuration":{"length":500,"unit":"MILLISECONDS"},"stopStreamEncryptionDuration":{"length":500,"unit":"MILLISECONDS"},"updateShardCountDuration":{"length":500,"unit":"MILLISECONDS"}} Logging Cache Config
localstack | 2023-08-06T03:56:32.986 INFO --- [uncthread166] l.s.k.kinesis_mock_server : [info] kinesis.mock.KinesisMockService$ 2023-08-06T03:56:32.986197Z Starting Kinesis TLS Mock Service on port 34279
localstack | 2023-08-06T03:56:32.987 INFO --- [uncthread166] l.s.k.kinesis_mock_server : [info] kinesis.mock.KinesisMockService$ 2023-08-06T03:56:32.986862Z Starting Kinesis Plain Mock Service on port 42209
localstack | 2023-08-06T03:56:32.994 INFO --- [uncthread166] l.s.k.kinesis_mock_server : [info] kinesis.mock.KinesisMockService$ 2023-08-06T03:56:32.994001Z contextId=1d81ef53-1648-4fbc-8b16-f09375d77ece Starting persist data loop
localstack | 2023-08-06T03:56:33.215 DEBUG --- [uncthread158] l.s.c.resource_provider : Executing callback method for AWS::DynamoDB::Table:ddbFooTabletable735E488F
localstack | 2023-08-06T03:56:33.330 DEBUG --- [uncthread158] l.s.c.e.template_deployer : Error applying changes for CloudFormation stack "stack-ddbFooTableNestedStackdd-f1d922ad": 'NoneType' object has no attribute 'get' Traceback (most recent call last):
localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/engine/template_deployer.py", line 965, in _run
localstack | self.do_apply_changes_in_loop(changes, stack)
localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/engine/template_deployer.py", line 1011, in do_apply_changes_in_loop
localstack | should_deploy = self.prepare_should_deploy_change(
localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/engine/template_deployer.py", line 1105, in prepare_should_deploy_change
localstack | resolve_refs_recursively(
localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack_ext/services/cloudformation/cloudformation_extended.py.enc", line 35, in resolve_refs_recursively
localstack | A=resolve_refs_recursively_orig(*E,**F)
localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/utils/functions.py", line 80, in func
localstack | return wrapped(*args, **kwargs)
localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/engine/template_deployer.py", line 178, in resolve_refs_recursively
localstack | result = _resolve_refs_recursively(
localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/utils/functions.py", line 80, in func
localstack | return wrapped(*args, **kwargs)
localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/engine/template_deployer.py", line 478, in _resolve_refs_recursively
localstack | value[key] = resolve_refs_recursively(
localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack_ext/services/cloudformation/cloudformation_extended.py.enc", line 35, in resolve_refs_recursively
localstack | A=resolve_refs_recursively_orig(*E,**F)
localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/utils/functions.py", line 80, in func
localstack | return wrapped(*args, **kwargs)
localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/engine/template_deployer.py", line 178, in resolve_refs_recursively
localstack | result = _resolve_refs_recursively(
localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/utils/functions.py", line 80, in func
localstack | return wrapped(*args, **kwargs)
localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/engine/template_deployer.py", line 478, in _resolve_refs_recursively
localstack | value[key] = resolve_refs_recursively(
localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack_ext/services/cloudformation/cloudformation_extended.py.enc", line 35, in resolve_refs_recursively
localstack | A=resolve_refs_recursively_orig(*E,**F)
localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/utils/functions.py", line 80, in func
localstack | return wrapped(*args, **kwargs)
localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/engine/template_deployer.py", line 178, in resolve_refs_recursively
localstack | result = _resolve_refs_recursively(
localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/utils/functions.py", line 80, in func
localstack | return wrapped(*args, **kwargs)
localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/engine/template_deployer.py", line 497, in _resolve_refs_recursively
localstack | value[i] = resolve_refs_recursively(
localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack_ext/services/cloudformation/cloudformation_extended.py.enc", line 35, in resolve_refs_recursively
localstack | A=resolve_refs_recursively_orig(*E,**F)
localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/utils/functions.py", line 80, in func
localstack | return wrapped(*args, **kwargs)
localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/engine/template_deployer.py", line 178, in resolve_refs_recursively
localstack | result = _resolve_refs_recursively(
localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/utils/functions.py", line 80, in func
localstack | return wrapped(*args, **kwargs)
localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/engine/template_deployer.py", line 290, in _resolve_refs_recursively
localstack | resolved_getatt = get_attr_from_model_instance(
localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/engine/template_deployer.py", line 102, in get_attr_from_model_instance
localstack | attribute = attribute.get(part)
localstack | AttributeError: 'NoneType' object has no attribute 'get'
localstack |
localstack |
localstack | 2023-08-06T03:56:33.426 INFO --- [ asgi_gw_3] localstack.request.aws : AWS cloudformation.DescribeStackEvents => 200
localstack | 2023-08-06T03:56:33.445 INFO --- [ asgi_gw_2] localstack.request.aws : AWS cloudformation.DescribeStacks => 200
localstack | 2023-08-06T03:56:38.447 INFO --- [ asgi_gw_4] localstack.request.aws : AWS cloudformation.DescribeStackEvents => 200
localstack | 2023-08-06T03:56:38.458 INFO --- [ asgi_gw_3] localstack.request.aws : AWS cloudformation.DescribeStacks => 200
localstack | 2023-08-06T03:56:43.464 INFO --- [ asgi_gw_2] localstack.request.aws : AWS cloudformation.DescribeStackEvents => 200
localstack | 2023-08-06T03:56:43.473 INFO --- [ asgi_gw_4] localstack.request.aws : AWS cloudformation.DescribeStacks => 200
localstack | 2023-08-06T03:56:48.479 INFO --- [ asgi_gw_2] localstack.request.aws : AWS cloudformation.DescribeStackEvents => 200
localstack | 2023-08-06T03:56:48.489 INFO --- [ asgi_gw_4] localstack.request.aws : AWS cloudformation.DescribeStacks => 200
localstack | 2023-08-06T03:56:53.495 INFO --- [ asgi_gw_3] localstack.request.aws : AWS cloudformation.DescribeStackEvents => 200
localstack | 2023-08-06T03:56:53.502 INFO --- [ asgi_gw_2] localstack.request.aws : AWS cloudformation.DescribeStacks => 200
localstack | 2023-08-06T03:56:58.487 INFO --- [ asgi_gw_3] localstack.request.aws : AWS cloudformation.DescribeStackEvents => 200
localstack | 2023-08-06T03:56:58.495 INFO --- [ asgi_gw_2] localstack.request.aws : AWS cloudformation.DescribeStacks => 200
localstack | 2023-08-06T03:56:58.743 DEBUG --- [uncthread154] l.s.c.e.template_deployer : Error applying changes for CloudFormation stack "stack": Waiter StackCreateComplete failed: Waiter encountered a terminal failure state: For expression "Stacks[].StackStatus" we matched expected path: "CREATE_FAILED" at least once Traceback (most recent call last):
localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/engine/template_deployer.py", line 965, in _run
localstack | self.do_apply_changes_in_loop(changes, stack)
localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/engine/template_deployer.py", line 1039, in do_apply_changes_in_loop
localstack | self.apply_change(change, stack=stack)
localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/engine/template_deployer.py", line 1152, in apply_change
localstack | progress_event = executor.deploy_loop(resource_provider_payload) # noqa
localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/resource_provider.py", line 572, in deploy_loop
localstack | event = self.execute_action(resource_provider, payload)
localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/resource_provider.py", line 638, in execute_action
localstack | return resource_provider.create(request)
localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/resource_provider.py", line 350, in create
localstack | return self.create_or_delete(request)
localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/resource_provider.py", line 499, in create_or_delete
localstack | result_handler(
localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/localstack/services/cloudformation/models/cloudformation.py", line 55, in _handle_result
localstack | connect_to().cloudformation.get_waiter("stack_create_complete").wait(
localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/botocore/waiter.py", line 55, in wait
localstack | Waiter.wait(self, **kwargs)
localstack | File "/opt/code/localstack/.venv/lib/python3.10/site-packages/botocore/waiter.py", line 375, in wait
localstack | raise WaiterError(
localstack | botocore.exceptions.WaiterError: Waiter StackCreateComplete failed: Waiter encountered a terminal failure state: For expression "Stacks[].StackStatus" we matched expected path: "CREATE_FAILED" at least once
localstack |
localstack |
localstack | 2023-08-06T03:57:03.504 INFO --- [ asgi_gw_3] localstack.request.aws : AWS cloudformation.DescribeStackEvents => 200
localstack | 2023-08-06T03:57:03.511 INFO --- [ asgi_gw_2] localstack.request.aws : AWS cloudformation.DescribeStacks => 200
localstack | 2023-08-06T03:57:03.520 INFO --- [ asgi_gw_4] localstack.request.aws : AWS cloudformation.DescribeStackEvents => 200
```
I am using the CDK Table class with a stream enabled, which requires replicas to enable the stream. There's no creative way around this issue that I'm aware of, due to the stack failing before the fully deployment of the Table class.
### Expected Behavior
I'm expecting the build to succeed locally, so I can continue with our stack deployment chain for our local environments.
The behavior is not present on `localstack/localstack-pro:2.2.0` or any release prior to the update this week. I am successfully able to deploy to live environments with the same CDK stack without error.
### How are you starting LocalStack?
With a docker-compose file
### Steps To Reproduce
#### How are you starting localstack (e.g., `bin/localstack` command, arguments, or `docker-compose.yml`)
docker-compose up
```
version: "3.8"
networks:
ls:
name: ls
services:
localstack:
container_name: "${LOCALSTACK_DOCKER_NAME-localstack}"
environment:
- DEBUG=${DEBUG-1}
- DISABLE_CORS_CHECKS=1
- DISABLE_CUSTOM_CORS_APIGATEWAY=1
- DOCKER_HOST=unix:///var/run/docker.sock
- EXTRA_CORS_ALLOWED_ORIGINS=*
- MAIN_DOCKER_NETWORK=ls
- PERSISTENCE=${PERSISTENCE-}
env_file:
- ./localstack.local.env
image: "localstack/localstack-pro:${LOCALSTACK_VERSION-latest}"
networks:
- ls
ports:
- "127.0.0.1:4566:4566" # LocalStack Gateway
- "127.0.0.1:4510-4559:4510-4559" # external services port range
- "127.0.0.1:53:53" # DNS config (required for Pro)
- "127.0.0.1:53:53/udp" # DNS config (required for Pro)
- "127.0.0.1:443:443" # LocalStack HTTPS Gateway (required for Pro)
volumes:
- "/var/run/docker.sock:/var/run/docker.sock"
```
.env file contains:
```
LOCALSTACK_API_KEY=xxxxxxxxxx
```
#### Client commands (e.g., AWS SDK code snippet, or sequence of "awslocal" commands)
npx cdklocal deploy api
```
const account = process.env.CDK_ACCOUNT || "000000000000";
const region = "us-west-2";
new APIStack(app, "api", {
crossRegionReferences: true,
env: { account, region },
stackName: "api",
});
```
```
new DynamoStack(this, "ddbFooTable", {
billingMode: BillingMode.PROVISIONED,
deletionProtection: false,
encryption: TableEncryption.AWS_MANAGED,
partitionKey: { name: "id", type: AttributeType.STRING },
pointInTimeRecovery: false,
removalPolicy: false
? RemovalPolicy.RETAIN
: RemovalPolicy.DESTROY,
replicationRegions: ["us-west-1"],
stream: StreamViewType.NEW_AND_OLD_IMAGES,
tableName: "foo",
});
```
### Environment
```markdown
- OS: OSX 13.4
- LocalStack: latest (70e077bf43491cc0954698c1240159caa9cecc0ac6652b890b52aaf0801d5fcb)
- aws-cdk-lib: 2.90.0
- aws-cdk-local: 2.18.0
```
### Anything else?
Unfortunately, I'm in a position where I need the latest update due to Cognito User Pool domains not being functional in prior releases. I'm trying to get our devs to authenticate locally with an Oauth2 IDP flow instead of forcing them to authenticate with a live-stable environment. More information [here](https://github.com/localstack/localstack/issues/8700). | null | https://github.com/localstack/localstack/pull/8882 | null | {'base_commit': '78f635ad3a8f819645f3991dfd244ff09f06a7f0', 'files': [{'path': 'localstack/services/cloudformation/engine/template_deployer.py', 'status': 'modified', 'Loc': {"(None, 'get_attr_from_model_instance', 75)": {'add': [100, 101]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"localstack/services/cloudformation/engine/template_deployer.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | null |
localstack | localstack | f4a188b6d51155a0831a3246f1d8e4f4be835861 | https://github.com/localstack/localstack/issues/4652 | type: bug
status: triage needed | bug: LAMBDA_DOCKER_FLAGS doesn't work with -e | ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Current Behavior
Attempting to add environment variables to containers created to service lambda requests no longer works in the latest version of localstack. This works in localStack version 0.12.16
### Expected Behavior
Setting the environment variable `LAMBDA_DOCKER_FLAGS=-e TEST_VAL=True` on localstack's docker container will result in spawned containers created for serving lambda functions having the environment variable TEST_VAL set to True.
### How are you starting LocalStack?
With a docker-compose file
### Steps To Reproduce
1. Run `docker-compose up -d` with the following docker-compose.yml
```yaml
version: '2.1'
services:
localstack_ltest:
container_name: "ltest"
image: localstack/localstack:0.12.18
ports:
- "4566:4566"
environment:
- DOCKER_HOST=unix:///var/run/docker.sock
- LOCALSTACK_API_KEY=<removed>
- LAMBDA_EXECUTOR=docker-reuse
- LAMBDA_DOCKER_FLAGS=-e TEST_VAL=True
volumes:
- "/var/run/docker.sock:/var/run/docker.sock"
restart: always
```
2. Create the file logs-template.yml
```yaml
---
AWSTemplateFormatVersion: '2010-09-09'
Resources:
LambdaFunctionLogGroup:
Type: AWS::Logs::LogGroup
Properties:
RetentionInDays: 60
LogGroupName: !Join ["", ["/aws/lambda/", !Ref LambdaFunction]]
LambdaFunctionRole:
Type: AWS::IAM::Role
Properties:
AssumeRolePolicyDocument:
Version: '2012-10-17'
Statement:
- Effect: Allow
Principal:
Service:
- lambda.amazonaws.com
Action:
- sts:AssumeRole
Path: /
Policies:
- PolicyName: LambdaRolePolicy
PolicyDocument:
Statement:
- Effect: Allow
Action:
- 'logs:*'
Resource: 'arn:aws:logs:*:*:*'
LambdaFunction:
Type: AWS::Lambda::Function
Properties:
FunctionName: "test-function"
Role: !GetAtt LambdaFunctionRole.Arn
Handler: index.lambda_handler
Runtime: python3.8
Code:
ZipFile: |
import os
def lambda_handler(event, context):
print("environ: " + str(os.environ))
```
3. Run ` aws cloudformation deploy --stack-name test --template-file .\logs-template.yml --endpoint-url http://127.0.0.1:4566 --region us-east-1`
4. Run `aws --endpoint-url http://127.0.0.1:4566 --region us-east-1 lambda invoke --function-name test-function out.txt`
5. Run `aws --endpoint-url=http://localhost:4566 --region us-east-1 logs tail /aws/lambda/test-function`
6. Check for `"TEST_VAL": "True",` being in the output of the above command.
### Environment
```markdown
Current configuration (broken)
- OS: Ubuntu 20.04
- LocalStack version: 0.12.18
- LocalStack build date: 2021-09-27
- LocalStack build git hash: 00797f9e
Working configuration
- OS: Ubuntu 20.04
- LocalStack version: 0.12.16
- LocalStack Docker container id: b0137bad2045
- LocalStack build date: 2021-07-31
- LocalStack build git hash: f1262f74
```
### Anything else?
_No response_ | null | null | https://github.com/localstack/localstack/commit/f4a188b6d51155a0831a3246f1d8e4f4be835861 | {'base_commit': 'f4a188b6d51155a0831a3246f1d8e4f4be835861', 'files': [{'path': 'localstack/services/awslambda/lambda_executors.py', 'status': 'modified', 'Loc': {"('LambdaExecutorContainers', 'run_lambda_executor', 495)": {'mod': [543, 544]}}}, {'path': 'tests/integration/test_lambda.py', 'status': 'modified', 'Loc': {"('TestLambdaBaseFeatures', 'test_large_payloads', 477)": {'add': [492]}}}]} | [] | [] | [] | {
"iss_type": "2",
"iss_reason": "1",
"loc_way": "commit",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"localstack/services/awslambda/lambda_executors.py"
],
"doc": [],
"test": [
"tests/integration/test_lambda.py"
],
"config": [],
"asset": []
} | null |
localstack | localstack | dec1ba1b94153a4380cc94a0c8bd805f8922b6e3 | https://github.com/localstack/localstack/issues/3202 | status: triage needed
area: configuration | Illegal path is passed into the HEAD request during the object download | # Type of request: This is a ...
[ ] bug report
[ ] feature request
# Detailed description
I have tried to upgrade localstack to 0.11.5(and after) and use the service port 4566.
On local, I got passes to all tests we have been doing.
But on CircleCI, I got errors when download object from s3.
```
An error occurred (404) when calling the HeadObject operation: Not Found
```
It goes through to 0.11.4, so I'm sure it's a bug, but what do you think?
Would you have any advice?
## Expected behavior
My test does as below
1. upload object to s3 bucket named "test-bucket"
1. list v2 for the bucket
1. download it
So, I expected to succeed to this.
Of course, this test has been passed until localstack upgrade.
## Actual behavior
At CircleCI, I got below.
Somehow "/test-bucket/test-bucket" is being passed as a HEAD request parameter at download time.
This path is double of bucket name "/test-bucket".
```
:
2020-10-30 05:47:54,523:API: 172.20.0.5 - - [30/Oct/2020 05:47:54] "PUT /test-bucket HTTP/1.1" 200 -
2020-10-30 05:47:54,536:API: 172.20.0.5 - - [30/Oct/2020 05:47:54] "PUT /test-bucket/loadable/2020/04/06/000000_2_e462a109-916a-4b8f-b393-f5b01a6a8c10 HTTP/1.1" 200 -
2020-10-30 05:47:54,554:API: 172.20.0.5 - - [30/Oct/2020 05:47:54] "GET /test-bucket?list-type=2&max-keys=200&prefix=loadable%2F HTTP/1.1" 200 -
2020-10-30 05:47:54,566:API: 172.20.0.5 - - [30/Oct/2020 05:47:54] "GET /test-bucket/loadable/2020/04/06/000000_2_e462a109-916a-4b8f-b393-f5b01a6a8c10 HTTP/1.1" 206 -
2020-10-30 05:47:54,578:API: 172.20.0.5 - - [30/Oct/2020 05:47:54] "HEAD /test-bucket/test-bucket HTTP/1.1" 404 -
2020-10-30T05:47:54:WARNING:bootstrap.py: Thread run method <function AdaptiveThreadPool.submit.<locals>._run at 0x7f333a234f70>(None) failed: An error occurred (404) when calling the HeadObject operation: Not Found Traceback (most recent call last):
File "/opt/code/localstack/localstack/utils/bootstrap.py", line 534, in run
result = self.func(self.params)
File "/opt/code/localstack/localstack/utils/async_utils.py", line 28, in _run
return fn(*args, **kwargs)
File "/opt/code/localstack/localstack/services/generic_proxy.py", line 560, in handler
response = modify_and_forward(method=method, path=path_with_params, data_bytes=data, headers=headers,
File "/opt/code/localstack/localstack/services/generic_proxy.py", line 333, in modify_and_forward
listener_result = listener.forward_request(method=method,
File "/opt/code/localstack/localstack/services/edge.py", line 81, in forward_request
return do_forward_request(api, port, method, path, data, headers)
File "/opt/code/localstack/localstack/services/edge.py", line 86, in do_forward_request
result = do_forward_request_inmem(api, port, method, path, data, headers)
File "/opt/code/localstack/localstack/services/edge.py", line 106, in do_forward_request_inmem
response = modify_and_forward(method=method, path=path, data_bytes=data, headers=headers,
File "/opt/code/localstack/localstack/services/generic_proxy.py", line 401, in modify_and_forward
updated_response = update_listener.return_response(**kwargs)
File "/opt/code/localstack/localstack/services/s3/s3_listener.py", line 1254, in return_response
fix_range_content_type(bucket_name, path, headers, response)
File "/opt/code/localstack/localstack/services/s3/s3_listener.py", line 465, in fix_range_content_type
result = s3_client.head_object(Bucket=bucket_name, Key=key_name)
File "/opt/code/localstack/.venv/lib/python3.8/site-packages/botocore/client.py", line 357, in _api_call
return self._make_api_call(operation_name, kwargs)
File "/opt/code/localstack/.venv/lib/python3.8/site-packages/botocore/client.py", line 676, in _make_api_call
raise error_class(parsed_response, operation_name)
botocore.exceptions.ClientError: An error occurred (404) when calling the HeadObject operation: Not Found
```
# Steps to reproduce
## Command used to start LocalStack
sorry...
## Client code (AWS SDK code snippet, or sequence of "awslocal" commands)
sorry...
┆Issue is synchronized with this [Jira Task](https://localstack.atlassian.net/browse/LOC-71) by [Unito](https://www.unito.io/learn-more)
| null | https://github.com/localstack/localstack/pull/3370 | null | {'base_commit': 'dec1ba1b94153a4380cc94a0c8bd805f8922b6e3', 'files': [{'path': 'localstack/services/s3/s3_listener.py', 'status': 'modified', 'Loc': {"(None, 'uses_path_addressing', 891)": {'mod': [892]}}}, {'path': 'tests/integration/test_s3.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [36]}, "('S3ListenerTest', None, 59)": {'add': [1454]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"localstack/services/s3/s3_listener.py"
],
"doc": [],
"test": [
"tests/integration/test_s3.py"
],
"config": [],
"asset": []
} | null |
OpenInterpreter | open-interpreter | 7a53ba3dad9f7b2e31dac3fbb3162838eb9441c6 | https://github.com/OpenInterpreter/open-interpreter/issues/58 | Please support Azure OpenAI | null | null | https://github.com/openinterpreter/open-interpreter/pull/62 | null | {'base_commit': '7a53ba3dad9f7b2e31dac3fbb3162838eb9441c6', 'files': [{'path': 'interpreter/cli.py', 'status': 'modified', 'Loc': {"(None, 'cli', 4)": {'add': [28, 39]}}}, {'path': 'interpreter/interpreter.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [52]}, "('Interpreter', '__init__', 62)": {'add': [69]}, "('Interpreter', 'verify_api_key', 255)": {'add': [263], 'mod': [260, 262, 271, 272, 275, 276, 278, 279, 280, 281, 284, 285, 286, 287, 289]}, "('Interpreter', 'respond', 296)": {'add': [340], 'mod': [312, 313, 314, 315, 316, 317, 318]}}}]} | [] | [] | [] | {
"iss_type": "4",
"iss_reason": "2",
"loc_way": "pr",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"interpreter/cli.py",
"interpreter/interpreter.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | null | |
abi | screenshot-to-code | e69269d844b7089dec636516d6edb4f70911ebf6 | https://github.com/abi/screenshot-to-code/issues/54 | Support OPENAI_ API_ BASE for proxy URLs | How to add OPENAI_ API_ BASE code to use other proxy keys? | null | null | https://github.com/abi/screenshot-to-code/commit/e69269d844b7089dec636516d6edb4f70911ebf6 | {'base_commit': 'e69269d844b7089dec636516d6edb4f70911ebf6', 'files': [{'path': 'backend/image_generation.py', 'status': 'modified', 'Loc': {"(None, 'process_tasks', 8)": {'mod': [8, 9]}, "(None, 'generate_image', 23)": {'mod': [23, 24]}, "(None, 'generate_images', 63)": {'mod': [63, 90]}}}, {'path': 'backend/llm.py', 'status': 'modified', 'Loc': {"(None, 'stream_openai_response', 8)": {'mod': [9, 11]}}}, {'path': 'backend/main.py', 'status': 'modified', 'Loc': {"(None, 'stream_code_test', 62)": {'add': [75, 85, 119], 'mod': [132]}}}, {'path': 'frontend/src/App.tsx', 'status': 'modified', 'Loc': {'(None, None, 39)': {'add': [39]}}}, {'path': 'frontend/src/components/SettingsDialog.tsx', 'status': 'modified', 'Loc': {'(None, None, 78)': {'add': [78]}}}, {'path': 'frontend/src/types.ts', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [2]}}}]} | [] | [] | [] | {
"iss_type": "3",
"iss_reason": "5",
"loc_way": "commit",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"frontend/src/App.tsx",
"frontend/src/types.ts",
"backend/main.py",
"frontend/src/components/SettingsDialog.tsx",
"backend/image_generation.py",
"backend/llm.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | null | |
pytorch | pytorch | d363cf4639aacdaefbb8f69919f3c787a4519b7b | https://github.com/pytorch/pytorch/issues/38479 | triaged
module: numpy | torch.einsum does not pass equation argument to __torch_function__ API | ## 🐛 Bug
when delegating torch.einsum call to an object which implements
`__torch_function__` API the equation argument is not passed resulting in the error.
```TypeError: einsum(): argument 'equation' (position 1) must be str, not Tensor```
this was tested on pytorch 1.5.0
I've actually found the cause of this bug and have written a fix.
the following script illustrates the problem and the proposed solution
## To Reproduce
```python
import torch
class Wrapper():
def __init__(self,data):
self.data = data
def __torch_function__(self, func, types, args=(), kwargs=None):
if kwargs is None:
kwargs = {}
#unwrap inputs if necessary
def unwrap(v):
return v.data if isinstance(v,Wrapper) else v
args = map(unwrap,args)
kwargs = {k:unwrap(v) for k,v in kwargs.items()}
return func(*args, **kwargs)
# fixed einsum implementation
from torch import Tensor,_VF
from torch._overrides import has_torch_function,handle_torch_function
def fixed_einsum(equation,*operands):
if not torch.jit.is_scripting():
if any(type(t) is not Tensor for t in operands) and has_torch_function(operands):
# equation is not passed
# return handle_torch_function(einsum, operands, *operands)
return handle_torch_function(fixed_einsum, operands, equation,*operands)
if len(operands) == 1 and isinstance(operands[0], (list, tuple)):
# the old interface of passing the operands as one list argument
operands = operands[0]
# recurse incase operands contains value that has torch function
#in the original implementation this line is omitted
return fixed_einsum(equation,*operands)
return _VF.einsum(equation, operands)
if __name__ == "__main__":
print(torch.__version__)
# uncomment to use fixed einsum
# torch.einsum = fixed_einsum
#operands are wrapped
x = Wrapper(torch.randn(5))
y = Wrapper(torch.randn(4))
assert torch.allclose(torch.einsum('i,j->ij',x, y),torch.ger(x,y)) # outer product
print("works with wrapped inputs")
#old interface operands is a list
a = Wrapper(torch.randn(2,3))
b = Wrapper(torch.randn(5,3,7))
c = Wrapper(torch.randn(2,7))
assert torch.allclose(torch.einsum('ik,jkl,il->ij', [a, b, c]),torch.nn.functional.bilinear(a,c,b)) # bilinear interpolation
print("works with old API operands is list")
#equation is wrapped
As = Wrapper(torch.randn(3,2,5))
Bs = Wrapper(torch.randn(3,5,4))
equation = Wrapper('bij,bjk->bik')
assert torch.allclose(torch.einsum(equation, As, Bs),torch.matmul(As,Bs)) # batch matrix multiplication
print("works with equation wrapped")
#see that it also works with plain tensors
x = torch.randn(5)
y = torch.randn(4)
assert torch.allclose(torch.einsum('i,j->ij',x, y),torch.ger(x,y))
print("works with no wrapped values")
```
cc @albanD @mruberry | null | null | https://github.com/pytorch/pytorch/commit/d363cf4639aacdaefbb8f69919f3c787a4519b7b | {'base_commit': 'd363cf4639aacdaefbb8f69919f3c787a4519b7b', 'files': [{'path': 'test/test_overrides.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [490]}}}, {'path': 'torch/_overrides.py', 'status': 'modified', 'Loc': {"(None, 'get_testing_overrides', 144)": {'mod': [264]}}}, {'path': 'torch/functional.py', 'status': 'modified', 'Loc': {"(None, 'einsum', 222)": {'add': [300], 'mod': [297]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "5",
"loc_way": "commit",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"torch/functional.py",
"torch/_overrides.py"
],
"doc": [],
"test": [
"test/test_overrides.py"
],
"config": [],
"asset": []
} | null |
pytorch | pytorch | 81a4aeabdf9d550ceda52a5060f19568de61b265 | https://github.com/pytorch/pytorch/issues/93667 | triaged
tracker
oncall: pt2
module: dynamo | 14k github models on PyTorch 2.0 pass rates dashboard | We are weekly running dynamo-eager, dynamo-eager-fullgraph, export and inductor on 14k ```nn.Modules``` crawled from 1.4k github projects to get coverage level, find and fix bugs. This dashboard page is used to track the pass rates of different backends.
How to repro:
* Checkout https://github.com/jansel/pytorch-jit-paritybench
* Batch evaluation with different backends:
* dynamo-eager: ```python main.py --backend eager```
* dynamo-eager-fullgraph: ```python main.py --backend eager --fullgraph```
* export: ```python main.py --compile_mode export```
* inductor: ```python main.py```
* Adhoc evaluation:
* ```pytest ./generated/{filename}.py -k test_{n}``` (e.g, ```pytest ./generated/test_KunpengLi1994_VSRN.py -k test_002```)
* ```python -e ./generated/{filename}.py --backend eager```
Bugs umbrella task(#92670)
cc @ezyang @msaroufim @bdhirsh @anijain2305 @zou3519 @voznesenskym @penguinwu @EikanWang @jgong5 @Guobing-Chen @XiaobingSuper @zhuhaozhe @blzheng @wenzhe-nrv @jiayisunx @chenyang78 @aakhundov @kadeng @soumith @wconstab @ngimel @Xia-Weiwen @desertfire @davidberard98 | null | null | https://github.com/pytorch/pytorch/commit/81a4aeabdf9d550ceda52a5060f19568de61b265 | {'base_commit': '81a4aeabdf9d550ceda52a5060f19568de61b265', 'files': [{'path': 'test/dynamo/test_misc.py', 'status': 'modified', 'Loc': {"('MiscTests', None, 40)": {'add': [2965]}, "('MiscTests', 'fn', 421)": {'mod': [422]}, "('MiscTests', 'test_numel', 420)": {'mod': [425]}}}, {'path': 'torch/_dynamo/variables/tensor.py', 'status': 'modified', 'Loc': {"('TensorVariable', 'call_method', 178)": {'mod': [206]}}}, {'path': 'torch/_dynamo/variables/torch.py', 'status': 'modified', 'Loc': {"('TorchVariable', 'can_constant_fold_through', 159)": {'add': [165]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "commit",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"torch/_dynamo/variables/torch.py",
"torch/_dynamo/variables/tensor.py"
],
"doc": [],
"test": [
"test/dynamo/test_misc.py"
],
"config": [],
"asset": []
} | null |
pytorch | pytorch | aac9e5288f7a9666884705e2b716c260cb5f9afc | https://github.com/pytorch/pytorch/issues/67002 | module: windows
module: multiprocessing
triaged
skipped | DISABLED test_fs_sharing (__main__.TestMultiprocessing) | Flaky failures in the last week: https://fburl.com/scuba/opensource_ci_jobs/inmj698k. They only appear to be on windows
Platforms: rocm
cc @peterjc123 @mszhanyi @skyline75489 @nbcsm @VitalyFedyunin | null | null | https://github.com/pytorch/pytorch/commit/aac9e5288f7a9666884705e2b716c260cb5f9afc | {'base_commit': 'aac9e5288f7a9666884705e2b716c260cb5f9afc', 'files': [{'path': 'test/test_multiprocessing.py', 'status': 'modified', 'Loc': {"('TestMultiprocessing', 'test_receive', 289)": {'add': [293]}, '(None, None, None)': {'mod': [19, 27]}, "('TestMultiprocessing', 'test_fill', 259)": {'mod': [269, 270]}, "('TestMultiprocessing', None, 251)": {'mod': [361]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "2",
"loc_way": "commit",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [],
"doc": [],
"test": [
"test/test_multiprocessing.py"
],
"config": [],
"asset": []
} | null |
pytorch | pytorch | e37a22128eca7ccac6e289659587a9e1bfe6d242 | https://github.com/pytorch/pytorch/issues/15052 | oncall: jit | Tracer doesn't work with join/wait | Reported error: `RuntimeError: output 0 of traced region did not have observable data dependence with trace inputs; this probably indicates your program cannot be understood by the tracer.`
To reproduce:
```python
def test_async_script_trace(self):
class Module(torch.jit.ScriptModule):
def __init__(self):
super(Module, self).__init__(False)
@torch.jit.script_method
def forward(self, x):
future = torch.jit._fork(torch.neg, x)
outputs = []
outputs.append(torch.jit._wait(future))
return outputs
class Tuple(nn.Module):
def __init__(self):
super(Tuple, self).__init__()
self.module = Module()
def forward(self, x):
return tuple(self.module(x))
x = torch.rand(3, 4)
module = torch.jit.trace(Tuple(), (x), _force_outplace=True)
self.assertEqual(module(x), torch.neg(x)) | null | null | https://github.com/pytorch/pytorch/commit/e37a22128eca7ccac6e289659587a9e1bfe6d242 | {'base_commit': 'e37a22128eca7ccac6e289659587a9e1bfe6d242', 'files': [{'path': 'aten/src/ATen/core/jit_type.h', 'status': 'modified', 'Loc': {"(None, 'FutureType', 516)": {'add': [534]}}}, {'path': 'test/test_jit.py', 'status': 'modified', 'Loc': {"('TestAsync', None, 11055)": {'add': [11224]}}}, {'path': 'torch/csrc/jit/graph_executor.cpp', 'status': 'modified', 'Loc': {'(None, None, 505)': {'mod': [516, 530, 531, 532, 533]}}}, {'path': 'torch/csrc/jit/tracer.cpp', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [39]}}}, {'path': 'torch/csrc/jit/tracer.h', 'status': 'modified', 'Loc': {"(None, 'function', 42)": {'add': [44]}, "(None, 'tracer', 24)": {'mod': [35, 36, 37, 38]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "commit",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"torch/csrc/jit/tracer.h",
"torch/csrc/jit/tracer.cpp",
"aten/src/ATen/core/jit_type.h",
"torch/csrc/jit/graph_executor.cpp"
],
"doc": [],
"test": [
"test/test_jit.py"
],
"config": [],
"asset": []
} | null |
pytorch | pytorch | 0c091380cc03b23e68dde7368f3b910c21deb010 | https://github.com/pytorch/pytorch/issues/21680 | high priority
module: cudnn
module: nn
triaged
small | Disable nondeterministic CTCLoss from cuDNN | ## 🐛 Bug
<!-- A clear and concise description of what the bug is. -->
## To Reproduce
Steps to reproduce the behavior:
1.I i updated pytorch version and ctc,use pytorch_nightly, but in my train ,nn.CTCloss() is still zero,so,i would like to ask if the version pytorch(nightly) has been solved this problem
1.
1.
<!-- If you have a code sample, error messages, stack traces, please provide it here as well -->
## Expected behavior
<!-- A clear and concise description of what you expected to happen. -->
## Environment
Please copy and paste the output from our
[environment collection script](https://raw.githubusercontent.com/pytorch/pytorch/master/torch/utils/collect_env.py)
(or fill out the checklist below manually).
You can get the script and run it with:
```
wget https://raw.githubusercontent.com/pytorch/pytorch/master/torch/utils/collect_env.py
# For security purposes, please check the contents of collect_env.py before running it.
python collect_env.py
```
- PyTorch Version (e.g., 1.0):
- OS (e.g., Linux):
- How you installed PyTorch (`conda`, `pip`, source):
- Build command you used (if compiling from source):
- Python version:
- CUDA/cuDNN version:
- GPU models and configuration:
- Any other relevant information:
## Additional context
<!-- Add any other context about the problem here. -->
| null | null | https://github.com/pytorch/pytorch/commit/0c091380cc03b23e68dde7368f3b910c21deb010 | {'base_commit': '0c091380cc03b23e68dde7368f3b910c21deb010', 'files': [{'path': 'aten/src/ATen/native/LossCTC.cpp', 'status': 'modified', 'Loc': {"(None, 'ctc_loss', 341)": {'mod': [367]}}}]} | [] | [] | [] | {
"iss_type": "2",
"iss_reason": "2",
"loc_way": "commit",
"loc_scope": "0",
"info_type": "Code\nC"
} | {
"code": [
"aten/src/ATen/native/LossCTC.cpp"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | null |
xtekky | gpt4free | 5d08c7201fa5b4641f4277bf248c944b2c297b94 | https://github.com/xtekky/gpt4free/issues/843 | bug | permission denied | **Bug description**
I'm running basic G4F code from README:
```
import g4f
response = g4f.ChatCompletion.create(
model=g4f.models.gpt_4,
messages=[{"role": "user", "content": "hi"}],
) # alterative model setting
print(response)
```
Error:
```
Traceback (most recent call last):
File "C:\Users\IonE\Desktop\main.py", line 3, in <module>
import g4f
File "D:\Program Files\Python399\lib\site-packages\g4f\__init__.py", line 1, in <module>
from . import models
File "D:\Program Files\Python399\lib\site-packages\g4f\models.py", line 3, in <module>
from .Provider import Bard, BaseProvider, GetGpt, H2o, Liaobots, Vercel, Equing
File "D:\Program Files\Python399\lib\site-packages\g4f\Provider\__init__.py", line 6, in <module>
from .Bard import Bard
File "D:\Program Files\Python399\lib\site-packages\g4f\Provider\Bard.py", line 11, in <module>
class Bard(AsyncProvider):
File "D:\Program Files\Python399\lib\site-packages\g4f\Provider\Bard.py", line 22, in Bard
cookies: dict = get_cookies(".google.com"),
File "D:\Program Files\Python399\lib\site-packages\g4f\Provider\base_provider.py", line 45, in get_cookies
for cookie in browser_cookie3.load(cookie_domain):
File "D:\Program Files\Python399\lib\site-packages\browser_cookie3\__init__.py", line 1233, in load
for cookie in cookie_fn(domain_name=domain_name):
File "D:\Program Files\Python399\lib\site-packages\browser_cookie3\__init__.py", line 1160, in chrome
return Chrome(cookie_file, domain_name, key_file).load()
File "D:\Program Files\Python399\lib\site-packages\browser_cookie3\__init__.py", line 489, in load
with _DatabaseConnetion(self.cookie_file) as con:
File "D:\Program Files\Python399\lib\site-packages\browser_cookie3\__init__.py", line 349, in __enter__
return self.get_connection()
File "D:\Program Files\Python399\lib\site-packages\browser_cookie3\__init__.py", line 383, in get_connection
con = method()
File "D:\Program Files\Python399\lib\site-packages\browser_cookie3\__init__.py", line 374, in __get_connection_legacy
shutil.copyfile(self.__database_file, self.__temp_cookie_file)
File "D:\Program Files\Python399\lib\shutil.py", line 264, in copyfile
with open(src, 'rb') as fsrc:
PermissionError: [Errno 13] Permission denied: 'C:\\Users\\IonE\\AppData\\Roaming\\..\\Local\\Google\\Chrome\\User Data\\Default\\Network\\Cookies'
```
**Environement**
- python 3.9.9
- ukraine | null | https://github.com/xtekky/gpt4free/pull/847 | null | {'base_commit': '5d08c7201fa5b4641f4277bf248c944b2c297b94', 'files': [{'path': 'g4f/Provider/Bard.py', 'status': 'modified', 'Loc': {"('Bard', 'create_async', 17)": {'add': [33], 'mod': [22]}}}, {'path': 'g4f/Provider/Bing.py', 'status': 'modified', 'Loc': {"('Bing', 'create_async_generator', 21)": {'add': [34], 'mod': [24]}}}, {'path': 'g4f/Provider/Hugchat.py', 'status': 'modified', 'Loc': {"('Hugchat', 'create_completion', 17)": {'add': [25], 'mod': [23]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "2",
"loc_way": "pr",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"g4f/Provider/Bing.py",
"g4f/Provider/Hugchat.py",
"g4f/Provider/Bard.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | null |
xtekky | gpt4free | 3430b04f870d982d7fba34e3b9d6e5cf3bd3b847 | https://github.com/xtekky/gpt4free/issues/1003 | bug | please delete site chat.aivvm.com | **Known Issues** // delete this
please delete site `chat.aivvm.com`
**Delete site description**
Gpt4free maintainers, I am the administrator of chat.aivvm.com and request to delete this site. My website is already under high load
| null | null | https://github.com/xtekky/gpt4free/commit/3430b04f870d982d7fba34e3b9d6e5cf3bd3b847 | {'base_commit': '3430b04f870d982d7fba34e3b9d6e5cf3bd3b847', 'files': [{'path': 'README.md', 'status': 'modified', 'Loc': {'(None, None, 218)': {'mod': [218]}, '(None, None, 281)': {'mod': [281]}, '(None, None, 374)': {'mod': [374]}}}, {'path': 'etc/testing/test_chat_completion.py', 'status': 'modified', 'Loc': {"(None, 'run_async', 19)": {'mod': [22]}}}, {'path': 'g4f/Provider/__init__.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [9]}}}, {'path': 'g4f/Provider/Aivvm.py', 'status': 'renamed', 'Loc': {'(None, None, None)': {'mod': [3, 4, 5, 21, 22]}}}, {'path': 'g4f/Provider/deprecated/__init__.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [14]}}}, {'path': 'g4f/models.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [23, 47, 56, 66, 170, 171, 172, 173, 178, 179, 183, 184, 188, 189]}}}]} | [] | [] | [] | {
"iss_type": "3\n用户请求项目删除自己网址",
"iss_reason": "2",
"loc_way": "commit",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"g4f/Provider/Aivvm.py",
"g4f/Provider/deprecated/__init__.py",
"g4f/Provider/__init__.py",
"g4f/models.py"
],
"doc": [
"README.md"
],
"test": [
"etc/testing/test_chat_completion.py"
],
"config": [],
"asset": []
} | null |
Z4nzu | hackingtool | 0a4faeac9c4f93a61c937b0e57023b693beeca6f | https://github.com/Z4nzu/hackingtool/issues/188 | fixed | can not run hackingtool !!!!! SyntaxError: invalid syntax ??? | # python3 ./hackingtool.py
Traceback (most recent call last):
File "/root/hackingtool/./hackingtool.py", line 11, in <module>
from tools.ddos import DDOSTools
File "/root/hackingtool/tools/ddos.py", line 29
"sudo", "python3 ddos", method, url, socks_type5.4.1, threads, proxylist, multiple, timer])
^
SyntaxError: invalid syntax
__________________________________________________________________________________
| null | https://github.com/Z4nzu/hackingtool/pull/176 | null | {'base_commit': '0a4faeac9c4f93a61c937b0e57023b693beeca6f', 'files': [{'path': 'tools/ddos.py', 'status': 'modified', 'Loc': {"('ddos', 'run', 20)": {'mod': [29]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"tools/ddos.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | null |
Z4nzu | hackingtool | 0a4faeac9c4f93a61c937b0e57023b693beeca6f | https://github.com/Z4nzu/hackingtool/issues/177 | Help me please | Traceback (most recent call last):
File "/usr/share/doc/hackingtool/hackingtool.py", line 11, in <module>
from tools.ddos import DDOSTools
File "/usr/share/doc/hackingtool/tools/ddos.py", line 29
"sudo", "python3 ddos", method, url, socks_type5.4.1, threads, proxylist, multiple, timer])
^
SyntaxError: invalid syntax
| null | https://github.com/Z4nzu/hackingtool/pull/176 | null | {'base_commit': '0a4faeac9c4f93a61c937b0e57023b693beeca6f', 'files': [{'path': 'tools/ddos.py', 'status': 'modified', 'Loc': {"('ddos', 'run', 20)": {'mod': [29]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"tools/ddos.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | null | |
Z4nzu | hackingtool | 0a4faeac9c4f93a61c937b0e57023b693beeca6f | https://github.com/Z4nzu/hackingtool/issues/189 | look | 
| null | https://github.com/Z4nzu/hackingtool/pull/176 | null | {'base_commit': '0a4faeac9c4f93a61c937b0e57023b693beeca6f', 'files': [{'path': 'tools/ddos.py', 'status': 'modified', 'Loc': {"('ddos', 'run', 20)": {'mod': [29]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"tools/ddos.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | null | |
Z4nzu | hackingtool | 0a4faeac9c4f93a61c937b0e57023b693beeca6f | https://github.com/Z4nzu/hackingtool/issues/209 | Running Issue | **Describe the bug**
A clear and concise description of what the bug is.
I have installed this tool successfully though when i go to run it with <sudo hackingtool> or <hackingtool> it comes up with the error I have attached as a screenshot. Why would this be happening?
**To Reproduce**
Steps to reproduce the behavior:
1. Go to '...'
2. Click on '....'
3. Scroll down to '....'
4. See error
**Expected behavior**
A clear and concise description of what you expected to happen.
**Screenshots**
If applicable, add screenshots to help explain your problem.

**Desktop (please complete the following information):**
- OS: Kali
- Browser [e.g. chrome, safari]
- Version [e.g. 22]
**Smartphone (please complete the following information):**
- Device: rpi4
- OS: [e.g. iOS8.1]
- Browser [e.g. stock browser, safari]
- Version [e.g. 22]
**Additional context**
Add any other context about the problem here. | null | https://github.com/Z4nzu/hackingtool/pull/176 | null | {'base_commit': '0a4faeac9c4f93a61c937b0e57023b693beeca6f', 'files': [{'path': 'tools/ddos.py', 'status': 'modified', 'Loc': {"('ddos', 'run', 20)": {'mod': [29]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"tools/ddos.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | null | |
Z4nzu | hackingtool | 0a4faeac9c4f93a61c937b0e57023b693beeca6f | https://github.com/Z4nzu/hackingtool/issues/185 | fixed | SyntaxError: invalid syntax traceback most recent call last | ┌──(root💀localhost)-[~]
└─# hackingtool
Traceback (most recent call last):
File "/usr/share/doc/hackingtool/hackingtool.py", line 11, in <module>
from tools.ddos import DDOSTools
File "/usr/share/doc/hackingtool/tools/ddos.py", line 29
"sudo", "python3 ddos", method, url, socks_type5.4.1, threads, proxylist, multiple, timer])
^
SyntaxError: invalid syntax
this happens when i type in "hackingtool" in terminal. any fixes? | null | https://github.com/Z4nzu/hackingtool/pull/176 | null | {'base_commit': '0a4faeac9c4f93a61c937b0e57023b693beeca6f', 'files': [{'path': 'tools/ddos.py', 'status': 'modified', 'Loc': {"('ddos', 'run', 20)": {'mod': [29]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"tools/ddos.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | null |
Z4nzu | hackingtool | 0a4faeac9c4f93a61c937b0e57023b693beeca6f | https://github.com/Z4nzu/hackingtool/issues/187 | fixed | from tools.ddos import DDOSTools | # sudo hackingtool
Traceback (most recent call last):
File "/usr/share/doc/hackingtool/hackingtool.py", line 11, in <module>
from tools.ddos import DDOSTools
File "/usr/share/doc/hackingtool/tools/ddos.py", line 29
"sudo", "python3 ddos", method, url, socks_type5.4.1, threads, proxylist, multiple, timer])
^
SyntaxError: invalid syntax
| null | https://github.com/Z4nzu/hackingtool/pull/176 | null | {'base_commit': '0a4faeac9c4f93a61c937b0e57023b693beeca6f', 'files': [{'path': 'tools/ddos.py', 'status': 'modified', 'Loc': {"('ddos', 'run', 20)": {'mod': [29]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"tools/ddos.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | null |
scikit-learn | scikit-learn | bf0886bae0ccbc8c5d285b6e2affe7e40474f970 | https://github.com/scikit-learn/scikit-learn/issues/16924 | Bug
Easy
module:metrics | Matthews correlation coefficient metric throws misleading division by zero RuntimeWarning | #### Description
With tested values all equal, `sklearn.metrics.matthews_corrcoef` throws a `RuntimeWarning` reporting a division by zero. This behavior was already reported in #1937 and reported fixed, but reappears in recent versions.
#### Steps/Code to Reproduce
The snippet below reproduces the warning.
```python
import sklearn.metrics
trues = [1,0,1,1,0]
preds = [0,0,0,0,0]
sklearn.metrics.matthews_corrcoef(trues, preds)
```
#### Expected Results
No warning is thrown.
#### Actual Results
The following warning is thrown:
```
C:\anaconda\envs\sklearn-test\lib\site-packages\sklearn\metrics\_classification.py:900: RuntimeWarning: invalid value encountered in double_scalars
mcc = cov_ytyp / np.sqrt(cov_ytyt * cov_ypyp)
```
#### Versions
```
System:
python: 3.8.2 (default, Mar 25 2020, 08:56:29) [MSC v.1916 64 bit (AMD64)]
executable: C:\anaconda\envs\sklearn-test\python.exe
machine: Windows-10-10.0.18362-SP0
Python dependencies:
pip: 20.0.2
setuptools: 46.1.3.post20200330
sklearn: 0.22.1
numpy: 1.18.1
scipy: 1.4.1
Cython: None
pandas: None
matplotlib: None
joblib: 0.14.1
```
| null | https://github.com/scikit-learn/scikit-learn/pull/19977 | null | {'base_commit': 'bf0886bae0ccbc8c5d285b6e2affe7e40474f970', 'files': [{'path': 'sklearn/metrics/_classification.py', 'status': 'modified', 'Loc': {"(None, 'matthews_corrcoef', 800)": {'mod': [881, 883, 886]}}}, {'path': 'sklearn/metrics/tests/test_classification.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [23, 625]}, "(None, 'test_matthews_corrcoef', 671)": {'mod': [687, 688, 690, 691, 694, 696, 697]}, "(None, 'test_matthews_corrcoef_multiclass', 713)": {'mod': [734, 737, 738, 739, 757, 761, 762, 763, 765, 766]}}}, {'path': 'sklearn/utils/_testing.py', 'status': 'modified', 'Loc': {"(None, 'assert_warns_div0', 190)": {'mod': [190, 191, 193, 194, 196, 197, 198, 199, 200, 201, 203, 204, 205, 206, 207, 208, 209, 210, 211]}}}]} | [] | [] | [] | {
"iss_type": "2",
"iss_reason": "2",
"loc_way": "pr",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"sklearn/metrics/_classification.py",
"sklearn/utils/_testing.py"
],
"doc": [],
"test": [
"sklearn/metrics/tests/test_classification.py"
],
"config": [],
"asset": []
} | null |
scikit-learn | scikit-learn | 4ac6a90a82e4a8d7b5338c18ae8a16559c98ba10 | https://github.com/scikit-learn/scikit-learn/issues/5101 | LatentDirichletAllocation has superfluous attributes | It has `dirichlet_component_` (undocumented) and `exp_dirichlet_component_` (exponential of same). I propose to get rid of at least the latter.
| null | https://github.com/scikit-learn/scikit-learn/pull/5111 | null | {'base_commit': '4ac6a90a82e4a8d7b5338c18ae8a16559c98ba10', 'files': [{'path': 'sklearn/decomposition/online_lda.py', 'status': 'modified', 'Loc': {"('LatentDirichletAllocation', '_approx_bound', 542)": {'add': [579], 'mod': [597, 612]}, "('LatentDirichletAllocation', '_init_latent_vars', 283)": {'mod': [305, 306, 308]}, "('LatentDirichletAllocation', '_em_step', 366)": {'mod': [407, 408]}}}]} | [] | [] | [] | {
"iss_type": "1\n多余的选项",
"iss_reason": "2",
"loc_way": "pr",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"sklearn/decomposition/online_lda.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | null | |
pandas-dev | pandas | 05123af1b2f8db1bc4f05c22515ef378cbeefbd3 | https://github.com/pandas-dev/pandas/issues/76 | Bug | Sparse cumsum functions do not work | e.g. SparseSeries.cumsum
| null | null | https://github.com/wesm/pandas/commit/05123af1b2f8db1bc4f05c22515ef378cbeefbd3 | {'base_commit': '05123af1b2f8db1bc4f05c22515ef378cbeefbd3', 'files': [{'path': 'pandas/core/frame.py', 'status': 'modified', 'Loc': {"('DataFrame', None, 97)": {'mod': [1962, 1963, 1964, 1966, 1967, 1968, 1969, 1971, 1972, 1973, 1974, 1975, 1976, 1977, 1978, 1979, 1980, 1981, 1982, 1983, 1984, 1985, 2021, 2022, 2023, 2025, 2026, 2027, 2028, 2030, 2031, 2032, 2033, 2034, 2035, 2036, 2037, 2038, 2039, 2041, 2043]}}}, {'path': 'pandas/core/generic.py', 'status': 'modified', 'Loc': {"('PandasGeneric', '_reindex_axis', 162)": {'add': [168]}}}, {'path': 'pandas/core/series.py', 'status': 'modified', 'Loc': {"('Series', 'cumsum', 570)": {'mod': [580, 581, 582, 583, 584, 585, 591, 592, 593, 594]}}}, {'path': 'pandas/core/sparse.py', 'status': 'modified', 'Loc': {"('SparseSeries', None, 152)": {'add': [512]}, "('SparseDataFrame', 'count', 1058)": {'add': [1059]}, '(None, None, None)': {'mod': [13]}}}, {'path': 'pandas/tests/test_frame.py', 'status': 'modified', 'Loc': {"('TestDataFrame', None, 539)": {'add': [2271]}, "('TestDataFrame', 'test_cumsum', 2271)": {'add': [2276, 2283, 2284], 'mod': [2273, 2274, 2286, 2287]}}}, {'path': 'pandas/tests/test_sparse.py', 'status': 'modified', 'Loc': {"('TestSparseSeries', None, 111)": {'add': [577]}, "('TestSparseDataFrame', 'test_count', 1065)": {'add': [1068]}}}]} | [] | [] | [] | {
"iss_type": "2",
"iss_reason": "1",
"loc_way": "commit",
"loc_scope": "0",
"info_type": "Code"
} | {
"code": [
"pandas/core/frame.py",
"pandas/core/generic.py",
"pandas/core/series.py",
"pandas/core/sparse.py"
],
"doc": [],
"test": [
"pandas/tests/test_frame.py",
"pandas/tests/test_sparse.py"
],
"config": [],
"asset": []
} | null |
pandas-dev | pandas | 65c0441a41b2dcaeebb648274d30978419a8661a | https://github.com/pandas-dev/pandas/issues/16607 | Datetime
Compat | to_datetime should support ISO week year | `to_datetime` does not currently seem to support `ISO week year` like `strptime` does:
```
In [38]: datetime.date(2016, 1, 1).strftime('%G-%V')
Out[38]: '2015-53'
In [39]: datetime.datetime.strptime(datetime.date(2016, 1, 1).strftime('%G-%V')+'-1', '%G-%V-%u')
Out[39]: datetime.datetime(2015, 12, 28, 0, 0)
In [41]: pd.to_datetime(datetime.date(2016, 1, 1).strftime('%G-%V')+'-1', format='%G-%V-%u')
---------------------------------------------------------------------------
TypeError Traceback (most recent call last)
/Users/Robin/.pyenv/versions/3.6.1/lib/python3.6/site-packages/pandas/core/tools/datetimes.py in _convert_listlike(arg, box, format, name, tz)
443 try:
--> 444 values, tz = tslib.datetime_to_datetime64(arg)
445 return DatetimeIndex._simple_new(values, name=name, tz=tz)
pandas/_libs/tslib.pyx in pandas._libs.tslib.datetime_to_datetime64 (pandas/_libs/tslib.c:33275)()
TypeError: Unrecognized value type: <class 'str'>
During handling of the above exception, another exception occurred:
ValueError Traceback (most recent call last)
<ipython-input-41-7ce30c959690> in <module>()
----> 1 pd.to_datetime(datetime.date(2016, 1, 1).strftime('%G-%V')+'-1', format='%G-%V-%u')
/Users/Robin/.pyenv/versions/3.6.1/lib/python3.6/site-packages/pandas/core/tools/datetimes.py in to_datetime(arg, errors, dayfirst, yearfirst, utc, box, format, exact, unit, infer_datetime_format, origin)
516 result = _convert_listlike(arg, box, format)
517 else:
--> 518 result = _convert_listlike(np.array([arg]), box, format)[0]
519
520 return result
/Users/Robin/.pyenv/versions/3.6.1/lib/python3.6/site-packages/pandas/core/tools/datetimes.py in _convert_listlike(arg, box, format, name, tz)
445 return DatetimeIndex._simple_new(values, name=name, tz=tz)
446 except (ValueError, TypeError):
--> 447 raise e
448
449 if arg is None:
/Users/Robin/.pyenv/versions/3.6.1/lib/python3.6/site-packages/pandas/core/tools/datetimes.py in _convert_listlike(arg, box, format, name, tz)
412 try:
413 result = tslib.array_strptime(arg, format, exact=exact,
--> 414 errors=errors)
415 except tslib.OutOfBoundsDatetime:
416 if errors == 'raise':
pandas/_libs/tslib.pyx in pandas._libs.tslib.array_strptime (pandas/_libs/tslib.c:63124)()
pandas/_libs/tslib.pyx in pandas._libs.tslib.array_strptime (pandas/_libs/tslib.c:63003)()
ValueError: 'G' is a bad directive in format '%G-%V-%u'
```
<details>
INSTALLED VERSIONS
------------------
commit: None
pandas: 0.20.1
pytest: 3.1.0
pip: 9.0.1
setuptools: 28.8.0
Cython: 0.25.2
numpy: 1.12.1
scipy: 0.19.0
xarray: None
IPython: 6.0.0
sphinx: None
patsy: 0.4.1
dateutil: 2.6.0
pytz: 2017.2
blosc: None
bottleneck: None
tables: 3.4.2
numexpr: 2.6.2
feather: None
matplotlib: 2.0.2
openpyxl: None
xlrd: None
xlwt: None
xlsxwriter: None
lxml: None
bs4: None
html5lib: 0.999999999
sqlalchemy: 1.1.10
pymysql: None
psycopg2: 2.7.1 (dt dec pq3 ext lo64)
jinja2: 2.9.6
s3fs: None
pandas_gbq: None
pandas_datareader: None
</details>
| null | https://github.com/pandas-dev/pandas/pull/25541 | null | {'base_commit': '65c0441a41b2dcaeebb648274d30978419a8661a', 'files': [{'path': 'doc/source/whatsnew/v0.25.0.rst', 'status': 'modified', 'Loc': {'(None, None, 21)': {'add': [21]}}}, {'path': 'pandas/_libs/tslibs/strptime.pyx', 'status': 'modified', 'Loc': {'(None, None, 79)': {'add': [79]}, '(None, None, 171)': {'add': [171]}, '(None, None, 267)': {'add': [267]}, '(None, None, 513)': {'add': [513]}, '(None, None, 520)': {'add': [520]}, '(None, None, 521)': {'add': [521]}, '(None, None, 622)': {'add': [622]}, '(None, None, 57)': {'mod': [57]}, '(None, None, 178)': {'mod': [178]}, '(None, None, 271)': {'mod': [271, 272, 273, 274]}, '(None, None, 596)': {'mod': [596, 597]}, '(None, None, 600)': {'mod': [600]}}}, {'path': 'pandas/core/tools/datetimes.py', 'status': 'modified', 'Loc': {"(None, 'to_datetime', 403)": {'add': [457]}}}, {'path': 'pandas/tests/indexes/datetimes/test_tools.py', 'status': 'modified', 'Loc': {"('TestToDatetime', None, 246)": {'add': [246]}}}]} | [] | [] | [] | {
"iss_type": "4",
"iss_reason": "2",
"loc_way": "pr",
"loc_scope": "0",
"info_type": "Code/Doc"
} | {
"code": [
"pandas/core/tools/datetimes.py",
"pandas/_libs/tslibs/strptime.pyx"
],
"doc": [
"doc/source/whatsnew/v0.25.0.rst"
],
"test": [
"pandas/tests/indexes/datetimes/test_tools.py"
],
"config": [],
"asset": []
} | null |
pallets | flask | 6ed68f015a50ab35b84a8ea71b0f846ca6a75281 | https://github.com/pallets/flask/issues/3074 | send_file doesn't urlencode ':/' in unicode attachment_filename | ### Expected Behavior
When sending files with unicode filename (with `:` or `/`) they should be downloaded with name from `filename*` field.
```python
# -*- coding: utf-8 -*-
import os
from flask import Flask, send_from_directory
app = Flask(__name__)
@app.route('/test/', methods=['GET'])
def test_route():
tmp_dir = os.getcwd()
tmp_filename = __file__
attachment_filename = u'тест:тест_тест.py'
return send_from_directory(
tmp_dir,
tmp_filename,
as_attachment=True,
attachment_filename=attachment_filename
)
if __name__ == '__main__':
app.run(host='::', port=5000)
```
### Actual Behavior
Some browsers (Chrome-based/Safari) ignore `filename*` field when it contains colon or slash. For example file `тест:тест_тест.py` gets downloaded in Chrome/Safari as `__.py` but in Firefox as `тест_тест_тест.py` which is acceptable in my opinion.
Flask response:
`Content-Disposition: attachment; filename*="UTF-8''%D1%82%D0%B5%D1%81%D1%82:%D1%82%D0%B5%D1%81%D1%82_%D1%82%D0%B5%D1%81%D1%82.py"; filename=":_.py"`
### Environment
* Python version: 2.7.15
* Flask version: 1.0.2
* Werkzeug version: 0.14.1
| null | https://github.com/pallets/flask/pull/3273 | null | {'base_commit': '6ed68f015a50ab35b84a8ea71b0f846ca6a75281', 'files': [{'path': 'CHANGES.rst', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [10]}}}, {'path': 'flask/helpers.py', 'status': 'modified', 'Loc': {"(None, 'send_file', 454)": {'mod': [579]}}}, {'path': 'tests/test_helpers.py', 'status': 'modified', 'Loc': {"('TestSendfile', None, 436)": {'add': [648]}}}]} | [] | [] | [] | {
"iss_type": "2",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"flask/helpers.py"
],
"doc": [
"CHANGES.rst"
],
"test": [
"tests/test_helpers.py"
],
"config": [],
"asset": []
} | 1 | |
pallets | flask | 50b7dcbab343c93bb6738bbf116a177e72b1d9ec | https://github.com/pallets/flask/issues/4099 | docs | Harmless race condition in tutorial | I was browsing the flaskr tutorial when I noticed an (admittedly quite unlikely) race condition in the `register` view, specifically:
```py
if not username:
error = 'Username is required.'
elif not password:
error = 'Password is required.'
elif db.execute(
'SELECT id FROM user WHERE username = ?', (username,)
).fetchone() is not None:
error = f"User {username} is already registered."
if error is None:
db.execute(
'INSERT INTO user (username, password) VALUES (?, ?)',
(username, generate_password_hash(password))
)
db.commit()
return redirect(url_for('auth.login'))
```
If two requests arrive with the right timing, the following can happen:
```
Request 1: Request 2:
SELECT id
FROM user
WHERE username = abc
|
v
empty, no such user
SELECT id
FROM user
WHERE username = abc
|
v
empty, no such user
INSERT INTO user (username, password)
VALUES (abc, 123)
|
v
ok
INSERT INTO user (username, password)
VALUES (abc, 456)
|
v
failed UNIQUE constraint ->
-> sqlite3.IntegrityError ->
-> user gets HTTP 500
```
While the likelihood of this happening is pretty small and the harm practically zero (user gets HTTP 500 and has to manually login/choose a different username), I feel like this is not really the sort of good practice the tutorial should teach. I also believe it's important the developer understands that it's the UNIQUE constraint that ensures their app works correctly and not the if condition in the application code (the tutorial mentions SQL injection attacks and explains what protects the developer against them, so I don't really feel this is out of scope).
In my own app I've modified the code to the following:
```py
if not username:
error = 'Username is required.'
elif not password:
error = 'Password is required.'
else:
try:
db.execute(
'INSERT INTO users (username, password) VALUES (?, ?)',
(username, generate_password_hash(password))
)
db.commit()
except IntegrityError:
error = f"User {username} is already registered."
else:
return redirect(url_for('auth.login'))
```
I suggest something similar be incorporated into the tutorial, with a short explanation (maybe a comment) of how the UNIQUE constraint does the work for the developer and maybe a note about the principle that one should "ask forgiveness, not permission." I'm not sure on how it's better worded, so I'm making this an issue instead of a pull request.
Cheers, and thank you for your great work! | null | https://github.com/pallets/flask/pull/4139 | null | {'base_commit': '50b7dcbab343c93bb6738bbf116a177e72b1d9ec', 'files': [{'path': 'docs/tutorial/views.rst', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [202], 'mod': [94, 95, 96, 97, 100, 101, 102, 103, 104, 105, 128, 129, 130, 131, 132, 133, 134, 136, 137, 138, 139, 141, 142, 143, 144, 145, 146, 147]}}}, {'path': 'examples/tutorial/flaskr/auth.py', 'status': 'modified', 'Loc': {"(None, 'register', 47)": {'mod': [63, 64, 65, 66, 67, 70, 71, 72, 73, 74, 75, 76, 77]}}}]} | [] | [] | [] | {
"iss_type": "4",
"iss_reason": "2",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"examples/tutorial/flaskr/auth.py"
],
"doc": [
"docs/tutorial/views.rst"
],
"test": [],
"config": [],
"asset": []
} | 1 |
pallets | flask | f17e6061fcffdc290f615d3fdc9d949e9e719574 | https://github.com/pallets/flask/issues/1443 | json_encoder not invoked from flask.jsonify | I created a custom JSON encoder class extended from flask.json.JSONEncoder but it is not called when calling flask.jsonify. Additionally, I removed my custom JSON encoder and confirmed that flask.json.JSONEncoder isn't called either via a break statement in Pycharm.
```
from flask import Flask
from flask import jsonify
from flask.json import JSONEncoder
class MyEncoder(JSONEncoder):
def default(self, obj):
if hasattr(obj, '__json__'):
return obj.__json__()
else:
try:
iterable = iter(obj)
except TypeError:
pass
else:
return list(iterable)
return JSONEncoder.default(self, obj)
class MyClass(object):
key = 'a'
value = 'b'
def __json__(self):
return {'key': self.key, 'value': self.value}
app = Flask(__name__)
app.json_encoder = MyEncoder
@app.route('/')
def hello_world():
return jsonify(MyClass())
if __name__ == '__main__':
app.run(debug=True)
```
| null | https://github.com/pallets/flask/pull/1671 | null | {'base_commit': 'f17e6061fcffdc290f615d3fdc9d949e9e719574', 'files': [{'path': 'AUTHORS', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [17, 20], 'mod': [35]}}}, {'path': 'CHANGES', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [10]}}}, {'path': 'docs/security.rst', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [98, 100, 101, 102, 103, 105, 106, 107, 108, 109, 111, 112, 113, 114, 115, 116, 117, 119, 120, 121, 122, 123, 125, 127, 128, 129, 130, 132, 133, 134, 135, 137, 138, 140, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 161, 162, 163, 164, 165, 166, 167, 168, 170, 171, 172, 173, 174, 175]}}}, {'path': 'flask/json.py', 'status': 'modified', 'Loc': {"(None, 'jsonify', 201)": {'add': [244], 'mod': [202, 203, 204, 205, 225, 226, 248, 249]}}}, {'path': 'tests/test_helpers.py', 'status': 'modified', 'Loc': {"('TestJSON', 'test_json_as_unicode', 121)": {'add': [122], 'mod': [124, 125, 126, 127, 129, 130, 131, 132]}, "('TestJSON', None, 32)": {'mod': [34, 35, 37, 38, 39, 40, 42, 43, 45, 46, 47, 48, 49, 50, 106, 107, 121]}}}]} | [] | [] | [] | {
"iss_type": "2",
"iss_reason": "2",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"flask/json.py"
],
"doc": [
"docs/security.rst",
"CHANGES"
],
"test": [
"tests/test_helpers.py"
],
"config": [],
"asset": [
"AUTHORS"
]
} | 1 | |
pallets | flask | f808c20139649b747f604492bc33b61a7dd3e13a | https://github.com/pallets/flask/issues/2731 | bug
blueprints | Flask 1.0 backwards-incompat with double-slash/no-slash re. #2629 | This is a major backwards-compat breaking change, but I suspect not the intended design and hopefully easy to fix.
The issue is related to PR #2629, and this example follows from that:
Given blueprint `bp` and app `app`:
```python
@bp.route('b/')
def tmp():
return "URI should be '/a/b/"
app.register_blueprint(bp, url_prefix='/a/')
```
In Flask 0.12 the URL is correctly `/a/b`, but in Flask 1.0 it's `/ab`.
Since issue #2629 relates to resolve double-slashes, I imagine this is a bug (and not a design decision) - and the correct solution would be to remove a slash only when there are two.
| null | https://github.com/pallets/flask/pull/2629 | null | {'base_commit': 'f808c20139649b747f604492bc33b61a7dd3e13a', 'files': [{'path': 'CHANGES.rst', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [145, 188]}}}, {'path': 'flask/blueprints.py', 'status': 'modified', 'Loc': {"('BlueprintSetupState', '__init__', 25)": {'add': [55]}}}, {'path': 'tests/test_blueprints.py', 'status': 'modified', 'Loc': {"(None, 'test_blueprint_url_definitions', 117)": {'mod': [117]}}}]} | [] | [] | [] | {
"iss_type": "2",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"flask/blueprints.py"
],
"doc": [
"CHANGES.rst"
],
"test": [
"tests/test_blueprints.py"
],
"config": [],
"asset": []
} | 1 |
pallets | flask | 22708b048d224a5590fa28d86ca02bac52294f90 | https://github.com/pallets/flask/issues/2594 | cli | add ssl_context option to `flask run` | ### Expected Behaviour
I expect to be able to pass the `flask run` command any of the options which are valid for the `Flask.run()` method:
```sh
$ FLASK_APP=myapp/run.py FLASK_DEBUG=1 flask run --host=0.0.0.0 --ssl_context=adhoc
* Running on https://0.0.0.0:5000/ (Press CTRL+C to quit)
```
Specifically, I want to pass `ssl_context=adhoc`, but it seems sensible to extend the command to accept all valid keyword arguments for `Flask.run()` / `werkzeug.serving.run_simple()`.
### Actual Behaviour
```
Error: no such option: --ssl_context
flask run --host=0.0.0.0 --ssl_context=adhoc exited with code 2
```
### Environment
* Python version: 3.5.2
* Flask version: 0.12.2
* Werkzeug version: 0.12.2
| null | https://github.com/pallets/flask/pull/2606 | null | {'base_commit': '22708b048d224a5590fa28d86ca02bac52294f90', 'files': [{'path': 'CHANGES', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [120, 156]}}}, {'path': 'flask/cli.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [16, 23, 601, 606], 'mod': [26, 608, 611, 614]}, "(None, 'run_command', 619)": {'mod': [620, 645]}}}, {'path': 'tests/test_cli.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [16, 17], 'mod': [27, 28]}, "(None, 'test_dotenv_optional', 462)": {'add': [466]}}}]} | [] | [] | [] | {
"iss_type": "4",
"iss_reason": "2",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"flask/cli.py"
],
"doc": [
"CHANGES"
],
"test": [
"tests/test_cli.py"
],
"config": [],
"asset": []
} | 1 |
pallets | flask | e4c712ffd2682f963906e1d0d27e67b7f83d95ce | https://github.com/pallets/flask/issues/266 | docs
good first issue
blueprints | Blueprint template lookup not documented enough | The new blueprint template lookup scheme where the templates folder is just added to the searchpath instead of doing some weird stuff with the names as before. The documentation has to be clearer about that.
| null | https://github.com/pallets/flask/pull/1843 | null | {'base_commit': 'e4c712ffd2682f963906e1d0d27e67b7f83d95ce', 'files': [{'path': 'docs/blueprints.rst', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [179, 180, 181, 182, 183, 188]}}}]} | [] | [] | [] | {
"iss_type": "4",
"iss_reason": "2",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [],
"doc": [
"docs/blueprints.rst"
],
"test": [],
"config": [],
"asset": []
} | 1 |
pallets | flask | 8cd0b03beeac4a41c398ea365475c651c484a9ee | https://github.com/pallets/flask/issues/2118 | bug | config.from_pyfile crashes on Python 3 when source isn't encoded in default encoding | when I read my instance config file, I get an error.
> exec(compile(config_file.read(), filename, 'exec'), d.__dict__)
> UnicodeDecodeError: 'gbk' codec can't decode byte 0x80 in position 437: illegal multibyte sequence
Then I modify the code of config.from_pyfile to this
> with open(filename, 'rb') as config_file:
The problem is resolved.
| null | https://github.com/pallets/flask/pull/2123 | null | {'base_commit': '8cd0b03beeac4a41c398ea365475c651c484a9ee', 'files': [{'path': 'CHANGES', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [10]}}}, {'path': 'flask/config.py', 'status': 'modified', 'Loc': {"('Config', 'from_pyfile', 111)": {'mod': [129]}}}, {'path': 'tests/test_config.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [13, 14], 'mod': [10, 12]}, "(None, 'test_get_namespace', 168)": {'add': [189]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"flask/config.py"
],
"doc": [
"CHANGES"
],
"test": [
"tests/test_config.py"
],
"config": [],
"asset": []
} | 1 |
pallets | flask | 85fa8aabf5a7bd0adf204f0c2dacbba1fa6683de | https://github.com/pallets/flask/issues/2023 | discussion
logging | How should logging in Flask look like? | Flask started to ship with a default, hardcoded logging handler. Unfortunately this setup makes it harder to install custom logging setups, because then you'll have to undo all the things Flask did to the app logger, or replace the `app.logger` entirely. A symptom of this is #1993, where Flask's own logger had to be tweaked yet again such that messages didn't get logged twice (once via Flask's setup, once via the custom one).
My question is: **Do we even want Flask to do any logging setup?** It appears that this sort of default logging is only useful during development, so maybe it makes sense to set up a default logging handler in the new Flask CLI instead of from within the application.
| null | https://github.com/pallets/flask/pull/2436 | null | {'base_commit': '85fa8aabf5a7bd0adf204f0c2dacbba1fa6683de', 'files': [{'path': 'CHANGES', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [108]}}}, {'path': 'docs/config.rst', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [331], 'mod': [202, 204, 205, 207, 209, 211, 212, 213, 215]}}}, {'path': 'docs/contents.rst.inc', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [18]}}}, {'path': 'docs/errorhandling.rst', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [146, 147, 149, 150, 151, 152, 153, 155, 156, 157, 158, 159, 161, 162, 163, 165, 166, 167, 168, 169, 170, 171, 172, 173, 175, 176, 177, 178, 179, 180, 181, 183, 184, 185, 187, 188, 189, 192, 193, 195, 196, 197, 198, 199, 200, 202, 203, 204, 206, 207, 208, 209, 210, 211, 212, 213, 214, 216, 217, 218, 220, 221, 222, 223, 224, 225, 227, 229, 230, 232, 233, 234, 235, 236, 238, 239, 240, 242, 244, 245, 247, 249, 250, 251, 252, 253, 254, 255, 257, 259, 260, 262, 263, 265, 267, 268, 269, 270, 271, 274, 275, 277, 278, 279, 281, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 314, 315, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 329, 332, 333, 335, 336, 337, 338, 339, 340, 341, 343, 344, 345, 347, 348, 349, 350, 351, 352]}}}, {'path': 'flask/app.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [32], 'mod': [19, 20, 21, 22, 40, 41]}, "('Flask', None, 70)": {'mod': [267, 268, 269, 270, 271, 297, 298, 616]}, "('Flask', '__init__', 352)": {'mod': [395, 396, 397]}, "('Flask', 'logger', 617)": {'mod': [618, 619, 620, 621, 623, 624, 625, 629, 630, 631, 632, 633, 634, 635, 636]}}}, {'path': 'flask/logging.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [13, 41, 49], 'mod': [1, 2, 3, 4, 6, 8, 9, 10, 17, 18, 19, 22, 23, 24, 25, 26, 27, 28]}, "(None, '_proxy_stream', 32)": {'mod': [32, 33, 34, 35, 37, 38, 39, 40]}, "(None, '_should_log_for', 43)": {'mod': [43, 44, 45, 46]}, "(None, 'create_logger', 50)": {'mod': [51, 52, 53, 54, 55, 57, 59, 60, 61, 62, 63, 65, 66, 67, 68, 70, 71, 72, 73, 75, 76, 77, 79, 80, 81, 83, 84, 85, 86, 87, 88, 89, 91, 92]}}}, {'path': 'tests/test_basic.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [1023]}, "(None, 'test_teardown_request_handler_error', 739)": {'mod': [741]}, "(None, 'test_error_handling', 816)": {'mod': [817]}, "(None, 'test_error_handling_processing', 862)": {'mod': [863]}, "(None, 'test_baseexception_error_handling', 884)": {'mod': [885]}, "(None, 'apprunner', 1427)": {'mod': [1428]}}}, {'path': 'tests/test_helpers.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [12, 16, 18, 19, 22, 23, 25]}, "('TestLogging', None, 663)": {'mod': [663, 664, 665, 666, 667, 668, 669, 670, 672, 673, 675, 676, 677, 678, 679, 681, 682, 683, 685, 686, 687, 688, 689, 690, 691, 693, 694, 696, 697, 698, 699, 700, 702, 703, 704, 705, 706, 707, 709, 710, 711, 713, 714, 715, 717, 718, 719, 720, 721, 723, 724, 725, 727, 728, 729, 730, 732, 733, 734, 735, 736, 738, 739, 740, 742, 743, 744, 746, 747, 748, 749]}}}, {'path': 'tests/test_subclassing.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [13]}, "(None, 'test_suppressed_exception_logging', 18)": {'mod': [25, 26, 32, 36, 37]}, "(None, 'index', 29)": {'mod': [30]}}}, {'path': 'tests/test_templating.py', 'status': 'modified', 'Loc': {"(None, 'test_template_loader_debugging', 402)": {'mod': [402, 422, 423, 424, 425, 426, 428, 429, 431, 432, 433, 434]}}}, {'path': 'tests/test_testing.py', 'status': 'modified', 'Loc': {"(None, 'test_test_client_context_binding', 209)": {'mod': [210]}}}]} | [] | [] | [] | {
"iss_type": "4",
"iss_reason": "2",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"flask/logging.py",
"flask/app.py"
],
"doc": [
"docs/contents.rst.inc",
"docs/config.rst",
"docs/errorhandling.rst",
"CHANGES"
],
"test": [
"tests/test_templating.py",
"tests/test_testing.py",
"tests/test_basic.py",
"tests/test_helpers.py",
"tests/test_subclassing.py"
],
"config": [],
"asset": []
} | 1 |
pallets | flask | 465da9f610a04d379bb39a0ff03fb6c0b0ea1c45 | https://github.com/pallets/flask/issues/2866 | logging | DispatcherMiddleware with different loggers per app in flask 1.0 | After upgrading to flask 1.0 logging from different apps using DispatcherMiddleware, each log emitted is written to all handlers in the different apps. I assume this caused by `app.logger` always having the name `flask.app`, maybe?
Here is a example:
```
from werkzeug.wsgi import DispatcherMiddleware
from flask import Flask
from logging.handlers import RotatingFileHandler
handler1 = RotatingFileHandler('app1.log')
app1 = Flask('app1')
app1.logger.addHandler(handler1)
handler2 = RotatingFileHandler('app2.log')
app2 = Flask('app2')
app2.logger.addHandler(handler2)
@app1.route("/")
def hello():
app1.logger.error("from app1")
return ''
@app2.route("/")
def hello2():
app2.logger.error("from app2")
return ''
app = DispatcherMiddleware(app1, {
'/app2': app2
})
```
Run with
```
uwsgi --socket 0.0.0.0:8000 --protocol=http -w app --callable app
```
And then make a request to / and /app2/. Each error log will be written in both logfiles.
### Environment
* Python version: 3.6.5
* Flask version: 1.0.2
* Werkzeug version: 0.14.1
My actual app is using `current_app.logger` with blueprints with the same behaviour, but I assume it the same issue. | null | https://github.com/pallets/flask/pull/3282 | null | {'base_commit': '465da9f610a04d379bb39a0ff03fb6c0b0ea1c45', 'files': [{'path': 'CHANGES.rst', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [20]}}}, {'path': 'docs/config.rst', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [385]}}}, {'path': 'docs/errorhandling.rst', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [234]}}}, {'path': 'docs/logging.rst', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [1, 6, 7, 8, 9]}}}, {'path': 'src/flask/app.py', 'status': 'modified', 'Loc': {"('Flask', 'logger', 655)": {'mod': [656, 657, 659, 660, 662, 663, 665, 667, 668, 669, 670, 671]}}}, {'path': 'src/flask/logging.py', 'status': 'modified', 'Loc': {"(None, 'create_logger', 59)": {'mod': [60, 69]}}}, {'path': 'tests/test_logging.py', 'status': 'modified', 'Loc': {"(None, 'reset_logging', 21)": {'mod': [26]}, "(None, 'test_logger', 44)": {'mod': [45]}}}, {'path': 'tests/test_templating.py', 'status': 'modified', 'Loc': {"(None, 'test_template_loader_debugging', 409)": {'mod': [433]}}}]} | [] | [] | [] | {
"iss_type": "2",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"src/flask/logging.py",
"src/flask/app.py"
],
"doc": [
"docs/config.rst",
"docs/logging.rst",
"docs/errorhandling.rst",
"CHANGES.rst"
],
"test": [
"tests/test_templating.py",
"tests/test_logging.py"
],
"config": [],
"asset": []
} | 1 |
pallets | flask | c8cf4694c60f0d81809468a1b45ec730496cc546 | https://github.com/pallets/flask/issues/5160 | Switch to importlib breaks scripts with `app.run()` | With a trivial script [using `app.run()`](https://flask.palletsprojects.com/en/2.3.x/server/#in-code) such as:
```python3
from flask import Flask
app = Flask(__name__)
if __name__ == "__main__":
app.run(debug=True)
```
The current git `main` breaks with:
```pytb
Traceback (most recent call last):
File "/home/florian/tmp/flask/app.py", line 3, in <module>
app = Flask(__name__)
^^^^^^^^^^^^^^^
File "/home/florian/tmp/flask/src/flask/app.py", line 376, in __init__
instance_path = self.auto_find_instance_path()
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/florian/tmp/flask/src/flask/app.py", line 630, in auto_find_instance_path
prefix, package_path = find_package(self.import_name)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/florian/tmp/flask/src/flask/scaffold.py", line 898, in find_package
package_path = _find_package_path(import_name)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/florian/tmp/flask/src/flask/scaffold.py", line 858, in _find_package_path
spec = importlib.util.find_spec(root_mod_name)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "<frozen importlib.util>", line 114, in find_spec
ValueError: __main__.__spec__ is None
```
This seems to be a regression due to 84e11a1e827c0f55f9b9ee15952eddcf8a6492e0 from #5157.
Environment:
- Python version: 3.11.4
- Flask version: git main
| null | https://github.com/pallets/flask/pull/5161 | null | {'base_commit': 'c8cf4694c60f0d81809468a1b45ec730496cc546', 'files': [{'path': 'CHANGES.rst', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [7]}}}, {'path': 'src/flask/helpers.py', 'status': 'modified', 'Loc': {"(None, 'get_root_path', 562)": {'mod': [578, 579, 584]}}}, {'path': 'src/flask/scaffold.py', 'status': 'modified', 'Loc': {"(None, '_matching_loader_thinks_module_is_package', 782)": {'mod': [782, 783, 785, 786, 787, 788, 789, 790, 792, 794, 795, 796, 797, 799, 800, 801, 802, 803, 804]}, "(None, '_find_package_path', 816)": {'mod': [825, 826, 827, 828, 829, 831, 832, 833, 834, 835, 836, 837, 838, 839, 840, 841, 842, 843, 844, 845, 846, 847, 848, 849, 850, 851, 852, 853, 854, 855, 857, 858, 859, 861, 862, 865, 866, 867, 868, 869, 870, 871, 872, 873, 875, 877, 878, 879, 880, 882]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"src/flask/helpers.py",
"src/flask/scaffold.py"
],
"doc": [
"CHANGES.rst"
],
"test": [],
"config": [],
"asset": []
} | 1 | |
sherlock-project | sherlock | a2655907389f1625540c0c643aed33bc26e63da8 | https://github.com/sherlock-project/sherlock/issues/762 | false positive | linkedin false positive | null | https://github.com/sherlock-project/sherlock/pull/773 | null | {'base_commit': 'a2655907389f1625540c0c643aed33bc26e63da8', 'files': [{'path': 'removed_sites.json', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [484], 'mod': [149, 150]}}}, {'path': 'removed_sites.md', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [956]}}}, {'path': 'sherlock/resources/data.json', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [891, 892, 893, 894, 895, 896, 897, 898, 899, 1521, 1522, 1523, 1524, 1525, 1526, 1527, 1528, 2259, 2260, 2261, 2262, 2263, 2264, 2265, 2266, 2267]}}}, {'path': 'sites.md', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [1, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300]}}}]} | [] | [] | [] | {
"iss_type": "2",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"removed_sites.json",
"sherlock/resources/data.json"
],
"doc": [
"removed_sites.md",
"sites.md"
],
"test": [],
"config": [],
"asset": []
} | 1 | |
sherlock-project | sherlock | b9ff06593288bcf94211878e41e5d010eb1e00f2 | https://github.com/sherlock-project/sherlock/issues/30 | enhancement | Optional csv Result Output | While the text output file is good to have, it only shows the sites that do have an account with the username. It would be useful to also have a list of the sites that still had an opening.
I think that it would be useful to have a csv output as well. It could be optional, so the user would have to add a `--csv` switch before the file would be output. Then the output would be available in LibreOffice or Excel for users to sort and organize as they wish.
I am thinking of the following columns:
- Username
The user name that is being queried.
- Social Network Name
For example, "Twitter", "GitHub", ...
- Social Network Main URL
For example, "https://twitter.com/" for Twitter
- Social Network User URL
This is the URL that Sherlock would be trying to check the existence of. For example, "https://twitter.com/user123" for Twitter.
- Username Exists
Will be either "True" or "False".
- Request Response Data
I think that this would be useful for debug of the program. Especially for those sites where Sherlock depends on the response text (as opposed to an explicit HTTP response code). | null | https://github.com/sherlock-project/sherlock/pull/32 | null | {'base_commit': 'b9ff06593288bcf94211878e41e5d010eb1e00f2', 'files': [{'path': 'sherlock.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [0, 6, 20], 'mod': [9]}, "(None, 'print_error', 26)": {'mod': [26]}, "(None, 'make_request', 33)": {'mod': [33, 37, 46]}, "(None, 'sherlock', 49)": {'mod': [57, 58, 60, 61, 62, 63, 64, 65, 67, 68, 69, 70, 72, 73, 74, 75, 77, 79, 80, 81, 82, 83, 84, 86, 87, 89, 90, 91, 92, 93, 95, 96, 98, 99, 100, 101, 102, 103, 104, 105, 107, 108]}, "(None, 'main', 115)": {'mod': [134, 155, 157, 158]}}}]} | [] | [] | [] | {
"iss_type": "4",
"iss_reason": "2",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"sherlock.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | 1 |
sherlock-project | sherlock | e781ed5e260e1fdb58410a3103adb2b95ce547c8 | https://github.com/sherlock-project/sherlock/issues/503 | site support request | Add search "kali community" | ```
"kali community": {
"errorType": "status_code",
"rank": 10313,
"url": "https://forums.kali.org/member.php?username={}",
"urlMain": "https://forums.kali.org/",
"username_claimed": "blue",
"username_unclaimed": "noonewouldeverusethis7"
},
``` | null | https://github.com/sherlock-project/sherlock/pull/567 | null | {'base_commit': 'e781ed5e260e1fdb58410a3103adb2b95ce547c8', 'files': [{'path': 'data.json', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [288, 545, 665, 939, 974, 1374, 1390, 1628]}}}]} | [] | [] | [] | {
"iss_type": "4",
"iss_reason": "2",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"data.json"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | 1 |
sherlock-project | sherlock | 7b3cf0aaf28e0e4051dffedb6d1416b9fa9ef456 | https://github.com/sherlock-project/sherlock/issues/1305 | bug | [ISSUE] Sherlock stops after Countable search | - [X] I'm reporting a bug in Sherlock's functionality
- [X] The bug I'm reporting is not a false positive or a false negative
- [X] I've verified that I'm running the latest version of Sherlock
- [X] I've checked for similar bug reports including closed ones
- [X] I've checked for pull requests that attempt to fix this bug
## Description
Upon running `python3 sherlock username`, the checks stop right after a return on Countable. (See screenshot below)
As I am unsure of if this is a social media downtime issue, I am consulting here.
<img width="401" alt="Screen_Shot_2022-04-16_at_11 00 42_PM" src="https://user-images.githubusercontent.com/60671860/163698401-dd7c35e0-b6ec-468c-9b9d-6ac29806065c.png">
Suggested fix:
Improve 404 messages and insert timeouts if sites are not reachable within a few seconds.
| null | https://github.com/sherlock-project/sherlock/pull/1307 | null | {'base_commit': '7b3cf0aaf28e0e4051dffedb6d1416b9fa9ef456', 'files': [{'path': 'removed_sites.json', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [655]}}}, {'path': 'removed_sites.md', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [1267]}}}, {'path': 'sherlock/resources/data.json', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [545, 546, 547, 548, 549, 550, 551]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"removed_sites.json",
"sherlock/resources/data.json"
],
"doc": [
"removed_sites.md"
],
"test": [],
"config": [],
"asset": []
} | 1 |
sherlock-project | sherlock | 588616d6155edb4184e5ff8f4cf0104e0a0e2468 | https://github.com/sherlock-project/sherlock/issues/326 | bug | Codepen False Negative/Trip False Positive | 

| null | https://github.com/sherlock-project/sherlock/pull/328 | null | {'base_commit': '588616d6155edb4184e5ff8f4cf0104e0a0e2468', 'files': [{'path': 'README.md', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [71]}}}, {'path': 'data.json', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [1257, 1258]}}}, {'path': 'sherlock.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [29]}}}]} | [] | [] | [] | {
"iss_type": "2",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"sherlock.py",
"data.json"
],
"doc": [
"README.md"
],
"test": [],
"config": [],
"asset": []
} | 1 |
sherlock-project | sherlock | 508eb88724dbe20dedf07dc00527ad4f32c93a77 | https://github.com/sherlock-project/sherlock/issues/5 | bug | Formatting on windows Powershell very off | [37;1m[[92;1m+[37;1m][92;1m Unsplash:[0m https://unsplash.com/@User
[1;92m[[0m[1;77m*[0m[1;92m] Saved: [37;1mUser.txt[0m
That's how all the lines look like in windows powershell. I guess the print is designed for a linux platform? | null | https://github.com/sherlock-project/sherlock/pull/71 | null | {'base_commit': '508eb88724dbe20dedf07dc00527ad4f32c93a77', 'files': [{'path': 'requirements.txt', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [3]}}}, {'path': 'sherlock.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [16]}, "(None, 'main', 247)": {'add': [247]}}}]} | [] | [] | [] | {
"iss_type": "2",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"sherlock.py"
],
"doc": [],
"test": [],
"config": [
"requirements.txt"
],
"asset": []
} | 1 |
sherlock-project | sherlock | e6642737462aaadf2d1f36ea0d417b8cf8a40541 | https://github.com/sherlock-project/sherlock/issues/301 | bug | Error Connecting for most websites (Failed to establish a new connection: [Errno -2] Name or service not known' in urllib3 | I believe this might be a partial duplicate of #230, however, because of lack of information and error output I made my own issue just to be safe.
With the latest python3.7.3 and the correct requirements I constantly run into the issue of websites not connecting.
Specifically this happens after trying to connect to Bandcamp, subsequently, all other urls following bandcamp fail as well.
I printed the `requests.exceptions.ConnectionError` at line 128 as well as the future object class.
For some reason, urllib3 seems to fail finding the name or service (so this might be a DNS issue?).
Further details below (snippet from e.g. YouTube):
```
[-] Error Connecting: YouTube
[-] YouTube: Error!
{'errorType': 'response_url', 'errorUrl': 'https://www.zhihu.com/people/{}', 'rank': 0, 'url': 'https://www.zhihu.com/people/{}', 'urlMain': 'https://www.zhihu.com/', 'username_claimed': 'blue', 'username_unclaimed': 'noonewouldeverusethis7', 'request_future': <Future at 0x7f318072ce80 state=finished raised ConnectionError>}
<Future at 0x7f318072ce80 state=finished raised ConnectionError>
HTTPSConnectionPool(host='www.zhihu.com', port=443): Max retries exceeded with url: /people/berghopper (Caused by NewConnectionError('<urllib3.connection.VerifiedHTTPSConnection object at 0x7f318070c2e8>: Failed to establish a new connection: [Errno -2] Name or service not known'))
<Future at 0x7f318072ce80 state=finished raised ConnectionError>
``` | null | https://github.com/sherlock-project/sherlock/pull/471 | null | {'base_commit': 'e6642737462aaadf2d1f36ea0d417b8cf8a40541', 'files': [{'path': 'README.md', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [72]}}}, {'path': 'data.json', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [50, 315, 714, 722, 1032, 1118, 1302, 1617, 1872, 1880, 1896, 1939, 1971, 2005, 2013], 'mod': [4, 13, 21, 29, 37, 45, 53, 61, 69, 78, 87, 95, 104, 112, 120, 128, 137, 146, 154, 162, 171, 179, 189, 197, 205, 215, 224, 232, 241, 250, 259, 267, 276, 284, 293, 302, 310, 318, 326, 334, 343, 351, 359, 368, 377, 386, 395, 403, 411, 419, 428, 444, 453, 461, 469, 477, 487, 497, 506, 514, 523, 531, 539, 549, 557, 565, 574, 582, 590, 598, 608, 616, 624, 633, 641, 649, 657, 665, 675, 684, 692, 700, 709, 717, 726, 735, 744, 753, 762, 770, 779, 788, 797, 805, 814, 832, 840, 848, 856, 865, 873, 881, 889, 898, 907, 916, 925, 933, 941, 950, 958, 968, 976, 985, 993, 1002, 1010, 1019, 1027, 1035, 1043, 1052, 1061, 1070, 1078, 1097, 1105, 1113, 1121, 1130, 1139, 1148, 1156, 1164, 1172, 1180, 1188, 1196, 1204, 1221, 1229, 1237, 1245, 1254, 1262, 1271, 1280, 1289, 1297, 1323, 1332, 1342, 1351, 1359, 1367, 1384, 1393, 1401, 1410, 1418, 1426, 1434, 1442, 1451, 1459, 1467, 1475, 1484, 1493, 1501, 1510, 1519, 1527, 1535, 1544, 1552, 1561, 1569, 1578, 1587, 1596, 1612, 1621, 1630, 1639, 1657, 1665, 1673, 1682, 1690, 1698, 1706, 1714, 1723, 1732, 1741, 1750, 1758, 1767, 1776, 1784, 1793, 1803, 1811, 1820, 1829, 1838, 1856, 1867, 1875, 1883, 1891, 1900, 1908, 1916, 1925, 1934, 1942, 1950, 1958, 1966, 1975, 1984, 1992, 2000, 2008, 2016, 2025, 2033, 2042, 2051, 2059, 2068, 2076, 2082, 2083, 2084, 2085, 2086, 2087, 2088, 2089, 2090, 2091, 2092, 2093, 2094, 2095, 2096, 2097, 2098, 2099, 2100, 2101, 2102, 2103, 2104, 2105, 2106, 2107, 2108, 2109, 2110, 2111, 2112, 2113, 2114, 2115, 2116, 2117, 2118, 2119, 2120, 2121, 2122, 2123, 2124, 2125, 2126, 2127, 2128, 2129, 2130, 2131, 2132, 2133, 2134, 2135, 2136, 2137, 2138, 2139, 2140, 2141, 2142, 2143, 2144, 2145, 2146, 2147, 2148, 2149, 2150, 2151, 2152, 2153, 2154, 2155, 2156, 2157, 2158, 2159, 2160, 2161, 2162, 2163, 2164, 2165, 2166, 2167, 2168, 2169, 2170, 2171, 2172, 2173, 2174, 2175, 2176, 2177, 2178, 2179, 2180, 2181, 2182, 2183, 2184, 2185, 2186, 2187, 2188, 2189, 2190, 2191, 2192, 2193, 2194, 2195, 2196, 2197, 2198, 2199, 2200, 2201, 2202, 2203, 2204, 2205, 2206, 2207, 2208, 2209, 2210, 2211, 2212, 2213, 2216, 2222, 2224, 2225, 2226, 2227, 2228, 2229, 2230, 2231, 2232, 2233, 2234, 2235, 2236, 2242, 2248, 2249, 2250, 2251, 2252, 2253, 2254, 2255, 2256, 2258, 2259, 2260, 2264, 2266, 2267, 2268, 2272, 2273, 2274, 2275, 2276, 2277, 2278, 2281, 2283, 2284, 2285, 2287]}}}, {'path': 'sherlock.py', 'status': 'modified', 'Loc': {"(None, 'sherlock', 151)": {'add': [189, 191], 'mod': [183, 184, 187, 188, 193, 194, 195]}, '(None, None, None)': {'mod': [29]}}}, {'path': 'sites.md', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [1, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 248]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "3",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"sherlock.py",
"data.json"
],
"doc": [
"README.md",
"sites.md"
],
"test": [],
"config": [],
"asset": []
} | 1 |
keras-team | keras | 96130040540e1405ffe746ddf2b2cceb9b8b8f65 | https://github.com/keras-team/keras/issues/16453 | type:docs | The vocabulary_size method of preprocessing layers does not work in graph mode | **System information**.
- Have I written custom code (as opposed to using a stock example script provided in Keras): **yes**
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): **Colab**
- TensorFlow installed from (source or binary): **binary**
- TensorFlow version (use command below): **TF 2.8, nightly 2.10.0-dev20220422**
**Describe the problem**.
Using the `vocabulary_size()` method of preprocessing layers (like `tf.keras.layers.StringLookup`) fails in graph mode, because the implementation looks like this: https://github.com/keras-team/keras/blob/aea9728313bcaa8262774699c21976288171b209/keras/layers/preprocessing/index_lookup.py#L346-L352
**Describe the current behavior**.
Calling `vocabulary_size()` in graph mode fails.
**Describe the expected behavior**.
Calling `vocabulary_size()` in graph mode succeeds and returns a `tf.Tensor` with the size.
**Standalone code to reproduce the issue**.
Colab notebook showing the issue is at https://colab.research.google.com/drive/1Mq9G8eUvNLw6ykk4ARKf6jurLREAvRhu?usp=sharing both for TF 2.8 and TF nightly.
**Source code / logs**.
An obvious fix is to avoid using the `numpy()` method when in graph mode, i.e., instead of
```python
return int(self.lookup_table.size().numpy()) + self._token_start_index()
```
use just
```python
return self.lookup_table.size() + self._token_start_index()
```
Note that the above notebook also shows that this implementation works in the graph mode. | null | https://github.com/keras-team/keras/pull/16460 | null | {'base_commit': '96130040540e1405ffe746ddf2b2cceb9b8b8f65', 'files': [{'path': 'keras/layers/preprocessing/index_lookup.py', 'status': 'modified', 'Loc': {"('IndexLookup', 'vocabulary_size', 346)": {'mod': [352]}}}]} | [] | [] | [] | {
"iss_type": "2",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"keras/layers/preprocessing/index_lookup.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | 1 |
keras-team | keras | 5b299743442b64afaeeec01e925ddbeb112aad3c | https://github.com/keras-team/keras/issues/20712 | type:support | Tensorflow BackupAndRestore method does not work | I copied the [example code here](https://keras.io/api/callbacks/backup_and_restore/) and it raises a ValueError with Python 3.11 and Tensorflow 2.17:
```
import keras
import numpy as np
class InterruptingCallback(keras.callbacks.Callback):
def on_epoch_begin(self, epoch, logs=None):
if epoch == 4:
raise RuntimeError('Interrupting!')
callback = keras.callbacks.BackupAndRestore(backup_dir="/tmp/backup")
model = keras.models.Sequential([keras.layers.Dense(10)])
model.compile(keras.optimizers.SGD(), loss='mse')
try:
model.fit(np.arange(100).reshape(5, 20), np.zeros(5), epochs=10,
batch_size=1, callbacks=[callback, InterruptingCallback()],
verbose=0)
except Exception as e:
print(e)
history = model.fit(np.arange(100).reshape(5, 20), np.zeros(5),
epochs=10, batch_size=1, callbacks=[callback],
verbose=0)
len(history.history['loss'])
ValueError: To use the BackupAndRestore method, your model must be built before you call `fit()`. Model is unbuilt. You can build it beforehand by calling it on a batch of data.
```
Doesn't that defeat the purpose of backupAndRestore? | null | https://github.com/keras-team/keras/pull/20714 | null | {'base_commit': '5b299743442b64afaeeec01e925ddbeb112aad3c', 'files': [{'path': 'keras/src/callbacks/backup_and_restore.py', 'status': 'modified', 'Loc': {"('BackupAndRestore', None, 9)": {'add': [39]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"keras/src/callbacks/backup_and_restore.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | 1 |
keras-team | keras | a56b16fffec6e4a431bf14e13e7dabeeb5904cd8 | https://github.com/keras-team/keras/issues/8426 | vgg16 predictions: class probabilities change significantly in 2.0.9 | Please make sure that the boxes below are checked before you submit your issue. If your issue is an implementation question, please ask your question on [StackOverflow](http://stackoverflow.com/questions/tagged/keras) or [join the Keras Slack channel](https://keras-slack-autojoin.herokuapp.com/) and ask there instead of filing a GitHub issue.
Thank you!
- [x] Check that you are up-to-date with the master branch of Keras. You can update with:
pip install git+git://github.com/fchollet/keras.git --upgrade --no-deps
- [x] If running on TensorFlow, check that you are up-to-date with the latest version. The installation instructions can be found [here](https://www.tensorflow.org/get_started/os_setup).
- [x] If running on Theano, check that you are up-to-date with the master branch of Theano. You can update with:
pip install git+git://github.com/Theano/Theano.git --upgrade --no-deps
- [x] Provide a link to a GitHub Gist of a Python script that can reproduce your issue (or just copy the script here if it is short).
I've noticed when updating to Keras 2.0.9 that the class probabilities for predictions on the "creative_commons_elephant.jpg" image used in tutorials change significantly (i.e. African_elephant goes from `0.909` to `0.789`).
Here is my script:
```python
from keras.applications.vgg16 import VGG16
model = VGG16(weights='imagenet')
from keras.preprocessing import image
from keras.applications.vgg16 import preprocess_input, decode_predictions
import numpy as np
img_path = 'creative_commons_elephant.jpg'
img = image.load_img(img_path, target_size=(224, 224))
x = image.img_to_array(img)
x = np.expand_dims(x, axis=0)
x = preprocess_input(x)
preds = model.predict(x)
print('Predicted:', decode_predictions(preds, top=3)[0])
```
Under Keras 2.0.8 this gives:
```
('Predicted:', [(u'n02504458', u'African_elephant', 0.90942073), (u'n01871265', u'tusker', 0.086183183), (u'n02504013', u'Indian_elephant', 0.0043545808)])
```
However under Keras 2.0.9 it gives:
```
('Predicted:', [(u'n02504458', u'African_elephant', 0.78988522), (u'n01871265', u'tusker', 0.1987267), (u'n02504013', u'Indian_elephant', 0.011142471)])
``` | null | https://github.com/keras-team/keras/pull/8435 | null | {'base_commit': 'a56b16fffec6e4a431bf14e13e7dabeeb5904cd8', 'files': [{'path': 'keras/preprocessing/image.py', 'status': 'modified', 'Loc': {"(None, 'load_img', 321)": {'mod': [322, 335]}}}]} | [] | [] | [] | {
"iss_type": "2",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"keras/preprocessing/image.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | 1 | |
keras-team | keras | ffb8b813818221a9b43d51e251d489a40b116607 | https://github.com/keras-team/keras/issues/9434 | type:bug/performance | Error with multiprocessing on Sequence in fit_generator() | I'm trying to use a `Sequence` as the generator for `model.fit_generator()`.
With `use_multiprocessing=False` it works fine, but with `use_multiprocessing=True` an error is produced.
**Minimal working example:**
```python
from keras.utils import Sequence
from keras.models import Sequential
from keras.layers import Dense
from keras.utils import to_categorical
import numpy as np
class DummySequence(Sequence):
def __init__(self, x_set, y_set, batch_size):
self.x, self.y = x_set, y_set
self.batch_size = batch_size
def __len__(self):
return int(np.ceil(len(self.x) / float(self.batch_size)))
def __getitem__(self, idx):
batch_x = self.x[idx * self.batch_size:(idx + 1) * self.batch_size]
batch_y = self.y[idx * self.batch_size:(idx + 1) * self.batch_size]
return np.array(batch_x), np.array(batch_y)
if __name__ == '__main__':
x = np.random.random((100, 3))
y = to_categorical(np.random.random(100) > .5).astype(int)
seq = DummySequence(x, y, 10)
model = Sequential()
model.add(Dense(32, input_dim=3))
model.add(Dense(2, activation='softmax'))
model.compile(optimizer='rmsprop',
loss='categorical_crossentropy',
metrics=['accuracy'])
model.fit_generator(generator=seq, workers=2, use_multiprocessing=True)
```
**Error:**
```
Traceback (most recent call last):
File "C:\Users\elcombato\AppData\Local\Continuum\Anaconda3\envs\ml\lib\multiprocessing\pool.py", line 119, in worker
result = (True, func(*args, **kwds))
File "C:\Users\elcombato\AppData\Local\Continuum\Anaconda3\envs\ml\lib\site-packages\keras\utils\data_utils.py", line 392, in get_index
return _SHARED_SEQUENCES[uid][i]
KeyError: 0
```
**Setup**
Windows 10
Python 3.6.4
Keras 2.1.3
Tensorflow 1.2.1 | null | https://github.com/keras-team/keras/pull/9436 | null | {'base_commit': 'ffb8b813818221a9b43d51e251d489a40b116607', 'files': [{'path': 'keras/utils/data_utils.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [375]}, "('OrderedEnqueuer', 'start', 500)": {'mod': [509, 511]}, "('OrderedEnqueuer', '_run', 526)": {'mod': [534]}}}, {'path': 'tests/keras/utils/data_utils_test.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [9, 11, 26, 221, 240, 251, 295]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"keras/utils/data_utils.py",
"tests/keras/utils/data_utils_test.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | 1 |
keras-team | keras | fed28a7357e13aeb955f891747a1f9b26d5bc581 | https://github.com/keras-team/keras/issues/19520 | type:docs
stat:awaiting response from contributor | kl divergence outputs confusing result when the label contains negative | Here are some APIs implemented for kl divergence:
```
tf.losses.kullback_leibler_divergence
tf.keras.losses.KLDivergence
tf.keras.metrics.kl_divergence
```
However, when the `y_true` or `y_pred` is negative, these APIs output incorrect result which is inconsistent with the documentation (https://www.tensorflow.org/api_docs/python/tf/keras/losses/KLDivergence).
Here is the code to reproduce:
```
import os
os.environ['TF_CPP_MIN_LOG_LEVEL'] = '2'
os.environ['OPENBLAS_NUM_THREADS'] = '1'
os.environ['CUDA_VISIBLE_DEVICES'] = ''
import numpy as np
np.random.seed(35)
import tensorflow as tf
y_true = tf.constant(np.random.randn(1), dtype='float32')
y_pred = tf.constant(np.random.randn(1), dtype='float32')
print(y_true, y_pred)
out1 = tf.losses.kullback_leibler_divergence(y_true,y_pred)
out2 = tf.keras.losses.KLDivergence(
reduction='sum_over_batch_size', name='kl_divergence'
)(y_true, y_pred)
out3 = tf.keras.metrics.kl_divergence(y_true, y_pred)
print(out1, out2, out3) # 0.0, 0.0, 0.0
print(f"Expected result: {y_true*(np.log(y_true/y_pred))}") # Expected result following the equation: [-2.8711002]
```
I notice that the current code will silently clip the negative value close to zero:
https://github.com/keras-team/keras/blob/61bbff593a0914f5a2c426c14caadb7372f56da0/keras/losses/losses.py#L1466-L1467
If negative value is not allowed for these concerned APIs, maybe some descriptions on the documentation or validation check inside the source code can provide more information.
| null | https://github.com/keras-team/keras/pull/19526 | null | {'base_commit': 'fed28a7357e13aeb955f891747a1f9b26d5bc581', 'files': [{'path': 'keras/losses/losses.py', 'status': 'modified', 'Loc': {"(None, 'kl_divergence', 1441)": {'mod': [1464]}}}]} | [] | [] | [] | {
"iss_type": "2",
"iss_reason": "2",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"keras/losses/losses.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | 1 |
keras-team | keras | 4bb90a5bfd762269adeee772e233a733a6d318a9 | https://github.com/keras-team/keras/issues/12013 | To investigate | Nasnet weight errors (no_top version) | I"m running into this error:
ValueError: You are trying to load a weight file containing 532 layers into a model with 526 layers.
My keras version is:
`>>> keras.__version__
'2.2.4'`
`Tensorflow version:
'1.12.0'`
Bascially, when the model tries to load I get this value error
` ValueError: You are trying to load a weight file containing 532 layers into a model with 526 layers. `
I"ve looked at this thread:
https://github.com/keras-team/keras/issues/10109
However, I"m trying to run the `no_top version`, so It shouldn't matter what my input vector is.
base_model(weights='imagenet', include_top=False)
Thank you.
Here is the script Im using:
################IMPORTS########################
#--Keras imports--#
from keras.applications import resnet50, xception,inception_v3,inception_resnet_v2, densenet, nasnet, imagenet_utils
from keras.preprocessing import image
from keras.preprocessing.image import ImageDataGenerator
from keras.layers import Input, Dense
from keras import backend as k
from keras.models import Model, clone_model
from keras.layers import Dense, GlobalAveragePooling2D,Dropout, BatchNormalization
from keras import optimizers
from keras.optimizers import Adam
from keras.callbacks import ModelCheckpoint,ProgbarLogger
from keras.utils import print_summary
from keras import __version__
from keras.datasets import cifar10
#--python imports--#
import os
import numpy as np
import datetime
import h5py
import json
import time
################# -- PARAMETERS -- ##################################
img_width, img_height = 480, 480
(x_train, y_train), _ = cifar10.load_data()
classes = len(y_train[0])
##------initial training parameters -----##
i_epochs = 10
i_batch_size = 20
i_steps_per_epoch = 100
i_optimizer = optimizers.SGD(lr=0.0001, momentum=0.9)
#################### MODELS ######################################
def basemodel():
base_model = nasnet.NASNetLarge(weights='imagenet', include_top=False)
preprocess = nasnet.preprocess_input
return base_model, preprocess
def full_model():
base_model, preprocess = basemodel()
x = base_model.output
x = GlobalAveragePooling2D()(x)
x = Dense(2048, activation='relu')(x)
x = Dropout(.60)(x)
x= Dense(512, activation='relu')(x)
predictions = Dense(classes, activation='softmax')(x)
model = Model(inputs= base_model.input, outputs=predictions)
return model,preprocess
def initial_training_full():
model, preprocess = full_model()
for layer in model.layers[:-5]:
layer.trainable = False
model.compile(optimizer= i_optimizer,
loss='sparse_categorical_crossentropy', metrics = ['accuracy'])
print('Starting model training')
history = model.fit(x_train, y_train,
steps_per_epoch = i_steps_per_epoch,
epochs = i_epochs,
shuffle= True,
verbose = 1)
return history
if __name__ == "__main__":
initial_training_full()
| null | https://github.com/keras-team/keras/pull/62 | null | {'base_commit': '4bb90a5bfd762269adeee772e233a733a6d318a9', 'files': [{'path': 'keras/datasets/data_utils.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [3, 4]}, "(None, 'get_file', 7)": {'mod': [32]}}}, {'path': 'setup.py', 'status': 'modified', 'Loc': {}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"keras/datasets/data_utils.py",
"setup.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | 1 |
keras-team | keras | 3d9428d3445a429b535a247168d93b8a5910219d | https://github.com/keras-team/keras/issues/6298 | Performance Using HDF5 File vs In-Memory Dataset | I am trying out using the HDF5 file as the format of training/testing data, so that I can avoid the problem of big dataset which cannot fit into the memory.
As a test, I used a dataset with about 20 million rows and 10 columns as my input, and my response variable is a real-valued vector. The model is a simple sequential model with 2 hidden layers and 50 neurons per hidden layer, and the training batch size is 10000. I am still using Keras version 1.2.0.
When the dataset is fully loaded into the memory, one epoch of training takes about 30 seconds. However, if I used HDF5 matrix with HDF5Matrix, one epoch of training takes about 360 seconds. Is this huge slow down common due to the I/O bottleneck? Shall I implement some perhaps better data loading process with fit_generator by myself by considering the specifications of my dataset? | null | https://github.com/keras-team/keras/pull/6891 | null | {'base_commit': '3d9428d3445a429b535a247168d93b8a5910219d', 'files': [{'path': 'keras/engine/training.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [12], 'mod': [7, 9, 10]}, "('Model', 'fit_generator', 1715)": {'add': [1732, 1856], 'mod': [1723, 1725, 1734, 1759, 1762, 1807, 1860, 1861, 1869, 1870, 1871, 1872, 1873, 1874, 1875, 1926, 1928]}, "('Model', 'evaluate_generator', 1956)": {'add': [1965, 1994], 'mod': [1957, 1968, 1971, 1998, 1999, 2002, 2003, 2004, 2005, 2006, 2007, 2008]}, "('Model', 'predict_generator', 2056)": {'add': [2091], 'mod': [2057, 2058, 2065, 2068, 2071, 2095, 2096, 2102, 2103, 2104, 2105, 2106, 2107, 2108]}, "('GeneratorEnqueuer', None, 582)": {'mod': [582, 583, 585, 587, 588, 589, 590, 592, 593, 594, 595, 596, 597, 599, 600, 602, 603, 604, 605, 606, 608, 609, 610, 611, 612, 613, 614, 615, 616, 617, 618, 620, 621, 622, 623, 624, 625, 626, 627, 628, 629, 630, 632, 633, 634, 635, 636, 637, 638, 639, 640, 641, 642, 643, 644, 645, 647, 648, 650, 651, 653, 655, 656, 657, 658, 659, 661, 662, 663, 664, 665, 666, 668, 669, 670, 672, 673, 674]}}}, {'path': 'keras/legacy/interfaces.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [604]}}}, {'path': 'keras/models.py', 'status': 'modified', 'Loc': {"('Sequential', 'fit_generator', 1028)": {'mod': [1036, 1038, 1075, 1077, 1121, 1123]}, "('Sequential', 'evaluate_generator', 1127)": {'mod': [1128, 1129, 1140, 1142, 1162, 1164]}, "('Sequential', 'predict_generator', 1167)": {'mod': [1168, 1169, 1179, 1181, 1194, 1196]}}}, {'path': 'keras/utils/__init__.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [10]}}}, {'path': 'keras/utils/data_utils.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [14, 17], 'mod': [5, 6, 8, 10, 12, 13, 16]}, "(None, 'validate_file', 261)": {'add': [284]}}}, {'path': 'tests/keras/engine/test_training.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [3, 10, 14, 299]}, "(None, 'test_model_methods', 16)": {'mod': [83, 207]}}}, {'path': 'tests/keras/legacy/interface_test.py', 'status': 'modified', 'Loc': {"(None, 'test_generator_methods_interface', 781)": {'mod': [810, 811, 812, 813, 814, 815, 818, 821]}}}, {'path': 'tests/keras/test_sequential_model.py', 'status': 'modified', 'Loc': {"(None, 'test_sequential_fit_generator', 69)": {'mod': [101]}, "(None, 'test_sequential', 106)": {'mod': [136, 137]}}}, {'path': 'tests/keras/utils/data_utils_test.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [5, 8, 66], 'mod': [4, 7, 11]}}}, {'path': 'tests/test_multiprocessing.py', 'status': 'modified', 'Loc': {"(None, 'test_multiprocessing_training', 22)": {'mod': [52, 54, 60, 61]}, "(None, 'test_multiprocessing_training_fromfile', 105)": {'mod': [134, 136, 142, 143]}, "(None, 'test_multiprocessing_predicting', 149)": {'mod': [169, 171, 174, 175]}, "(None, 'test_multiprocessing_evaluating', 179)": {'mod': [202, 204, 207, 208]}, "(None, 'test_multiprocessing_fit_error', 212)": {'mod': [229, 232, 235, 238]}, "(None, 'test_multiprocessing_evaluate_error', 243)": {'mod': [258, 261, 264, 267]}, "(None, 'test_multiprocessing_predict_error', 272)": {'mod': [287, 290, 293, 296]}}}]} | [] | [] | [] | {
"iss_type": "3",
"iss_reason": "3",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"tests/keras/utils/data_utils_test.py",
"keras/utils/data_utils.py",
"keras/utils/__init__.py",
"tests/keras/legacy/interface_test.py",
"keras/legacy/interfaces.py",
"keras/engine/training.py",
"keras/models.py"
],
"doc": [],
"test": [
"tests/keras/test_sequential_model.py",
"tests/test_multiprocessing.py",
"tests/keras/engine/test_training.py"
],
"config": [],
"asset": []
} | 1 | |
keras-team | keras | 2bfd1f2c950df5fc3f40b903c1966f1b0a48bee4 | https://github.com/keras-team/keras/issues/10855 | fit_generator crashes though keras.utils.data_utils.Sequence was used | When `model.fit_generator` is used with `workers=0` and subclasses of [`keras.utils.data_utils.Sequence`](https://keras.io/utils/#sequence) for both training and validation data, API of `Sequence` is not recognized inside `evaluate_generator`, it raises:
```python
File ".../keras/legacy/interfaces.py", line 91, in wrapper
return func(*args, **kwargs)
File ".../keras/engine/training.py", line 1415, in fit_generator
initial_epoch=initial_epoch)
File ".../keras/engine/training_generator.py", line 230, in fit_generator
validation_steps,
File ".../keras/legacy/interfaces.py", line 91, in wrapper
return func(*args, **kwargs)
File ".../keras/engine/training.py", line 1469, in evaluate_generator
verbose=verbose)
File ".../keras/engine/training_generator.py", line 298, in evaluate_generator
else:
ValueError: `steps=None` is only valid for a generator based on the `keras.utils.Sequence` class. Please specify `steps` or use the `keras.utils.Sequence` class.
```
Example code:
```python
from keras import Sequential
from keras.layers import Dense
from keras.utils.data_utils import Sequence
import numpy as np
class Dataset(Sequence):
def __getitem__(self, index):
return np.random.uniform(size=(16, 8)), np.random.uniform(size=(16, 1))
def __len__(self):
return 128
model = Sequential([Dense(4, activation='relu', input_shape=(8,)),
Dense(1, activation='sigmoid')])
model.compile(loss='mse', optimizer='adam')
model.fit_generator(generator=Dataset(), validation_data=Dataset(),
workers=0)
```
Issue can be fixed [here](https://github.com/keras-team/keras/blob/7205d903fbc079bb99fbae0e3c02e6d2b4d227f0/keras/engine/training_generator.py#L124) by replacing:
```python
if isinstance(val_data, Sequence):
val_enqueuer_gen = iter(val_data)
```
with
```python
if isinstance(val_data, Sequence):
val_enqueuer_gen = iter(val_data)
validation_steps = len(val_data)
``` | null | https://github.com/keras-team/keras/pull/11285 | null | {'base_commit': '2bfd1f2c950df5fc3f40b903c1966f1b0a48bee4', 'files': [{'path': 'keras/engine/training_generator.py', 'status': 'modified', 'Loc': {"(None, 'fit_generator', 21)": {'mod': [127]}}}, {'path': 'tests/keras/engine/test_training.py', 'status': 'modified', 'Loc': {"(None, 'test_model_methods', 126)": {'add': [470]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "3",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"keras/engine/training_generator.py"
],
"doc": [],
"test": [
"tests/keras/engine/test_training.py"
],
"config": [],
"asset": []
} | 1 | |
keras-team | keras | c0d95fd6c2cd8ffc0738819825c3065e3c89977c | https://github.com/keras-team/keras/issues/3859 | clipnorm doesn't work with Embedding | I'm getting a Traceback every time "clipnorm" is used in NN with Embedding layer.
Here is a simple script where the problem is obvious:
``` python
import numpy as np
from keras.layers import Input, Embedding
from keras.optimizers import Adam
from keras.models import Model
input_layer = Input(shape = (1,) )
embedding = Embedding(input_dim = 1,
output_dim = 1)(input_layer)
model = Model(input = input_layer, output = embedding)
model.compile(optimizer = Adam(clipnorm = 1.0), loss = 'mse')
X = np.array([[1]])
Y = np.array([[[0.5]]])
model.fit(X, Y, nb_epoch = 1)
```
Failure:
``` shell
I tensorflow/core/common_runtime/gpu/gpu_device.cc:867] Creating TensorFlow device (/gpu:0) -> (device: 0, name: GeForce GTX TITAN X, pci bus id: 0000:01:00.0)
I tensorflow/core/common_runtime/gpu/gpu_device.cc:867] Creating TensorFlow device (/gpu:1) -> (device: 1, name: GeForce GTX TITAN X, pci bus id: 0000:02:00.0)
Traceback (most recent call last):
File "./clipnorm-bug.py", line 20, in <module>
model.fit(X, Y, nb_epoch = 1)
File "/usr/local/lib/python3.5/dist-packages/keras/engine/training.py", line 1079, in fit
self._make_train_function()
File "/usr/local/lib/python3.5/dist-packages/keras/engine/training.py", line 696, in _make_train_function
self.total_loss)
File "/usr/local/lib/python3.5/dist-packages/keras/optimizers.py", line 379, in get_updates
grads = self.get_gradients(loss, params)
File "/usr/local/lib/python3.5/dist-packages/keras/optimizers.py", line 71, in get_gradients
grads = [clip_norm(g, self.clipnorm, norm) for g in grads]
File "/usr/local/lib/python3.5/dist-packages/keras/optimizers.py", line 71, in <listcomp>
grads = [clip_norm(g, self.clipnorm, norm) for g in grads]
File "/usr/local/lib/python3.5/dist-packages/keras/optimizers.py", line 9, in clip_norm
g = K.switch(n >= c, g * c / n, g)
TypeError: unsupported operand type(s) for *: 'IndexedSlices' and 'float'
```
Keras version is 1.1.0, TensorFlow is 0.10rc
clipvalue on the other hand works fine.
| null | https://github.com/keras-team/keras/pull/4815 | null | {'base_commit': 'c0d95fd6c2cd8ffc0738819825c3065e3c89977c', 'files': [{'path': '.travis.yml', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [52, 54]}}}, {'path': 'keras/backend/tensorflow_backend.py', 'status': 'modified', 'Loc': {"(None, 'rnn', 1640)": {'add': [1699], 'mod': [1720, 1738, 1745, 1746, 1747]}, "(None, 'batch_dot', 811)": {'mod': [870, 871, 872, 873]}, "(None, '_step', 1798)": {'mod': [1808, 1809]}, "(None, 'elu', 1923)": {'mod': [1934]}, "(None, 'random_binomial', 2402)": {'mod': [2407, 2408, 2409]}}}, {'path': 'keras/optimizers.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [1], 'mod': [4]}}}, {'path': 'tests/keras/test_optimizers.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [76]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"keras/optimizers.py",
"keras/backend/tensorflow_backend.py"
],
"doc": [],
"test": [
"tests/keras/test_optimizers.py"
],
"config": [
".travis.yml"
],
"asset": []
} | 1 | |
keras-team | keras | 80fbbc3a6a2a30f391bad2aa85e7558c50ca0709 | https://github.com/keras-team/keras/issues/4452 | PReLU should be channelwise | ```
from keras.models import Sequential
from keras.layers.advanced_activations import PReLU
from keras.layers import Convolution2D, MaxPooling2D
model = Sequential()
model.add(Convolution2D(32, 5, 5, input_shape=(28,28,1)))
model.add(PReLU())
model.summary()
```
This script produces the PReLU layer with 18432 parameters. If the PReLU was implemented according to the paper (https://arxiv.org/abs/1502.01852), number of parameters would be 32. Shouldn't this not be implemented that way? It adds an unnecessarily large number of parameters to the model. | null | https://github.com/keras-team/keras/pull/4141 | null | {'base_commit': '80fbbc3a6a2a30f391bad2aa85e7558c50ca0709', 'files': [{'path': 'keras/backend/theano_backend.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [726]}}}, {'path': 'keras/layers/advanced_activations.py', 'status': 'modified', 'Loc': {"('PReLU', None, 38)": {'add': [54], 'mod': [59]}, "('PReLU', '__init__', 59)": {'add': [62]}, "('ParametricSoftplus', None, 118)": {'add': [133]}, "('ParametricSoftplus', '__init__', 138)": {'add': [143], 'mod': [139]}, "('SReLU', None, 201)": {'add': [216]}, "('SReLU', '__init__', 221)": {'add': [227], 'mod': [222]}, "('PReLU', 'build', 65)": {'mod': [66]}, "('PReLU', 'call', 74)": {'mod': [76]}, "('ParametricSoftplus', 'build', 146)": {'mod': [147, 148, 150]}, "('ParametricSoftplus', 'call', 158)": {'mod': [159]}, "('SReLU', 'build', 230)": {'mod': [231, 238, 240, 242, 244]}, "('SReLU', 'call', 251)": {'mod': [252, 253, 254, 255]}}}, {'path': 'tests/keras/layers/test_advanced_activations.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [19, 51]}, "(None, 'test_parametric_softplus', 29)": {'mod': [31, 32, 33, 34, 35]}}}]} | [] | [] | [] | {
"iss_type": "2",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"keras/backend/theano_backend.py",
"keras/layers/advanced_activations.py"
],
"doc": [],
"test": [
"tests/keras/layers/test_advanced_activations.py"
],
"config": [],
"asset": []
} | 1 | |
nvbn | thefuck | 070bb2ff28b108e5fc627efa1d95ded00ca067c8 | https://github.com/nvbn/thefuck/issues/537 | Args are lost with git_push | From the README:
> `git_push` – adds `--set-upstream origin $branch` to previous failed `git push`;
This isn't actually the case, `--set-upstream` isn't added to the previous push, the entire command is replaced with just `git push --set-upstream origin $branch` regardless of whether there were any args.
Example:
```
$ git push --quiet
fatal: The current branch new-branch has no upstream branch.
To push the current branch and set the remote as upstream, use
git push --set-upstream origin new-branch
$ fuck
git push --set-upstream origin new-branch [enter/↑/↓/ctrl+c]
```
The `--quiet` arg is lost, ideally `git push --set-upstream origin new-branch --quiet` would be suggested.
| null | https://github.com/nvbn/thefuck/pull/538 | null | {'base_commit': '070bb2ff28b108e5fc627efa1d95ded00ca067c8', 'files': [{'path': 'tests/rules/test_git_push.py', 'status': 'modified', 'Loc': {"(None, 'test_match', 16)": {'add': [16]}, "(None, 'test_get_new_command', 22)": {'add': [24]}}}, {'path': 'thefuck/rules/git_push.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [0]}, "(None, 'get_new_command', 11)": {'mod': [12]}}}]} | [] | [] | [] | {
"iss_type": "2",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"thefuck/rules/git_push.py"
],
"doc": [],
"test": [
"tests/rules/test_git_push.py"
],
"config": [],
"asset": []
} | 1 | |
nvbn | thefuck | 86efc6a252c39526c643ea3335db02c4621798e9 | https://github.com/nvbn/thefuck/issues/718 | next release | Cannot run fuck in zsh without export errors | Installed via pip. Ran eval "$(thefuck --alias)" in both bash and zsh. $ fuck then produces 'No fucks given' in bash while in zsh it complains about invalid arguments.
```
$ eval "$(thefuck --alias)"
$ fuck
fuck:export:3: not valid in this context: -'
fuck:3: not an identifier: [skip
```
Also FWIW:
```
$ zsh --version
zsh 4.3.11 (x86_64-redhat-linux-gnu)
$ cat /etc/redhat-release
Red Hat Enterprise Linux Server release 6.8 (Santiago)
``` | null | https://github.com/nvbn/thefuck/pull/820 | null | {'base_commit': '86efc6a252c39526c643ea3335db02c4621798e9', 'files': [{'path': 'thefuck/shells/zsh.py', 'status': 'modified', 'Loc': {"('Zsh', 'app_alias', 12)": {'mod': [19, 20]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"thefuck/shells/zsh.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | 1 |
nvbn | thefuck | f700b23f5725e14f8ee6ffb0a5c44ab9eaf42b53 | https://github.com/nvbn/thefuck/issues/717 | alias produces invalid code for fish as of 3.24 | As of 3.24 the output from `thefuck --alias` is incompatible with fish
**Output from 3.24**
```
$ thefuck --alias
function fuck () {
TF_PYTHONIOENCODING=$PYTHONIOENCODING;
export TF_ALIAS=fuck;
export TF_SHELL_ALIASES=$(alias);
export TF_HISTORY=$(fc -ln -10);
export PYTHONIOENCODING=utf-8;
TF_CMD=$(
thefuck THEFUCK_ARGUMENT_PLACEHOLDER $@
) && eval $TF_CMD;
unset TF_HISTORY;
export PYTHONIOENCODING=$TF_PYTHONIOENCODING;
history -s $TF_CMD;
}
$ thefuck --version
The Fuck 3.24 using Python 3.5.2
```
**Output from 3.23**
```
$ thefuck --alias
function fuck -d "Correct your previous console command"
set -l fucked_up_command $history[1]
env TF_ALIAS=fuck PYTHONIOENCODING=utf-8 thefuck $fucked_up_command | read -l unfucked_command
if [ "$unfucked_command" != "" ]
eval $unfucked_command
builtin history delete --exact --case-sensitive -- $fucked_up_command
builtin history merge ^ /dev/null
end
end
$ thefuck --version
The Fuck 3.23 using Python 3.5.2
```
**Fish version output**
```
$ fish --version
fish, version 2.6.0
``` | null | https://github.com/nvbn/thefuck/pull/744 | null | {'base_commit': 'f700b23f5725e14f8ee6ffb0a5c44ab9eaf42b53', 'files': [{'path': 'tests/shells/test_bash.py', 'status': 'modified', 'Loc': {"('TestBash', 'test_app_alias_variables_correctly_set', 51)": {'add': [53]}}}, {'path': 'tests/shells/test_fish.py', 'status': 'modified', 'Loc': {"('TestFish', 'test_app_alias', 70)": {'add': [73]}}}, {'path': 'tests/shells/test_tcsh.py', 'status': 'modified', 'Loc': {"('TestTcsh', None, 8)": {'add': [46]}}}, {'path': 'tests/shells/test_zsh.py', 'status': 'modified', 'Loc': {"('TestZsh', 'test_app_alias_variables_correctly_set', 51)": {'add': [53]}}}, {'path': 'thefuck/shells/fish.py', 'status': 'modified', 'Loc': {"('Fish', 'app_alias', 28)": {'mod': [38]}}}, {'path': 'thefuck/shells/tcsh.py', 'status': 'modified', 'Loc': {"('Tcsh', 'app_alias', 9)": {'mod': [10]}}}]} | [] | [] | [] | {
"iss_type": "2",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"thefuck/shells/tcsh.py",
"thefuck/shells/fish.py"
],
"doc": [],
"test": [
"tests/shells/test_zsh.py",
"tests/shells/test_fish.py",
"tests/shells/test_tcsh.py",
"tests/shells/test_bash.py"
],
"config": [],
"asset": []
} | 1 | |
nvbn | thefuck | 18992f246a84331832c399283ab930408db21d86 | https://github.com/nvbn/thefuck/issues/652 | git push to branch not of same name: use git's suggestion | I checked out a branch with a different name than the remote counterpart:
```
$ git checkout -b local-branch myupstream/remote-branch
```
When I attempted to push my changes upstream, I got the following error:
```
$ git push -f
fatal: The upstream branch of your current branch does not match
the name of your current branch. To push to the upstream branch
on the remote, use
git push myupstream HEAD:remote-branch
To push to the branch of the same name on the remote, use
git push myupstream local-branch
```
However, thefuck didn't use this information to provide a suggestion:
```
$ fuck
git push -d [enter/↑/↓/ctrl+c]
``` | null | https://github.com/nvbn/thefuck/pull/703 | null | {'base_commit': '18992f246a84331832c399283ab930408db21d86', 'files': [{'path': 'README.md', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [198]}}}]} | [] | [] | [] | {
"iss_type": "4",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [],
"doc": [
"README.md"
],
"test": [],
"config": [],
"asset": []
} | 1 | |
nvbn | thefuck | f604756cb7639ab3efb1cd50c85e633a18cd9f9d | https://github.com/nvbn/thefuck/issues/486 | windows | werid behavior | python 2.7.10
windows powershell
when I type:
``` bash
> git comit -am 'test'
> fuck
```
it gives me something that is completely weird:
```
C:\Python27\lib\site-packages\win_unicode_console\__init__.py:27: RuntimeWarning: sys.stdin.encoding == 'utf-8', whereas sys.stdout.encoding == None, readline hook consumer may assume they are the same
readline_hook.enable(use_pyreadline=use_pyreadline)
[1mgit commit -am 'test'[0m [[32menter[0m/[34m↑[0m/[34m↓[0m/[31mctrl+c[0m]
```
and then give an exception:
```
Traceback (most recent call last):
File "C:\Python27\lib\runpy.py", line 162, in _run_module_as_main
"__main__", fname, loader, pkg_name)
File "C:\Python27\lib\runpy.py", line 72, in _run_code
exec code in run_globals
File "C:\Python27\Scripts\thefuck.exe\__main__.py", line 9, in <module>
File "C:\Python27\lib\site-packages\thefuck\main.py", line 80, in main
fix_command()
File "C:\Python27\lib\site-packages\thefuck\main.py", line 32, in fix_command
selected_command = select_command(corrected_commands)
File "C:\Python27\lib\site-packages\thefuck\ui.py", line 80, in select_command
for action in read_actions():
File "C:\Python27\lib\site-packages\thefuck\ui.py", line 13, in read_actions
key = get_key()
File "C:\Python27\lib\site-packages\thefuck\system\win32.py", line 25, in get_key
return ch.decode(sys.stdout.encoding)
TypeError: decode() argument 1 must be string, not None
```
I think this is related to UTF-8 encoding in python2...
I cannot add python3 to path, because my vim plugins will require python2 in the path...
Is there any way I can fix this?
| null | https://github.com/nvbn/thefuck/pull/487 | null | {'base_commit': 'f604756cb7639ab3efb1cd50c85e633a18cd9f9d', 'files': [{'path': 'thefuck/system/win32.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [0]}, "(None, 'get_key', 13)": {'mod': [25]}}}]} | [] | [] | [] | {
"iss_type": "3",
"iss_reason": "2",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"thefuck/system/win32.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | 1 |
nvbn | thefuck | 926e9ef963464075184f5c2e04ffdce9cb971997 | https://github.com/nvbn/thefuck/issues/809 | windows | Powershell, The term 'thefuck' is not recognized as the name of a cmdlet | **The output of `thefuck --version` (something like `The Fuck 3.1 using Python 3.5.0`):**
The Fuck 3.26 using Python 3.6.5
**Your shell and its version (`bash`, `zsh`, *Windows PowerShell*, etc.):**
Windows Powershell, 5.1.16299.251
**Your system (Debian 7, ArchLinux, Windows, etc.):**
Windows 10
**How to reproduce the bug:**
- install thefuck
`py -m pip install thefuck`
- put the alias into your powershell profile
```
$env:PYTHONIOENCODING="utf-8"
iex "$(thefuck --alias)"
```
- start a new powershell
Result:
```
thefuck : The term 'thefuck' is not recognized as the name of a cmdlet, function, script file,
or operable program. Check the spelling of the name, or if a path was included, verify that the
path is correct and try again.
```
**The output of The Fuck with `THEFUCK_DEBUG=true` exported (typically execute `export THEFUCK_DEBUG=true` in your shell before The Fuck):**
Hmm I can't manage to output this.
**Anything else you think is relevant:**
I also tried it with the following in my profile:
```
...
iex "$(C:\Users\MyUserName\AppData\Local\Programs\Python\Python36-32\Scripts\thefuck.exe --alias)"
```
Then I get a similar error after executing `fuck`:
```
thefuck : The term 'thefuck' is not recognized as the name of a cmdlet, function, script file,
or operable program. Check the spelling of the name, or if a path was included, verify that the
path is correct and try again.
At line:1 char:141
+ ... g]::IsNullOrWhiteSpace($history)) { $fuck = $(thefuck $args $ ...
+ ~~~~~~~
+ CategoryInfo : ObjectNotFound: (thefuck:String) [], CommandNotFoundException
+ FullyQualifiedErrorId : CommandNotFoundException
```
| null | https://github.com/nvbn/thefuck/pull/844 | null | {'base_commit': '926e9ef963464075184f5c2e04ffdce9cb971997', 'files': [{'path': 'thefuck/shells/__init__.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [19]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "2",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"thefuck/shells/__init__.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | 1 |
nvbn | thefuck | 18992f246a84331832c399283ab930408db21d86 | https://github.com/nvbn/thefuck/issues/658 | hacktoberfest | exception while "run fuck second time for configuring it automatically." | ```
luca@g550jk ~> fuck
Seems like fuck alias isn't configured!
Please put eval (thefuck --alias | tr '
' ';') in your ~/.config/fish/config.fish and apply changes with fish or restart your shell.
Or run fuck second time for configuring it automatically.
More details - https://github.com/nvbn/thefuck#manual-installation
luca@g550jk ~> fuck
Traceback (most recent call last):
File "/usr/bin/fuck", line 11, in <module>
load_entry_point('thefuck==3.18', 'console_scripts', 'fuck')()
File "/usr/lib/python3.6/site-packages/thefuck/not_configured.py", line 80, in main
elif _is_second_run():
File "/usr/lib/python3.6/site-packages/thefuck/not_configured.py", line 40, in _is_second_run
if not tracker_path.exists() or not shell.get_history()[-1] == 'fuck':
IndexError: list index out of range
luca@g550jk ~> thefuck --version
The Fuck 3.18 using Python 3.6.1
luca@g550jk ~> lsb_release -d
Description: Arch Linux
luca@g550jk ~> echo $SHELL
/usr/bin/fish
luca@g550jk ~> fish --version
fish, version 2.5.0
``` | null | https://github.com/nvbn/thefuck/pull/704 | null | {'base_commit': '18992f246a84331832c399283ab930408db21d86', 'files': [{'path': 'thefuck/shells/fish.py', 'status': 'modified', 'Loc': {"('Fish', 'how_to_configure', 72)": {'mod': [74]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"thefuck/shells/fish.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | 1 |
nvbn | thefuck | 328e65179e8a886b44611850b908109868884c14 | https://github.com/nvbn/thefuck/issues/229 | Should suggest 'git tag' when typing 'git tags' instead of 'git stage' | ```
$ git tags
git: 'tags' is not a git command. See 'git --help'.
Did you mean one of these?
stage
tag
$ fuck
git stage
Nothing specified, nothing added.
Maybe you wanted to say 'git add .'?
```
Any way to improve the suggestion in this case?
| null | https://github.com/nvbn/thefuck/pull/285 | null | {'base_commit': '328e65179e8a886b44611850b908109868884c14', 'files': [{'path': 'tests/rules/test_tmux.py', 'status': 'modified', 'Loc': {"(None, 'test_get_new_command', 17)": {'mod': [19]}}}, {'path': 'thefuck/rules/tmux.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [0]}, "(None, 'get_new_command', 10)": {'mod': [11, 14]}}}]} | [] | [] | [] | {
"iss_type": "4",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"thefuck/rules/tmux.py"
],
"doc": [],
"test": [
"tests/rules/test_tmux.py"
],
"config": [],
"asset": []
} | 1 | |
nvbn | thefuck | 470c0ef699f31bba76445f18d7cbb289af388565 | https://github.com/nvbn/thefuck/issues/439 | Windows - ImportError: No module named dbm | I have installed thefuck with pip using Python 2.7.11 in Windows. When running `thefuck` from Powershell, I get the error message `ImportError: No module named dbm`. Any ideas?
| null | https://github.com/nvbn/thefuck/pull/458 | null | {'base_commit': '470c0ef699f31bba76445f18d7cbb289af388565', 'files': [{'path': 'thefuck/utils.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [1, 19, 21, 22]}, "(None, '_cache', 211)": {'mod': [230]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"thefuck/utils.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | 1 | |
nvbn | thefuck | d92765d5df6607cb2f2fb67cee7b63f64ac7aa6b | https://github.com/nvbn/thefuck/issues/301 | bug | UnicodeEncodeError on non-english systems | When a command outputs a sequence with non-ascii characters, no fucks are given with a dirty UnicodeEncodeError followed by a grep --compact not recognized option.
Example :
```
~$ aptget blah
La commande « aptget » est introuvable, vouliez-vous dire :
La commande « apt-get » du paquet « apt » (main)
aptget : commande introuvable
~$ fuck
apt-get blah [enter/ctrl+c]
E: L'opération blah n'est pas valable
~$ fuck
Traceback (most recent call last):
File "/usr/local/lib/python2.7/dist-packages/thefuck-2.1-py2.7.egg/thefuck/main.py", line 100, in get_matched_rule
if rule.match(command, settings):
File "/usr/local/lib/python2.7/dist-packages/thefuck-2.1-py2.7.egg/thefuck/rules/history.py", line 28, in match
_history_of_exists_without_current(command)))
File "/usr/local/lib/python2.7/dist-packages/thefuck-2.1-py2.7.egg/thefuck/utils.py", line 109, in wrapper
memo[key] = fn(*args, **kwargs)
File "/usr/local/lib/python2.7/dist-packages/thefuck-2.1-py2.7.egg/thefuck/rules/history.py", line 19, in _history_of_exists_without_current
history = get_history()
File "/usr/local/lib/python2.7/dist-packages/thefuck-2.1-py2.7.egg/thefuck/utils.py", line 109, in wrapper
memo[key] = fn(*args, **kwargs)
File "/usr/local/lib/python2.7/dist-packages/thefuck-2.1-py2.7.egg/thefuck/shells.py", line 267, in get_history
return list(_get_shell().get_history())
File "/usr/local/lib/python2.7/dist-packages/thefuck-2.1-py2.7.egg/thefuck/shells.py", line 67, in get_history
prepared = self._script_from_history(line)\
File "/usr/local/lib/python2.7/dist-packages/thefuck-2.1-py2.7.egg/thefuck/shells.py", line 103, in _script_from_history
print(line)
UnicodeEncodeError: 'ascii' codec can't encode character u'\xe9' in position 4: ordinal not in range(128)
----------------------------
No fuck given
grep : option non reconnue « --compact »
Utilisation : grep [OPTION]... MOTIF [FICHIER]...
Exécutez « grep --help » pour obtenir des renseignements complémentaires.
sync: ignorer tous les arguments
```
| null | https://github.com/nvbn/thefuck/pull/458 | null | {'base_commit': '470c0ef699f31bba76445f18d7cbb289af388565', 'files': [{'path': 'thefuck/utils.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [1, 19, 21, 22]}, "(None, '_cache', 211)": {'mod': [230]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"thefuck/utils.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | 1 |
home-assistant | core | 7d71a2c979bc9ae300e177cbedb059a6133174b1 | https://github.com/home-assistant/core/issues/34897 | integration: homekit | homekit integration: work but error in log | <!-- READ THIS FIRST:
- If you need additional help with this template, please refer to https://www.home-assistant.io/help/reporting_issues/
- Make sure you are running the latest version of Home Assistant before reporting an issue: https://github.com/home-assistant/core/releases
- Do not report issues for integrations if you are using custom components or integrations.
- Provide as many details as possible. Paste logs, configuration samples and code into the backticks.
DO NOT DELETE ANY TEXT from this template! Otherwise, your issue may be closed without comment.
-->
## The problem
<!--
Describe the issue you are experiencing here to communicate to the
maintainers. Tell us what you were trying to do and what happened.
-->No problem for functionality but a lot of same error in log
## Environment
<!--
Provide details about the versions you are using, which helps us to reproduce
and find the issue quicker. Version information is found in the
Home Assistant frontend: Developer tools -> Info.
-->0.109
- Home Assistant Core release with the issue: 0.109
- Last working Home Assistant Core release (if known): 0.108.9
- Operating environment (Home Assistant/Supervised/Docker/venv): Home Assistant in venv
- Integration causing this issue: homekit
- Link to integration documentation on our website: integration:homekit
## Problem-relevant `configuration.yaml`
<!--
An example configuration that caused the problem for you. Fill this out even
if it seems unimportant to you. Please be sure to remove personal information
like passwords, private URLs and other credentials.
-->
```
homekit:
filter:
exclude_domains:
- automation
- device_tracker
- person
- script
- group
exclude_entities:
- climate.netatmo_ingresso
- climate.netatmo_cucina
- binary_sensor.telecomando_a
- binary_sensor.telecomando_b
- binary_sensor.telecomando_c
- binary_sensor.telecomando_d
- light.led_ingresso2
- light.led_ingresso3
- light.led_ingresso4
- light.led_ingresso5
- switch.ovunque_do_not_disturb_switch
- switch.ovunque_repeat_switch
- switch.altoparlante_do_not_disturb_switch
- switch.davide_s_reverb_ai_do_not_disturb_switch
- switch.echo_studio_repeat_switch
- switch.dappertutto_do_not_disturb_switch
- switch.echo_dot_ingresso_repeat_switch
- switch.davide_s_reverb_ai_shuffle_switch
- switch.dappertutto_repeat_switch
- switch.this_device_do_not_disturb_switch
- switch.altoparlante_repeat_switch
- switch.davide_s_alexa_apps_do_not_disturb_switch
- switch.echo_dot_camera_letto_do_not_disturb_switch
- switch.echo_dot_camera_letto_shuffle_switch
- switch.echo_dot_ingresso_do_not_disturb_switch
- switch.echo_show_shuffle_switch
- switch.davide_s_reverb_ai_repeat_switch
- switch.echo_dot_camera_letto_repeat_switch
- switch.echo_dot_ingresso_shuffle_switch
- switch.altoparlante_shuffle_switch
- switch.dappertutto_shuffle_switch
- switch.ovunque_shuffle_switch
- switch.echo_show_repeat_switch
- switch.echo_studio_shuffle_switch
- switch.echo_show_do_not_disturb_switch
- switch.echo_studio_do_not_disturb_switch
- binary_sensor.updater
entity_config:
media_player.lg_tv:
name: TV
```
## Traceback/Error logs
<!--
If you come across any trace or error logs, please provide them.
-->
```
Apr 29 23:03:37 raspberrypi hass[6324]: 2020-04-29 23:03:37 ERROR (MainThread) [homeassistant.core] Error doing job: Future exception was never retrieved
Apr 29 23:03:37 raspberrypi hass[6324]: Traceback (most recent call last):
Apr 29 23:03:37 raspberrypi hass[6324]: File "/usr/lib/python3.7/concurrent/futures/thread.py", line 57, in run
Apr 29 23:03:37 raspberrypi hass[6324]: result = self.fn(*self.args, **self.kwargs)
Apr 29 23:03:37 raspberrypi hass[6324]: File "/srv/homeassistant/lib/python3.7/site-packages/homeassistant/components/homekit/accessories.py", line 252, in update_battery
Apr 29 23:03:37 raspberrypi hass[6324]: if self._char_battery.value != battery_level:
Apr 29 23:03:37 raspberrypi hass[6324]: AttributeError: 'TemperatureSensor' object has no attribute '_char_battery'
```
## Additional information
| null | https://github.com/home-assistant/core/pull/34906 | null | {'base_commit': '7d71a2c979bc9ae300e177cbedb059a6133174b1', 'files': [{'path': 'homeassistant/components/homekit/accessories.py', 'status': 'modified', 'Loc': {"('HomeAccessory', '__init__', 85)": {'add': [101]}, "('HomeAccessory', 'update_battery', 245)": {'add': [249], 'mod': [261]}}}, {'path': 'tests/components/homekit/test_accessories.py', 'status': 'modified', 'Loc': {"(None, 'test_missing_linked_battery_sensor', 344)": {'mod': [366, 367, 368]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"homeassistant/components/homekit/accessories.py"
],
"doc": [],
"test": [
"tests/components/homekit/test_accessories.py"
],
"config": [],
"asset": []
} | 1 |
home-assistant | core | acf41d03db6293b9897b8194f5530793de89952f | https://github.com/home-assistant/core/issues/26146 | integration: mopar | mopar component still does not work | Home Assistant release with the issue:
0.97.2
Last working Home Assistant release (if known):
Operating environment (Hass.io/Docker/Windows/etc.):
arch | x86_64
dev | false
docker | true
hassio | true
os_name | Linux
python_version | 3.7.4
timezone | America/Denver
version | 0.97.2
virtualenv | false
**Description of problem:**
Componets show up but nothing works.
```
'utf-8' codec can't decode byte 0x80 in position 0: invalid start byte
```
**Problem-relevant `configuration.yaml` entries and (fill out even if it seems unimportant):**
```yaml
mopar:
username: !secret mopar_username
password: !secret mopar_password
pin: !secret mopar_pin
```
**Traceback (if applicable):**
```
File "/usr/local/lib/python3.7/site-packages/requests/models.py", line 897, in json
return complexjson.loads(self.text, **kwargs)
File "/usr/local/lib/python3.7/site-packages/simplejson/__init__.py", line 518, in loads
return _default_decoder.decode(s)
File "/usr/local/lib/python3.7/site-packages/simplejson/decoder.py", line 370, in decode
obj, end = self.raw_decode(s)
File "/usr/local/lib/python3.7/site-packages/simplejson/decoder.py", line 400, in raw_decode
return self.scan_once(s, idx=_w(s, idx).end())
simplejson.errors.JSONDecodeError: Expecting value: line 1 column 1 (char 0)
```
**Additional information:** | null | https://github.com/home-assistant/core/pull/33066 | null | {'base_commit': 'acf41d03db6293b9897b8194f5530793de89952f', 'files': [{'path': '.coveragerc', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [435]}}}, {'path': 'homeassistant/components/mopar/__init__.py', 'status': 'removed', 'Loc': {}}, {'path': 'homeassistant/components/mopar/lock.py', 'status': 'removed', 'Loc': {}}, {'path': 'homeassistant/components/mopar/manifest.json', 'status': 'removed', 'Loc': {}}, {'path': 'homeassistant/components/mopar/sensor.py', 'status': 'removed', 'Loc': {}}, {'path': 'homeassistant/components/mopar/services.yaml', 'status': 'removed', 'Loc': {}}, {'path': 'homeassistant/components/mopar/switch.py', 'status': 'removed', 'Loc': {}}, {'path': 'requirements_all.txt', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [885, 886]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"homeassistant/components/mopar/sensor.py",
".coveragerc",
"homeassistant/components/mopar/manifest.json",
"homeassistant/components/mopar/__init__.py",
"homeassistant/components/mopar/lock.py",
"homeassistant/components/mopar/switch.py"
],
"doc": [],
"test": [],
"config": [
"homeassistant/components/mopar/services.yaml",
"requirements_all.txt"
],
"asset": []
} | 1 |
home-assistant | core | f020d65416177e4647a846ec017f441fe08c0696 | https://github.com/home-assistant/core/issues/54258 | integration: fronius
energy | Fronius energy sensors missing device_class and last_reset | ### The problem
The Fronius integration with `sensor_type: inverter` has energy sensors, but these are not marked with any device_class or state_class and there's no last_reset.
eg the sensor `sensor.energy_day_fronius_inverter_1_http_solar` value `12455` and only the following attributes:
```
unit_of_measurement: Wh
friendly_name: Energy day Fronius Inverter 1 http://solar
```
The energy sensor from `sensor_type: power_flow` has the state_class and device_class but no last_reset. eg `sensor.energy_day_fronius_power_flow_0_http_solar`. Also the device_class is power but should be energy.
```
state_class: measurement
unit_of_measurement: Wh
friendly_name: Energy day Fronius Power flow 0 http://solar
device_class: power
```
The total energy from the inverter sensor `sensor.energy_total_fronius_inverter_1_http_solar` is correct though.
### What is version of Home Assistant Core has the issue?
core-2021.8.3
### What was the last working version of Home Assistant Core?
_No response_
### What type of installation are you running?
Home Assistant Container
### Integration causing the issue
fronius
### Link to integration documentation on our website
_No response_
### Example YAML snippet
```yaml
sensor:
- platform: fronius
resource: http://solar
monitored_conditions:
- sensor_type: power_flow
- sensor_type: meter
- sensor_type: inverter
```
```
### Anything in the logs that might be useful for us?
_No response_
### Additional information
_No response_ | null | https://github.com/home-assistant/core/pull/54758 | null | {'base_commit': '013b998974c889c8a80d04637a9ea8e43c7e2fc3', 'files': [{'path': 'homeassistant/components/fronius/manifest.json', 'status': 'modified', 'Loc': {'(None, None, 5)': {'mod': [5]}}}, {'path': 'requirements_all.txt', 'status': 'modified', 'Loc': {'(None, None, 1469)': {'mod': [1469]}}}]} | [] | [] | [] | {
"iss_type": "4",
"iss_reason": "2",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"homeassistant/components/fronius/manifest.json"
],
"doc": [],
"test": [],
"config": [
"requirements_all.txt"
],
"asset": []
} | 1 |
home-assistant | core | 832bc15daacfe1294538bb4a17600e4ada36a47e | https://github.com/home-assistant/core/issues/83076 | integration: twinkly | Bug: Twinkly unavailable | ### The problem
Twinkly device unavailable after update to latest beta.
### What version of Home Assistant Core has the issue?
core-2022.12.0b0
### What was the last working version of Home Assistant Core?
core-2022.11.5
### What type of installation are you running?
Home Assistant Container
### Integration causing the issue
twinkly
### Link to integration documentation on our website
https://www.home-assistant.io/integrations/twinkly/
### Diagnostics information
n/a
### Example YAML snippet
```yaml
n/a
```
### Anything in the logs that might be useful for us?
```txt
2022-12-01 14:11:15.238 DEBUG (MainThread) [ttls.client] Authentication token refreshed
2022-12-01 14:11:15.239 DEBUG (MainThread) [ttls.client] GET endpoint gestalt
2022-12-01 14:11:15.264 DEBUG (MainThread) [ttls.client] GET response 200
2022-12-01 14:11:15.265 DEBUG (MainThread) [homeassistant.components.twinkly.light] Updating 'Twinkly_6A5D49.collective.lan'
2022-12-01 14:11:15.265 DEBUG (MainThread) [ttls.client] Authentication token still valid
2022-12-01 14:11:15.265 DEBUG (MainThread) [ttls.client] GET endpoint led/mode
2022-12-01 14:11:15.290 DEBUG (MainThread) [ttls.client] GET response 200
2022-12-01 14:11:15.290 DEBUG (MainThread) [ttls.client] Authentication token still valid
2022-12-01 14:11:15.290 DEBUG (MainThread) [ttls.client] GET endpoint led/out/brightness
2022-12-01 14:11:15.358 DEBUG (MainThread) [ttls.client] GET response 200
2022-12-01 14:11:15.359 DEBUG (MainThread) [ttls.client] Authentication token still valid
2022-12-01 14:11:15.359 DEBUG (MainThread) [ttls.client] GET endpoint gestalt
2022-12-01 14:11:15.374 DEBUG (MainThread) [ttls.client] GET response 200
2022-12-01 14:11:15.374 DEBUG (MainThread) [ttls.client] Authentication token still valid
2022-12-01 14:11:15.374 DEBUG (MainThread) [ttls.client] GET endpoint movies
```
### Additional information
The device reports `unavailable`.
I am able to control it via the Twinkly app. | null | https://github.com/home-assistant/core/pull/83145 | null | {'base_commit': '832bc15daacfe1294538bb4a17600e4ada36a47e', 'files': [{'path': 'homeassistant/components/twinkly/const.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [11, 29]}}}, {'path': 'homeassistant/components/twinkly/light.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [9, 27, 39]}, "('TwinklyLight', '__init__', 63)": {'add': [98]}, "('TwinklyLight', 'device_info', 126)": {'add': [132]}, "('TwinklyLight', None, 60)": {'add': [167], 'mod': [135, 136, 137, 138]}, "('TwinklyLight', 'async_turn_on', 168)": {'mod': [181, 182, 183, 185, 187, 188, 189, 190, 191, 192, 193, 194, 196, 197, 199, 200, 201, 203, 205, 206, 207, 208, 210]}, "('TwinklyLight', 'async_update', 226)": {'mod': [271, 272]}}}, {'path': 'tests/components/twinkly/__init__.py', 'status': 'modified', 'Loc': {"('ClientMock', '__init__', 19)": {'add': [27]}, "('ClientMock', 'turn_on', 53)": {'add': [57]}, "('ClientMock', 'set_static_colour', 81)": {'add': [83]}, "('ClientMock', 'set_mode', 100)": {'mod': [103, 105]}}}, {'path': 'tests/components/twinkly/test_light.py', 'status': 'modified', 'Loc': {"(None, 'test_turn_on_with_brightness', 68)": {'add': [80, 92], 'mod': [82, 94]}, "(None, 'test_turn_on_with_color_rgbw', 101)": {'add': [109, 114, 122], 'mod': [102, 116]}, "(None, 'test_turn_on_with_color_rgb', 125)": {'add': [133, 138, 146], 'mod': [126, 140]}, "(None, 'test_turn_on_with_effect', 149)": {'add': [163, 171], 'mod': [150, 158, 165]}, '(None, None, None)': {'mod': [6]}, "(None, 'test_turn_on_off', 48)": {'mod': [58, 60]}, "(None, 'test_turn_off', 174)": {'mod': [181, 183]}, "(None, 'test_update_name', 190)": {'mod': [202, 204]}}}]} | [] | [] | [] | {
"iss_type": "2",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"homeassistant/components/twinkly/light.py",
"tests/components/twinkly/__init__.py",
"homeassistant/components/twinkly/const.py"
],
"doc": [],
"test": [
"tests/components/twinkly/test_light.py"
],
"config": [],
"asset": []
} | 1 |
home-assistant | core | 20f94d7ad47082ec2f1c204130abe318c4b014dd | https://github.com/home-assistant/core/issues/54678 | integration: myq | MyQ Login Issue | ### The problem
Myq fails to initialize with "Authentication failed: Form containing fields for email, password and submit not found.Unable to continue login process.". Issue started happening today for me. There are more people reporting it here: https://community.home-assistant.io/t/myq-login-unavailable/330322
### What is version of Home Assistant Core has the issue?
2021.7.4
### What was the last working version of Home Assistant Core?
_No response_
### What type of installation are you running?
Home Assistant Container
### Integration causing the issue
myq
### Link to integration documentation on our website
https://www.home-assistant.io/integrations/myq/
### Example YAML snippet
_No response_
### Anything in the logs that might be useful for us?
```txt
Logger: pymyq.api
Source: /usr/local/lib/python3.9/site-packages/pymyq/api.py:701
First occurred: 9:24:30 PM (13 occurrences)
Last logged: 9:37:04 PM
Authentication failed: Form containing fields for email, password and submit not found.Unable to continue login process.
```
### Additional information
_No response_ | null | https://github.com/home-assistant/core/pull/55099 | null | {'base_commit': '20f94d7ad47082ec2f1c204130abe318c4b014dd', 'files': [{'path': 'homeassistant/components/myq/manifest.json', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [5]}}}, {'path': 'requirements_all.txt', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [1626]}}}, {'path': 'requirements_test_all.txt', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [939]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"homeassistant/components/myq/manifest.json"
],
"doc": [],
"test": [],
"config": [
"requirements_test_all.txt",
"requirements_all.txt"
],
"asset": []
} | 1 |
home-assistant | core | b960ebeb8bbf791a8072e8503b417474fec134ac | https://github.com/home-assistant/core/issues/2290 | Smoke detector are not changing from "Panic" state | Make sure you are running the latest version of Home Assistant before reporting an issue.
You should only file an issue if you found a bug. Feature and enhancement requests should go in [the Feature Requests section](https://community.home-assistant.io/c/feature-requests) of our community forum:
**Home Assistant release (`hass --version`): 0.21.1
**Component/platform: Rfxtrx
**Description of problem: First time I press the test button on my Nexa KD101 smoke detector it appears in HASS UI, but after that the UI is never changed, It keep showing panic and nothing happen in the UI after first test. The log are showing the smoke detector signal, so it is received.


| null | https://github.com/home-assistant/core/pull/2498 | null | {'base_commit': '09a4336bc5d9a2f01f475b786492bb517ca36b5f', 'files': [{'path': 'homeassistant/components/sensor/rfxtrx.py', 'status': 'modified', 'Loc': {"(None, 'setup_platform', 29)": {'mod': [50]}, "('RfxtrxSensor', 'state', 111)": {'mod': [113]}, "('RfxtrxSensor', 'device_state_attributes', 123)": {'mod': [125]}}}, {'path': 'tests/components/sensor/test_rfxtrx.py', 'status': 'modified', 'Loc': {"('TestSensorRfxtrx', 'test_several_sensors', 93)": {'add': [122, 135], 'mod': [115, 116, 117, 118, 119, 120, 125, 126, 127, 128, 129, 130, 137, 138, 139, 140, 141, 142, 143]}, "('TestSensorRfxtrx', 'test_update_of_sensors', 233)": {'add': [262, 275], 'mod': [255, 256, 257, 258, 259, 260, 265, 266, 267, 268, 269, 270, 277, 278, 279, 280, 281, 282, 283]}, "('TestSensorRfxtrx', 'test_old_config_sensor', 33)": {'mod': [47, 48, 49, 50, 51, 52]}, "('TestSensorRfxtrx', 'test_one_sensor', 54)": {'mod': [67, 68, 69, 70, 71, 72]}, "('TestSensorRfxtrx', 'test_one_sensor_no_datatype', 74)": {'mod': [86, 87, 88, 89, 90, 91]}}}]} | [] | [] | [] | {
"iss_type": "2",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"homeassistant/components/sensor/rfxtrx.py"
],
"doc": [],
"test": [
"tests/components/sensor/test_rfxtrx.py"
],
"config": [],
"asset": []
} | 1 | |
home-assistant | core | d9e3c02df3a2690e74d1b606e8db0a4dd686e872 | https://github.com/home-assistant/core/issues/79275 | integration: unifi | Ubiquiti POE toggle (off) : aiounifi.models.event:Unsupported event key EVT_SW_PoeDisconnect | ### The problem
as title, toggling a POE off throws the Warning
they do toggle in the end, although it can take up to 45 secs or so.
which is the same behavior we see on the camera's issue, extremely slow response times on the toggle
### What version of Home Assistant Core has the issue?
2022.10.0b0
### What was the last working version of Home Assistant Core?
_No response_
### What type of installation are you running?
Home Assistant OS
### Integration causing the issue
UniFi Network
### Link to integration documentation on our website
https://www.home-assistant.io/integrations/unifi/
### Diagnostics information
_No response_
### Example YAML snippet
_No response_
### Anything in the logs that might be useful for us?
```txt
2022-09-29 09:31:08.184 WARNING (MainThread) [aiounifi.models.event] Unsupported event key EVT_SW_PoeDisconnect
2022-09-29 09:31:08.190 WARNING (MainThread) [aiounifi.models.event] Unsupported event key EVT_SW_PoeDisconnect
2022-09-29 09:31:08.192 WARNING (MainThread) [aiounifi.models.event] Unsupported event key EVT_SW_PoeDisconnect
2022-09-29 09:31:08.196 WARNING (MainThread) [aiounifi.models.event] Unsupported event key EVT_SW_PoeDisconnect
```
or:
```
Logger: aiounifi.models.event
Source: components/unifi/controller.py:232
First occurred: 09:31:08 (1 occurrences)
Last logged: 09:31:08
Unsupported event key EVT_SW_PoeDisconnect
```
### Additional information
these are the bt proxies I have in the integration:
<img width="811" alt="Schermafbeelding 2022-09-29 om 09 44 02" src="https://user-images.githubusercontent.com/33354141/192971228-1269e885-6e6e-499b-8261-734ce122fcec.png">
have [another issue on POE switching](https://github.com/home-assistant/core/issues/79088), Ubiquiti Cams not responding correctly), not sure if its related
| null | https://github.com/home-assistant/core/pull/79487 | null | {'base_commit': '069818940e5a16c412544e1e14e69b5d0964f157', 'files': [{'path': 'homeassistant/components/unifi/manifest.json', 'status': 'modified', 'Loc': {'(None, None, 6)': {'mod': [6]}}}, {'path': 'requirements_all.txt', 'status': 'modified', 'Loc': {'(None, None, 279)': {'mod': [279]}}}, {'path': 'requirements_test_all.txt', 'status': 'modified', 'Loc': {'(None, None, 254)': {'mod': [254]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"homeassistant/components/unifi/manifest.json"
],
"doc": [],
"test": [],
"config": [
"requirements_test_all.txt",
"requirements_all.txt"
],
"asset": []
} | 1 |
home-assistant | core | b71a0c5d4bb22f1d7ec84e892cff851ad1d6f283 | https://github.com/home-assistant/core/issues/54573 | integration: ambiclimate | Ambi Climate Integration not listing any Entities | ### The problem
The ambiclimate integration is appearing in the Integration Screen but not showing any entities even though 1 is available and was working correctly in the previous version
### What is version of Home Assistant Core has the issue?
Home Assistant 2021.8.6
### What was the last working version of Home Assistant Core?
Home Assistant 2021.8.1
### What type of installation are you running?
Home Assistant Core
### Integration causing the issue
Ambiclimate
### Link to integration documentation on our website
https://www.home-assistant.io/integrations/ambiclimate/
### Example YAML snippet
_No response_
### Anything in the logs that might be useful for us?
```txt
Logger: homeassistant.components.climate
Source: components/ambiclimate/climate.py:157
Integration: Climate (documentation, issues)
First occurred: 2:32:41 PM (1 occurrences)
Last logged: 2:32:41 PM
Error while setting up ambiclimate platform for climate
Traceback (most recent call last):
File "/usr/src/homeassistant/homeassistant/helpers/entity_platform.py", line 249, in _async_setup_platform
await asyncio.shield(task)
File "/usr/src/homeassistant/homeassistant/components/ambiclimate/climate.py", line 94, in async_setup_entry
devs.append(AmbiclimateEntity(heater, store))
File "/usr/src/homeassistant/homeassistant/components/ambiclimate/climate.py", line 157, in __init__
self._attr_min_temp = heater.get_min_temp()
File "/usr/local/lib/python3.9/site-packages/ambiclimate/__init__.py", line 251, in get_min_temp
data = self.ir_features['data'][self.ac_data[0].get('mode').lower()]['temperature']['value']
TypeError: 'NoneType' object is not subscriptable
```
### Additional information
_No response_ | null | https://github.com/home-assistant/core/pull/54579 | null | {'base_commit': 'b71a0c5d4bb22f1d7ec84e892cff851ad1d6f283', 'files': [{'path': 'homeassistant/components/ambiclimate/climate.py', 'status': 'modified', 'Loc': {"('AmbiclimateEntity', 'async_update', 175)": {'add': [186]}, "('AmbiclimateEntity', '__init__', 146)": {'mod': [157, 158]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"homeassistant/components/ambiclimate/climate.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | 1 |
home-assistant | core | f18c6ae72ccf9075c3f5442b1aff2381379ca9d6 | https://github.com/home-assistant/core/issues/39472 | integration: arlo | arlo integration is broken | <!-- READ THIS FIRST:
- If you need additional help with this template, please refer to https://www.home-assistant.io/help/reporting_issues/
- Make sure you are running the latest version of Home Assistant before reporting an issue: https://github.com/home-assistant/core/releases
- Do not report issues for integrations if you are using custom components or integrations.
- Provide as many details as possible. Paste logs, configuration samples and code into the backticks.
DO NOT DELETE ANY TEXT from this template! Otherwise, your issue may be closed without comment.
-->
## The problem
<!--
Describe the issue you are experiencing here to communicate to the
maintainers. Tell us what you were trying to do and what happened.
-->
`arlo` integration fails to initialize with message
```txt
ERROR (MainThread) [homeassistant.setup] Setup failed for arlo: Integration failed to initialize.
```
## Environment
<!--
Provide details about the versions you are using, which helps us to reproduce
and find the issue quicker. Version information is found in the
Home Assistant frontend: Configuration -> Info.
-->
- Home Assistant Core release with the issue: 0.114.1
- Last working Home Assistant Core release (if known): N/A
- Operating environment (OS/Container/Supervised/Core): Container
- Integration causing this issue: arlo
- Link to integration documentation on our website: https://www.home-assistant.io/integrations/arlo/
## Problem-relevant `configuration.yaml`
<!--
An example configuration that caused the problem for you. Fill this out even
if it seems unimportant to you. Please be sure to remove personal information
like passwords, private URLs and other credentials.
-->
```yaml
arlo: !include arlo.yaml
```
`arlo.yaml`
```yaml
username: !secret arlo_username
password: !secret arlo_password
```
`secrets.yaml`
```yaml
arlo_username: REDACTED
arlo_password: REDACTED
```
## Traceback/Error logs
<!--
If you come across any trace or error logs, please provide them.
-->
```txt
> grep arlo home-assistant.log
2020-08-31 18:29:28 INFO (MainThread) [homeassistant.bootstrap] Setting up stage 2: {'noonlight', 'met', 'sun', 'mobile_app', 'input_datetime', 'hue', 'ffmpeg', 'default_config', 'map', 'input_text', 'alarm_control_panel', 'notify', 'input_number', 'arlo', 'scene', 'system_health', 'input_boolean', 'zeroconf', 'weather', 'light', 'automation', 'zha', 'updater', 'tts', 'ssdp', 'input_select', 'history', 'zone', 'logbook', 'zwave', 'group', 'script'}
2020-08-31 18:29:29 DEBUG (MainThread) [homeassistant.setup] Dependency arlo will wait for ['ffmpeg']
2020-08-31 18:29:34 INFO (MainThread) [homeassistant.setup] Setting up arlo
2020-08-31 18:29:34 DEBUG (SyncWorker_6) [pyarlo] Creating Arlo session
2020-08-31 18:29:34 DEBUG (SyncWorker_6) [pyarlo] Params: {'email': 'REDACTED', 'password': 'REDACTED'}
2020-08-31 18:29:34 DEBUG (SyncWorker_6) [pyarlo] Headers: {'Content-Type': 'application/json', 'Authorization': None}
2020-08-31 18:29:34 DEBUG (SyncWorker_6) [pyarlo] Querying https://arlo.netgear.com/hmsweb/login/v2 on attempt: 0/3
2020-08-31 18:29:34 DEBUG (SyncWorker_6) [pyarlo] Params: {'email': 'REDACTED', 'password': 'REDACTED'}
2020-08-31 18:29:34 DEBUG (SyncWorker_6) [pyarlo] Headers: {'Content-Type': 'application/json', 'Authorization': None}
2020-08-31 18:29:34 DEBUG (SyncWorker_6) [pyarlo] Querying https://arlo.netgear.com/hmsweb/login/v2 on attempt: 1/3
2020-08-31 18:29:34 DEBUG (SyncWorker_6) [pyarlo] Params: {'email': 'REDACTED', 'password': 'REDACTED'}
2020-08-31 18:29:34 DEBUG (SyncWorker_6) [pyarlo] Headers: {'Content-Type': 'application/json', 'Authorization': None}
2020-08-31 18:29:34 DEBUG (SyncWorker_6) [pyarlo] Querying https://arlo.netgear.com/hmsweb/login/v2 on attempt: 2/3
2020-08-31 18:29:35 DEBUG (SyncWorker_6) [pyarlo] Params: {'email': 'REDACTED', 'password': 'REDACTED'}
2020-08-31 18:29:35 DEBUG (SyncWorker_6) [pyarlo] Headers: {'Content-Type': 'application/json', 'Authorization': None}
2020-08-31 18:29:35 DEBUG (SyncWorker_6) [pyarlo] Querying https://arlo.netgear.com/hmsweb/login/v2 on attempt: 3/3
2020-08-31 18:29:35 INFO (MainThread) [homeassistant.setup] Setup of domain arlo took 0.6 seconds
2020-08-31 18:29:35 ERROR (MainThread) [homeassistant.setup] Setup failed for arlo: Integration failed to initialize.
```
## Additional information
I have tested the exact same configuration using https://github.com/twrecked/hass-aarlo instead (by renaming config references from `arlo` to `aarlo`) and the integration initializes without issue. My Arlo account does not have 2FA enabled. | null | https://github.com/home-assistant/core/pull/44034 | null | {'base_commit': 'f18c6ae72ccf9075c3f5442b1aff2381379ca9d6', 'files': [{'path': 'homeassistant/components/arlo/manifest.json', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [5]}}}, {'path': 'requirements_all.txt', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [1274]}}}, {'path': 'requirements_test_all.txt', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [643]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"homeassistant/components/arlo/manifest.json"
],
"doc": [],
"test": [],
"config": [
"requirements_test_all.txt",
"requirements_all.txt"
],
"asset": []
} | 1 |
home-assistant | core | 91962e2681dde1b23612df06633b16aa0867c950 | https://github.com/home-assistant/core/issues/14989 | Mysensor HVAC frontend shows multiple values | I think since 0.6x mysensors hvac node shows several temp settings

| null | https://github.com/home-assistant/core/pull/15110 | null | {'base_commit': '91962e2681dde1b23612df06633b16aa0867c950', 'files': [{'path': 'homeassistant/components/climate/mysensors.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [29, 30, 31]}, "('MySensorsHVAC', 'supported_features', 46)": {'mod': [48]}, "('MySensorsHVAC', 'operation_list', 104)": {'mod': [106]}, "('MySensorsHVAC', 'fan_list', 114)": {'mod': [116]}}}]} | [] | [] | [] | {
"iss_type": "2",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"homeassistant/components/climate/mysensors.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | 1 | |
home-assistant | core | 62054b84336e4d4a8a6a1ea649724e9eb86876b4 | https://github.com/home-assistant/core/issues/35779 | integration: denon | Error while setting up denon platform for media_player with Marantz | <!-- READ THIS FIRST:
- If you need additional help with this template, please refer to https://www.home-assistant.io/help/reporting_issues/
- Make sure you are running the latest version of Home Assistant before reporting an issue: https://github.com/home-assistant/core/releases
- Do not report issues for integrations if you are using custom components or integrations.
- Provide as many details as possible. Paste logs, configuration samples and code into the backticks.
DO NOT DELETE ANY TEXT from this template! Otherwise, your issue may be closed without comment.
-->
## The problem
I've used the denon integration (not DenonAVR) to connect to my Marantz NR1609. It worked for a while, but doesn't now
## Environment
- Home Assistant Core release with the issue: 0.109.6
- Last working Home Assistant Core release (if known): 0.99 worked. Don't know when it was switched over.
- Operating environment (Home Assistant/Supervised/Docker/venv): Supervised
- Integration causing this issue: Denon
- Link to integration documentation on our website: https://www.home-assistant.io/integrations/denon/
## Problem-relevant `configuration.yaml`
Didn't change this configuration recently, since it only has one variable (and a name)
```
media_player:
...
- platform: denon
host: 192.168.xx.xx
name: Marantz
```
## Traceback/Error logs
```
Logger: homeassistant.components.media_player
Source: components/denon/media_player.py:117
Traceback (most recent call last):
File "/usr/src/homeassistant/homeassistant/helpers/entity_platform.py", line 178, in _async_setup_platform
await asyncio.wait_for(asyncio.shield(task), SLOW_SETUP_MAX_WAIT)
File "/usr/local/lib/python3.7/asyncio/tasks.py", line 442, in wait_for
return fut.result()
File "/usr/local/lib/python3.7/concurrent/futures/thread.py", line 57, in run
result = self.fn(*self.args, **self.kwargs)
File "/usr/src/homeassistant/homeassistant/components/denon/media_player.py", line 85, in setup_platform
if denon.update():
File "/usr/src/homeassistant/homeassistant/components/denon/media_player.py", line 162, in update
self._setup_sources(telnet)
File "/usr/src/homeassistant/homeassistant/components/denon/media_player.py", line 117, in _setup_sources
source, configured_name = line[len("SSFUN") :].split(" ", 1)
ValueError: not enough values to unpack (expected 2, got 1)
```
## Additional information
| null | https://github.com/home-assistant/core/pull/40514 | null | {'base_commit': '62054b84336e4d4a8a6a1ea649724e9eb86876b4', 'files': [{'path': 'homeassistant/components/denon/media_player.py', 'status': 'modified', 'Loc': {"('DenonDevice', '_setup_sources', 108)": {'mod': [114, 117]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"homeassistant/components/denon/media_player.py"
],
"doc": [],
"test": [],
"config": [],
"asset": []
} | 1 |
home-assistant | core | e11ead410bbd9179ded05ba5e07b6de9919ec0ea | https://github.com/home-assistant/core/issues/128046 | integration: gree | Gree climate does not update tha data (status) | ### The problem
The HA does not update the climate status (switch) if the change is made with a remote control. It does not follow the climate continuously.
### What version of Home Assistant Core has the issue?
2024.10.1
### What was the last working version of Home Assistant Core?
2024.09.
### What type of installation are you running?
Home Assistant OS
### Integration causing the issue
Gree
### Link to integration documentation on our website
https://www.home-assistant.io/integrations/gree
### Diagnostics information
_No response_
### Example YAML snippet
_No response_
### Anything in the logs that might be useful for us?
_No response_
### Additional information
_No response_ | null | https://github.com/home-assistant/core/pull/139469 | null | {'base_commit': 'e11ead410bbd9179ded05ba5e07b6de9919ec0ea', 'files': [{'path': 'homeassistant/components/gree/const.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [22]}}}, {'path': 'homeassistant/components/gree/coordinator.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [4, 26]}, "('DeviceDataUpdateCoordinator', '_async_update_data', 65)": {'add': [102], 'mod': [91, 99, 108]}, "('DeviceDataUpdateCoordinator', '__init__', 38)": {'mod': [51]}}}, {'path': 'tests/components/gree/test_climate.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [54]}, "(None, 'run_update', 348)": {'mod': [349]}}}]} | [] | [] | [] | {
"iss_type": "2",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"homeassistant/components/gree/coordinator.py",
"homeassistant/components/gree/const.py"
],
"doc": [],
"test": [
"tests/components/gree/test_climate.py"
],
"config": [],
"asset": []
} | 1 |
home-assistant | core | 80af2f4279828419a9112a5478029f2c506896e7 | https://github.com/home-assistant/core/issues/53678 | integration: sonos | Sonos Alarm switch changes immediatly after click | ### The problem
I have sonos alarm switches. When I click on it, the status changes, but immediately back again.
After a while (approx. 60s) the status changes to the correct state.
In the sonos app, the alarm is switched correctly with the first click.
See the GIF below. The behavior when switching on and then switching off.

### What is version of Home Assistant Core has the issue?
core-2021.7.4
### What was the last working version of Home Assistant Core?
_No response_
### What type of installation are you running?
Home Assistant Core
### Integration causing the issue
Sonos
### Link to integration documentation on our website
https://github.com/home-assistant/core/tree/dev/homeassistant/components/sonos
### Example YAML snippet
_No response_
### Anything in the logs that might be useful for us?
_No response_
### Additional information
_No response_ | null | https://github.com/home-assistant/core/pull/55529 | null | {'base_commit': '80af2f4279828419a9112a5478029f2c506896e7', 'files': [{'path': 'homeassistant/components/sonos/alarms.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [11], 'mod': [9]}, "('SonosAlarms', 'async_update_entities', 37)": {'add': [38], 'mod': [40, 42, 45, 46, 47, 48, 49, 50, 51, 53, 54, 56, 57, 58, 59, 61, 62, 63, 64]}, "('SonosAlarms', '__init__', 23)": {'mod': [26]}, "('SonosAlarms', '__iter__', 28)": {'mod': [30, 31]}, "('SonosAlarms', 'get', 33)": {'mod': [35]}, "('SonosAlarms', None, 20)": {'mod': [37]}, "('SonosAlarms', 'update_cache', 53)": {'mod': [66, 67, 68, 70]}}}, {'path': 'homeassistant/components/sonos/favorites.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [5, 11]}, "('SonosFavorites', '__init__', 23)": {'add': [26]}, "('SonosFavorites', 'update_cache', 46)": {'add': [50, 63, 64], 'mod': [49, 61]}, "('SonosFavorites', None, 20)": {'mod': [33, 46, 47]}, "('SonosFavorites', 'async_update_entities', 33)": {'mod': [35, 36, 37, 38, 39, 44]}}}, {'path': 'homeassistant/components/sonos/household_coordinator.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [9], 'mod': [4, 7, 11]}, "('SonosHouseholdCoordinator', '__init__', 22)": {'add': [27], 'mod': [26]}, "('SonosHouseholdCoordinator', None, 19)": {'add': [46], 'mod': [34, 35, 37, 38, 60, 61, 62, 63, 64, 65, 66, 68, 72, 73]}, "('SonosHouseholdCoordinator', 'setup', 29)": {'mod': [32]}, "('SonosHouseholdCoordinator', '_async_poll', 47)": {'mod': [52, 53, 55]}}}, {'path': 'homeassistant/components/sonos/manifest.json', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [6]}}}, {'path': 'homeassistant/components/sonos/speaker.py', 'status': 'modified', 'Loc': {"('SonosSpeaker', 'async_dispatch_favorites', 449)": {'mod': [453]}}}, {'path': 'homeassistant/components/sonos/switch.py', 'status': 'modified', 'Loc': {"(None, '_async_create_entity', 38)": {'add': [39, 40, 41]}}}, {'path': 'requirements_all.txt', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [2175]}}}, {'path': 'requirements_test_all.txt', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [1215]}}}, {'path': 'tests/components/sonos/conftest.py', 'status': 'modified', 'Loc': {"('SonosMockEvent', 'increment_variable', 34)": {'add': [41]}, "(None, 'alarm_clock_fixture', 111)": {'add': [115], 'mod': [121]}, "(None, 'alarm_clock_fixture_extended', 127)": {'add': [131], 'mod': [141]}, "(None, 'music_library_fixture', 103)": {'mod': [105, 106]}}}, {'path': 'tests/components/sonos/test_switch.py', 'status': 'modified', 'Loc': {"(None, 'test_alarm_create_delete', 53)": {'add': [71], 'mod': [78]}}}]} | [] | [] | [] | {
"iss_type": "2",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"homeassistant/components/sonos/alarms.py",
"homeassistant/components/sonos/household_coordinator.py",
"homeassistant/components/sonos/manifest.json",
"tests/components/sonos/conftest.py",
"homeassistant/components/sonos/switch.py",
"homeassistant/components/sonos/speaker.py",
"homeassistant/components/sonos/favorites.py"
],
"doc": [],
"test": [
"tests/components/sonos/test_switch.py"
],
"config": [
"requirements_test_all.txt",
"requirements_all.txt"
],
"asset": []
} | 1 |
home-assistant | core | f29e0bf53eec22dca65e3a62336cf84797e36bf2 | https://github.com/home-assistant/core/issues/7306 | Problems sending push message with HTML5 notify | **Home Assistant release (`hass --version`):**
0.43.0
**Python release (`python3 --version`):**
Python 3.4.2
**Component/platform:**
HTML5 notify on Raspberry PI 3
**Description of problem:**
I've followed the installation instructions here: https://home-assistant.io/components/notify.html5/
When I try to send a message, manually triggering it through the services pane, I get the error below (see traceback).
At the bottom of the trace it says:
`AttributeError: /usr/lib/arm-linux-gnueabihf/libcrypto.so.1.0.0: undefined symbol: EVP_CIPHER_CTX_reset`
**Expected:**
**Problem-relevant `configuration.yaml` entries and steps to reproduce:**
I followed the installation instructions and installed `libffi-dev`, `libpython-dev`, and `libssl-dev` before installing `pywebpush`. I've tried `pywebpush` 0.5, 0.6, 0.6.1 and 0.8. All resulted in the same error, which leads me to believe the error is elsewhere.
I'm using a SSL certificate from let's encrypt and duckdns.org.
```yaml
notify:
- name: html5
platform: html5
gcm_api_key: !secret gcm_api_key
gcm_sender_id: !secret gcm_sender_id
```
1. Go to services pane
2. Use the `notify.html5push` service with data `{"title": "Title", "message": "Message"}`
3. Call service, and the error pops up in the HASS logs.
**Traceback (if applicable):**
```bash
Apr 25 20:13:36 raspberrypi hass[2828]: INFO:homeassistant.core:Bus:Handling <Event call_service[L]: service_call_id=1978163280-1, service_data=title=hei, message=hei, service=html5push, domain=notify>
Apr 25 20:13:37 raspberrypi hass[2828]: ERROR:homeassistant.core:Error doing job: Task exception was never retrieved
Apr 25 20:13:37 raspberrypi hass[2828]: Traceback (most recent call last):
Apr 25 20:13:37 raspberrypi hass[2828]: File "/usr/lib/python3.4/asyncio/tasks.py", line 233, in _step
Apr 25 20:13:37 raspberrypi hass[2828]: result = coro.throw(exc)
Apr 25 20:13:37 raspberrypi hass[2828]: File "/srv/homeassistant/lib/python3.4/site-packages/homeassistant/core.py", line 1015, in _event_to_service_call
Apr 25 20:13:37 raspberrypi hass[2828]: yield from service_handler.func(service_call)
Apr 25 20:13:37 raspberrypi hass[2828]: File "/srv/homeassistant/lib/python3.4/site-packages/homeassistant/components/notify/__init__.py", line 136, in async_notify_message
Apr 25 20:13:37 raspberrypi hass[2828]: yield from notify_service.async_send_message(**kwargs)
Apr 25 20:13:37 raspberrypi hass[2828]: File "/usr/lib/python3.4/asyncio/futures.py", line 388, in __iter__
Apr 25 20:13:37 raspberrypi hass[2828]: yield self # This tells Task to wait for completion.
Apr 25 20:13:37 raspberrypi hass[2828]: File "/usr/lib/python3.4/asyncio/tasks.py", line 286, in _wakeup
Apr 25 20:13:37 raspberrypi hass[2828]: value = future.result()
Apr 25 20:13:37 raspberrypi hass[2828]: File "/usr/lib/python3.4/asyncio/futures.py", line 277, in result
Apr 25 20:13:37 raspberrypi hass[2828]: raise self._exception
Apr 25 20:13:37 raspberrypi hass[2828]: File "/usr/lib/python3.4/concurrent/futures/thread.py", line 54, in run
Apr 25 20:13:37 raspberrypi hass[2828]: result = self.fn(*self.args, **self.kwargs)
Apr 25 20:13:37 raspberrypi hass[2828]: File "/srv/homeassistant/lib/python3.4/site-packages/homeassistant/components/notify/html5.py", line 349, in send_message
Apr 25 20:13:37 raspberrypi hass[2828]: from pywebpush import WebPusher
Apr 25 20:13:37 raspberrypi hass[2828]: File "/home/homeassistant/.homeassistant/deps/pywebpush/__init__.py", line 10, in <module>
Apr 25 20:13:37 raspberrypi hass[2828]: import http_ece
Apr 25 20:13:37 raspberrypi hass[2828]: File "/home/homeassistant/.homeassistant/deps/http_ece/__init__.py", line 8, in <module>
Apr 25 20:13:37 raspberrypi hass[2828]: import pyelliptic
Apr 25 20:13:37 raspberrypi hass[2828]: File "/home/homeassistant/.homeassistant/deps/pyelliptic/__init__.py", line 43, in <module>
Apr 25 20:13:37 raspberrypi hass[2828]: from .openssl import OpenSSL
Apr 25 20:13:37 raspberrypi hass[2828]: File "/home/homeassistant/.homeassistant/deps/pyelliptic/openssl.py", line 310, in <module>
Apr 25 20:13:37 raspberrypi hass[2828]: OpenSSL = _OpenSSL(libname)
Apr 25 20:13:37 raspberrypi hass[2828]: File "/home/homeassistant/.homeassistant/deps/pyelliptic/openssl.py", line 144, in __init__
Apr 25 20:13:37 raspberrypi hass[2828]: self.EVP_CIPHER_CTX_reset = self._lib.EVP_CIPHER_CTX_reset
Apr 25 20:13:37 raspberrypi hass[2828]: File "/usr/lib/python3.4/ctypes/__init__.py", line 364, in __getattr__
Apr 25 20:13:37 raspberrypi hass[2828]: func = self.__getitem__(name)
Apr 25 20:13:37 raspberrypi hass[2828]: File "/usr/lib/python3.4/ctypes/__init__.py", line 369, in __getitem__
Apr 25 20:13:37 raspberrypi hass[2828]: func = self._FuncPtr((name_or_ordinal, self))
Apr 25 20:13:37 raspberrypi hass[2828]: AttributeError: /usr/lib/arm-linux-gnueabihf/libcrypto.so.1.0.0: undefined symbol: EVP_CIPHER_CTX_reset
```
**Additional info:**
| null | https://github.com/home-assistant/core/pull/7310 | null | {'base_commit': 'f29e0bf53eec22dca65e3a62336cf84797e36bf2', 'files': [{'path': 'homeassistant/components/notify/html5.py', 'status': 'modified', 'Loc': {'(None, None, None)': {'mod': [28]}}}, {'path': 'requirements_all.txt', 'status': 'modified', 'Loc': {'(None, None, None)': {'add': [526]}}}]} | [] | [] | [] | {
"iss_type": "1",
"iss_reason": "1",
"loc_way": "pr",
"loc_scope": null,
"info_type": null
} | {
"code": [
"homeassistant/components/notify/html5.py"
],
"doc": [],
"test": [],
"config": [
"requirements_all.txt"
],
"asset": []
} | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.