Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 855 | labels stringlengths 4 721 | body stringlengths 1 261k | index stringclasses 13 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 240k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
466,654 | 13,430,857,528 | IssuesEvent | 2020-09-07 05:51:52 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | app.getpocket.com - desktop site instead of mobile site | browser-firefox engine-gecko ml-needsdiagnosis-false ml-probability-high priority-normal | <!-- @browser: Firefox 81.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:81.0) Gecko/20100101 Firefox/81.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/57745 -->
**URL**: https://app.getpocket.com/
**Browser / Version**: Firefox 81.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Edge
**Problem type**: Desktop site instead of mobile site
**Description**: Desktop site instead of mobile site
**Steps to Reproduce**:
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2020/9/9c82f483-788f-487f-858e-38568f16cd23.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200829200810</li><li>channel: aurora</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/9/9489329d-ab4f-4b90-a113-6c1eeb9764c0)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | app.getpocket.com - desktop site instead of mobile site - <!-- @browser: Firefox 81.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:81.0) Gecko/20100101 Firefox/81.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/57745 -->
**URL**: https://app.getpocket.com/
**Browser / Version**: Firefox 81.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Edge
**Problem type**: Desktop site instead of mobile site
**Description**: Desktop site instead of mobile site
**Steps to Reproduce**:
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2020/9/9c82f483-788f-487f-858e-38568f16cd23.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200829200810</li><li>channel: aurora</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/9/9489329d-ab4f-4b90-a113-6c1eeb9764c0)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | app getpocket com desktop site instead of mobile site url browser version firefox operating system windows tested another browser yes edge problem type desktop site instead of mobile site description desktop site instead of mobile site steps to reproduce view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel aurora hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️ | 1 |
397,380 | 11,727,643,889 | IssuesEvent | 2020-03-10 16:15:16 | privacytoolsIO/privacytools.io | https://api.github.com/repos/privacytoolsIO/privacytools.io | closed | Add steps to disable WebRTC in Safari | high priority 👁️ browsers 📝 correction | Please update the webRTC section about the Safari support. Current version of Safari just started working with webRTC.
EDIT: Sorry for the bad labeling, but you don’t have a good label for outdated info on the site, and this seemed like a close enough label 😅
| 1.0 | Add steps to disable WebRTC in Safari - Please update the webRTC section about the Safari support. Current version of Safari just started working with webRTC.
EDIT: Sorry for the bad labeling, but you don’t have a good label for outdated info on the site, and this seemed like a close enough label 😅
| priority | add steps to disable webrtc in safari please update the webrtc section about the safari support current version of safari just started working with webrtc edit sorry for the bad labeling but you don’t have a good label for outdated info on the site and this seemed like a close enough label 😅 | 1 |
153,467 | 5,892,761,782 | IssuesEvent | 2017-05-17 20:16:35 | denniscarr/dada | https://api.github.com/repos/denniscarr/dada | closed | More tasks | enhancement High priority | Task with more diversity, like killing NPCs, equipping objects, storing a certain number of things or a specific object in visor | 1.0 | More tasks - Task with more diversity, like killing NPCs, equipping objects, storing a certain number of things or a specific object in visor | priority | more tasks task with more diversity like killing npcs equipping objects storing a certain number of things or a specific object in visor | 1 |
623,208 | 19,663,214,363 | IssuesEvent | 2022-01-10 19:17:48 | Thorium-Sim/thorium | https://api.github.com/repos/Thorium-Sim/thorium | opened | Selecting Video Cards Crashes the Server | type/bug priority/high | ### Requested By: Alex DeBirk
### Priority: High
### Version: 3.5.1
When creating a timeline step that changes the viewscreen cards, selecting the card typically resets the server until it crashes.
### Steps to Reproduce
Using Linux Server, Windows Thorium editor, and Chrome. I create a timeline step that uses the Viewscreen: Change Viewscreen Card, and try to select the video. The card does not select. Any attempts to do so results in the last previous step that I created being reset. Eventually the server just crashes. | 1.0 | Selecting Video Cards Crashes the Server - ### Requested By: Alex DeBirk
### Priority: High
### Version: 3.5.1
When creating a timeline step that changes the viewscreen cards, selecting the card typically resets the server until it crashes.
### Steps to Reproduce
Using Linux Server, Windows Thorium editor, and Chrome. I create a timeline step that uses the Viewscreen: Change Viewscreen Card, and try to select the video. The card does not select. Any attempts to do so results in the last previous step that I created being reset. Eventually the server just crashes. | priority | selecting video cards crashes the server requested by alex debirk priority high version when creating a timeline step that changes the viewscreen cards selecting the card typically resets the server until it crashes steps to reproduce using linux server windows thorium editor and chrome i create a timeline step that uses the viewscreen change viewscreen card and try to select the video the card does not select any attempts to do so results in the last previous step that i created being reset eventually the server just crashes | 1 |
562,392 | 16,659,011,773 | IssuesEvent | 2021-06-06 02:39:54 | ut-code/utmap-times | https://api.github.com/repos/ut-code/utmap-times | closed | イベント記事タイプ/インターン記事タイプで、imageが表示されていない | Priority: High enhancement | イベント記事タイプ/インターン記事タイプで、imageが表示されていないので、表示させるようお願いいたします。
<img width="491" alt="スクリーンショット 2021-04-28 13 44 55" src="https://user-images.githubusercontent.com/80612908/116347829-ebe86900-a827-11eb-965a-a2d9e87ac65d.png">
| 1.0 | イベント記事タイプ/インターン記事タイプで、imageが表示されていない - イベント記事タイプ/インターン記事タイプで、imageが表示されていないので、表示させるようお願いいたします。
<img width="491" alt="スクリーンショット 2021-04-28 13 44 55" src="https://user-images.githubusercontent.com/80612908/116347829-ebe86900-a827-11eb-965a-a2d9e87ac65d.png">
| priority | イベント記事タイプ インターン記事タイプで、imageが表示されていない イベント記事タイプ インターン記事タイプで、imageが表示されていないので、表示させるようお願いいたします。 img width alt スクリーンショット src | 1 |
157,463 | 6,001,310,433 | IssuesEvent | 2017-06-05 08:49:25 | commons-app/apps-android-commons | https://api.github.com/repos/commons-app/apps-android-commons | closed | Tutorial messages use line breaks instead of '\n' after being translated | bug high priority localization | Tutorial messages that use `\n` for stylistic purposes actually appear in https://translatewiki.net as if having actual line breaks. Here's an example original message:
```
<string name="tutorial_2_subtext">- Natural objects (flowers, animals, mountains)\n- Useful objects (bicycles, train stations)\n- Famous people (your mayor, Olympic athletes you met)</string>
```
Shows as https://i.imgur.com/PYVYX4A.png
And naturally gets translated into:
```
<string name="tutorial_2_subtext">- Objets naturels (fleurs, animaux, montagnes)
- Objets utiles (bicyclettes, gares ferroviaires)
- Personnes célèbres (votre maire, les athlètes olympiques que vous avez rencontrés)</string>
```
The problem? It seems that the thing that lays out text in android only respects `\n` and not actual line breaks, which means that the dashes that supposed to be list bullets appear _mid-line_.
Live example (using a different string, because it looks worse): https://i.imgur.com/SYYvbU7.png
Since translations come from https://translatewiki.net, it could be a problem on their side, but it may be easy to work around this for now. | 1.0 | Tutorial messages use line breaks instead of '\n' after being translated - Tutorial messages that use `\n` for stylistic purposes actually appear in https://translatewiki.net as if having actual line breaks. Here's an example original message:
```
<string name="tutorial_2_subtext">- Natural objects (flowers, animals, mountains)\n- Useful objects (bicycles, train stations)\n- Famous people (your mayor, Olympic athletes you met)</string>
```
Shows as https://i.imgur.com/PYVYX4A.png
And naturally gets translated into:
```
<string name="tutorial_2_subtext">- Objets naturels (fleurs, animaux, montagnes)
- Objets utiles (bicyclettes, gares ferroviaires)
- Personnes célèbres (votre maire, les athlètes olympiques que vous avez rencontrés)</string>
```
The problem? It seems that the thing that lays out text in android only respects `\n` and not actual line breaks, which means that the dashes that supposed to be list bullets appear _mid-line_.
Live example (using a different string, because it looks worse): https://i.imgur.com/SYYvbU7.png
Since translations come from https://translatewiki.net, it could be a problem on their side, but it may be easy to work around this for now. | priority | tutorial messages use line breaks instead of n after being translated tutorial messages that use n for stylistic purposes actually appear in as if having actual line breaks here s an example original message natural objects flowers animals mountains n useful objects bicycles train stations n famous people your mayor olympic athletes you met shows as and naturally gets translated into objets naturels fleurs animaux montagnes objets utiles bicyclettes gares ferroviaires personnes célèbres votre maire les athlètes olympiques que vous avez rencontrés the problem it seems that the thing that lays out text in android only respects n and not actual line breaks which means that the dashes that supposed to be list bullets appear mid line live example using a different string because it looks worse since translations come from it could be a problem on their side but it may be easy to work around this for now | 1 |
651,981 | 21,517,411,500 | IssuesEvent | 2022-04-28 11:14:09 | ProsperityGH/VrijwilligersHuis-Opdracht | https://api.github.com/repos/ProsperityGH/VrijwilligersHuis-Opdracht | closed | Form frontend | enhancement High Priority | - [x] opmaak formulier css
- [x] contactgegevens
- [x] categories
- [x] locatie
- [x] hoeveelheid
- [x] opmerkingen | 1.0 | Form frontend - - [x] opmaak formulier css
- [x] contactgegevens
- [x] categories
- [x] locatie
- [x] hoeveelheid
- [x] opmerkingen | priority | form frontend opmaak formulier css contactgegevens categories locatie hoeveelheid opmerkingen | 1 |
311,130 | 9,528,991,865 | IssuesEvent | 2019-04-29 09:58:08 | Abwasserrohr/SKYBLOCK.SK | https://api.github.com/repos/Abwasserrohr/SKYBLOCK.SK | closed | Island creation process still uses old lore newline | bug priority:high | Because of this, lava is unusable on new islands. MMake a newline instead of || using a list. | 1.0 | Island creation process still uses old lore newline - Because of this, lava is unusable on new islands. MMake a newline instead of || using a list. | priority | island creation process still uses old lore newline because of this lava is unusable on new islands mmake a newline instead of using a list | 1 |
740,042 | 25,733,704,703 | IssuesEvent | 2022-12-07 22:30:54 | coder/code-server | https://api.github.com/repos/coder/code-server | closed | [Bug]: Invalid version in about window after 4.9.0 upgrade | bug high-priority | ### Is there an existing issue for this?
- [X] I have searched the existing issues
### OS/Web Information
- Web Browser: Chrome
- Local OS: Arch Linux
- Remote OS: Debian sid
- Remote Architecture: amd64
- `code-server --version`: 4.9.0 0502dfa1ff42ab8a43adb911f7bf21f8b09ee25f with Code 1.73.1
### Steps to Reproduce
1. Open code-server
2. Navigate to Menu->Help->About
### Expected
Expected to see `code-server: v4.9.0` in about dialog.
Expected to no see the blue icon in status bar suggesting to download code-server.
### Actual
There a `code-server: v$VERSION`

Also there's a blue icon in status bar suggesting to download code-server, despite having launched with `--disable-update-check`

### Logs
_No response_
### Screenshot/Video
_No response_
### Does this issue happen in VS Code or GitHub Codespaces?
- [X] I cannot reproduce this in VS Code.
- [X] I cannot reproduce this in GitHub Codespaces.
### Are you accessing code-server over HTTPS?
- [X] I am using HTTPS.
### Notes
_No response_ | 1.0 | [Bug]: Invalid version in about window after 4.9.0 upgrade - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### OS/Web Information
- Web Browser: Chrome
- Local OS: Arch Linux
- Remote OS: Debian sid
- Remote Architecture: amd64
- `code-server --version`: 4.9.0 0502dfa1ff42ab8a43adb911f7bf21f8b09ee25f with Code 1.73.1
### Steps to Reproduce
1. Open code-server
2. Navigate to Menu->Help->About
### Expected
Expected to see `code-server: v4.9.0` in about dialog.
Expected to no see the blue icon in status bar suggesting to download code-server.
### Actual
There a `code-server: v$VERSION`

Also there's a blue icon in status bar suggesting to download code-server, despite having launched with `--disable-update-check`

### Logs
_No response_
### Screenshot/Video
_No response_
### Does this issue happen in VS Code or GitHub Codespaces?
- [X] I cannot reproduce this in VS Code.
- [X] I cannot reproduce this in GitHub Codespaces.
### Are you accessing code-server over HTTPS?
- [X] I am using HTTPS.
### Notes
_No response_ | priority | invalid version in about window after upgrade is there an existing issue for this i have searched the existing issues os web information web browser chrome local os arch linux remote os debian sid remote architecture code server version with code steps to reproduce open code server navigate to menu help about expected expected to see code server in about dialog expected to no see the blue icon in status bar suggesting to download code server actual there a code server v version also there s a blue icon in status bar suggesting to download code server despite having launched with disable update check logs no response screenshot video no response does this issue happen in vs code or github codespaces i cannot reproduce this in vs code i cannot reproduce this in github codespaces are you accessing code server over https i am using https notes no response | 1 |
58,168 | 3,087,882,916 | IssuesEvent | 2015-08-25 14:11:48 | PeerSay/Atlas | https://api.github.com/repos/PeerSay/Atlas | opened | Add and Avg. PeerSay Grade columon for every product in the decision table | enhancement Priority: High | To the right of the current Grade field. | 1.0 | Add and Avg. PeerSay Grade columon for every product in the decision table - To the right of the current Grade field. | priority | add and avg peersay grade columon for every product in the decision table to the right of the current grade field | 1 |
372,592 | 11,017,387,510 | IssuesEvent | 2019-12-05 08:18:48 | ballerina-platform/ballerina-lang | https://api.github.com/repos/ballerina-platform/ballerina-lang | closed | Add goto definition support for annotations and errors | Area/Tooling Component/LanguageServer Points/3 Priority/High Type/Task | **Description**
Add the goto Definition support for annotations and errors
**Affected Versions**
v1.0.0 at least | 1.0 | Add goto definition support for annotations and errors - **Description**
Add the goto Definition support for annotations and errors
**Affected Versions**
v1.0.0 at least | priority | add goto definition support for annotations and errors description add the goto definition support for annotations and errors affected versions at least | 1 |
101,278 | 4,112,019,170 | IssuesEvent | 2016-06-07 08:55:03 | scalan/scalan | https://api.github.com/repos/scalan/scalan | closed | Generate separate nodes and short names for StructElems in GraphVizExport | high priority | Non-trivial structs make graph nodes very wide. | 1.0 | Generate separate nodes and short names for StructElems in GraphVizExport - Non-trivial structs make graph nodes very wide. | priority | generate separate nodes and short names for structelems in graphvizexport non trivial structs make graph nodes very wide | 1 |
782,044 | 27,484,810,055 | IssuesEvent | 2023-03-04 01:21:29 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | closed | Fused AdamW causes NaN loss | high priority module: optimizer triaged | ### 🐛 Describe the bug
There already has been an extended discussion of this issue over on the nanoGPT repository:
- karpathy/nanoGPT#167
I have been encouraged to submit a separate bug report here, since the issue seems to lie with pytorch. I have prepared a preconfigured [fork of nanoGPT](https://github.com/oddlama/nanoGPT_nan) where I provided my configuration and part of my dataset which causes the issue to appear. It now immediately produces NaNs after the first training step, but only if fused adam is in use.
The settings are basically the nanoGPT shakespeare configuration, but using a blocksize of 343 and vocab size of 2006. The data is quite sparse (only 1,3% is not 0, which significantly accelerates getting to the issue). I've included 1000 batches of my own real data so you will have the same conditions as I do. I'm training on a single 2080Ti, please look at a diff of the latest commit in this repository to see what I changed exactly - it's not much.
```bash
> git clone https://github.com/oddlama/nanoGPT_nan
> cd nanoGPT_nan
```
```bash
> python train.py config/nan.py --allow_fused=True
[...]
number of parameters: 11.39M
using fused AdamW: True
compiling the model... (takes a ~minute)
step 0: train loss 6.6010, val loss 6.6009
[2023-03-01 15:19:43,485] torch._inductor.utils: [WARNING] using triton random, expect difference from eager
/projects/venv/lib/python3.10/site-packages/torch/_functorch/aot_autograd.py:1251: UserWarning: Your compiler for AOTAutograd is returning a a function that doesn't take boxed arguments. Please wrap it with functorch.compile.make_boxed_func or handle the boxed arguments yourself. See https://github.com/pytorch/pytorch/pull/83137#issuecomment-1211320670 for rationale.
warnings.warn(
iter 0: loss 6.6793, time 22316.39ms, mfu -100.00%
iter 1: loss nan, time 3249.91ms, mfu -100.00%
```
```bash
> python train.py config/nan.py --allow_fused=False
[...]
number of parameters: 11.39M
using fused AdamW: False
compiling the model... (takes a ~minute)
step 0: train loss 6.6010, val loss 6.6009
[2023-03-01 15:20:26,110] torch._inductor.utils: [WARNING] using triton random, expect difference from eager
/projects/nanoGPT_nan/venv/lib/python3.10/site-packages/torch/_functorch/aot_autograd.py:1251: UserWarning: Your compiler for AOTAutograd is returning a a function that doesn't take boxed arguments. Please wrap it with functorch.compile.make_boxed_func or handle the boxed arguments yourself. See https://github.com/pytorch/pytorch/pull/83137#issuecomment-1211320670 for rationale.
warnings.warn(
iter 0: loss 6.6793, time 22286.14ms, mfu -100.00%
iter 1: loss 6.6811, time 3534.44ms, mfu -100.00%
iter 2: loss 6.6770, time 3365.06ms, mfu -100.00%
iter 3: loss 4.7208, time 3366.52ms, mfu -100.00%
iter 4: loss 3.0927, time 3365.22ms, mfu -100.00%
iter 5: loss 2.0694, time 3362.80ms, mfu 6.51%
iter 6: loss 1.4752, time 3360.81ms, mfu 6.51%
# continues to work fine
```
Hope this helps.
P.S.:
Can anyone explain what the warning means? Doesn't seem to be related to the issue, but is it a user error that should be fixed?
```
[2023-03-01 15:19:43,485] torch._inductor.utils: [WARNING] using triton random, expect difference from eager
/projects/venv/lib/python3.10/site-packages/torch/_functorch/aot_autograd.py:1251: UserWarning: Your compiler for AOTAutograd is returning a a function that doesn't take boxed arguments. Please wrap it with functorch.compile.make_boxed_func or handle the boxed arguments yourself. See https://github.com/pytorch/pytorch/pull/83137#issuecomment-1211320670 for rationale.
warnings.warn(
```
### Versions
```
Collecting environment information...
PyTorch version: 2.0.0.dev20230220+cu118
Is debug build: False
CUDA used to build PyTorch: 11.8
ROCM used to build PyTorch: N/A
OS: Arch Linux (x86_64)
GCC version: (GCC) 12.2.1 20230201
Clang version: 15.0.7
CMake version: version 3.25.0
Libc version: glibc-2.37
Python version: 3.10.9 (main, Dec 19 2022, 17:35:49) [GCC 12.2.0] (64-bit runtime)
Python platform: Linux-5.15.79.1-microsoft-standard-WSL2-x86_64-with-glibc2.37
Is CUDA available: True
CUDA runtime version: 11.8.89
CUDA_MODULE_LOADING set to: LAZY
GPU models and configuration: GPU 0: NVIDIA GeForce RTX 2080 Ti
Nvidia driver version: 516.94
cuDNN version: Could not collect
HIP runtime version: N/A
MIOpen runtime version: N/A
Is XNNPACK available: True
CPU:
Architecture: x86_64
CPU op-mode(s): 32-bit, 64-bit
Address sizes: 48 bits physical, 48 bits virtual
Byte Order: Little Endian
CPU(s): 24
On-line CPU(s) list: 0-23
Vendor ID: AuthenticAMD
Model name: AMD Ryzen 9 5900X 12-Core Processor
CPU family: 25
Model: 33
Thread(s) per core: 2
Core(s) per socket: 12
Socket(s): 1
Stepping: 0
BogoMIPS: 7400.01
Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc rep_good nopl tsc_reliable nonstop_tsc cpuid extd_apicid pni pclmulqdq ssse3 fma cx16 sse4_1 sse4_2 movbe popcnt aes xsave avx f16c rdrand hypervisor lahf_lm cmp_legacy cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw topoext ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 erms rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 xsaves clzero xsaveerptr arat umip vaes vpclmulqdq rdpid fsrm
Hypervisor vendor: Microsoft
Virtualization type: full
L1d cache: 384 KiB (12 instances)
L1i cache: 384 KiB (12 instances)
L2 cache: 6 MiB (12 instances)
L3 cache: 32 MiB (1 instance)
Vulnerability Itlb multihit: Not affected
Vulnerability L1tf: Not affected
Vulnerability Mds: Not affected
Vulnerability Meltdown: Not affected
Vulnerability Mmio stale data: Not affected
Vulnerability Retbleed: Not affected
Vulnerability Spec store bypass: Vulnerable
Vulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization
Vulnerability Spectre v2: Mitigation; Retpolines, IBPB conditional, IBRS_FW, STIBP conditional, RSB filling, PBRSB-eIBRS Not affected
Vulnerability Srbds: Not affected
Vulnerability Tsx async abort: Not affected
Versions of relevant libraries:
[pip3] mypy-extensions==1.0.0
[pip3] numpy==1.24.1
[pip3] pytorch-triton==2.0.0+d54c04abe2
[pip3] torch==2.0.0.dev20230220+cu118
[pip3] torchaudio==2.0.0.dev20230223+cu118
[pip3] torchinfo==1.7.2
[pip3] torchsummary==1.5.1
[pip3] torchvision==0.15.0.dev20230223+cu118
[conda] Could not collect
```
cc @ezyang @gchanan @zou3519 @vincentqb @jbschlosser @albanD @janeyx99 | 1.0 | Fused AdamW causes NaN loss - ### 🐛 Describe the bug
There already has been an extended discussion of this issue over on the nanoGPT repository:
- karpathy/nanoGPT#167
I have been encouraged to submit a separate bug report here, since the issue seems to lie with pytorch. I have prepared a preconfigured [fork of nanoGPT](https://github.com/oddlama/nanoGPT_nan) where I provided my configuration and part of my dataset which causes the issue to appear. It now immediately produces NaNs after the first training step, but only if fused adam is in use.
The settings are basically the nanoGPT shakespeare configuration, but using a blocksize of 343 and vocab size of 2006. The data is quite sparse (only 1,3% is not 0, which significantly accelerates getting to the issue). I've included 1000 batches of my own real data so you will have the same conditions as I do. I'm training on a single 2080Ti, please look at a diff of the latest commit in this repository to see what I changed exactly - it's not much.
```bash
> git clone https://github.com/oddlama/nanoGPT_nan
> cd nanoGPT_nan
```
```bash
> python train.py config/nan.py --allow_fused=True
[...]
number of parameters: 11.39M
using fused AdamW: True
compiling the model... (takes a ~minute)
step 0: train loss 6.6010, val loss 6.6009
[2023-03-01 15:19:43,485] torch._inductor.utils: [WARNING] using triton random, expect difference from eager
/projects/venv/lib/python3.10/site-packages/torch/_functorch/aot_autograd.py:1251: UserWarning: Your compiler for AOTAutograd is returning a a function that doesn't take boxed arguments. Please wrap it with functorch.compile.make_boxed_func or handle the boxed arguments yourself. See https://github.com/pytorch/pytorch/pull/83137#issuecomment-1211320670 for rationale.
warnings.warn(
iter 0: loss 6.6793, time 22316.39ms, mfu -100.00%
iter 1: loss nan, time 3249.91ms, mfu -100.00%
```
```bash
> python train.py config/nan.py --allow_fused=False
[...]
number of parameters: 11.39M
using fused AdamW: False
compiling the model... (takes a ~minute)
step 0: train loss 6.6010, val loss 6.6009
[2023-03-01 15:20:26,110] torch._inductor.utils: [WARNING] using triton random, expect difference from eager
/projects/nanoGPT_nan/venv/lib/python3.10/site-packages/torch/_functorch/aot_autograd.py:1251: UserWarning: Your compiler for AOTAutograd is returning a a function that doesn't take boxed arguments. Please wrap it with functorch.compile.make_boxed_func or handle the boxed arguments yourself. See https://github.com/pytorch/pytorch/pull/83137#issuecomment-1211320670 for rationale.
warnings.warn(
iter 0: loss 6.6793, time 22286.14ms, mfu -100.00%
iter 1: loss 6.6811, time 3534.44ms, mfu -100.00%
iter 2: loss 6.6770, time 3365.06ms, mfu -100.00%
iter 3: loss 4.7208, time 3366.52ms, mfu -100.00%
iter 4: loss 3.0927, time 3365.22ms, mfu -100.00%
iter 5: loss 2.0694, time 3362.80ms, mfu 6.51%
iter 6: loss 1.4752, time 3360.81ms, mfu 6.51%
# continues to work fine
```
Hope this helps.
P.S.:
Can anyone explain what the warning means? Doesn't seem to be related to the issue, but is it a user error that should be fixed?
```
[2023-03-01 15:19:43,485] torch._inductor.utils: [WARNING] using triton random, expect difference from eager
/projects/venv/lib/python3.10/site-packages/torch/_functorch/aot_autograd.py:1251: UserWarning: Your compiler for AOTAutograd is returning a a function that doesn't take boxed arguments. Please wrap it with functorch.compile.make_boxed_func or handle the boxed arguments yourself. See https://github.com/pytorch/pytorch/pull/83137#issuecomment-1211320670 for rationale.
warnings.warn(
```
### Versions
```
Collecting environment information...
PyTorch version: 2.0.0.dev20230220+cu118
Is debug build: False
CUDA used to build PyTorch: 11.8
ROCM used to build PyTorch: N/A
OS: Arch Linux (x86_64)
GCC version: (GCC) 12.2.1 20230201
Clang version: 15.0.7
CMake version: version 3.25.0
Libc version: glibc-2.37
Python version: 3.10.9 (main, Dec 19 2022, 17:35:49) [GCC 12.2.0] (64-bit runtime)
Python platform: Linux-5.15.79.1-microsoft-standard-WSL2-x86_64-with-glibc2.37
Is CUDA available: True
CUDA runtime version: 11.8.89
CUDA_MODULE_LOADING set to: LAZY
GPU models and configuration: GPU 0: NVIDIA GeForce RTX 2080 Ti
Nvidia driver version: 516.94
cuDNN version: Could not collect
HIP runtime version: N/A
MIOpen runtime version: N/A
Is XNNPACK available: True
CPU:
Architecture: x86_64
CPU op-mode(s): 32-bit, 64-bit
Address sizes: 48 bits physical, 48 bits virtual
Byte Order: Little Endian
CPU(s): 24
On-line CPU(s) list: 0-23
Vendor ID: AuthenticAMD
Model name: AMD Ryzen 9 5900X 12-Core Processor
CPU family: 25
Model: 33
Thread(s) per core: 2
Core(s) per socket: 12
Socket(s): 1
Stepping: 0
BogoMIPS: 7400.01
Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc rep_good nopl tsc_reliable nonstop_tsc cpuid extd_apicid pni pclmulqdq ssse3 fma cx16 sse4_1 sse4_2 movbe popcnt aes xsave avx f16c rdrand hypervisor lahf_lm cmp_legacy cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw topoext ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 erms rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 xsaves clzero xsaveerptr arat umip vaes vpclmulqdq rdpid fsrm
Hypervisor vendor: Microsoft
Virtualization type: full
L1d cache: 384 KiB (12 instances)
L1i cache: 384 KiB (12 instances)
L2 cache: 6 MiB (12 instances)
L3 cache: 32 MiB (1 instance)
Vulnerability Itlb multihit: Not affected
Vulnerability L1tf: Not affected
Vulnerability Mds: Not affected
Vulnerability Meltdown: Not affected
Vulnerability Mmio stale data: Not affected
Vulnerability Retbleed: Not affected
Vulnerability Spec store bypass: Vulnerable
Vulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization
Vulnerability Spectre v2: Mitigation; Retpolines, IBPB conditional, IBRS_FW, STIBP conditional, RSB filling, PBRSB-eIBRS Not affected
Vulnerability Srbds: Not affected
Vulnerability Tsx async abort: Not affected
Versions of relevant libraries:
[pip3] mypy-extensions==1.0.0
[pip3] numpy==1.24.1
[pip3] pytorch-triton==2.0.0+d54c04abe2
[pip3] torch==2.0.0.dev20230220+cu118
[pip3] torchaudio==2.0.0.dev20230223+cu118
[pip3] torchinfo==1.7.2
[pip3] torchsummary==1.5.1
[pip3] torchvision==0.15.0.dev20230223+cu118
[conda] Could not collect
```
cc @ezyang @gchanan @zou3519 @vincentqb @jbschlosser @albanD @janeyx99 | priority | fused adamw causes nan loss 🐛 describe the bug there already has been an extended discussion of this issue over on the nanogpt repository karpathy nanogpt i have been encouraged to submit a separate bug report here since the issue seems to lie with pytorch i have prepared a preconfigured where i provided my configuration and part of my dataset which causes the issue to appear it now immediately produces nans after the first training step but only if fused adam is in use the settings are basically the nanogpt shakespeare configuration but using a blocksize of and vocab size of the data is quite sparse only is not which significantly accelerates getting to the issue i ve included batches of my own real data so you will have the same conditions as i do i m training on a single please look at a diff of the latest commit in this repository to see what i changed exactly it s not much bash git clone cd nanogpt nan bash python train py config nan py allow fused true number of parameters using fused adamw true compiling the model takes a minute step train loss val loss torch inductor utils using triton random expect difference from eager projects venv lib site packages torch functorch aot autograd py userwarning your compiler for aotautograd is returning a a function that doesn t take boxed arguments please wrap it with functorch compile make boxed func or handle the boxed arguments yourself see for rationale warnings warn iter loss time mfu iter loss nan time mfu bash python train py config nan py allow fused false number of parameters using fused adamw false compiling the model takes a minute step train loss val loss torch inductor utils using triton random expect difference from eager projects nanogpt nan venv lib site packages torch functorch aot autograd py userwarning your compiler for aotautograd is returning a a function that doesn t take boxed arguments please wrap it with functorch compile make boxed func or handle the boxed arguments yourself see for rationale warnings warn iter loss time mfu iter loss time mfu iter loss time mfu iter loss time mfu iter loss time mfu iter loss time mfu iter loss time mfu continues to work fine hope this helps p s can anyone explain what the warning means doesn t seem to be related to the issue but is it a user error that should be fixed torch inductor utils using triton random expect difference from eager projects venv lib site packages torch functorch aot autograd py userwarning your compiler for aotautograd is returning a a function that doesn t take boxed arguments please wrap it with functorch compile make boxed func or handle the boxed arguments yourself see for rationale warnings warn versions collecting environment information pytorch version is debug build false cuda used to build pytorch rocm used to build pytorch n a os arch linux gcc version gcc clang version cmake version version libc version glibc python version main dec bit runtime python platform linux microsoft standard with is cuda available true cuda runtime version cuda module loading set to lazy gpu models and configuration gpu nvidia geforce rtx ti nvidia driver version cudnn version could not collect hip runtime version n a miopen runtime version n a is xnnpack available true cpu architecture cpu op mode s bit bit address sizes bits physical bits virtual byte order little endian cpu s on line cpu s list vendor id authenticamd model name amd ryzen core processor cpu family model thread s per core core s per socket socket s stepping bogomips flags fpu vme de pse tsc msr pae mce apic sep mtrr pge mca cmov pat clflush mmx fxsr sse ht syscall nx mmxext fxsr opt rdtscp lm constant tsc rep good nopl tsc reliable nonstop tsc cpuid extd apicid pni pclmulqdq fma movbe popcnt aes xsave avx rdrand hypervisor lahf lm cmp legacy legacy abm misalignsse osvw topoext ibrs ibpb stibp vmmcall fsgsbase smep erms rdseed adx smap clflushopt clwb sha ni xsaveopt xsavec xsaves clzero xsaveerptr arat umip vaes vpclmulqdq rdpid fsrm hypervisor vendor microsoft virtualization type full cache kib instances cache kib instances cache mib instances cache mib instance vulnerability itlb multihit not affected vulnerability not affected vulnerability mds not affected vulnerability meltdown not affected vulnerability mmio stale data not affected vulnerability retbleed not affected vulnerability spec store bypass vulnerable vulnerability spectre mitigation usercopy swapgs barriers and user pointer sanitization vulnerability spectre mitigation retpolines ibpb conditional ibrs fw stibp conditional rsb filling pbrsb eibrs not affected vulnerability srbds not affected vulnerability tsx async abort not affected versions of relevant libraries mypy extensions numpy pytorch triton torch torchaudio torchinfo torchsummary torchvision could not collect cc ezyang gchanan vincentqb jbschlosser alband | 1 |
173,671 | 6,529,399,966 | IssuesEvent | 2017-08-30 11:28:35 | donniedarkoparko/app | https://api.github.com/repos/donniedarkoparko/app | closed | [Both] - Place - 2x messages when signing out and back in | bug priority: high | It seems like the Place listeners aren't getting detached on sign out. | 1.0 | [Both] - Place - 2x messages when signing out and back in - It seems like the Place listeners aren't getting detached on sign out. | priority | place messages when signing out and back in it seems like the place listeners aren t getting detached on sign out | 1 |
117,568 | 4,718,547,436 | IssuesEvent | 2016-10-17 03:08:26 | google/paco | https://api.github.com/repos/google/paco | closed | iOS - After joining show Post Join Install Instructions | Priority-High | The experiment provides a field called postJoinInstallInstructions that contains instructions to show the user once they consent to the informed consent.
This text should be shown above the Edit Alert Times text. It might scroll, but, we still need to show the Edit Alert Times button. Also, this is where we might ask for the location permission for an experiment. | 1.0 | iOS - After joining show Post Join Install Instructions - The experiment provides a field called postJoinInstallInstructions that contains instructions to show the user once they consent to the informed consent.
This text should be shown above the Edit Alert Times text. It might scroll, but, we still need to show the Edit Alert Times button. Also, this is where we might ask for the location permission for an experiment. | priority | ios after joining show post join install instructions the experiment provides a field called postjoininstallinstructions that contains instructions to show the user once they consent to the informed consent this text should be shown above the edit alert times text it might scroll but we still need to show the edit alert times button also this is where we might ask for the location permission for an experiment | 1 |
562,662 | 16,666,305,913 | IssuesEvent | 2021-06-07 04:44:25 | adirh3/Fluent-Search | https://api.github.com/repos/adirh3/Fluent-Search | closed | High Priority: Web Searches broken | High Priority bug | **Describe the bug**
Web searches are not triggering even if they're enabled. It might have broke the version _before_ 0.9.85.1
**To Reproduce**
Steps to reproduce the behavior:
1. Type "g <search string>
2. Observe that FS doesn't search google.
I have google in web search:

Search turned on:

**Expected behavior**
Google search should be triggered
**Screenshots**
Instead, I get this:

100% reproducible
**Desktop (please complete the following information):**
- Windows 10 LTSC v1809

- Fluent Search Version [e.g. 0.9.85.1] although it might have broken version before.
| 1.0 | High Priority: Web Searches broken - **Describe the bug**
Web searches are not triggering even if they're enabled. It might have broke the version _before_ 0.9.85.1
**To Reproduce**
Steps to reproduce the behavior:
1. Type "g <search string>
2. Observe that FS doesn't search google.
I have google in web search:

Search turned on:

**Expected behavior**
Google search should be triggered
**Screenshots**
Instead, I get this:

100% reproducible
**Desktop (please complete the following information):**
- Windows 10 LTSC v1809

- Fluent Search Version [e.g. 0.9.85.1] although it might have broken version before.
| priority | high priority web searches broken describe the bug web searches are not triggering even if they re enabled it might have broke the version before to reproduce steps to reproduce the behavior type g observe that fs doesn t search google i have google in web search search turned on expected behavior google search should be triggered screenshots instead i get this reproducible desktop please complete the following information windows ltsc fluent search version although it might have broken version before | 1 |
726,016 | 24,984,619,235 | IssuesEvent | 2022-11-02 14:15:03 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | opened | ASAN shard 4 started to OOM after unrelated commit | high priority module: ci triaged | ### 🐛 Describe the bug
See https://hud.pytorch.org/hud/pytorch/pytorch/master/1?per_page=50&name_filter=asan%20%2F%20test%20(default%2C%204%2C%205
Error looks as follows:
```
test_serialization_offset_filelike_weights_only_False (__main__.TestOldSerialization) ... =================================================================
==1033==ERROR: AddressSanitizer: allocator is out of memory trying to allocate 0x80000000 bytes
```
First commit (clearly unrelated): https://hud.pytorch.org/pytorch/pytorch/commit/a51da28551e9f13a7afca5bbc829a8d9abced44e
### Versions
CI | 1.0 | ASAN shard 4 started to OOM after unrelated commit - ### 🐛 Describe the bug
See https://hud.pytorch.org/hud/pytorch/pytorch/master/1?per_page=50&name_filter=asan%20%2F%20test%20(default%2C%204%2C%205
Error looks as follows:
```
test_serialization_offset_filelike_weights_only_False (__main__.TestOldSerialization) ... =================================================================
==1033==ERROR: AddressSanitizer: allocator is out of memory trying to allocate 0x80000000 bytes
```
First commit (clearly unrelated): https://hud.pytorch.org/pytorch/pytorch/commit/a51da28551e9f13a7afca5bbc829a8d9abced44e
### Versions
CI | priority | asan shard started to oom after unrelated commit 🐛 describe the bug see error looks as follows test serialization offset filelike weights only false main testoldserialization error addresssanitizer allocator is out of memory trying to allocate bytes first commit clearly unrelated versions ci | 1 |
354,255 | 10,564,549,243 | IssuesEvent | 2019-10-05 02:56:06 | python/mypy | https://api.github.com/repos/python/mypy | closed | mypy 0.730 no longer recognizes x**y literals to be ints | bug needs discussion priority-0-high | I noticed a regression when switching from 0.720 to 0.730. Here's the code:
```py
def f() -> int:
size = 2 ** 20
reveal_type(size)
return size
```
**Expected behavior**
With mypy 0.720 (no matter whether the new or the old semantic analyzer is used), I get:
```
$ mypy --strict f.py
f.py:3: note: Revealed type is 'builtins.int'
```
**Actual behavior**
With mypy 0.730 (Python 3.6.8), I get:
```
$ mypy --strict f.py
f.py:3: note: Revealed type is 'Any'
f.py:4: error: Returning Any from function declared to return "int"
```
**Notes**
The behavior from 0.720 is what I'd expect.
One thing I noticed is that the typeshed indeed declares that `__pow__` for two `int`s returns `Any`. This makes sense because `2 ** -2` is a `float`. The only recent change which I can see in the typeshed though is the following:
```diff
--- a/stdlib/2and3/builtins.pyi
+++ b/stdlib/2and3/builtins.pyi
@@ -168,7 +168,7 @@ class int:
def __rtruediv__(self, x: int) -> float: ...
def __rmod__(self, x: int) -> int: ...
def __rdivmod__(self, x: int) -> Tuple[int, int]: ...
- def __pow__(self, x: int) -> Any: ... # Return type can be int or float, depending on x.
+ def __pow__(self, __x: int, __modulo: Optional[int] = ...) -> Any: ... # Return type can be int or float, depending on x.
def __rpow__(self, x: int) -> Any: ...
def __and__(self, n: int) -> int: ...
```
(see: https://github.com/python/typeshed/commit/b2cd972b1760d850694772fe763e3027f131f223)
This means that the `Any` return type has always been there but `mypy` used to correctly recognize that `2 ** 20` is an `int`. It's difficult for me to tell whether the behavior in 0.730 is a regression compared to 0.720 or actually the desired behavior – it seems to me that `mypy` follows the typeshed more precisely here. However, maybe it's worth to handle literals like `2 ** 30` in a way that `mypy` knows that this is always `int`? | 1.0 | mypy 0.730 no longer recognizes x**y literals to be ints - I noticed a regression when switching from 0.720 to 0.730. Here's the code:
```py
def f() -> int:
size = 2 ** 20
reveal_type(size)
return size
```
**Expected behavior**
With mypy 0.720 (no matter whether the new or the old semantic analyzer is used), I get:
```
$ mypy --strict f.py
f.py:3: note: Revealed type is 'builtins.int'
```
**Actual behavior**
With mypy 0.730 (Python 3.6.8), I get:
```
$ mypy --strict f.py
f.py:3: note: Revealed type is 'Any'
f.py:4: error: Returning Any from function declared to return "int"
```
**Notes**
The behavior from 0.720 is what I'd expect.
One thing I noticed is that the typeshed indeed declares that `__pow__` for two `int`s returns `Any`. This makes sense because `2 ** -2` is a `float`. The only recent change which I can see in the typeshed though is the following:
```diff
--- a/stdlib/2and3/builtins.pyi
+++ b/stdlib/2and3/builtins.pyi
@@ -168,7 +168,7 @@ class int:
def __rtruediv__(self, x: int) -> float: ...
def __rmod__(self, x: int) -> int: ...
def __rdivmod__(self, x: int) -> Tuple[int, int]: ...
- def __pow__(self, x: int) -> Any: ... # Return type can be int or float, depending on x.
+ def __pow__(self, __x: int, __modulo: Optional[int] = ...) -> Any: ... # Return type can be int or float, depending on x.
def __rpow__(self, x: int) -> Any: ...
def __and__(self, n: int) -> int: ...
```
(see: https://github.com/python/typeshed/commit/b2cd972b1760d850694772fe763e3027f131f223)
This means that the `Any` return type has always been there but `mypy` used to correctly recognize that `2 ** 20` is an `int`. It's difficult for me to tell whether the behavior in 0.730 is a regression compared to 0.720 or actually the desired behavior – it seems to me that `mypy` follows the typeshed more precisely here. However, maybe it's worth to handle literals like `2 ** 30` in a way that `mypy` knows that this is always `int`? | priority | mypy no longer recognizes x y literals to be ints i noticed a regression when switching from to here s the code py def f int size reveal type size return size expected behavior with mypy no matter whether the new or the old semantic analyzer is used i get mypy strict f py f py note revealed type is builtins int actual behavior with mypy python i get mypy strict f py f py note revealed type is any f py error returning any from function declared to return int notes the behavior from is what i d expect one thing i noticed is that the typeshed indeed declares that pow for two int s returns any this makes sense because is a float the only recent change which i can see in the typeshed though is the following diff a stdlib builtins pyi b stdlib builtins pyi class int def rtruediv self x int float def rmod self x int int def rdivmod self x int tuple def pow self x int any return type can be int or float depending on x def pow self x int modulo optional any return type can be int or float depending on x def rpow self x int any def and self n int int see this means that the any return type has always been there but mypy used to correctly recognize that is an int it s difficult for me to tell whether the behavior in is a regression compared to or actually the desired behavior – it seems to me that mypy follows the typeshed more precisely here however maybe it s worth to handle literals like in a way that mypy knows that this is always int | 1 |
173,854 | 6,533,216,377 | IssuesEvent | 2017-08-31 04:49:11 | JujaLabs/gamification-slack-bot | https://api.github.com/repos/JujaLabs/gamification-slack-bot | closed | Refactoring SlackCommandController and DefaultGamificationService | enhancement High priority | move slackNameHandlerService from GamificationSlackCommandController to DefaultGamificationService
sample code in the keeper-slack-bot | 1.0 | Refactoring SlackCommandController and DefaultGamificationService - move slackNameHandlerService from GamificationSlackCommandController to DefaultGamificationService
sample code in the keeper-slack-bot | priority | refactoring slackcommandcontroller and defaultgamificationservice move slacknamehandlerservice from gamificationslackcommandcontroller to defaultgamificationservice sample code in the keeper slack bot | 1 |
304,844 | 9,336,446,078 | IssuesEvent | 2019-03-28 21:15:38 | thebarrensissues/Issues | https://api.github.com/repos/thebarrensissues/Issues | closed | Deeprun Tram movement bugged. Trams moving through walls/other Trams | bug gamebreaking gameobject high priority | **Description:**
The movement of the Deeprun Trams is bugged, causing trams to disappear into walls or fly through other trams before disappearing into walls.
This will cause players to be flung into walls and dropped to the ground.

Fig1. _map of stormwind side tram station. trams numbered._
**Current behaviour:**
Please refer to Fig1.
Trams 1,2,4,5 all have correct movement and behaviour.
Tram 3 will move to the **west** when on the ironforge side, causing it to move into a wall.
Tram 6 will move to the **east** when on the stormwind side, causing it to clip through trams 4 and 5 before moving into a wall.
Tram 3 will not be present on the stormwind side station.
Tram 6 will not be present on the ironforge side station.


**Expected behaviour:**
All trams should not intercept each other nor move into walls.
**Steps to reproduce the problem:**
1. Go to the Deeprun tram, either at ironforge or stormwind
2. Observe the northern-most tram (1,2,3)
3. Observe the southern-most tram (4,5,6)
4. notice how trams 3 and 6 are bugged
**rev. hash/commit:**
**rev. 10725** | 1.0 | Deeprun Tram movement bugged. Trams moving through walls/other Trams - **Description:**
The movement of the Deeprun Trams is bugged, causing trams to disappear into walls or fly through other trams before disappearing into walls.
This will cause players to be flung into walls and dropped to the ground.

Fig1. _map of stormwind side tram station. trams numbered._
**Current behaviour:**
Please refer to Fig1.
Trams 1,2,4,5 all have correct movement and behaviour.
Tram 3 will move to the **west** when on the ironforge side, causing it to move into a wall.
Tram 6 will move to the **east** when on the stormwind side, causing it to clip through trams 4 and 5 before moving into a wall.
Tram 3 will not be present on the stormwind side station.
Tram 6 will not be present on the ironforge side station.


**Expected behaviour:**
All trams should not intercept each other nor move into walls.
**Steps to reproduce the problem:**
1. Go to the Deeprun tram, either at ironforge or stormwind
2. Observe the northern-most tram (1,2,3)
3. Observe the southern-most tram (4,5,6)
4. notice how trams 3 and 6 are bugged
**rev. hash/commit:**
**rev. 10725** | priority | deeprun tram movement bugged trams moving through walls other trams description the movement of the deeprun trams is bugged causing trams to disappear into walls or fly through other trams before disappearing into walls this will cause players to be flung into walls and dropped to the ground map of stormwind side tram station trams numbered current behaviour please refer to trams all have correct movement and behaviour tram will move to the west when on the ironforge side causing it to move into a wall tram will move to the east when on the stormwind side causing it to clip through trams and before moving into a wall tram will not be present on the stormwind side station tram will not be present on the ironforge side station expected behaviour all trams should not intercept each other nor move into walls steps to reproduce the problem go to the deeprun tram either at ironforge or stormwind observe the northern most tram observe the southern most tram notice how trams and are bugged rev hash commit rev | 1 |
122,242 | 4,828,699,874 | IssuesEvent | 2016-11-07 16:54:29 | meumobi/sitebuilder | https://api.github.com/repos/meumobi/sitebuilder | opened | Push not received | bug high priority push notification | I've sent push on various sites, with segmented audience and not, but never received push on my ios.
Can't digg further on my investigation due to issue #431 | 1.0 | Push not received - I've sent push on various sites, with segmented audience and not, but never received push on my ios.
Can't digg further on my investigation due to issue #431 | priority | push not received i ve sent push on various sites with segmented audience and not but never received push on my ios can t digg further on my investigation due to issue | 1 |
771,727 | 27,090,642,002 | IssuesEvent | 2023-02-14 20:44:34 | larray-project/larray | https://api.github.com/repos/larray-project/larray | closed | add some axes info in error message on invalid label | enhancement difficulty: low priority: high work in progress size: small | Having something **similar** to AxisCollection.info added to the "%r is not a valid label for any axis" message would very helpful. I have actually hacked my version several times to do just that and found it very helpful. I would only avoid displaying the first line with the array shape. | 1.0 | add some axes info in error message on invalid label - Having something **similar** to AxisCollection.info added to the "%r is not a valid label for any axis" message would very helpful. I have actually hacked my version several times to do just that and found it very helpful. I would only avoid displaying the first line with the array shape. | priority | add some axes info in error message on invalid label having something similar to axiscollection info added to the r is not a valid label for any axis message would very helpful i have actually hacked my version several times to do just that and found it very helpful i would only avoid displaying the first line with the array shape | 1 |
807,032 | 29,932,709,297 | IssuesEvent | 2023-06-22 10:34:57 | NikkelM/Random-YouTube-Video | https://api.github.com/repos/NikkelM/Random-YouTube-Video | closed | [Bug] If fetching videos takes too long, the background worker shuts down | Bug Priority: High | Example channel as reported on the Chrome Web store: https://www.youtube.com/@EpicSkillshot | 1.0 | [Bug] If fetching videos takes too long, the background worker shuts down - Example channel as reported on the Chrome Web store: https://www.youtube.com/@EpicSkillshot | priority | if fetching videos takes too long the background worker shuts down example channel as reported on the chrome web store | 1 |
493,708 | 14,236,987,033 | IssuesEvent | 2020-11-18 16:39:15 | craftercms/craftercms | https://api.github.com/repos/craftercms/craftercms | closed | CRAFTERCMS-2185: Cropping an image should create a new image with a name that indicates the crop details | new feature priority: high | Original JIRA Fix Versions:
2.5.2, Original JIRA Components:
Studio,
----------
Original JIRA Description: :
----------
Original JIRA Comments:
Update Crop Service code to create a custom filename if supplied
Update dialog to use custom name, if name exists offer to rename (but pre-populate field with value)
=======================
import java.awt.image.BufferedImage
import java.io.ByteArrayOutputStream
import java.io.ByteArrayInputStream
import java.io.File
import scripts.api.ContentServices
import javax.imageio.ImageIO
def req = request
def site = params.site
def imgPath = params.path
def newName = params.newname
def t = params.t.toInteger()
def l = params.l.toInteger()
def w = params.w.toInteger()
def h = params.h.toInteger()
def newName = params.newFileName
def imgToCrop = null
def imgCropped = null
def imgCroppedOutStream = null
def imgCroppedInStream = null
def imgType = imgPath.substring(imgPath.indexOf(".")+1)
def imgPathOnly = imgPath.substring(0, imgPath.lastIndexOf("/"))
def imgFilename = imgPath.substring(imgPath.lastIndexOf("/")+1)
def newImgFilename =(newFileName!=null) ? newFileName: imgFilename
if (newName) {
imgFilename = newName;
}
def context = ContentServices.createContext(applicationContext, request)
imgToCrop = ImageIO.read(ContentServices.getContentAsStream(site, imgPath, context))
imgCropped = imgToCrop.getSubimage(l, t, w, h)
imgCroppedOutStream = new ByteArrayOutputStream()
ImageIO.write(imgCropped, imgType, imgCroppedOutStream)
imgCroppedInStream = new ByteArrayInputStream(imgCroppedOutStream.toByteArray())
def result = ContentServices.writeContentAsset(context, site, imgPathOnly, newImgFilename, imgCroppedInStream, "true", "", "", "", "false", "true", null);
return result
----------
UI should recommend file name
imgFilenameNoExt+"-t"+t+"-l"+l+"-h"+h+"-w"+w+"."+imgType
----------
Original JIRA:
http://issues.craftercms.org/browse/CRAFTERCMS-2185
---------- | 1.0 | CRAFTERCMS-2185: Cropping an image should create a new image with a name that indicates the crop details - Original JIRA Fix Versions:
2.5.2, Original JIRA Components:
Studio,
----------
Original JIRA Description: :
----------
Original JIRA Comments:
Update Crop Service code to create a custom filename if supplied
Update dialog to use custom name, if name exists offer to rename (but pre-populate field with value)
=======================
import java.awt.image.BufferedImage
import java.io.ByteArrayOutputStream
import java.io.ByteArrayInputStream
import java.io.File
import scripts.api.ContentServices
import javax.imageio.ImageIO
def req = request
def site = params.site
def imgPath = params.path
def newName = params.newname
def t = params.t.toInteger()
def l = params.l.toInteger()
def w = params.w.toInteger()
def h = params.h.toInteger()
def newName = params.newFileName
def imgToCrop = null
def imgCropped = null
def imgCroppedOutStream = null
def imgCroppedInStream = null
def imgType = imgPath.substring(imgPath.indexOf(".")+1)
def imgPathOnly = imgPath.substring(0, imgPath.lastIndexOf("/"))
def imgFilename = imgPath.substring(imgPath.lastIndexOf("/")+1)
def newImgFilename =(newFileName!=null) ? newFileName: imgFilename
if (newName) {
imgFilename = newName;
}
def context = ContentServices.createContext(applicationContext, request)
imgToCrop = ImageIO.read(ContentServices.getContentAsStream(site, imgPath, context))
imgCropped = imgToCrop.getSubimage(l, t, w, h)
imgCroppedOutStream = new ByteArrayOutputStream()
ImageIO.write(imgCropped, imgType, imgCroppedOutStream)
imgCroppedInStream = new ByteArrayInputStream(imgCroppedOutStream.toByteArray())
def result = ContentServices.writeContentAsset(context, site, imgPathOnly, newImgFilename, imgCroppedInStream, "true", "", "", "", "false", "true", null);
return result
----------
UI should recommend file name
imgFilenameNoExt+"-t"+t+"-l"+l+"-h"+h+"-w"+w+"."+imgType
----------
Original JIRA:
http://issues.craftercms.org/browse/CRAFTERCMS-2185
---------- | priority | craftercms cropping an image should create a new image with a name that indicates the crop details original jira fix versions original jira components studio original jira description original jira comments update crop service code to create a custom filename if supplied update dialog to use custom name if name exists offer to rename but pre populate field with value import java awt image bufferedimage import java io bytearrayoutputstream import java io bytearrayinputstream import java io file import scripts api contentservices import javax imageio imageio def req request def site params site def imgpath params path def newname params newname def t params t tointeger def l params l tointeger def w params w tointeger def h params h tointeger def newname params newfilename def imgtocrop null def imgcropped null def imgcroppedoutstream null def imgcroppedinstream null def imgtype imgpath substring imgpath indexof def imgpathonly imgpath substring imgpath lastindexof def imgfilename imgpath substring imgpath lastindexof def newimgfilename newfilename null newfilename imgfilename if newname imgfilename newname def context contentservices createcontext applicationcontext request imgtocrop imageio read contentservices getcontentasstream site imgpath context imgcropped imgtocrop getsubimage l t w h imgcroppedoutstream new bytearrayoutputstream imageio write imgcropped imgtype imgcroppedoutstream imgcroppedinstream new bytearrayinputstream imgcroppedoutstream tobytearray def result contentservices writecontentasset context site imgpathonly newimgfilename imgcroppedinstream true false true null return result ui should recommend file name imgfilenamenoext t t l l h h w w imgtype original jira | 1 |
752,170 | 26,275,853,657 | IssuesEvent | 2023-01-06 21:55:26 | Team-Betise/Lunchline | https://api.github.com/repos/Team-Betise/Lunchline | reopened | Table for menu | enhancement high-priority | Table to store the items currently in menu
- ItemName
- ItemCost
- Desc
- Rating
- CurrentAvailability (Set automatically according to Availability, can be changed by vendor)
- AvailabilityTimes (Array with start and end times) | 1.0 | Table for menu - Table to store the items currently in menu
- ItemName
- ItemCost
- Desc
- Rating
- CurrentAvailability (Set automatically according to Availability, can be changed by vendor)
- AvailabilityTimes (Array with start and end times) | priority | table for menu table to store the items currently in menu itemname itemcost desc rating currentavailability set automatically according to availability can be changed by vendor availabilitytimes array with start and end times | 1 |
145,595 | 5,578,391,441 | IssuesEvent | 2017-03-28 12:18:00 | w3c/wai-people-use-web | https://api.github.com/repos/w3c/wai-people-use-web | closed | [Principles] update references | High Priority | From Judy Brewer:
> The references to UAAG 1.0 and ATAG 1.0 are seriously outdated and should be updated. Additionally, in general comments on this page it states that WAI develops UAAG and ATAG. That also needs to be updated. | 1.0 | [Principles] update references - From Judy Brewer:
> The references to UAAG 1.0 and ATAG 1.0 are seriously outdated and should be updated. Additionally, in general comments on this page it states that WAI develops UAAG and ATAG. That also needs to be updated. | priority | update references from judy brewer the references to uaag and atag are seriously outdated and should be updated additionally in general comments on this page it states that wai develops uaag and atag that also needs to be updated | 1 |
544,983 | 15,933,216,389 | IssuesEvent | 2021-04-14 07:08:22 | rootless-containers/rootlesskit | https://api.github.com/repos/rootless-containers/rootlesskit | opened | [regression in Docker 20.10.6] slirp4netns port driver fails: "Timed out proxy starting the userland proxy." | bug priority/high | Rootless Docker 20.10.6 + RootlessKit v0.14.1 + slirp4netns port driver fails
```console
$ cat ~/.config/systemd/user/docker.service.d/override.conf
[Service]
Environment=DOCKERD_ROOTLESS_ROOTLESSKIT_PORT_DRIVER="slirp4netns"
$ docker --context=rootless run --rm -p 8080:80 nginx:alpine
docker: Error response from daemon: driver failed programming external connectivity on endpoint dreamy_gauss (93092ec62bc18d4190f18c05f188505c11acce4feef7ea2eb259805c18edfd85): Timed out proxy starting the userland proxy.
```
builtin port driver works.
The both drivers were working with Docker v20.10.5.
| 1.0 | [regression in Docker 20.10.6] slirp4netns port driver fails: "Timed out proxy starting the userland proxy." - Rootless Docker 20.10.6 + RootlessKit v0.14.1 + slirp4netns port driver fails
```console
$ cat ~/.config/systemd/user/docker.service.d/override.conf
[Service]
Environment=DOCKERD_ROOTLESS_ROOTLESSKIT_PORT_DRIVER="slirp4netns"
$ docker --context=rootless run --rm -p 8080:80 nginx:alpine
docker: Error response from daemon: driver failed programming external connectivity on endpoint dreamy_gauss (93092ec62bc18d4190f18c05f188505c11acce4feef7ea2eb259805c18edfd85): Timed out proxy starting the userland proxy.
```
builtin port driver works.
The both drivers were working with Docker v20.10.5.
| priority | port driver fails timed out proxy starting the userland proxy rootless docker rootlesskit port driver fails console cat config systemd user docker service d override conf environment dockerd rootless rootlesskit port driver docker context rootless run rm p nginx alpine docker error response from daemon driver failed programming external connectivity on endpoint dreamy gauss timed out proxy starting the userland proxy builtin port driver works the both drivers were working with docker | 1 |
211,186 | 7,199,048,842 | IssuesEvent | 2018-02-05 14:51:08 | wso2/cloudformation-apim | https://api.github.com/repos/wso2/cloudformation-apim | opened | Current Egress policy is set to allowAll for 9763/9443/8243 ports | Priority/High Severity/Major Type/Bug | **Description:**
ATM, AWS::EC2::SecurityGroup Egress policy in cloudformation-apim scripts [1] is set to allowAll (0.0.0.0/0 CIDR block) for 9763/9443/8243 ports.
Since this allows complete internet access via these ports, we need to restrict it, right?

[1] https://github.com/wso2/cloudformation-apim/blob/master/patterns/pattern-1/pattern-1-cloudformation.template.yaml#L815
**Suggested Labels:**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees:**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
**Affected Product Version:**
**OS, DB, other environment details and versions:**
**Steps to reproduce:**
**Related Issues:**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. --> | 1.0 | Current Egress policy is set to allowAll for 9763/9443/8243 ports - **Description:**
ATM, AWS::EC2::SecurityGroup Egress policy in cloudformation-apim scripts [1] is set to allowAll (0.0.0.0/0 CIDR block) for 9763/9443/8243 ports.
Since this allows complete internet access via these ports, we need to restrict it, right?

[1] https://github.com/wso2/cloudformation-apim/blob/master/patterns/pattern-1/pattern-1-cloudformation.template.yaml#L815
**Suggested Labels:**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees:**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
**Affected Product Version:**
**OS, DB, other environment details and versions:**
**Steps to reproduce:**
**Related Issues:**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. --> | priority | current egress policy is set to allowall for ports description atm aws securitygroup egress policy in cloudformation apim scripts is set to allowall cidr block for ports since this allows complete internet access via these ports we need to restrict it right suggested labels suggested assignees affected product version os db other environment details and versions steps to reproduce related issues | 1 |
810,111 | 30,225,627,844 | IssuesEvent | 2023-07-06 00:04:02 | shieldworks/aegis | https://api.github.com/repos/shieldworks/aegis | closed | Ability to swap age encryption with a FIPS-compliant alternative | priority:high | Age encryption is not FIPS-complinat.
If that part is turned into a plugin, then we can offer different implementations at various compliancy levels. | 1.0 | Ability to swap age encryption with a FIPS-compliant alternative - Age encryption is not FIPS-complinat.
If that part is turned into a plugin, then we can offer different implementations at various compliancy levels. | priority | ability to swap age encryption with a fips compliant alternative age encryption is not fips complinat if that part is turned into a plugin then we can offer different implementations at various compliancy levels | 1 |
175,821 | 6,554,270,275 | IssuesEvent | 2017-09-06 04:35:11 | wso2/product-ei | https://api.github.com/repos/wso2/product-ei | opened | Changing port offset in deployment.yaml does not change the broker bind addresses | Component/Broker Priority/High Severity/Major Type/Bug | ### Steps tp reproduce
1. Unzip wso2ei-7.0.0-SNAPSHOT.zip to instance-1 and instance-2
1. Change the `wso2.carbon.offset` to 1 in instance-2
Following exception was observed even after changing the `wso2.carbon.offset` to 1 in instance-2.
```
[2017-09-06 09:37:38,155] ERROR {org.wso2.andes.server.Main} - Exception during startup. Triggering shutdown org.wso2.andes.kernel.AndesException: Unable to initialise application registry
at org.wso2.andes.server.Broker.startupImpl(Broker.java:307)
at org.wso2.andes.server.Broker.startup(Broker.java:110)
at org.wso2.andes.server.Main.startBroker(Main.java:217)
at org.wso2.andes.server.Main.execute(Main.java:206)
at org.wso2.andes.server.Main.<init>(Main.java:54)
at org.wso2.andes.server.Main.main(Main.java:47)
at org.wso2.carbon.business.messaging.core.internal.BrokerServiceComponent.startAndesBroker(BrokerServiceComponent.java:280)
at org.wso2.carbon.business.messaging.core.internal.BrokerServiceComponent.start(BrokerServiceComponent.java:142)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.eclipse.equinox.internal.ds.model.ServiceComponent.activate(ServiceComponent.java:235)
at org.eclipse.equinox.internal.ds.model.ServiceComponentProp.activate(ServiceComponentProp.java:146)
at org.eclipse.equinox.internal.ds.model.ServiceComponentProp.build(ServiceComponentProp.java:345)
at org.eclipse.equinox.internal.ds.InstanceProcess.buildComponent(InstanceProcess.java:620)
at org.eclipse.equinox.internal.ds.InstanceProcess.buildComponents(InstanceProcess.java:197)
at org.eclipse.equinox.internal.ds.Resolver.getEligible(Resolver.java:343)
at org.eclipse.equinox.internal.ds.SCRManager.serviceChanged(SCRManager.java:222)
at org.eclipse.osgi.internal.serviceregistry.FilteredServiceListener.serviceChanged(FilteredServiceListener.java:109)
at org.eclipse.osgi.internal.framework.BundleContextImpl.dispatchEvent(BundleContextImpl.java:915)
at org.eclipse.osgi.framework.eventmgr.EventManager.dispatchEvent(EventManager.java:230)
at org.eclipse.osgi.framework.eventmgr.ListenerQueue.dispatchEventSynchronous(ListenerQueue.java:148)
at org.eclipse.osgi.internal.serviceregistry.ServiceRegistry.publishServiceEventPrivileged(ServiceRegistry.java:862)
at org.eclipse.osgi.internal.serviceregistry.ServiceRegistry.publishServiceEvent(ServiceRegistry.java:801)
at org.eclipse.osgi.internal.serviceregistry.ServiceRegistrationImpl.register(ServiceRegistrationImpl.java:127)
at org.eclipse.osgi.internal.serviceregistry.ServiceRegistry.registerService(ServiceRegistry.java:225)
at org.eclipse.osgi.internal.framework.BundleContextImpl.registerService(BundleContextImpl.java:464)
at org.eclipse.osgi.internal.framework.BundleContextImpl.registerService(BundleContextImpl.java:482)
at org.eclipse.osgi.internal.framework.BundleContextImpl.registerService(BundleContextImpl.java:999)
at org.wso2.carbon.datasource.core.internal.DataSourceListenerComponent.onAllRequiredCapabilitiesAvailable(DataSourceListenerComponent.java:101)
at org.wso2.carbon.kernel.internal.startupresolver.StartupComponentManager.lambda$notifySatisfiableComponents$51(StartupComponentManager.java:238)
at java.util.ArrayList.forEach(ArrayList.java:1249)
at org.wso2.carbon.kernel.internal.startupresolver.StartupComponentManager.notifySatisfiableComponents(StartupComponentManager.java:224)
at org.wso2.carbon.kernel.internal.startupresolver.StartupOrderResolver$1.run(StartupOrderResolver.java:204)
at java.util.TimerThread.mainLoop(Timer.java:555)
at java.util.TimerThread.run(Timer.java:505)
Caused by: java.rmi.server.ExportException: Port already in use: 8999; nested exception is:
java.net.BindException: Address already in use (Bind failed)
at sun.rmi.transport.tcp.TCPTransport.listen(TCPTransport.java:341)
at sun.rmi.transport.tcp.TCPTransport.exportObject(TCPTransport.java:249)
at sun.rmi.transport.tcp.TCPEndpoint.exportObject(TCPEndpoint.java:411)
at sun.rmi.transport.LiveRef.exportObject(LiveRef.java:147)
at sun.rmi.server.UnicastServerRef.exportObject(UnicastServerRef.java:236)
at sun.rmi.registry.RegistryImpl.setup(RegistryImpl.java:213)
at sun.rmi.registry.RegistryImpl.<init>(RegistryImpl.java:173)
at sun.rmi.registry.RegistryImpl.<init>(RegistryImpl.java:144)
at java.rmi.registry.LocateRegistry.createRegistry(LocateRegistry.java:239)
at org.wso2.andes.server.management.JMXManagedObjectRegistry.start(JMXManagedObjectRegistry.java:205)
at org.wso2.andes.server.registry.ApplicationRegistry.initialise(ApplicationRegistry.java:251)
at org.wso2.andes.server.registry.ApplicationRegistry.initialise(ApplicationRegistry.java:147)
at org.wso2.andes.server.Broker.startupImpl(Broker.java:274)
... 36 more
Caused by: java.net.BindException: Address already in use (Bind failed)
at java.net.PlainSocketImpl.socketBind(Native Method)
at java.net.AbstractPlainSocketImpl.bind(AbstractPlainSocketImpl.java:387)
at java.net.ServerSocket.bind(ServerSocket.java:375)
at java.net.ServerSocket.<init>(ServerSocket.java:237)
at java.net.ServerSocket.<init>(ServerSocket.java:128)
at org.wso2.andes.server.management.JMXManagedObjectRegistry$CustomRMIServerSocketFactory$NoLocalAddressServerSocket.<init>(JMXManagedObjectRegistry.java:332)
at org.wso2.andes.server.management.JMXManagedObjectRegistry$CustomRMIServerSocketFactory.createServerSocket(JMXManagedObjectRegistry.java:325)
at sun.rmi.transport.tcp.TCPEndpoint.newServerSocket(TCPEndpoint.java:666)
at sun.rmi.transport.tcp.TCPTransport.listen(TCPTransport.java:330)
... 48 more
``` | 1.0 | Changing port offset in deployment.yaml does not change the broker bind addresses - ### Steps tp reproduce
1. Unzip wso2ei-7.0.0-SNAPSHOT.zip to instance-1 and instance-2
1. Change the `wso2.carbon.offset` to 1 in instance-2
Following exception was observed even after changing the `wso2.carbon.offset` to 1 in instance-2.
```
[2017-09-06 09:37:38,155] ERROR {org.wso2.andes.server.Main} - Exception during startup. Triggering shutdown org.wso2.andes.kernel.AndesException: Unable to initialise application registry
at org.wso2.andes.server.Broker.startupImpl(Broker.java:307)
at org.wso2.andes.server.Broker.startup(Broker.java:110)
at org.wso2.andes.server.Main.startBroker(Main.java:217)
at org.wso2.andes.server.Main.execute(Main.java:206)
at org.wso2.andes.server.Main.<init>(Main.java:54)
at org.wso2.andes.server.Main.main(Main.java:47)
at org.wso2.carbon.business.messaging.core.internal.BrokerServiceComponent.startAndesBroker(BrokerServiceComponent.java:280)
at org.wso2.carbon.business.messaging.core.internal.BrokerServiceComponent.start(BrokerServiceComponent.java:142)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.eclipse.equinox.internal.ds.model.ServiceComponent.activate(ServiceComponent.java:235)
at org.eclipse.equinox.internal.ds.model.ServiceComponentProp.activate(ServiceComponentProp.java:146)
at org.eclipse.equinox.internal.ds.model.ServiceComponentProp.build(ServiceComponentProp.java:345)
at org.eclipse.equinox.internal.ds.InstanceProcess.buildComponent(InstanceProcess.java:620)
at org.eclipse.equinox.internal.ds.InstanceProcess.buildComponents(InstanceProcess.java:197)
at org.eclipse.equinox.internal.ds.Resolver.getEligible(Resolver.java:343)
at org.eclipse.equinox.internal.ds.SCRManager.serviceChanged(SCRManager.java:222)
at org.eclipse.osgi.internal.serviceregistry.FilteredServiceListener.serviceChanged(FilteredServiceListener.java:109)
at org.eclipse.osgi.internal.framework.BundleContextImpl.dispatchEvent(BundleContextImpl.java:915)
at org.eclipse.osgi.framework.eventmgr.EventManager.dispatchEvent(EventManager.java:230)
at org.eclipse.osgi.framework.eventmgr.ListenerQueue.dispatchEventSynchronous(ListenerQueue.java:148)
at org.eclipse.osgi.internal.serviceregistry.ServiceRegistry.publishServiceEventPrivileged(ServiceRegistry.java:862)
at org.eclipse.osgi.internal.serviceregistry.ServiceRegistry.publishServiceEvent(ServiceRegistry.java:801)
at org.eclipse.osgi.internal.serviceregistry.ServiceRegistrationImpl.register(ServiceRegistrationImpl.java:127)
at org.eclipse.osgi.internal.serviceregistry.ServiceRegistry.registerService(ServiceRegistry.java:225)
at org.eclipse.osgi.internal.framework.BundleContextImpl.registerService(BundleContextImpl.java:464)
at org.eclipse.osgi.internal.framework.BundleContextImpl.registerService(BundleContextImpl.java:482)
at org.eclipse.osgi.internal.framework.BundleContextImpl.registerService(BundleContextImpl.java:999)
at org.wso2.carbon.datasource.core.internal.DataSourceListenerComponent.onAllRequiredCapabilitiesAvailable(DataSourceListenerComponent.java:101)
at org.wso2.carbon.kernel.internal.startupresolver.StartupComponentManager.lambda$notifySatisfiableComponents$51(StartupComponentManager.java:238)
at java.util.ArrayList.forEach(ArrayList.java:1249)
at org.wso2.carbon.kernel.internal.startupresolver.StartupComponentManager.notifySatisfiableComponents(StartupComponentManager.java:224)
at org.wso2.carbon.kernel.internal.startupresolver.StartupOrderResolver$1.run(StartupOrderResolver.java:204)
at java.util.TimerThread.mainLoop(Timer.java:555)
at java.util.TimerThread.run(Timer.java:505)
Caused by: java.rmi.server.ExportException: Port already in use: 8999; nested exception is:
java.net.BindException: Address already in use (Bind failed)
at sun.rmi.transport.tcp.TCPTransport.listen(TCPTransport.java:341)
at sun.rmi.transport.tcp.TCPTransport.exportObject(TCPTransport.java:249)
at sun.rmi.transport.tcp.TCPEndpoint.exportObject(TCPEndpoint.java:411)
at sun.rmi.transport.LiveRef.exportObject(LiveRef.java:147)
at sun.rmi.server.UnicastServerRef.exportObject(UnicastServerRef.java:236)
at sun.rmi.registry.RegistryImpl.setup(RegistryImpl.java:213)
at sun.rmi.registry.RegistryImpl.<init>(RegistryImpl.java:173)
at sun.rmi.registry.RegistryImpl.<init>(RegistryImpl.java:144)
at java.rmi.registry.LocateRegistry.createRegistry(LocateRegistry.java:239)
at org.wso2.andes.server.management.JMXManagedObjectRegistry.start(JMXManagedObjectRegistry.java:205)
at org.wso2.andes.server.registry.ApplicationRegistry.initialise(ApplicationRegistry.java:251)
at org.wso2.andes.server.registry.ApplicationRegistry.initialise(ApplicationRegistry.java:147)
at org.wso2.andes.server.Broker.startupImpl(Broker.java:274)
... 36 more
Caused by: java.net.BindException: Address already in use (Bind failed)
at java.net.PlainSocketImpl.socketBind(Native Method)
at java.net.AbstractPlainSocketImpl.bind(AbstractPlainSocketImpl.java:387)
at java.net.ServerSocket.bind(ServerSocket.java:375)
at java.net.ServerSocket.<init>(ServerSocket.java:237)
at java.net.ServerSocket.<init>(ServerSocket.java:128)
at org.wso2.andes.server.management.JMXManagedObjectRegistry$CustomRMIServerSocketFactory$NoLocalAddressServerSocket.<init>(JMXManagedObjectRegistry.java:332)
at org.wso2.andes.server.management.JMXManagedObjectRegistry$CustomRMIServerSocketFactory.createServerSocket(JMXManagedObjectRegistry.java:325)
at sun.rmi.transport.tcp.TCPEndpoint.newServerSocket(TCPEndpoint.java:666)
at sun.rmi.transport.tcp.TCPTransport.listen(TCPTransport.java:330)
... 48 more
``` | priority | changing port offset in deployment yaml does not change the broker bind addresses steps tp reproduce unzip snapshot zip to instance and instance change the carbon offset to in instance following exception was observed even after changing the carbon offset to in instance error org andes server main exception during startup triggering shutdown org andes kernel andesexception unable to initialise application registry at org andes server broker startupimpl broker java at org andes server broker startup broker java at org andes server main startbroker main java at org andes server main execute main java at org andes server main main java at org andes server main main main java at org carbon business messaging core internal brokerservicecomponent startandesbroker brokerservicecomponent java at org carbon business messaging core internal brokerservicecomponent start brokerservicecomponent java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org eclipse equinox internal ds model servicecomponent activate servicecomponent java at org eclipse equinox internal ds model servicecomponentprop activate servicecomponentprop java at org eclipse equinox internal ds model servicecomponentprop build servicecomponentprop java at org eclipse equinox internal ds instanceprocess buildcomponent instanceprocess java at org eclipse equinox internal ds instanceprocess buildcomponents instanceprocess java at org eclipse equinox internal ds resolver geteligible resolver java at org eclipse equinox internal ds scrmanager servicechanged scrmanager java at org eclipse osgi internal serviceregistry filteredservicelistener servicechanged filteredservicelistener java at org eclipse osgi internal framework bundlecontextimpl dispatchevent bundlecontextimpl java at org eclipse osgi framework eventmgr eventmanager dispatchevent eventmanager java at org eclipse osgi framework eventmgr listenerqueue dispatcheventsynchronous listenerqueue java at org eclipse osgi internal serviceregistry serviceregistry publishserviceeventprivileged serviceregistry java at org eclipse osgi internal serviceregistry serviceregistry publishserviceevent serviceregistry java at org eclipse osgi internal serviceregistry serviceregistrationimpl register serviceregistrationimpl java at org eclipse osgi internal serviceregistry serviceregistry registerservice serviceregistry java at org eclipse osgi internal framework bundlecontextimpl registerservice bundlecontextimpl java at org eclipse osgi internal framework bundlecontextimpl registerservice bundlecontextimpl java at org eclipse osgi internal framework bundlecontextimpl registerservice bundlecontextimpl java at org carbon datasource core internal datasourcelistenercomponent onallrequiredcapabilitiesavailable datasourcelistenercomponent java at org carbon kernel internal startupresolver startupcomponentmanager lambda notifysatisfiablecomponents startupcomponentmanager java at java util arraylist foreach arraylist java at org carbon kernel internal startupresolver startupcomponentmanager notifysatisfiablecomponents startupcomponentmanager java at org carbon kernel internal startupresolver startuporderresolver run startuporderresolver java at java util timerthread mainloop timer java at java util timerthread run timer java caused by java rmi server exportexception port already in use nested exception is java net bindexception address already in use bind failed at sun rmi transport tcp tcptransport listen tcptransport java at sun rmi transport tcp tcptransport exportobject tcptransport java at sun rmi transport tcp tcpendpoint exportobject tcpendpoint java at sun rmi transport liveref exportobject liveref java at sun rmi server unicastserverref exportobject unicastserverref java at sun rmi registry registryimpl setup registryimpl java at sun rmi registry registryimpl registryimpl java at sun rmi registry registryimpl registryimpl java at java rmi registry locateregistry createregistry locateregistry java at org andes server management jmxmanagedobjectregistry start jmxmanagedobjectregistry java at org andes server registry applicationregistry initialise applicationregistry java at org andes server registry applicationregistry initialise applicationregistry java at org andes server broker startupimpl broker java more caused by java net bindexception address already in use bind failed at java net plainsocketimpl socketbind native method at java net abstractplainsocketimpl bind abstractplainsocketimpl java at java net serversocket bind serversocket java at java net serversocket serversocket java at java net serversocket serversocket java at org andes server management jmxmanagedobjectregistry customrmiserversocketfactory nolocaladdressserversocket jmxmanagedobjectregistry java at org andes server management jmxmanagedobjectregistry customrmiserversocketfactory createserversocket jmxmanagedobjectregistry java at sun rmi transport tcp tcpendpoint newserversocket tcpendpoint java at sun rmi transport tcp tcptransport listen tcptransport java more | 1 |
679,041 | 23,219,578,025 | IssuesEvent | 2022-08-02 16:51:34 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | closed | `torch.nn.functional.gumbel_softmax` returns different results on cpu and on cuda with the same input | high priority module: nn module: cuda triaged module: correctness (silent) | ### 🐛 Describe the bug
`torch.nn.functional.gumbel_softmax` returns different results on cpu and on cuda with the same input
```
import torch
torch.random.manual_seed(420)
input = torch.randn(2,3,requires_grad=True)
res_cpu = torch.nn.functional.gumbel_softmax(input, hard=True)
print("res_cpu: ", res_cpu)
input2 = input.clone().detach().to('cuda')
res_gpu = torch.nn.functional.gumbel_softmax(input2, hard=True)
print("res_gpu: ", res_gpu)
```
Outputs:
```
res_cpu: tensor([[0., 1., 0.],
[1., 0., 0.]], grad_fn=<AddBackward0>)
res_gpu: tensor([[0., 1., 0.],
[0., 1., 0.]], device='cuda:0')
```
### Versions
pytorch: 1.12.0
cc @ezyang @gchanan @zou3519 @albanD @mruberry @jbschlosser @walterddr @kshitij12345 @saketh-are @ngimel | 1.0 | `torch.nn.functional.gumbel_softmax` returns different results on cpu and on cuda with the same input - ### 🐛 Describe the bug
`torch.nn.functional.gumbel_softmax` returns different results on cpu and on cuda with the same input
```
import torch
torch.random.manual_seed(420)
input = torch.randn(2,3,requires_grad=True)
res_cpu = torch.nn.functional.gumbel_softmax(input, hard=True)
print("res_cpu: ", res_cpu)
input2 = input.clone().detach().to('cuda')
res_gpu = torch.nn.functional.gumbel_softmax(input2, hard=True)
print("res_gpu: ", res_gpu)
```
Outputs:
```
res_cpu: tensor([[0., 1., 0.],
[1., 0., 0.]], grad_fn=<AddBackward0>)
res_gpu: tensor([[0., 1., 0.],
[0., 1., 0.]], device='cuda:0')
```
### Versions
pytorch: 1.12.0
cc @ezyang @gchanan @zou3519 @albanD @mruberry @jbschlosser @walterddr @kshitij12345 @saketh-are @ngimel | priority | torch nn functional gumbel softmax returns different results on cpu and on cuda with the same input 🐛 describe the bug torch nn functional gumbel softmax returns different results on cpu and on cuda with the same input import torch torch random manual seed input torch randn requires grad true res cpu torch nn functional gumbel softmax input hard true print res cpu res cpu input clone detach to cuda res gpu torch nn functional gumbel softmax hard true print res gpu res gpu outputs res cpu tensor grad fn res gpu tensor device cuda versions pytorch cc ezyang gchanan alband mruberry jbschlosser walterddr saketh are ngimel | 1 |
450,795 | 13,019,395,198 | IssuesEvent | 2020-07-26 22:21:00 | marcus67/little_brother | https://api.github.com/repos/marcus67/little_brother | closed | Same user appears several times in admin view | bug high priority | Apparently, for each change made to the user configuration, that user is duplicated in the administration view. After restart everything is back to normal which means that the user is not actually stored multiple times but this bug seems to be related to caching. | 1.0 | Same user appears several times in admin view - Apparently, for each change made to the user configuration, that user is duplicated in the administration view. After restart everything is back to normal which means that the user is not actually stored multiple times but this bug seems to be related to caching. | priority | same user appears several times in admin view apparently for each change made to the user configuration that user is duplicated in the administration view after restart everything is back to normal which means that the user is not actually stored multiple times but this bug seems to be related to caching | 1 |
635,250 | 20,382,533,833 | IssuesEvent | 2022-02-22 00:43:25 | Sage-Bionetworks/projectLive_NF | https://api.github.com/repos/Sage-Bionetworks/projectLive_NF | closed | Populate plots in `data focus` and `study timeline` if data is uploaded to a project but not annotated | Priority:High | **Is your feature request related to a problem? Please describe.**
The sections `Study Timeline` and `Data Focus` do not show any plots if the project has files but they are not annotated properly.
**Describe the solution you'd like**
Populate the sections mentioned above with plots but showing that files are `Pending Annotation`. `Study Timeline` section will mirror the current `Annotation Activity` Section.
**Describe alternatives you've considered**
1. A clear and concise message to the user saying that there are files present in the project but they have not been annotated.
**Additional context**
Add any other context or screenshots about the feature request here.
| 1.0 | Populate plots in `data focus` and `study timeline` if data is uploaded to a project but not annotated - **Is your feature request related to a problem? Please describe.**
The sections `Study Timeline` and `Data Focus` do not show any plots if the project has files but they are not annotated properly.
**Describe the solution you'd like**
Populate the sections mentioned above with plots but showing that files are `Pending Annotation`. `Study Timeline` section will mirror the current `Annotation Activity` Section.
**Describe alternatives you've considered**
1. A clear and concise message to the user saying that there are files present in the project but they have not been annotated.
**Additional context**
Add any other context or screenshots about the feature request here.
| priority | populate plots in data focus and study timeline if data is uploaded to a project but not annotated is your feature request related to a problem please describe the sections study timeline and data focus do not show any plots if the project has files but they are not annotated properly describe the solution you d like populate the sections mentioned above with plots but showing that files are pending annotation study timeline section will mirror the current annotation activity section describe alternatives you ve considered a clear and concise message to the user saying that there are files present in the project but they have not been annotated additional context add any other context or screenshots about the feature request here | 1 |
788,039 | 27,741,406,607 | IssuesEvent | 2023-03-15 14:30:39 | renovatebot/renovate | https://api.github.com/repos/renovatebot/renovate | closed | ExecError updating package (yarn up may need quotes?) | type:bug priority-2-high manager:npm status:in-progress | ### How are you running Renovate?
Mend Renovate hosted app on github.com
### Was this something which used to work for you, and then stopped?
I never saw this working
### Describe the bug
Renovate won't update this package: https://github.com/recitalsoftware/yarn3/blob/main/package.json#L5. Appears to only happen when specifying range strategy "in-range-only" (didn't test other range strategies, just the absence of that, so whatever the default would be). [Yes, there's a minimal repro :)](https://github.com/recitalsoftware/yarn3)
[Looking at the logs](https://app.renovatebot.com/dashboard#github/recitalsoftware/yarn3/1021810149) (included below), the problem is that the upgrade command is returning an ExecError, with an exit code of 1. There's no output to suggest what part of the command failed.
This may be a red herring (especially since I'm using zsh, not bash), but:
I've tried to run the same command locally (just the part sidecar runs) and it works fine. I do, however, need to add quotes around the package version:
```
yarn install --mode=update-lockfile && yarn up @vue/apollo-composable@>=4.0.0-beta.1 --mode=update-lockfile ~/dev/oss/repros/yarn3
➤ YN0000: ┌ Resolution step
... snip ...
➤ YN0000: Done with warnings in 0s 52ms
zsh: 4.0.0-beta.1 not found
```
Adding quotes around the package in the upgrade command fixes that problem: `yarn up "@vue/apollo-composable@>=4.0.0-beta.1" --mode=update-lockfile`
### Relevant debug logs
<details><summary>Logs</summary>
```
DEBUG: Executing command(branch="renovate/vue-apollo-composable-4.x-lockfile")
{
"command": "docker run --rm --name=renovate_sidecar --label=renovate_child -v \"/mnt/renovate/gh/recitalsoftware/yarn3\":\"/mnt/renovate/gh/recitalsoftware/yarn3\" -v \"/tmp/renovate-cache\":\"/tmp/renovate-cache\" -v \"/tmp/containerbase\":\"/tmp/containerbase\" -e NPM_CONFIG_CACHE -e npm_config_store -e CI -e YARN_ENABLE_IMMUTABLE_INSTALLS -e YARN_HTTP_TIMEOUT -e YARN_GLOBAL_FOLDER -e YARN_ENABLE_GLOBAL_CACHE -e BUILDPACK_CACHE_DIR -e CONTAINERBASE_CACHE_DIR -w \"/mnt/renovate/gh/recitalsoftware/yarn3\" docker.io/renovate/sidecar bash -l -c \"install-tool node v18.15.0 && install-tool corepack 0.17.0 && yarn install --mode=update-lockfile && yarn up @vue/apollo-composable@>=4.0.0-beta.1 --mode=update-lockfile\""
}
DEBUG: rawExec err(branch="renovate/vue-apollo-composable-4.x-lockfile")
{
"err": {
"name": "ExecError",
"cmd": "/bin/sh -c docker run --rm --name=renovate_sidecar --label=renovate_child -v \"/mnt/renovate/gh/recitalsoftware/yarn3\":\"/mnt/renovate/gh/recitalsoftware/yarn3\" -v \"/tmp/renovate-cache\":\"/tmp/renovate-cache\" -v \"/tmp/containerbase\":\"/tmp/containerbase\" -e NPM_CONFIG_CACHE -e npm_config_store -e CI -e YARN_ENABLE_IMMUTABLE_INSTALLS -e YARN_HTTP_TIMEOUT -e YARN_GLOBAL_FOLDER -e YARN_ENABLE_GLOBAL_CACHE -e BUILDPACK_CACHE_DIR -e CONTAINERBASE_CACHE_DIR -w \"/mnt/renovate/gh/recitalsoftware/yarn3\" docker.io/renovate/sidecar bash -l -c \"install-tool node v18.15.0 && install-tool corepack 0.17.0 && yarn install --mode=update-lockfile && yarn up @vue/apollo-composable@>=4.0.0-beta.1 --mode=update-lockfile\"",
"stderr": "",
"stdout": "installing v2 tool node v18.15.0\nlinking tool node v18.15.0\nnode: v18.15.0 /usr/local/bin/node\nnpm: 9.5.0 /usr/local/bin/npm\nInstalled v2 /usr/local/buildpack/tools/v2/node.sh in 8 seconds\nskip cleanup, not a docker build: 37be2fc221c0\ninstalling v2 tool corepack v0.17.0\nlinking tool corepack v0.17.0\n0.17.0\nInstalled v2 /usr/local/buildpack/tools/v2/corepack.sh in 4 seconds\nskip cleanup, not a docker build: 37be2fc221c0\n➤ YN0000: ┌ Resolution step\n➤ YN0002: │ @vue/apollo-composable@npm:4.0.0-beta.3 [dd12c] doesn't provide typescript (pae615), requested by ts-essentials\n➤ YN0002: │ yarn3@workspace:. doesn't provide @apollo/client (pf3d3e), requested by @vue/apollo-composable\n➤ YN0002: │ yarn3@workspace:. doesn't provide graphql (pe1da9), requested by @vue/apollo-composable\n➤ YN0002: │ yarn3@workspace:. doesn't provide vue (p36305), requested by @vue/apollo-composable\n➤ YN0000: │ Some peer dependencies are incorrectly met; run yarn explain peer-requirements <hash> for details, where <hash> is the six-letter p-prefixed code\n➤ YN0000: └ Completed\n➤ YN0000: ┌ Fetch step\n➤ YN0000: └ Completed\n➤ YN0000: ┌ Link step\n➤ YN0073: │ Skipped due to mode=update-lockfile\n➤ YN0000: └ Completed\n➤ YN0000: Done with warnings in 0s 114ms\n",
"options": {
"cwd": "/mnt/renovate/gh/recitalsoftware/yarn3",
"encoding": "utf-8",
"env": {
"NPM_CONFIG_CACHE": "/tmp/renovate-cache/others/npm",
"npm_config_store": "/tmp/renovate-cache/others/pnpm",
"CI": "true",
"YARN_ENABLE_IMMUTABLE_INSTALLS": "false",
"YARN_HTTP_TIMEOUT": "100000",
"YARN_GLOBAL_FOLDER": "/tmp/renovate-cache/others/berry",
"YARN_ENABLE_GLOBAL_CACHE": "1",
"HOME": "/home/ubuntu",
"PATH": "/home/ubuntu/.local/bin:/home/ubuntu/bin:/opt/buildpack/tools/python/3.9.3/bin:/home/ubuntu/.npm-global/bin:/home/ubuntu/renovateapp/node_modules/.bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin",
"LC_ALL": "C.UTF-8",
"LANG": "C.UTF-8",
"BUILDPACK_CACHE_DIR": "/tmp/containerbase",
"CONTAINERBASE_CACHE_DIR": "/tmp/containerbase"
},
"maxBuffer": 10485760,
"timeout": 900000
},
"exitCode": 1,
"message": "Command failed: docker run --rm --name=renovate_sidecar --label=renovate_child -v \"/mnt/renovate/gh/recitalsoftware/yarn3\":\"/mnt/renovate/gh/recitalsoftware/yarn3\" -v \"/tmp/renovate-cache\":\"/tmp/renovate-cache\" -v \"/tmp/containerbase\":\"/tmp/containerbase\" -e NPM_CONFIG_CACHE -e npm_config_store -e CI -e YARN_ENABLE_IMMUTABLE_INSTALLS -e YARN_HTTP_TIMEOUT -e YARN_GLOBAL_FOLDER -e YARN_ENABLE_GLOBAL_CACHE -e BUILDPACK_CACHE_DIR -e CONTAINERBASE_CACHE_DIR -w \"/mnt/renovate/gh/recitalsoftware/yarn3\" docker.io/renovate/sidecar bash -l -c \"install-tool node v18.15.0 && install-tool corepack 0.17.0 && yarn install --mode=update-lockfile && yarn up @vue/apollo-composable@>=4.0.0-beta.1 --mode=update-lockfile\"\n",
"stack": "ExecError: Command failed: docker run --rm --name=renovate_sidecar --label=renovate_child -v \"/mnt/renovate/gh/recitalsoftware/yarn3\":\"/mnt/renovate/gh/recitalsoftware/yarn3\" -v \"/tmp/renovate-cache\":\"/tmp/renovate-cache\" -v \"/tmp/containerbase\":\"/tmp/containerbase\" -e NPM_CONFIG_CACHE -e npm_config_store -e CI -e YARN_ENABLE_IMMUTABLE_INSTALLS -e YARN_HTTP_TIMEOUT -e YARN_GLOBAL_FOLDER -e YARN_ENABLE_GLOBAL_CACHE -e BUILDPACK_CACHE_DIR -e CONTAINERBASE_CACHE_DIR -w \"/mnt/renovate/gh/recitalsoftware/yarn3\" docker.io/renovate/sidecar bash -l -c \"install-tool node v18.15.0 && install-tool corepack 0.17.0 && yarn install --mode=update-lockfile && yarn up @vue/apollo-composable@>=4.0.0-beta.1 --mode=update-lockfile\"\n\n at ChildProcess.<anonymous> (/home/ubuntu/renovateapp/node_modules/renovate/dist/util/exec/common.js:87:24)\n at ChildProcess.emit (node:events:525:35)\n at ChildProcess.emit (node:domain:489:12)\n at Process.ChildProcess._handle.onexit (node:internal/child_process:293:12)"
}
}
```
</details>
### Have you created a minimal reproduction repository?
I have read the minimal reproductions documentation and linked to such a repository in the bug description | 1.0 | ExecError updating package (yarn up may need quotes?) - ### How are you running Renovate?
Mend Renovate hosted app on github.com
### Was this something which used to work for you, and then stopped?
I never saw this working
### Describe the bug
Renovate won't update this package: https://github.com/recitalsoftware/yarn3/blob/main/package.json#L5. Appears to only happen when specifying range strategy "in-range-only" (didn't test other range strategies, just the absence of that, so whatever the default would be). [Yes, there's a minimal repro :)](https://github.com/recitalsoftware/yarn3)
[Looking at the logs](https://app.renovatebot.com/dashboard#github/recitalsoftware/yarn3/1021810149) (included below), the problem is that the upgrade command is returning an ExecError, with an exit code of 1. There's no output to suggest what part of the command failed.
This may be a red herring (especially since I'm using zsh, not bash), but:
I've tried to run the same command locally (just the part sidecar runs) and it works fine. I do, however, need to add quotes around the package version:
```
yarn install --mode=update-lockfile && yarn up @vue/apollo-composable@>=4.0.0-beta.1 --mode=update-lockfile ~/dev/oss/repros/yarn3
➤ YN0000: ┌ Resolution step
... snip ...
➤ YN0000: Done with warnings in 0s 52ms
zsh: 4.0.0-beta.1 not found
```
Adding quotes around the package in the upgrade command fixes that problem: `yarn up "@vue/apollo-composable@>=4.0.0-beta.1" --mode=update-lockfile`
### Relevant debug logs
<details><summary>Logs</summary>
```
DEBUG: Executing command(branch="renovate/vue-apollo-composable-4.x-lockfile")
{
"command": "docker run --rm --name=renovate_sidecar --label=renovate_child -v \"/mnt/renovate/gh/recitalsoftware/yarn3\":\"/mnt/renovate/gh/recitalsoftware/yarn3\" -v \"/tmp/renovate-cache\":\"/tmp/renovate-cache\" -v \"/tmp/containerbase\":\"/tmp/containerbase\" -e NPM_CONFIG_CACHE -e npm_config_store -e CI -e YARN_ENABLE_IMMUTABLE_INSTALLS -e YARN_HTTP_TIMEOUT -e YARN_GLOBAL_FOLDER -e YARN_ENABLE_GLOBAL_CACHE -e BUILDPACK_CACHE_DIR -e CONTAINERBASE_CACHE_DIR -w \"/mnt/renovate/gh/recitalsoftware/yarn3\" docker.io/renovate/sidecar bash -l -c \"install-tool node v18.15.0 && install-tool corepack 0.17.0 && yarn install --mode=update-lockfile && yarn up @vue/apollo-composable@>=4.0.0-beta.1 --mode=update-lockfile\""
}
DEBUG: rawExec err(branch="renovate/vue-apollo-composable-4.x-lockfile")
{
"err": {
"name": "ExecError",
"cmd": "/bin/sh -c docker run --rm --name=renovate_sidecar --label=renovate_child -v \"/mnt/renovate/gh/recitalsoftware/yarn3\":\"/mnt/renovate/gh/recitalsoftware/yarn3\" -v \"/tmp/renovate-cache\":\"/tmp/renovate-cache\" -v \"/tmp/containerbase\":\"/tmp/containerbase\" -e NPM_CONFIG_CACHE -e npm_config_store -e CI -e YARN_ENABLE_IMMUTABLE_INSTALLS -e YARN_HTTP_TIMEOUT -e YARN_GLOBAL_FOLDER -e YARN_ENABLE_GLOBAL_CACHE -e BUILDPACK_CACHE_DIR -e CONTAINERBASE_CACHE_DIR -w \"/mnt/renovate/gh/recitalsoftware/yarn3\" docker.io/renovate/sidecar bash -l -c \"install-tool node v18.15.0 && install-tool corepack 0.17.0 && yarn install --mode=update-lockfile && yarn up @vue/apollo-composable@>=4.0.0-beta.1 --mode=update-lockfile\"",
"stderr": "",
"stdout": "installing v2 tool node v18.15.0\nlinking tool node v18.15.0\nnode: v18.15.0 /usr/local/bin/node\nnpm: 9.5.0 /usr/local/bin/npm\nInstalled v2 /usr/local/buildpack/tools/v2/node.sh in 8 seconds\nskip cleanup, not a docker build: 37be2fc221c0\ninstalling v2 tool corepack v0.17.0\nlinking tool corepack v0.17.0\n0.17.0\nInstalled v2 /usr/local/buildpack/tools/v2/corepack.sh in 4 seconds\nskip cleanup, not a docker build: 37be2fc221c0\n➤ YN0000: ┌ Resolution step\n➤ YN0002: │ @vue/apollo-composable@npm:4.0.0-beta.3 [dd12c] doesn't provide typescript (pae615), requested by ts-essentials\n➤ YN0002: │ yarn3@workspace:. doesn't provide @apollo/client (pf3d3e), requested by @vue/apollo-composable\n➤ YN0002: │ yarn3@workspace:. doesn't provide graphql (pe1da9), requested by @vue/apollo-composable\n➤ YN0002: │ yarn3@workspace:. doesn't provide vue (p36305), requested by @vue/apollo-composable\n➤ YN0000: │ Some peer dependencies are incorrectly met; run yarn explain peer-requirements <hash> for details, where <hash> is the six-letter p-prefixed code\n➤ YN0000: └ Completed\n➤ YN0000: ┌ Fetch step\n➤ YN0000: └ Completed\n➤ YN0000: ┌ Link step\n➤ YN0073: │ Skipped due to mode=update-lockfile\n➤ YN0000: └ Completed\n➤ YN0000: Done with warnings in 0s 114ms\n",
"options": {
"cwd": "/mnt/renovate/gh/recitalsoftware/yarn3",
"encoding": "utf-8",
"env": {
"NPM_CONFIG_CACHE": "/tmp/renovate-cache/others/npm",
"npm_config_store": "/tmp/renovate-cache/others/pnpm",
"CI": "true",
"YARN_ENABLE_IMMUTABLE_INSTALLS": "false",
"YARN_HTTP_TIMEOUT": "100000",
"YARN_GLOBAL_FOLDER": "/tmp/renovate-cache/others/berry",
"YARN_ENABLE_GLOBAL_CACHE": "1",
"HOME": "/home/ubuntu",
"PATH": "/home/ubuntu/.local/bin:/home/ubuntu/bin:/opt/buildpack/tools/python/3.9.3/bin:/home/ubuntu/.npm-global/bin:/home/ubuntu/renovateapp/node_modules/.bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin",
"LC_ALL": "C.UTF-8",
"LANG": "C.UTF-8",
"BUILDPACK_CACHE_DIR": "/tmp/containerbase",
"CONTAINERBASE_CACHE_DIR": "/tmp/containerbase"
},
"maxBuffer": 10485760,
"timeout": 900000
},
"exitCode": 1,
"message": "Command failed: docker run --rm --name=renovate_sidecar --label=renovate_child -v \"/mnt/renovate/gh/recitalsoftware/yarn3\":\"/mnt/renovate/gh/recitalsoftware/yarn3\" -v \"/tmp/renovate-cache\":\"/tmp/renovate-cache\" -v \"/tmp/containerbase\":\"/tmp/containerbase\" -e NPM_CONFIG_CACHE -e npm_config_store -e CI -e YARN_ENABLE_IMMUTABLE_INSTALLS -e YARN_HTTP_TIMEOUT -e YARN_GLOBAL_FOLDER -e YARN_ENABLE_GLOBAL_CACHE -e BUILDPACK_CACHE_DIR -e CONTAINERBASE_CACHE_DIR -w \"/mnt/renovate/gh/recitalsoftware/yarn3\" docker.io/renovate/sidecar bash -l -c \"install-tool node v18.15.0 && install-tool corepack 0.17.0 && yarn install --mode=update-lockfile && yarn up @vue/apollo-composable@>=4.0.0-beta.1 --mode=update-lockfile\"\n",
"stack": "ExecError: Command failed: docker run --rm --name=renovate_sidecar --label=renovate_child -v \"/mnt/renovate/gh/recitalsoftware/yarn3\":\"/mnt/renovate/gh/recitalsoftware/yarn3\" -v \"/tmp/renovate-cache\":\"/tmp/renovate-cache\" -v \"/tmp/containerbase\":\"/tmp/containerbase\" -e NPM_CONFIG_CACHE -e npm_config_store -e CI -e YARN_ENABLE_IMMUTABLE_INSTALLS -e YARN_HTTP_TIMEOUT -e YARN_GLOBAL_FOLDER -e YARN_ENABLE_GLOBAL_CACHE -e BUILDPACK_CACHE_DIR -e CONTAINERBASE_CACHE_DIR -w \"/mnt/renovate/gh/recitalsoftware/yarn3\" docker.io/renovate/sidecar bash -l -c \"install-tool node v18.15.0 && install-tool corepack 0.17.0 && yarn install --mode=update-lockfile && yarn up @vue/apollo-composable@>=4.0.0-beta.1 --mode=update-lockfile\"\n\n at ChildProcess.<anonymous> (/home/ubuntu/renovateapp/node_modules/renovate/dist/util/exec/common.js:87:24)\n at ChildProcess.emit (node:events:525:35)\n at ChildProcess.emit (node:domain:489:12)\n at Process.ChildProcess._handle.onexit (node:internal/child_process:293:12)"
}
}
```
</details>
### Have you created a minimal reproduction repository?
I have read the minimal reproductions documentation and linked to such a repository in the bug description | priority | execerror updating package yarn up may need quotes how are you running renovate mend renovate hosted app on github com was this something which used to work for you and then stopped i never saw this working describe the bug renovate won t update this package appears to only happen when specifying range strategy in range only didn t test other range strategies just the absence of that so whatever the default would be included below the problem is that the upgrade command is returning an execerror with an exit code of there s no output to suggest what part of the command failed this may be a red herring especially since i m using zsh not bash but i ve tried to run the same command locally just the part sidecar runs and it works fine i do however need to add quotes around the package version yarn install mode update lockfile yarn up vue apollo composable beta mode update lockfile dev oss repros ➤ ┌ resolution step snip ➤ done with warnings in zsh beta not found adding quotes around the package in the upgrade command fixes that problem yarn up vue apollo composable beta mode update lockfile relevant debug logs logs debug executing command branch renovate vue apollo composable x lockfile command docker run rm name renovate sidecar label renovate child v mnt renovate gh recitalsoftware mnt renovate gh recitalsoftware v tmp renovate cache tmp renovate cache v tmp containerbase tmp containerbase e npm config cache e npm config store e ci e yarn enable immutable installs e yarn http timeout e yarn global folder e yarn enable global cache e buildpack cache dir e containerbase cache dir w mnt renovate gh recitalsoftware docker io renovate sidecar bash l c install tool node install tool corepack yarn install mode update lockfile yarn up vue apollo composable beta mode update lockfile debug rawexec err branch renovate vue apollo composable x lockfile err name execerror cmd bin sh c docker run rm name renovate sidecar label renovate child v mnt renovate gh recitalsoftware mnt renovate gh recitalsoftware v tmp renovate cache tmp renovate cache v tmp containerbase tmp containerbase e npm config cache e npm config store e ci e yarn enable immutable installs e yarn http timeout e yarn global folder e yarn enable global cache e buildpack cache dir e containerbase cache dir w mnt renovate gh recitalsoftware docker io renovate sidecar bash l c install tool node install tool corepack yarn install mode update lockfile yarn up vue apollo composable beta mode update lockfile stderr stdout installing tool node nlinking tool node nnode usr local bin node nnpm usr local bin npm ninstalled usr local buildpack tools node sh in seconds nskip cleanup not a docker build ninstalling tool corepack nlinking tool corepack ninstalled usr local buildpack tools corepack sh in seconds nskip cleanup not a docker build n➤ ┌ resolution step n➤ │ vue apollo composable npm beta doesn t provide typescript requested by ts essentials n➤ │ workspace doesn t provide apollo client requested by vue apollo composable n➤ │ workspace doesn t provide graphql requested by vue apollo composable n➤ │ workspace doesn t provide vue requested by vue apollo composable n➤ │ some peer dependencies are incorrectly met run yarn explain peer requirements for details where is the six letter p prefixed code n➤ └ completed n➤ ┌ fetch step n➤ └ completed n➤ ┌ link step n➤ │ skipped due to mode update lockfile n➤ └ completed n➤ done with warnings in n options cwd mnt renovate gh recitalsoftware encoding utf env npm config cache tmp renovate cache others npm npm config store tmp renovate cache others pnpm ci true yarn enable immutable installs false yarn http timeout yarn global folder tmp renovate cache others berry yarn enable global cache home home ubuntu path home ubuntu local bin home ubuntu bin opt buildpack tools python bin home ubuntu npm global bin home ubuntu renovateapp node modules bin usr local sbin usr local bin usr sbin usr bin sbin bin lc all c utf lang c utf buildpack cache dir tmp containerbase containerbase cache dir tmp containerbase maxbuffer timeout exitcode message command failed docker run rm name renovate sidecar label renovate child v mnt renovate gh recitalsoftware mnt renovate gh recitalsoftware v tmp renovate cache tmp renovate cache v tmp containerbase tmp containerbase e npm config cache e npm config store e ci e yarn enable immutable installs e yarn http timeout e yarn global folder e yarn enable global cache e buildpack cache dir e containerbase cache dir w mnt renovate gh recitalsoftware docker io renovate sidecar bash l c install tool node install tool corepack yarn install mode update lockfile yarn up vue apollo composable beta mode update lockfile n stack execerror command failed docker run rm name renovate sidecar label renovate child v mnt renovate gh recitalsoftware mnt renovate gh recitalsoftware v tmp renovate cache tmp renovate cache v tmp containerbase tmp containerbase e npm config cache e npm config store e ci e yarn enable immutable installs e yarn http timeout e yarn global folder e yarn enable global cache e buildpack cache dir e containerbase cache dir w mnt renovate gh recitalsoftware docker io renovate sidecar bash l c install tool node install tool corepack yarn install mode update lockfile yarn up vue apollo composable beta mode update lockfile n n at childprocess home ubuntu renovateapp node modules renovate dist util exec common js n at childprocess emit node events n at childprocess emit node domain n at process childprocess handle onexit node internal child process have you created a minimal reproduction repository i have read the minimal reproductions documentation and linked to such a repository in the bug description | 1 |
499,673 | 14,475,750,853 | IssuesEvent | 2020-12-10 02:23:42 | Thorium-Sim/thorium | https://api.github.com/repos/Thorium-Sim/thorium | opened | Lighting can't be controlled on blackout | priority/high type/bug | ### Requested By: Jordan
### Priority: High
### Version: 3.0.7
When blacking out all stations this also blacks out the lighting and keyboard clients. (They're one in the same on my setup right now). However those ones cannot be individually turned back on without online-ing all stations again. And lights cannot be controlled with that client blacked out. Can it not black out with everything else please?
### Steps to Reproduce
Have a lighting client controlling a DMX
Black out all stations
Attempt to control lights. | 1.0 | Lighting can't be controlled on blackout - ### Requested By: Jordan
### Priority: High
### Version: 3.0.7
When blacking out all stations this also blacks out the lighting and keyboard clients. (They're one in the same on my setup right now). However those ones cannot be individually turned back on without online-ing all stations again. And lights cannot be controlled with that client blacked out. Can it not black out with everything else please?
### Steps to Reproduce
Have a lighting client controlling a DMX
Black out all stations
Attempt to control lights. | priority | lighting can t be controlled on blackout requested by jordan priority high version when blacking out all stations this also blacks out the lighting and keyboard clients they re one in the same on my setup right now however those ones cannot be individually turned back on without online ing all stations again and lights cannot be controlled with that client blacked out can it not black out with everything else please steps to reproduce have a lighting client controlling a dmx black out all stations attempt to control lights | 1 |
451,881 | 13,042,859,940 | IssuesEvent | 2020-07-28 23:43:18 | phetsims/axon | https://api.github.com/repos/phetsims/axon | opened | PropertyStateHandler unregisterOrderDependenciesForProperty is a performance bottleneck | dev:phet-io priority:2-high type:performance | In https://github.com/phetsims/natural-selection/issues/140, we're investigating big performance differences in `brand=phet` vs `brand=phet-io` for Natural Selection.
In https://github.com/phetsims/natural-selection/issues/140#issuecomment-665348980:
> PropertyStateHandler.js `unregisterOrderDependenciesForProperty` contains a loop that is O(N2). Percentage varies by platform, but was ~25%. | 1.0 | PropertyStateHandler unregisterOrderDependenciesForProperty is a performance bottleneck - In https://github.com/phetsims/natural-selection/issues/140, we're investigating big performance differences in `brand=phet` vs `brand=phet-io` for Natural Selection.
In https://github.com/phetsims/natural-selection/issues/140#issuecomment-665348980:
> PropertyStateHandler.js `unregisterOrderDependenciesForProperty` contains a loop that is O(N2). Percentage varies by platform, but was ~25%. | priority | propertystatehandler unregisterorderdependenciesforproperty is a performance bottleneck in we re investigating big performance differences in brand phet vs brand phet io for natural selection in propertystatehandler js unregisterorderdependenciesforproperty contains a loop that is o percentage varies by platform but was | 1 |
510,421 | 14,790,401,782 | IssuesEvent | 2021-01-12 11:59:18 | KomodoPlatform/atomicDEX-Desktop | https://api.github.com/repos/KomodoPlatform/atomicDEX-Desktop | closed | [FEATURE REQUEST]: new mm2 | 0.3.1 priority: high | **Is your feature request related to a problem? Please describe.**
Add the new version of mm2 and add all the required tests
PR: #533 | 1.0 | [FEATURE REQUEST]: new mm2 - **Is your feature request related to a problem? Please describe.**
Add the new version of mm2 and add all the required tests
PR: #533 | priority | new is your feature request related to a problem please describe add the new version of and add all the required tests pr | 1 |
390,101 | 11,524,782,981 | IssuesEvent | 2020-02-15 02:54:02 | MolSnoo/Alter-Ego | https://api.github.com/repos/MolSnoo/Alter-Ego | closed | Remake use_player command | feature high priority | New use_player command should allow the player to use an item in their inventory. If the item has limited uses, it should turn into its next stage(s) when all its uses have been depleted. If no next stages exist, it should be removed from the player's inventory. | 1.0 | Remake use_player command - New use_player command should allow the player to use an item in their inventory. If the item has limited uses, it should turn into its next stage(s) when all its uses have been depleted. If no next stages exist, it should be removed from the player's inventory. | priority | remake use player command new use player command should allow the player to use an item in their inventory if the item has limited uses it should turn into its next stage s when all its uses have been depleted if no next stages exist it should be removed from the player s inventory | 1 |
95,080 | 3,933,839,769 | IssuesEvent | 2016-04-25 20:31:22 | eXistSolutions/hsg-shell | https://api.github.com/repos/eXistSolutions/hsg-shell | closed | Fix locale for date formats | high priority | On pages where dates are rendered, the locale appears to be affecting the date format. For example, see https://52.20.198.114/departmenthistory/travels/president/roosevelt-franklin-d:
> 
FYI the code that generates these dates is https://github.com/eXistSolutions/hsg-shell/blob/master/modules/travels-html.xqm#L159-L171. | 1.0 | Fix locale for date formats - On pages where dates are rendered, the locale appears to be affecting the date format. For example, see https://52.20.198.114/departmenthistory/travels/president/roosevelt-franklin-d:
> 
FYI the code that generates these dates is https://github.com/eXistSolutions/hsg-shell/blob/master/modules/travels-html.xqm#L159-L171. | priority | fix locale for date formats on pages where dates are rendered the locale appears to be affecting the date format for example see fyi the code that generates these dates is | 1 |
432,355 | 12,491,887,823 | IssuesEvent | 2020-06-01 05:41:06 | OrangeJuice7/SDL-OpenGL-Game-Framework | https://api.github.com/repos/OrangeJuice7/SDL-OpenGL-Game-Framework | opened | v1.0 demo | area.Model area.Scene priority.medium work.high | v1.0 will demo a shoot 'em up. Specifically:
- The map will consist of a large rectangular arena with walls at the borders.
- This arena is populated by the player character, and enemies.
- Enemies regularly spawn in the playing field, at a rate dependent on the current number of enemies already spawned. (Big or small enemies may contribute different weights to this total.)
- There is an internal difficulty value that increases with time. At higher difficulties, more enemies an more difficult varieties of enemies will spawn. (Enemies come in discrete upgraded versions?)
- Examples of enemies: Basic melee attacker; basic ranged attacker; suicide bomber; melee speeder; melee swarmer; melee tank; short-ranged acid cloud sprayer; ranged artillery bomber.
- Enemies strive to kill the player. The player's goal is to survive endlessly and rack up a high score.
- The player has access to a variety of weapons to kill enemies: A sweeping melee weapon; a pistol; shotgun; machine gun; sniper rifle; grenades; flamethrower. Only the melee weapon and pistol are available at the start. (Other weapons have finite ammo?)
- Killing enemies rewards the player with score, and money drops from the enemy's corpse. Rarely, upgrades may drop from the enemies as well.
- At any time, the player may bring up the shop screen, which pauses the game. This screen allows the player to exchange money for upgrades.
- Upgrades will improve the player's health, defense, and weapons (damage, fire rate, range, splash area etc.).
New devs can experiment with implementing weapon and enemy types. | 1.0 | v1.0 demo - v1.0 will demo a shoot 'em up. Specifically:
- The map will consist of a large rectangular arena with walls at the borders.
- This arena is populated by the player character, and enemies.
- Enemies regularly spawn in the playing field, at a rate dependent on the current number of enemies already spawned. (Big or small enemies may contribute different weights to this total.)
- There is an internal difficulty value that increases with time. At higher difficulties, more enemies an more difficult varieties of enemies will spawn. (Enemies come in discrete upgraded versions?)
- Examples of enemies: Basic melee attacker; basic ranged attacker; suicide bomber; melee speeder; melee swarmer; melee tank; short-ranged acid cloud sprayer; ranged artillery bomber.
- Enemies strive to kill the player. The player's goal is to survive endlessly and rack up a high score.
- The player has access to a variety of weapons to kill enemies: A sweeping melee weapon; a pistol; shotgun; machine gun; sniper rifle; grenades; flamethrower. Only the melee weapon and pistol are available at the start. (Other weapons have finite ammo?)
- Killing enemies rewards the player with score, and money drops from the enemy's corpse. Rarely, upgrades may drop from the enemies as well.
- At any time, the player may bring up the shop screen, which pauses the game. This screen allows the player to exchange money for upgrades.
- Upgrades will improve the player's health, defense, and weapons (damage, fire rate, range, splash area etc.).
New devs can experiment with implementing weapon and enemy types. | priority | demo will demo a shoot em up specifically the map will consist of a large rectangular arena with walls at the borders this arena is populated by the player character and enemies enemies regularly spawn in the playing field at a rate dependent on the current number of enemies already spawned big or small enemies may contribute different weights to this total there is an internal difficulty value that increases with time at higher difficulties more enemies an more difficult varieties of enemies will spawn enemies come in discrete upgraded versions examples of enemies basic melee attacker basic ranged attacker suicide bomber melee speeder melee swarmer melee tank short ranged acid cloud sprayer ranged artillery bomber enemies strive to kill the player the player s goal is to survive endlessly and rack up a high score the player has access to a variety of weapons to kill enemies a sweeping melee weapon a pistol shotgun machine gun sniper rifle grenades flamethrower only the melee weapon and pistol are available at the start other weapons have finite ammo killing enemies rewards the player with score and money drops from the enemy s corpse rarely upgrades may drop from the enemies as well at any time the player may bring up the shop screen which pauses the game this screen allows the player to exchange money for upgrades upgrades will improve the player s health defense and weapons damage fire rate range splash area etc new devs can experiment with implementing weapon and enemy types | 1 |
564,762 | 16,740,459,810 | IssuesEvent | 2021-06-11 09:08:25 | codetapacademy/codetap.academy | https://api.github.com/repos/codetapacademy/codetap.academy | closed | feat: improve the looks of the toggle button on the video description of a lesson in a course | Priority: High Status: Available Type: Enhancement | improve the looks of the toggle button on the video description of a lesson in a course. He suggested to do it with some fancy CSS that will look awesome! We're working on it right now! | 1.0 | feat: improve the looks of the toggle button on the video description of a lesson in a course - improve the looks of the toggle button on the video description of a lesson in a course. He suggested to do it with some fancy CSS that will look awesome! We're working on it right now! | priority | feat improve the looks of the toggle button on the video description of a lesson in a course improve the looks of the toggle button on the video description of a lesson in a course he suggested to do it with some fancy css that will look awesome we re working on it right now | 1 |
220,191 | 7,354,002,764 | IssuesEvent | 2018-03-09 03:59:18 | CS2103JAN2018-W13-B4/main | https://api.github.com/repos/CS2103JAN2018-W13-B4/main | opened | 7. As a user I want to delete a subtask from a task | priority.high type.story | ... so that I can remove a subtask I no longer need. | 1.0 | 7. As a user I want to delete a subtask from a task - ... so that I can remove a subtask I no longer need. | priority | as a user i want to delete a subtask from a task so that i can remove a subtask i no longer need | 1 |
192,205 | 6,847,613,116 | IssuesEvent | 2017-11-13 15:55:28 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | Flashing EM Starterkit with EM7D fails on master | area: Flashing bug priority: high | This works fine on 1.9.x but fails on master with the new flashing scripts | 1.0 | Flashing EM Starterkit with EM7D fails on master - This works fine on 1.9.x but fails on master with the new flashing scripts | priority | flashing em starterkit with fails on master this works fine on x but fails on master with the new flashing scripts | 1 |
400,065 | 11,768,290,572 | IssuesEvent | 2020-03-15 09:34:15 | Seniru/merchant | https://api.github.com/repos/Seniru/merchant | closed | Display non-qualified jobs too | affects: structure difficulty: medium priority: high type: enhancement type: feature request | Many players bugged me that they can't see the jobs they've created (probably becaused they are not qualifed). Therefore it seems it d be good to display those unqualified jobs as well.
Here are some possible solutions:
- Create a new tab in jobs list ~~(this would be an overkill though)~~
- Blur the non-qualified jobs (making the text grey and non-selectable) | 1.0 | Display non-qualified jobs too - Many players bugged me that they can't see the jobs they've created (probably becaused they are not qualifed). Therefore it seems it d be good to display those unqualified jobs as well.
Here are some possible solutions:
- Create a new tab in jobs list ~~(this would be an overkill though)~~
- Blur the non-qualified jobs (making the text grey and non-selectable) | priority | display non qualified jobs too many players bugged me that they can t see the jobs they ve created probably becaused they are not qualifed therefore it seems it d be good to display those unqualified jobs as well here are some possible solutions create a new tab in jobs list this would be an overkill though blur the non qualified jobs making the text grey and non selectable | 1 |
432,441 | 12,493,212,472 | IssuesEvent | 2020-06-01 08:50:07 | geosolutions-it/testbed16 | https://api.github.com/repos/geosolutions-it/testbed16 | opened | Add TLS certificate | DevOps Priority: High Task | TLS must be enabled.
TODO:
- [ ] Buy a certificate (@randomorder please confirm we won't reuse a star certificate for this server)
- [ ] Edit ansible playbooks to include the **encrypted** certificate
- [ ] Update documentation and credentials document
- [ ] Deploy | 1.0 | Add TLS certificate - TLS must be enabled.
TODO:
- [ ] Buy a certificate (@randomorder please confirm we won't reuse a star certificate for this server)
- [ ] Edit ansible playbooks to include the **encrypted** certificate
- [ ] Update documentation and credentials document
- [ ] Deploy | priority | add tls certificate tls must be enabled todo buy a certificate randomorder please confirm we won t reuse a star certificate for this server edit ansible playbooks to include the encrypted certificate update documentation and credentials document deploy | 1 |
205,035 | 7,093,536,962 | IssuesEvent | 2018-01-12 20:57:02 | Lutes/lisnd | https://api.github.com/repos/Lutes/lisnd | opened | Make sure photo is square | FEATURE High Priority | Either crop the photo to make it square, or run a check on the image to make sure it is square. | 1.0 | Make sure photo is square - Either crop the photo to make it square, or run a check on the image to make sure it is square. | priority | make sure photo is square either crop the photo to make it square or run a check on the image to make sure it is square | 1 |
451,609 | 13,039,251,873 | IssuesEvent | 2020-07-28 16:26:19 | department-of-veterans-affairs/caseflow | https://api.github.com/repos/department-of-veterans-affairs/caseflow | closed | Upstream Error | investigate failed DecisionDocument jobs | Priority: High Product: IDT Source: Upstream-error Team: Foxtrot 🦊 | ## Description
DecisionDocument jobs created when IDT users attempt to upload a decision document are failing with the error "FILENUMBER does not exist within the system". This ticket exists to document the investigation and resolution of this error.
## Impacted Jobs
DDs: [15694, 7127, 3851, 1973, 30261, 28780, 38738, 40343, 43313, 46651, 47806, 49382, 70035, 67498]
## Technical notes
```ruby
# get all the DecisionDocuments affected by stuck jobs with this error
> dds = DecisionDocument.potentially_stuck.where("error LIKE '%FILENUMBER does not exist within the system%'").order(:attempted_at)
# examine one DecisionDocument
> dd = dds.last
# check the error message
> dd.error.match(/message>([^<]+)</)[1]
=> "FILENUMBER does not exist within the system."
# when was the job last attempted?
> helper.time_ago_in_words(dd.attempted_at)
=> "1 day"
# get the file number sent in the request to BGS
> rfn = dd.error.match(/file_number="(\d+)"/)[1]
# that number is the same as the appeal's veteran_file_number
# (see ExternalApi::VBMSService.initialize_upload)
> rfn == dd.appeal.veteran_file_number
=> true
# get BGS's file number on record for the veteran
> bfn = BGSService.new.fetch_file_number_by_ssn(dd.veteran.ssn)
# do the two match?
> rfn == bfn
=> false
# was the ssn sent as the veteran's file number?
> rfn == dd.veteran.ssn
=> true
```
### Things to check
1. Is there a duplicate Caseflow veteran record with the file number we got from BGS?
```ruby
> Veteran.where(file_number: bfn)
```
If there is a duplicate record, and you can't find evidence of [duplicate veteran records in the corporate database](https://github.com/department-of-veterans-affairs/appeals-team/wiki/How-to-resolve-duplicate-veteran-records), you may be able to "merge" the two records in Caseflow by updating the `file_number` column on one and its associated records, and deleting the other.
If there's no duplicate Caseflow veteran record, you may be able to update the `file_number` column on the veteran with the file number you pulled from BGS. Unfortunately, this number is used in at least 11 other objects in Caseflow:
```
a. Appeal .where(veteran_file_number: fn)
a. AvailableHearingLocations .where(veteran_file_number: fn)
b. BgsPowerOfAttorney .where(file_number: fn)
d. Document .where(file_number: fn)
e. EndProductEstablishment .where(veteran_file_number: fn)
f. Form8 .where(file_number: fn)
h. HigherLevelReview .where(veteran_file_number: fn)
i. Intake .where(veteran_file_number: fn)
r. RampElection .where(veteran_file_number: fn)
r. RampRefiling .where(veteran_file_number: fn)
s. SupplementalClaim .where(veteran_file_number: fn)
v. Veteran .where(file_number: fn)
```
Here's a method you can paste in the console and use to quickly figure out where a file number is being used in the database:
```ruby
def objects_with_file_number(fn)
"a.#{Appeal.where(veteran_file_number: fn).count} a.#{AvailableHearingLocations.where(veteran_file_number: fn).count} b.#{BgsPowerOfAttorney.where(file_number: fn).count} d.#{Document.where(file_number: fn).count} e.#{EndProductEstablishment.where(veteran_file_number: fn).count} f.#{Form8.where(file_number: fn).count} h.#{HigherLevelReview.where(veteran_file_number: fn).count} i.#{Intake.where(veteran_file_number: fn).count} r.#{RampElection.where(veteran_file_number: fn).count} r.#{RampRefiling.where(veteran_file_number: fn).count} s.#{SupplementalClaim.where(veteran_file_number: fn).count} v.#{Veteran.where(file_number: fn).count}"
end
```
output looks like this (use the letters next to model names in the list above for reference)
```ruby
> objects_with_file_number(v.file_number)
=> "a.1 a.0 b.1 d.84 e.0 f.0 h.0 i.1 r.0 r.0 s.0 v.1"
# this file_number is associated with 1 appeal, 1 BgsPowerOfAttorney, 84 Documents, 1 Intake, and 1 Veteran
```
Once you know which object contain the file number you need to update, you can update them like this:
```ruby
> ds = Document.where(file_number: v.file_number)
> ds.update_all(file_number: bfn)
```
See [this wiki page](https://github.com/department-of-veterans-affairs/appeals-deployment/wiki/Bat-Team-Quick-Ref#ssn-stored-as-file-number) for additional guidance.
## Background/context/resources
[Slack thread where the issue was first raised](https://dsva.slack.com/archives/CD5DAQNCU/p1586263187020200)
### Out of scope
- Create a nightly job that checks stuck DecisionDocument jobs and alerts us to ones in this state, or resolves them, if the process can be automated. | 1.0 | Upstream Error | investigate failed DecisionDocument jobs - ## Description
DecisionDocument jobs created when IDT users attempt to upload a decision document are failing with the error "FILENUMBER does not exist within the system". This ticket exists to document the investigation and resolution of this error.
## Impacted Jobs
DDs: [15694, 7127, 3851, 1973, 30261, 28780, 38738, 40343, 43313, 46651, 47806, 49382, 70035, 67498]
## Technical notes
```ruby
# get all the DecisionDocuments affected by stuck jobs with this error
> dds = DecisionDocument.potentially_stuck.where("error LIKE '%FILENUMBER does not exist within the system%'").order(:attempted_at)
# examine one DecisionDocument
> dd = dds.last
# check the error message
> dd.error.match(/message>([^<]+)</)[1]
=> "FILENUMBER does not exist within the system."
# when was the job last attempted?
> helper.time_ago_in_words(dd.attempted_at)
=> "1 day"
# get the file number sent in the request to BGS
> rfn = dd.error.match(/file_number="(\d+)"/)[1]
# that number is the same as the appeal's veteran_file_number
# (see ExternalApi::VBMSService.initialize_upload)
> rfn == dd.appeal.veteran_file_number
=> true
# get BGS's file number on record for the veteran
> bfn = BGSService.new.fetch_file_number_by_ssn(dd.veteran.ssn)
# do the two match?
> rfn == bfn
=> false
# was the ssn sent as the veteran's file number?
> rfn == dd.veteran.ssn
=> true
```
### Things to check
1. Is there a duplicate Caseflow veteran record with the file number we got from BGS?
```ruby
> Veteran.where(file_number: bfn)
```
If there is a duplicate record, and you can't find evidence of [duplicate veteran records in the corporate database](https://github.com/department-of-veterans-affairs/appeals-team/wiki/How-to-resolve-duplicate-veteran-records), you may be able to "merge" the two records in Caseflow by updating the `file_number` column on one and its associated records, and deleting the other.
If there's no duplicate Caseflow veteran record, you may be able to update the `file_number` column on the veteran with the file number you pulled from BGS. Unfortunately, this number is used in at least 11 other objects in Caseflow:
```
a. Appeal .where(veteran_file_number: fn)
a. AvailableHearingLocations .where(veteran_file_number: fn)
b. BgsPowerOfAttorney .where(file_number: fn)
d. Document .where(file_number: fn)
e. EndProductEstablishment .where(veteran_file_number: fn)
f. Form8 .where(file_number: fn)
h. HigherLevelReview .where(veteran_file_number: fn)
i. Intake .where(veteran_file_number: fn)
r. RampElection .where(veteran_file_number: fn)
r. RampRefiling .where(veteran_file_number: fn)
s. SupplementalClaim .where(veteran_file_number: fn)
v. Veteran .where(file_number: fn)
```
Here's a method you can paste in the console and use to quickly figure out where a file number is being used in the database:
```ruby
def objects_with_file_number(fn)
"a.#{Appeal.where(veteran_file_number: fn).count} a.#{AvailableHearingLocations.where(veteran_file_number: fn).count} b.#{BgsPowerOfAttorney.where(file_number: fn).count} d.#{Document.where(file_number: fn).count} e.#{EndProductEstablishment.where(veteran_file_number: fn).count} f.#{Form8.where(file_number: fn).count} h.#{HigherLevelReview.where(veteran_file_number: fn).count} i.#{Intake.where(veteran_file_number: fn).count} r.#{RampElection.where(veteran_file_number: fn).count} r.#{RampRefiling.where(veteran_file_number: fn).count} s.#{SupplementalClaim.where(veteran_file_number: fn).count} v.#{Veteran.where(file_number: fn).count}"
end
```
output looks like this (use the letters next to model names in the list above for reference)
```ruby
> objects_with_file_number(v.file_number)
=> "a.1 a.0 b.1 d.84 e.0 f.0 h.0 i.1 r.0 r.0 s.0 v.1"
# this file_number is associated with 1 appeal, 1 BgsPowerOfAttorney, 84 Documents, 1 Intake, and 1 Veteran
```
Once you know which object contain the file number you need to update, you can update them like this:
```ruby
> ds = Document.where(file_number: v.file_number)
> ds.update_all(file_number: bfn)
```
See [this wiki page](https://github.com/department-of-veterans-affairs/appeals-deployment/wiki/Bat-Team-Quick-Ref#ssn-stored-as-file-number) for additional guidance.
## Background/context/resources
[Slack thread where the issue was first raised](https://dsva.slack.com/archives/CD5DAQNCU/p1586263187020200)
### Out of scope
- Create a nightly job that checks stuck DecisionDocument jobs and alerts us to ones in this state, or resolves them, if the process can be automated. | priority | upstream error investigate failed decisiondocument jobs description decisiondocument jobs created when idt users attempt to upload a decision document are failing with the error filenumber does not exist within the system this ticket exists to document the investigation and resolution of this error impacted jobs dds technical notes ruby get all the decisiondocuments affected by stuck jobs with this error dds decisiondocument potentially stuck where error like filenumber does not exist within the system order attempted at examine one decisiondocument dd dds last check the error message dd error match message filenumber does not exist within the system when was the job last attempted helper time ago in words dd attempted at day get the file number sent in the request to bgs rfn dd error match file number d that number is the same as the appeal s veteran file number see externalapi vbmsservice initialize upload rfn dd appeal veteran file number true get bgs s file number on record for the veteran bfn bgsservice new fetch file number by ssn dd veteran ssn do the two match rfn bfn false was the ssn sent as the veteran s file number rfn dd veteran ssn true things to check is there a duplicate caseflow veteran record with the file number we got from bgs ruby veteran where file number bfn if there is a duplicate record and you can t find evidence of you may be able to merge the two records in caseflow by updating the file number column on one and its associated records and deleting the other if there s no duplicate caseflow veteran record you may be able to update the file number column on the veteran with the file number you pulled from bgs unfortunately this number is used in at least other objects in caseflow a appeal where veteran file number fn a availablehearinglocations where veteran file number fn b bgspowerofattorney where file number fn d document where file number fn e endproductestablishment where veteran file number fn f where file number fn h higherlevelreview where veteran file number fn i intake where veteran file number fn r rampelection where veteran file number fn r ramprefiling where veteran file number fn s supplementalclaim where veteran file number fn v veteran where file number fn here s a method you can paste in the console and use to quickly figure out where a file number is being used in the database ruby def objects with file number fn a appeal where veteran file number fn count a availablehearinglocations where veteran file number fn count b bgspowerofattorney where file number fn count d document where file number fn count e endproductestablishment where veteran file number fn count f where file number fn count h higherlevelreview where veteran file number fn count i intake where veteran file number fn count r rampelection where veteran file number fn count r ramprefiling where veteran file number fn count s supplementalclaim where veteran file number fn count v veteran where file number fn count end output looks like this use the letters next to model names in the list above for reference ruby objects with file number v file number a a b d e f h i r r s v this file number is associated with appeal bgspowerofattorney documents intake and veteran once you know which object contain the file number you need to update you can update them like this ruby ds document where file number v file number ds update all file number bfn see for additional guidance background context resources out of scope create a nightly job that checks stuck decisiondocument jobs and alerts us to ones in this state or resolves them if the process can be automated | 1 |
208,069 | 7,135,579,577 | IssuesEvent | 2018-01-23 01:51:29 | neuropoly/spinalcordtoolbox | https://api.github.com/repos/neuropoly/spinalcordtoolbox | closed | sct_fmri_compute_tsnr does not compute tSNR | bug priority:HIGH | ### Description
Nothing happens when sct_fmri_compute_tsnr is run.
### Steps to Reproduce
[fmri.nii.gz](https://github.com/neuropoly/spinalcordtoolbox/files/1472903/fmri.nii.gz)
sct_fmri_compute_tsnr -i fmri.nii.gz
### State of spinalcordtoolbox
--
Spinal Cord Toolbox (master/6976c73a4318a0acd7eb88da2d190fde73b6beb0)
Running /Users/stall_admin/spinalcordtoolbox/scripts/sct_fmri_compute_tsnr.py -i fmri.nii.gz
| 1.0 | sct_fmri_compute_tsnr does not compute tSNR - ### Description
Nothing happens when sct_fmri_compute_tsnr is run.
### Steps to Reproduce
[fmri.nii.gz](https://github.com/neuropoly/spinalcordtoolbox/files/1472903/fmri.nii.gz)
sct_fmri_compute_tsnr -i fmri.nii.gz
### State of spinalcordtoolbox
--
Spinal Cord Toolbox (master/6976c73a4318a0acd7eb88da2d190fde73b6beb0)
Running /Users/stall_admin/spinalcordtoolbox/scripts/sct_fmri_compute_tsnr.py -i fmri.nii.gz
| priority | sct fmri compute tsnr does not compute tsnr description nothing happens when sct fmri compute tsnr is run steps to reproduce sct fmri compute tsnr i fmri nii gz state of spinalcordtoolbox spinal cord toolbox master running users stall admin spinalcordtoolbox scripts sct fmri compute tsnr py i fmri nii gz | 1 |
243,236 | 7,854,548,231 | IssuesEvent | 2018-06-20 21:08:39 | SCAII/SCAII | https://api.github.com/repos/SCAII/SCAII | closed | agent tank doesn't move during stepping forward sometimes | priority: high | seen on q-and-a branch
Can defer if provide at least two replays where this does not occur. (one for tutorial and one for q&a for sandboxing) | 1.0 | agent tank doesn't move during stepping forward sometimes - seen on q-and-a branch
Can defer if provide at least two replays where this does not occur. (one for tutorial and one for q&a for sandboxing) | priority | agent tank doesn t move during stepping forward sometimes seen on q and a branch can defer if provide at least two replays where this does not occur one for tutorial and one for q a for sandboxing | 1 |
590,676 | 17,784,340,684 | IssuesEvent | 2021-08-31 09:14:16 | EscolaLMS/Cart | https://api.github.com/repos/EscolaLMS/Cart | closed | orders pagination & search | priority high | - [ ] results paginated with default laravel paginator
- [ ] swagger update, reference https://escolalms.github.io/Courses/#/Course/9663f6d2e6826935f1abd0fa2734f590 (query params & Order Definition)
- [ ] search by course (if possible)
- [ ] search by user
- [ ] search by dates
- [ ] sort by ID
- [ ] sort by update/date
- [ ] what does student/tutor/admin see - managing roles | 1.0 | orders pagination & search - - [ ] results paginated with default laravel paginator
- [ ] swagger update, reference https://escolalms.github.io/Courses/#/Course/9663f6d2e6826935f1abd0fa2734f590 (query params & Order Definition)
- [ ] search by course (if possible)
- [ ] search by user
- [ ] search by dates
- [ ] sort by ID
- [ ] sort by update/date
- [ ] what does student/tutor/admin see - managing roles | priority | orders pagination search results paginated with default laravel paginator swagger update reference query params order definition search by course if possible search by user search by dates sort by id sort by update date what does student tutor admin see managing roles | 1 |
696,646 | 23,909,342,055 | IssuesEvent | 2022-09-09 06:29:58 | signum-network/signum-node | https://api.github.com/repos/signum-network/signum-node | closed | Get closed orders per account and/or asset | priority-high API | There is no easy way to get the closed orders for an account and/or asset | 1.0 | Get closed orders per account and/or asset - There is no easy way to get the closed orders for an account and/or asset | priority | get closed orders per account and or asset there is no easy way to get the closed orders for an account and or asset | 1 |
157,630 | 6,010,060,212 | IssuesEvent | 2017-06-06 12:17:10 | metasfresh/metasfresh-webui-frontend | https://api.github.com/repos/metasfresh/metasfresh-webui-frontend | closed | Kickstart Avatar button | branch:master priority:high | This issue it's a split off https://github.com/metasfresh/metasfresh-webui-frontend/issues/802 which shall deliver only the minimal functionaly.
### Type of issue
Feature request
### Expected behavior
* have a Avatar button on the top right side of the browser window
* for now, the avatar shall be just a "no name" picture.
* when clicked, it shall show following options (which basically shall be moved from Action menu)
** Settings
** Logout
| 1.0 | Kickstart Avatar button - This issue it's a split off https://github.com/metasfresh/metasfresh-webui-frontend/issues/802 which shall deliver only the minimal functionaly.
### Type of issue
Feature request
### Expected behavior
* have a Avatar button on the top right side of the browser window
* for now, the avatar shall be just a "no name" picture.
* when clicked, it shall show following options (which basically shall be moved from Action menu)
** Settings
** Logout
| priority | kickstart avatar button this issue it s a split off which shall deliver only the minimal functionaly type of issue feature request expected behavior have a avatar button on the top right side of the browser window for now the avatar shall be just a no name picture when clicked it shall show following options which basically shall be moved from action menu settings logout | 1 |
83,594 | 3,637,779,871 | IssuesEvent | 2016-02-12 12:42:45 | mantidproject/mantid | https://api.github.com/repos/mantidproject/mantid | opened | Fix bug in vtkMDHexFactory which causes 3D slices of higher dimensional workspaces to not contain any points | Group: Vates Priority: High | vtkMDHexFactory expects zero thickness slices to contain data, due to the strange implementation of MDPlanes before #15209. Previous to that PR it would actually return a slice 2 bins thick if the two MDPlanes defining the slice coincide with a bin boundary and 1 bin thick otherwise.
Defining the higher bound plane (the one with the -ve normal) as lying on the next highest bin boundary would solve the problem. NB efficiency is important as the slicing procedure is repeated many times, for example to create a movie when the 4th dimension is time. | 1.0 | Fix bug in vtkMDHexFactory which causes 3D slices of higher dimensional workspaces to not contain any points - vtkMDHexFactory expects zero thickness slices to contain data, due to the strange implementation of MDPlanes before #15209. Previous to that PR it would actually return a slice 2 bins thick if the two MDPlanes defining the slice coincide with a bin boundary and 1 bin thick otherwise.
Defining the higher bound plane (the one with the -ve normal) as lying on the next highest bin boundary would solve the problem. NB efficiency is important as the slicing procedure is repeated many times, for example to create a movie when the 4th dimension is time. | priority | fix bug in vtkmdhexfactory which causes slices of higher dimensional workspaces to not contain any points vtkmdhexfactory expects zero thickness slices to contain data due to the strange implementation of mdplanes before previous to that pr it would actually return a slice bins thick if the two mdplanes defining the slice coincide with a bin boundary and bin thick otherwise defining the higher bound plane the one with the ve normal as lying on the next highest bin boundary would solve the problem nb efficiency is important as the slicing procedure is repeated many times for example to create a movie when the dimension is time | 1 |
191,460 | 6,829,178,112 | IssuesEvent | 2017-11-08 23:10:42 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | Halting server after placing WorldObject can leave occupancy but no worldobject | High Priority | This is because they are saved in different databases. PlantSim handles this by checking every plant and only adding those that have the matching world blocks, should do same for worldobjects. | 1.0 | Halting server after placing WorldObject can leave occupancy but no worldobject - This is because they are saved in different databases. PlantSim handles this by checking every plant and only adding those that have the matching world blocks, should do same for worldobjects. | priority | halting server after placing worldobject can leave occupancy but no worldobject this is because they are saved in different databases plantsim handles this by checking every plant and only adding those that have the matching world blocks should do same for worldobjects | 1 |
723,005 | 24,880,696,084 | IssuesEvent | 2022-10-28 00:35:47 | AY2223S1-CS2103T-T12-4/tp | https://api.github.com/repos/AY2223S1-CS2103T-T12-4/tp | closed | Fix options-based command syntax | type.Bug priority.High severity.Medium | Instead of considering `/` as the delimiter for options-based command (i.e., only consider options before the first `/` character), the delimiter should be a valid prefix in `CliSyntax`. | 1.0 | Fix options-based command syntax - Instead of considering `/` as the delimiter for options-based command (i.e., only consider options before the first `/` character), the delimiter should be a valid prefix in `CliSyntax`. | priority | fix options based command syntax instead of considering as the delimiter for options based command i e only consider options before the first character the delimiter should be a valid prefix in clisyntax | 1 |
632,873 | 20,237,493,254 | IssuesEvent | 2022-02-14 04:48:55 | nbbang-sinchon/backend | https://api.github.com/repos/nbbang-sinchon/backend | opened | 수정된 데이터를 받아 DB에 저장한다. | BE high priority party-update | 처음에 기존 제목, 내용, 위치, 태그, 모집 인원수 정보를 제공한다. 태그 정보는 list로 전송한다.
수정된 데이터를 받아서 DB에 저장한다. | 1.0 | 수정된 데이터를 받아 DB에 저장한다. - 처음에 기존 제목, 내용, 위치, 태그, 모집 인원수 정보를 제공한다. 태그 정보는 list로 전송한다.
수정된 데이터를 받아서 DB에 저장한다. | priority | 수정된 데이터를 받아 db에 저장한다 처음에 기존 제목 내용 위치 태그 모집 인원수 정보를 제공한다 태그 정보는 list로 전송한다 수정된 데이터를 받아서 db에 저장한다 | 1 |
272,637 | 8,515,734,278 | IssuesEvent | 2018-10-31 22:46:34 | swe-ms-boun/2018fall-swe574-g1 | https://api.github.com/repos/swe-ms-boun/2018fall-swe574-g1 | opened | Stackoverflow exception | priority.high tier.api type.bug | Hi,
While the code block is active:
``` java
@Override
protected void configure(final HttpSecurity http) throws Exception {
http
.authorizeRequests()
.antMatchers("/resources/**", "/", "/login", "/register").permitAll()
.antMatchers("/welcome/**").hasRole("USER")
.anyRequest().authenticated();
}
```
And when I try to access the login page, I am getting the exception below:
``` java
2018-11-01 00:42:28.845 ERROR 31832 --- [nio-8080-exec-1] o.a.c.c.C.[.[.[/].[dispatcherServlet] : Servlet.service() for servlet [dispatcherServlet] in context with path [] threw exception [Handler processing failed; nested exception is java.lang.StackOverflowError] with root cause
java.lang.StackOverflowError: null
at org.springframework.web.servlet.support.RequestContext$JstlLocaleResolver.getJstlTimeZone(RequestContext.java:929) ~[spring-webmvc-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.springframework.web.servlet.support.RequestContext.getFallbackTimeZone(RequestContext.java:360) ~[spring-webmvc-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.springframework.web.servlet.support.RequestContext.initContext(RequestContext.java:259) ~[spring-webmvc-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.springframework.web.servlet.support.RequestContext.<init>(RequestContext.java:198) ~[spring-webmvc-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.thymeleaf.spring4.view.ThymeleafView.renderFragment(ThymeleafView.java:236) ~[thymeleaf-spring4-2.1.6.RELEASE.jar:2.1.6.RELEASE]
at org.thymeleaf.spring4.view.ThymeleafView.render(ThymeleafView.java:190) ~[thymeleaf-spring4-2.1.6.RELEASE.jar:2.1.6.RELEASE]
at org.springframework.web.servlet.DispatcherServlet.render(DispatcherServlet.java:1286) ~[spring-webmvc-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.springframework.web.servlet.DispatcherServlet.processDispatchResult(DispatcherServlet.java:1041) ~[spring-webmvc-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.springframework.web.servlet.DispatcherServlet.doDispatch(DispatcherServlet.java:984) ~[spring-webmvc-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.springframework.web.servlet.DispatcherServlet.doService(DispatcherServlet.java:901) ~[spring-webmvc-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.springframework.web.servlet.FrameworkServlet.processRequest(FrameworkServlet.java:970) ~[spring-webmvc-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.springframework.web.servlet.FrameworkServlet.doGet(FrameworkServlet.java:861) ~[spring-webmvc-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at javax.servlet.http.HttpServlet.service(HttpServlet.java:635) ~[tomcat-embed-core-8.5.34.jar:8.5.34]
at org.springframework.web.servlet.FrameworkServlet.service(FrameworkServlet.java:846) ~[spring-webmvc-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at javax.servlet.http.HttpServlet.service(HttpServlet.java:742) ~[tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:231) ~[tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) ~[tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.tomcat.websocket.server.WsFilter.doFilter(WsFilter.java:52) ~[tomcat-embed-websocket-8.5.34.jar:8.5.34]
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) ~[tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) ~[tomcat-embed-core-8.5.34.jar:8.5.34]
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:317) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.access.intercept.FilterSecurityInterceptor.invoke(FilterSecurityInterceptor.java:127) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.access.intercept.FilterSecurityInterceptor.doFilter(FilterSecurityInterceptor.java:91) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.access.ExceptionTranslationFilter.doFilter(ExceptionTranslationFilter.java:114) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.session.SessionManagementFilter.doFilter(SessionManagementFilter.java:137) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.authentication.AnonymousAuthenticationFilter.doFilter(AnonymousAuthenticationFilter.java:111) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.servletapi.SecurityContextHolderAwareRequestFilter.doFilter(SecurityContextHolderAwareRequestFilter.java:170) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.savedrequest.RequestCacheAwareFilter.doFilter(RequestCacheAwareFilter.java:63) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.authentication.logout.LogoutFilter.doFilter(LogoutFilter.java:116) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.csrf.CsrfFilter.doFilterInternal(CsrfFilter.java:100) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107) ~[spring-web-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.header.HeaderWriterFilter.doFilterInternal(HeaderWriterFilter.java:66) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107) ~[spring-web-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.context.SecurityContextPersistenceFilter.doFilter(SecurityContextPersistenceFilter.java:105) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.context.request.async.WebAsyncManagerIntegrationFilter.doFilterInternal(WebAsyncManagerIntegrationFilter.java:56) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107) ~[spring-web-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.FilterChainProxy.doFilterInternal(FilterChainProxy.java:214) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.FilterChainProxy.doFilter(FilterChainProxy.java:177) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.web.filter.DelegatingFilterProxy.invokeDelegate(DelegatingFilterProxy.java:347) ~[spring-web-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.springframework.web.filter.DelegatingFilterProxy.doFilter(DelegatingFilterProxy.java:263) ~[spring-web-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) ~[tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) ~[tomcat-embed-core-8.5.34.jar:8.5.34]
at org.springframework.web.filter.RequestContextFilter.doFilterInternal(RequestContextFilter.java:99) ~[spring-web-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107) ~[spring-web-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) ~[tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) ~[tomcat-embed-core-8.5.34.jar:8.5.34]
at org.springframework.web.filter.HttpPutFormContentFilter.doFilterInternal(HttpPutFormContentFilter.java:109) ~[spring-web-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107) ~[spring-web-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) ~[tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) ~[tomcat-embed-core-8.5.34.jar:8.5.34]
at org.springframework.web.filter.HiddenHttpMethodFilter.doFilterInternal(HiddenHttpMethodFilter.java:93) ~[spring-web-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107) ~[spring-web-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) ~[tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) ~[tomcat-embed-core-8.5.34.jar:8.5.34]
at org.springframework.web.filter.CharacterEncodingFilter.doFilterInternal(CharacterEncodingFilter.java:197) ~[spring-web-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107) ~[spring-web-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) ~[tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) ~[tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:198) ~[tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:96) [tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:493) [tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:140) [tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:81) [tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:87) [tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:342) [tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.coyote.http11.Http11Processor.service(Http11Processor.java:800) [tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.coyote.AbstractProcessorLight.process(AbstractProcessorLight.java:66) [tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.coyote.AbstractProtocol$ConnectionHandler.process(AbstractProtocol.java:806) [tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1498) [tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.tomcat.util.net.SocketProcessorBase.run(SocketProcessorBase.java:49) [tomcat-embed-core-8.5.34.jar:8.5.34]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [na:1.8.0_20]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [na:1.8.0_20]
at org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) [tomcat-embed-core-8.5.34.jar:8.5.34]
at java.lang.Thread.run(Thread.java:745) [na:1.8.0_20]
```
After commenting the method above, the exception disappears. | 1.0 | Stackoverflow exception - Hi,
While the code block is active:
``` java
@Override
protected void configure(final HttpSecurity http) throws Exception {
http
.authorizeRequests()
.antMatchers("/resources/**", "/", "/login", "/register").permitAll()
.antMatchers("/welcome/**").hasRole("USER")
.anyRequest().authenticated();
}
```
And when I try to access the login page, I am getting the exception below:
``` java
2018-11-01 00:42:28.845 ERROR 31832 --- [nio-8080-exec-1] o.a.c.c.C.[.[.[/].[dispatcherServlet] : Servlet.service() for servlet [dispatcherServlet] in context with path [] threw exception [Handler processing failed; nested exception is java.lang.StackOverflowError] with root cause
java.lang.StackOverflowError: null
at org.springframework.web.servlet.support.RequestContext$JstlLocaleResolver.getJstlTimeZone(RequestContext.java:929) ~[spring-webmvc-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.springframework.web.servlet.support.RequestContext.getFallbackTimeZone(RequestContext.java:360) ~[spring-webmvc-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.springframework.web.servlet.support.RequestContext.initContext(RequestContext.java:259) ~[spring-webmvc-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.springframework.web.servlet.support.RequestContext.<init>(RequestContext.java:198) ~[spring-webmvc-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.thymeleaf.spring4.view.ThymeleafView.renderFragment(ThymeleafView.java:236) ~[thymeleaf-spring4-2.1.6.RELEASE.jar:2.1.6.RELEASE]
at org.thymeleaf.spring4.view.ThymeleafView.render(ThymeleafView.java:190) ~[thymeleaf-spring4-2.1.6.RELEASE.jar:2.1.6.RELEASE]
at org.springframework.web.servlet.DispatcherServlet.render(DispatcherServlet.java:1286) ~[spring-webmvc-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.springframework.web.servlet.DispatcherServlet.processDispatchResult(DispatcherServlet.java:1041) ~[spring-webmvc-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.springframework.web.servlet.DispatcherServlet.doDispatch(DispatcherServlet.java:984) ~[spring-webmvc-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.springframework.web.servlet.DispatcherServlet.doService(DispatcherServlet.java:901) ~[spring-webmvc-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.springframework.web.servlet.FrameworkServlet.processRequest(FrameworkServlet.java:970) ~[spring-webmvc-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.springframework.web.servlet.FrameworkServlet.doGet(FrameworkServlet.java:861) ~[spring-webmvc-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at javax.servlet.http.HttpServlet.service(HttpServlet.java:635) ~[tomcat-embed-core-8.5.34.jar:8.5.34]
at org.springframework.web.servlet.FrameworkServlet.service(FrameworkServlet.java:846) ~[spring-webmvc-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at javax.servlet.http.HttpServlet.service(HttpServlet.java:742) ~[tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:231) ~[tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) ~[tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.tomcat.websocket.server.WsFilter.doFilter(WsFilter.java:52) ~[tomcat-embed-websocket-8.5.34.jar:8.5.34]
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) ~[tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) ~[tomcat-embed-core-8.5.34.jar:8.5.34]
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:317) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.access.intercept.FilterSecurityInterceptor.invoke(FilterSecurityInterceptor.java:127) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.access.intercept.FilterSecurityInterceptor.doFilter(FilterSecurityInterceptor.java:91) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.access.ExceptionTranslationFilter.doFilter(ExceptionTranslationFilter.java:114) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.session.SessionManagementFilter.doFilter(SessionManagementFilter.java:137) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.authentication.AnonymousAuthenticationFilter.doFilter(AnonymousAuthenticationFilter.java:111) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.servletapi.SecurityContextHolderAwareRequestFilter.doFilter(SecurityContextHolderAwareRequestFilter.java:170) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.savedrequest.RequestCacheAwareFilter.doFilter(RequestCacheAwareFilter.java:63) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.authentication.logout.LogoutFilter.doFilter(LogoutFilter.java:116) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.csrf.CsrfFilter.doFilterInternal(CsrfFilter.java:100) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107) ~[spring-web-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.header.HeaderWriterFilter.doFilterInternal(HeaderWriterFilter.java:66) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107) ~[spring-web-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.context.SecurityContextPersistenceFilter.doFilter(SecurityContextPersistenceFilter.java:105) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.context.request.async.WebAsyncManagerIntegrationFilter.doFilterInternal(WebAsyncManagerIntegrationFilter.java:56) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107) ~[spring-web-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.FilterChainProxy.doFilterInternal(FilterChainProxy.java:214) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.security.web.FilterChainProxy.doFilter(FilterChainProxy.java:177) ~[spring-security-web-4.2.8.RELEASE.jar:4.2.8.RELEASE]
at org.springframework.web.filter.DelegatingFilterProxy.invokeDelegate(DelegatingFilterProxy.java:347) ~[spring-web-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.springframework.web.filter.DelegatingFilterProxy.doFilter(DelegatingFilterProxy.java:263) ~[spring-web-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) ~[tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) ~[tomcat-embed-core-8.5.34.jar:8.5.34]
at org.springframework.web.filter.RequestContextFilter.doFilterInternal(RequestContextFilter.java:99) ~[spring-web-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107) ~[spring-web-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) ~[tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) ~[tomcat-embed-core-8.5.34.jar:8.5.34]
at org.springframework.web.filter.HttpPutFormContentFilter.doFilterInternal(HttpPutFormContentFilter.java:109) ~[spring-web-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107) ~[spring-web-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) ~[tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) ~[tomcat-embed-core-8.5.34.jar:8.5.34]
at org.springframework.web.filter.HiddenHttpMethodFilter.doFilterInternal(HiddenHttpMethodFilter.java:93) ~[spring-web-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107) ~[spring-web-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) ~[tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) ~[tomcat-embed-core-8.5.34.jar:8.5.34]
at org.springframework.web.filter.CharacterEncodingFilter.doFilterInternal(CharacterEncodingFilter.java:197) ~[spring-web-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107) ~[spring-web-4.3.19.RELEASE.jar:4.3.19.RELEASE]
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) ~[tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) ~[tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:198) ~[tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:96) [tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:493) [tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:140) [tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:81) [tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:87) [tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:342) [tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.coyote.http11.Http11Processor.service(Http11Processor.java:800) [tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.coyote.AbstractProcessorLight.process(AbstractProcessorLight.java:66) [tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.coyote.AbstractProtocol$ConnectionHandler.process(AbstractProtocol.java:806) [tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1498) [tomcat-embed-core-8.5.34.jar:8.5.34]
at org.apache.tomcat.util.net.SocketProcessorBase.run(SocketProcessorBase.java:49) [tomcat-embed-core-8.5.34.jar:8.5.34]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [na:1.8.0_20]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [na:1.8.0_20]
at org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) [tomcat-embed-core-8.5.34.jar:8.5.34]
at java.lang.Thread.run(Thread.java:745) [na:1.8.0_20]
```
After commenting the method above, the exception disappears. | priority | stackoverflow exception hi while the code block is active java override protected void configure final httpsecurity http throws exception http authorizerequests antmatchers resources login register permitall antmatchers welcome hasrole user anyrequest authenticated and when i try to access the login page i am getting the exception below java error o a c c c servlet service for servlet in context with path threw exception with root cause java lang stackoverflowerror null at org springframework web servlet support requestcontext jstllocaleresolver getjstltimezone requestcontext java at org springframework web servlet support requestcontext getfallbacktimezone requestcontext java at org springframework web servlet support requestcontext initcontext requestcontext java at org springframework web servlet support requestcontext requestcontext java at org thymeleaf view thymeleafview renderfragment thymeleafview java at org thymeleaf view thymeleafview render thymeleafview java at org springframework web servlet dispatcherservlet render dispatcherservlet java at org springframework web servlet dispatcherservlet processdispatchresult dispatcherservlet java at org springframework web servlet dispatcherservlet dodispatch dispatcherservlet java at org springframework web servlet dispatcherservlet doservice dispatcherservlet java at org springframework web servlet frameworkservlet processrequest frameworkservlet java at org springframework web servlet frameworkservlet doget frameworkservlet java at javax servlet http httpservlet service httpservlet java at org springframework web servlet frameworkservlet service frameworkservlet java at javax servlet http httpservlet service httpservlet java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org apache tomcat websocket server wsfilter dofilter wsfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java at org springframework security web access intercept filtersecurityinterceptor invoke filtersecurityinterceptor java at org springframework security web access intercept filtersecurityinterceptor dofilter filtersecurityinterceptor java at org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java at org springframework security web access exceptiontranslationfilter dofilter exceptiontranslationfilter java at org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java at org springframework security web session sessionmanagementfilter dofilter sessionmanagementfilter java at org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java at org springframework security web authentication anonymousauthenticationfilter dofilter anonymousauthenticationfilter java at org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java at org springframework security web servletapi securitycontextholderawarerequestfilter dofilter securitycontextholderawarerequestfilter java at org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java at org springframework security web savedrequest requestcacheawarefilter dofilter requestcacheawarefilter java at org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java at org springframework security web authentication logout logoutfilter dofilter logoutfilter java at org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java at org springframework security web csrf csrffilter dofilterinternal csrffilter java at org springframework web filter onceperrequestfilter dofilter onceperrequestfilter java at org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java at org springframework security web header headerwriterfilter dofilterinternal headerwriterfilter java at org springframework web filter onceperrequestfilter dofilter onceperrequestfilter java at org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java at org springframework security web context securitycontextpersistencefilter dofilter securitycontextpersistencefilter java at org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java at org springframework security web context request async webasyncmanagerintegrationfilter dofilterinternal webasyncmanagerintegrationfilter java at org springframework web filter onceperrequestfilter dofilter onceperrequestfilter java at org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java at org springframework security web filterchainproxy dofilterinternal filterchainproxy java at org springframework security web filterchainproxy dofilter filterchainproxy java at org springframework web filter delegatingfilterproxy invokedelegate delegatingfilterproxy java at org springframework web filter delegatingfilterproxy dofilter delegatingfilterproxy java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org springframework web filter requestcontextfilter dofilterinternal requestcontextfilter java at org springframework web filter onceperrequestfilter dofilter onceperrequestfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org springframework web filter httpputformcontentfilter dofilterinternal httpputformcontentfilter java at org springframework web filter onceperrequestfilter dofilter onceperrequestfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org springframework web filter hiddenhttpmethodfilter dofilterinternal hiddenhttpmethodfilter java at org springframework web filter onceperrequestfilter dofilter onceperrequestfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org springframework web filter characterencodingfilter dofilterinternal characterencodingfilter java at org springframework web filter onceperrequestfilter dofilter onceperrequestfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org apache catalina core standardwrappervalve invoke standardwrappervalve java at org apache catalina core standardcontextvalve invoke standardcontextvalve java at org apache catalina authenticator authenticatorbase invoke authenticatorbase java at org apache catalina core standardhostvalve invoke standardhostvalve java at org apache catalina valves errorreportvalve invoke errorreportvalve java at org apache catalina core standardenginevalve invoke standardenginevalve java at org apache catalina connector coyoteadapter service coyoteadapter java at org apache coyote service java at org apache coyote abstractprocessorlight process abstractprocessorlight java at org apache coyote abstractprotocol connectionhandler process abstractprotocol java at org apache tomcat util net nioendpoint socketprocessor dorun nioendpoint java at org apache tomcat util net socketprocessorbase run socketprocessorbase java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at org apache tomcat util threads taskthread wrappingrunnable run taskthread java at java lang thread run thread java after commenting the method above the exception disappears | 1 |
675,762 | 23,104,713,359 | IssuesEvent | 2022-07-27 07:47:12 | haskell/cabal | https://api.github.com/repos/haskell/cabal | opened | Generating haddocks sometimes produces "index.json does not exist" | priority: high :fire: type: regression cabal-install: cmd/haddock regression on master | **Describe the bug**
See #hackage for an extensive testing session by multiple volunteers. It works with cabal 3.6.2, fails with cabal 3.8.0.RC1 (the "regression on master" label is a guess) . Failures have been reproduced with both GHC 9.2.3 and GHC 9.4.1-alpha3. A log from a failed attempt with GHC 9.4.1-alpha3 and head.hackage:
```
cabal v2-haddock --builddir=dir --haddock-for-hackage --enable-doc --haddock-options=--quickjump --allow-newer cabal-install-solver
...
haddock: internal error: /home/mikolaj/.cabal/store/ghc-9.4.0.20220623/directory-1.3.7.1-1b6710e8174fc4ee7baecc57c46ce782434603c3b3b58295a4fcbdc139bbd395/share/doc/html/doc-index.json: openBinaryFile: does not exist (No such file or directory)
```
but at some point (without head.hackage) there was also
```
haddock: internal error: /home/mikolaj/.cabal/store/ghc-9.4.0.20220623/edit-distance-0.2.2.1-a7ea0a2f3ddd1328ec73dacef20a73fff1fc5fe66f9f3edd881ded197bf6a796/share/doc/html/doc-index.json: openBinaryFile: does not exist (No such file or directory)
```
so the problem is not limited to packages shipped with GHC that we override and generate new haddocks for.
We should probably rip out all PRs that touch haddock merged after 3.6.2, in turn, and check if this fixes itself. We should also add a test that catches this problem.
**System information**
linux
| 1.0 | Generating haddocks sometimes produces "index.json does not exist" - **Describe the bug**
See #hackage for an extensive testing session by multiple volunteers. It works with cabal 3.6.2, fails with cabal 3.8.0.RC1 (the "regression on master" label is a guess) . Failures have been reproduced with both GHC 9.2.3 and GHC 9.4.1-alpha3. A log from a failed attempt with GHC 9.4.1-alpha3 and head.hackage:
```
cabal v2-haddock --builddir=dir --haddock-for-hackage --enable-doc --haddock-options=--quickjump --allow-newer cabal-install-solver
...
haddock: internal error: /home/mikolaj/.cabal/store/ghc-9.4.0.20220623/directory-1.3.7.1-1b6710e8174fc4ee7baecc57c46ce782434603c3b3b58295a4fcbdc139bbd395/share/doc/html/doc-index.json: openBinaryFile: does not exist (No such file or directory)
```
but at some point (without head.hackage) there was also
```
haddock: internal error: /home/mikolaj/.cabal/store/ghc-9.4.0.20220623/edit-distance-0.2.2.1-a7ea0a2f3ddd1328ec73dacef20a73fff1fc5fe66f9f3edd881ded197bf6a796/share/doc/html/doc-index.json: openBinaryFile: does not exist (No such file or directory)
```
so the problem is not limited to packages shipped with GHC that we override and generate new haddocks for.
We should probably rip out all PRs that touch haddock merged after 3.6.2, in turn, and check if this fixes itself. We should also add a test that catches this problem.
**System information**
linux
| priority | generating haddocks sometimes produces index json does not exist describe the bug see hackage for an extensive testing session by multiple volunteers it works with cabal fails with cabal the regression on master label is a guess failures have been reproduced with both ghc and ghc a log from a failed attempt with ghc and head hackage cabal haddock builddir dir haddock for hackage enable doc haddock options quickjump allow newer cabal install solver haddock internal error home mikolaj cabal store ghc directory share doc html doc index json openbinaryfile does not exist no such file or directory but at some point without head hackage there was also haddock internal error home mikolaj cabal store ghc edit distance share doc html doc index json openbinaryfile does not exist no such file or directory so the problem is not limited to packages shipped with ghc that we override and generate new haddocks for we should probably rip out all prs that touch haddock merged after in turn and check if this fixes itself we should also add a test that catches this problem system information linux | 1 |
263,793 | 8,302,010,010 | IssuesEvent | 2018-09-21 13:20:52 | openbankingspace/tpp-issues | https://api.github.com/repos/openbankingspace/tpp-issues | opened | The OBRI consent page is failing on IOS chrome | aspsp:ref_forgerock env:sandbox priority:high | We recently pushed a new UI for the consent and it appears this is not working well on IOS.
Note: It's been mentioned to us that on Android, it works well | 1.0 | The OBRI consent page is failing on IOS chrome - We recently pushed a new UI for the consent and it appears this is not working well on IOS.
Note: It's been mentioned to us that on Android, it works well | priority | the obri consent page is failing on ios chrome we recently pushed a new ui for the consent and it appears this is not working well on ios note it s been mentioned to us that on android it works well | 1 |
653,634 | 21,609,203,760 | IssuesEvent | 2022-05-04 08:20:23 | federicazoe/gradetools | https://api.github.com/repos/federicazoe/gradetools | closed | Comprehensive vignette | high priority | - [x] Create new vignette on project team grading with multiple files per team on GitHub
- [x] Add 2 team repos | 1.0 | Comprehensive vignette - - [x] Create new vignette on project team grading with multiple files per team on GitHub
- [x] Add 2 team repos | priority | comprehensive vignette create new vignette on project team grading with multiple files per team on github add team repos | 1 |
384,055 | 11,382,761,840 | IssuesEvent | 2020-01-29 03:08:39 | ucbtrans/opt | https://api.github.com/repos/ucbtrans/opt | closed | ControlFactory bug: multiple controllers for the same actuator must be allowed | High Priority bad/invalid solution bug | Gabriel,
ControlFactory does not allow creating multiple ramp meters for the same onramp and lane group.
Exception: **The ramp is already controlled. Please remove the controller first**.
Possible solution: You should re-use the actuator if it already exists instead of throwing the exception. | 1.0 | ControlFactory bug: multiple controllers for the same actuator must be allowed - Gabriel,
ControlFactory does not allow creating multiple ramp meters for the same onramp and lane group.
Exception: **The ramp is already controlled. Please remove the controller first**.
Possible solution: You should re-use the actuator if it already exists instead of throwing the exception. | priority | controlfactory bug multiple controllers for the same actuator must be allowed gabriel controlfactory does not allow creating multiple ramp meters for the same onramp and lane group exception the ramp is already controlled please remove the controller first possible solution you should re use the actuator if it already exists instead of throwing the exception | 1 |
32,235 | 2,751,047,202 | IssuesEvent | 2015-04-24 05:48:22 | DrupalCampWroclaw/organizacyjne_2015 | https://api.github.com/repos/DrupalCampWroclaw/organizacyjne_2015 | opened | Spotkanie organizacyjne z ASI w sprawie nagrywania | Czas: przed konferencją Priority: High | Trzeba się z nimi spotkać i ustalić plany co i kiedy nagrywają. | 1.0 | Spotkanie organizacyjne z ASI w sprawie nagrywania - Trzeba się z nimi spotkać i ustalić plany co i kiedy nagrywają. | priority | spotkanie organizacyjne z asi w sprawie nagrywania trzeba się z nimi spotkać i ustalić plany co i kiedy nagrywają | 1 |
668,986 | 22,606,712,051 | IssuesEvent | 2022-06-29 13:49:52 | COS301-SE-2022/Twitter-Summariser | https://api.github.com/repos/COS301-SE-2022/Twitter-Summariser | closed | (UI/Responsiveness): Make the application responsive | priority:high status:not-ready scope:frontend role:frontend-engineer type:enhance | @GO-Shoderu I recommend that you start working on the responsiveness of pages/components as making a non responsive application responsive is quite the challenge | 1.0 | (UI/Responsiveness): Make the application responsive - @GO-Shoderu I recommend that you start working on the responsiveness of pages/components as making a non responsive application responsive is quite the challenge | priority | ui responsiveness make the application responsive go shoderu i recommend that you start working on the responsiveness of pages components as making a non responsive application responsive is quite the challenge | 1 |
291,784 | 8,949,598,576 | IssuesEvent | 2019-01-25 08:14:04 | autopo-collective/autopo-app | https://api.github.com/repos/autopo-collective/autopo-app | closed | textbox weirdness | priority:high | tiny bug, when you type in the "write a comment" box at the top, the edge of the box stretches into the right banner (into team members, etc.) so sometimes you can't see some of the words you have typed.
| 1.0 | textbox weirdness - tiny bug, when you type in the "write a comment" box at the top, the edge of the box stretches into the right banner (into team members, etc.) so sometimes you can't see some of the words you have typed.
| priority | textbox weirdness tiny bug when you type in the write a comment box at the top the edge of the box stretches into the right banner into team members etc so sometimes you can t see some of the words you have typed | 1 |
52,117 | 3,021,633,089 | IssuesEvent | 2015-07-31 15:46:54 | meetinghouse/cms | https://api.github.com/repos/meetinghouse/cms | closed | Blog Post Publish checkbox functionality | High Priority | Similar to issue #61 for Projects, I notice that the publish checkbox for the Blog posts does not work. For instance, the blog post called "Test Post" is not marked as published, but does show up on the Builder's Notebook page:
http://shconstruction2.corbettresearchgroup.com/posts/9/edit

http://shconstruction2.corbettresearchgroup.com/posts

| 1.0 | Blog Post Publish checkbox functionality - Similar to issue #61 for Projects, I notice that the publish checkbox for the Blog posts does not work. For instance, the blog post called "Test Post" is not marked as published, but does show up on the Builder's Notebook page:
http://shconstruction2.corbettresearchgroup.com/posts/9/edit

http://shconstruction2.corbettresearchgroup.com/posts

| priority | blog post publish checkbox functionality similar to issue for projects i notice that the publish checkbox for the blog posts does not work for instance the blog post called test post is not marked as published but does show up on the builder s notebook page | 1 |
276,236 | 8,589,409,046 | IssuesEvent | 2018-11-14 16:22:48 | blackbaud/skyux2 | https://api.github.com/repos/blackbaud/skyux2 | opened | Document `debounceTime` input for autocomplete and lookup components | Priority: High sky-autocomplete sky-lookup | See https://github.com/blackbaud/skyux-lookup/pull/8.
@Blackbaud-SteveBrush, if you can add any details that you want included in the docs, it'd be much appreciated. Thanks! | 1.0 | Document `debounceTime` input for autocomplete and lookup components - See https://github.com/blackbaud/skyux-lookup/pull/8.
@Blackbaud-SteveBrush, if you can add any details that you want included in the docs, it'd be much appreciated. Thanks! | priority | document debouncetime input for autocomplete and lookup components see blackbaud stevebrush if you can add any details that you want included in the docs it d be much appreciated thanks | 1 |
795,184 | 28,064,985,020 | IssuesEvent | 2023-03-29 14:52:02 | clt313/SuperballVR | https://api.github.com/repos/clt313/SuperballVR | closed | Final build: build for multiple platforms | priority: high | This is more of a reminder for me to build the release for multiple platforms like Mac in addition to our usual Windows build. This way more people can play the game when it's finished. | 1.0 | Final build: build for multiple platforms - This is more of a reminder for me to build the release for multiple platforms like Mac in addition to our usual Windows build. This way more people can play the game when it's finished. | priority | final build build for multiple platforms this is more of a reminder for me to build the release for multiple platforms like mac in addition to our usual windows build this way more people can play the game when it s finished | 1 |
635,218 | 20,382,102,060 | IssuesEvent | 2022-02-21 23:48:59 | Team-4795/2022-Rapid-React | https://api.github.com/repos/Team-4795/2022-Rapid-React | closed | Issues from testing | priority:high | - Percent out not changing falcon speed
- Cleanup double inversions
- All motors going the wrong way | 1.0 | Issues from testing - - Percent out not changing falcon speed
- Cleanup double inversions
- All motors going the wrong way | priority | issues from testing percent out not changing falcon speed cleanup double inversions all motors going the wrong way | 1 |
650,282 | 21,367,206,379 | IssuesEvent | 2022-04-20 03:56:42 | bitsongofficial/sinfonia-ui | https://api.github.com/repos/bitsongofficial/sinfonia-ui | closed | Define abstract - Liquidity Mining | High Priority Review | We need a small text below the title just to introduce "Liquidity Mining" and "Earning" topics.
Character Limit: 100
<img width="667" alt="Schermata 2022-04-12 alle 13 38 39" src="https://user-images.githubusercontent.com/92384659/162952396-15b0dcb1-672e-414e-bb17-af2c1d49bbee.png">
. | 1.0 | Define abstract - Liquidity Mining - We need a small text below the title just to introduce "Liquidity Mining" and "Earning" topics.
Character Limit: 100
<img width="667" alt="Schermata 2022-04-12 alle 13 38 39" src="https://user-images.githubusercontent.com/92384659/162952396-15b0dcb1-672e-414e-bb17-af2c1d49bbee.png">
. | priority | define abstract liquidity mining we need a small text below the title just to introduce liquidity mining and earning topics character limit img width alt schermata alle src | 1 |
269,446 | 8,435,813,553 | IssuesEvent | 2018-10-17 14:02:17 | MARKETProtocol/dApp | https://api.github.com/repos/MARKETProtocol/dApp | closed | [BUG] Properly handle web3 connection states in the dApp | Priority: High Status: Review Needed Type: Bug | <!--
Hello!
Please use the template below for issue ideas or bugs found within MARKET Protocol.
If it is general support you need, reach out to us at
https://marketprotocol.io/discord
Provide a general summary of the issue in the title above and use relevant fields below to define the problem.
-->
<!--
## Before you `start work`
Please read our contribution [guidelines](https://docs.marketprotocol.io/#contributing) and if there is a bounty involved please also see [here](https://docs.marketprotocol.io/#gitcoin-and-bounties)
If you have ongoing work from other bounties with us where funding has not been released, please do not pick up a new issue. We would like to involve as many contributors as possible and parallelize the work flow as much as possible.
Please make sure to comment in the issue here immediately after starting work so we know your plans for implementation and a timeline.
Please also note that in order for work to be accepted, all code must be accompanied by test cases as well.
-->
### User Story
[comment]: # (As a <user type>, I want to <task> so that <goal>.)
### Why Is this Needed?
Currently if i'm not logged in, I can browse around the site and attempt to use features like Deploy Contract. Attempting to deploy the contract while not logged into meta mask blows up the entire dApp.
[comment]: # (Describe the problem and why this task is needed. Provide description of the current state, what you would like to happen, and what actually happen)
*Summary*:
### Description
[comment]: # (Feature or Bug? i.e Type: Bug)
*Type*:
### Current Behavior
[comment]: # (Describe what actually happened.)
When not logged into Meta Mask I can browse and attempt to use features of the website I shouldn't have access to.
### Expected Behavior
[comment]: # (Describe what you expected to happen.)
Prevent users from being able to access certain areas of the site when not logged into meta mask. A popup directing them to login to Meta Mask should be displayed.
### Reproduction
[comment]: # (Describe how we can replicate the bug step by step.)
Log out of meta mask > Deploy Contract
### Solution
[comment]: # (Provide a summary of the solution and a task list on what needs to be fixed.)
*Summary*:
### Definition of Done
[comment]: # (Any other information that would be useful, bullets are helpful.)
- [ ] Proper handling of logged in/logged out users
### Additional Information
[comment]: # (Any other information that would be useful, content, screenshots, etc.)
| 1.0 | [BUG] Properly handle web3 connection states in the dApp - <!--
Hello!
Please use the template below for issue ideas or bugs found within MARKET Protocol.
If it is general support you need, reach out to us at
https://marketprotocol.io/discord
Provide a general summary of the issue in the title above and use relevant fields below to define the problem.
-->
<!--
## Before you `start work`
Please read our contribution [guidelines](https://docs.marketprotocol.io/#contributing) and if there is a bounty involved please also see [here](https://docs.marketprotocol.io/#gitcoin-and-bounties)
If you have ongoing work from other bounties with us where funding has not been released, please do not pick up a new issue. We would like to involve as many contributors as possible and parallelize the work flow as much as possible.
Please make sure to comment in the issue here immediately after starting work so we know your plans for implementation and a timeline.
Please also note that in order for work to be accepted, all code must be accompanied by test cases as well.
-->
### User Story
[comment]: # (As a <user type>, I want to <task> so that <goal>.)
### Why Is this Needed?
Currently if i'm not logged in, I can browse around the site and attempt to use features like Deploy Contract. Attempting to deploy the contract while not logged into meta mask blows up the entire dApp.
[comment]: # (Describe the problem and why this task is needed. Provide description of the current state, what you would like to happen, and what actually happen)
*Summary*:
### Description
[comment]: # (Feature or Bug? i.e Type: Bug)
*Type*:
### Current Behavior
[comment]: # (Describe what actually happened.)
When not logged into Meta Mask I can browse and attempt to use features of the website I shouldn't have access to.
### Expected Behavior
[comment]: # (Describe what you expected to happen.)
Prevent users from being able to access certain areas of the site when not logged into meta mask. A popup directing them to login to Meta Mask should be displayed.
### Reproduction
[comment]: # (Describe how we can replicate the bug step by step.)
Log out of meta mask > Deploy Contract
### Solution
[comment]: # (Provide a summary of the solution and a task list on what needs to be fixed.)
*Summary*:
### Definition of Done
[comment]: # (Any other information that would be useful, bullets are helpful.)
- [ ] Proper handling of logged in/logged out users
### Additional Information
[comment]: # (Any other information that would be useful, content, screenshots, etc.)
| priority | properly handle connection states in the dapp hello please use the template below for issue ideas or bugs found within market protocol if it is general support you need reach out to us at provide a general summary of the issue in the title above and use relevant fields below to define the problem before you start work please read our contribution and if there is a bounty involved please also see if you have ongoing work from other bounties with us where funding has not been released please do not pick up a new issue we would like to involve as many contributors as possible and parallelize the work flow as much as possible please make sure to comment in the issue here immediately after starting work so we know your plans for implementation and a timeline please also note that in order for work to be accepted all code must be accompanied by test cases as well user story as a i want to so that why is this needed currently if i m not logged in i can browse around the site and attempt to use features like deploy contract attempting to deploy the contract while not logged into meta mask blows up the entire dapp describe the problem and why this task is needed provide description of the current state what you would like to happen and what actually happen summary description feature or bug i e type bug type current behavior describe what actually happened when not logged into meta mask i can browse and attempt to use features of the website i shouldn t have access to expected behavior describe what you expected to happen prevent users from being able to access certain areas of the site when not logged into meta mask a popup directing them to login to meta mask should be displayed reproduction describe how we can replicate the bug step by step log out of meta mask deploy contract solution provide a summary of the solution and a task list on what needs to be fixed summary definition of done any other information that would be useful bullets are helpful proper handling of logged in logged out users additional information any other information that would be useful content screenshots etc | 1 |
485,932 | 14,001,541,316 | IssuesEvent | 2020-10-28 13:46:59 | bigdataprocessor/bigdataprocessor2 | https://api.github.com/repos/bigdataprocessor/bigdataprocessor2 | opened | Benchmarking | enhancement high priority | @haesleinhuepf @nornil
I think I have everything in place now for some benchmarking. The issue is that there is an exponential number of parameter combinations and we thus have to limit ourselves.
### Dataset
For the initial benchmarking I suggest to only use one data set where the xyz dimensions are typical for light-sheet data.
```
FileType: Hdf5
BitDepth: 16
nX: 2048
nY: 2048
nZ: 101
nC: 1
nT: 3
```
### Readouts
- Loading of a single z-plane [ ms ]
- Loading and processing of a single z-plane [ ms ]
- Relevant to judge how interactive the application is
- To judge the pure processing time one can simply subtract the loading time (see above)
- Loading a single xyz volume [ ms ]
- `LPV`: Loading and processing of a single xyz volume [ ms ]
- Relevant to judge how fast a whole data set could be processed
- `SV`: Saving of a single xyz volume [ ms ]
- Relevant to judge how fast a whole data set could be re-saved
To estimate the total processing and saving (`TPS`) time of a whole data set on a single computer one would have to use the following formula:
`TPS = nC * nT * (LPV + SV)`
### Processing
In order to assess the processing we added a single processing step, namely a 3x3 binning in x&y.
This is a very typical processing step and should be the computationally most expensive one of the currently available.
Note that this significantly reduces the saving times as the data to be saved is 3x3=9 times smaller than the data to be loaded.
### Results
#### Tischi's Mac
##### Hardware
- 2.5 GHz Intel Core i7 (4 cores)
- Loading and saving from and to local SSD
##### Measurements
- Loading of a single z-plane [ ms ]: ~2
- Loading and processing of a single z-plane [ ms ]: ~100
- Loading a single xyz volume [ ms ]: ~200
- `LPV`: Loading and processing of a single xyz volume [ ms ]: ~6000
- `SV`: Saving of a single xyz volume [ ms ]: ~800
My plan would be to measure these numbers for a few different computer and data storage scenarios.
What do you think?
| 1.0 | Benchmarking - @haesleinhuepf @nornil
I think I have everything in place now for some benchmarking. The issue is that there is an exponential number of parameter combinations and we thus have to limit ourselves.
### Dataset
For the initial benchmarking I suggest to only use one data set where the xyz dimensions are typical for light-sheet data.
```
FileType: Hdf5
BitDepth: 16
nX: 2048
nY: 2048
nZ: 101
nC: 1
nT: 3
```
### Readouts
- Loading of a single z-plane [ ms ]
- Loading and processing of a single z-plane [ ms ]
- Relevant to judge how interactive the application is
- To judge the pure processing time one can simply subtract the loading time (see above)
- Loading a single xyz volume [ ms ]
- `LPV`: Loading and processing of a single xyz volume [ ms ]
- Relevant to judge how fast a whole data set could be processed
- `SV`: Saving of a single xyz volume [ ms ]
- Relevant to judge how fast a whole data set could be re-saved
To estimate the total processing and saving (`TPS`) time of a whole data set on a single computer one would have to use the following formula:
`TPS = nC * nT * (LPV + SV)`
### Processing
In order to assess the processing we added a single processing step, namely a 3x3 binning in x&y.
This is a very typical processing step and should be the computationally most expensive one of the currently available.
Note that this significantly reduces the saving times as the data to be saved is 3x3=9 times smaller than the data to be loaded.
### Results
#### Tischi's Mac
##### Hardware
- 2.5 GHz Intel Core i7 (4 cores)
- Loading and saving from and to local SSD
##### Measurements
- Loading of a single z-plane [ ms ]: ~2
- Loading and processing of a single z-plane [ ms ]: ~100
- Loading a single xyz volume [ ms ]: ~200
- `LPV`: Loading and processing of a single xyz volume [ ms ]: ~6000
- `SV`: Saving of a single xyz volume [ ms ]: ~800
My plan would be to measure these numbers for a few different computer and data storage scenarios.
What do you think?
| priority | benchmarking haesleinhuepf nornil i think i have everything in place now for some benchmarking the issue is that there is an exponential number of parameter combinations and we thus have to limit ourselves dataset for the initial benchmarking i suggest to only use one data set where the xyz dimensions are typical for light sheet data filetype bitdepth nx ny nz nc nt readouts loading of a single z plane loading and processing of a single z plane relevant to judge how interactive the application is to judge the pure processing time one can simply subtract the loading time see above loading a single xyz volume lpv loading and processing of a single xyz volume relevant to judge how fast a whole data set could be processed sv saving of a single xyz volume relevant to judge how fast a whole data set could be re saved to estimate the total processing and saving tps time of a whole data set on a single computer one would have to use the following formula tps nc nt lpv sv processing in order to assess the processing we added a single processing step namely a binning in x y this is a very typical processing step and should be the computationally most expensive one of the currently available note that this significantly reduces the saving times as the data to be saved is times smaller than the data to be loaded results tischi s mac hardware ghz intel core cores loading and saving from and to local ssd measurements loading of a single z plane loading and processing of a single z plane loading a single xyz volume lpv loading and processing of a single xyz volume sv saving of a single xyz volume my plan would be to measure these numbers for a few different computer and data storage scenarios what do you think | 1 |
462,417 | 13,246,780,085 | IssuesEvent | 2020-08-19 16:11:44 | FightPandemics/FightPandemics | https://api.github.com/repos/FightPandemics/FightPandemics | opened | Reset password condition and sign in password condition do not match | High Priority SignUp/Register | As a User, I should be able to use the password I reset to log in so that I can access the portal
Please check the conditions and importantly, the special character condition for logging in.
I received an error after resetting the password for logging in.
Expected behaviour:
The password setting criteria must be common for sign up, sign in and resetting links. | 1.0 | Reset password condition and sign in password condition do not match - As a User, I should be able to use the password I reset to log in so that I can access the portal
Please check the conditions and importantly, the special character condition for logging in.
I received an error after resetting the password for logging in.
Expected behaviour:
The password setting criteria must be common for sign up, sign in and resetting links. | priority | reset password condition and sign in password condition do not match as a user i should be able to use the password i reset to log in so that i can access the portal please check the conditions and importantly the special character condition for logging in i received an error after resetting the password for logging in expected behaviour the password setting criteria must be common for sign up sign in and resetting links | 1 |
479,714 | 13,804,919,437 | IssuesEvent | 2020-10-11 11:20:34 | HackYourFuture-CPH/chattie | https://api.github.com/repos/HackYourFuture-CPH/chattie | closed | Connect the roomList with the backend | High priority User story | Yuka's RoomList component should be connected to the backend endpoint | 1.0 | Connect the roomList with the backend - Yuka's RoomList component should be connected to the backend endpoint | priority | connect the roomlist with the backend yuka s roomlist component should be connected to the backend endpoint | 1 |
567,483 | 16,859,739,721 | IssuesEvent | 2021-06-21 11:29:50 | VirtusLab/akka-serialization-helper | https://api.github.com/repos/VirtusLab/akka-serialization-helper | closed | `spray.json.DeserializationException: Object is missing required member 'typeSymbol'` under Scala 2.12 | HIGH-PRIORITY bug | ```
[error] ## Exception when compiling 365 sources to /home/plipski/hydra-backend/services/target/scala-2.12/classes
[error] spray.json.DeserializationException: Object is missing required member 'typeSymbol'
[error] spray.json.package$.deserializationError(package.scala:23)
[error] spray.json.ProductFormats.fromField(ProductFormats.scala:61)
[error] spray.json.ProductFormats.fromField$(ProductFormats.scala:51)
[error] org.virtuslab.ash.writer.EventSchemaWriter.fromField(EventSchemaWriter.scala:10)
[error] spray.json.ProductFormatsInstances$$anon$5.read(ProductFormatsInstances.scala:133)
[error] spray.json.ProductFormatsInstances$$anon$5.read(ProductFormatsInstances.scala:121)
[error] spray.json.JsValue.convertTo(JsValue.scala:33)
[error] org.virtuslab.ash.writer.EventSchemaWriter.$anonfun$lastDump$2(EventSchemaWriter.scala:20)
[error] scala.collection.Iterator$$anon$10.next(Iterator.scala:461)
[error] scala.collection.Iterator$$anon$10.next(Iterator.scala:461)
[error] scala.collection.Iterator.foreach(Iterator.scala:943)
[error] scala.collection.Iterator.foreach$(Iterator.scala:943)
[error] scala.collection.AbstractIterator.foreach(Iterator.scala:1431)
[error] scala.collection.generic.Growable.$plus$plus$eq(Growable.scala:62)
[error] scala.collection.generic.Growable.$plus$plus$eq$(Growable.scala:53)
[error] scala.collection.immutable.Map$MapBuilderImpl.$plus$plus$eq(Map.scala:583)
[error] scala.collection.immutable.Map$MapBuilderImpl.$plus$plus$eq(Map.scala:533)
[error] scala.collection.TraversableOnce.toMap(TraversableOnce.scala:354)
[error] scala.collection.TraversableOnce.toMap$(TraversableOnce.scala:352)
[error] scala.collection.AbstractIterator.toMap(Iterator.scala:1431)
[error] org.virtuslab.ash.writer.EventSchemaWriter.lastDump$lzycompute(EventSchemaWriter.scala:22)
[error] org.virtuslab.ash.writer.EventSchemaWriter.lastDump(EventSchemaWriter.scala:16)
[error] org.virtuslab.ash.DumpEventSchemaCompilerPluginComponent$$anon$1$$anonfun$2.isDefinedAt(DumpEventSchemaCompilerPluginComponent.scala:46)
[error] org.virtuslab.ash.DumpEventSchemaCompilerPluginComponent$$anon$1$$anonfun$2.isDefinedAt(DumpEventSchemaCompilerPluginComponent.scala:45)
[error] scala.reflect.internal.Trees$CollectTreeTraverser.traverse(Trees.scala:1715)
[error] scala.reflect.internal.Trees$CollectTreeTraverser.traverse(Trees.scala:1712)
[error] scala.reflect.api.Trees$Traverser.$anonfun$traverseStats$2(Trees.scala:2506)
[error] scala.reflect.api.Trees$Traverser.atOwner(Trees.scala:2515)
[error] scala.reflect.api.Trees$Traverser.$anonfun$traverseStats$1(Trees.scala:2506)
[error] scala.reflect.api.Trees$Traverser.traverseStats(Trees.scala:2505)
[error] scala.reflect.internal.Trees.itraverse(Trees.scala:1390)
[error] scala.reflect.internal.Trees.itraverse$(Trees.scala:1264)
[error] scala.reflect.internal.SymbolTable.itraverse(SymbolTable.scala:28)
[error] scala.reflect.internal.SymbolTable.itraverse(SymbolTable.scala:28)
[error] scala.reflect.api.Trees$Traverser.traverse(Trees.scala:2483)
[error] scala.reflect.internal.Trees$CollectTreeTraverser.traverse(Trees.scala:1716)
[error] scala.reflect.internal.Trees$TreeContextApiImpl.collect(Trees.scala:124)
[error] org.virtuslab.ash.DumpEventSchemaCompilerPluginComponent$$anon$1.apply(DumpEventSchemaCompilerPluginComponent.scala:45)
[error] scala.tools.nsc.Global$GlobalPhase.applyPhase(Global.scala:454)
[error] scala.tools.nsc.Global$GlobalPhase.run(Global.scala:402)
[error] scala.tools.nsc.Global$Run.compileUnitsInternal(Global.scala:1511)
[error] scala.tools.nsc.Global$Run.compileUnits(Global.scala:1495)
[error] scala.tools.nsc.Global$Run.compileSources(Global.scala:1488)
[error] scala.tools.nsc.Global$Run.compileFiles(Global.scala:1596)
[error] xsbt.CachedCompiler0.run(CompilerBridge.scala:163)
[error] xsbt.CachedCompiler0.run(CompilerBridge.scala:134)
[error] xsbt.CompilerBridge.run(CompilerBridge.scala:39)
[error] sbt.internal.inc.AnalyzingCompiler.compile(AnalyzingCompiler.scala:92)
[error] sbt.internal.inc.MixedAnalyzingCompiler.$anonfun$compile$7(MixedAnalyzingCompiler.scala:186)
[error] scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
[error] sbt.internal.inc.MixedAnalyzingCompiler.timed(MixedAnalyzingCompiler.scala:241)
[error] sbt.internal.inc.MixedAnalyzingCompiler.$anonfun$compile$4(MixedAnalyzingCompiler.scala:176)
[error] sbt.internal.inc.MixedAnalyzingCompiler.$anonfun$compile$4$adapted(MixedAnalyzingCompiler.scala:157)
[error] sbt.internal.inc.JarUtils$.withPreviousJar(JarUtils.scala:239)
[error] sbt.internal.inc.MixedAnalyzingCompiler.compileScala$1(MixedAnalyzingCompiler.scala:157)
[error] sbt.internal.inc.MixedAnalyzingCompiler.compile(MixedAnalyzingCompiler.scala:204)
[error] sbt.internal.inc.IncrementalCompilerImpl.$anonfun$compileInternal$1(IncrementalCompilerImpl.scala:573)
[error] sbt.internal.inc.IncrementalCompilerImpl.$anonfun$compileInternal$1$adapted(IncrementalCompilerImpl.scala:573)
[error] sbt.internal.inc.Incremental$.$anonfun$apply$5(Incremental.scala:173)
[error] sbt.internal.inc.Incremental$.$anonfun$apply$5$adapted(Incremental.scala:171)
[error] sbt.internal.inc.Incremental$$anon$2.run(Incremental.scala:458)
[error] sbt.internal.inc.IncrementalCommon$CycleState.next(IncrementalCommon.scala:116)
[error] sbt.internal.inc.IncrementalCommon$$anon$1.next(IncrementalCommon.scala:56)
[error] sbt.internal.inc.IncrementalCommon$$anon$1.next(IncrementalCommon.scala:52)
[error] sbt.internal.inc.IncrementalCommon.cycle(IncrementalCommon.scala:261)
[error] sbt.internal.inc.Incremental$.$anonfun$incrementalCompile$8(Incremental.scala:413)
[error] sbt.internal.inc.Incremental$.withClassfileManager(Incremental.scala:498)
[error] sbt.internal.inc.Incremental$.incrementalCompile(Incremental.scala:400)
[error] sbt.internal.inc.Incremental$.apply(Incremental.scala:165)
[error] sbt.internal.inc.IncrementalCompilerImpl.compileInternal(IncrementalCompilerImpl.scala:573)
[error] sbt.internal.inc.IncrementalCompilerImpl.$anonfun$compileIncrementally$1(IncrementalCompilerImpl.scala:491)
[error] sbt.internal.inc.IncrementalCompilerImpl.handleCompilationError(IncrementalCompilerImpl.scala:332)
[error] sbt.internal.inc.IncrementalCompilerImpl.compileIncrementally(IncrementalCompilerImpl.scala:420)
[error] sbt.internal.inc.IncrementalCompilerImpl.compile(IncrementalCompilerImpl.scala:137)
[error] sbt.Defaults$.compileIncrementalTaskImpl(Defaults.scala:2176)
[error] sbt.Defaults$.$anonfun$compileIncrementalTask$2(Defaults.scala:2133)
[error] sbt.internal.io.Retry$.apply(Retry.scala:40)
[error] sbt.internal.io.Retry$.apply(Retry.scala:23)
[error] sbt.internal.server.BspCompileTask$.compute(BspCompileTask.scala:31)
[error] sbt.Defaults$.$anonfun$compileIncrementalTask$1(Defaults.scala:2129)
[error] scala.Function1.$anonfun$compose$1(Function1.scala:49)
[error] sbt.internal.util.$tilde$greater.$anonfun$$u2219$1(TypeFunctions.scala:62)
[error] sbt.std.Transform$$anon$4.work(Transform.scala:68)
[error] sbt.Execute.$anonfun$submit$2(Execute.scala:282)
[error] sbt.internal.util.ErrorHandling$.wideConvert(ErrorHandling.scala:23)
[error] sbt.Execute.work(Execute.scala:291)
[error] sbt.Execute.$anonfun$submit$1(Execute.scala:282)
[error] sbt.ConcurrentRestrictions$$anon$4.$anonfun$submitValid$1(ConcurrentRestrictions.scala:265)
[error] sbt.CompletionService$$anon$2.call(CompletionService.scala:64)
[error] java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
[error] java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
[error] java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
[error] java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
[error] java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
[error] java.base/java.lang.Thread.run(Thread.java:834)
[error]
[error] stack trace is suppressed; run last hydra-backend / Compile / compileIncremental for the full output
[error] (hydra-backend / Compile / compileIncremental) spray.json.DeserializationException: Object is missing required member 'typeSymbol'
[error] Total time: 111 s (01:51), completed 21 Jun 2021, 13:22:15
``` | 1.0 | `spray.json.DeserializationException: Object is missing required member 'typeSymbol'` under Scala 2.12 - ```
[error] ## Exception when compiling 365 sources to /home/plipski/hydra-backend/services/target/scala-2.12/classes
[error] spray.json.DeserializationException: Object is missing required member 'typeSymbol'
[error] spray.json.package$.deserializationError(package.scala:23)
[error] spray.json.ProductFormats.fromField(ProductFormats.scala:61)
[error] spray.json.ProductFormats.fromField$(ProductFormats.scala:51)
[error] org.virtuslab.ash.writer.EventSchemaWriter.fromField(EventSchemaWriter.scala:10)
[error] spray.json.ProductFormatsInstances$$anon$5.read(ProductFormatsInstances.scala:133)
[error] spray.json.ProductFormatsInstances$$anon$5.read(ProductFormatsInstances.scala:121)
[error] spray.json.JsValue.convertTo(JsValue.scala:33)
[error] org.virtuslab.ash.writer.EventSchemaWriter.$anonfun$lastDump$2(EventSchemaWriter.scala:20)
[error] scala.collection.Iterator$$anon$10.next(Iterator.scala:461)
[error] scala.collection.Iterator$$anon$10.next(Iterator.scala:461)
[error] scala.collection.Iterator.foreach(Iterator.scala:943)
[error] scala.collection.Iterator.foreach$(Iterator.scala:943)
[error] scala.collection.AbstractIterator.foreach(Iterator.scala:1431)
[error] scala.collection.generic.Growable.$plus$plus$eq(Growable.scala:62)
[error] scala.collection.generic.Growable.$plus$plus$eq$(Growable.scala:53)
[error] scala.collection.immutable.Map$MapBuilderImpl.$plus$plus$eq(Map.scala:583)
[error] scala.collection.immutable.Map$MapBuilderImpl.$plus$plus$eq(Map.scala:533)
[error] scala.collection.TraversableOnce.toMap(TraversableOnce.scala:354)
[error] scala.collection.TraversableOnce.toMap$(TraversableOnce.scala:352)
[error] scala.collection.AbstractIterator.toMap(Iterator.scala:1431)
[error] org.virtuslab.ash.writer.EventSchemaWriter.lastDump$lzycompute(EventSchemaWriter.scala:22)
[error] org.virtuslab.ash.writer.EventSchemaWriter.lastDump(EventSchemaWriter.scala:16)
[error] org.virtuslab.ash.DumpEventSchemaCompilerPluginComponent$$anon$1$$anonfun$2.isDefinedAt(DumpEventSchemaCompilerPluginComponent.scala:46)
[error] org.virtuslab.ash.DumpEventSchemaCompilerPluginComponent$$anon$1$$anonfun$2.isDefinedAt(DumpEventSchemaCompilerPluginComponent.scala:45)
[error] scala.reflect.internal.Trees$CollectTreeTraverser.traverse(Trees.scala:1715)
[error] scala.reflect.internal.Trees$CollectTreeTraverser.traverse(Trees.scala:1712)
[error] scala.reflect.api.Trees$Traverser.$anonfun$traverseStats$2(Trees.scala:2506)
[error] scala.reflect.api.Trees$Traverser.atOwner(Trees.scala:2515)
[error] scala.reflect.api.Trees$Traverser.$anonfun$traverseStats$1(Trees.scala:2506)
[error] scala.reflect.api.Trees$Traverser.traverseStats(Trees.scala:2505)
[error] scala.reflect.internal.Trees.itraverse(Trees.scala:1390)
[error] scala.reflect.internal.Trees.itraverse$(Trees.scala:1264)
[error] scala.reflect.internal.SymbolTable.itraverse(SymbolTable.scala:28)
[error] scala.reflect.internal.SymbolTable.itraverse(SymbolTable.scala:28)
[error] scala.reflect.api.Trees$Traverser.traverse(Trees.scala:2483)
[error] scala.reflect.internal.Trees$CollectTreeTraverser.traverse(Trees.scala:1716)
[error] scala.reflect.internal.Trees$TreeContextApiImpl.collect(Trees.scala:124)
[error] org.virtuslab.ash.DumpEventSchemaCompilerPluginComponent$$anon$1.apply(DumpEventSchemaCompilerPluginComponent.scala:45)
[error] scala.tools.nsc.Global$GlobalPhase.applyPhase(Global.scala:454)
[error] scala.tools.nsc.Global$GlobalPhase.run(Global.scala:402)
[error] scala.tools.nsc.Global$Run.compileUnitsInternal(Global.scala:1511)
[error] scala.tools.nsc.Global$Run.compileUnits(Global.scala:1495)
[error] scala.tools.nsc.Global$Run.compileSources(Global.scala:1488)
[error] scala.tools.nsc.Global$Run.compileFiles(Global.scala:1596)
[error] xsbt.CachedCompiler0.run(CompilerBridge.scala:163)
[error] xsbt.CachedCompiler0.run(CompilerBridge.scala:134)
[error] xsbt.CompilerBridge.run(CompilerBridge.scala:39)
[error] sbt.internal.inc.AnalyzingCompiler.compile(AnalyzingCompiler.scala:92)
[error] sbt.internal.inc.MixedAnalyzingCompiler.$anonfun$compile$7(MixedAnalyzingCompiler.scala:186)
[error] scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
[error] sbt.internal.inc.MixedAnalyzingCompiler.timed(MixedAnalyzingCompiler.scala:241)
[error] sbt.internal.inc.MixedAnalyzingCompiler.$anonfun$compile$4(MixedAnalyzingCompiler.scala:176)
[error] sbt.internal.inc.MixedAnalyzingCompiler.$anonfun$compile$4$adapted(MixedAnalyzingCompiler.scala:157)
[error] sbt.internal.inc.JarUtils$.withPreviousJar(JarUtils.scala:239)
[error] sbt.internal.inc.MixedAnalyzingCompiler.compileScala$1(MixedAnalyzingCompiler.scala:157)
[error] sbt.internal.inc.MixedAnalyzingCompiler.compile(MixedAnalyzingCompiler.scala:204)
[error] sbt.internal.inc.IncrementalCompilerImpl.$anonfun$compileInternal$1(IncrementalCompilerImpl.scala:573)
[error] sbt.internal.inc.IncrementalCompilerImpl.$anonfun$compileInternal$1$adapted(IncrementalCompilerImpl.scala:573)
[error] sbt.internal.inc.Incremental$.$anonfun$apply$5(Incremental.scala:173)
[error] sbt.internal.inc.Incremental$.$anonfun$apply$5$adapted(Incremental.scala:171)
[error] sbt.internal.inc.Incremental$$anon$2.run(Incremental.scala:458)
[error] sbt.internal.inc.IncrementalCommon$CycleState.next(IncrementalCommon.scala:116)
[error] sbt.internal.inc.IncrementalCommon$$anon$1.next(IncrementalCommon.scala:56)
[error] sbt.internal.inc.IncrementalCommon$$anon$1.next(IncrementalCommon.scala:52)
[error] sbt.internal.inc.IncrementalCommon.cycle(IncrementalCommon.scala:261)
[error] sbt.internal.inc.Incremental$.$anonfun$incrementalCompile$8(Incremental.scala:413)
[error] sbt.internal.inc.Incremental$.withClassfileManager(Incremental.scala:498)
[error] sbt.internal.inc.Incremental$.incrementalCompile(Incremental.scala:400)
[error] sbt.internal.inc.Incremental$.apply(Incremental.scala:165)
[error] sbt.internal.inc.IncrementalCompilerImpl.compileInternal(IncrementalCompilerImpl.scala:573)
[error] sbt.internal.inc.IncrementalCompilerImpl.$anonfun$compileIncrementally$1(IncrementalCompilerImpl.scala:491)
[error] sbt.internal.inc.IncrementalCompilerImpl.handleCompilationError(IncrementalCompilerImpl.scala:332)
[error] sbt.internal.inc.IncrementalCompilerImpl.compileIncrementally(IncrementalCompilerImpl.scala:420)
[error] sbt.internal.inc.IncrementalCompilerImpl.compile(IncrementalCompilerImpl.scala:137)
[error] sbt.Defaults$.compileIncrementalTaskImpl(Defaults.scala:2176)
[error] sbt.Defaults$.$anonfun$compileIncrementalTask$2(Defaults.scala:2133)
[error] sbt.internal.io.Retry$.apply(Retry.scala:40)
[error] sbt.internal.io.Retry$.apply(Retry.scala:23)
[error] sbt.internal.server.BspCompileTask$.compute(BspCompileTask.scala:31)
[error] sbt.Defaults$.$anonfun$compileIncrementalTask$1(Defaults.scala:2129)
[error] scala.Function1.$anonfun$compose$1(Function1.scala:49)
[error] sbt.internal.util.$tilde$greater.$anonfun$$u2219$1(TypeFunctions.scala:62)
[error] sbt.std.Transform$$anon$4.work(Transform.scala:68)
[error] sbt.Execute.$anonfun$submit$2(Execute.scala:282)
[error] sbt.internal.util.ErrorHandling$.wideConvert(ErrorHandling.scala:23)
[error] sbt.Execute.work(Execute.scala:291)
[error] sbt.Execute.$anonfun$submit$1(Execute.scala:282)
[error] sbt.ConcurrentRestrictions$$anon$4.$anonfun$submitValid$1(ConcurrentRestrictions.scala:265)
[error] sbt.CompletionService$$anon$2.call(CompletionService.scala:64)
[error] java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
[error] java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
[error] java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
[error] java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
[error] java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
[error] java.base/java.lang.Thread.run(Thread.java:834)
[error]
[error] stack trace is suppressed; run last hydra-backend / Compile / compileIncremental for the full output
[error] (hydra-backend / Compile / compileIncremental) spray.json.DeserializationException: Object is missing required member 'typeSymbol'
[error] Total time: 111 s (01:51), completed 21 Jun 2021, 13:22:15
``` | priority | spray json deserializationexception object is missing required member typesymbol under scala exception when compiling sources to home plipski hydra backend services target scala classes spray json deserializationexception object is missing required member typesymbol spray json package deserializationerror package scala spray json productformats fromfield productformats scala spray json productformats fromfield productformats scala org virtuslab ash writer eventschemawriter fromfield eventschemawriter scala spray json productformatsinstances anon read productformatsinstances scala spray json productformatsinstances anon read productformatsinstances scala spray json jsvalue convertto jsvalue scala org virtuslab ash writer eventschemawriter anonfun lastdump eventschemawriter scala scala collection iterator anon next iterator scala scala collection iterator anon next iterator scala scala collection iterator foreach iterator scala scala collection iterator foreach iterator scala scala collection abstractiterator foreach iterator scala scala collection generic growable plus plus eq growable scala scala collection generic growable plus plus eq growable scala scala collection immutable map mapbuilderimpl plus plus eq map scala scala collection immutable map mapbuilderimpl plus plus eq map scala scala collection traversableonce tomap traversableonce scala scala collection traversableonce tomap traversableonce scala scala collection abstractiterator tomap iterator scala org virtuslab ash writer eventschemawriter lastdump lzycompute eventschemawriter scala org virtuslab ash writer eventschemawriter lastdump eventschemawriter scala org virtuslab ash dumpeventschemacompilerplugincomponent anon anonfun isdefinedat dumpeventschemacompilerplugincomponent scala org virtuslab ash dumpeventschemacompilerplugincomponent anon anonfun isdefinedat dumpeventschemacompilerplugincomponent scala scala reflect internal trees collecttreetraverser traverse trees scala scala reflect internal trees collecttreetraverser traverse trees scala scala reflect api trees traverser anonfun traversestats trees scala scala reflect api trees traverser atowner trees scala scala reflect api trees traverser anonfun traversestats trees scala scala reflect api trees traverser traversestats trees scala scala reflect internal trees itraverse trees scala scala reflect internal trees itraverse trees scala scala reflect internal symboltable itraverse symboltable scala scala reflect internal symboltable itraverse symboltable scala scala reflect api trees traverser traverse trees scala scala reflect internal trees collecttreetraverser traverse trees scala scala reflect internal trees treecontextapiimpl collect trees scala org virtuslab ash dumpeventschemacompilerplugincomponent anon apply dumpeventschemacompilerplugincomponent scala scala tools nsc global globalphase applyphase global scala scala tools nsc global globalphase run global scala scala tools nsc global run compileunitsinternal global scala scala tools nsc global run compileunits global scala scala tools nsc global run compilesources global scala scala tools nsc global run compilefiles global scala xsbt run compilerbridge scala xsbt run compilerbridge scala xsbt compilerbridge run compilerbridge scala sbt internal inc analyzingcompiler compile analyzingcompiler scala sbt internal inc mixedanalyzingcompiler anonfun compile mixedanalyzingcompiler scala scala runtime mcv sp apply mcv sp java sbt internal inc mixedanalyzingcompiler timed mixedanalyzingcompiler scala sbt internal inc mixedanalyzingcompiler anonfun compile mixedanalyzingcompiler scala sbt internal inc mixedanalyzingcompiler anonfun compile adapted mixedanalyzingcompiler scala sbt internal inc jarutils withpreviousjar jarutils scala sbt internal inc mixedanalyzingcompiler compilescala mixedanalyzingcompiler scala sbt internal inc mixedanalyzingcompiler compile mixedanalyzingcompiler scala sbt internal inc incrementalcompilerimpl anonfun compileinternal incrementalcompilerimpl scala sbt internal inc incrementalcompilerimpl anonfun compileinternal adapted incrementalcompilerimpl scala sbt internal inc incremental anonfun apply incremental scala sbt internal inc incremental anonfun apply adapted incremental scala sbt internal inc incremental anon run incremental scala sbt internal inc incrementalcommon cyclestate next incrementalcommon scala sbt internal inc incrementalcommon anon next incrementalcommon scala sbt internal inc incrementalcommon anon next incrementalcommon scala sbt internal inc incrementalcommon cycle incrementalcommon scala sbt internal inc incremental anonfun incrementalcompile incremental scala sbt internal inc incremental withclassfilemanager incremental scala sbt internal inc incremental incrementalcompile incremental scala sbt internal inc incremental apply incremental scala sbt internal inc incrementalcompilerimpl compileinternal incrementalcompilerimpl scala sbt internal inc incrementalcompilerimpl anonfun compileincrementally incrementalcompilerimpl scala sbt internal inc incrementalcompilerimpl handlecompilationerror incrementalcompilerimpl scala sbt internal inc incrementalcompilerimpl compileincrementally incrementalcompilerimpl scala sbt internal inc incrementalcompilerimpl compile incrementalcompilerimpl scala sbt defaults compileincrementaltaskimpl defaults scala sbt defaults anonfun compileincrementaltask defaults scala sbt internal io retry apply retry scala sbt internal io retry apply retry scala sbt internal server bspcompiletask compute bspcompiletask scala sbt defaults anonfun compileincrementaltask defaults scala scala anonfun compose scala sbt internal util tilde greater anonfun typefunctions scala sbt std transform anon work transform scala sbt execute anonfun submit execute scala sbt internal util errorhandling wideconvert errorhandling scala sbt execute work execute scala sbt execute anonfun submit execute scala sbt concurrentrestrictions anon anonfun submitvalid concurrentrestrictions scala sbt completionservice anon call completionservice scala java base java util concurrent futuretask run futuretask java java base java util concurrent executors runnableadapter call executors java java base java util concurrent futuretask run futuretask java java base java util concurrent threadpoolexecutor runworker threadpoolexecutor java java base java util concurrent threadpoolexecutor worker run threadpoolexecutor java java base java lang thread run thread java stack trace is suppressed run last hydra backend compile compileincremental for the full output hydra backend compile compileincremental spray json deserializationexception object is missing required member typesymbol total time s completed jun | 1 |
692,359 | 23,731,382,830 | IssuesEvent | 2022-08-31 02:15:51 | EscherLabs/Graphene | https://api.github.com/repos/EscherLabs/Graphene | closed | GrapheneAPIGateway: Add Form-based Options to API/API Instance | enhancement high priority Upcoming Release API Gateway | Currently, whenever we want to configure / customize an API Instance, we are required to set up custom resources, and then map those back to an API. While that is correct in some situations (example: databases), it is not really correct for certain API Specific controls (turning functions on / off, storing simple configuration strings, etc). In those situations, we are arguably abusing the "resources" function, which has led to dozens of single-use resources that only have application with a very specific API (which was not the original intention).
Rather, it should be possible to have a simple "options" form, which is defined as part of the API. This should be nearly identical to the "options" form which is configured for MicroApps.
From within the API editor, I expect that there would be a new tab called "Options", which provides access to the GrapheneForm Builder, and generates a new "options" form.
From the API Instance viewer, I expect that there would be a new tab called "Options", where an admin can configure the various options which are available.
These options should then be available via an $options variable, which is accessible from within the API Code. (In addition to the existing $args, and $resources variables).
Note that this will require some modifications to the database for managing and storing these additional data fields. | 1.0 | GrapheneAPIGateway: Add Form-based Options to API/API Instance - Currently, whenever we want to configure / customize an API Instance, we are required to set up custom resources, and then map those back to an API. While that is correct in some situations (example: databases), it is not really correct for certain API Specific controls (turning functions on / off, storing simple configuration strings, etc). In those situations, we are arguably abusing the "resources" function, which has led to dozens of single-use resources that only have application with a very specific API (which was not the original intention).
Rather, it should be possible to have a simple "options" form, which is defined as part of the API. This should be nearly identical to the "options" form which is configured for MicroApps.
From within the API editor, I expect that there would be a new tab called "Options", which provides access to the GrapheneForm Builder, and generates a new "options" form.
From the API Instance viewer, I expect that there would be a new tab called "Options", where an admin can configure the various options which are available.
These options should then be available via an $options variable, which is accessible from within the API Code. (In addition to the existing $args, and $resources variables).
Note that this will require some modifications to the database for managing and storing these additional data fields. | priority | grapheneapigateway add form based options to api api instance currently whenever we want to configure customize an api instance we are required to set up custom resources and then map those back to an api while that is correct in some situations example databases it is not really correct for certain api specific controls turning functions on off storing simple configuration strings etc in those situations we are arguably abusing the resources function which has led to dozens of single use resources that only have application with a very specific api which was not the original intention rather it should be possible to have a simple options form which is defined as part of the api this should be nearly identical to the options form which is configured for microapps from within the api editor i expect that there would be a new tab called options which provides access to the grapheneform builder and generates a new options form from the api instance viewer i expect that there would be a new tab called options where an admin can configure the various options which are available these options should then be available via an options variable which is accessible from within the api code in addition to the existing args and resources variables note that this will require some modifications to the database for managing and storing these additional data fields | 1 |
144,280 | 5,537,622,256 | IssuesEvent | 2017-03-21 22:37:35 | fossology/fossology | https://api.github.com/repos/fossology/fossology | closed | Scheduler not stable when more than one schedulers running at the same time | bug Category: Scheduler Priority: High Status: New Tracker: Bug | ---
Author Name: **Dong Ma**
Original Redmine Issue: 6981, http://www.fossology.org/issues/6981
Original Date: 2014/05/07
Original Assignee: larry shi
---
The root cause of bug#5601 is scheduler not stable when more than one schedulers running at the same time, create a new bug related scheduler issue.
| 1.0 | Scheduler not stable when more than one schedulers running at the same time - ---
Author Name: **Dong Ma**
Original Redmine Issue: 6981, http://www.fossology.org/issues/6981
Original Date: 2014/05/07
Original Assignee: larry shi
---
The root cause of bug#5601 is scheduler not stable when more than one schedulers running at the same time, create a new bug related scheduler issue.
| priority | scheduler not stable when more than one schedulers running at the same time author name dong ma original redmine issue original date original assignee larry shi the root cause of bug is scheduler not stable when more than one schedulers running at the same time create a new bug related scheduler issue | 1 |
322,246 | 9,815,228,243 | IssuesEvent | 2019-06-13 12:10:24 | thewca/worldcubeassociation.org | https://api.github.com/repos/thewca/worldcubeassociation.org | closed | Add permissions for WEC | high-priority | From Ethan Pride via email:
>Dear WST,
>
>In discussion with the board we have requested the following permissions for WEC members to aid with current and future investigations:
1: Have the ability to view Delegate reports on the website
2: Have the ability to see the registration lists for competitions using the wca registration system as well as payment status
>
>Could you please advise the board which permissions are required to grant this access?
The first one is easy: [can_view_delegate_report?](https://github.com/thewca/worldcubeassociation.org/blob/master/WcaOnRails/app/models/user.rb#L654)
As for second one, I'm not sure. There is https://github.com/thewca/worldcubeassociation.org/blob/master/WcaOnRails/app/models/user.rb#L654 and https://github.com/thewca/worldcubeassociation.org/blob/master/WcaOnRails/app/models/user.rb#L654 | 1.0 | Add permissions for WEC - From Ethan Pride via email:
>Dear WST,
>
>In discussion with the board we have requested the following permissions for WEC members to aid with current and future investigations:
1: Have the ability to view Delegate reports on the website
2: Have the ability to see the registration lists for competitions using the wca registration system as well as payment status
>
>Could you please advise the board which permissions are required to grant this access?
The first one is easy: [can_view_delegate_report?](https://github.com/thewca/worldcubeassociation.org/blob/master/WcaOnRails/app/models/user.rb#L654)
As for second one, I'm not sure. There is https://github.com/thewca/worldcubeassociation.org/blob/master/WcaOnRails/app/models/user.rb#L654 and https://github.com/thewca/worldcubeassociation.org/blob/master/WcaOnRails/app/models/user.rb#L654 | priority | add permissions for wec from ethan pride via email dear wst in discussion with the board we have requested the following permissions for wec members to aid with current and future investigations have the ability to view delegate reports on the website have the ability to see the registration lists for competitions using the wca registration system as well as payment status could you please advise the board which permissions are required to grant this access the first one is easy as for second one i m not sure there is and | 1 |
101,679 | 4,128,217,307 | IssuesEvent | 2016-06-10 04:28:30 | damlaren/ogle | https://api.github.com/repos/damlaren/ogle | opened | Figure out how to update camera aspect ratio | priority:very high | This is something that should be done behind the scenes, not in an application. Suggests that more behavior is needed for components to do their job properly. | 1.0 | Figure out how to update camera aspect ratio - This is something that should be done behind the scenes, not in an application. Suggests that more behavior is needed for components to do their job properly. | priority | figure out how to update camera aspect ratio this is something that should be done behind the scenes not in an application suggests that more behavior is needed for components to do their job properly | 1 |
86,581 | 3,727,272,613 | IssuesEvent | 2016-03-06 05:41:17 | fit360/spott | https://api.github.com/repos/fit360/spott | closed | Models: Activities | backend feature priority:high v1 | Enums/Constants:
pick up references from sweatt.
```
Icon
Name
``` | 1.0 | Models: Activities - Enums/Constants:
pick up references from sweatt.
```
Icon
Name
``` | priority | models activities enums constants pick up references from sweatt icon name | 1 |
154,308 | 5,917,374,862 | IssuesEvent | 2017-05-22 13:07:28 | metasfresh/metasfresh-webui-api | https://api.github.com/repos/metasfresh/metasfresh-webui-api | closed | Document changes: provide the validStatus and saveStatus only when changed | priority:high status:integrated type:bug | ### Is this a bug or feature request?
Bug
### What is the current behavior?
The validStatus and saveStatus are provided almost all the time.
In some cases, because of that, the document changes list that is provided back to frontend is huge and the frontend on it's turn freeze chrome for a while (i.e. you are not able to close the browser tab).
#### Which are the steps to reproduce?
Try editing a role, e.g. https://w101.metasfresh.com:8443/window/111/1000055.
Try changing the name, or try activating/deactivating a window access or a task access or something.
Your browser tab will freeze.
Wait a (long) while and the chrome tab will work again.
If you inspect the response of the endpoint which was called to change the document, you will see that it has a huge amount of changes, almost all of them are only about valid status and save status.
### What is the expected or desired behavior?
Backend shall keep the response as small as possible.
It shall send only the valid and save statuses which really changed.
NOTE: might be that we also have to rethink when and how the valid and save status are changed.
| 1.0 | Document changes: provide the validStatus and saveStatus only when changed - ### Is this a bug or feature request?
Bug
### What is the current behavior?
The validStatus and saveStatus are provided almost all the time.
In some cases, because of that, the document changes list that is provided back to frontend is huge and the frontend on it's turn freeze chrome for a while (i.e. you are not able to close the browser tab).
#### Which are the steps to reproduce?
Try editing a role, e.g. https://w101.metasfresh.com:8443/window/111/1000055.
Try changing the name, or try activating/deactivating a window access or a task access or something.
Your browser tab will freeze.
Wait a (long) while and the chrome tab will work again.
If you inspect the response of the endpoint which was called to change the document, you will see that it has a huge amount of changes, almost all of them are only about valid status and save status.
### What is the expected or desired behavior?
Backend shall keep the response as small as possible.
It shall send only the valid and save statuses which really changed.
NOTE: might be that we also have to rethink when and how the valid and save status are changed.
| priority | document changes provide the validstatus and savestatus only when changed is this a bug or feature request bug what is the current behavior the validstatus and savestatus are provided almost all the time in some cases because of that the document changes list that is provided back to frontend is huge and the frontend on it s turn freeze chrome for a while i e you are not able to close the browser tab which are the steps to reproduce try editing a role e g try changing the name or try activating deactivating a window access or a task access or something your browser tab will freeze wait a long while and the chrome tab will work again if you inspect the response of the endpoint which was called to change the document you will see that it has a huge amount of changes almost all of them are only about valid status and save status what is the expected or desired behavior backend shall keep the response as small as possible it shall send only the valid and save statuses which really changed note might be that we also have to rethink when and how the valid and save status are changed | 1 |
83,465 | 3,635,221,007 | IssuesEvent | 2016-02-11 20:57:50 | kkennethlee/hacka-tictactoe-athon | https://api.github.com/repos/kkennethlee/hacka-tictactoe-athon | closed | create the css for the game (main game area) | High Priority | - gameboard
- cells
- different player contents (colors, X, O's) | 1.0 | create the css for the game (main game area) - - gameboard
- cells
- different player contents (colors, X, O's) | priority | create the css for the game main game area gameboard cells different player contents colors x o s | 1 |
639,436 | 20,753,751,360 | IssuesEvent | 2022-03-15 10:12:18 | banco-alimentar/alimentestaideia.pt | https://api.github.com/repos/banco-alimentar/alimentestaideia.pt | opened | Improve message in Email multibanco reminder | enhancement High.Priority | **Is your feature request related to a problem? Please describe.**
The email reminder, for lack of payment of multibanco is very generic and should include the payment indications, value, reference, etc. currently it is generic:
" Banco Alimentar: Doação por multibanco ainda não concretizada"
| 1.0 | Improve message in Email multibanco reminder - **Is your feature request related to a problem? Please describe.**
The email reminder, for lack of payment of multibanco is very generic and should include the payment indications, value, reference, etc. currently it is generic:
" Banco Alimentar: Doação por multibanco ainda não concretizada"
| priority | improve message in email multibanco reminder is your feature request related to a problem please describe the email reminder for lack of payment of multibanco is very generic and should include the payment indications value reference etc currently it is generic banco alimentar doação por multibanco ainda não concretizada | 1 |
252,653 | 8,038,627,394 | IssuesEvent | 2018-07-30 15:52:58 | curationexperts/laevigata | https://api.github.com/repos/curationexperts/laevigata | closed | Respect workflow field indicating hidden visibility in display | embargo high priority release priority review | - [x] Hidden ETDs don't display unless you are a superuser or workflow approver
- [x] When viewing a hidden work, you see a visual indication that the work is hidden | 2.0 | Respect workflow field indicating hidden visibility in display - - [x] Hidden ETDs don't display unless you are a superuser or workflow approver
- [x] When viewing a hidden work, you see a visual indication that the work is hidden | priority | respect workflow field indicating hidden visibility in display hidden etds don t display unless you are a superuser or workflow approver when viewing a hidden work you see a visual indication that the work is hidden | 1 |
278,898 | 8,652,129,346 | IssuesEvent | 2018-11-27 06:48:53 | iSosnitsky/DschinghisKhan | https://api.github.com/repos/iSosnitsky/DschinghisKhan | closed | Копирование машрутов | High priority | Маршруты тоже надо копировать (со всеми данными, включая их точки, но к названию добавлять дату копирования без точек в формате 18102018)
| 1.0 | Копирование машрутов - Маршруты тоже надо копировать (со всеми данными, включая их точки, но к названию добавлять дату копирования без точек в формате 18102018)
| priority | копирование машрутов маршруты тоже надо копировать со всеми данными включая их точки но к названию добавлять дату копирования без точек в формате | 1 |
366,610 | 10,824,676,760 | IssuesEvent | 2019-11-09 10:59:25 | crux-bphc/CMS-Android | https://api.github.com/repos/crux-bphc/CMS-Android | closed | NullPointerException at ForumFragment.onFailure() | bug easy pick high_priority | **Corresponds to version:** 7
**Relevant line in log:**
`at crux.bphc.cms.fragments.ForumFragment$4.onFailure (ForumFragment.java:233)`
**Full log:**
```
java.lang.NullPointerException:
at android.widget.Toast.<init> (Toast.java:103)
at android.widget.Toast.makeText (Toast.java:256)
at crux.bphc.cms.fragments.ForumFragment$4.onFailure (ForumFragment.java:233)
at retrofit2.ExecutorCallAdapterFactory$ExecutorCallbackCall$1$2.run (ExecutorCallAdapterFactory.java:79)
at android.os.Handler.handleCallback (Handler.java:754)
at android.os.Handler.dispatchMessage (Handler.java:95)
at android.os.Looper.loop (Looper.java:163)
at android.app.ActivityThread.main (ActivityThread.java:6228)
at java.lang.reflect.Method.invoke (Native Method)
at com.android.internal.os.ZygoteInit$MethodAndArgsCaller.run (ZygoteInit.java:904)
at com.android.internal.os.ZygoteInit.main (ZygoteInit.java:794)
``` | 1.0 | NullPointerException at ForumFragment.onFailure() - **Corresponds to version:** 7
**Relevant line in log:**
`at crux.bphc.cms.fragments.ForumFragment$4.onFailure (ForumFragment.java:233)`
**Full log:**
```
java.lang.NullPointerException:
at android.widget.Toast.<init> (Toast.java:103)
at android.widget.Toast.makeText (Toast.java:256)
at crux.bphc.cms.fragments.ForumFragment$4.onFailure (ForumFragment.java:233)
at retrofit2.ExecutorCallAdapterFactory$ExecutorCallbackCall$1$2.run (ExecutorCallAdapterFactory.java:79)
at android.os.Handler.handleCallback (Handler.java:754)
at android.os.Handler.dispatchMessage (Handler.java:95)
at android.os.Looper.loop (Looper.java:163)
at android.app.ActivityThread.main (ActivityThread.java:6228)
at java.lang.reflect.Method.invoke (Native Method)
at com.android.internal.os.ZygoteInit$MethodAndArgsCaller.run (ZygoteInit.java:904)
at com.android.internal.os.ZygoteInit.main (ZygoteInit.java:794)
``` | priority | nullpointerexception at forumfragment onfailure corresponds to version relevant line in log at crux bphc cms fragments forumfragment onfailure forumfragment java full log java lang nullpointerexception at android widget toast toast java at android widget toast maketext toast java at crux bphc cms fragments forumfragment onfailure forumfragment java at executorcalladapterfactory executorcallbackcall run executorcalladapterfactory java at android os handler handlecallback handler java at android os handler dispatchmessage handler java at android os looper loop looper java at android app activitythread main activitythread java at java lang reflect method invoke native method at com android internal os zygoteinit methodandargscaller run zygoteinit java at com android internal os zygoteinit main zygoteinit java | 1 |
715,230 | 24,591,569,847 | IssuesEvent | 2022-10-14 03:05:03 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | opened | Resnet classifier sample returns incorrect results on MPS | high priority module: mps | ### 🐛 Describe the bug
Consider the following example:
```python
from torchvision.io import read_image
from torchvision.models import resnet50, ResNet50_Weights
img = read_image("dog2.jpg").to("mps")
weights = ResNet50_Weights.DEFAULT
model = resnet50(weights=weights).to("mps")
model.eval()
preprocess = weights.transforms()
batch = preprocess(img).unsqueeze(0)
prediction = model(batch).squeeze(0).softmax(0)
class_id = prediction.argmax().item()
score = prediction[class_id].item()
```
When run on MPS following image is classified as necklace:

### Versions
Nightly/1.13.0 | 1.0 | Resnet classifier sample returns incorrect results on MPS - ### 🐛 Describe the bug
Consider the following example:
```python
from torchvision.io import read_image
from torchvision.models import resnet50, ResNet50_Weights
img = read_image("dog2.jpg").to("mps")
weights = ResNet50_Weights.DEFAULT
model = resnet50(weights=weights).to("mps")
model.eval()
preprocess = weights.transforms()
batch = preprocess(img).unsqueeze(0)
prediction = model(batch).squeeze(0).softmax(0)
class_id = prediction.argmax().item()
score = prediction[class_id].item()
```
When run on MPS following image is classified as necklace:

### Versions
Nightly/1.13.0 | priority | resnet classifier sample returns incorrect results on mps 🐛 describe the bug consider the following example python from torchvision io import read image from torchvision models import weights img read image jpg to mps weights weights default model weights weights to mps model eval preprocess weights transforms batch preprocess img unsqueeze prediction model batch squeeze softmax class id prediction argmax item score prediction item when run on mps following image is classified as necklace versions nightly | 1 |
559,022 | 16,548,084,050 | IssuesEvent | 2021-05-28 04:12:09 | SerpentDrago/skin.auramod | https://api.github.com/repos/SerpentDrago/skin.auramod | closed | TV Shows Network Logos don’t show up on next episodes or progress episodes but they do work on home screen widgets | High Priority bug | ERROR: type should be string, got "\r\n\r\n\r\nhttps://user-images.githubusercontent.com/49699869/119732321-0709c100-be46-11eb-89e3-69c51d40c3e6.MOV\r\n\r\n\r\nThis issue is a little hard for me to explain hope I make sense but the TV shows network logos on progress or next up episodes only work on home screen widgets but if I enter progress or next up episodes within the add-on The TV show network logos don’t show up but if I go to context menu browse series and come back out the TV show network logos show up on next up episodes or progress I made a short video to try to explain a little better not a big deal they work everywhere else like they’re supposed to, seems to only happen on next up or progress,, just trying to figure out if it’s on my end,, using Nvidia shield pro on the latest version of AuraMod 1.0.11 ,, Fen and Venom\r\n\r\n" | 1.0 | TV Shows Network Logos don’t show up on next episodes or progress episodes but they do work on home screen widgets -
https://user-images.githubusercontent.com/49699869/119732321-0709c100-be46-11eb-89e3-69c51d40c3e6.MOV
This issue is a little hard for me to explain hope I make sense but the TV shows network logos on progress or next up episodes only work on home screen widgets but if I enter progress or next up episodes within the add-on The TV show network logos don’t show up but if I go to context menu browse series and come back out the TV show network logos show up on next up episodes or progress I made a short video to try to explain a little better not a big deal they work everywhere else like they’re supposed to, seems to only happen on next up or progress,, just trying to figure out if it’s on my end,, using Nvidia shield pro on the latest version of AuraMod 1.0.11 ,, Fen and Venom
| priority | tv shows network logos don’t show up on next episodes or progress episodes but they do work on home screen widgets this issue is a little hard for me to explain hope i make sense but the tv shows network logos on progress or next up episodes only work on home screen widgets but if i enter progress or next up episodes within the add on the tv show network logos don’t show up but if i go to context menu browse series and come back out the tv show network logos show up on next up episodes or progress i made a short video to try to explain a little better not a big deal they work everywhere else like they’re supposed to seems to only happen on next up or progress just trying to figure out if it’s on my end using nvidia shield pro on the latest version of auramod fen and venom | 1 |
166,225 | 6,300,186,930 | IssuesEvent | 2017-07-21 02:27:52 | PacketCloud/ALVIS | https://api.github.com/repos/PacketCloud/ALVIS | opened | Do not allow uploading of audio files that go against naming convention. | Bug High Priority | Rather disallow users from uploading audio files with spaces or periods (or other breaking characters), or else remove these characters from the filename when saving. | 1.0 | Do not allow uploading of audio files that go against naming convention. - Rather disallow users from uploading audio files with spaces or periods (or other breaking characters), or else remove these characters from the filename when saving. | priority | do not allow uploading of audio files that go against naming convention rather disallow users from uploading audio files with spaces or periods or other breaking characters or else remove these characters from the filename when saving | 1 |
639,404 | 20,752,789,958 | IssuesEvent | 2022-03-15 09:21:25 | SE701-T1/backend | https://api.github.com/repos/SE701-T1/backend | closed | Timetable File Processing | Priority: High Status: In Progress Type: Feature Team: Timetable | **Describe the task that needs to be done.**
Work out how to take in timetable file (only .ics) and process it to get specific courses (and any other information).
Need to look into the file format and options for upload.
**Describe how a solution to your proposed task might look like (and any alternatives considered).**
.ics file can be read and relevant information stored in database seperately
**Notes**
Dependent on database solution being decided and framework implemented - #13 | 1.0 | Timetable File Processing - **Describe the task that needs to be done.**
Work out how to take in timetable file (only .ics) and process it to get specific courses (and any other information).
Need to look into the file format and options for upload.
**Describe how a solution to your proposed task might look like (and any alternatives considered).**
.ics file can be read and relevant information stored in database seperately
**Notes**
Dependent on database solution being decided and framework implemented - #13 | priority | timetable file processing describe the task that needs to be done work out how to take in timetable file only ics and process it to get specific courses and any other information need to look into the file format and options for upload describe how a solution to your proposed task might look like and any alternatives considered ics file can be read and relevant information stored in database seperately notes dependent on database solution being decided and framework implemented | 1 |
235,264 | 7,735,941,805 | IssuesEvent | 2018-05-27 20:37:27 | Loopring/loopr-ios | https://api.github.com/repos/Loopring/loopr-ios | closed | Face ID and Touch ID Authentication | high priority | The iOS app should have Face ID and Touch ID authentication. | 1.0 | Face ID and Touch ID Authentication - The iOS app should have Face ID and Touch ID authentication. | priority | face id and touch id authentication the ios app should have face id and touch id authentication | 1 |
237,314 | 7,758,469,758 | IssuesEvent | 2018-05-31 19:42:26 | teambit/bit | https://api.github.com/repos/teambit/bit | closed | missing breaklines in output | area/outputs priority/High type/bug | ## Expected Behavior
when listing missing packages/dependencies/etc, each list should have its own line
## Actual Behavior
<img width="1265" alt="screenshot 2018-02-13 12 58 48" src="https://user-images.githubusercontent.com/4623056/36146582-cf9ea3e0-10bd-11e8-9f2e-88ce08a68277.png">
## Steps to Reproduce the Problem
1. import component
1. modify component
1. run `bit status`
## Specifications
- Version: 0.12.5
- Platform: macOS
| 1.0 | missing breaklines in output - ## Expected Behavior
when listing missing packages/dependencies/etc, each list should have its own line
## Actual Behavior
<img width="1265" alt="screenshot 2018-02-13 12 58 48" src="https://user-images.githubusercontent.com/4623056/36146582-cf9ea3e0-10bd-11e8-9f2e-88ce08a68277.png">
## Steps to Reproduce the Problem
1. import component
1. modify component
1. run `bit status`
## Specifications
- Version: 0.12.5
- Platform: macOS
| priority | missing breaklines in output expected behavior when listing missing packages dependencies etc each list should have its own line actual behavior img width alt screenshot src steps to reproduce the problem import component modify component run bit status specifications version platform macos | 1 |
654,840 | 21,671,527,285 | IssuesEvent | 2022-05-08 02:40:30 | wilrnh/emergencywalletcards.com | https://api.github.com/repos/wilrnh/emergencywalletcards.com | closed | Add tooltips to form describing each field | type:enhancement priority:high status:stale | especially highligh freeform fields such as what can go in "information", etc. | 1.0 | Add tooltips to form describing each field - especially highligh freeform fields such as what can go in "information", etc. | priority | add tooltips to form describing each field especially highligh freeform fields such as what can go in information etc | 1 |
770,472 | 27,041,204,940 | IssuesEvent | 2023-02-13 05:34:11 | CMU-313/spring23-nodebb-team-dj-kew | https://api.github.com/repos/CMU-313/spring23-nodebb-team-dj-kew | closed | US2 - If is_anon is true, then remove avatar/pfp on post. | (P0) high priority medium effort | If is_anon field on the post is true, then the avatar/pfp of the poster's profile should be removed.
Effort: Medium
Depends on: https://github.com/CMU-313/spring23-nodebb-team-dj-kew/issues/1
Milestone: Sprint 1
Acceptance Criteria: Pass visual test, and tests that ensures the avatar is removed from the post when is_anon is true. | 1.0 | US2 - If is_anon is true, then remove avatar/pfp on post. - If is_anon field on the post is true, then the avatar/pfp of the poster's profile should be removed.
Effort: Medium
Depends on: https://github.com/CMU-313/spring23-nodebb-team-dj-kew/issues/1
Milestone: Sprint 1
Acceptance Criteria: Pass visual test, and tests that ensures the avatar is removed from the post when is_anon is true. | priority | if is anon is true then remove avatar pfp on post if is anon field on the post is true then the avatar pfp of the poster s profile should be removed effort medium depends on milestone sprint acceptance criteria pass visual test and tests that ensures the avatar is removed from the post when is anon is true | 1 |
151,585 | 5,824,502,236 | IssuesEvent | 2017-05-07 13:39:26 | NAT0P0TAT0/Genesis-modpack | https://api.github.com/repos/NAT0P0TAT0/Genesis-modpack | closed | trapdoor after beating serpent sometimes wont open while a pedestal is next to it | bug Priority: High Septic | is fine if you take the item, leave the room and return, but it can really screw you if it spawns an active item you don't want | 1.0 | trapdoor after beating serpent sometimes wont open while a pedestal is next to it - is fine if you take the item, leave the room and return, but it can really screw you if it spawns an active item you don't want | priority | trapdoor after beating serpent sometimes wont open while a pedestal is next to it is fine if you take the item leave the room and return but it can really screw you if it spawns an active item you don t want | 1 |
601,056 | 18,365,402,639 | IssuesEvent | 2021-10-10 00:35:41 | sglavoie/uol-grades-calculator-server | https://api.github.com/repos/sglavoie/uol-grades-calculator-server | closed | Allow users to upload their own grades file and remove local storage data as desired | Priority: high backend frontend | ## Acceptance
* [x] Users are first presented a screen to upload their grades config file if no data is found in the localStorage.
* [x] Users can access and use the existing "features" of the UI once their grades are loaded.
* [x] Users can remove all the local data by clicking on a link.
* [x] Users can upload new grades to overwrite existing data, being asked to confirmed before proceeding.
## Tasks
* [x] Implement a global state manager with Redux.
* [x] If no data is found in localStorage, display only a component to upload existing grades when loading the home page and prevent access to the other routes.
* [x] Add an `Upload existing grades` link to upload a file and cache the grades.
* [x] Validate grades file with `ugc` once uploaded before caching any data.
* [x] Add a `Clear grades data` link to remove all data from localStorage.
* [x] Check whether grades are already loaded before uploading a new file and ask the user whether they'd want to overwrite their existing grades.
## Analysis
* As it stands, the server is only useful if installed locally and launched with the ability to fetch an existing grades file. There is no reloading on the fly of new data nor is there the possibility to deploy somewhere else and be able to easily read existing grades without the grades file being present on the server.
* Using [localStorage](https://developer.mozilla.org/en-US/docs/Web/API/Window/localStorage) would be convenient to keep data intact on the website until the user cache is cleared.
* Having the ability to upload an existing config file is necessary since there's no way to build the config file from scratch from the UI at this moment. | 1.0 | Allow users to upload their own grades file and remove local storage data as desired - ## Acceptance
* [x] Users are first presented a screen to upload their grades config file if no data is found in the localStorage.
* [x] Users can access and use the existing "features" of the UI once their grades are loaded.
* [x] Users can remove all the local data by clicking on a link.
* [x] Users can upload new grades to overwrite existing data, being asked to confirmed before proceeding.
## Tasks
* [x] Implement a global state manager with Redux.
* [x] If no data is found in localStorage, display only a component to upload existing grades when loading the home page and prevent access to the other routes.
* [x] Add an `Upload existing grades` link to upload a file and cache the grades.
* [x] Validate grades file with `ugc` once uploaded before caching any data.
* [x] Add a `Clear grades data` link to remove all data from localStorage.
* [x] Check whether grades are already loaded before uploading a new file and ask the user whether they'd want to overwrite their existing grades.
## Analysis
* As it stands, the server is only useful if installed locally and launched with the ability to fetch an existing grades file. There is no reloading on the fly of new data nor is there the possibility to deploy somewhere else and be able to easily read existing grades without the grades file being present on the server.
* Using [localStorage](https://developer.mozilla.org/en-US/docs/Web/API/Window/localStorage) would be convenient to keep data intact on the website until the user cache is cleared.
* Having the ability to upload an existing config file is necessary since there's no way to build the config file from scratch from the UI at this moment. | priority | allow users to upload their own grades file and remove local storage data as desired acceptance users are first presented a screen to upload their grades config file if no data is found in the localstorage users can access and use the existing features of the ui once their grades are loaded users can remove all the local data by clicking on a link users can upload new grades to overwrite existing data being asked to confirmed before proceeding tasks implement a global state manager with redux if no data is found in localstorage display only a component to upload existing grades when loading the home page and prevent access to the other routes add an upload existing grades link to upload a file and cache the grades validate grades file with ugc once uploaded before caching any data add a clear grades data link to remove all data from localstorage check whether grades are already loaded before uploading a new file and ask the user whether they d want to overwrite their existing grades analysis as it stands the server is only useful if installed locally and launched with the ability to fetch an existing grades file there is no reloading on the fly of new data nor is there the possibility to deploy somewhere else and be able to easily read existing grades without the grades file being present on the server using would be convenient to keep data intact on the website until the user cache is cleared having the ability to upload an existing config file is necessary since there s no way to build the config file from scratch from the ui at this moment | 1 |
211,641 | 7,203,163,054 | IssuesEvent | 2018-02-06 08:08:10 | commons-app/apps-android-commons | https://api.github.com/repos/commons-app/apps-android-commons | closed | Logging in to app with two-factor authentication | IEG enhancement high priority | Comment on our review page:
> I can't login after enabling 2FA I've activated the Two Factor Authentication and the app shows "Incorrect login"
Has anyone tried this before? Also, is there a need to support 2FA in our app? | 1.0 | Logging in to app with two-factor authentication - Comment on our review page:
> I can't login after enabling 2FA I've activated the Two Factor Authentication and the app shows "Incorrect login"
Has anyone tried this before? Also, is there a need to support 2FA in our app? | priority | logging in to app with two factor authentication comment on our review page i can t login after enabling i ve activated the two factor authentication and the app shows incorrect login has anyone tried this before also is there a need to support in our app | 1 |
87,444 | 3,754,758,397 | IssuesEvent | 2016-03-12 06:07:36 | cs2103jan2016-f13-4j/main | https://api.github.com/repos/cs2103jan2016-f13-4j/main | closed | The user can get list of task sorted by priority | priority.high type.story | so that he/she can check how urgent his/her task is | 1.0 | The user can get list of task sorted by priority - so that he/she can check how urgent his/her task is | priority | the user can get list of task sorted by priority so that he she can check how urgent his her task is | 1 |
473,643 | 13,645,367,184 | IssuesEvent | 2020-09-25 20:39:00 | HackYourFuture-CPH/chattie | https://api.github.com/repos/HackYourFuture-CPH/chattie | closed | Backend: Getting channels with specific members | High priority User story | ## User story
**Who:** **As a** developer
**What:** **I want to** figure out which channels has specific user
## Implementation details
- Should be used like this: `/api/channels?memberIds=1,3,4` Will return the channels where users with id 1,3 and 4 are members
| 1.0 | Backend: Getting channels with specific members - ## User story
**Who:** **As a** developer
**What:** **I want to** figure out which channels has specific user
## Implementation details
- Should be used like this: `/api/channels?memberIds=1,3,4` Will return the channels where users with id 1,3 and 4 are members
| priority | backend getting channels with specific members user story who as a developer what i want to figure out which channels has specific user implementation details should be used like this api channels memberids will return the channels where users with id and are members | 1 |
462,660 | 13,251,030,875 | IssuesEvent | 2020-08-20 00:53:30 | qgis/QGIS | https://api.github.com/repos/qgis/QGIS | closed | If any of the "topology checks" options is enabled, then cliking on "save layer edits" toggles editing off | Bug Digitizing High Priority | Author Name: **Giovanni Manghi** (@gioman)
Original Redmine Issue: [20772](https://issues.qgis.org/issues/20772)
Affected QGIS version: 3.6.2
Redmine category:vectors
---
1) add a vector layer
2) in its options toggle one (or more) of the three "topology checks" options
3) toogle editing, draw a feature (tested polygons here)
4) click on "save layer edits" -> edits are saved, but editing is toggled off (not expected)
---
Related issue(s): #28757 (duplicates)
Redmine related issue(s): [20938](https://issues.qgis.org/issues/20938)
---
| 1.0 | If any of the "topology checks" options is enabled, then cliking on "save layer edits" toggles editing off - Author Name: **Giovanni Manghi** (@gioman)
Original Redmine Issue: [20772](https://issues.qgis.org/issues/20772)
Affected QGIS version: 3.6.2
Redmine category:vectors
---
1) add a vector layer
2) in its options toggle one (or more) of the three "topology checks" options
3) toogle editing, draw a feature (tested polygons here)
4) click on "save layer edits" -> edits are saved, but editing is toggled off (not expected)
---
Related issue(s): #28757 (duplicates)
Redmine related issue(s): [20938](https://issues.qgis.org/issues/20938)
---
| priority | if any of the topology checks options is enabled then cliking on save layer edits toggles editing off author name giovanni manghi gioman original redmine issue affected qgis version redmine category vectors add a vector layer in its options toggle one or more of the three topology checks options toogle editing draw a feature tested polygons here click on save layer edits edits are saved but editing is toggled off not expected related issue s duplicates redmine related issue s | 1 |
278,509 | 8,642,265,012 | IssuesEvent | 2018-11-25 04:30:19 | phonetworks/graphjs | https://api.github.com/repos/phonetworks/graphjs | opened | Blog module fixes | high priority bug | Please check the following on graphjs-website chaituvr-master-blog branch:
- [ ] Publish/unpublish needs to be checked and fixed (seems like unpublish doesn't work)
- [ ] Saved but unpublished work should be shown with a **[DRAFT]** prefix in its title, to its author.
- [ ] Delete is not working. Let's check if it's a server-side or client-side issue, and fix if it's client-side.
- [ ] when you click a link /blogs (go to a specific blog entry) it doesn't show the date. nor it shows in the /blogs homepage. It should show the date similarly how it shows in the blog-composer page.
- [ ] It should also show the permalink of the blog entry with hashtag. (please check if permalink works in the first place!!)
- [ ] In bullet lists, cursor goes down by a line automatically. To reproduce this, try using bullet points in composer. Check if it's graphjs-website only, or a general problem. Fix it in any case (if it's general, in graphjs repo, if it's graphjs-website only, then by overriding css on graphjs-website repo only)
* [ ] in /blogs back/forward buttons of the browser messes up the navigation. We should fix that.
And once these are fixed.
Add this component to graphjs-website docs in a new branch.
| 1.0 | Blog module fixes - Please check the following on graphjs-website chaituvr-master-blog branch:
- [ ] Publish/unpublish needs to be checked and fixed (seems like unpublish doesn't work)
- [ ] Saved but unpublished work should be shown with a **[DRAFT]** prefix in its title, to its author.
- [ ] Delete is not working. Let's check if it's a server-side or client-side issue, and fix if it's client-side.
- [ ] when you click a link /blogs (go to a specific blog entry) it doesn't show the date. nor it shows in the /blogs homepage. It should show the date similarly how it shows in the blog-composer page.
- [ ] It should also show the permalink of the blog entry with hashtag. (please check if permalink works in the first place!!)
- [ ] In bullet lists, cursor goes down by a line automatically. To reproduce this, try using bullet points in composer. Check if it's graphjs-website only, or a general problem. Fix it in any case (if it's general, in graphjs repo, if it's graphjs-website only, then by overriding css on graphjs-website repo only)
* [ ] in /blogs back/forward buttons of the browser messes up the navigation. We should fix that.
And once these are fixed.
Add this component to graphjs-website docs in a new branch.
| priority | blog module fixes please check the following on graphjs website chaituvr master blog branch publish unpublish needs to be checked and fixed seems like unpublish doesn t work saved but unpublished work should be shown with a prefix in its title to its author delete is not working let s check if it s a server side or client side issue and fix if it s client side when you click a link blogs go to a specific blog entry it doesn t show the date nor it shows in the blogs homepage it should show the date similarly how it shows in the blog composer page it should also show the permalink of the blog entry with hashtag please check if permalink works in the first place in bullet lists cursor goes down by a line automatically to reproduce this try using bullet points in composer check if it s graphjs website only or a general problem fix it in any case if it s general in graphjs repo if it s graphjs website only then by overriding css on graphjs website repo only in blogs back forward buttons of the browser messes up the navigation we should fix that and once these are fixed add this component to graphjs website docs in a new branch | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.