Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
18,716
24,606,555,643
IssuesEvent
2022-10-14 16:47:07
NationalSecurityAgency/ghidra
https://api.github.com/repos/NationalSecurityAgency/ghidra
closed
ARM A32 Thumb, VZIP and VUZP may be swapped
Reason: Internal effort Feature: Processor/ARM
**Describe the bug** I want to RE an ARM program, it uses vzip instructions, but when getting to a point of interest i came up with this: ![image](https://user-images.githubusercontent.com/30161277/195462105-0acd2d49-558f-4c1b-907b-0f6dbc09e559.png) ^^ This is what ghidra thinks the instruction is, however, the instruction is actually VZIP instead of VUZP For example, this is what a third party interpreter (https://github.com/merryhime/dynarmic) says about it: ![image](https://user-images.githubusercontent.com/30161277/195462305-acd0db2e-075e-4c51-a186-77a2b7b59bc5.png) Same with online decoder (https://armconverter.com/?disasm&code=b2ff8111) **To Reproduce** Steps to reproduce the behavior: 1. Dissasemble an ARM program 2. Go to an instruction that uses the opcode `b2ff8111` 3. See the wrong instruction **Expected behavior** The instruction matching the ones of other tools **Environment (please complete the following information):** - OS: Arch linux 5.19.13 - Java Version: `openjdk 18.0.2 2022-07-19` - Ghidra Version: 10.1.5 14 Aug 2022 - Ghidra Origin: https://archlinux.org/packages/community/x86_64/ghidra/
1.0
ARM A32 Thumb, VZIP and VUZP may be swapped - **Describe the bug** I want to RE an ARM program, it uses vzip instructions, but when getting to a point of interest i came up with this: ![image](https://user-images.githubusercontent.com/30161277/195462105-0acd2d49-558f-4c1b-907b-0f6dbc09e559.png) ^^ This is what ghidra thinks the instruction is, however, the instruction is actually VZIP instead of VUZP For example, this is what a third party interpreter (https://github.com/merryhime/dynarmic) says about it: ![image](https://user-images.githubusercontent.com/30161277/195462305-acd0db2e-075e-4c51-a186-77a2b7b59bc5.png) Same with online decoder (https://armconverter.com/?disasm&code=b2ff8111) **To Reproduce** Steps to reproduce the behavior: 1. Dissasemble an ARM program 2. Go to an instruction that uses the opcode `b2ff8111` 3. See the wrong instruction **Expected behavior** The instruction matching the ones of other tools **Environment (please complete the following information):** - OS: Arch linux 5.19.13 - Java Version: `openjdk 18.0.2 2022-07-19` - Ghidra Version: 10.1.5 14 Aug 2022 - Ghidra Origin: https://archlinux.org/packages/community/x86_64/ghidra/
process
arm thumb vzip and vuzp may be swapped describe the bug i want to re an arm program it uses vzip instructions but when getting to a point of interest i came up with this this is what ghidra thinks the instruction is however the instruction is actually vzip instead of vuzp for example this is what a third party interpreter says about it same with online decoder to reproduce steps to reproduce the behavior dissasemble an arm program go to an instruction that uses the opcode see the wrong instruction expected behavior the instruction matching the ones of other tools environment please complete the following information os arch linux java version openjdk ghidra version aug ghidra origin
1
6,420
9,522,841,305
IssuesEvent
2019-04-27 12:21:37
plazi/arcadia-project
https://api.github.com/repos/plazi/arcadia-project
opened
QC: quality control criteria listing
Article processing GoldenGate qualityControlTool
We need to list the critera to be implemented in the the Quality Control tool see: taxonomic treatment processing: https://docs.google.com/document/d/17tBAqwwdZWFNLQFTFWQeCi2kV0NTswbJy4nvDCpwT2M/edit#heading=h.y86qrq6o06fb input from EJT processing: https://docs.google.com/document/d/14ITiwfBHlBRjCZWpgssjH6DQQI39M2hyFRfkQqyQCOU/edit
1.0
QC: quality control criteria listing - We need to list the critera to be implemented in the the Quality Control tool see: taxonomic treatment processing: https://docs.google.com/document/d/17tBAqwwdZWFNLQFTFWQeCi2kV0NTswbJy4nvDCpwT2M/edit#heading=h.y86qrq6o06fb input from EJT processing: https://docs.google.com/document/d/14ITiwfBHlBRjCZWpgssjH6DQQI39M2hyFRfkQqyQCOU/edit
process
qc quality control criteria listing we need to list the critera to be implemented in the the quality control tool see taxonomic treatment processing input from ejt processing
1
5,928
8,752,450,915
IssuesEvent
2018-12-14 03:07:04
GoogleCloudPlatform/google-cloud-cpp
https://api.github.com/repos/GoogleCloudPlatform/google-cloud-cpp
closed
Enable additional clang-tidy warnings.
type: cleanup type: process
The following warnings are useful and should be turned on: - [x] misc-unused-using-decls - [x] bugprone-use-after-move - [x] misc-unused-alias-decls - [x] bugprone-move-forwarding-reference
1.0
Enable additional clang-tidy warnings. - The following warnings are useful and should be turned on: - [x] misc-unused-using-decls - [x] bugprone-use-after-move - [x] misc-unused-alias-decls - [x] bugprone-move-forwarding-reference
process
enable additional clang tidy warnings the following warnings are useful and should be turned on misc unused using decls bugprone use after move misc unused alias decls bugprone move forwarding reference
1
84,914
16,578,156,963
IssuesEvent
2021-05-31 08:09:49
CyBear-Jinni/CBJ_App
https://api.github.com/repos/CyBear-Jinni/CBJ_App
closed
wifi_info_flutter package is deprecated
Improving code organization good first issue not urgent
As the [wifi_info_flutter](https://pub.dev/packages/wifi_info_flutter) package in pub.dev say: > This plugin has been replaced by the Flutter Community Plus Plugins version, network_info_plus So we need to move to [network_info_plus](https://pub.dev/packages/network_info_plus)
1.0
wifi_info_flutter package is deprecated - As the [wifi_info_flutter](https://pub.dev/packages/wifi_info_flutter) package in pub.dev say: > This plugin has been replaced by the Flutter Community Plus Plugins version, network_info_plus So we need to move to [network_info_plus](https://pub.dev/packages/network_info_plus)
non_process
wifi info flutter package is deprecated as the package in pub dev say this plugin has been replaced by the flutter community plus plugins version network info plus so we need to move to
0
2,205
5,047,409,071
IssuesEvent
2016-12-20 09:20:04
hbz/lobid-resources
https://api.github.com/repos/hbz/lobid-resources
closed
Label with multiple entries (instead of just one)
processing review
Reported by @fsteeg in https://github.com/hbz/nwbib/pull/355#issuecomment-264499102. Snippet from http://lobid.org/resources/HT017324477: ```json "contributor" : [ { "altLabel" : "Westfälisches Heilpädagogisches Kinderheim", "id" : "http://d-nb.info/gnd/5183585-X", "label" : [ "Westfälisches Heilpädagogisches Kinderheim, Hamm (Westf)", "Westfälisches Heilpädagogisches Kinderheim (Hamm (Westf))" ], "type" : [ "CorporateBody" ] } ] ``` [Source](http://lobid.org/hbz01/HT017324477) snippet: ```xml <datafield tag="200" ind1="b" ind2="2"> <subfield code="k">Westfälisches Heilpädagogisches Kinderheim</subfield> <subfield code="h">Hamm (Westf)</subfield> <subfield code="9">(DE-588)5183585-X</subfield> </datafield> <datafield tag="902" ind1="-" ind2="2"> <subfield code="k">Westfälisches Heilpädagogisches Kinderheim</subfield> <subfield code="h">Hamm (Westf)</subfield> <subfield code="9">(DE-588)5183585-X</subfield> </datafield> ``` The problem seems to be that the resource is both contributor and subject as problems do not occur in similar entries (see e.g. http://lobid.org/resources/HT012319980). I guess, the label is created in a different way for contributors (with content from `h` in parentheses) and subjects (with content from `h` after comma). We should align these, using parentheses for both.
1.0
Label with multiple entries (instead of just one) - Reported by @fsteeg in https://github.com/hbz/nwbib/pull/355#issuecomment-264499102. Snippet from http://lobid.org/resources/HT017324477: ```json "contributor" : [ { "altLabel" : "Westfälisches Heilpädagogisches Kinderheim", "id" : "http://d-nb.info/gnd/5183585-X", "label" : [ "Westfälisches Heilpädagogisches Kinderheim, Hamm (Westf)", "Westfälisches Heilpädagogisches Kinderheim (Hamm (Westf))" ], "type" : [ "CorporateBody" ] } ] ``` [Source](http://lobid.org/hbz01/HT017324477) snippet: ```xml <datafield tag="200" ind1="b" ind2="2"> <subfield code="k">Westfälisches Heilpädagogisches Kinderheim</subfield> <subfield code="h">Hamm (Westf)</subfield> <subfield code="9">(DE-588)5183585-X</subfield> </datafield> <datafield tag="902" ind1="-" ind2="2"> <subfield code="k">Westfälisches Heilpädagogisches Kinderheim</subfield> <subfield code="h">Hamm (Westf)</subfield> <subfield code="9">(DE-588)5183585-X</subfield> </datafield> ``` The problem seems to be that the resource is both contributor and subject as problems do not occur in similar entries (see e.g. http://lobid.org/resources/HT012319980). I guess, the label is created in a different way for contributors (with content from `h` in parentheses) and subjects (with content from `h` after comma). We should align these, using parentheses for both.
process
label with multiple entries instead of just one reported by fsteeg in snippet from json contributor altlabel westfälisches heilpädagogisches kinderheim id label type snippet xml westfälisches heilpädagogisches kinderheim hamm westf de x westfälisches heilpädagogisches kinderheim hamm westf de x the problem seems to be that the resource is both contributor and subject as problems do not occur in similar entries see e g i guess the label is created in a different way for contributors with content from h in parentheses and subjects with content from h after comma we should align these using parentheses for both
1
8,207
2,611,471,060
IssuesEvent
2015-02-27 05:15:50
chrsmith/hedgewars
https://api.github.com/repos/chrsmith/hedgewars
closed
Hedgewars does not run on Mac OS X 10.7 Lion
auto-migrated Priority-Medium Type-Defect
``` I got this information from hezz in the #hedgewars channel, who downloaded the Developer Review 2 version of OS X Lion from the Apple developer website, and installed Hedgewars via the App Store. What steps will reproduce the problem? 1. Install Developer Review 2 of Mac OS X Lion. 2. Download and run Hedgewars from the Mac App Store 3. Run a game. Instead of the engine displaying a game window, it simply fails to load (I don't know whether a window pops up or it just gives an error message in the Qt window). The output in Terminal is as follows: hezkiels-macbook:applications hezkiel$ /Applications/Hedgewars.app/Contents/MacOS/hedgewars 2011-04-15 18:47:11.038 hedgewars[3645:e03] Error loading /Library/Audio/Plug-Ins/HAL/OSS 3D.plugin/Contents/MacOS/OSS 3D: dlopen(/Library/Audio/Plug-Ins/HAL/OSS 3D.plugin/Contents/MacOS/OSS 3D, 262): no suitable image found. Did find: /Library/Audio/Plug-Ins/HAL/OSS 3D.plugin/Contents/MacOS/OSS 3D: no matching architecture in universal wrapper 2011-04-15 18:47:11.044 hedgewars[3645:e03] Cannot find function pointer MyCFPlugInFactory for factory 8368FFAB-0147-4692-97ED-4D50A2483C6C in CFBundle/CFPlugIn 0x10949e160 </Library/Audio/Plug-Ins/HAL/OSS 3D.plugin> (bundle, not loaded) If you have any other questions, or things you want to try, I can e-mail him. :) ``` Original issue reported on code.google.com by `sjorsgie...@gmail.com` on 15 Apr 2011 at 11:54
1.0
Hedgewars does not run on Mac OS X 10.7 Lion - ``` I got this information from hezz in the #hedgewars channel, who downloaded the Developer Review 2 version of OS X Lion from the Apple developer website, and installed Hedgewars via the App Store. What steps will reproduce the problem? 1. Install Developer Review 2 of Mac OS X Lion. 2. Download and run Hedgewars from the Mac App Store 3. Run a game. Instead of the engine displaying a game window, it simply fails to load (I don't know whether a window pops up or it just gives an error message in the Qt window). The output in Terminal is as follows: hezkiels-macbook:applications hezkiel$ /Applications/Hedgewars.app/Contents/MacOS/hedgewars 2011-04-15 18:47:11.038 hedgewars[3645:e03] Error loading /Library/Audio/Plug-Ins/HAL/OSS 3D.plugin/Contents/MacOS/OSS 3D: dlopen(/Library/Audio/Plug-Ins/HAL/OSS 3D.plugin/Contents/MacOS/OSS 3D, 262): no suitable image found. Did find: /Library/Audio/Plug-Ins/HAL/OSS 3D.plugin/Contents/MacOS/OSS 3D: no matching architecture in universal wrapper 2011-04-15 18:47:11.044 hedgewars[3645:e03] Cannot find function pointer MyCFPlugInFactory for factory 8368FFAB-0147-4692-97ED-4D50A2483C6C in CFBundle/CFPlugIn 0x10949e160 </Library/Audio/Plug-Ins/HAL/OSS 3D.plugin> (bundle, not loaded) If you have any other questions, or things you want to try, I can e-mail him. :) ``` Original issue reported on code.google.com by `sjorsgie...@gmail.com` on 15 Apr 2011 at 11:54
non_process
hedgewars does not run on mac os x lion i got this information from hezz in the hedgewars channel who downloaded the developer review version of os x lion from the apple developer website and installed hedgewars via the app store what steps will reproduce the problem install developer review of mac os x lion download and run hedgewars from the mac app store run a game instead of the engine displaying a game window it simply fails to load i don t know whether a window pops up or it just gives an error message in the qt window the output in terminal is as follows hezkiels macbook applications hezkiel applications hedgewars app contents macos hedgewars hedgewars error loading library audio plug ins hal oss plugin contents macos oss dlopen library audio plug ins hal oss plugin contents macos oss no suitable image found did find library audio plug ins hal oss plugin contents macos oss no matching architecture in universal wrapper hedgewars cannot find function pointer mycfpluginfactory for factory in cfbundle cfplugin bundle not loaded if you have any other questions or things you want to try i can e mail him original issue reported on code google com by sjorsgie gmail com on apr at
0
261,060
19,696,636,569
IssuesEvent
2022-01-12 12:51:45
spring-projects/spring-data-mongodb
https://api.github.com/repos/spring-projects/spring-data-mongodb
opened
No longer default UUID representation to JAVA_LEGACY
type: documentation type: task theme: 4.0
Turn defaulting off and enforce an explicit choice from the user.
1.0
No longer default UUID representation to JAVA_LEGACY - Turn defaulting off and enforce an explicit choice from the user.
non_process
no longer default uuid representation to java legacy turn defaulting off and enforce an explicit choice from the user
0
7,020
10,169,093,076
IssuesEvent
2019-08-07 22:56:35
processing-r/Processing.R
https://api.github.com/repos/processing-r/Processing.R
closed
Processing.R mode not detected in contributions manager
community/processing priority/p0
## Problem Currently Processing.R is not showing up in PDE Contributions Manager. ## Reason This file is missing -- PDE contributions manager doesn't show a listing and doesn't link to the zip file without it. http://gaocegege.com/Processing.R/RLangMode.txt ## Solution @gaocegege do you have a copy of RLangMode.txt ? If so: 1. attach RangLangeMode.txt to the 1.0.6 release here: https://github.com/processing-r/Processing.R/releases/edit/v1.0.6 2. I will notify the Processing Foundation to update sources.conf here: https://raw.githubusercontent.com/processing/processing-docs/master/contrib_generate/sources.conf from this entry: 220 \ http://gaocegege.com/Processing.R/RLangMode.txt to this: 220 \ https://github.com/processing-r/Processing.R/releases/download/latest/RLangMode.txt This is how other github-hosted libraries do it, for example: https://github.com/gohai/processing-uploadtopi/releases/ https://github.com/gohai/processing-uploadtopi/releases/latest/download/UploadToPiTool.txt
1.0
Processing.R mode not detected in contributions manager - ## Problem Currently Processing.R is not showing up in PDE Contributions Manager. ## Reason This file is missing -- PDE contributions manager doesn't show a listing and doesn't link to the zip file without it. http://gaocegege.com/Processing.R/RLangMode.txt ## Solution @gaocegege do you have a copy of RLangMode.txt ? If so: 1. attach RangLangeMode.txt to the 1.0.6 release here: https://github.com/processing-r/Processing.R/releases/edit/v1.0.6 2. I will notify the Processing Foundation to update sources.conf here: https://raw.githubusercontent.com/processing/processing-docs/master/contrib_generate/sources.conf from this entry: 220 \ http://gaocegege.com/Processing.R/RLangMode.txt to this: 220 \ https://github.com/processing-r/Processing.R/releases/download/latest/RLangMode.txt This is how other github-hosted libraries do it, for example: https://github.com/gohai/processing-uploadtopi/releases/ https://github.com/gohai/processing-uploadtopi/releases/latest/download/UploadToPiTool.txt
process
processing r mode not detected in contributions manager problem currently processing r is not showing up in pde contributions manager reason this file is missing pde contributions manager doesn t show a listing and doesn t link to the zip file without it solution gaocegege do you have a copy of rlangmode txt if so attach ranglangemode txt to the release here i will notify the processing foundation to update sources conf here from this entry to this this is how other github hosted libraries do it for example
1
5,322
8,136,258,936
IssuesEvent
2018-08-20 07:47:25
openvstorage/framework-alba-plugin
https://api.github.com/repos/openvstorage/framework-alba-plugin
closed
error claim disk to local backend linked to global backend
process_cantreproduce state_question
Hi all. I install my cluster on OpenVStorage version : fargo-update-1 > > ii openvstorage 2.9.5-1 amd64 openvStorage > ii openvstorage-backend 1.9.1-1 amd64 openvStorage Backend plugin > ii openvstorage-backend-core 1.9.1-1 amd64 openvStorage Backend plugin core > ii openvstorage-backend-webapps 1.9.1-1 amd64 openvStorage Backend plugin Web Applications > ii openvstorage-core 2.9.5-1 amd64 openvStorage core > ii openvstorage-webapps 2.9.5-1 amd64 openvStorage Web Applications After add peformance node 1,2 and 3, i add node 4 5. All disk on node performance 4,5 initialized , but click to claim all disk it do nothing. I tried remove node, add new nodes but not work. Thanks.
1.0
error claim disk to local backend linked to global backend - Hi all. I install my cluster on OpenVStorage version : fargo-update-1 > > ii openvstorage 2.9.5-1 amd64 openvStorage > ii openvstorage-backend 1.9.1-1 amd64 openvStorage Backend plugin > ii openvstorage-backend-core 1.9.1-1 amd64 openvStorage Backend plugin core > ii openvstorage-backend-webapps 1.9.1-1 amd64 openvStorage Backend plugin Web Applications > ii openvstorage-core 2.9.5-1 amd64 openvStorage core > ii openvstorage-webapps 2.9.5-1 amd64 openvStorage Web Applications After add peformance node 1,2 and 3, i add node 4 5. All disk on node performance 4,5 initialized , but click to claim all disk it do nothing. I tried remove node, add new nodes but not work. Thanks.
process
error claim disk to local backend linked to global backend hi all i install my cluster on openvstorage version fargo update ii openvstorage openvstorage ii openvstorage backend openvstorage backend plugin ii openvstorage backend core openvstorage backend plugin core ii openvstorage backend webapps openvstorage backend plugin web applications ii openvstorage core openvstorage core ii openvstorage webapps openvstorage web applications after add peformance node and i add node all disk on node performance initialized but click to claim all disk it do nothing i tried remove node add new nodes but not work thanks
1
88,875
17,689,359,847
IssuesEvent
2021-08-24 08:02:43
microsoft/vscode
https://api.github.com/repos/microsoft/vscode
closed
The quick fix short cut
keybindings editor-code-actions notebook
We have written the needed data into your clipboard because it was too large to send. Please paste. Issue Type: <b>Bug</b> I change the quick fix shortcut to alt +enter(using intellij keymap), however, it is same as the 'note book cell and insert below'. I want to change the short cut of note book cell but is not actually changed. That is, the note book short cut is unchangeble. More over, the trigger condition of for quick fix is not working properly, i press 'alt enter' on the underlined code and it swith the cusor to below line. Previous version has no such problem. VS Code version: Code 1.55.2 (3c4e3df9e89829dce27b7b5c24508306b151f30d, 2021-04-13T09:35:57.887Z) OS version: Windows_NT x64 10.0.19041 <details> <summary>System Info</summary> |Item|Value| |---|---| |CPUs|Intel(R) Core(TM) i7-9750H CPU @ 2.60GHz (12 x 2592)| |GPU Status|2d_canvas: enabled<br>gpu_compositing: enabled<br>multiple_raster_threads: enabled_on<br>oop_rasterization: enabled<br>opengl: enabled_on<br>protected_video_decode: unavailable_off<br>rasterization: enabled<br>skia_renderer: enabled_on<br>video_decode: enabled<br>vulkan: disabled_off<br>webgl: enabled<br>webgl2: enabled| |Load (avg)|undefined| |Memory (System)|15.88GB (8.92GB free)| |Process Argv|--crash-reporter-id bc31d28a-bd8f-4fc4-af71-959b15922be9| |Screen Reader|no| |VM|57%| </details><details><summary>Extensions (29)</summary> Extension|Author (truncated)|Version ---|---|--- matlab-formatter|Aff|2.8.7 matlab-interactive-terminal|apo|0.3.4 vscode-browser-preview|auc|0.7.1 matlab-extension-pack|bat|0.1.0 matlab-code-run|bra|1.0.2 bracket-pair-colorizer-2|Coe|0.2.0 sql-formatter|cym|1.0.2 vuerd-vscode|din|0.7.19 matlab|Gim|2.0.1 latex-workshop|Jam|8.17.0 vscode-language-pack-zh-hans|MS-|1.55.2 remote-wsl|ms-|0.54.6 cpptools|ms-|1.3.1 wordcount|ms-|0.1.0 java|red|0.77.0 vscode-commons|red|0.0.6 vscode-xml|red|0.16.0 matlab-complete|Sla|1.1.1 code-spell-checker|str|1.10.2 tabnine-vscode|Tab|3.4.4 language-pde|Tob|1.4.6 vscodeintellicode|Vis|1.2.12 vscode-java-debug|vsc|0.33.0 vscode-java-dependency|vsc|0.18.3 vscode-java-pack|vsc|0.14.0 vscode-java-test|vsc|0.28.1 vscode-maven|vsc|0.30.0 markdown-pdf|yza|1.4.4 markdown-all-in-one|yzh|3.4.0 (1 theme extensions excluded) </details><details> <summary>A/B Experiments</summary> ``` vsliv368:30146709 vsreu685:30147344 python383:30185418 pythonvspyt700cf:30270857 pythonvspyt602:30294772 vspor879:30202332 vspor708:30202333 vspor363:30204092 vswsl492cf:30256860 vstry914:30276682 vstes516:30244333 pythonvspyt639:30291489 pythontb:30283811 vspre833:30267464 pythonptprofiler:30281270 vshan820:30294714 vscoreces:30290705 pythondataviewer:30285071 vscus158:30286553 ``` </details> <!-- generated by issue reporter -->
1.0
The quick fix short cut - We have written the needed data into your clipboard because it was too large to send. Please paste. Issue Type: <b>Bug</b> I change the quick fix shortcut to alt +enter(using intellij keymap), however, it is same as the 'note book cell and insert below'. I want to change the short cut of note book cell but is not actually changed. That is, the note book short cut is unchangeble. More over, the trigger condition of for quick fix is not working properly, i press 'alt enter' on the underlined code and it swith the cusor to below line. Previous version has no such problem. VS Code version: Code 1.55.2 (3c4e3df9e89829dce27b7b5c24508306b151f30d, 2021-04-13T09:35:57.887Z) OS version: Windows_NT x64 10.0.19041 <details> <summary>System Info</summary> |Item|Value| |---|---| |CPUs|Intel(R) Core(TM) i7-9750H CPU @ 2.60GHz (12 x 2592)| |GPU Status|2d_canvas: enabled<br>gpu_compositing: enabled<br>multiple_raster_threads: enabled_on<br>oop_rasterization: enabled<br>opengl: enabled_on<br>protected_video_decode: unavailable_off<br>rasterization: enabled<br>skia_renderer: enabled_on<br>video_decode: enabled<br>vulkan: disabled_off<br>webgl: enabled<br>webgl2: enabled| |Load (avg)|undefined| |Memory (System)|15.88GB (8.92GB free)| |Process Argv|--crash-reporter-id bc31d28a-bd8f-4fc4-af71-959b15922be9| |Screen Reader|no| |VM|57%| </details><details><summary>Extensions (29)</summary> Extension|Author (truncated)|Version ---|---|--- matlab-formatter|Aff|2.8.7 matlab-interactive-terminal|apo|0.3.4 vscode-browser-preview|auc|0.7.1 matlab-extension-pack|bat|0.1.0 matlab-code-run|bra|1.0.2 bracket-pair-colorizer-2|Coe|0.2.0 sql-formatter|cym|1.0.2 vuerd-vscode|din|0.7.19 matlab|Gim|2.0.1 latex-workshop|Jam|8.17.0 vscode-language-pack-zh-hans|MS-|1.55.2 remote-wsl|ms-|0.54.6 cpptools|ms-|1.3.1 wordcount|ms-|0.1.0 java|red|0.77.0 vscode-commons|red|0.0.6 vscode-xml|red|0.16.0 matlab-complete|Sla|1.1.1 code-spell-checker|str|1.10.2 tabnine-vscode|Tab|3.4.4 language-pde|Tob|1.4.6 vscodeintellicode|Vis|1.2.12 vscode-java-debug|vsc|0.33.0 vscode-java-dependency|vsc|0.18.3 vscode-java-pack|vsc|0.14.0 vscode-java-test|vsc|0.28.1 vscode-maven|vsc|0.30.0 markdown-pdf|yza|1.4.4 markdown-all-in-one|yzh|3.4.0 (1 theme extensions excluded) </details><details> <summary>A/B Experiments</summary> ``` vsliv368:30146709 vsreu685:30147344 python383:30185418 pythonvspyt700cf:30270857 pythonvspyt602:30294772 vspor879:30202332 vspor708:30202333 vspor363:30204092 vswsl492cf:30256860 vstry914:30276682 vstes516:30244333 pythonvspyt639:30291489 pythontb:30283811 vspre833:30267464 pythonptprofiler:30281270 vshan820:30294714 vscoreces:30290705 pythondataviewer:30285071 vscus158:30286553 ``` </details> <!-- generated by issue reporter -->
non_process
the quick fix short cut we have written the needed data into your clipboard because it was too large to send please paste issue type bug i change the quick fix shortcut to alt enter using intellij keymap however it is same as the note book cell and insert below i want to change the short cut of note book cell but is not actually changed that is the note book short cut is unchangeble more over the trigger condition of for quick fix is not working properly i press alt enter on the underlined code and it swith the cusor to below line previous version has no such problem vs code version code os version windows nt system info item value cpus intel r core tm cpu x gpu status canvas enabled gpu compositing enabled multiple raster threads enabled on oop rasterization enabled opengl enabled on protected video decode unavailable off rasterization enabled skia renderer enabled on video decode enabled vulkan disabled off webgl enabled enabled load avg undefined memory system free process argv crash reporter id screen reader no vm extensions extension author truncated version matlab formatter aff matlab interactive terminal apo vscode browser preview auc matlab extension pack bat matlab code run bra bracket pair colorizer coe sql formatter cym vuerd vscode din matlab gim latex workshop jam vscode language pack zh hans ms remote wsl ms cpptools ms wordcount ms java red vscode commons red vscode xml red matlab complete sla code spell checker str tabnine vscode tab language pde tob vscodeintellicode vis vscode java debug vsc vscode java dependency vsc vscode java pack vsc vscode java test vsc vscode maven vsc markdown pdf yza markdown all in one yzh theme extensions excluded a b experiments pythontb pythonptprofiler vscoreces pythondataviewer
0
32,016
26,349,848,380
IssuesEvent
2023-01-11 03:08:14
fosscord/fosscord-server
https://api.github.com/repos/fosscord/fosscord-server
closed
TypeORM broke its own junction table
bug infrastructure
When a new account is created the first created guild is still there from previous accounts. You can leave but the main problem is that it's there in the first place
1.0
TypeORM broke its own junction table - When a new account is created the first created guild is still there from previous accounts. You can leave but the main problem is that it's there in the first place
non_process
typeorm broke its own junction table when a new account is created the first created guild is still there from previous accounts you can leave but the main problem is that it s there in the first place
0
1,242
2,714,849,415
IssuesEvent
2015-04-10 08:32:48
debrief/debrief
https://api.github.com/repos/debrief/debrief
closed
Make standard distribution "portable"
build new users
I'd like to be able to hand out Debrief CDs or memory sticks - or leave a supply of these with interested users ("fans"). Debrief would be in a runnable form on the device - so it can just run (without being unzipped / installed). We effectively support this already, but I'd add a JRE and Natural Earth to the CD. I'd also like Debrief to start off with the "Sample Data" project already in the navigator. Could you consider how we'd do this? I know we have the "Create Project" button - which we could auto-run if there aren't any projects visible. I'm wondering if its possible to do by configuration? So, instead of creating a new empty project that gets a copied version of the sample data - we would just have a sample project/workspace in the root level of the distribution that opens. Hmm, wasn't the default behaviour of Eclipse to create a .metadata folder in the root of the installed folder? If that works for a read-only folder (since we're on a CD-ROM), then that may be an acceptable solution.
1.0
Make standard distribution "portable" - I'd like to be able to hand out Debrief CDs or memory sticks - or leave a supply of these with interested users ("fans"). Debrief would be in a runnable form on the device - so it can just run (without being unzipped / installed). We effectively support this already, but I'd add a JRE and Natural Earth to the CD. I'd also like Debrief to start off with the "Sample Data" project already in the navigator. Could you consider how we'd do this? I know we have the "Create Project" button - which we could auto-run if there aren't any projects visible. I'm wondering if its possible to do by configuration? So, instead of creating a new empty project that gets a copied version of the sample data - we would just have a sample project/workspace in the root level of the distribution that opens. Hmm, wasn't the default behaviour of Eclipse to create a .metadata folder in the root of the installed folder? If that works for a read-only folder (since we're on a CD-ROM), then that may be an acceptable solution.
non_process
make standard distribution portable i d like to be able to hand out debrief cds or memory sticks or leave a supply of these with interested users fans debrief would be in a runnable form on the device so it can just run without being unzipped installed we effectively support this already but i d add a jre and natural earth to the cd i d also like debrief to start off with the sample data project already in the navigator could you consider how we d do this i know we have the create project button which we could auto run if there aren t any projects visible i m wondering if its possible to do by configuration so instead of creating a new empty project that gets a copied version of the sample data we would just have a sample project workspace in the root level of the distribution that opens hmm wasn t the default behaviour of eclipse to create a metadata folder in the root of the installed folder if that works for a read only folder since we re on a cd rom then that may be an acceptable solution
0
12,385
14,900,172,859
IssuesEvent
2021-01-21 15:08:04
researchdata-sheffield/dataviz-hub2-comments
https://api.github.com/repos/researchdata-sheffield/dataviz-hub2-comments
opened
Data Processing for data visualisations in R
/blog/15/01/2021/Data-Processing-In-R Gitalk
https://dataviz.shef.ac.uk/blog/15/01/2021/Data-Processing-In-R | Part 1 | Exploring packages in R using data collected as part of the EU funded 'Collaborative research and development of green roof system technology' project.
1.0
Data Processing for data visualisations in R - https://dataviz.shef.ac.uk/blog/15/01/2021/Data-Processing-In-R | Part 1 | Exploring packages in R using data collected as part of the EU funded 'Collaborative research and development of green roof system technology' project.
process
data processing for data visualisations in r part exploring packages in r using data collected as part of the eu funded collaborative research and development of green roof system technology project
1
219,642
7,344,596,482
IssuesEvent
2018-03-07 15:07:03
ngageoint/hootenanny
https://api.github.com/repos/ngageoint/hootenanny
opened
Rename core commands for consistency
Category: Core Priority: Low Status: Defined Type: Maintenance
The list has grown pretty long. Its easier to follow them when the first part of the command name contains the category they are in. That way, like commands are grouped together when you list them. Here are some proposed changes: * conflate-cumulative --> conflate --cumulative (rnd) * diff-conflate --> conflate --differential * apply-changeset --> changeset-apply * derive-changeset --> changeset-derive * apply-rubber-sheet --> rubber-sheet-apply * derive-rubber-sheet --> rubber-sheet-derive * delete-map --> map-delete * calculate-tiles --> tile-calculate (rnd) * calculate-random-tile --> tile-calculate-random (rnd) * attr-co-occurence --> attribute-co-occurrence * optimize-conflate-settings --> optimize-network-config-options * remove big-ogr2osm (not really useful w/o big-merge-nodes, which is gone with Hadoop) * add-pbf-header --> pbf-add-header
1.0
Rename core commands for consistency - The list has grown pretty long. Its easier to follow them when the first part of the command name contains the category they are in. That way, like commands are grouped together when you list them. Here are some proposed changes: * conflate-cumulative --> conflate --cumulative (rnd) * diff-conflate --> conflate --differential * apply-changeset --> changeset-apply * derive-changeset --> changeset-derive * apply-rubber-sheet --> rubber-sheet-apply * derive-rubber-sheet --> rubber-sheet-derive * delete-map --> map-delete * calculate-tiles --> tile-calculate (rnd) * calculate-random-tile --> tile-calculate-random (rnd) * attr-co-occurence --> attribute-co-occurrence * optimize-conflate-settings --> optimize-network-config-options * remove big-ogr2osm (not really useful w/o big-merge-nodes, which is gone with Hadoop) * add-pbf-header --> pbf-add-header
non_process
rename core commands for consistency the list has grown pretty long its easier to follow them when the first part of the command name contains the category they are in that way like commands are grouped together when you list them here are some proposed changes conflate cumulative conflate cumulative rnd diff conflate conflate differential apply changeset changeset apply derive changeset changeset derive apply rubber sheet rubber sheet apply derive rubber sheet rubber sheet derive delete map map delete calculate tiles tile calculate rnd calculate random tile tile calculate random rnd attr co occurence attribute co occurrence optimize conflate settings optimize network config options remove big not really useful w o big merge nodes which is gone with hadoop add pbf header pbf add header
0
26,151
7,787,462,386
IssuesEvent
2018-06-06 22:36:48
couchbase/couchbase-lite-ios
https://api.github.com/repos/couchbase/couchbase-lite-ios
closed
Incompatible swift version
Build P1: high
Hi, I am using carthage as package manager, and the lastest XCode version (iOS 11.4 with XCode 9.3). Unfortunately, couchbase-lite-ios is a binary-only release with Carthage (recomended way to get it in the readme) and is incompatible with the latest swift version: ```bash $ carthage update --platform iOS *** Downloading binary-only framework CouchbaseLite-Community at "https://raw.githubusercontent.com/couchbase/couchbase-lite-ios/master/Release/Carthage/CouchbaseLite-Community.json" *** xcodebuild output can be found in /var/folders/f9/33ntg5xj38n_5xrdj7rc7bqh0000gq/T/carthage-xcodebuild.9T8SGO.log Incompatible Swift version - framework was built with 4.1 (swiftlang-902.0.48 clang-902.0.37.1) and the local version is 4.1.2 (swiftlang-902.0.54 clang-902.0.39.2). ``` -------- * Version: 2.0.0 * Client OS: macOS High Sierra 10.13.4 * Server: n/a
1.0
Incompatible swift version - Hi, I am using carthage as package manager, and the lastest XCode version (iOS 11.4 with XCode 9.3). Unfortunately, couchbase-lite-ios is a binary-only release with Carthage (recomended way to get it in the readme) and is incompatible with the latest swift version: ```bash $ carthage update --platform iOS *** Downloading binary-only framework CouchbaseLite-Community at "https://raw.githubusercontent.com/couchbase/couchbase-lite-ios/master/Release/Carthage/CouchbaseLite-Community.json" *** xcodebuild output can be found in /var/folders/f9/33ntg5xj38n_5xrdj7rc7bqh0000gq/T/carthage-xcodebuild.9T8SGO.log Incompatible Swift version - framework was built with 4.1 (swiftlang-902.0.48 clang-902.0.37.1) and the local version is 4.1.2 (swiftlang-902.0.54 clang-902.0.39.2). ``` -------- * Version: 2.0.0 * Client OS: macOS High Sierra 10.13.4 * Server: n/a
non_process
incompatible swift version hi i am using carthage as package manager and the lastest xcode version ios with xcode unfortunately couchbase lite ios is a binary only release with carthage recomended way to get it in the readme and is incompatible with the latest swift version bash carthage update platform ios downloading binary only framework couchbaselite community at xcodebuild output can be found in var folders t carthage xcodebuild log incompatible swift version framework was built with swiftlang clang and the local version is swiftlang clang version client os macos high sierra server n a
0
10,164
13,044,162,681
IssuesEvent
2020-07-29 03:47:35
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
UCP: Migrate scalar function `ValuesString` from TiDB
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
## Description Port the scalar function `ValuesString` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @iosmanthus ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
2.0
UCP: Migrate scalar function `ValuesString` from TiDB - ## Description Port the scalar function `ValuesString` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @iosmanthus ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
process
ucp migrate scalar function valuesstring from tidb description port the scalar function valuesstring from tidb to coprocessor score mentor s iosmanthus recommended skills rust programming learning materials already implemented expressions ported from tidb
1
14,824
18,166,073,109
IssuesEvent
2021-09-27 14:44:18
googleapis/python-pubsub
https://api.github.com/repos/googleapis/python-pubsub
closed
Add type annotation checks to CI
api: pubsub type: process
We do not run static type checks on the code, but we probably should, which will allow us to safely incrementally annotate the library code, i.e. without incorrect annotation that would break users' checks. Something similar to the [pull request](https://github.com/googleapis/python-bigquery/pull/613/files) that we merged a few months ago in BigQuery. ~~(the scope of the issue can be extended to also include annotating all public methods)~~ Will be done separately.
1.0
Add type annotation checks to CI - We do not run static type checks on the code, but we probably should, which will allow us to safely incrementally annotate the library code, i.e. without incorrect annotation that would break users' checks. Something similar to the [pull request](https://github.com/googleapis/python-bigquery/pull/613/files) that we merged a few months ago in BigQuery. ~~(the scope of the issue can be extended to also include annotating all public methods)~~ Will be done separately.
process
add type annotation checks to ci we do not run static type checks on the code but we probably should which will allow us to safely incrementally annotate the library code i e without incorrect annotation that would break users checks something similar to the that we merged a few months ago in bigquery the scope of the issue can be extended to also include annotating all public methods will be done separately
1
33,620
12,216,770,397
IssuesEvent
2020-05-01 15:48:55
habusha/CIOIL
https://api.github.com/repos/habusha/CIOIL
opened
CVE-2020-10969 (High) detected in jackson-databind-2.9.7.jar
security vulnerability
## CVE-2020-10969 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.7.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /tmp/ws-scm/CIOIL/infra_github/pom.xml</p> <p>Path to vulnerable library: /tmp/ws-ua_20200501140025_KHMIDU/downloadResource_IFJBLS/20200501140117/jackson-databind-2.9.7.jar</p> <p> Dependency Hierarchy: - logstash-logback-encoder-5.2.jar (Root Library) - :x: **jackson-databind-2.9.7.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/habusha/CIOIL/commit/bbaa61e2fd7a1837b81f9827e715dc8c1817cd31">bbaa61e2fd7a1837b81f9827e715dc8c1817cd31</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to javax.swing.JEditorPane. <p>Publish Date: 2020-03-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10969>CVE-2020-10969</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-10969">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-10969</a></p> <p>Release Date: 2020-03-26</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.8.11.6;com.fasterxml.jackson.core:jackson-databind:2.7.9.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-10969 (High) detected in jackson-databind-2.9.7.jar - ## CVE-2020-10969 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.7.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /tmp/ws-scm/CIOIL/infra_github/pom.xml</p> <p>Path to vulnerable library: /tmp/ws-ua_20200501140025_KHMIDU/downloadResource_IFJBLS/20200501140117/jackson-databind-2.9.7.jar</p> <p> Dependency Hierarchy: - logstash-logback-encoder-5.2.jar (Root Library) - :x: **jackson-databind-2.9.7.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/habusha/CIOIL/commit/bbaa61e2fd7a1837b81f9827e715dc8c1817cd31">bbaa61e2fd7a1837b81f9827e715dc8c1817cd31</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to javax.swing.JEditorPane. <p>Publish Date: 2020-03-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10969>CVE-2020-10969</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-10969">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-10969</a></p> <p>Release Date: 2020-03-26</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.8.11.6;com.fasterxml.jackson.core:jackson-databind:2.7.9.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file tmp ws scm cioil infra github pom xml path to vulnerable library tmp ws ua khmidu downloadresource ifjbls jackson databind jar dependency hierarchy logstash logback encoder jar root library x jackson databind jar vulnerable library found in head commit a href vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to javax swing jeditorpane publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind com fasterxml jackson core jackson databind step up your open source security game with whitesource
0
123,661
17,772,288,191
IssuesEvent
2021-08-30 14:56:06
kapseliboi/platform-status
https://api.github.com/repos/kapseliboi/platform-status
opened
CVE-2017-16116 (High) detected in underscore.string-2.2.1.tgz
security vulnerability
## CVE-2017-16116 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>underscore.string-2.2.1.tgz</b></p></summary> <p>String manipulation extensions for Underscore.js javascript library.</p> <p>Library home page: <a href="https://registry.npmjs.org/underscore.string/-/underscore.string-2.2.1.tgz">https://registry.npmjs.org/underscore.string/-/underscore.string-2.2.1.tgz</a></p> <p> Dependency Hierarchy: - oghliner-1.1.1.tgz (Root Library) - travis-ci-2.1.1.tgz - :x: **underscore.string-2.2.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kapseliboi/platform-status/commit/f1843ccb4f9fa8cac219c196c9bcceb734286e98">f1843ccb4f9fa8cac219c196c9bcceb734286e98</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The string module is a module that provides extra string operations. The string module is vulnerable to regular expression denial of service when specifically crafted untrusted user input is passed into the underscore or unescapeHTML methods. <p>Publish Date: 2018-06-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16116>CVE-2017-16116</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/745">https://www.npmjs.com/advisories/745</a></p> <p>Release Date: 2018-06-07</p> <p>Fix Resolution: 3.3.5</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2017-16116 (High) detected in underscore.string-2.2.1.tgz - ## CVE-2017-16116 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>underscore.string-2.2.1.tgz</b></p></summary> <p>String manipulation extensions for Underscore.js javascript library.</p> <p>Library home page: <a href="https://registry.npmjs.org/underscore.string/-/underscore.string-2.2.1.tgz">https://registry.npmjs.org/underscore.string/-/underscore.string-2.2.1.tgz</a></p> <p> Dependency Hierarchy: - oghliner-1.1.1.tgz (Root Library) - travis-ci-2.1.1.tgz - :x: **underscore.string-2.2.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kapseliboi/platform-status/commit/f1843ccb4f9fa8cac219c196c9bcceb734286e98">f1843ccb4f9fa8cac219c196c9bcceb734286e98</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The string module is a module that provides extra string operations. The string module is vulnerable to regular expression denial of service when specifically crafted untrusted user input is passed into the underscore or unescapeHTML methods. <p>Publish Date: 2018-06-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16116>CVE-2017-16116</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/745">https://www.npmjs.com/advisories/745</a></p> <p>Release Date: 2018-06-07</p> <p>Fix Resolution: 3.3.5</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in underscore string tgz cve high severity vulnerability vulnerable library underscore string tgz string manipulation extensions for underscore js javascript library library home page a href dependency hierarchy oghliner tgz root library travis ci tgz x underscore string tgz vulnerable library found in head commit a href found in base branch master vulnerability details the string module is a module that provides extra string operations the string module is vulnerable to regular expression denial of service when specifically crafted untrusted user input is passed into the underscore or unescapehtml methods publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
3,752
6,733,153,401
IssuesEvent
2017-10-18 14:00:31
york-region-tpss/stp
https://api.github.com/repos/york-region-tpss/stp
closed
Tree Planting Deficiency Repair Status - Assign Inspector
form process workflow
Create a form to assign an inspector to a contract
1.0
Tree Planting Deficiency Repair Status - Assign Inspector - Create a form to assign an inspector to a contract
process
tree planting deficiency repair status assign inspector create a form to assign an inspector to a contract
1
213,846
16,541,099,435
IssuesEvent
2021-05-27 16:53:53
dotnet/efcore
https://api.github.com/repos/dotnet/efcore
opened
Add code coverage report to AzDo
area-test type-enhancement
https://docs.microsoft.com/azure/devops/pipelines/tasks/test/publish-code-coverage-results https://docs.microsoft.com/dotnet/core/testing/unit-testing-code-coverage?tabs=windows ```dotnet test --collect:"XPlat Code Coverage"``` ```dotnet test /p:CollectCoverage=true /p:CoverletOutputFormat=cobertura``` Related: https://github.com/dotnet/efcore/issues/22589
1.0
Add code coverage report to AzDo - https://docs.microsoft.com/azure/devops/pipelines/tasks/test/publish-code-coverage-results https://docs.microsoft.com/dotnet/core/testing/unit-testing-code-coverage?tabs=windows ```dotnet test --collect:"XPlat Code Coverage"``` ```dotnet test /p:CollectCoverage=true /p:CoverletOutputFormat=cobertura``` Related: https://github.com/dotnet/efcore/issues/22589
non_process
add code coverage report to azdo dotnet test collect xplat code coverage dotnet test p collectcoverage true p coverletoutputformat cobertura related
0
12,689
15,054,223,932
IssuesEvent
2021-02-03 17:10:33
dzhw/zofar
https://api.github.com/repos/dzhw/zofar
closed
Storing paradata from access logs
category: secondary.exporter category: service.processes category: zofar status: discussion type: backlog.item
The complete access logs will be routinely stored in the future (either via web administrator or save directly in Zofar). The storage must not be made by default. It must be selectable, since a corresponding informed consent must be present for this.
1.0
Storing paradata from access logs - The complete access logs will be routinely stored in the future (either via web administrator or save directly in Zofar). The storage must not be made by default. It must be selectable, since a corresponding informed consent must be present for this.
process
storing paradata from access logs the complete access logs will be routinely stored in the future either via web administrator or save directly in zofar the storage must not be made by default it must be selectable since a corresponding informed consent must be present for this
1
260,865
22,674,472,542
IssuesEvent
2022-07-04 01:55:50
infor-design/enterprise
https://api.github.com/repos/infor-design/enterprise
closed
Cards: Convert Protractor Tests to Puppeteer
[2] type: unit testing for: dev
**Is your feature request related to a problem or use case? Please describe.** Convert Cards component tests to puppeteer. **Describe the solution you'd like** Convert all test scenarios in protractor to puppeteer. The naming can be, `{component}.puppeteer-spec.js` **Describe alternatives you've considered** N/A
1.0
Cards: Convert Protractor Tests to Puppeteer - **Is your feature request related to a problem or use case? Please describe.** Convert Cards component tests to puppeteer. **Describe the solution you'd like** Convert all test scenarios in protractor to puppeteer. The naming can be, `{component}.puppeteer-spec.js` **Describe alternatives you've considered** N/A
non_process
cards convert protractor tests to puppeteer is your feature request related to a problem or use case please describe convert cards component tests to puppeteer describe the solution you d like convert all test scenarios in protractor to puppeteer the naming can be component puppeteer spec js describe alternatives you ve considered n a
0
12,711
15,082,830,259
IssuesEvent
2021-02-05 15:04:04
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
4.0.0: StackOverflowError in error-prone
P4 team-Rules-Java type: support / not a bug (process)
### Description of the problem / feature request: Trying out [bazel 4.0.0](https://github.com/bazelbuild/bazel/issues/12455), with `--javabase=@bazel_tools//tools/jdk:remote_jdk11 --java_toolchain=@bazel_tools//tools/jdk:toolchain_java11` and hit a bug in error-prone 2.4.0 (bazel 3.3.1 has 2.3.2 I think, and it works) ### Bugs: what's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible. Haven't yet reproduced in a small setup but likely is triggered by code generated by https://github.com/immutables/immutables 2.8.8 ### Any other information, logs, or outputs that you want to share? Most recent error_prone is 2.5.1, but haven't tested if it'd fix the issue https://github.com/google/error-prone/releases ``` some_class.java:192: error: An unhandled exception was thrown by the Error Prone static analysis plugin. SomeClass.builder().from(someInstance).build()) Please report this at https://github.com/google/error-prone/issues/new and include the following: error-prone version: 2.4.0 BugPattern: ArrayToString Stack Trace: java.lang.StackOverflowError at jdk.compiler/com.sun.tools.javac.code.Types$UnaryVisitor.visit(Types.java:4937) at jdk.compiler/com.sun.tools.javac.code.Types.supertype(Types.java:2462) at jdk.compiler/com.sun.tools.javac.code.Types$12.visitClassType(Types.java:2128) at jdk.compiler/com.sun.tools.javac.code.Types$12.visitClassType(Types.java:2117) at jdk.compiler/com.sun.tools.javac.code.Type$ClassType.accept(Type.java:993) at jdk.compiler/com.sun.tools.javac.code.Types$DefaultTypeVisitor.visit(Types.java:4857) at jdk.compiler/com.sun.tools.javac.code.Types.asSuper(Types.java:2114) at jdk.compiler/com.sun.tools.javac.code.Types$12.visitClassType(Types.java:2130) at jdk.compiler/com.sun.tools.javac.code.Types$12.visitClassType(Types.java:2117) at jdk.compiler/com.sun.tools.javac.code.Type$ClassType.accept(Type.java:993) at jdk.compiler/com.sun.tools.javac.code.Types$DefaultTypeVisitor.visit(Types.java:4857) at jdk.compiler/com.sun.tools.javac.code.Types.asSuper(Types.java:2114) at jdk.compiler/com.sun.tools.javac.code.Types$12.visitClassType(Types.java:2130) at jdk.compiler/com.sun.tools.javac.code.Types$12.visitClassType(Types.java:2117) at jdk.compiler/com.sun.tools.javac.code.Type$ClassType.accept(Type.java:993) at jdk.compiler/com.sun.tools.javac.code.Types$DefaultTypeVisitor.visit(Types.java:4857) ... ``` Not sure if there's an easy way to upgrade error_prone version without the need to rebuild bazel itself Bump of error_prone commit https://github.com/bazelbuild/bazel/commit/239b2aab17cc1f007b2221ada9074bbe0c58db88 Some issues with earlies attempts of upgrade 2.3.2->2.3.3 https://github.com/bazelbuild/bazel/issues/9286
1.0
4.0.0: StackOverflowError in error-prone - ### Description of the problem / feature request: Trying out [bazel 4.0.0](https://github.com/bazelbuild/bazel/issues/12455), with `--javabase=@bazel_tools//tools/jdk:remote_jdk11 --java_toolchain=@bazel_tools//tools/jdk:toolchain_java11` and hit a bug in error-prone 2.4.0 (bazel 3.3.1 has 2.3.2 I think, and it works) ### Bugs: what's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible. Haven't yet reproduced in a small setup but likely is triggered by code generated by https://github.com/immutables/immutables 2.8.8 ### Any other information, logs, or outputs that you want to share? Most recent error_prone is 2.5.1, but haven't tested if it'd fix the issue https://github.com/google/error-prone/releases ``` some_class.java:192: error: An unhandled exception was thrown by the Error Prone static analysis plugin. SomeClass.builder().from(someInstance).build()) Please report this at https://github.com/google/error-prone/issues/new and include the following: error-prone version: 2.4.0 BugPattern: ArrayToString Stack Trace: java.lang.StackOverflowError at jdk.compiler/com.sun.tools.javac.code.Types$UnaryVisitor.visit(Types.java:4937) at jdk.compiler/com.sun.tools.javac.code.Types.supertype(Types.java:2462) at jdk.compiler/com.sun.tools.javac.code.Types$12.visitClassType(Types.java:2128) at jdk.compiler/com.sun.tools.javac.code.Types$12.visitClassType(Types.java:2117) at jdk.compiler/com.sun.tools.javac.code.Type$ClassType.accept(Type.java:993) at jdk.compiler/com.sun.tools.javac.code.Types$DefaultTypeVisitor.visit(Types.java:4857) at jdk.compiler/com.sun.tools.javac.code.Types.asSuper(Types.java:2114) at jdk.compiler/com.sun.tools.javac.code.Types$12.visitClassType(Types.java:2130) at jdk.compiler/com.sun.tools.javac.code.Types$12.visitClassType(Types.java:2117) at jdk.compiler/com.sun.tools.javac.code.Type$ClassType.accept(Type.java:993) at jdk.compiler/com.sun.tools.javac.code.Types$DefaultTypeVisitor.visit(Types.java:4857) at jdk.compiler/com.sun.tools.javac.code.Types.asSuper(Types.java:2114) at jdk.compiler/com.sun.tools.javac.code.Types$12.visitClassType(Types.java:2130) at jdk.compiler/com.sun.tools.javac.code.Types$12.visitClassType(Types.java:2117) at jdk.compiler/com.sun.tools.javac.code.Type$ClassType.accept(Type.java:993) at jdk.compiler/com.sun.tools.javac.code.Types$DefaultTypeVisitor.visit(Types.java:4857) ... ``` Not sure if there's an easy way to upgrade error_prone version without the need to rebuild bazel itself Bump of error_prone commit https://github.com/bazelbuild/bazel/commit/239b2aab17cc1f007b2221ada9074bbe0c58db88 Some issues with earlies attempts of upgrade 2.3.2->2.3.3 https://github.com/bazelbuild/bazel/issues/9286
process
stackoverflowerror in error prone description of the problem feature request trying out with javabase bazel tools tools jdk remote java toolchain bazel tools tools jdk toolchain and hit a bug in error prone bazel has i think and it works bugs what s the simplest easiest way to reproduce this bug please provide a minimal example if possible haven t yet reproduced in a small setup but likely is triggered by code generated by any other information logs or outputs that you want to share most recent error prone is but haven t tested if it d fix the issue some class java error an unhandled exception was thrown by the error prone static analysis plugin someclass builder from someinstance build please report this at and include the following error prone version bugpattern arraytostring stack trace java lang stackoverflowerror at jdk compiler com sun tools javac code types unaryvisitor visit types java at jdk compiler com sun tools javac code types supertype types java at jdk compiler com sun tools javac code types visitclasstype types java at jdk compiler com sun tools javac code types visitclasstype types java at jdk compiler com sun tools javac code type classtype accept type java at jdk compiler com sun tools javac code types defaulttypevisitor visit types java at jdk compiler com sun tools javac code types assuper types java at jdk compiler com sun tools javac code types visitclasstype types java at jdk compiler com sun tools javac code types visitclasstype types java at jdk compiler com sun tools javac code type classtype accept type java at jdk compiler com sun tools javac code types defaulttypevisitor visit types java at jdk compiler com sun tools javac code types assuper types java at jdk compiler com sun tools javac code types visitclasstype types java at jdk compiler com sun tools javac code types visitclasstype types java at jdk compiler com sun tools javac code type classtype accept type java at jdk compiler com sun tools javac code types defaulttypevisitor visit types java not sure if there s an easy way to upgrade error prone version without the need to rebuild bazel itself bump of error prone commit some issues with earlies attempts of upgrade
1
16,434
21,316,756,882
IssuesEvent
2022-04-16 12:16:17
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
Topics excluded using filter specified with ditavalref still appear in output folder
bug preprocess preprocess/branch-filtering stale
I'm attaching a sample DITA project. Although the DITAVAL excludes the "excluded_topic.dita", the generated HTML output folder contains an "excluded_topic.html". Tested with DITA OT 3.2.1 [ditaval issue sample.zip](https://github.com/dita-ot/dita-ot/files/2781611/ditaval.issue.sample.zip)
2.0
Topics excluded using filter specified with ditavalref still appear in output folder - I'm attaching a sample DITA project. Although the DITAVAL excludes the "excluded_topic.dita", the generated HTML output folder contains an "excluded_topic.html". Tested with DITA OT 3.2.1 [ditaval issue sample.zip](https://github.com/dita-ot/dita-ot/files/2781611/ditaval.issue.sample.zip)
process
topics excluded using filter specified with ditavalref still appear in output folder i m attaching a sample dita project although the ditaval excludes the excluded topic dita the generated html output folder contains an excluded topic html tested with dita ot
1
616,801
19,321,166,818
IssuesEvent
2021-12-14 05:51:33
kitabisa/teler
https://api.github.com/repos/kitabisa/teler
opened
[FEATURE] Zinc search engine
Type: Enhancement Status: On Hold Priority: Medium
**Is your feature request related to a problem? Please describe.** Supporting https://github.com/prabhatsharma/zinc **Describe the solution you'd like** A clear and concise description of what you want to happen. **Describe alternatives you've considered** A clear and concise description of any alternative solutions or features you've considered. **Additional context** Add any other context or screenshots about the feature request here.
1.0
[FEATURE] Zinc search engine - **Is your feature request related to a problem? Please describe.** Supporting https://github.com/prabhatsharma/zinc **Describe the solution you'd like** A clear and concise description of what you want to happen. **Describe alternatives you've considered** A clear and concise description of any alternative solutions or features you've considered. **Additional context** Add any other context or screenshots about the feature request here.
non_process
zinc search engine is your feature request related to a problem please describe supporting describe the solution you d like a clear and concise description of what you want to happen describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered additional context add any other context or screenshots about the feature request here
0
11,759
14,592,571,545
IssuesEvent
2020-12-19 18:15:32
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
QGIS / “create grid” consumes far more memory on linux than on windows
Bug Feedback Processing
I am using QGIS 3.10 both on windows and on Linux (3.10.8+28bionic) with the same RAM and CPU resources. When running the tool "create grid" QGIS behaves strange on Linux: For a certain extent (~ 100x100km), when creating a grid with 20m-100m cell size, memory consuming doesn't go beyond ~2,5-3GB. With 15m cell size, the calculations seem to exponentiate a lot - and it takes up to 30GB memory (not a steady increase, but rather abrupt) and it won't finish either. As my goal was initially a 10m grid - with this cellcize QGIS seems to crash a few minutes after running the tool, as memory goes up abruptly, however I don't any messages or warnings. On windows, running the same tool with the same parameters behaves differently: for the 10m grid only half of the memory (max. 7-8GB from 16GB) is used - I don't know what causes this strange behavior on Linux ... (?)
1.0
QGIS / “create grid” consumes far more memory on linux than on windows - I am using QGIS 3.10 both on windows and on Linux (3.10.8+28bionic) with the same RAM and CPU resources. When running the tool "create grid" QGIS behaves strange on Linux: For a certain extent (~ 100x100km), when creating a grid with 20m-100m cell size, memory consuming doesn't go beyond ~2,5-3GB. With 15m cell size, the calculations seem to exponentiate a lot - and it takes up to 30GB memory (not a steady increase, but rather abrupt) and it won't finish either. As my goal was initially a 10m grid - with this cellcize QGIS seems to crash a few minutes after running the tool, as memory goes up abruptly, however I don't any messages or warnings. On windows, running the same tool with the same parameters behaves differently: for the 10m grid only half of the memory (max. 7-8GB from 16GB) is used - I don't know what causes this strange behavior on Linux ... (?)
process
qgis “create grid” consumes far more memory on linux than on windows i am using qgis both on windows and on linux with the same ram and cpu resources when running the tool create grid qgis behaves strange on linux for a certain extent when creating a grid with cell size memory consuming doesn t go beyond with cell size the calculations seem to exponentiate a lot and it takes up to memory not a steady increase but rather abrupt and it won t finish either as my goal was initially a grid with this cellcize qgis seems to crash a few minutes after running the tool as memory goes up abruptly however i don t any messages or warnings on windows running the same tool with the same parameters behaves differently for the grid only half of the memory max from is used i don t know what causes this strange behavior on linux
1
331,808
10,077,038,474
IssuesEvent
2019-07-24 17:44:20
googleapis/google-cloud-dotnet
https://api.github.com/repos/googleapis/google-cloud-dotnet
closed
Synthesis failed for Google.LongRunning
autosynth failure priority: p1 type: bug
Hello! Autosynth couldn't regenerate Google.LongRunning. :broken_heart: Here's the output from running `synth.py`: ``` Cloning into 'working_repo'... Switched to branch 'autosynth-Google.LongRunning' Running synthtool ['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', 'synth.py', '--'] synthtool > Executing /tmpfs/src/git/autosynth/working_repo/apis/Google.LongRunning/synth.py. Cloning into 'gapic-generator'... Cloning into 'gapic-generator-csharp'... Submodule 'api-common-protos' (https://github.com/googleapis/api-common-protos.git) registered for path 'api-common-protos' Submodule 'protobuf' (https://github.com/protocolbuffers/protobuf.git) registered for path 'protobuf' Cloning into '/tmpfs/src/git/autosynth/working_repo/gapic-generator-csharp/api-common-protos'... Cloning into '/tmpfs/src/git/autosynth/working_repo/gapic-generator-csharp/protobuf'... Submodule path 'api-common-protos': checked out '4c0a203e3658ae0e56d47c817c2c5904116c0ae0' Submodule path 'protobuf': checked out '815ff7e1fb2d417d5aebcbf5fc46e626b18dc834' Submodule 'third_party/benchmark' (https://github.com/google/benchmark.git) registered for path 'protobuf/third_party/benchmark' Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'protobuf/third_party/googletest' Cloning into '/tmpfs/src/git/autosynth/working_repo/gapic-generator-csharp/protobuf/third_party/benchmark'... Cloning into '/tmpfs/src/git/autosynth/working_repo/gapic-generator-csharp/protobuf/third_party/googletest'... Submodule path 'protobuf/third_party/benchmark': checked out '5b7683f49e1e9223cf9927b24f6fd3d6bd82e3f8' Submodule path 'protobuf/third_party/googletest': checked out '5ec7f0c4a113e2f18ac2c6cc7df51ad6afc24081' Cloning into 'googleapis'... > Task :buildSrc:compileJava NO-SOURCE > Task :buildSrc:compileGroovy > Task :buildSrc:processResources NO-SOURCE > Task :buildSrc:classes > Task :buildSrc:jar > Task :buildSrc:assemble > Task :buildSrc:compileTestJava NO-SOURCE > Task :buildSrc:compileTestGroovy NO-SOURCE > Task :buildSrc:processTestResources NO-SOURCE > Task :buildSrc:testClasses UP-TO-DATE > Task :buildSrc:test NO-SOURCE > Task :buildSrc:check UP-TO-DATE > Task :buildSrc:build > Task :extractIncludeProto > Task :extractProto > Task :generateProto > Task :compileJava Note: Some input files use or override a deprecated API. Note: Recompile with -Xlint:deprecation for details. > Task :processResources > Task :createProperties > Task :classes > Task :shadowJar BUILD SUCCESSFUL in 13s 7 actionable tasks: 7 executed Generating Google.LongRunning Traceback (most recent call last): File "tools/getapifield.py", line 27, in <module> catalog = json.load(file) File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/json/__init__.py", line 299, in load parse_constant=parse_constant, object_pairs_hook=object_pairs_hook, **kw) File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/json/__init__.py", line 354, in loads return _default_decoder.decode(s) File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/json/decoder.py", line 339, in decode obj, end = self.raw_decode(s, idx=_w(s, 0).end()) File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/json/decoder.py", line 355, in raw_decode obj, end = self.scan_once(s, idx) json.decoder.JSONDecodeError: Expecting property name enclosed in double quotes: line 273 column 5 (char 10035) synthtool > Failed executing /bin/bash generateapis.sh Google.LongRunning: None Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/__main__.py", line 87, in <module> main() File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 764, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 717, in main rv = self.invoke(ctx) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 956, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 555, in invoke return callback(*args, **kwargs) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/__main__.py", line 79, in main spec.loader.exec_module(synth_module) # type: ignore File "<frozen importlib._bootstrap_external>", line 678, in exec_module File "<frozen importlib._bootstrap>", line 205, in _call_with_frames_removed File "/tmpfs/src/git/autosynth/working_repo/apis/Google.LongRunning/synth.py", line 20, in <module> hide_output = False) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/shell.py", line 39, in run raise exc File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/shell.py", line 33, in run encoding="utf-8", File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/subprocess.py", line 418, in run output=stdout, stderr=stderr) subprocess.CalledProcessError: Command '('/bin/bash', 'generateapis.sh', 'Google.LongRunning')' returned non-zero exit status 1. synthtool > Wrote metadata to synth.metadata. Synthesis failed ``` Google internal developers can see the full log [here](https://sponge/1abcf0a3-8980-4536-ac8b-8c3ce51b7c49).
1.0
Synthesis failed for Google.LongRunning - Hello! Autosynth couldn't regenerate Google.LongRunning. :broken_heart: Here's the output from running `synth.py`: ``` Cloning into 'working_repo'... Switched to branch 'autosynth-Google.LongRunning' Running synthtool ['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', 'synth.py', '--'] synthtool > Executing /tmpfs/src/git/autosynth/working_repo/apis/Google.LongRunning/synth.py. Cloning into 'gapic-generator'... Cloning into 'gapic-generator-csharp'... Submodule 'api-common-protos' (https://github.com/googleapis/api-common-protos.git) registered for path 'api-common-protos' Submodule 'protobuf' (https://github.com/protocolbuffers/protobuf.git) registered for path 'protobuf' Cloning into '/tmpfs/src/git/autosynth/working_repo/gapic-generator-csharp/api-common-protos'... Cloning into '/tmpfs/src/git/autosynth/working_repo/gapic-generator-csharp/protobuf'... Submodule path 'api-common-protos': checked out '4c0a203e3658ae0e56d47c817c2c5904116c0ae0' Submodule path 'protobuf': checked out '815ff7e1fb2d417d5aebcbf5fc46e626b18dc834' Submodule 'third_party/benchmark' (https://github.com/google/benchmark.git) registered for path 'protobuf/third_party/benchmark' Submodule 'third_party/googletest' (https://github.com/google/googletest.git) registered for path 'protobuf/third_party/googletest' Cloning into '/tmpfs/src/git/autosynth/working_repo/gapic-generator-csharp/protobuf/third_party/benchmark'... Cloning into '/tmpfs/src/git/autosynth/working_repo/gapic-generator-csharp/protobuf/third_party/googletest'... Submodule path 'protobuf/third_party/benchmark': checked out '5b7683f49e1e9223cf9927b24f6fd3d6bd82e3f8' Submodule path 'protobuf/third_party/googletest': checked out '5ec7f0c4a113e2f18ac2c6cc7df51ad6afc24081' Cloning into 'googleapis'... > Task :buildSrc:compileJava NO-SOURCE > Task :buildSrc:compileGroovy > Task :buildSrc:processResources NO-SOURCE > Task :buildSrc:classes > Task :buildSrc:jar > Task :buildSrc:assemble > Task :buildSrc:compileTestJava NO-SOURCE > Task :buildSrc:compileTestGroovy NO-SOURCE > Task :buildSrc:processTestResources NO-SOURCE > Task :buildSrc:testClasses UP-TO-DATE > Task :buildSrc:test NO-SOURCE > Task :buildSrc:check UP-TO-DATE > Task :buildSrc:build > Task :extractIncludeProto > Task :extractProto > Task :generateProto > Task :compileJava Note: Some input files use or override a deprecated API. Note: Recompile with -Xlint:deprecation for details. > Task :processResources > Task :createProperties > Task :classes > Task :shadowJar BUILD SUCCESSFUL in 13s 7 actionable tasks: 7 executed Generating Google.LongRunning Traceback (most recent call last): File "tools/getapifield.py", line 27, in <module> catalog = json.load(file) File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/json/__init__.py", line 299, in load parse_constant=parse_constant, object_pairs_hook=object_pairs_hook, **kw) File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/json/__init__.py", line 354, in loads return _default_decoder.decode(s) File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/json/decoder.py", line 339, in decode obj, end = self.raw_decode(s, idx=_w(s, 0).end()) File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/json/decoder.py", line 355, in raw_decode obj, end = self.scan_once(s, idx) json.decoder.JSONDecodeError: Expecting property name enclosed in double quotes: line 273 column 5 (char 10035) synthtool > Failed executing /bin/bash generateapis.sh Google.LongRunning: None Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/__main__.py", line 87, in <module> main() File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 764, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 717, in main rv = self.invoke(ctx) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 956, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 555, in invoke return callback(*args, **kwargs) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/__main__.py", line 79, in main spec.loader.exec_module(synth_module) # type: ignore File "<frozen importlib._bootstrap_external>", line 678, in exec_module File "<frozen importlib._bootstrap>", line 205, in _call_with_frames_removed File "/tmpfs/src/git/autosynth/working_repo/apis/Google.LongRunning/synth.py", line 20, in <module> hide_output = False) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/shell.py", line 39, in run raise exc File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/shell.py", line 33, in run encoding="utf-8", File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/subprocess.py", line 418, in run output=stdout, stderr=stderr) subprocess.CalledProcessError: Command '('/bin/bash', 'generateapis.sh', 'Google.LongRunning')' returned non-zero exit status 1. synthtool > Wrote metadata to synth.metadata. Synthesis failed ``` Google internal developers can see the full log [here](https://sponge/1abcf0a3-8980-4536-ac8b-8c3ce51b7c49).
non_process
synthesis failed for google longrunning hello autosynth couldn t regenerate google longrunning broken heart here s the output from running synth py cloning into working repo switched to branch autosynth google longrunning running synthtool synthtool executing tmpfs src git autosynth working repo apis google longrunning synth py cloning into gapic generator cloning into gapic generator csharp submodule api common protos registered for path api common protos submodule protobuf registered for path protobuf cloning into tmpfs src git autosynth working repo gapic generator csharp api common protos cloning into tmpfs src git autosynth working repo gapic generator csharp protobuf submodule path api common protos checked out submodule path protobuf checked out submodule third party benchmark registered for path protobuf third party benchmark submodule third party googletest registered for path protobuf third party googletest cloning into tmpfs src git autosynth working repo gapic generator csharp protobuf third party benchmark cloning into tmpfs src git autosynth working repo gapic generator csharp protobuf third party googletest submodule path protobuf third party benchmark checked out submodule path protobuf third party googletest checked out cloning into googleapis task buildsrc compilejava no source task buildsrc compilegroovy task buildsrc processresources no source task buildsrc classes task buildsrc jar task buildsrc assemble task buildsrc compiletestjava no source task buildsrc compiletestgroovy no source task buildsrc processtestresources no source task buildsrc testclasses up to date task buildsrc test no source task buildsrc check up to date task buildsrc build task extractincludeproto task extractproto task generateproto task compilejava note some input files use or override a deprecated api note recompile with xlint deprecation for details task processresources task createproperties task classes task shadowjar build successful in actionable tasks executed generating google longrunning traceback most recent call last file tools getapifield py line in catalog json load file file home kbuilder pyenv versions lib json init py line in load parse constant parse constant object pairs hook object pairs hook kw file home kbuilder pyenv versions lib json init py line in loads return default decoder decode s file home kbuilder pyenv versions lib json decoder py line in decode obj end self raw decode s idx w s end file home kbuilder pyenv versions lib json decoder py line in raw decode obj end self scan once s idx json decoder jsondecodeerror expecting property name enclosed in double quotes line column char synthtool failed executing bin bash generateapis sh google longrunning none traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src git autosynth env lib site packages synthtool main py line in main file tmpfs src git autosynth env lib site packages click core py line in call return self main args kwargs file tmpfs src git autosynth env lib site packages click core py line in main rv self invoke ctx file tmpfs src git autosynth env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src git autosynth env lib site packages click core py line in invoke return callback args kwargs file tmpfs src git autosynth env lib site packages synthtool main py line in main spec loader exec module synth module type ignore file line in exec module file line in call with frames removed file tmpfs src git autosynth working repo apis google longrunning synth py line in hide output false file tmpfs src git autosynth env lib site packages synthtool shell py line in run raise exc file tmpfs src git autosynth env lib site packages synthtool shell py line in run encoding utf file home kbuilder pyenv versions lib subprocess py line in run output stdout stderr stderr subprocess calledprocesserror command bin bash generateapis sh google longrunning returned non zero exit status synthtool wrote metadata to synth metadata synthesis failed google internal developers can see the full log
0
15,896
20,101,994,071
IssuesEvent
2022-02-07 06:02:57
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Copy/paste entry model in the graphical modeler keep old name even it has been changed when pasting it
Processing Bug Modeller
Hi, There is a problem in graphical modeler. I copy-paste a bit of model from Model 1 to Model 2. I renamed all the model input and algorithm's algorithm in the model 2 but it's impossible to launch the model due to an error of finding an input layer corresponding of my Model 1. It ask me 'Table name Model 1' when I put 'Table name Model 2'. Exporting the model in script format, the name of the table is 'Table name Model 1'. Here we can see that in the script for this table, it registers both names : self.addParameter(QgsProcessingParameterVectorLayer('Table name Model 1', 'Table name Model 2', types=[QgsProcessing.TypeVector], defaultValue=None)) But for launching the model, it takes only the first name ('Model 1'). Why, is there keep the first name value ? In the attached model, 'Table name Model 1 = 'tableFITTINGdinera' and [QgisBug.zip](https://github.com/qgis/QGIS/files/5475309/QgisBug.zip) 'Table name Model 2 = 'Excel Canalyse'
1.0
Copy/paste entry model in the graphical modeler keep old name even it has been changed when pasting it - Hi, There is a problem in graphical modeler. I copy-paste a bit of model from Model 1 to Model 2. I renamed all the model input and algorithm's algorithm in the model 2 but it's impossible to launch the model due to an error of finding an input layer corresponding of my Model 1. It ask me 'Table name Model 1' when I put 'Table name Model 2'. Exporting the model in script format, the name of the table is 'Table name Model 1'. Here we can see that in the script for this table, it registers both names : self.addParameter(QgsProcessingParameterVectorLayer('Table name Model 1', 'Table name Model 2', types=[QgsProcessing.TypeVector], defaultValue=None)) But for launching the model, it takes only the first name ('Model 1'). Why, is there keep the first name value ? In the attached model, 'Table name Model 1 = 'tableFITTINGdinera' and [QgisBug.zip](https://github.com/qgis/QGIS/files/5475309/QgisBug.zip) 'Table name Model 2 = 'Excel Canalyse'
process
copy paste entry model in the graphical modeler keep old name even it has been changed when pasting it hi there is a problem in graphical modeler i copy paste a bit of model from model to model i renamed all the model input and algorithm s algorithm in the model but it s impossible to launch the model due to an error of finding an input layer corresponding of my model it ask me table name model when i put table name model exporting the model in script format the name of the table is table name model here we can see that in the script for this table it registers both names self addparameter qgsprocessingparametervectorlayer table name model table name model types defaultvalue none but for launching the model it takes only the first name model why is there keep the first name value in the attached model table name model tablefittingdinera and table name model excel canalyse
1
15,059
18,763,412,741
IssuesEvent
2021-11-05 19:28:01
ORNL-AMO/AMO-Tools-Desktop
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop
opened
Process Heat Flue gas unit conversion
bug Process Heating
Ambient Air Temp and Fuel Temp do not have unit conversions for when converting the assessment (should be same as other temperatures in form F to C) Honestly, Fuel temp has been around long enough and not had a unit conversion, if this doesn't get done before the release, it is not the end of the world
1.0
Process Heat Flue gas unit conversion - Ambient Air Temp and Fuel Temp do not have unit conversions for when converting the assessment (should be same as other temperatures in form F to C) Honestly, Fuel temp has been around long enough and not had a unit conversion, if this doesn't get done before the release, it is not the end of the world
process
process heat flue gas unit conversion ambient air temp and fuel temp do not have unit conversions for when converting the assessment should be same as other temperatures in form f to c honestly fuel temp has been around long enough and not had a unit conversion if this doesn t get done before the release it is not the end of the world
1
45,576
7,190,699,372
IssuesEvent
2018-02-02 18:12:27
openshift/origin
https://api.github.com/repos/openshift/origin
closed
Move CONTRIBUTING.adoc to a series of getting-started guides.
area/documentation kind/enhancement priority/P3
_Problem_ - The simplest way to setup openshift from source is DIND... however, its at the bottom of the CONTRIBUTING.adoc. - The running tests section is at the bottom of the doc, so anyone wanting to run tests and setup a cluster has to read the top of the doc, and then skip to the bottom. Its not clear where dev instructions begin and test instructions begin. - Some mechanisms (like vagrant-dind) can be done super easily, and are unfortunately buried in the same text as the complex local dev setup instructions. _Solution_ The common solution to all of this: Separate the dev workflows into separate files, like this: I'd suggest something like ``` - contributing - local.md - dind.md - vagrant-dind.md - openshift-ansible.md - running-tests.md ``` ... @marun sounds ok to you? This would make it super easy to add new wacky testing scenarios (i.e. ec2, openstack-vagrant, whatever) without making the existing instructions increasingly complex.
1.0
Move CONTRIBUTING.adoc to a series of getting-started guides. - _Problem_ - The simplest way to setup openshift from source is DIND... however, its at the bottom of the CONTRIBUTING.adoc. - The running tests section is at the bottom of the doc, so anyone wanting to run tests and setup a cluster has to read the top of the doc, and then skip to the bottom. Its not clear where dev instructions begin and test instructions begin. - Some mechanisms (like vagrant-dind) can be done super easily, and are unfortunately buried in the same text as the complex local dev setup instructions. _Solution_ The common solution to all of this: Separate the dev workflows into separate files, like this: I'd suggest something like ``` - contributing - local.md - dind.md - vagrant-dind.md - openshift-ansible.md - running-tests.md ``` ... @marun sounds ok to you? This would make it super easy to add new wacky testing scenarios (i.e. ec2, openstack-vagrant, whatever) without making the existing instructions increasingly complex.
non_process
move contributing adoc to a series of getting started guides problem the simplest way to setup openshift from source is dind however its at the bottom of the contributing adoc the running tests section is at the bottom of the doc so anyone wanting to run tests and setup a cluster has to read the top of the doc and then skip to the bottom its not clear where dev instructions begin and test instructions begin some mechanisms like vagrant dind can be done super easily and are unfortunately buried in the same text as the complex local dev setup instructions solution the common solution to all of this separate the dev workflows into separate files like this i d suggest something like contributing local md dind md vagrant dind md openshift ansible md running tests md marun sounds ok to you this would make it super easy to add new wacky testing scenarios i e openstack vagrant whatever without making the existing instructions increasingly complex
0
5,649
8,513,602,253
IssuesEvent
2018-10-31 16:25:50
easy-software-ufal/annotations_repos
https://api.github.com/repos/easy-software-ufal/annotations_repos
opened
aspnet/Routing No route candidate coming back to Mvc when it should be resolvable
C# RPV test wrong processing
Issue: `https://github.com/aspnet/Routing/issues/316` PR: `https://github.com/aspnet/Routing/commit/dd79d0b88ccca57a45452c6630f8ed1056829227`
1.0
aspnet/Routing No route candidate coming back to Mvc when it should be resolvable - Issue: `https://github.com/aspnet/Routing/issues/316` PR: `https://github.com/aspnet/Routing/commit/dd79d0b88ccca57a45452c6630f8ed1056829227`
process
aspnet routing no route candidate coming back to mvc when it should be resolvable issue pr
1
12,358
14,887,329,885
IssuesEvent
2021-01-20 18:10:46
panther-labs/panther
https://api.github.com/repos/panther-labs/panther
opened
Upgrade AWS Go SDK to v2
p2 team:core infra team:data analytics team:data processing team:security engineering
### Description AWS Go SDK v2 is now [generally available](https://aws.amazon.com/about-aws/whats-new/2021/01/aws-sdk-for-go-version-2-now-generally-available/)! There are substantial performance improvements in the new library (which, in turn, lead to cost savings), so it's definitely worth our time to switch sooner than later. We'll wait a few weeks for them to patch any issues before we start upgrading Panther services. The upgrade need not happen all at once - each service owner can migrate their own service when they have a few spare cycles to do so. This epic is not necessarily anyone's dedicated responsibility yet; it's just a convenient place to track which services have been upgraded and any roadblocks we encounter along the way. See [migration docs](https://aws.github.io/aws-sdk-go-v2/docs/migrating/) ### Acceptance Criteria - Eventually, all aws go sdk imports are switched to `github.com/aws/aws-sdk-go-v2` Check off each service once the migration is complete: #### `cmd` - [ ] `devtools` - [ ] `opstools` #### `internal/compliance` - [ ] `alert-forwarder` - [ ] `alert-processor` - [ ] `aws-event-processor` - [ ] `compliance-api` - [ ] `datalake-forwarder` - [ ] `remediation-api` - [ ] `remediation-processor` - [ ] `resources-api` - [ ] `snapshot-poller` - [ ] `snapshot-scheduler` #### `internal/core` - [ ] `alert-delivery` - [ ] `analysis-api` - [ ] `custom-resources` - [ ] `database-api` (enterprise) - [ ] `layer-manager` - [ ] `logtypesapi` - [ ] `metrics-api` - [ ] `organization-api` - [ ] `outputs-api` - [ ] `source-api` - [ ] `users-api` #### `internal/log-analysis` - [ ] `alert-forwarder` - [ ] `alerts-api` - [ ] `datacatalog-compactor` (enterprise) - [ ] `datacatalog-compactor-reaper` (enterprise) - [ ] `datacatalog-updater` - [ ] `log-processor` - [ ] `log-puller` (enterprise) - [ ] `message-forwarder` #### `pkg` - [ ] `awsathena` - [ ] `awsbatch` - [ ] `awscfn` - [ ] `awscostexplorer` - [ ] `awsretry` - [ ] `awssqs` - [ ] `awsutils` - [ ] `box` - [ ] `encryption` - [ ] `extract` - [ ] `gatewayapi` - [ ] `genericapi` - [ ] `lambdalogger` - [ ] `metrics` - [ ] `oplog` - [ ] `priorityq` - [ ] `prompt` - [ ] `shutil` - [ ] `stringset` - [ ] `testutils` - [ ] `unbox` - [ ] `x` #### `tools` - [ ] `mage`
1.0
Upgrade AWS Go SDK to v2 - ### Description AWS Go SDK v2 is now [generally available](https://aws.amazon.com/about-aws/whats-new/2021/01/aws-sdk-for-go-version-2-now-generally-available/)! There are substantial performance improvements in the new library (which, in turn, lead to cost savings), so it's definitely worth our time to switch sooner than later. We'll wait a few weeks for them to patch any issues before we start upgrading Panther services. The upgrade need not happen all at once - each service owner can migrate their own service when they have a few spare cycles to do so. This epic is not necessarily anyone's dedicated responsibility yet; it's just a convenient place to track which services have been upgraded and any roadblocks we encounter along the way. See [migration docs](https://aws.github.io/aws-sdk-go-v2/docs/migrating/) ### Acceptance Criteria - Eventually, all aws go sdk imports are switched to `github.com/aws/aws-sdk-go-v2` Check off each service once the migration is complete: #### `cmd` - [ ] `devtools` - [ ] `opstools` #### `internal/compliance` - [ ] `alert-forwarder` - [ ] `alert-processor` - [ ] `aws-event-processor` - [ ] `compliance-api` - [ ] `datalake-forwarder` - [ ] `remediation-api` - [ ] `remediation-processor` - [ ] `resources-api` - [ ] `snapshot-poller` - [ ] `snapshot-scheduler` #### `internal/core` - [ ] `alert-delivery` - [ ] `analysis-api` - [ ] `custom-resources` - [ ] `database-api` (enterprise) - [ ] `layer-manager` - [ ] `logtypesapi` - [ ] `metrics-api` - [ ] `organization-api` - [ ] `outputs-api` - [ ] `source-api` - [ ] `users-api` #### `internal/log-analysis` - [ ] `alert-forwarder` - [ ] `alerts-api` - [ ] `datacatalog-compactor` (enterprise) - [ ] `datacatalog-compactor-reaper` (enterprise) - [ ] `datacatalog-updater` - [ ] `log-processor` - [ ] `log-puller` (enterprise) - [ ] `message-forwarder` #### `pkg` - [ ] `awsathena` - [ ] `awsbatch` - [ ] `awscfn` - [ ] `awscostexplorer` - [ ] `awsretry` - [ ] `awssqs` - [ ] `awsutils` - [ ] `box` - [ ] `encryption` - [ ] `extract` - [ ] `gatewayapi` - [ ] `genericapi` - [ ] `lambdalogger` - [ ] `metrics` - [ ] `oplog` - [ ] `priorityq` - [ ] `prompt` - [ ] `shutil` - [ ] `stringset` - [ ] `testutils` - [ ] `unbox` - [ ] `x` #### `tools` - [ ] `mage`
process
upgrade aws go sdk to description aws go sdk is now there are substantial performance improvements in the new library which in turn lead to cost savings so it s definitely worth our time to switch sooner than later we ll wait a few weeks for them to patch any issues before we start upgrading panther services the upgrade need not happen all at once each service owner can migrate their own service when they have a few spare cycles to do so this epic is not necessarily anyone s dedicated responsibility yet it s just a convenient place to track which services have been upgraded and any roadblocks we encounter along the way see acceptance criteria eventually all aws go sdk imports are switched to github com aws aws sdk go check off each service once the migration is complete cmd devtools opstools internal compliance alert forwarder alert processor aws event processor compliance api datalake forwarder remediation api remediation processor resources api snapshot poller snapshot scheduler internal core alert delivery analysis api custom resources database api enterprise layer manager logtypesapi metrics api organization api outputs api source api users api internal log analysis alert forwarder alerts api datacatalog compactor enterprise datacatalog compactor reaper enterprise datacatalog updater log processor log puller enterprise message forwarder pkg awsathena awsbatch awscfn awscostexplorer awsretry awssqs awsutils box encryption extract gatewayapi genericapi lambdalogger metrics oplog priorityq prompt shutil stringset testutils unbox x tools mage
1
90,921
11,452,000,747
IssuesEvent
2020-02-06 12:50:29
rl-institut/E_Metrobus
https://api.github.com/repos/rl-institut/E_Metrobus
opened
Answer page's layout is not optimal
design
- [ ] Add space between text and images - [ ] Fix issue with button overlapping content ![localhost_3000_answer_(iPhone 6_7_8)](https://user-images.githubusercontent.com/33999424/73938423-8721b580-48e7-11ea-98f8-3a01f323d66e.png)
1.0
Answer page's layout is not optimal - - [ ] Add space between text and images - [ ] Fix issue with button overlapping content ![localhost_3000_answer_(iPhone 6_7_8)](https://user-images.githubusercontent.com/33999424/73938423-8721b580-48e7-11ea-98f8-3a01f323d66e.png)
non_process
answer page s layout is not optimal add space between text and images fix issue with button overlapping content
0
167,007
20,725,679,713
IssuesEvent
2022-03-14 01:21:51
rgordon95/ConFusionBootStrap
https://api.github.com/repos/rgordon95/ConFusionBootStrap
opened
CVE-2021-37701 (High) detected in tar-2.2.1.tgz
security vulnerability
## CVE-2021-37701 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-2.2.1.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-2.2.1.tgz">https://registry.npmjs.org/tar/-/tar-2.2.1.tgz</a></p> <p>Path to dependency file: /ConFusionBootStrap/conFusion/package.json</p> <p>Path to vulnerable library: /conFusion/node_modules/tar/package.json</p> <p> Dependency Hierarchy: - node-sass-4.12.0.tgz (Root Library) - node-gyp-3.8.0.tgz - :x: **tar-2.2.1.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The npm package "tar" (aka node-tar) before versions 4.4.16, 5.0.8, and 6.1.7 has an arbitrary file creation/overwrite and arbitrary code execution vulnerability. node-tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted. This is, in part, achieved by ensuring that extracted directories are not symlinks. Additionally, in order to prevent unnecessary stat calls to determine whether a given path is a directory, paths are cached when directories are created. This logic was insufficient when extracting tar files that contained both a directory and a symlink with the same name as the directory, where the symlink and directory names in the archive entry used backslashes as a path separator on posix systems. The cache checking logic used both `\` and `/` characters as path separators, however `\` is a valid filename character on posix systems. By first creating a directory, and then replacing that directory with a symlink, it was thus possible to bypass node-tar symlink checks on directories, essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location, thus allowing arbitrary file creation and overwrite. Additionally, a similar confusion could arise on case-insensitive filesystems. If a tar archive contained a directory at `FOO`, followed by a symbolic link named `foo`, then on case-insensitive file systems, the creation of the symbolic link would remove the directory from the filesystem, but _not_ from the internal directory cache, as it would not be treated as a cache hit. A subsequent file entry within the `FOO` directory would then be placed in the target of the symbolic link, thinking that the directory had already been created. These issues were addressed in releases 4.4.16, 5.0.8 and 6.1.7. The v3 branch of node-tar has been deprecated and did not receive patches for these issues. If you are still using a v3 release we recommend you update to a more recent version of node-tar. If this is not possible, a workaround is available in the referenced GHSA-9r2w-394v-53qc. <p>Publish Date: 2021-08-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37701>CVE-2021-37701</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-9r2w-394v-53qc">https://github.com/npm/node-tar/security/advisories/GHSA-9r2w-394v-53qc</a></p> <p>Release Date: 2021-08-31</p> <p>Fix Resolution (tar): 4.4.16</p> <p>Direct dependency fix Resolution (node-sass): 5.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-37701 (High) detected in tar-2.2.1.tgz - ## CVE-2021-37701 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-2.2.1.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-2.2.1.tgz">https://registry.npmjs.org/tar/-/tar-2.2.1.tgz</a></p> <p>Path to dependency file: /ConFusionBootStrap/conFusion/package.json</p> <p>Path to vulnerable library: /conFusion/node_modules/tar/package.json</p> <p> Dependency Hierarchy: - node-sass-4.12.0.tgz (Root Library) - node-gyp-3.8.0.tgz - :x: **tar-2.2.1.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The npm package "tar" (aka node-tar) before versions 4.4.16, 5.0.8, and 6.1.7 has an arbitrary file creation/overwrite and arbitrary code execution vulnerability. node-tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted. This is, in part, achieved by ensuring that extracted directories are not symlinks. Additionally, in order to prevent unnecessary stat calls to determine whether a given path is a directory, paths are cached when directories are created. This logic was insufficient when extracting tar files that contained both a directory and a symlink with the same name as the directory, where the symlink and directory names in the archive entry used backslashes as a path separator on posix systems. The cache checking logic used both `\` and `/` characters as path separators, however `\` is a valid filename character on posix systems. By first creating a directory, and then replacing that directory with a symlink, it was thus possible to bypass node-tar symlink checks on directories, essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location, thus allowing arbitrary file creation and overwrite. Additionally, a similar confusion could arise on case-insensitive filesystems. If a tar archive contained a directory at `FOO`, followed by a symbolic link named `foo`, then on case-insensitive file systems, the creation of the symbolic link would remove the directory from the filesystem, but _not_ from the internal directory cache, as it would not be treated as a cache hit. A subsequent file entry within the `FOO` directory would then be placed in the target of the symbolic link, thinking that the directory had already been created. These issues were addressed in releases 4.4.16, 5.0.8 and 6.1.7. The v3 branch of node-tar has been deprecated and did not receive patches for these issues. If you are still using a v3 release we recommend you update to a more recent version of node-tar. If this is not possible, a workaround is available in the referenced GHSA-9r2w-394v-53qc. <p>Publish Date: 2021-08-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37701>CVE-2021-37701</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-9r2w-394v-53qc">https://github.com/npm/node-tar/security/advisories/GHSA-9r2w-394v-53qc</a></p> <p>Release Date: 2021-08-31</p> <p>Fix Resolution (tar): 4.4.16</p> <p>Direct dependency fix Resolution (node-sass): 5.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in tar tgz cve high severity vulnerability vulnerable library tar tgz tar for node library home page a href path to dependency file confusionbootstrap confusion package json path to vulnerable library confusion node modules tar package json dependency hierarchy node sass tgz root library node gyp tgz x tar tgz vulnerable library vulnerability details the npm package tar aka node tar before versions and has an arbitrary file creation overwrite and arbitrary code execution vulnerability node tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted this is in part achieved by ensuring that extracted directories are not symlinks additionally in order to prevent unnecessary stat calls to determine whether a given path is a directory paths are cached when directories are created this logic was insufficient when extracting tar files that contained both a directory and a symlink with the same name as the directory where the symlink and directory names in the archive entry used backslashes as a path separator on posix systems the cache checking logic used both and characters as path separators however is a valid filename character on posix systems by first creating a directory and then replacing that directory with a symlink it was thus possible to bypass node tar symlink checks on directories essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location thus allowing arbitrary file creation and overwrite additionally a similar confusion could arise on case insensitive filesystems if a tar archive contained a directory at foo followed by a symbolic link named foo then on case insensitive file systems the creation of the symbolic link would remove the directory from the filesystem but not from the internal directory cache as it would not be treated as a cache hit a subsequent file entry within the foo directory would then be placed in the target of the symbolic link thinking that the directory had already been created these issues were addressed in releases and the branch of node tar has been deprecated and did not receive patches for these issues if you are still using a release we recommend you update to a more recent version of node tar if this is not possible a workaround is available in the referenced ghsa publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tar direct dependency fix resolution node sass step up your open source security game with whitesource
0
7,160
10,308,668,303
IssuesEvent
2019-08-29 11:32:23
Hypertopic/Argos
https://api.github.com/repos/Hypertopic/Argos
closed
Tests should pass on Travis CI
Process
A security fix for CouchDB (1.6.0 -> 1.6.1) broke the continuous integration of Argos. The reason for that is that using HTTP for settings is now disabled. In Travis docs, I found no other way to change the settings of CouchDB. A solution could be to use a dockerized version of CouchDB instead of the default one.
1.0
Tests should pass on Travis CI - A security fix for CouchDB (1.6.0 -> 1.6.1) broke the continuous integration of Argos. The reason for that is that using HTTP for settings is now disabled. In Travis docs, I found no other way to change the settings of CouchDB. A solution could be to use a dockerized version of CouchDB instead of the default one.
process
tests should pass on travis ci a security fix for couchdb broke the continuous integration of argos the reason for that is that using http for settings is now disabled in travis docs i found no other way to change the settings of couchdb a solution could be to use a dockerized version of couchdb instead of the default one
1
1,100
3,575,207,759
IssuesEvent
2016-01-27 15:11:31
nodejs/node
https://api.github.com/repos/nodejs/node
closed
Add child_process.spawnShell
child_process feature request windows
In the child_process module currently there is two flavors of exec function: exec and execFile, which is good, but for spawn, there is only one version, would've been great to have something like child_process.spawnShell. Currently I use this code to emulate it: ```javascript child_process.spawn('cmd', ['/s', '/c', command], {windowsVerbatimArguments:true}) ``` but it is platform specific and requires the use of undocumented flag windowsVerbatimArguments
1.0
Add child_process.spawnShell - In the child_process module currently there is two flavors of exec function: exec and execFile, which is good, but for spawn, there is only one version, would've been great to have something like child_process.spawnShell. Currently I use this code to emulate it: ```javascript child_process.spawn('cmd', ['/s', '/c', command], {windowsVerbatimArguments:true}) ``` but it is platform specific and requires the use of undocumented flag windowsVerbatimArguments
process
add child process spawnshell in the child process module currently there is two flavors of exec function exec and execfile which is good but for spawn there is only one version would ve been great to have something like child process spawnshell currently i use this code to emulate it javascript child process spawn cmd windowsverbatimarguments true but it is platform specific and requires the use of undocumented flag windowsverbatimarguments
1
321,046
9,792,727,249
IssuesEvent
2019-06-10 18:08:27
DFO-Ocean-Navigator/Ocean-Data-Map-Project
https://api.github.com/repos/DFO-Ocean-Navigator/Ocean-Data-Map-Project
closed
Python environment clean up
Priority: Medium Python
The environment.yml file listing the python packages does not currently work (some specific versions are no longer available).
1.0
Python environment clean up - The environment.yml file listing the python packages does not currently work (some specific versions are no longer available).
non_process
python environment clean up the environment yml file listing the python packages does not currently work some specific versions are no longer available
0
7,688
10,774,730,303
IssuesEvent
2019-11-03 08:55:34
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Export to python doesn't follow the "Advanced" panel settings in QGIS 3.10 Graphical Modeler
Bug Processing
I have created a model with some fields marked as "Advanced" in QGIS 3.10 Graphical Modeler. After conversion to a Python script the "Advanced" flags are missing.
1.0
Export to python doesn't follow the "Advanced" panel settings in QGIS 3.10 Graphical Modeler - I have created a model with some fields marked as "Advanced" in QGIS 3.10 Graphical Modeler. After conversion to a Python script the "Advanced" flags are missing.
process
export to python doesn t follow the advanced panel settings in qgis graphical modeler i have created a model with some fields marked as advanced in qgis graphical modeler after conversion to a python script the advanced flags are missing
1
366,495
10,821,538,772
IssuesEvent
2019-11-08 18:55:39
oppia/oppia-android
https://api.github.com/repos/oppia/oppia-android
opened
Paused activity does not get LiveData update onResume
Priority: Important Type: Bug
In ProfileActivity, ProfileChooserFragment's list of profiles does not get updated when returning to the activity when another activity updates ProfileDatabase.
1.0
Paused activity does not get LiveData update onResume - In ProfileActivity, ProfileChooserFragment's list of profiles does not get updated when returning to the activity when another activity updates ProfileDatabase.
non_process
paused activity does not get livedata update onresume in profileactivity profilechooserfragment s list of profiles does not get updated when returning to the activity when another activity updates profiledatabase
0
501,329
14,525,752,035
IssuesEvent
2020-12-14 13:22:16
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.whatsapp.com - desktop site instead of mobile site
browser-focus-geckoview engine-gecko priority-critical
<!-- @browser: Firefox Mobile 84.0 --> <!-- @ua_header: Mozilla/5.0 (Android 7.0; Mobile; rv:84.0) Gecko/84.0 Firefox/84.0 --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/63567 --> <!-- @extra_labels: browser-focus-geckoview --> **URL**: https://www.whatsapp.com/ **Browser / Version**: Firefox Mobile 84.0 **Operating System**: Android 7.0 **Tested Another Browser**: Yes Chrome **Problem type**: Desktop site instead of mobile site **Description**: Desktop site instead of mobile site **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.whatsapp.com - desktop site instead of mobile site - <!-- @browser: Firefox Mobile 84.0 --> <!-- @ua_header: Mozilla/5.0 (Android 7.0; Mobile; rv:84.0) Gecko/84.0 Firefox/84.0 --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/63567 --> <!-- @extra_labels: browser-focus-geckoview --> **URL**: https://www.whatsapp.com/ **Browser / Version**: Firefox Mobile 84.0 **Operating System**: Android 7.0 **Tested Another Browser**: Yes Chrome **Problem type**: Desktop site instead of mobile site **Description**: Desktop site instead of mobile site **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
desktop site instead of mobile site url browser version firefox mobile operating system android tested another browser yes chrome problem type desktop site instead of mobile site description desktop site instead of mobile site steps to reproduce browser configuration none from with ❤️
0
9,922
12,962,304,318
IssuesEvent
2020-07-20 16:56:41
department-of-veterans-affairs/notification-api
https://api.github.com/repos/department-of-veterans-affairs/notification-api
opened
Design VA Profile and MPI integration
Process Task
As a developer I want to ensure alignment on technical design for a new integration with internal VA teams As VANotify I want to enable other teams ... ... As VANotify
We want to ensure we build an API 
So that interested teams can integrate with VANotify to send notifications
 **Outcomes:** - Create diagrams that depict sequence and data flow - Review and discuss with the team - Open API documentation of the new API that utilizes VA identifiers
1.0
Design VA Profile and MPI integration - As a developer I want to ensure alignment on technical design for a new integration with internal VA teams As VANotify I want to enable other teams ... ... As VANotify
We want to ensure we build an API 
So that interested teams can integrate with VANotify to send notifications
 **Outcomes:** - Create diagrams that depict sequence and data flow - Review and discuss with the team - Open API documentation of the new API that utilizes VA identifiers
process
design va profile and mpi integration as a developer i want to ensure alignment on technical design for a new integration with internal va teams as vanotify i want to enable other teams as vanotify
we want to ensure we build an api 
so that interested teams can integrate with vanotify to send notifications
 outcomes create diagrams that depict sequence and data flow review and discuss with the team open api documentation of the new api that utilizes va identifiers
1
729
3,214,305,776
IssuesEvent
2015-10-07 00:40:31
broadinstitute/hellbender-dataflow
https://api.github.com/repos/broadinstitute/hellbender-dataflow
opened
Support interval input in dataflow read pre-processing pipeline
Dataflow DataflowPreprocessingPipeline
_From @droazen on April 27, 2015 19:20_ _Copied from original issue: broadinstitute/hellbender#461_
1.0
Support interval input in dataflow read pre-processing pipeline - _From @droazen on April 27, 2015 19:20_ _Copied from original issue: broadinstitute/hellbender#461_
process
support interval input in dataflow read pre processing pipeline from droazen on april copied from original issue broadinstitute hellbender
1
83,382
7,869,705,387
IssuesEvent
2018-06-24 17:00:51
CodeWarriorsOrganization/CodeWarriorsFinalYearProject
https://api.github.com/repos/CodeWarriorsOrganization/CodeWarriorsFinalYearProject
closed
CW0031/T0031.5 -Create Test Cases (Priority-4)
test cases
### User Story : > AS a player, > I NEED to be able to view the overview of the historical values of all stocks, SO THAT I can get a overall idea about how the prices have fluctuated according to the turn on each company. ### Task : > **Create Test Cases**
1.0
CW0031/T0031.5 -Create Test Cases (Priority-4) - ### User Story : > AS a player, > I NEED to be able to view the overview of the historical values of all stocks, SO THAT I can get a overall idea about how the prices have fluctuated according to the turn on each company. ### Task : > **Create Test Cases**
non_process
create test cases priority user story as a player i need to be able to view the overview of the historical values of all stocks so that i can get a overall idea about how the prices have fluctuated according to the turn on each company task create test cases
0
10,975
13,780,139,792
IssuesEvent
2020-10-08 14:34:51
GoogleCloudPlatform/nodejs-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/nodejs-docs-samples
closed
Decrease Renovate Frequency
samples type: process
Hey repo folks 👋, I'm very thankful we have bots automating some mundane tasks like updating dependency semvers. However, the frequency of these updates is causing micro delays (a minute a week maybe) and not very beneficial for the health of our system. ## Suggestion To increase the usefulness of this bot, and be considerate of our time, I suggest we change the frequency of Renovate ([`renovate.json`](https://github.com/GoogleCloudPlatform/nodejs-docs-samples/blob/master/renovate.json)) to once a quarter: ```json { "schedule": ["every 3 months on the first day of the month"] } ``` https://docs.renovatebot.com/configuration-options/#schedule That way we can ensure we have healthy dependencies and reduce GitHub notifications. At least let's try it for a bit? --- @JustinBeckwith @ace-n Original PR: https://github.com/GoogleCloudPlatform/nodejs-docs-samples/pull/778
1.0
Decrease Renovate Frequency - Hey repo folks 👋, I'm very thankful we have bots automating some mundane tasks like updating dependency semvers. However, the frequency of these updates is causing micro delays (a minute a week maybe) and not very beneficial for the health of our system. ## Suggestion To increase the usefulness of this bot, and be considerate of our time, I suggest we change the frequency of Renovate ([`renovate.json`](https://github.com/GoogleCloudPlatform/nodejs-docs-samples/blob/master/renovate.json)) to once a quarter: ```json { "schedule": ["every 3 months on the first day of the month"] } ``` https://docs.renovatebot.com/configuration-options/#schedule That way we can ensure we have healthy dependencies and reduce GitHub notifications. At least let's try it for a bit? --- @JustinBeckwith @ace-n Original PR: https://github.com/GoogleCloudPlatform/nodejs-docs-samples/pull/778
process
decrease renovate frequency hey repo folks 👋 i m very thankful we have bots automating some mundane tasks like updating dependency semvers however the frequency of these updates is causing micro delays a minute a week maybe and not very beneficial for the health of our system suggestion to increase the usefulness of this bot and be considerate of our time i suggest we change the frequency of renovate to once a quarter json schedule that way we can ensure we have healthy dependencies and reduce github notifications at least let s try it for a bit justinbeckwith ace n original pr
1
232,154
7,655,520,791
IssuesEvent
2018-05-10 13:32:02
BuckleScript/bucklescript
https://api.github.com/repos/BuckleScript/bucklescript
closed
Weak returns the wrong length
PRIORITY:HIGH discussion
Example: `Js.log(Weak.length (Weak.create 0))` prints `-1`. I'm running OCaml 4.02.3 and BuckleScript 2.2.3.
1.0
Weak returns the wrong length - Example: `Js.log(Weak.length (Weak.create 0))` prints `-1`. I'm running OCaml 4.02.3 and BuckleScript 2.2.3.
non_process
weak returns the wrong length example js log weak length weak create prints i m running ocaml and bucklescript
0
11,394
14,230,190,834
IssuesEvent
2020-11-18 07:40:42
EcrituresNumeriques/stylo
https://api.github.com/repos/EcrituresNumeriques/stylo
closed
template latex via process
process :factory:
- [x] forker le nouveau repo de template pour un template-process-sp et modifier le template latex en reprenant l'ancien de @marviro
1.0
template latex via process - - [x] forker le nouveau repo de template pour un template-process-sp et modifier le template latex en reprenant l'ancien de @marviro
process
template latex via process forker le nouveau repo de template pour un template process sp et modifier le template latex en reprenant l ancien de marviro
1
8,274
21,785,659,452
IssuesEvent
2022-05-14 04:47:33
transclude-me/extension
https://api.github.com/repos/transclude-me/extension
closed
User link-summoner library as a backend
enhancement infrastructure architecture
https://github.com/Stvad/link-summoner was forked from this extension to provide previews when embedded into websites. it's architecture affords reuse/reintegration by this extension though
1.0
User link-summoner library as a backend - https://github.com/Stvad/link-summoner was forked from this extension to provide previews when embedded into websites. it's architecture affords reuse/reintegration by this extension though
non_process
user link summoner library as a backend was forked from this extension to provide previews when embedded into websites it s architecture affords reuse reintegration by this extension though
0
13,604
16,191,015,336
IssuesEvent
2021-05-04 08:29:50
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Incorrect cmdlet used in 'Prevent concurrent jobs' example
Pri2 automation/svc cxp doc-enhancement process-automation/subsvc triaged
The snippet in the 'Prevent concurrent jobs' section starts as follows.... ``` # Authenticate to Azure $connection = Get-AutomationConnection -Name AzureRunAsConnection $cnParams = @{ ServicePrincipal = $true Tenant = $connection.TenantId ApplicationId = $connection.ApplicationId CertificateThumbprint = $connection.CertificateThumbprint } Connect-AzAccount @cnParams **$AzureContext = Get-AzSubscription -SubscriptionId $connection.SubscriptionID** ... ``` Instead of using the **Get-AzSubscription** cmdlet in the emboldened line, the example should use **Set-AzContext**. The working snippet would then instead read as follows: ``` # Authenticate to Azure $connection = Get-AutomationConnection -Name AzureRunAsConnection $cnParams = @{ ServicePrincipal = $true Tenant = $connection.TenantId ApplicationId = $connection.ApplicationId CertificateThumbprint = $connection.CertificateThumbprint } Connect-AzAccount @cnParams **$AzureContext = Get-AzContext -SubscriptionId $connection.SubscriptionID** ... ``` --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 42cfca50-2141-d684-cbc1-86fe39f57df6 * Version Independent ID: 38879f45-d225-1fab-86ba-75bed87d7afd * Content: [Manage runbooks in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/manage-runbooks#prevent-concurrent-jobs) * Content Source: [articles/automation/manage-runbooks.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/manage-runbooks.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @MGoedtel * Microsoft Alias: **magoedte**
1.0
Incorrect cmdlet used in 'Prevent concurrent jobs' example - The snippet in the 'Prevent concurrent jobs' section starts as follows.... ``` # Authenticate to Azure $connection = Get-AutomationConnection -Name AzureRunAsConnection $cnParams = @{ ServicePrincipal = $true Tenant = $connection.TenantId ApplicationId = $connection.ApplicationId CertificateThumbprint = $connection.CertificateThumbprint } Connect-AzAccount @cnParams **$AzureContext = Get-AzSubscription -SubscriptionId $connection.SubscriptionID** ... ``` Instead of using the **Get-AzSubscription** cmdlet in the emboldened line, the example should use **Set-AzContext**. The working snippet would then instead read as follows: ``` # Authenticate to Azure $connection = Get-AutomationConnection -Name AzureRunAsConnection $cnParams = @{ ServicePrincipal = $true Tenant = $connection.TenantId ApplicationId = $connection.ApplicationId CertificateThumbprint = $connection.CertificateThumbprint } Connect-AzAccount @cnParams **$AzureContext = Get-AzContext -SubscriptionId $connection.SubscriptionID** ... ``` --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 42cfca50-2141-d684-cbc1-86fe39f57df6 * Version Independent ID: 38879f45-d225-1fab-86ba-75bed87d7afd * Content: [Manage runbooks in Azure Automation](https://docs.microsoft.com/en-us/azure/automation/manage-runbooks#prevent-concurrent-jobs) * Content Source: [articles/automation/manage-runbooks.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/manage-runbooks.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @MGoedtel * Microsoft Alias: **magoedte**
process
incorrect cmdlet used in prevent concurrent jobs example the snippet in the prevent concurrent jobs section starts as follows authenticate to azure connection get automationconnection name azurerunasconnection cnparams serviceprincipal true tenant connection tenantid applicationid connection applicationid certificatethumbprint connection certificatethumbprint connect azaccount cnparams azurecontext get azsubscription subscriptionid connection subscriptionid instead of using the get azsubscription cmdlet in the emboldened line the example should use set azcontext the working snippet would then instead read as follows authenticate to azure connection get automationconnection name azurerunasconnection cnparams serviceprincipal true tenant connection tenantid applicationid connection applicationid certificatethumbprint connection certificatethumbprint connect azaccount cnparams azurecontext get azcontext subscriptionid connection subscriptionid document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login mgoedtel microsoft alias magoedte
1
12,067
14,739,749,266
IssuesEvent
2021-01-07 07:51:11
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
Updating the CDR and flat file process to streamline and improve
anc-process anp-3 ant-enhancement
In GitLab by @kdjstudios on Sep 18, 2018, 09:43 When we upload the usage the system will create the flat file. Invoices then are created from this flat file. Once this is done, this flat file should be imported directly into the CDR(database). This way the CDR and flat file should match The CDR should include all the original fields in the upload and all of the calculated fields. regardless of what is being billed/invoiced.
1.0
Updating the CDR and flat file process to streamline and improve - In GitLab by @kdjstudios on Sep 18, 2018, 09:43 When we upload the usage the system will create the flat file. Invoices then are created from this flat file. Once this is done, this flat file should be imported directly into the CDR(database). This way the CDR and flat file should match The CDR should include all the original fields in the upload and all of the calculated fields. regardless of what is being billed/invoiced.
process
updating the cdr and flat file process to streamline and improve in gitlab by kdjstudios on sep when we upload the usage the system will create the flat file invoices then are created from this flat file once this is done this flat file should be imported directly into the cdr database this way the cdr and flat file should match the cdr should include all the original fields in the upload and all of the calculated fields regardless of what is being billed invoiced
1
17,634
23,451,705,335
IssuesEvent
2022-08-16 03:59:21
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
opened
DISABLED test_exception_raises (__main__.SpawnTest)
module: multiprocessing triaged module: flaky-tests skipped
Platforms: linux This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_exception_raises&suite=SpawnTest&file=test_multiprocessing_spawn.py) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/7848861561). Over the past 3 hours, it has been determined flaky in 1 workflow(s) with 1 failures and 1 successes. **Debugging instructions (after clicking on the recent samples link):** DO NOT BE ALARMED THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs. To find relevant log snippets: 1. Click on the workflow logs linked above 2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work. 3. Grep for `test_exception_raises` 4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs.
1.0
DISABLED test_exception_raises (__main__.SpawnTest) - Platforms: linux This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_exception_raises&suite=SpawnTest&file=test_multiprocessing_spawn.py) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/7848861561). Over the past 3 hours, it has been determined flaky in 1 workflow(s) with 1 failures and 1 successes. **Debugging instructions (after clicking on the recent samples link):** DO NOT BE ALARMED THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs. To find relevant log snippets: 1. Click on the workflow logs linked above 2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work. 3. Grep for `test_exception_raises` 4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs.
process
disabled test exception raises main spawntest platforms linux this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has been determined flaky in workflow s with failures and successes debugging instructions after clicking on the recent samples link do not be alarmed the ci is green we now shield flaky tests from developers so ci will thus be green but it will be harder to parse the logs to find relevant log snippets click on the workflow logs linked above click on the test step of the job so that it is expanded otherwise the grepping will not work grep for test exception raises there should be several instances run as flaky tests are rerun in ci from which you can study the logs
1
12,603
15,007,912,126
IssuesEvent
2021-01-31 07:30:42
PKJrod/CPW213-eCommerceSite
https://api.github.com/repos/PKJrod/CPW213-eCommerceSite
closed
Add Cl Pipeline
developer process
Add Continuous integration that will check to make sure code in a pull request compiles successfully
1.0
Add Cl Pipeline - Add Continuous integration that will check to make sure code in a pull request compiles successfully
process
add cl pipeline add continuous integration that will check to make sure code in a pull request compiles successfully
1
2,538
5,299,988,021
IssuesEvent
2017-02-10 02:26:10
mitchellh/packer
https://api.github.com/repos/mitchellh/packer
closed
Packer push causes inconsitent behavior in atlas builds
docs post-processor/atlas
I started with the [Atlas Packer Vagrant Tutorial](https://github.com/hashicorp/atlas-packer-vagrant-tutorial.git) template and successfully pushed to atlas and completed a build with no changes following the tutorial steps: ``` packer push -name udev/ceapi template.json ``` After the successful build I added a push section to the template to include a directory created in the project root: ``` json "push": { "name": "udev/ceapi", "vcs": true, "include": [ "directory" ] } ``` The build fails with the following log: ``` ---- Started new build at 2015-08-16 16:56:26.660917832 +0000 UTC ---- Packer v0.8.2 vmware-iso output will be in this color. 7 error(s) occurred: * Bad script 'scripts/base.sh': stat scripts/base.sh: no such file or directory * Bad script 'scripts/virtualbox.sh': stat scripts/virtualbox.sh: no such file or directory * Bad script 'scripts/vmware.sh': stat scripts/vmware.sh: no such file or directory * Bad script 'scripts/vagrant.sh': stat scripts/vagrant.sh: no such file or directory * Bad script 'scripts/dep.sh': stat scripts/dep.sh: no such file or directory * Bad script 'scripts/cleanup.sh': stat scripts/cleanup.sh: no such file or directory * Bad script 'scripts/zerodisk.sh': stat scripts/zerodisk.sh: no such file or directory ``` expected: adding a push section with an array of included files/directories does not cause files in vcs to not be found. This issue occurred on 0.8.5 and 0.8.2 (installed via brew)
1.0
Packer push causes inconsitent behavior in atlas builds - I started with the [Atlas Packer Vagrant Tutorial](https://github.com/hashicorp/atlas-packer-vagrant-tutorial.git) template and successfully pushed to atlas and completed a build with no changes following the tutorial steps: ``` packer push -name udev/ceapi template.json ``` After the successful build I added a push section to the template to include a directory created in the project root: ``` json "push": { "name": "udev/ceapi", "vcs": true, "include": [ "directory" ] } ``` The build fails with the following log: ``` ---- Started new build at 2015-08-16 16:56:26.660917832 +0000 UTC ---- Packer v0.8.2 vmware-iso output will be in this color. 7 error(s) occurred: * Bad script 'scripts/base.sh': stat scripts/base.sh: no such file or directory * Bad script 'scripts/virtualbox.sh': stat scripts/virtualbox.sh: no such file or directory * Bad script 'scripts/vmware.sh': stat scripts/vmware.sh: no such file or directory * Bad script 'scripts/vagrant.sh': stat scripts/vagrant.sh: no such file or directory * Bad script 'scripts/dep.sh': stat scripts/dep.sh: no such file or directory * Bad script 'scripts/cleanup.sh': stat scripts/cleanup.sh: no such file or directory * Bad script 'scripts/zerodisk.sh': stat scripts/zerodisk.sh: no such file or directory ``` expected: adding a push section with an array of included files/directories does not cause files in vcs to not be found. This issue occurred on 0.8.5 and 0.8.2 (installed via brew)
process
packer push causes inconsitent behavior in atlas builds i started with the template and successfully pushed to atlas and completed a build with no changes following the tutorial steps packer push name udev ceapi template json after the successful build i added a push section to the template to include a directory created in the project root json push name udev ceapi vcs true include directory the build fails with the following log started new build at utc packer vmware iso output will be in this color error s occurred bad script scripts base sh stat scripts base sh no such file or directory bad script scripts virtualbox sh stat scripts virtualbox sh no such file or directory bad script scripts vmware sh stat scripts vmware sh no such file or directory bad script scripts vagrant sh stat scripts vagrant sh no such file or directory bad script scripts dep sh stat scripts dep sh no such file or directory bad script scripts cleanup sh stat scripts cleanup sh no such file or directory bad script scripts zerodisk sh stat scripts zerodisk sh no such file or directory expected adding a push section with an array of included files directories does not cause files in vcs to not be found this issue occurred on and installed via brew
1
11,323
14,140,143,326
IssuesEvent
2020-11-10 10:47:01
ESMValGroup/ESMValCore
https://api.github.com/repos/ESMValGroup/ESMValCore
closed
Preprocessor function multi_model_statistics depends on the name of the vertical coordinate
bug preprocessor
See the code here: https://github.com/ESMValGroup/ESMValCore/blob/574ad007a03ad640619c653eb4210c849142db6c/esmvalcore/preprocessor/_multimodel.py#L170-L184 This does not work if the coordinate has a different name. The code could e.g. be changed so it uses `cube.coord(axis='Z')`.
1.0
Preprocessor function multi_model_statistics depends on the name of the vertical coordinate - See the code here: https://github.com/ESMValGroup/ESMValCore/blob/574ad007a03ad640619c653eb4210c849142db6c/esmvalcore/preprocessor/_multimodel.py#L170-L184 This does not work if the coordinate has a different name. The code could e.g. be changed so it uses `cube.coord(axis='Z')`.
process
preprocessor function multi model statistics depends on the name of the vertical coordinate see the code here this does not work if the coordinate has a different name the code could e g be changed so it uses cube coord axis z
1
21,265
28,438,616,440
IssuesEvent
2023-04-15 16:21:11
sulton-max/profile.todoapp
https://api.github.com/repos/sulton-max/profile.todoapp
opened
Create Account Processing Service
processing
# The Ask Develop an Account Processing Service # How to Complete this Task Here's some steps to complete this deliverable. - Create a service contract - Implement following features
1.0
Create Account Processing Service - # The Ask Develop an Account Processing Service # How to Complete this Task Here's some steps to complete this deliverable. - Create a service contract - Implement following features
process
create account processing service the ask develop an account processing service how to complete this task here s some steps to complete this deliverable create a service contract implement following features
1
203,472
23,156,026,939
IssuesEvent
2022-07-29 13:05:58
turkdevops/training-kit
https://api.github.com/repos/turkdevops/training-kit
closed
CVE-2022-30123 (High) detected in rack-2.2.3.gem - autoclosed
security vulnerability
## CVE-2022-30123 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>rack-2.2.3.gem</b></p></summary> <p>Rack provides a minimal, modular and adaptable interface for developing web applications in Ruby. By wrapping HTTP requests and responses in the simplest way possible, it unifies and distills the API for web servers, web frameworks, and software in between (the so-called middleware) into a single method call. </p> <p>Library home page: <a href="https://rubygems.org/gems/rack-2.2.3.gem">https://rubygems.org/gems/rack-2.2.3.gem</a></p> <p>Path to dependency file: /Gemfile.lock</p> <p>Path to vulnerable library: /home/wss-scanner/.gem/ruby/2.7.0/cache/rack-2.2.3.gem</p> <p> Dependency Hierarchy: - rack-contrib-2.3.0.gem (Root Library) - :x: **rack-2.2.3.gem** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> There is a possible shell escape sequence injection vulnerability in the Lint and CommonLogger components of Rack before 2.0.9.1,2.1.4.1,2.2.3.1 <p>Publish Date: 2022-05-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-30123>CVE-2022-30123</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-wq4h-7r42-5hrr">https://github.com/advisories/GHSA-wq4h-7r42-5hrr</a></p> <p>Release Date: 2022-05-03</p> <p>Fix Resolution: rack - 2.0.9.1,2.1.4.1,2.2.3.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-30123 (High) detected in rack-2.2.3.gem - autoclosed - ## CVE-2022-30123 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>rack-2.2.3.gem</b></p></summary> <p>Rack provides a minimal, modular and adaptable interface for developing web applications in Ruby. By wrapping HTTP requests and responses in the simplest way possible, it unifies and distills the API for web servers, web frameworks, and software in between (the so-called middleware) into a single method call. </p> <p>Library home page: <a href="https://rubygems.org/gems/rack-2.2.3.gem">https://rubygems.org/gems/rack-2.2.3.gem</a></p> <p>Path to dependency file: /Gemfile.lock</p> <p>Path to vulnerable library: /home/wss-scanner/.gem/ruby/2.7.0/cache/rack-2.2.3.gem</p> <p> Dependency Hierarchy: - rack-contrib-2.3.0.gem (Root Library) - :x: **rack-2.2.3.gem** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> There is a possible shell escape sequence injection vulnerability in the Lint and CommonLogger components of Rack before 2.0.9.1,2.1.4.1,2.2.3.1 <p>Publish Date: 2022-05-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-30123>CVE-2022-30123</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-wq4h-7r42-5hrr">https://github.com/advisories/GHSA-wq4h-7r42-5hrr</a></p> <p>Release Date: 2022-05-03</p> <p>Fix Resolution: rack - 2.0.9.1,2.1.4.1,2.2.3.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in rack gem autoclosed cve high severity vulnerability vulnerable library rack gem rack provides a minimal modular and adaptable interface for developing web applications in ruby by wrapping http requests and responses in the simplest way possible it unifies and distills the api for web servers web frameworks and software in between the so called middleware into a single method call library home page a href path to dependency file gemfile lock path to vulnerable library home wss scanner gem ruby cache rack gem dependency hierarchy rack contrib gem root library x rack gem vulnerable library found in base branch master vulnerability details there is a possible shell escape sequence injection vulnerability in the lint and commonlogger components of rack before publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rack step up your open source security game with mend
0
1,384
2,742,699,936
IssuesEvent
2015-04-21 17:45:52
latex3/svn-mirror
https://api.github.com/repos/latex3/svn-mirror
closed
normalise luatex missing character message
l3build
luatex uses a different format for missing character messages ``` ! Missing character: There is no ^^c5 in font cmr10! ! Missing character: There is no ^^c5 (U+00C5) in font cmr10! ``` This seems to fix it (just do two `^^` version that could also be in etex.) ``` $ svn diff l3build.lua Index: l3build.lua =================================================================== --- l3build.lua (revision 5568) +++ l3build.lua (working copy) @@ -593,6 +593,11 @@ line = string.gsub ( line, "I''m going to assume", "I'm going to assume" ) + -- LuaTeX Missing Character message includes U+ notation. + line = string.gsub ( + line, "Missing character: There is no (%^%^..) %(U%+(....)%) in font", + "Missing character: There is no %1 in font" + ) return line end local newlog = "" ```
1.0
normalise luatex missing character message - luatex uses a different format for missing character messages ``` ! Missing character: There is no ^^c5 in font cmr10! ! Missing character: There is no ^^c5 (U+00C5) in font cmr10! ``` This seems to fix it (just do two `^^` version that could also be in etex.) ``` $ svn diff l3build.lua Index: l3build.lua =================================================================== --- l3build.lua (revision 5568) +++ l3build.lua (working copy) @@ -593,6 +593,11 @@ line = string.gsub ( line, "I''m going to assume", "I'm going to assume" ) + -- LuaTeX Missing Character message includes U+ notation. + line = string.gsub ( + line, "Missing character: There is no (%^%^..) %(U%+(....)%) in font", + "Missing character: There is no %1 in font" + ) return line end local newlog = "" ```
non_process
normalise luatex missing character message luatex uses a different format for missing character messages missing character there is no in font missing character there is no u in font this seems to fix it just do two version that could also be in etex svn diff lua index lua lua revision lua working copy line string gsub line i m going to assume i m going to assume luatex missing character message includes u notation line string gsub line missing character there is no u in font missing character there is no in font return line end local newlog
0
9,960
12,991,959,862
IssuesEvent
2020-07-23 05:30:49
GoogleCloudPlatform/stackdriver-sandbox
https://api.github.com/repos/GoogleCloudPlatform/stackdriver-sandbox
closed
Inject error reporting in demo code
lang: shell priority: p2 type: process
To use the Error Reporting UI requires us to either call a client library to report errors, or format the log so that it can be digested by Cloud Logging. Both need us to explicitly inject it in the demo code. One way to simulate the error generation is to trigger an error in some function of some service with probability. Another way is to create a service specifically causing errors when requests are made.
1.0
Inject error reporting in demo code - To use the Error Reporting UI requires us to either call a client library to report errors, or format the log so that it can be digested by Cloud Logging. Both need us to explicitly inject it in the demo code. One way to simulate the error generation is to trigger an error in some function of some service with probability. Another way is to create a service specifically causing errors when requests are made.
process
inject error reporting in demo code to use the error reporting ui requires us to either call a client library to report errors or format the log so that it can be digested by cloud logging both need us to explicitly inject it in the demo code one way to simulate the error generation is to trigger an error in some function of some service with probability another way is to create a service specifically causing errors when requests are made
1
102,170
12,748,873,843
IssuesEvent
2020-06-26 21:09:33
Hubvet8/hubvet-web
https://api.github.com/repos/Hubvet8/hubvet-web
opened
[Tarefa] Layout para gerenciamento de Workspaces (Visualização e Gerenciamento)
Tarefa UX Design
## Descrição Breve resumo do que é essa tarefa, descreva o trabalho a ser realizado - Precisamos de uma gerenciamento de workspaces e Organizações [Vide: ZenHub] Dados | Tipo ----|---- Nome do Workspace | String Nome da Organização | Boolean True:False ## Ações - Selecionar - O usuário deverá ter a possibilidade de habilitar o acrregamento dos dados da organização dentro do workspace - Editar: - Desvincular organizações do seu workspace e trocar de workspace - Pesquisa: - Filtro de busca para organizações Sugestões: ![image.png](https://images.zenhubusercontent.com/5e9a31126a44049a3a9b9e85/9f62a846-5091-4ede-9349-12cb5aafbaf9) ![image.png](https://images.zenhubusercontent.com/5e9a31126a44049a3a9b9e85/962dfcfe-6714-45b5-a893-0aa59e9311be) ![image.png](https://images.zenhubusercontent.com/5e9a31126a44049a3a9b9e85/2914cfe6-ffd3-4cbe-8d1d-1b98bb85fd8a) ![image.png](https://images.zenhubusercontent.com/5e9a31126a44049a3a9b9e85/fd5140c0-590c-4b09-85e7-436d087188b5) Layout:
1.0
[Tarefa] Layout para gerenciamento de Workspaces (Visualização e Gerenciamento) - ## Descrição Breve resumo do que é essa tarefa, descreva o trabalho a ser realizado - Precisamos de uma gerenciamento de workspaces e Organizações [Vide: ZenHub] Dados | Tipo ----|---- Nome do Workspace | String Nome da Organização | Boolean True:False ## Ações - Selecionar - O usuário deverá ter a possibilidade de habilitar o acrregamento dos dados da organização dentro do workspace - Editar: - Desvincular organizações do seu workspace e trocar de workspace - Pesquisa: - Filtro de busca para organizações Sugestões: ![image.png](https://images.zenhubusercontent.com/5e9a31126a44049a3a9b9e85/9f62a846-5091-4ede-9349-12cb5aafbaf9) ![image.png](https://images.zenhubusercontent.com/5e9a31126a44049a3a9b9e85/962dfcfe-6714-45b5-a893-0aa59e9311be) ![image.png](https://images.zenhubusercontent.com/5e9a31126a44049a3a9b9e85/2914cfe6-ffd3-4cbe-8d1d-1b98bb85fd8a) ![image.png](https://images.zenhubusercontent.com/5e9a31126a44049a3a9b9e85/fd5140c0-590c-4b09-85e7-436d087188b5) Layout:
non_process
layout para gerenciamento de workspaces visualização e gerenciamento descrição breve resumo do que é essa tarefa descreva o trabalho a ser realizado precisamos de uma gerenciamento de workspaces e organizações dados tipo nome do workspace string nome da organização boolean true false ações selecionar o usuário deverá ter a possibilidade de habilitar o acrregamento dos dados da organização dentro do workspace editar desvincular organizações do seu workspace e trocar de workspace pesquisa filtro de busca para organizações sugestões layout
0
12,761
3,090,589,273
IssuesEvent
2015-08-26 07:51:43
excelsior-oss/restler
https://api.github.com/repos/excelsior-oss/restler
opened
Introduce modules system
design enhancement
Looks like Restler may become general purpose REST client generation library so we need to provide ability for us and for customers to easily extend it with new functionality which may be plugged in at client side.
1.0
Introduce modules system - Looks like Restler may become general purpose REST client generation library so we need to provide ability for us and for customers to easily extend it with new functionality which may be plugged in at client side.
non_process
introduce modules system looks like restler may become general purpose rest client generation library so we need to provide ability for us and for customers to easily extend it with new functionality which may be plugged in at client side
0
18,126
24,166,578,798
IssuesEvent
2022-09-22 15:28:47
zammad/zammad
https://api.github.com/repos/zammad/zammad
closed
Unable to add Microsoft 365 account when admin consent for enterprise applications is enabled
enhancement question verified mail processing
<!-- Hi there - thanks for filing an issue. Please ensure the following things before creating an issue - thank you! 🤓 Since november 15th we handle all requests, except real bugs, at our community board. Full explanation: https://community.zammad.org/t/major-change-regarding-github-issues-community-board/21 Please post: - Feature requests - Development questions - Technical questions on the board -> https://community.zammad.org ! If you think you hit a bug, please continue: - Search existing issues and the CHANGELOG.md for your issue - there might be a solution already - Make sure to use the latest version of Zammad if possible - Add the `log/production.log` file from your system. Attention: Make sure no confidential data is in it! - Please write the issue in english - Don't remove the template - otherwise we will close the issue without further comments - Ask questions about Zammad configuration and usage at our mailinglist. See: https://zammad.org/participate Note: We always do our best. Unfortunately, sometimes there are too many requests and we can't handle everything at once. If you want to prioritize/escalate your issue, you can do so by means of a support contract (see https://zammad.com/pricing#selfhosted). * The upper textblock will be removed automatically when you submit your issue * --> ### Infos: * Used Zammad version: 5.0 * Installation method (source, package, ..): deb * Operating system: * Database + version: * Elasticsearch version: * Browser + version: ### Expected behavior: Adds Account to zammad once the application is approved by an admin ### Actual behavior: The "Approval required" prompt keeps appearing even if the application has all permissions approved by an admin ![image](https://user-images.githubusercontent.com/33467318/153863861-8b2019c0-3606-4dcd-a3e3-8374c689a65d.png) ### Steps to reproduce the behavior: - Enable Admin consent for enterprise applications in Azure - create application for zammad and set it all up - then try to add the account in zammad I found out that it is working if you delete the "**_&prompt=consent_**" part of the request URL: https://login.microsoftonline.com/{TENEANTID}/oauth2/v2.0/authorize?access_type=offline&client_id={ID}&prompt=consent&redirect_uri=... Yes I'm sure this is a bug and no feature request or a general question.
1.0
Unable to add Microsoft 365 account when admin consent for enterprise applications is enabled - <!-- Hi there - thanks for filing an issue. Please ensure the following things before creating an issue - thank you! 🤓 Since november 15th we handle all requests, except real bugs, at our community board. Full explanation: https://community.zammad.org/t/major-change-regarding-github-issues-community-board/21 Please post: - Feature requests - Development questions - Technical questions on the board -> https://community.zammad.org ! If you think you hit a bug, please continue: - Search existing issues and the CHANGELOG.md for your issue - there might be a solution already - Make sure to use the latest version of Zammad if possible - Add the `log/production.log` file from your system. Attention: Make sure no confidential data is in it! - Please write the issue in english - Don't remove the template - otherwise we will close the issue without further comments - Ask questions about Zammad configuration and usage at our mailinglist. See: https://zammad.org/participate Note: We always do our best. Unfortunately, sometimes there are too many requests and we can't handle everything at once. If you want to prioritize/escalate your issue, you can do so by means of a support contract (see https://zammad.com/pricing#selfhosted). * The upper textblock will be removed automatically when you submit your issue * --> ### Infos: * Used Zammad version: 5.0 * Installation method (source, package, ..): deb * Operating system: * Database + version: * Elasticsearch version: * Browser + version: ### Expected behavior: Adds Account to zammad once the application is approved by an admin ### Actual behavior: The "Approval required" prompt keeps appearing even if the application has all permissions approved by an admin ![image](https://user-images.githubusercontent.com/33467318/153863861-8b2019c0-3606-4dcd-a3e3-8374c689a65d.png) ### Steps to reproduce the behavior: - Enable Admin consent for enterprise applications in Azure - create application for zammad and set it all up - then try to add the account in zammad I found out that it is working if you delete the "**_&prompt=consent_**" part of the request URL: https://login.microsoftonline.com/{TENEANTID}/oauth2/v2.0/authorize?access_type=offline&client_id={ID}&prompt=consent&redirect_uri=... Yes I'm sure this is a bug and no feature request or a general question.
process
unable to add microsoft account when admin consent for enterprise applications is enabled hi there thanks for filing an issue please ensure the following things before creating an issue thank you 🤓 since november we handle all requests except real bugs at our community board full explanation please post feature requests development questions technical questions on the board if you think you hit a bug please continue search existing issues and the changelog md for your issue there might be a solution already make sure to use the latest version of zammad if possible add the log production log file from your system attention make sure no confidential data is in it please write the issue in english don t remove the template otherwise we will close the issue without further comments ask questions about zammad configuration and usage at our mailinglist see note we always do our best unfortunately sometimes there are too many requests and we can t handle everything at once if you want to prioritize escalate your issue you can do so by means of a support contract see the upper textblock will be removed automatically when you submit your issue infos used zammad version installation method source package deb operating system database version elasticsearch version browser version expected behavior adds account to zammad once the application is approved by an admin actual behavior the approval required prompt keeps appearing even if the application has all permissions approved by an admin steps to reproduce the behavior enable admin consent for enterprise applications in azure create application for zammad and set it all up then try to add the account in zammad i found out that it is working if you delete the prompt consent part of the request url yes i m sure this is a bug and no feature request or a general question
1
9,864
12,877,814,040
IssuesEvent
2020-07-11 13:18:44
elves/elvish
https://api.github.com/repos/elves/elvish
closed
suspending external command with [ctrl-z] does not yield a shell prompt
C:Process Control t:bug
Using elvish built from HEAD run `vim` then press [ctrl-z]. The vim process is apparently suspended but the elvish process doesn't provide a new prompt. Typing is echoed but commands aren't executed. Sending SIGUSR1 results in the following stack dumps. I can reproduce on macOS and Linux. ``` goroutine 5 [running]: github.com/elves/elvish/pkg/sys.DumpStack(0xc0001922d0, 0x15d49c8) /Users/krader/projects/3rd-party/elvish/pkg/sys/dumpstack.go:10 +0x9d github.com/elves/elvish/pkg/program/shell.handleSignal(0x1672de0, 0xc000026660, 0xc0000a2010) /Users/krader/projects/3rd-party/elvish/pkg/program/shell/signal_unix.go:19 +0x6b github.com/elves/elvish/pkg/program/shell.setupShell.func1(0xc0001b6120, 0xc0000a2000, 0xc0000a2008, 0xc0000a2010) /Users/krader/projects/3rd-party/elvish/pkg/program/shell/shell.go:28 +0xb0 created by github.com/elves/elvish/pkg/program/shell.setupShell /Users/krader/projects/3rd-party/elvish/pkg/program/shell/shell.go:25 +0x12d goroutine 1 [semacquire]: sync.runtime_Semacquire(0xc000196c88) /usr/local/Cellar/go/1.14.1/libexec/src/runtime/sema.go:56 +0x42 sync.(*WaitGroup).Wait(0xc000196c80) /usr/local/Cellar/go/1.14.1/libexec/src/sync/waitgroup.go:130 +0x64 github.com/elves/elvish/pkg/eval.(*pipelineOp).invoke(0xc000697220, 0xc000697270, 0xc0007af5e8, 0x100e856) /Users/krader/projects/3rd-party/elvish/pkg/eval/compile_effect.go:173 +0x4c9 github.com/elves/elvish/pkg/eval.effectOp.exec(0x166ba00, 0xc000697220, 0x0, 0x3, 0xc000697270, 0x0, 0xc000200050) /Users/krader/projects/3rd-party/elvish/pkg/eval/op.go:32 +0x69 github.com/elves/elvish/pkg/eval.chunkOp.invoke(0xc000b101a0, 0x1, 0x1, 0xc000697270, 0xc000b102c0, 0xc0007af6b0) /Users/krader/projects/3rd-party/elvish/pkg/eval/compile_effect.go:29 +0x7e github.com/elves/elvish/pkg/eval.effectOp.exec(0x166c9a0, 0xc000b10200, 0x0, 0x3, 0xc000697270, 0xc000b10240, 0xc000b10260) /Users/krader/projects/3rd-party/elvish/pkg/eval/op.go:32 +0x69 github.com/elves/elvish/pkg/eval.(*Evaler).Eval(0xc0001ec000, 0x166c9a0, 0xc000b10200, 0x0, 0x3, 0xc000316060, 0xc000b10220, 0x3, 0x3, 0x15d3748, ...) /Users/krader/projects/3rd-party/elvish/pkg/eval/eval.go:281 +0x1e1 github.com/elves/elvish/pkg/program/shell.evalInTTY(0xc0001ec000, 0x166c9a0, 0xc000b10200, 0x0, 0x3, 0xc000316060, 0xc0000a2000, 0xc0000a2008, 0xc0000a2010, 0x0, ...) /Users/krader/projects/3rd-party/elvish/pkg/program/shell/shell.go:43 +0x1b4 github.com/elves/elvish/pkg/program/shell.Interact(0xc0000a2000, 0xc0000a2008, 0xc0000a2010, 0xc00011bd88) /Users/krader/projects/3rd-party/elvish/pkg/program/shell/interact.go:85 +0x965 github.com/elves/elvish/pkg/program.(*shellProgram).Main(0xc00018c840, 0xc0000a2000, 0xc0000a2008, 0xc0000a2010, 0xc000098190, 0x0, 0x0, 0xc00011bf18) /Users/krader/projects/3rd-party/elvish/pkg/program/shell.go:31 +0x27c github.com/elves/elvish/pkg/program.Main(0xc0000a2000, 0xc0000a2008, 0xc0000a2010, 0xc000098190, 0x1, 0x1, 0x0) /Users/krader/projects/3rd-party/elvish/pkg/program/program.go:103 +0x1c8 main.main() /Users/krader/projects/3rd-party/elvish/main.go:14 +0x69 goroutine 34 [chan receive]: github.com/elves/elvish/pkg/eval.getBlackholeChan.func1(0xc0001b6060) /Users/krader/projects/3rd-party/elvish/pkg/eval/port.go:59 +0x47 created by github.com/elves/elvish/pkg/eval.getBlackholeChan /Users/krader/projects/3rd-party/elvish/pkg/eval/port.go:58 +0x58 goroutine 35 [syscall]: os/signal.signal_recv(0x1672de0) /usr/local/Cellar/go/1.14.1/libexec/src/runtime/sigqueue.go:144 +0x96 os/signal.loop() /usr/local/Cellar/go/1.14.1/libexec/src/os/signal/signal_unix.go:23 +0x22 created by os/signal.Notify.func1 /usr/local/Cellar/go/1.14.1/libexec/src/os/signal/signal.go:127 +0x44 goroutine 6 [chan receive]: github.com/elves/elvish/pkg/cli/prompt.(*Prompt).loop(0xc0002841c0) /Users/krader/projects/3rd-party/elvish/pkg/cli/prompt/prompt.go:77 +0xc8 created by github.com/elves/elvish/pkg/cli/prompt.New /Users/krader/projects/3rd-party/elvish/pkg/cli/prompt/prompt.go:70 +0x15f goroutine 7 [chan receive]: github.com/elves/elvish/pkg/cli/prompt.(*Prompt).loop(0xc0002842a0) /Users/krader/projects/3rd-party/elvish/pkg/cli/prompt/prompt.go:77 +0xc8 created by github.com/elves/elvish/pkg/cli/prompt.New /Users/krader/projects/3rd-party/elvish/pkg/cli/prompt/prompt.go:70 +0x15f goroutine 15127 [chan receive]: github.com/elves/elvish/pkg/eval.relayChanToFile(0xc0000a4e40, 0xc0000a2008, 0xc0005aa4c0, 0xc, 0xc000196bd4) /Users/krader/projects/3rd-party/elvish/pkg/eval/port_helper.go:47 +0x1ae created by github.com/elves/elvish/pkg/eval.portsFromFiles /Users/krader/projects/3rd-party/elvish/pkg/eval/port_helper.go:31 +0xfd goroutine 15129 [select]: github.com/elves/elvish/pkg/eval.ListenInterrupts.func1(0xc00008e0c0, 0xc00008e180, 0xc00008e1e0, 0xc00008e240) /Users/krader/projects/3rd-party/elvish/pkg/eval/interrupts.go:46 +0xad created by github.com/elves/elvish/pkg/eval.ListenInterrupts /Users/krader/projects/3rd-party/elvish/pkg/eval/interrupts.go:42 +0x143 goroutine 13879 [syscall]: syscall.syscall6(0x107dbd0, 0xf99a, 0xc00025fa9c, 0x0, 0xc0008e8090, 0x0, 0x0, 0x0, 0x0, 0x0) /usr/local/Cellar/go/1.14.1/libexec/src/runtime/sys_darwin.go:74 +0x2e syscall.wait4(0xf99a, 0xc00025fa9c, 0x0, 0xc0008e8090, 0x90, 0x1593fa0, 0x1) /usr/local/Cellar/go/1.14.1/libexec/src/syscall/zsyscall_darwin_amd64.go:44 +0x87 syscall.Wait4(0xf99a, 0xc00025faec, 0x0, 0xc0008e8090, 0x37, 0xc0008db560, 0x6) /usr/local/Cellar/go/1.14.1/libexec/src/syscall/syscall_bsd.go:129 +0x51 os.(*Process).wait(0xc00018e330, 0x37, 0xc0008db560, 0x6) /usr/local/Cellar/go/1.14.1/libexec/src/os/exec_unix.go:38 +0x7b os.(*Process).Wait(...) /usr/local/Cellar/go/1.14.1/libexec/src/os/exec.go:125 github.com/elves/elvish/pkg/eval.ExternalCmd.Call(0xc000230b80, 0x37, 0xc0008f4050, 0xc000281d80, 0x5, 0x8, 0xc0003df440, 0xc0002a5800, 0xc0004840a0) /Users/krader/projects/3rd-party/elvish/pkg/eval/external_cmd.go:91 +0x37e github.com/elves/elvish/pkg/eval.(*formOp).invoke(0xc0002fb5f0, 0xc0008f4050, 0x0, 0x0) /Users/krader/projects/3rd-party/elvish/pkg/eval/compile_effect.go:380 +0x8be github.com/elves/elvish/pkg/eval.effectOp.exec(0x166b960, 0xc0002fb5f0, 0xe93, 0x1003, 0xc0008f4050, 0xffffffffffffffff, 0xffffffffffffffff) /Users/krader/projects/3rd-party/elvish/pkg/eval/op.go:32 +0x69 github.com/elves/elvish/pkg/eval.(*pipelineOp).invoke.func1(0x166b960, 0xc0002fb5f0, 0xe93, 0x1003, 0xc0008f4050, 0xc000010de0, 0xc0009df700, 0x0) /Users/krader/projects/3rd-party/elvish/pkg/eval/compile_effect.go:136 +0x60 created by github.com/elves/elvish/pkg/eval.(*pipelineOp).invoke /Users/krader/projects/3rd-party/elvish/pkg/eval/compile_effect.go:135 +0x424 goroutine 15128 [chan receive]: github.com/elves/elvish/pkg/eval.relayChanToFile(0xc000169740, 0xc0000a2010, 0xc0005aa4c0, 0xc, 0xc000196bd4) /Users/krader/projects/3rd-party/elvish/pkg/eval/port_helper.go:47 +0x1ae created by github.com/elves/elvish/pkg/eval.portsFromFiles /Users/krader/projects/3rd-party/elvish/pkg/eval/port_helper.go:32 +0x150 goroutine 15130 [syscall]: syscall.syscall6(0x107dbd0, 0xf99e, 0xc00005fa9c, 0x0, 0xc0008e8000, 0x0, 0x0, 0x0, 0x0, 0x0) /usr/local/Cellar/go/1.14.1/libexec/src/runtime/sys_darwin.go:74 +0x2e syscall.wait4(0xf99e, 0xc00005fa9c, 0x0, 0xc0008e8000, 0x90, 0x1593fa0, 0x1) /usr/local/Cellar/go/1.14.1/libexec/src/syscall/zsyscall_darwin_amd64.go:44 +0x87 syscall.Wait4(0xf99e, 0xc00005faec, 0x0, 0xc0008e8000, 0x12, 0xc0003ce0d0, 0x1) /usr/local/Cellar/go/1.14.1/libexec/src/syscall/syscall_bsd.go:129 +0x51 os.(*Process).wait(0xc0006f6120, 0x12, 0xc0003ce0d0, 0x1) /usr/local/Cellar/go/1.14.1/libexec/src/os/exec_unix.go:38 +0x7b os.(*Process).Wait(...) /usr/local/Cellar/go/1.14.1/libexec/src/os/exec.go:125 github.com/elves/elvish/pkg/eval.ExternalCmd.Call(0xc00060e0d8, 0x3, 0xc0006972c0, 0x0, 0x0, 0x0, 0xc000316180, 0xc00060e0d8, 0x15a6d71) /Users/krader/projects/3rd-party/elvish/pkg/eval/external_cmd.go:91 +0x37e github.com/elves/elvish/pkg/eval.(*formOp).invoke(0xc0007bb040, 0xc0006972c0, 0x0, 0x0) /Users/krader/projects/3rd-party/elvish/pkg/eval/compile_effect.go:380 +0x8be github.com/elves/elvish/pkg/eval.effectOp.exec(0x166b960, 0xc0007bb040, 0x0, 0x3, 0xc0006972c0, 0x14d9040, 0xc0008dd4d8) /Users/krader/projects/3rd-party/elvish/pkg/eval/op.go:32 +0x69 github.com/elves/elvish/pkg/eval.(*pipelineOp).invoke.func1(0x166b960, 0xc0007bb040, 0x0, 0x3, 0xc0006972c0, 0xc0004dc030, 0xc000196c80, 0xc00060e000) /Users/krader/projects/3rd-party/elvish/pkg/eval/compile_effect.go:136 +0x60 created by github.com/elves/elvish/pkg/eval.(*pipelineOp).invoke /Users/krader/projects/3rd-party/elvish/pkg/eval/compile_effect.go:135 +0x424 goroutine 13880 [semacquire]: sync.runtime_Semacquire(0xc0009df708) /usr/local/Cellar/go/1.14.1/libexec/src/runtime/sema.go:56 +0x42 sync.(*WaitGroup).Wait(0xc0009df700) /usr/local/Cellar/go/1.14.1/libexec/src/sync/waitgroup.go:130 +0x64 github.com/elves/elvish/pkg/eval.(*pipelineOp).invoke.func2(0xc0009df700, 0xc0008f4000, 0xc0002df630, 0xc000010de0, 0x1, 0x1) /Users/krader/projects/3rd-party/elvish/pkg/eval/compile_effect.go:156 +0x40 created by github.com/elves/elvish/pkg/eval.(*pipelineOp).invoke /Users/krader/projects/3rd-party/elvish/pkg/eval/compile_effect.go:155 +0x49d ```
1.0
suspending external command with [ctrl-z] does not yield a shell prompt - Using elvish built from HEAD run `vim` then press [ctrl-z]. The vim process is apparently suspended but the elvish process doesn't provide a new prompt. Typing is echoed but commands aren't executed. Sending SIGUSR1 results in the following stack dumps. I can reproduce on macOS and Linux. ``` goroutine 5 [running]: github.com/elves/elvish/pkg/sys.DumpStack(0xc0001922d0, 0x15d49c8) /Users/krader/projects/3rd-party/elvish/pkg/sys/dumpstack.go:10 +0x9d github.com/elves/elvish/pkg/program/shell.handleSignal(0x1672de0, 0xc000026660, 0xc0000a2010) /Users/krader/projects/3rd-party/elvish/pkg/program/shell/signal_unix.go:19 +0x6b github.com/elves/elvish/pkg/program/shell.setupShell.func1(0xc0001b6120, 0xc0000a2000, 0xc0000a2008, 0xc0000a2010) /Users/krader/projects/3rd-party/elvish/pkg/program/shell/shell.go:28 +0xb0 created by github.com/elves/elvish/pkg/program/shell.setupShell /Users/krader/projects/3rd-party/elvish/pkg/program/shell/shell.go:25 +0x12d goroutine 1 [semacquire]: sync.runtime_Semacquire(0xc000196c88) /usr/local/Cellar/go/1.14.1/libexec/src/runtime/sema.go:56 +0x42 sync.(*WaitGroup).Wait(0xc000196c80) /usr/local/Cellar/go/1.14.1/libexec/src/sync/waitgroup.go:130 +0x64 github.com/elves/elvish/pkg/eval.(*pipelineOp).invoke(0xc000697220, 0xc000697270, 0xc0007af5e8, 0x100e856) /Users/krader/projects/3rd-party/elvish/pkg/eval/compile_effect.go:173 +0x4c9 github.com/elves/elvish/pkg/eval.effectOp.exec(0x166ba00, 0xc000697220, 0x0, 0x3, 0xc000697270, 0x0, 0xc000200050) /Users/krader/projects/3rd-party/elvish/pkg/eval/op.go:32 +0x69 github.com/elves/elvish/pkg/eval.chunkOp.invoke(0xc000b101a0, 0x1, 0x1, 0xc000697270, 0xc000b102c0, 0xc0007af6b0) /Users/krader/projects/3rd-party/elvish/pkg/eval/compile_effect.go:29 +0x7e github.com/elves/elvish/pkg/eval.effectOp.exec(0x166c9a0, 0xc000b10200, 0x0, 0x3, 0xc000697270, 0xc000b10240, 0xc000b10260) /Users/krader/projects/3rd-party/elvish/pkg/eval/op.go:32 +0x69 github.com/elves/elvish/pkg/eval.(*Evaler).Eval(0xc0001ec000, 0x166c9a0, 0xc000b10200, 0x0, 0x3, 0xc000316060, 0xc000b10220, 0x3, 0x3, 0x15d3748, ...) /Users/krader/projects/3rd-party/elvish/pkg/eval/eval.go:281 +0x1e1 github.com/elves/elvish/pkg/program/shell.evalInTTY(0xc0001ec000, 0x166c9a0, 0xc000b10200, 0x0, 0x3, 0xc000316060, 0xc0000a2000, 0xc0000a2008, 0xc0000a2010, 0x0, ...) /Users/krader/projects/3rd-party/elvish/pkg/program/shell/shell.go:43 +0x1b4 github.com/elves/elvish/pkg/program/shell.Interact(0xc0000a2000, 0xc0000a2008, 0xc0000a2010, 0xc00011bd88) /Users/krader/projects/3rd-party/elvish/pkg/program/shell/interact.go:85 +0x965 github.com/elves/elvish/pkg/program.(*shellProgram).Main(0xc00018c840, 0xc0000a2000, 0xc0000a2008, 0xc0000a2010, 0xc000098190, 0x0, 0x0, 0xc00011bf18) /Users/krader/projects/3rd-party/elvish/pkg/program/shell.go:31 +0x27c github.com/elves/elvish/pkg/program.Main(0xc0000a2000, 0xc0000a2008, 0xc0000a2010, 0xc000098190, 0x1, 0x1, 0x0) /Users/krader/projects/3rd-party/elvish/pkg/program/program.go:103 +0x1c8 main.main() /Users/krader/projects/3rd-party/elvish/main.go:14 +0x69 goroutine 34 [chan receive]: github.com/elves/elvish/pkg/eval.getBlackholeChan.func1(0xc0001b6060) /Users/krader/projects/3rd-party/elvish/pkg/eval/port.go:59 +0x47 created by github.com/elves/elvish/pkg/eval.getBlackholeChan /Users/krader/projects/3rd-party/elvish/pkg/eval/port.go:58 +0x58 goroutine 35 [syscall]: os/signal.signal_recv(0x1672de0) /usr/local/Cellar/go/1.14.1/libexec/src/runtime/sigqueue.go:144 +0x96 os/signal.loop() /usr/local/Cellar/go/1.14.1/libexec/src/os/signal/signal_unix.go:23 +0x22 created by os/signal.Notify.func1 /usr/local/Cellar/go/1.14.1/libexec/src/os/signal/signal.go:127 +0x44 goroutine 6 [chan receive]: github.com/elves/elvish/pkg/cli/prompt.(*Prompt).loop(0xc0002841c0) /Users/krader/projects/3rd-party/elvish/pkg/cli/prompt/prompt.go:77 +0xc8 created by github.com/elves/elvish/pkg/cli/prompt.New /Users/krader/projects/3rd-party/elvish/pkg/cli/prompt/prompt.go:70 +0x15f goroutine 7 [chan receive]: github.com/elves/elvish/pkg/cli/prompt.(*Prompt).loop(0xc0002842a0) /Users/krader/projects/3rd-party/elvish/pkg/cli/prompt/prompt.go:77 +0xc8 created by github.com/elves/elvish/pkg/cli/prompt.New /Users/krader/projects/3rd-party/elvish/pkg/cli/prompt/prompt.go:70 +0x15f goroutine 15127 [chan receive]: github.com/elves/elvish/pkg/eval.relayChanToFile(0xc0000a4e40, 0xc0000a2008, 0xc0005aa4c0, 0xc, 0xc000196bd4) /Users/krader/projects/3rd-party/elvish/pkg/eval/port_helper.go:47 +0x1ae created by github.com/elves/elvish/pkg/eval.portsFromFiles /Users/krader/projects/3rd-party/elvish/pkg/eval/port_helper.go:31 +0xfd goroutine 15129 [select]: github.com/elves/elvish/pkg/eval.ListenInterrupts.func1(0xc00008e0c0, 0xc00008e180, 0xc00008e1e0, 0xc00008e240) /Users/krader/projects/3rd-party/elvish/pkg/eval/interrupts.go:46 +0xad created by github.com/elves/elvish/pkg/eval.ListenInterrupts /Users/krader/projects/3rd-party/elvish/pkg/eval/interrupts.go:42 +0x143 goroutine 13879 [syscall]: syscall.syscall6(0x107dbd0, 0xf99a, 0xc00025fa9c, 0x0, 0xc0008e8090, 0x0, 0x0, 0x0, 0x0, 0x0) /usr/local/Cellar/go/1.14.1/libexec/src/runtime/sys_darwin.go:74 +0x2e syscall.wait4(0xf99a, 0xc00025fa9c, 0x0, 0xc0008e8090, 0x90, 0x1593fa0, 0x1) /usr/local/Cellar/go/1.14.1/libexec/src/syscall/zsyscall_darwin_amd64.go:44 +0x87 syscall.Wait4(0xf99a, 0xc00025faec, 0x0, 0xc0008e8090, 0x37, 0xc0008db560, 0x6) /usr/local/Cellar/go/1.14.1/libexec/src/syscall/syscall_bsd.go:129 +0x51 os.(*Process).wait(0xc00018e330, 0x37, 0xc0008db560, 0x6) /usr/local/Cellar/go/1.14.1/libexec/src/os/exec_unix.go:38 +0x7b os.(*Process).Wait(...) /usr/local/Cellar/go/1.14.1/libexec/src/os/exec.go:125 github.com/elves/elvish/pkg/eval.ExternalCmd.Call(0xc000230b80, 0x37, 0xc0008f4050, 0xc000281d80, 0x5, 0x8, 0xc0003df440, 0xc0002a5800, 0xc0004840a0) /Users/krader/projects/3rd-party/elvish/pkg/eval/external_cmd.go:91 +0x37e github.com/elves/elvish/pkg/eval.(*formOp).invoke(0xc0002fb5f0, 0xc0008f4050, 0x0, 0x0) /Users/krader/projects/3rd-party/elvish/pkg/eval/compile_effect.go:380 +0x8be github.com/elves/elvish/pkg/eval.effectOp.exec(0x166b960, 0xc0002fb5f0, 0xe93, 0x1003, 0xc0008f4050, 0xffffffffffffffff, 0xffffffffffffffff) /Users/krader/projects/3rd-party/elvish/pkg/eval/op.go:32 +0x69 github.com/elves/elvish/pkg/eval.(*pipelineOp).invoke.func1(0x166b960, 0xc0002fb5f0, 0xe93, 0x1003, 0xc0008f4050, 0xc000010de0, 0xc0009df700, 0x0) /Users/krader/projects/3rd-party/elvish/pkg/eval/compile_effect.go:136 +0x60 created by github.com/elves/elvish/pkg/eval.(*pipelineOp).invoke /Users/krader/projects/3rd-party/elvish/pkg/eval/compile_effect.go:135 +0x424 goroutine 15128 [chan receive]: github.com/elves/elvish/pkg/eval.relayChanToFile(0xc000169740, 0xc0000a2010, 0xc0005aa4c0, 0xc, 0xc000196bd4) /Users/krader/projects/3rd-party/elvish/pkg/eval/port_helper.go:47 +0x1ae created by github.com/elves/elvish/pkg/eval.portsFromFiles /Users/krader/projects/3rd-party/elvish/pkg/eval/port_helper.go:32 +0x150 goroutine 15130 [syscall]: syscall.syscall6(0x107dbd0, 0xf99e, 0xc00005fa9c, 0x0, 0xc0008e8000, 0x0, 0x0, 0x0, 0x0, 0x0) /usr/local/Cellar/go/1.14.1/libexec/src/runtime/sys_darwin.go:74 +0x2e syscall.wait4(0xf99e, 0xc00005fa9c, 0x0, 0xc0008e8000, 0x90, 0x1593fa0, 0x1) /usr/local/Cellar/go/1.14.1/libexec/src/syscall/zsyscall_darwin_amd64.go:44 +0x87 syscall.Wait4(0xf99e, 0xc00005faec, 0x0, 0xc0008e8000, 0x12, 0xc0003ce0d0, 0x1) /usr/local/Cellar/go/1.14.1/libexec/src/syscall/syscall_bsd.go:129 +0x51 os.(*Process).wait(0xc0006f6120, 0x12, 0xc0003ce0d0, 0x1) /usr/local/Cellar/go/1.14.1/libexec/src/os/exec_unix.go:38 +0x7b os.(*Process).Wait(...) /usr/local/Cellar/go/1.14.1/libexec/src/os/exec.go:125 github.com/elves/elvish/pkg/eval.ExternalCmd.Call(0xc00060e0d8, 0x3, 0xc0006972c0, 0x0, 0x0, 0x0, 0xc000316180, 0xc00060e0d8, 0x15a6d71) /Users/krader/projects/3rd-party/elvish/pkg/eval/external_cmd.go:91 +0x37e github.com/elves/elvish/pkg/eval.(*formOp).invoke(0xc0007bb040, 0xc0006972c0, 0x0, 0x0) /Users/krader/projects/3rd-party/elvish/pkg/eval/compile_effect.go:380 +0x8be github.com/elves/elvish/pkg/eval.effectOp.exec(0x166b960, 0xc0007bb040, 0x0, 0x3, 0xc0006972c0, 0x14d9040, 0xc0008dd4d8) /Users/krader/projects/3rd-party/elvish/pkg/eval/op.go:32 +0x69 github.com/elves/elvish/pkg/eval.(*pipelineOp).invoke.func1(0x166b960, 0xc0007bb040, 0x0, 0x3, 0xc0006972c0, 0xc0004dc030, 0xc000196c80, 0xc00060e000) /Users/krader/projects/3rd-party/elvish/pkg/eval/compile_effect.go:136 +0x60 created by github.com/elves/elvish/pkg/eval.(*pipelineOp).invoke /Users/krader/projects/3rd-party/elvish/pkg/eval/compile_effect.go:135 +0x424 goroutine 13880 [semacquire]: sync.runtime_Semacquire(0xc0009df708) /usr/local/Cellar/go/1.14.1/libexec/src/runtime/sema.go:56 +0x42 sync.(*WaitGroup).Wait(0xc0009df700) /usr/local/Cellar/go/1.14.1/libexec/src/sync/waitgroup.go:130 +0x64 github.com/elves/elvish/pkg/eval.(*pipelineOp).invoke.func2(0xc0009df700, 0xc0008f4000, 0xc0002df630, 0xc000010de0, 0x1, 0x1) /Users/krader/projects/3rd-party/elvish/pkg/eval/compile_effect.go:156 +0x40 created by github.com/elves/elvish/pkg/eval.(*pipelineOp).invoke /Users/krader/projects/3rd-party/elvish/pkg/eval/compile_effect.go:155 +0x49d ```
process
suspending external command with does not yield a shell prompt using elvish built from head run vim then press the vim process is apparently suspended but the elvish process doesn t provide a new prompt typing is echoed but commands aren t executed sending results in the following stack dumps i can reproduce on macos and linux goroutine github com elves elvish pkg sys dumpstack users krader projects party elvish pkg sys dumpstack go github com elves elvish pkg program shell handlesignal users krader projects party elvish pkg program shell signal unix go github com elves elvish pkg program shell setupshell users krader projects party elvish pkg program shell shell go created by github com elves elvish pkg program shell setupshell users krader projects party elvish pkg program shell shell go goroutine sync runtime semacquire usr local cellar go libexec src runtime sema go sync waitgroup wait usr local cellar go libexec src sync waitgroup go github com elves elvish pkg eval pipelineop invoke users krader projects party elvish pkg eval compile effect go github com elves elvish pkg eval effectop exec users krader projects party elvish pkg eval op go github com elves elvish pkg eval chunkop invoke users krader projects party elvish pkg eval compile effect go github com elves elvish pkg eval effectop exec users krader projects party elvish pkg eval op go github com elves elvish pkg eval evaler eval users krader projects party elvish pkg eval eval go github com elves elvish pkg program shell evalintty users krader projects party elvish pkg program shell shell go github com elves elvish pkg program shell interact users krader projects party elvish pkg program shell interact go github com elves elvish pkg program shellprogram main users krader projects party elvish pkg program shell go github com elves elvish pkg program main users krader projects party elvish pkg program program go main main users krader projects party elvish main go goroutine github com elves elvish pkg eval getblackholechan users krader projects party elvish pkg eval port go created by github com elves elvish pkg eval getblackholechan users krader projects party elvish pkg eval port go goroutine os signal signal recv usr local cellar go libexec src runtime sigqueue go os signal loop usr local cellar go libexec src os signal signal unix go created by os signal notify usr local cellar go libexec src os signal signal go goroutine github com elves elvish pkg cli prompt prompt loop users krader projects party elvish pkg cli prompt prompt go created by github com elves elvish pkg cli prompt new users krader projects party elvish pkg cli prompt prompt go goroutine github com elves elvish pkg cli prompt prompt loop users krader projects party elvish pkg cli prompt prompt go created by github com elves elvish pkg cli prompt new users krader projects party elvish pkg cli prompt prompt go goroutine github com elves elvish pkg eval relaychantofile users krader projects party elvish pkg eval port helper go created by github com elves elvish pkg eval portsfromfiles users krader projects party elvish pkg eval port helper go goroutine github com elves elvish pkg eval listeninterrupts users krader projects party elvish pkg eval interrupts go created by github com elves elvish pkg eval listeninterrupts users krader projects party elvish pkg eval interrupts go goroutine syscall usr local cellar go libexec src runtime sys darwin go syscall usr local cellar go libexec src syscall zsyscall darwin go syscall usr local cellar go libexec src syscall syscall bsd go os process wait usr local cellar go libexec src os exec unix go os process wait usr local cellar go libexec src os exec go github com elves elvish pkg eval externalcmd call users krader projects party elvish pkg eval external cmd go github com elves elvish pkg eval formop invoke users krader projects party elvish pkg eval compile effect go github com elves elvish pkg eval effectop exec users krader projects party elvish pkg eval op go github com elves elvish pkg eval pipelineop invoke users krader projects party elvish pkg eval compile effect go created by github com elves elvish pkg eval pipelineop invoke users krader projects party elvish pkg eval compile effect go goroutine github com elves elvish pkg eval relaychantofile users krader projects party elvish pkg eval port helper go created by github com elves elvish pkg eval portsfromfiles users krader projects party elvish pkg eval port helper go goroutine syscall usr local cellar go libexec src runtime sys darwin go syscall usr local cellar go libexec src syscall zsyscall darwin go syscall usr local cellar go libexec src syscall syscall bsd go os process wait usr local cellar go libexec src os exec unix go os process wait usr local cellar go libexec src os exec go github com elves elvish pkg eval externalcmd call users krader projects party elvish pkg eval external cmd go github com elves elvish pkg eval formop invoke users krader projects party elvish pkg eval compile effect go github com elves elvish pkg eval effectop exec users krader projects party elvish pkg eval op go github com elves elvish pkg eval pipelineop invoke users krader projects party elvish pkg eval compile effect go created by github com elves elvish pkg eval pipelineop invoke users krader projects party elvish pkg eval compile effect go goroutine sync runtime semacquire usr local cellar go libexec src runtime sema go sync waitgroup wait usr local cellar go libexec src sync waitgroup go github com elves elvish pkg eval pipelineop invoke users krader projects party elvish pkg eval compile effect go created by github com elves elvish pkg eval pipelineop invoke users krader projects party elvish pkg eval compile effect go
1
3,266
6,343,454,127
IssuesEvent
2017-07-27 17:40:26
WSDOT-GIS/wsdot-traffic-gp
https://api.github.com/repos/WSDOT-GIS/wsdot-traffic-gp
opened
File size keeps increasing after each truncate+update
bug geoprocessing
...the file size eventually gets so large that uploading via [overwrite-hosted-features](https://github.com/WSDOT-GIS/overwrite-hosted-features/tree/add-setup) fails.
1.0
File size keeps increasing after each truncate+update - ...the file size eventually gets so large that uploading via [overwrite-hosted-features](https://github.com/WSDOT-GIS/overwrite-hosted-features/tree/add-setup) fails.
process
file size keeps increasing after each truncate update the file size eventually gets so large that uploading via fails
1
8,595
11,758,883,803
IssuesEvent
2020-03-13 16:12:38
nltk/nltk
https://api.github.com/repos/nltk/nltk
closed
Use multiprocessing package for ParallelProverBuilder and ParallelProverBuilderCommand
inactive multithread / multiprocessing
ParallelProverBuilder and ParallelProverBuilderCommand currently use the threading package to spawn two threads, one with a prover, and one with a model builder. Threads are not stoppable, so when one thread returns first, the other is kept alive until it finishes; it might never finish, as Mace4 can be set to have no maximum number of models to try. By using the multiprocessing package, we can replace the two threads with two processes instead, and once one process finishes, the other process can be terminated. Does this seem like a good idea? I would be happy to try making this change and opening up a PR.
1.0
Use multiprocessing package for ParallelProverBuilder and ParallelProverBuilderCommand - ParallelProverBuilder and ParallelProverBuilderCommand currently use the threading package to spawn two threads, one with a prover, and one with a model builder. Threads are not stoppable, so when one thread returns first, the other is kept alive until it finishes; it might never finish, as Mace4 can be set to have no maximum number of models to try. By using the multiprocessing package, we can replace the two threads with two processes instead, and once one process finishes, the other process can be terminated. Does this seem like a good idea? I would be happy to try making this change and opening up a PR.
process
use multiprocessing package for parallelproverbuilder and parallelproverbuildercommand parallelproverbuilder and parallelproverbuildercommand currently use the threading package to spawn two threads one with a prover and one with a model builder threads are not stoppable so when one thread returns first the other is kept alive until it finishes it might never finish as can be set to have no maximum number of models to try by using the multiprocessing package we can replace the two threads with two processes instead and once one process finishes the other process can be terminated does this seem like a good idea i would be happy to try making this change and opening up a pr
1
7,208
10,343,294,765
IssuesEvent
2019-09-04 08:39:03
heim-rs/heim
https://api.github.com/repos/heim-rs/heim
closed
Implement std::cmp::Eq for process::Process
A-process C-enhancement P-low
It would be a nice addition to be able comparing processes with the usual `==` and `!=`.
1.0
Implement std::cmp::Eq for process::Process - It would be a nice addition to be able comparing processes with the usual `==` and `!=`.
process
implement std cmp eq for process process it would be a nice addition to be able comparing processes with the usual and
1
9,506
12,494,094,267
IssuesEvent
2020-06-01 10:30:26
sysown/proxysql
https://api.github.com/repos/sysown/proxysql
closed
Transaction savepoints do not work
CONNECTION POOL PROTOCOL QUERY PROCESSOR ROUTING bug
After we started using transaction savepoints in business logic, we noticed that RELEASE SAVEPOINT statement began to fail when proxysql is used with error "no savepoint exists". On the other hand, when connecting directly to MySQL (Percona server in our case) all works fine. Some examination of proxysql' repository gave me a hint that the problem could be possibly workarounded by disabling connection multiplexing, and this actually worked. But obviously there should be a better solution for this problem. FYI: all statements are issued from client, no stored procedures are used at all.
1.0
Transaction savepoints do not work - After we started using transaction savepoints in business logic, we noticed that RELEASE SAVEPOINT statement began to fail when proxysql is used with error "no savepoint exists". On the other hand, when connecting directly to MySQL (Percona server in our case) all works fine. Some examination of proxysql' repository gave me a hint that the problem could be possibly workarounded by disabling connection multiplexing, and this actually worked. But obviously there should be a better solution for this problem. FYI: all statements are issued from client, no stored procedures are used at all.
process
transaction savepoints do not work after we started using transaction savepoints in business logic we noticed that release savepoint statement began to fail when proxysql is used with error no savepoint exists on the other hand when connecting directly to mysql percona server in our case all works fine some examination of proxysql repository gave me a hint that the problem could be possibly workarounded by disabling connection multiplexing and this actually worked but obviously there should be a better solution for this problem fyi all statements are issued from client no stored procedures are used at all
1
9,903
12,907,819,910
IssuesEvent
2020-07-15 06:04:36
cypress-io/cypress
https://api.github.com/repos/cypress-io/cypress
closed
Flaky internal test: offsetTop error
browser: chrome browser: electron process: flaky test stage: needs investigating
### Current behavior: `Cannot read property 'offsetTop' of null` error happens on Chrome and Electron. [Link 1](https://app.circleci.com/pipelines/github/cypress-io/cypress/12329/workflows/41ae1df2-54c9-466d-a6aa-6fed94430e26/jobs/394189/steps) [Link 2](https://app.circleci.com/pipelines/github/cypress-io/cypress/12326/workflows/f7762876-613b-4dbb-badd-5b6803126a42/jobs/394009/steps) [Link 3](https://app.circleci.com/pipelines/github/cypress-io/cypress/12314/workflows/5dd5c1d6-a1b4-4d53-b7b2-6c897baa808f/jobs/393439/steps) **Failed Test File:** https://github.com/cypress-io/cypress/blob/develop/packages/server/test/e2e/3_runnable_execution_spec.ts ### Desired behavior: Test should not be flaky. ### Test code to reproduce N/A ### Versions Cypress 4.10.0+
1.0
Flaky internal test: offsetTop error - ### Current behavior: `Cannot read property 'offsetTop' of null` error happens on Chrome and Electron. [Link 1](https://app.circleci.com/pipelines/github/cypress-io/cypress/12329/workflows/41ae1df2-54c9-466d-a6aa-6fed94430e26/jobs/394189/steps) [Link 2](https://app.circleci.com/pipelines/github/cypress-io/cypress/12326/workflows/f7762876-613b-4dbb-badd-5b6803126a42/jobs/394009/steps) [Link 3](https://app.circleci.com/pipelines/github/cypress-io/cypress/12314/workflows/5dd5c1d6-a1b4-4d53-b7b2-6c897baa808f/jobs/393439/steps) **Failed Test File:** https://github.com/cypress-io/cypress/blob/develop/packages/server/test/e2e/3_runnable_execution_spec.ts ### Desired behavior: Test should not be flaky. ### Test code to reproduce N/A ### Versions Cypress 4.10.0+
process
flaky internal test offsettop error current behavior cannot read property offsettop of null error happens on chrome and electron failed test file desired behavior test should not be flaky test code to reproduce n a versions cypress
1
264,353
8,308,791,986
IssuesEvent
2018-09-24 00:33:08
RITlug/tasks
https://api.github.com/repos/RITlug/tasks
closed
Write a blog post on TigerOS recap / overview of Flock 2018
priority:high projects
# Summary Writing about TigerOS in the public is important to build awareness. We should establish a precedent for regular blog posts / updates about TigerOS. # Description This could be done in a few different ways. ### Our own blog We could host our own blog, like `blog.ritlug.com`, where we could write content about the club to share with the world. **Pros**: * Clearly associated with RITlug * Flexible to publish as often and as little or as much as we want **Cons**: * Harder to promote (starting from ground zero) * Requires us to host more infrastructure ### Personal blogs Team members can use their own personal blogs for updates about TigerOS. **Pros**: * RITlug doesn't host infrastructure * People can write about what they're most interested in (more personal focus than team focus) * Ideally, more content this way! **Cons**: * Harder to aggregate / collect (i.e. "_Where can I keep up with TigerOS news?_") * Possibility that someone might decide to delete their blog later / risk of lost content ### Publishing elsewhere We could use sites like Fedora Magazine or Opensource.com to talk about the work we're doing with TigerOS. **Pros**: * _Much_ larger audience, better chances of being noticed * Professional team of editors to help polish content and also promote it **Cons**: * Takes more time to publish content * Writing needs to be more complete than short, tidbit updates. # Assignees I'm assigning this to @axk4545 primarily to take into consideration as Project Coordinator. This is important for TigerOS planning and he should discuss with the team to figure out frequency and what works best for the team. I'm also assigned to this issue to help support and coordinate the promotion of written material on TigerOS.
1.0
Write a blog post on TigerOS recap / overview of Flock 2018 - # Summary Writing about TigerOS in the public is important to build awareness. We should establish a precedent for regular blog posts / updates about TigerOS. # Description This could be done in a few different ways. ### Our own blog We could host our own blog, like `blog.ritlug.com`, where we could write content about the club to share with the world. **Pros**: * Clearly associated with RITlug * Flexible to publish as often and as little or as much as we want **Cons**: * Harder to promote (starting from ground zero) * Requires us to host more infrastructure ### Personal blogs Team members can use their own personal blogs for updates about TigerOS. **Pros**: * RITlug doesn't host infrastructure * People can write about what they're most interested in (more personal focus than team focus) * Ideally, more content this way! **Cons**: * Harder to aggregate / collect (i.e. "_Where can I keep up with TigerOS news?_") * Possibility that someone might decide to delete their blog later / risk of lost content ### Publishing elsewhere We could use sites like Fedora Magazine or Opensource.com to talk about the work we're doing with TigerOS. **Pros**: * _Much_ larger audience, better chances of being noticed * Professional team of editors to help polish content and also promote it **Cons**: * Takes more time to publish content * Writing needs to be more complete than short, tidbit updates. # Assignees I'm assigning this to @axk4545 primarily to take into consideration as Project Coordinator. This is important for TigerOS planning and he should discuss with the team to figure out frequency and what works best for the team. I'm also assigned to this issue to help support and coordinate the promotion of written material on TigerOS.
non_process
write a blog post on tigeros recap overview of flock summary writing about tigeros in the public is important to build awareness we should establish a precedent for regular blog posts updates about tigeros description this could be done in a few different ways our own blog we could host our own blog like blog ritlug com where we could write content about the club to share with the world pros clearly associated with ritlug flexible to publish as often and as little or as much as we want cons harder to promote starting from ground zero requires us to host more infrastructure personal blogs team members can use their own personal blogs for updates about tigeros pros ritlug doesn t host infrastructure people can write about what they re most interested in more personal focus than team focus ideally more content this way cons harder to aggregate collect i e where can i keep up with tigeros news possibility that someone might decide to delete their blog later risk of lost content publishing elsewhere we could use sites like fedora magazine or opensource com to talk about the work we re doing with tigeros pros much larger audience better chances of being noticed professional team of editors to help polish content and also promote it cons takes more time to publish content writing needs to be more complete than short tidbit updates assignees i m assigning this to primarily to take into consideration as project coordinator this is important for tigeros planning and he should discuss with the team to figure out frequency and what works best for the team i m also assigned to this issue to help support and coordinate the promotion of written material on tigeros
0
13,681
16,439,880,068
IssuesEvent
2021-05-20 13:19:24
Open-EO/openeo-processes
https://api.github.com/repos/Open-EO/openeo-processes
closed
Date Math processes
new process
I just got this feature request: > Hi, I have a question on the usage of parameters in a UDP. Is it possible to do some date calculations on a parameter before using it later in the UDP? For example, this is a Python snippet that shows what I want to achieve: > > def pixel_inputs(vito_connection, bbox, start, end) -> DataCube: > start_dt = pandas.to_datetime(start) - pandas.to_timedelta("90D") > end_dt = pandas.to_datetime(end) + pandas.to_timedelta("90D") > S2bands = vito_connection.load_collection( > 'TERRASCOPE_S2_TOC_V2', > temporal_extent=[start_dt, end_dt] > .... > > This works fine when I run this as a normal Python script. But when I try to include it in a UDP, the start and end are UDP parameters and the pandas.to_datetime() method doesn't know what to do with a Parameter object. Is there another way to achieve this? I think this is an interesting use case to start introducing "date math" processes, so that they can be used inside the UDP of this use case. A good one to start seems to be something like `date_shift` (or `date_delta`, `date_offset`), taking arguments: - `date` (string): reference date to work from - `delta` (string): a string expressing a time shift, like "90D" (90 days) in the example (also see https://pandas.pydata.org/pandas-docs/stable/reference/api/pandas.to_timedelta.html). The delta probably also have to include a sign to indicated if the delta has to be added or subtracted. and returning the shifted date (as string)
1.0
Date Math processes - I just got this feature request: > Hi, I have a question on the usage of parameters in a UDP. Is it possible to do some date calculations on a parameter before using it later in the UDP? For example, this is a Python snippet that shows what I want to achieve: > > def pixel_inputs(vito_connection, bbox, start, end) -> DataCube: > start_dt = pandas.to_datetime(start) - pandas.to_timedelta("90D") > end_dt = pandas.to_datetime(end) + pandas.to_timedelta("90D") > S2bands = vito_connection.load_collection( > 'TERRASCOPE_S2_TOC_V2', > temporal_extent=[start_dt, end_dt] > .... > > This works fine when I run this as a normal Python script. But when I try to include it in a UDP, the start and end are UDP parameters and the pandas.to_datetime() method doesn't know what to do with a Parameter object. Is there another way to achieve this? I think this is an interesting use case to start introducing "date math" processes, so that they can be used inside the UDP of this use case. A good one to start seems to be something like `date_shift` (or `date_delta`, `date_offset`), taking arguments: - `date` (string): reference date to work from - `delta` (string): a string expressing a time shift, like "90D" (90 days) in the example (also see https://pandas.pydata.org/pandas-docs/stable/reference/api/pandas.to_timedelta.html). The delta probably also have to include a sign to indicated if the delta has to be added or subtracted. and returning the shifted date (as string)
process
date math processes i just got this feature request hi i have a question on the usage of parameters in a udp is it possible to do some date calculations on a parameter before using it later in the udp for example this is a python snippet that shows what i want to achieve def pixel inputs vito connection bbox start end datacube start dt pandas to datetime start pandas to timedelta end dt pandas to datetime end pandas to timedelta vito connection load collection terrascope toc temporal extent this works fine when i run this as a normal python script but when i try to include it in a udp the start and end are udp parameters and the pandas to datetime method doesn t know what to do with a parameter object is there another way to achieve this i think this is an interesting use case to start introducing date math processes so that they can be used inside the udp of this use case a good one to start seems to be something like date shift or date delta date offset taking arguments date string reference date to work from delta string a string expressing a time shift like days in the example also see the delta probably also have to include a sign to indicated if the delta has to be added or subtracted and returning the shifted date as string
1
6,132
8,997,730,196
IssuesEvent
2019-02-02 15:18:46
Team2470/FRC-2019
https://api.github.com/repos/Team2470/FRC-2019
reopened
Alignment Function / Vision Assisted Driving
Feature Vision Processing help wanted major
**Is your feature request related to a problem? Please describe.** Drivers might have a hard time lining up the robot so that we can place hatches. **Describe the solution you'd like** A function that would use vision processing to align with the reflective tape above the target. **Describe alternatives you've considered** We could just use a camera to let the drivers see, but that still relies on them being good at driving and the camera functioning well. **Additional context** Example of how we could use the limelight: ![lime](https://user-images.githubusercontent.com/16826817/52147270-c873eb00-262b-11e9-9ea8-14442f957a85.gif) ([http://docs.limelightvision.io/en/latest/software_change_log.html](http://docs.limelightvision.io/en/latest/software_change_log.html))
1.0
Alignment Function / Vision Assisted Driving - **Is your feature request related to a problem? Please describe.** Drivers might have a hard time lining up the robot so that we can place hatches. **Describe the solution you'd like** A function that would use vision processing to align with the reflective tape above the target. **Describe alternatives you've considered** We could just use a camera to let the drivers see, but that still relies on them being good at driving and the camera functioning well. **Additional context** Example of how we could use the limelight: ![lime](https://user-images.githubusercontent.com/16826817/52147270-c873eb00-262b-11e9-9ea8-14442f957a85.gif) ([http://docs.limelightvision.io/en/latest/software_change_log.html](http://docs.limelightvision.io/en/latest/software_change_log.html))
process
alignment function vision assisted driving is your feature request related to a problem please describe drivers might have a hard time lining up the robot so that we can place hatches describe the solution you d like a function that would use vision processing to align with the reflective tape above the target describe alternatives you ve considered we could just use a camera to let the drivers see but that still relies on them being good at driving and the camera functioning well additional context example of how we could use the limelight
1
109,731
11,648,237,696
IssuesEvent
2020-03-01 19:34:49
DNNCommunity/DNNDocs
https://api.github.com/repos/DNNCommunity/DNNDocs
closed
Correct the DNN Platform Suggested Upgrade Path
documentation
In many cases, one can upgrade directly to 09.05.00 from 08.00.04. However, we recommend an intermediate upgrade to 09.01.01 due to significant changes for 09.02.00. Likewise, we recommend an intermediate "stepping-stone" upgrade to 09.03.02 due to the Dependency Injection implementation in 09.04.00. There is no need to upgrade to 09.04.04 first though. --- #### Document Details ⚠ *Do not edit this section. It is required for dnndocs.com ➟ Core Team processing.* * Content Source: [setup-upgrades-suggested-upgrade-path](https://github.com/DNNCommunity/DNNDocs/blob/master/content/getting-started/setup/upgrades/suggested-upgrade-path/index.md/#L1)
1.0
Correct the DNN Platform Suggested Upgrade Path - In many cases, one can upgrade directly to 09.05.00 from 08.00.04. However, we recommend an intermediate upgrade to 09.01.01 due to significant changes for 09.02.00. Likewise, we recommend an intermediate "stepping-stone" upgrade to 09.03.02 due to the Dependency Injection implementation in 09.04.00. There is no need to upgrade to 09.04.04 first though. --- #### Document Details ⚠ *Do not edit this section. It is required for dnndocs.com ➟ Core Team processing.* * Content Source: [setup-upgrades-suggested-upgrade-path](https://github.com/DNNCommunity/DNNDocs/blob/master/content/getting-started/setup/upgrades/suggested-upgrade-path/index.md/#L1)
non_process
correct the dnn platform suggested upgrade path in many cases one can upgrade directly to from however we recommend an intermediate upgrade to due to significant changes for likewise we recommend an intermediate stepping stone upgrade to due to the dependency injection implementation in there is no need to upgrade to first though document details ⚠ do not edit this section it is required for dnndocs com ➟ core team processing content source
0
2,887
5,865,382,105
IssuesEvent
2017-05-13 03:24:18
P0cL4bs/WiFi-Pumpkin
https://api.github.com/repos/P0cL4bs/WiFi-Pumpkin
closed
TCP-Proxy plugin imageCap Error
in process priority
Running plugin SSLStrip+Dns2-Proxy ****Testing shows HTTPS not removed?**** ***Output*** root@kali:~# wifi-pumpkin Loading GUI... WiFi-Pumpkin Running! [*] Loading debugging mode [*] Current Session::ID [OTI2MTk=] [*] Configuring hostapd... [*] enable forwarding in iptables... [*] Configuring dhcpd... [*] Sharing Internet Connections with NAT... [New Thread 5521 (hostapd)] [New Thread 5523 (Dns2Proxy)] SSLstrip v0.9 + POC by Leonardo Nve [*] TCPProxy running on port 80/8080: TCPProxy::ftp status:On TCPProxy::hexdump status:On TCPProxy::kerberos status:On TCPProxy::summary status:On TCPProxy::imageCap status:On TCPProxy::httpCap status:On TCPProxy::emails status:On ------------------------------- AP::[PumpAP] Running... AP::BSSID::[60:e3:27:0a:9d:ac] CH 11 libpng warning: iCCP: known incorrect sRGB profile [New Thread 1756623984 (Dns2proxy::Capture)] [New Thread 1748235376 (Netcreds::Credentials)] [New Thread 1739846768 (Netcreds::Urls)] Traceback (most recent call last): File "/usr/share/WiFi-Pumpkin/core/servers/proxy/tcp/intercept.py", line 43, in run self.main() File "/usr/share/WiFi-Pumpkin/core/servers/proxy/tcp/intercept.py", line 94, in main self.plugins[Active].filterPackets(pkt) File "/usr/share/WiFi-Pumpkin/plugins/analyzers/image.py", line 38, in filterPackets file_name = 'logs/ImagesCap/%s_%s%s' % (self.session,self.random_char(5), file_extension) File "/usr/share/WiFi-Pumpkin/plugins/analyzers/image.py", line 43, in random_char return ''.join(random.choice(string.ascii_letters) for x in range(y)) File "/usr/share/WiFi-Pumpkin/plugins/analyzers/image.py", line 43, in <genexpr> return ''.join(random.choice(string.ascii_letters) for x in range(y)) NameError: global name 'string' is not defined #### Please tell us details about your environment. * Card wireless adapters name (please check if support AP/mode):TL-WN823N * Version used tool: 0.8.5 * Virtual Machine (yes or no and which): no * Operating System and version: Kali Linux
1.0
TCP-Proxy plugin imageCap Error - Running plugin SSLStrip+Dns2-Proxy ****Testing shows HTTPS not removed?**** ***Output*** root@kali:~# wifi-pumpkin Loading GUI... WiFi-Pumpkin Running! [*] Loading debugging mode [*] Current Session::ID [OTI2MTk=] [*] Configuring hostapd... [*] enable forwarding in iptables... [*] Configuring dhcpd... [*] Sharing Internet Connections with NAT... [New Thread 5521 (hostapd)] [New Thread 5523 (Dns2Proxy)] SSLstrip v0.9 + POC by Leonardo Nve [*] TCPProxy running on port 80/8080: TCPProxy::ftp status:On TCPProxy::hexdump status:On TCPProxy::kerberos status:On TCPProxy::summary status:On TCPProxy::imageCap status:On TCPProxy::httpCap status:On TCPProxy::emails status:On ------------------------------- AP::[PumpAP] Running... AP::BSSID::[60:e3:27:0a:9d:ac] CH 11 libpng warning: iCCP: known incorrect sRGB profile [New Thread 1756623984 (Dns2proxy::Capture)] [New Thread 1748235376 (Netcreds::Credentials)] [New Thread 1739846768 (Netcreds::Urls)] Traceback (most recent call last): File "/usr/share/WiFi-Pumpkin/core/servers/proxy/tcp/intercept.py", line 43, in run self.main() File "/usr/share/WiFi-Pumpkin/core/servers/proxy/tcp/intercept.py", line 94, in main self.plugins[Active].filterPackets(pkt) File "/usr/share/WiFi-Pumpkin/plugins/analyzers/image.py", line 38, in filterPackets file_name = 'logs/ImagesCap/%s_%s%s' % (self.session,self.random_char(5), file_extension) File "/usr/share/WiFi-Pumpkin/plugins/analyzers/image.py", line 43, in random_char return ''.join(random.choice(string.ascii_letters) for x in range(y)) File "/usr/share/WiFi-Pumpkin/plugins/analyzers/image.py", line 43, in <genexpr> return ''.join(random.choice(string.ascii_letters) for x in range(y)) NameError: global name 'string' is not defined #### Please tell us details about your environment. * Card wireless adapters name (please check if support AP/mode):TL-WN823N * Version used tool: 0.8.5 * Virtual Machine (yes or no and which): no * Operating System and version: Kali Linux
process
tcp proxy plugin imagecap error running plugin sslstrip proxy testing shows https not removed output root kali wifi pumpkin loading gui wifi pumpkin running loading debugging mode current session id configuring hostapd enable forwarding in iptables configuring dhcpd sharing internet connections with nat sslstrip poc by leonardo nve tcpproxy running on port tcpproxy ftp status on tcpproxy hexdump status on tcpproxy kerberos status on tcpproxy summary status on tcpproxy imagecap status on tcpproxy httpcap status on tcpproxy emails status on ap running ap bssid ch libpng warning iccp known incorrect srgb profile traceback most recent call last file usr share wifi pumpkin core servers proxy tcp intercept py line in run self main file usr share wifi pumpkin core servers proxy tcp intercept py line in main self plugins filterpackets pkt file usr share wifi pumpkin plugins analyzers image py line in filterpackets file name logs imagescap s s s self session self random char file extension file usr share wifi pumpkin plugins analyzers image py line in random char return join random choice string ascii letters for x in range y file usr share wifi pumpkin plugins analyzers image py line in return join random choice string ascii letters for x in range y nameerror global name string is not defined please tell us details about your environment card wireless adapters name please check if support ap mode tl version used tool virtual machine yes or no and which no operating system and version kali linux
1
608,933
18,851,544,832
IssuesEvent
2021-11-11 21:36:57
adirh3/Fluent-Search
https://api.github.com/repos/adirh3/Fluent-Search
closed
Can't Open Folder When Focused on Desktop
bug High Priority
**Describe the bug** FS can't open folders when invoked on desktop. **To Reproduce** Steps to reproduce the behavior: 1. Close all windows and go to desktop; 2. Click on desktop to make sure desktop is in focus; 3. Use hotkey to invoke FS; 4. Search for a folder and try to open it with enter key. **Expected behavior** FS should be able to open the folder regardless of focus. **Screenshots** None. **Desktop (please complete the following information):** - Windows 10 Version: 21H1 - Fluent Search Version 0.9.89.50 **Additional context** The bug is consistent throughout the last few nightly releases.
1.0
Can't Open Folder When Focused on Desktop - **Describe the bug** FS can't open folders when invoked on desktop. **To Reproduce** Steps to reproduce the behavior: 1. Close all windows and go to desktop; 2. Click on desktop to make sure desktop is in focus; 3. Use hotkey to invoke FS; 4. Search for a folder and try to open it with enter key. **Expected behavior** FS should be able to open the folder regardless of focus. **Screenshots** None. **Desktop (please complete the following information):** - Windows 10 Version: 21H1 - Fluent Search Version 0.9.89.50 **Additional context** The bug is consistent throughout the last few nightly releases.
non_process
can t open folder when focused on desktop describe the bug fs can t open folders when invoked on desktop to reproduce steps to reproduce the behavior close all windows and go to desktop click on desktop to make sure desktop is in focus use hotkey to invoke fs search for a folder and try to open it with enter key expected behavior fs should be able to open the folder regardless of focus screenshots none desktop please complete the following information windows version fluent search version additional context the bug is consistent throughout the last few nightly releases
0
304,684
23,078,253,165
IssuesEvent
2022-07-26 03:20:39
sbip-sg/llvm-project
https://api.github.com/repos/sbip-sg/llvm-project
closed
Rename branches `sbip-llvm-***` to `llvm-sbip-***`
documentation
I renamed the branches `sbip-llvm-13`, `sbip-llvm-14` to `llvm-sbip-13`, `llvm-sbip-14` to match the branch names with the release names. @thanhtoantnt: you may need to update your local code base.
1.0
Rename branches `sbip-llvm-***` to `llvm-sbip-***` - I renamed the branches `sbip-llvm-13`, `sbip-llvm-14` to `llvm-sbip-13`, `llvm-sbip-14` to match the branch names with the release names. @thanhtoantnt: you may need to update your local code base.
non_process
rename branches sbip llvm to llvm sbip i renamed the branches sbip llvm sbip llvm to llvm sbip llvm sbip to match the branch names with the release names thanhtoantnt you may need to update your local code base
0
19,152
25,232,446,745
IssuesEvent
2022-11-14 21:04:27
benthosdev/benthos
https://api.github.com/repos/benthosdev/benthos
closed
Support to lua scripts in the Redis processor
enhancement processors
Hello. Would be nice to be able to execute scripts in the Redis processor. I will implement a first version and submit a PR.
1.0
Support to lua scripts in the Redis processor - Hello. Would be nice to be able to execute scripts in the Redis processor. I will implement a first version and submit a PR.
process
support to lua scripts in the redis processor hello would be nice to be able to execute scripts in the redis processor i will implement a first version and submit a pr
1
14,415
9,306,057,784
IssuesEvent
2019-03-25 08:42:10
computer-technik-krause/ctk-warehouse-management
https://api.github.com/repos/computer-technik-krause/ctk-warehouse-management
opened
Simple Benutzeradministration anlegen
enhancement security
-> Benutzerübersicht -> Benutzer anlegen -> Benutzer bearbeiten -> Benutzer deaktivieren
True
Simple Benutzeradministration anlegen - -> Benutzerübersicht -> Benutzer anlegen -> Benutzer bearbeiten -> Benutzer deaktivieren
non_process
simple benutzeradministration anlegen benutzerübersicht benutzer anlegen benutzer bearbeiten benutzer deaktivieren
0
492,981
14,223,854,080
IssuesEvent
2020-11-17 18:48:23
aiidateam/aiida-core
https://api.github.com/repos/aiidateam/aiida-core
opened
Small change in the behavior of `verdi group list`
priority/nice-to-have topic/groups type/feature request
After using groups to organize my nodes, I find that the `verdi group list` may be improved by the two following small changes: - List order by label instead of pk (as this has a better chance of listing related group together). - List the count of nodes by default (I often find this considerable more relevant than the `User` or `Type string`). - Allow to include the "begining" of the label as an argument and filter results accordingly (ease of just listing the groups you are interested in, specially if you have a lot of groups). Example: ``` $ verdi group list PK Label Type string User Node count ---- -------------------------------- ------------- ------------------------- ------------ 1 structures/3ddb_2020/all core myname@mail.com 24901 3 structures/3ddb_2020/electrides core myname@mail.com 45 2 workflows/restartgen/0started core myname@mail.com 0 4 workflows/restartgen/finished core myname@mail.com 0 $ verdi group list struct PK Label Type string User Node count ---- -------------------------------- ------------- ------------------------- ------------ 1 structures/3ddb_2020/all core myname@mail.com 24901 3 structures/3ddb_2020/electrides core myname@mail.com 45 ```
1.0
Small change in the behavior of `verdi group list` - After using groups to organize my nodes, I find that the `verdi group list` may be improved by the two following small changes: - List order by label instead of pk (as this has a better chance of listing related group together). - List the count of nodes by default (I often find this considerable more relevant than the `User` or `Type string`). - Allow to include the "begining" of the label as an argument and filter results accordingly (ease of just listing the groups you are interested in, specially if you have a lot of groups). Example: ``` $ verdi group list PK Label Type string User Node count ---- -------------------------------- ------------- ------------------------- ------------ 1 structures/3ddb_2020/all core myname@mail.com 24901 3 structures/3ddb_2020/electrides core myname@mail.com 45 2 workflows/restartgen/0started core myname@mail.com 0 4 workflows/restartgen/finished core myname@mail.com 0 $ verdi group list struct PK Label Type string User Node count ---- -------------------------------- ------------- ------------------------- ------------ 1 structures/3ddb_2020/all core myname@mail.com 24901 3 structures/3ddb_2020/electrides core myname@mail.com 45 ```
non_process
small change in the behavior of verdi group list after using groups to organize my nodes i find that the verdi group list may be improved by the two following small changes list order by label instead of pk as this has a better chance of listing related group together list the count of nodes by default i often find this considerable more relevant than the user or type string allow to include the begining of the label as an argument and filter results accordingly ease of just listing the groups you are interested in specially if you have a lot of groups example verdi group list pk label type string user node count structures all core myname mail com structures electrides core myname mail com workflows restartgen core myname mail com workflows restartgen finished core myname mail com verdi group list struct pk label type string user node count structures all core myname mail com structures electrides core myname mail com
0
21,352
29,184,535,044
IssuesEvent
2023-05-19 14:25:54
microsoft/vscode
https://api.github.com/repos/microsoft/vscode
opened
Add isProposedApiEnabled check for env var scope
bug api terminal-process
It's used in https://github.com/microsoft/vscode/pull/182883, this would be easiest after that's merged.
1.0
Add isProposedApiEnabled check for env var scope - It's used in https://github.com/microsoft/vscode/pull/182883, this would be easiest after that's merged.
process
add isproposedapienabled check for env var scope it s used in this would be easiest after that s merged
1
19,780
26,162,822,901
IssuesEvent
2022-12-31 21:14:16
allinurl/goaccess
https://api.github.com/repos/allinurl/goaccess
closed
Custom log format help - strip URL parameters
question log-processing
Hi, my $request is _"GET /pix.png?url=/blog/it/2020-04-18-tips-disabling_bluetooth_on_raspberry_pi/,uid=589299321 HTTP/1.1"_ and I want to transform it to _"GET /blog/it/2020-04-18-tips-disabling_bluetooth_on_raspberry_pi/ HTTP/1.1"_ (i.e without **"/pix.png?url="** and without "**,uid=58929xxx"**). It's possible to do this ? with a custom log-format for example ... NGINX access.log sample : ``` 92.50.87.22 - - [28/Dec/2022:14:44:32 +0000] "GET /pix.png?url=/blog/it/2020-04-18-tips-disabling_bluetooth_on_raspberry_pi/,uid=5892OO61 HTTP/1.1" 200 98 "https://exemple.net/" "Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:108.0) Gecko/20100101 Firefox/108.0" 2.81.89.87 - - [28/Dec/2022:14:44:42 +0000] "GET /pix.png?url=/blog/it/2020-04-18-tips-disabling_bluetooth_on_raspberry_pi/,uid=589299321 HTTP/1.1" 200 98 "https://example.net/" "Mozilla/5.0 (Windows NT 6.3; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/108.0.0.0 Safari/537.36" 123.241.138.100 - - [28/Dec/2022:14:44:44 +0000] "GET /pix.png?url=/blog/hi-fi/2020-10-04-naim_audio_nait_2/,uid=589495349 HTTP/1.1" 200 98 "https://example.net/" "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_6) AppleWebKit/605.1.15 (KHTML, like Gecko) Version/14.1.1 Safari/605.1.15" ``` Many thank for your help.
1.0
Custom log format help - strip URL parameters - Hi, my $request is _"GET /pix.png?url=/blog/it/2020-04-18-tips-disabling_bluetooth_on_raspberry_pi/,uid=589299321 HTTP/1.1"_ and I want to transform it to _"GET /blog/it/2020-04-18-tips-disabling_bluetooth_on_raspberry_pi/ HTTP/1.1"_ (i.e without **"/pix.png?url="** and without "**,uid=58929xxx"**). It's possible to do this ? with a custom log-format for example ... NGINX access.log sample : ``` 92.50.87.22 - - [28/Dec/2022:14:44:32 +0000] "GET /pix.png?url=/blog/it/2020-04-18-tips-disabling_bluetooth_on_raspberry_pi/,uid=5892OO61 HTTP/1.1" 200 98 "https://exemple.net/" "Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:108.0) Gecko/20100101 Firefox/108.0" 2.81.89.87 - - [28/Dec/2022:14:44:42 +0000] "GET /pix.png?url=/blog/it/2020-04-18-tips-disabling_bluetooth_on_raspberry_pi/,uid=589299321 HTTP/1.1" 200 98 "https://example.net/" "Mozilla/5.0 (Windows NT 6.3; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/108.0.0.0 Safari/537.36" 123.241.138.100 - - [28/Dec/2022:14:44:44 +0000] "GET /pix.png?url=/blog/hi-fi/2020-10-04-naim_audio_nait_2/,uid=589495349 HTTP/1.1" 200 98 "https://example.net/" "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_6) AppleWebKit/605.1.15 (KHTML, like Gecko) Version/14.1.1 Safari/605.1.15" ``` Many thank for your help.
process
custom log format help strip url parameters hi my request is get pix png url blog it tips disabling bluetooth on raspberry pi uid http and i want to transform it to get blog it tips disabling bluetooth on raspberry pi http i e without pix png url and without uid it s possible to do this with a custom log format for example nginx access log sample get pix png url blog it tips disabling bluetooth on raspberry pi uid http mozilla windows nt rv gecko firefox get pix png url blog it tips disabling bluetooth on raspberry pi uid http mozilla windows nt applewebkit khtml like gecko chrome safari get pix png url blog hi fi naim audio nait uid http mozilla macintosh intel mac os x applewebkit khtml like gecko version safari many thank for your help
1
240,664
7,804,359,348
IssuesEvent
2018-06-11 07:07:32
BuiltBrokenModding/Atomic-Science
https://api.github.com/repos/BuiltBrokenModding/Atomic-Science
opened
[1.7] World mechanic: world generation
enhancement low priority
Custom world generation, disabled by default. To allow random areas to spawn: - [ ] Radioactive barrels - [ ] Radioactive bunker - [ ] Radioactive pools of water - [ ] Radioactive pools of sludge - [ ] Radioactive bomb blasts This will be used as examples of world generation for modpack makers. Should be intergrated with VoltzEngine in some way for reuse and done via JSON templates. Each area will have some risk vs reward mechanic. In which cleaning up the area could give rare resources. At the risk of radiation exposure and risk of death. For this new content should be added as decoration.
1.0
[1.7] World mechanic: world generation - Custom world generation, disabled by default. To allow random areas to spawn: - [ ] Radioactive barrels - [ ] Radioactive bunker - [ ] Radioactive pools of water - [ ] Radioactive pools of sludge - [ ] Radioactive bomb blasts This will be used as examples of world generation for modpack makers. Should be intergrated with VoltzEngine in some way for reuse and done via JSON templates. Each area will have some risk vs reward mechanic. In which cleaning up the area could give rare resources. At the risk of radiation exposure and risk of death. For this new content should be added as decoration.
non_process
world mechanic world generation custom world generation disabled by default to allow random areas to spawn radioactive barrels radioactive bunker radioactive pools of water radioactive pools of sludge radioactive bomb blasts this will be used as examples of world generation for modpack makers should be intergrated with voltzengine in some way for reuse and done via json templates each area will have some risk vs reward mechanic in which cleaning up the area could give rare resources at the risk of radiation exposure and risk of death for this new content should be added as decoration
0
18,521
24,551,807,216
IssuesEvent
2022-10-12 13:10:07
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[iOS] [Offline indicator] Issues related to offline message in consent PDF while enrolling into study
Bug P1 iOS Process: Fixed Process: Tested QA Process: Tested dev
Steps: 1. Signup/login 2. Enroll into the study 3. Navigate till enrollment completion screen 4. Click on View consent 5. Switch off the internet Actual: 1. You are offline message is displayed on top 2. Share button is enabled Expected: 1. Toast message should be displayed instead of message on top 2. Share button should be disabled Refer screenshot ![IMG_3117](https://user-images.githubusercontent.com/60386291/180207794-cc2e2880-be72-449b-927b-7488e9c88a39.PNG)
3.0
[iOS] [Offline indicator] Issues related to offline message in consent PDF while enrolling into study - Steps: 1. Signup/login 2. Enroll into the study 3. Navigate till enrollment completion screen 4. Click on View consent 5. Switch off the internet Actual: 1. You are offline message is displayed on top 2. Share button is enabled Expected: 1. Toast message should be displayed instead of message on top 2. Share button should be disabled Refer screenshot ![IMG_3117](https://user-images.githubusercontent.com/60386291/180207794-cc2e2880-be72-449b-927b-7488e9c88a39.PNG)
process
issues related to offline message in consent pdf while enrolling into study steps signup login enroll into the study navigate till enrollment completion screen click on view consent switch off the internet actual you are offline message is displayed on top share button is enabled expected toast message should be displayed instead of message on top share button should be disabled refer screenshot
1
14,794
18,069,283,276
IssuesEvent
2021-09-20 23:35:55
allinurl/goaccess
https://api.github.com/repos/allinurl/goaccess
closed
avif (and probably jxl) are treated as "requested files (URLs)"
add log-processing
These should be in the STATIC requests section, however they're currently in the URLs section Looks like avif and jxl would just go here: https://github.com/allinurl/goaccess/blob/master/src/settings.c#L169-L205
1.0
avif (and probably jxl) are treated as "requested files (URLs)" - These should be in the STATIC requests section, however they're currently in the URLs section Looks like avif and jxl would just go here: https://github.com/allinurl/goaccess/blob/master/src/settings.c#L169-L205
process
avif and probably jxl are treated as requested files urls these should be in the static requests section however they re currently in the urls section looks like avif and jxl would just go here
1
44,566
5,846,546,761
IssuesEvent
2017-05-10 16:24:29
18F/crime-data-explorer
https://api.github.com/repos/18F/crime-data-explorer
closed
Explore & refine pathways for agency navigation
design
Explore the usefulness of being able to narrow down searches by population size.
1.0
Explore & refine pathways for agency navigation - Explore the usefulness of being able to narrow down searches by population size.
non_process
explore refine pathways for agency navigation explore the usefulness of being able to narrow down searches by population size
0
24,120
3,917,070,283
IssuesEvent
2016-04-21 06:23:36
irnawansuprapti/openbiz-cubi
https://api.github.com/repos/irnawansuprapti/openbiz-cubi
closed
What Doctors Hate to Tell You About Skin Care
auto-migrated Priority-Medium spam Type-Defect
``` Now, men wound work is gaining solon tending from varied parties regarding men pare protection, but how can men comprehend the unexcelled set for them.Men and women screw diametrical tegument but how can we specialize between them? Men do not pass untold money on toiletries, facial watercolor but trim statesman than women. There are solon products for women but there are whatever for men. Yet, the content of tegument upkeep products for men is unimportant compared to women's.Here are both echt tips that men can persist to insure white and ruddy tackling. Use facial cleaner formerly in the forenoon and erst in the evening, try to avoid ghb as it can make your tegument dry. When shaving, try using moderate depilation toiletries or gel, depilation at least 1-3 times a week. http://nitroshredadvice.com/novus-serum/ ``` Original issue reported on code.google.com by `ChunAnt...@gmail.com` on 17 Apr 2015 at 7:01
1.0
What Doctors Hate to Tell You About Skin Care - ``` Now, men wound work is gaining solon tending from varied parties regarding men pare protection, but how can men comprehend the unexcelled set for them.Men and women screw diametrical tegument but how can we specialize between them? Men do not pass untold money on toiletries, facial watercolor but trim statesman than women. There are solon products for women but there are whatever for men. Yet, the content of tegument upkeep products for men is unimportant compared to women's.Here are both echt tips that men can persist to insure white and ruddy tackling. Use facial cleaner formerly in the forenoon and erst in the evening, try to avoid ghb as it can make your tegument dry. When shaving, try using moderate depilation toiletries or gel, depilation at least 1-3 times a week. http://nitroshredadvice.com/novus-serum/ ``` Original issue reported on code.google.com by `ChunAnt...@gmail.com` on 17 Apr 2015 at 7:01
non_process
what doctors hate to tell you about skin care now men wound work is gaining solon tending from varied parties regarding men pare protection but how can men comprehend the unexcelled set for them men and women screw diametrical tegument but how can we specialize between them men do not pass untold money on toiletries facial watercolor but trim statesman than women there are solon products for women but there are whatever for men yet the content of tegument upkeep products for men is unimportant compared to women s here are both echt tips that men can persist to insure white and ruddy tackling use facial cleaner formerly in the forenoon and erst in the evening try to avoid ghb as it can make your tegument dry when shaving try using moderate depilation toiletries or gel depilation at least times a week original issue reported on code google com by chunant gmail com on apr at
0
11,428
14,248,189,867
IssuesEvent
2020-11-19 12:35:01
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
AskTUG: Coprocessor support `bson` and `protocobuffers`
feature/review again sig/coprocessor status/help-wanted type/req-from-ask-tug
## Feature Request Could we support to push down functions for `bson` and `protocobuffers`? With the above feature, tikv will be more friendly to users from No-SQL such as MongoDB. Then the No-SQL users can use tikv's computation source as `TiDB` did. This is from [ask-tug](https://asktug.com/t/topic/34614/49?u=shirly-pingcap)
1.0
AskTUG: Coprocessor support `bson` and `protocobuffers` - ## Feature Request Could we support to push down functions for `bson` and `protocobuffers`? With the above feature, tikv will be more friendly to users from No-SQL such as MongoDB. Then the No-SQL users can use tikv's computation source as `TiDB` did. This is from [ask-tug](https://asktug.com/t/topic/34614/49?u=shirly-pingcap)
process
asktug coprocessor support bson and protocobuffers feature request could we support to push down functions for bson and protocobuffers with the above feature tikv will be more friendly to users from no sql such as mongodb then the no sql users can use tikv s computation source as tidb did this is from
1
330,823
24,279,011,916
IssuesEvent
2022-09-28 15:48:41
ouvroir/righettino
https://api.github.com/repos/ouvroir/righettino
opened
Fiche récapitualitive du projet
:screwdriver: documentation
Salut, j'ai rangé les notes du midi-causerie dans le repo [dahnotes](https://github.com/ouvroir/dahnotes/blob/master/ouvroir_Righettino.md). Est-ce que je pourrais te demander de corriger les erreurs que tu vois? C'est possible que des personnes me demandent les notes puisque nous n'avons pas enregistré la séance :) merci
1.0
Fiche récapitualitive du projet - Salut, j'ai rangé les notes du midi-causerie dans le repo [dahnotes](https://github.com/ouvroir/dahnotes/blob/master/ouvroir_Righettino.md). Est-ce que je pourrais te demander de corriger les erreurs que tu vois? C'est possible que des personnes me demandent les notes puisque nous n'avons pas enregistré la séance :) merci
non_process
fiche récapitualitive du projet salut j ai rangé les notes du midi causerie dans le repo est ce que je pourrais te demander de corriger les erreurs que tu vois c est possible que des personnes me demandent les notes puisque nous n avons pas enregistré la séance merci
0
104,736
4,218,120,155
IssuesEvent
2016-06-30 15:05:08
dminca/drupal-docker
https://api.github.com/repos/dminca/drupal-docker
closed
Provide persistent database storage
database docker feature Priority: High
After **rebooting/stopping/removing** docker containers, let the database maintain persistence, thus avoiding any errors caused by the reboot (ie. unable to link cache, or 'content not found'). ### Useful resources: - [How to create a Docker image for PostgreSQL and persist data][1] - [Data-only container madness][2] - [Persistent PostgreSQL inside Docker][3] - [Advanced Docker Volumes][4] - [Tiny Docker Pieces, Loosely Joined][5] - [tianon/true docker image for Data only containers][6] [1]: https://www.andreagrandi.it/2015/02/21/how-to-create-a-docker-image-for-postgresql-and-persist-data/ [2]: http://container42.com/2014/11/18/data-only-container-madness/ [3]: https://crondev.com/persistent-postgresql-inside-docker/ [4]: http://crosbymichael.com/advanced-docker-volumes.html [5]: http://www.offermann.us/2013/12/tiny-docker-pieces-loosely-joined.html [6]: https://hub.docker.com/r/tianon/true/
1.0
Provide persistent database storage - After **rebooting/stopping/removing** docker containers, let the database maintain persistence, thus avoiding any errors caused by the reboot (ie. unable to link cache, or 'content not found'). ### Useful resources: - [How to create a Docker image for PostgreSQL and persist data][1] - [Data-only container madness][2] - [Persistent PostgreSQL inside Docker][3] - [Advanced Docker Volumes][4] - [Tiny Docker Pieces, Loosely Joined][5] - [tianon/true docker image for Data only containers][6] [1]: https://www.andreagrandi.it/2015/02/21/how-to-create-a-docker-image-for-postgresql-and-persist-data/ [2]: http://container42.com/2014/11/18/data-only-container-madness/ [3]: https://crondev.com/persistent-postgresql-inside-docker/ [4]: http://crosbymichael.com/advanced-docker-volumes.html [5]: http://www.offermann.us/2013/12/tiny-docker-pieces-loosely-joined.html [6]: https://hub.docker.com/r/tianon/true/
non_process
provide persistent database storage after rebooting stopping removing docker containers let the database maintain persistence thus avoiding any errors caused by the reboot ie unable to link cache or content not found useful resources
0
507,265
14,679,956,758
IssuesEvent
2020-12-31 08:36:29
k8smeetup/website-tasks
https://api.github.com/repos/k8smeetup/website-tasks
opened
/docs/reference/glossary/annotation.md
lang/zh priority/P0 sync/update version/master welcome
Source File: [/docs/reference/glossary/annotation.md](https://github.com/kubernetes/website/blob/master/content/en/docs/reference/glossary/annotation.md) Diff 命令参考: ```bash # 查看原始文档与翻译文档更新差异 git diff --no-index -- content/en/docs/reference/glossary/annotation.md content/zh/docs/reference/glossary/annotation.md # 跨分支持查看原始文档更新差异 git diff release-1.19 master -- content/en/docs/reference/glossary/annotation.md ```
1.0
/docs/reference/glossary/annotation.md - Source File: [/docs/reference/glossary/annotation.md](https://github.com/kubernetes/website/blob/master/content/en/docs/reference/glossary/annotation.md) Diff 命令参考: ```bash # 查看原始文档与翻译文档更新差异 git diff --no-index -- content/en/docs/reference/glossary/annotation.md content/zh/docs/reference/glossary/annotation.md # 跨分支持查看原始文档更新差异 git diff release-1.19 master -- content/en/docs/reference/glossary/annotation.md ```
non_process
docs reference glossary annotation md source file diff 命令参考 bash 查看原始文档与翻译文档更新差异 git diff no index content en docs reference glossary annotation md content zh docs reference glossary annotation md 跨分支持查看原始文档更新差异 git diff release master content en docs reference glossary annotation md
0
292,597
25,224,953,554
IssuesEvent
2022-11-14 15:21:08
Princeton-CDH/geniza
https://api.github.com/repos/Princeton-CDH/geniza
closed
As a content editor, I want to view source URLs when I download the sources CSV in order to more easily find/update external sources.
🗜️ awaiting testing
Context: there are a few Asana cleaning tasks pertaining to the cleaning/checking of external links (particularly Marina's google doc translations). When creating a work flow for the RAs, I discovered that the current sources CSV download does not include the URLs, and therefore it is difficult to develop a proper workflow. In conversation with @rlskoeser on 8/17/21, she said it should be easy to add the existing URL field to the CSV. Thanks!
1.0
As a content editor, I want to view source URLs when I download the sources CSV in order to more easily find/update external sources. - Context: there are a few Asana cleaning tasks pertaining to the cleaning/checking of external links (particularly Marina's google doc translations). When creating a work flow for the RAs, I discovered that the current sources CSV download does not include the URLs, and therefore it is difficult to develop a proper workflow. In conversation with @rlskoeser on 8/17/21, she said it should be easy to add the existing URL field to the CSV. Thanks!
non_process
as a content editor i want to view source urls when i download the sources csv in order to more easily find update external sources context there are a few asana cleaning tasks pertaining to the cleaning checking of external links particularly marina s google doc translations when creating a work flow for the ras i discovered that the current sources csv download does not include the urls and therefore it is difficult to develop a proper workflow in conversation with rlskoeser on she said it should be easy to add the existing url field to the csv thanks
0
33,640
14,163,944,015
IssuesEvent
2020-11-12 03:40:45
Azure/azure-cli
https://api.github.com/repos/Azure/azure-cli
closed
az aks update: "unrecognized arguments: -—load-balancer-outbound-ips"
AKS Service Attention question
Trying to update the public IP for my aks cluster via the az aks update command I'm receiving the error: `unrecognized arguments: -—load-balancer-outbound-ips` The command I am replicating from documentation is: `az aks update \ --resource-group myResourceGroup \ --name myAKSCluster \ --load-balancer-outbound-ips <publicIpId1>,<publicIpId2> ` Originally I was on az cli version 2.12 and aks preview 0.4.33. I upgraded to az cli version 2.14 and aks preview 0.4.65, but that didn't help. I tried to remove az extension aks-preview but that didn't work either. Is the documentation wrong or is there a bug here? Thanks. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: d75c3a8d-1187-bea1-a7d4-7ebba71c1790 * Version Independent ID: 75b0c352-762e-16ae-5cc6-d807ffc1dc3f * Content: [az aks](https://docs.microsoft.com/en-us/cli/azure/aks?view=azure-cli-latest) * Content Source: [latest/docs-ref-autogen/aks.yml](https://github.com/MicrosoftDocs/azure-docs-cli/blob/master/latest/docs-ref-autogen/aks.yml) * Service: **container-service** * GitHub Login: @rloutlaw * Microsoft Alias: **routlaw**
1.0
az aks update: "unrecognized arguments: -—load-balancer-outbound-ips" - Trying to update the public IP for my aks cluster via the az aks update command I'm receiving the error: `unrecognized arguments: -—load-balancer-outbound-ips` The command I am replicating from documentation is: `az aks update \ --resource-group myResourceGroup \ --name myAKSCluster \ --load-balancer-outbound-ips <publicIpId1>,<publicIpId2> ` Originally I was on az cli version 2.12 and aks preview 0.4.33. I upgraded to az cli version 2.14 and aks preview 0.4.65, but that didn't help. I tried to remove az extension aks-preview but that didn't work either. Is the documentation wrong or is there a bug here? Thanks. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: d75c3a8d-1187-bea1-a7d4-7ebba71c1790 * Version Independent ID: 75b0c352-762e-16ae-5cc6-d807ffc1dc3f * Content: [az aks](https://docs.microsoft.com/en-us/cli/azure/aks?view=azure-cli-latest) * Content Source: [latest/docs-ref-autogen/aks.yml](https://github.com/MicrosoftDocs/azure-docs-cli/blob/master/latest/docs-ref-autogen/aks.yml) * Service: **container-service** * GitHub Login: @rloutlaw * Microsoft Alias: **routlaw**
non_process
az aks update unrecognized arguments —load balancer outbound ips trying to update the public ip for my aks cluster via the az aks update command i m receiving the error unrecognized arguments —load balancer outbound ips the command i am replicating from documentation is az aks update resource group myresourcegroup name myakscluster load balancer outbound ips originally i was on az cli version and aks preview i upgraded to az cli version and aks preview but that didn t help i tried to remove az extension aks preview but that didn t work either is the documentation wrong or is there a bug here thanks document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service container service github login rloutlaw microsoft alias routlaw
0
289,133
8,855,179,735
IssuesEvent
2019-01-09 05:09:13
visit-dav/issues-test
https://api.github.com/repos/visit-dav/issues-test
closed
An LC web page on VisIt has incorrect email addresses for the various visit email lists
bug crash likelihood medium priority reviewed severity high wrong results
The web page https://computing.llnl.gov/vis/visit.shtml#Help has incorrect email addresses for all the visit related email lists. It has them as visit-xxx@email.ornl.gov instead of visit-xxx@elist.ornl.gov We should have LC change it. Looking at the page further, it has many errors. The spaces between words appear to be missing and the link to VisIt's web site is also incorrect. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 2241 Status: Resolved Project: VisIt Tracker: Bug Priority: High Subject: An LC web page on VisIt has incorrect email addresses for the various visit email lists Assigned to: Eric Brugger Category: - Target version: 2.10 Author: Eric Brugger Start: 04/23/2015 Due date: % Done: 100% Estimated time: 2.00 hours Created: 04/23/2015 12:40 pm Updated: 09/10/2015 01:26 pm Likelihood: 3 - Occasional Severity: 4 - Crash / Wrong Results Found in version: 2.9.0 Impact: Expected Use: OS: All Support Group: Any Description: The web page https://computing.llnl.gov/vis/visit.shtml#Help has incorrect email addresses for all the visit related email lists. It has them as visit-xxx@email.ornl.gov instead of visit-xxx@elist.ornl.gov We should have LC change it. Looking at the page further, it has many errors. The spaces between words appear to be missing and the link to VisIt's web site is also incorrect. Comments: I submitted a front range ticket about having the LC fix the web page. I was contacted by Rich Cook and he told me where I could check out the web page from the git repository. I have made the changes to the web page and sent Rich an e-mail to that effect. Rich Cook put the new page on the web server and I checked it. All the links are correct and the description of VisIt now matches what is on the VisIt web site.
1.0
An LC web page on VisIt has incorrect email addresses for the various visit email lists - The web page https://computing.llnl.gov/vis/visit.shtml#Help has incorrect email addresses for all the visit related email lists. It has them as visit-xxx@email.ornl.gov instead of visit-xxx@elist.ornl.gov We should have LC change it. Looking at the page further, it has many errors. The spaces between words appear to be missing and the link to VisIt's web site is also incorrect. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 2241 Status: Resolved Project: VisIt Tracker: Bug Priority: High Subject: An LC web page on VisIt has incorrect email addresses for the various visit email lists Assigned to: Eric Brugger Category: - Target version: 2.10 Author: Eric Brugger Start: 04/23/2015 Due date: % Done: 100% Estimated time: 2.00 hours Created: 04/23/2015 12:40 pm Updated: 09/10/2015 01:26 pm Likelihood: 3 - Occasional Severity: 4 - Crash / Wrong Results Found in version: 2.9.0 Impact: Expected Use: OS: All Support Group: Any Description: The web page https://computing.llnl.gov/vis/visit.shtml#Help has incorrect email addresses for all the visit related email lists. It has them as visit-xxx@email.ornl.gov instead of visit-xxx@elist.ornl.gov We should have LC change it. Looking at the page further, it has many errors. The spaces between words appear to be missing and the link to VisIt's web site is also incorrect. Comments: I submitted a front range ticket about having the LC fix the web page. I was contacted by Rich Cook and he told me where I could check out the web page from the git repository. I have made the changes to the web page and sent Rich an e-mail to that effect. Rich Cook put the new page on the web server and I checked it. All the links are correct and the description of VisIt now matches what is on the VisIt web site.
non_process
an lc web page on visit has incorrect email addresses for the various visit email lists the web page has incorrect email addresses for all the visit related email lists it has them as visit xxx email ornl gov instead of visit xxx elist ornl gov we should have lc change it looking at the page further it has many errors the spaces between words appear to be missing and the link to visit s web site is also incorrect redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority high subject an lc web page on visit has incorrect email addresses for the various visit email lists assigned to eric brugger category target version author eric brugger start due date done estimated time hours created pm updated pm likelihood occasional severity crash wrong results found in version impact expected use os all support group any description the web page has incorrect email addresses for all the visit related email lists it has them as visit xxx email ornl gov instead of visit xxx elist ornl gov we should have lc change it looking at the page further it has many errors the spaces between words appear to be missing and the link to visit s web site is also incorrect comments i submitted a front range ticket about having the lc fix the web page i was contacted by rich cook and he told me where i could check out the web page from the git repository i have made the changes to the web page and sent rich an e mail to that effect rich cook put the new page on the web server and i checked it all the links are correct and the description of visit now matches what is on the visit web site
0
402,329
27,363,666,819
IssuesEvent
2023-02-27 17:29:57
rubocop/rubocop
https://api.github.com/repos/rubocop/rubocop
closed
Does sider no longer exist? If so perhaps we can remove from docs?
documentation
<img width="750" alt="image" src="https://user-images.githubusercontent.com/4121835/221633445-721e536b-1c64-43e8-90cc-c4d14a8fba86.png"> <img width="1392" alt="image" src="https://user-images.githubusercontent.com/4121835/221633519-f5c897d6-2176-46aa-a3df-bb4b03d42a4d.png">
1.0
Does sider no longer exist? If so perhaps we can remove from docs? - <img width="750" alt="image" src="https://user-images.githubusercontent.com/4121835/221633445-721e536b-1c64-43e8-90cc-c4d14a8fba86.png"> <img width="1392" alt="image" src="https://user-images.githubusercontent.com/4121835/221633519-f5c897d6-2176-46aa-a3df-bb4b03d42a4d.png">
non_process
does sider no longer exist if so perhaps we can remove from docs img width alt image src img width alt image src
0
404,972
27,500,207,491
IssuesEvent
2023-03-05 16:03:32
tidyverse/ggplot2
https://api.github.com/repos/tidyverse/ggplot2
closed
geom_contour undefined behaviour for multiple observations per grid-point
documentation
If there are multiple observations per grid point for `geom_contour*`, several aspects of behaviour are undefined. I was expecting the mean value for each unique grid point to be used in the contour, but it was not. It seems both the contours themselves and the computed bins/levels are affected. ```r library('ggplot2') M = pracma::meshgrid(seq(0,1,l=10)) X1 = data.frame(x=c(M$X),y=c(M$Y),z=c(M$X * M$Y)) X2 = X1; X2$z = 0 X3 = X1; X3$z = -X1$z plot.one = function(X.){ ggplot(X.,aes(x=x,y=y,z=z)) + geom_contour_filled() } plot.one(X1) plot.one(X2) plot.one(X3) plot.one(rbind(X1,X2)) plot.one(rbind(X1,X2,X3)) plot.one(rbind(X3,X2,X1)) ``` I think the relevant code is [here](https://github.com/tidyverse/ggplot2/blob/main/R/stat-contour.r#L211) I don't necessarily think the aggregation should be performed by `geom_contour`, but perhaps some explanation of how the data will be used can be added to the docs --- or even just a warning that the behaviour is undefined for multiple observations per grid-point.
1.0
geom_contour undefined behaviour for multiple observations per grid-point - If there are multiple observations per grid point for `geom_contour*`, several aspects of behaviour are undefined. I was expecting the mean value for each unique grid point to be used in the contour, but it was not. It seems both the contours themselves and the computed bins/levels are affected. ```r library('ggplot2') M = pracma::meshgrid(seq(0,1,l=10)) X1 = data.frame(x=c(M$X),y=c(M$Y),z=c(M$X * M$Y)) X2 = X1; X2$z = 0 X3 = X1; X3$z = -X1$z plot.one = function(X.){ ggplot(X.,aes(x=x,y=y,z=z)) + geom_contour_filled() } plot.one(X1) plot.one(X2) plot.one(X3) plot.one(rbind(X1,X2)) plot.one(rbind(X1,X2,X3)) plot.one(rbind(X3,X2,X1)) ``` I think the relevant code is [here](https://github.com/tidyverse/ggplot2/blob/main/R/stat-contour.r#L211) I don't necessarily think the aggregation should be performed by `geom_contour`, but perhaps some explanation of how the data will be used can be added to the docs --- or even just a warning that the behaviour is undefined for multiple observations per grid-point.
non_process
geom contour undefined behaviour for multiple observations per grid point if there are multiple observations per grid point for geom contour several aspects of behaviour are undefined i was expecting the mean value for each unique grid point to be used in the contour but it was not it seems both the contours themselves and the computed bins levels are affected r library m pracma meshgrid seq l data frame x c m x y c m y z c m x m y z z z plot one function x ggplot x aes x x y y z z geom contour filled plot one plot one plot one plot one rbind plot one rbind plot one rbind i think the relevant code is i don t necessarily think the aggregation should be performed by geom contour but perhaps some explanation of how the data will be used can be added to the docs or even just a warning that the behaviour is undefined for multiple observations per grid point
0
824
2,516,619,890
IssuesEvent
2015-01-16 06:16:18
waterbearlang/waterbear
https://api.github.com/repos/waterbearlang/waterbear
closed
After showing trashcan, menu jumps to the top
Bug User Friendliness UX / Design
We should maintain the scroll position. Perhaps instead of hiding the menu we should just overlay the trashcan view on top?
1.0
After showing trashcan, menu jumps to the top - We should maintain the scroll position. Perhaps instead of hiding the menu we should just overlay the trashcan view on top?
non_process
after showing trashcan menu jumps to the top we should maintain the scroll position perhaps instead of hiding the menu we should just overlay the trashcan view on top
0
7,043
10,198,617,487
IssuesEvent
2019-08-13 06:05:19
okTurtles/group-income-simple
https://api.github.com/repos/okTurtles/group-income-simple
closed
Double-check changes in PR 559
App:Frontend Kind:Process Priority:High
### Problem My review for PR #559 was rushed as we decided to move forward quickly, and there might be issues in there that I missed. ### Solution For the next major frontend PR that's submitted, I should finish reviewing all the changes in [559](https://github.com/okTurtles/group-income-simple/pull/559) and add anything I find as part of my feedback
1.0
Double-check changes in PR 559 - ### Problem My review for PR #559 was rushed as we decided to move forward quickly, and there might be issues in there that I missed. ### Solution For the next major frontend PR that's submitted, I should finish reviewing all the changes in [559](https://github.com/okTurtles/group-income-simple/pull/559) and add anything I find as part of my feedback
process
double check changes in pr problem my review for pr was rushed as we decided to move forward quickly and there might be issues in there that i missed solution for the next major frontend pr that s submitted i should finish reviewing all the changes in and add anything i find as part of my feedback
1
16,382
21,104,743,622
IssuesEvent
2022-04-04 17:35:58
pycaret/pycaret
https://api.github.com/repos/pycaret/pycaret
closed
[preprocessing revamp] Add Boruta feature selection
enhancement preprocessing
**Is your feature request related to a problem? Please describe.** A clear and concise description of what the problem is. **Describe the solution you'd like** A clear and concise description of what you want to happen, ideally taking into consideration the existing toolbox design, classes and methods. **Describe alternatives you've considered** A clear and concise description of any alternative solutions or features you've considered. **Additional context** Add any other context or screenshots about the feature request here.
1.0
[preprocessing revamp] Add Boruta feature selection - **Is your feature request related to a problem? Please describe.** A clear and concise description of what the problem is. **Describe the solution you'd like** A clear and concise description of what you want to happen, ideally taking into consideration the existing toolbox design, classes and methods. **Describe alternatives you've considered** A clear and concise description of any alternative solutions or features you've considered. **Additional context** Add any other context or screenshots about the feature request here.
process
add boruta feature selection is your feature request related to a problem please describe a clear and concise description of what the problem is describe the solution you d like a clear and concise description of what you want to happen ideally taking into consideration the existing toolbox design classes and methods describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered additional context add any other context or screenshots about the feature request here
1
432,360
12,491,981,516
IssuesEvent
2020-06-01 05:57:49
LSFLK/request-management
https://api.github.com/repos/LSFLK/request-management
closed
Create issue for reports selected by language
Priority 1
The daily summary report - Total no. of requests by category. - [x] include tamil and english for subcategories - [x] include all three languages for category also
1.0
Create issue for reports selected by language - The daily summary report - Total no. of requests by category. - [x] include tamil and english for subcategories - [x] include all three languages for category also
non_process
create issue for reports selected by language the daily summary report total no of requests by category include tamil and english for subcategories include all three languages for category also
0
40,372
12,792,301,990
IssuesEvent
2020-07-02 01:01:25
LevyForchh/stat-engine
https://api.github.com/repos/LevyForchh/stat-engine
opened
CVE-2020-15084 (High) detected in express-jwt-5.3.1.tgz
security vulnerability
## CVE-2020-15084 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>express-jwt-5.3.1.tgz</b></p></summary> <p>JWT authentication middleware.</p> <p>Library home page: <a href="https://registry.npmjs.org/express-jwt/-/express-jwt-5.3.1.tgz">https://registry.npmjs.org/express-jwt/-/express-jwt-5.3.1.tgz</a></p> <p> Dependency Hierarchy: - :x: **express-jwt-5.3.1.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In express-jwt (NPM package) up and including version 5.3.3, the algorithms entry to be specified in the configuration is not being enforced. When algorithms is not specified in the configuration, with the combination of jwks-rsa, it may lead to authorization bypass. You are affected by this vulnerability if all of the following conditions apply: - You are using express-jwt - You do not have **algorithms** configured in your express-jwt configuration. - You are using libraries such as jwks-rsa as the **secret**. You can fix this by specifying **algorithms** in the express-jwt configuration. See linked GHSA for example. This is also fixed in version 6.0.0. <p>Publish Date: 2020-06-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15084>CVE-2020-15084</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/auth0/express-jwt/security/advisories/GHSA-6g6m-m6h5-w9gf">https://github.com/auth0/express-jwt/security/advisories/GHSA-6g6m-m6h5-w9gf</a></p> <p>Release Date: 2020-06-30</p> <p>Fix Resolution: 6.0.0</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END --> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"express-jwt","packageVersion":"5.3.1","isTransitiveDependency":false,"dependencyTree":"express-jwt:5.3.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"6.0.0"}],"vulnerabilityIdentifier":"CVE-2020-15084","vulnerabilityDetails":"In express-jwt (NPM package) up and including version 5.3.3, the algorithms entry to be specified in the configuration is not being enforced. When algorithms is not specified in the configuration, with the combination of jwks-rsa, it may lead to authorization bypass. You are affected by this vulnerability if all of the following conditions apply: - You are using express-jwt - You do not have **algorithms** configured in your express-jwt configuration. - You are using libraries such as jwks-rsa as the **secret**. You can fix this by specifying **algorithms** in the express-jwt configuration. See linked GHSA for example. This is also fixed in version 6.0.0.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15084","cvss3Severity":"high","cvss3Score":"7.7","cvss3Metrics":{"A":"N/A","AC":"N/A","PR":"N/A","S":"N/A","C":"N/A","UI":"N/A","AV":"N/A","I":"N/A"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-15084 (High) detected in express-jwt-5.3.1.tgz - ## CVE-2020-15084 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>express-jwt-5.3.1.tgz</b></p></summary> <p>JWT authentication middleware.</p> <p>Library home page: <a href="https://registry.npmjs.org/express-jwt/-/express-jwt-5.3.1.tgz">https://registry.npmjs.org/express-jwt/-/express-jwt-5.3.1.tgz</a></p> <p> Dependency Hierarchy: - :x: **express-jwt-5.3.1.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In express-jwt (NPM package) up and including version 5.3.3, the algorithms entry to be specified in the configuration is not being enforced. When algorithms is not specified in the configuration, with the combination of jwks-rsa, it may lead to authorization bypass. You are affected by this vulnerability if all of the following conditions apply: - You are using express-jwt - You do not have **algorithms** configured in your express-jwt configuration. - You are using libraries such as jwks-rsa as the **secret**. You can fix this by specifying **algorithms** in the express-jwt configuration. See linked GHSA for example. This is also fixed in version 6.0.0. <p>Publish Date: 2020-06-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15084>CVE-2020-15084</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/auth0/express-jwt/security/advisories/GHSA-6g6m-m6h5-w9gf">https://github.com/auth0/express-jwt/security/advisories/GHSA-6g6m-m6h5-w9gf</a></p> <p>Release Date: 2020-06-30</p> <p>Fix Resolution: 6.0.0</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END --> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"express-jwt","packageVersion":"5.3.1","isTransitiveDependency":false,"dependencyTree":"express-jwt:5.3.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"6.0.0"}],"vulnerabilityIdentifier":"CVE-2020-15084","vulnerabilityDetails":"In express-jwt (NPM package) up and including version 5.3.3, the algorithms entry to be specified in the configuration is not being enforced. When algorithms is not specified in the configuration, with the combination of jwks-rsa, it may lead to authorization bypass. You are affected by this vulnerability if all of the following conditions apply: - You are using express-jwt - You do not have **algorithms** configured in your express-jwt configuration. - You are using libraries such as jwks-rsa as the **secret**. You can fix this by specifying **algorithms** in the express-jwt configuration. See linked GHSA for example. This is also fixed in version 6.0.0.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15084","cvss3Severity":"high","cvss3Score":"7.7","cvss3Metrics":{"A":"N/A","AC":"N/A","PR":"N/A","S":"N/A","C":"N/A","UI":"N/A","AV":"N/A","I":"N/A"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in express jwt tgz cve high severity vulnerability vulnerable library express jwt tgz jwt authentication middleware library home page a href dependency hierarchy x express jwt tgz vulnerable library vulnerability details in express jwt npm package up and including version the algorithms entry to be specified in the configuration is not being enforced when algorithms is not specified in the configuration with the combination of jwks rsa it may lead to authorization bypass you are affected by this vulnerability if all of the following conditions apply you are using express jwt you do not have algorithms configured in your express jwt configuration you are using libraries such as jwks rsa as the secret you can fix this by specifying algorithms in the express jwt configuration see linked ghsa for example this is also fixed in version publish date url a href cvss score details base score metrics exploitability metrics attack vector n a attack complexity n a privileges required n a user interaction n a scope n a impact metrics confidentiality impact n a integrity impact n a availability impact n a for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution check this box to open an automated fix pr isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails in express jwt npm package up and including version the algorithms entry to be specified in the configuration is not being enforced when algorithms is not specified in the configuration with the combination of jwks rsa it may lead to authorization bypass you are affected by this vulnerability if all of the following conditions apply you are using express jwt you do not have algorithms configured in your express jwt configuration you are using libraries such as jwks rsa as the secret you can fix this by specifying algorithms in the express jwt configuration see linked ghsa for example this is also fixed in version vulnerabilityurl
0
160,975
12,528,087,784
IssuesEvent
2020-06-04 09:00:30
aliasrobotics/RVD
https://api.github.com/repos/aliasrobotics/RVD
opened
Use of possibly insecure function - consider using safer ast, ./src/rqt_pose_view/src/rqt_pose_view/pose_view_widget.py:100
bandit bug components software robot component: ROS static analysis testing triage version: noetic
```yaml { "id": 1, "title": "Use of possibly insecure function - consider using safer ast, ./src/rqt_pose_view/src/rqt_pose_view/pose_view_widget.py:100", "type": "bug", "description": "HIGH confidence of MEDIUM severity bug. Use of possibly insecure function - consider using safer ast.literal_eval. ./src/rqt_pose_view/src/rqt_pose_view/pose_view_widget.py:100. See links for more info on the bug.", "cwe": "None", "cve": "None", "keywords": [ "bandit", "bug", "static analysis", "testing", "triage", "bug", "version: noetic", "robot component: ROS", "components software" ], "system": "", "vendor": null, "severity": { "rvss-score": 0, "rvss-vector": "", "severity-description": "", "cvss-score": 0, "cvss-vector": "" }, "links": "", "flaw": { "phase": "testing", "specificity": "subject-specific", "architectural-location": "application-specific", "application": "N/A", "subsystem": "N/A", "package": "N/A", "languages": "None", "date-detected": "2020-06-04 (09:00)", "detected-by": "Alias Robotics", "detected-by-method": "testing static", "date-reported": "2020-06-04 (09:00)", "reported-by": "Alias Robotics", "reported-by-relationship": "automatic", "issue": "", "reproducibility": "always", "trace": "./src/rqt_pose_view/src/rqt_pose_view/pose_view_widget.py:100", "reproduction": "See artifacts below (if available)", "reproduction-image": "" }, "exploitation": { "description": "", "exploitation-image": "", "exploitation-vector": "" }, "mitigation": { "description": "", "pull-request": "", "date-mitigation": "" } } ```
1.0
Use of possibly insecure function - consider using safer ast, ./src/rqt_pose_view/src/rqt_pose_view/pose_view_widget.py:100 - ```yaml { "id": 1, "title": "Use of possibly insecure function - consider using safer ast, ./src/rqt_pose_view/src/rqt_pose_view/pose_view_widget.py:100", "type": "bug", "description": "HIGH confidence of MEDIUM severity bug. Use of possibly insecure function - consider using safer ast.literal_eval. ./src/rqt_pose_view/src/rqt_pose_view/pose_view_widget.py:100. See links for more info on the bug.", "cwe": "None", "cve": "None", "keywords": [ "bandit", "bug", "static analysis", "testing", "triage", "bug", "version: noetic", "robot component: ROS", "components software" ], "system": "", "vendor": null, "severity": { "rvss-score": 0, "rvss-vector": "", "severity-description": "", "cvss-score": 0, "cvss-vector": "" }, "links": "", "flaw": { "phase": "testing", "specificity": "subject-specific", "architectural-location": "application-specific", "application": "N/A", "subsystem": "N/A", "package": "N/A", "languages": "None", "date-detected": "2020-06-04 (09:00)", "detected-by": "Alias Robotics", "detected-by-method": "testing static", "date-reported": "2020-06-04 (09:00)", "reported-by": "Alias Robotics", "reported-by-relationship": "automatic", "issue": "", "reproducibility": "always", "trace": "./src/rqt_pose_view/src/rqt_pose_view/pose_view_widget.py:100", "reproduction": "See artifacts below (if available)", "reproduction-image": "" }, "exploitation": { "description": "", "exploitation-image": "", "exploitation-vector": "" }, "mitigation": { "description": "", "pull-request": "", "date-mitigation": "" } } ```
non_process
use of possibly insecure function consider using safer ast src rqt pose view src rqt pose view pose view widget py yaml id title use of possibly insecure function consider using safer ast src rqt pose view src rqt pose view pose view widget py type bug description high confidence of medium severity bug use of possibly insecure function consider using safer ast literal eval src rqt pose view src rqt pose view pose view widget py see links for more info on the bug cwe none cve none keywords bandit bug static analysis testing triage bug version noetic robot component ros components software system vendor null severity rvss score rvss vector severity description cvss score cvss vector links flaw phase testing specificity subject specific architectural location application specific application n a subsystem n a package n a languages none date detected detected by alias robotics detected by method testing static date reported reported by alias robotics reported by relationship automatic issue reproducibility always trace src rqt pose view src rqt pose view pose view widget py reproduction see artifacts below if available reproduction image exploitation description exploitation image exploitation vector mitigation description pull request date mitigation
0
19,071
25,099,055,854
IssuesEvent
2022-11-08 12:20:51
anitsh/til
https://api.github.com/repos/anitsh/til
opened
Vivid Vision
process thinking
Vivid Vision helps in planning long-term strategy, coming up with a plan, and communicating that plan to the team. Vivid Vision is a method that helps companies bring their ideas together to generate a clear plan for what their future looks like. It’s a detailed overview of what the business will look like, feel like, and act like three years down the road. Oftentimes, leaders have ideas in their mind, but it’s hard to put it into words. Most business “visions” are just words or vague ideas; they’re not specific goals that the company is working towards. Vivid Vision allows those leaders to craft a story around the vision. This makes it easier for everyone to understand and work towards. - https://drive.google.com/file/d/1_ckwLlWM4Cvt-1j4_pfffLePLdBhj-Yo/view Download Vivid Vision Mindmap - https://www.youtube.com/watch?v=6w-Hsgonrcc - How use above document - https://thewoweffect.com/PDF/CreativeWorksVividVision.pdf Example
1.0
Vivid Vision - Vivid Vision helps in planning long-term strategy, coming up with a plan, and communicating that plan to the team. Vivid Vision is a method that helps companies bring their ideas together to generate a clear plan for what their future looks like. It’s a detailed overview of what the business will look like, feel like, and act like three years down the road. Oftentimes, leaders have ideas in their mind, but it’s hard to put it into words. Most business “visions” are just words or vague ideas; they’re not specific goals that the company is working towards. Vivid Vision allows those leaders to craft a story around the vision. This makes it easier for everyone to understand and work towards. - https://drive.google.com/file/d/1_ckwLlWM4Cvt-1j4_pfffLePLdBhj-Yo/view Download Vivid Vision Mindmap - https://www.youtube.com/watch?v=6w-Hsgonrcc - How use above document - https://thewoweffect.com/PDF/CreativeWorksVividVision.pdf Example
process
vivid vision vivid vision helps in planning long term strategy coming up with a plan and communicating that plan to the team vivid vision is a method that helps companies bring their ideas together to generate a clear plan for what their future looks like it’s a detailed overview of what the business will look like feel like and act like three years down the road oftentimes leaders have ideas in their mind but it’s hard to put it into words most business “visions” are just words or vague ideas they’re not specific goals that the company is working towards vivid vision allows those leaders to craft a story around the vision this makes it easier for everyone to understand and work towards download vivid vision mindmap how use above document example
1